Compare commits
2 Commits
55069d6845
...
ci/add-che
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
5d87d386c3 | ||
|
|
f65e3887b2 |
@@ -12,5 +12,15 @@ jobs:
|
|||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@34e114876b0b11c390a56381ad16ebd13914f8d5 # v4
|
- uses: actions/checkout@34e114876b0b11c390a56381ad16ebd13914f8d5 # v4
|
||||||
|
|
||||||
- name: Build (runs make check via Dockerfile)
|
- uses: actions/setup-go@40f1582b2485089dde7abd97c1529aa768e1baff # v5
|
||||||
run: docker build .
|
with:
|
||||||
|
go-version-file: go.mod
|
||||||
|
|
||||||
|
- name: Install golangci-lint
|
||||||
|
run: go install github.com/golangci/golangci-lint/cmd/golangci-lint@latest
|
||||||
|
|
||||||
|
- name: Install goimports
|
||||||
|
run: go install golang.org/x/tools/cmd/goimports@latest
|
||||||
|
|
||||||
|
- name: Run make check
|
||||||
|
run: make check
|
||||||
|
|||||||
@@ -1,11 +1,11 @@
|
|||||||
# Build stage
|
# Build stage
|
||||||
FROM golang@sha256:f6751d823c26342f9506c03797d2527668d095b0a15f1862cddb4d927a7a4ced AS builder # golang:1.25-alpine
|
FROM golang:1.25-alpine AS builder
|
||||||
|
|
||||||
RUN apk add --no-cache git make gcc musl-dev
|
RUN apk add --no-cache git make gcc musl-dev
|
||||||
|
|
||||||
# Install golangci-lint v2
|
# Install golangci-lint v2
|
||||||
RUN go install github.com/golangci/golangci-lint/v2/cmd/golangci-lint@5d1e709b7be35cb2025444e19de266b056b7b7ee # v2.10.1
|
RUN go install github.com/golangci/golangci-lint/v2/cmd/golangci-lint@latest
|
||||||
RUN go install golang.org/x/tools/cmd/goimports@009367f5c17a8d4c45a961a3a509277190a9a6f0 # v0.42.0
|
RUN go install golang.org/x/tools/cmd/goimports@latest
|
||||||
|
|
||||||
WORKDIR /src
|
WORKDIR /src
|
||||||
COPY go.mod go.sum ./
|
COPY go.mod go.sum ./
|
||||||
@@ -20,7 +20,7 @@ RUN make check
|
|||||||
RUN make build
|
RUN make build
|
||||||
|
|
||||||
# Runtime stage
|
# Runtime stage
|
||||||
FROM alpine@sha256:6baf43584bcb78f2e5847d1de515f23499913ac9f12bdf834811a3145eb11ca1 # alpine:3.19
|
FROM alpine:3.19
|
||||||
|
|
||||||
RUN apk add --no-cache ca-certificates tzdata git openssh-client docker-cli
|
RUN apk add --no-cache ca-certificates tzdata git openssh-client docker-cli
|
||||||
|
|
||||||
|
|||||||
@@ -51,7 +51,7 @@ type Config struct {
|
|||||||
MaintenanceMode bool
|
MaintenanceMode bool
|
||||||
MetricsUsername string
|
MetricsUsername string
|
||||||
MetricsPassword string
|
MetricsPassword string
|
||||||
SessionSecret string `json:"-"`
|
SessionSecret string
|
||||||
CORSOrigins string
|
CORSOrigins string
|
||||||
params *Params
|
params *Params
|
||||||
log *slog.Logger
|
log *slog.Logger
|
||||||
|
|||||||
@@ -1,41 +0,0 @@
|
|||||||
package database
|
|
||||||
|
|
||||||
import (
|
|
||||||
"log/slog"
|
|
||||||
"os"
|
|
||||||
"testing"
|
|
||||||
|
|
||||||
"git.eeqj.de/sneak/upaas/internal/config"
|
|
||||||
"git.eeqj.de/sneak/upaas/internal/logger"
|
|
||||||
)
|
|
||||||
|
|
||||||
// NewTestDatabase creates an in-memory Database for testing.
|
|
||||||
// It runs migrations so all tables are available.
|
|
||||||
func NewTestDatabase(t *testing.T) *Database {
|
|
||||||
t.Helper()
|
|
||||||
|
|
||||||
tmpDir := t.TempDir()
|
|
||||||
|
|
||||||
cfg := &config.Config{
|
|
||||||
DataDir: tmpDir,
|
|
||||||
}
|
|
||||||
|
|
||||||
log := slog.New(slog.NewTextHandler(os.Stderr, nil))
|
|
||||||
logWrapper := logger.NewForTest(log)
|
|
||||||
|
|
||||||
db, err := New(nil, Params{
|
|
||||||
Logger: logWrapper,
|
|
||||||
Config: cfg,
|
|
||||||
})
|
|
||||||
if err != nil {
|
|
||||||
t.Fatalf("failed to create test database: %v", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
t.Cleanup(func() {
|
|
||||||
if db.database != nil {
|
|
||||||
_ = db.database.Close()
|
|
||||||
}
|
|
||||||
})
|
|
||||||
|
|
||||||
return db
|
|
||||||
}
|
|
||||||
@@ -480,20 +480,6 @@ func (c *Client) CloneRepo(
|
|||||||
return c.performClone(ctx, cfg)
|
return c.performClone(ctx, cfg)
|
||||||
}
|
}
|
||||||
|
|
||||||
// RemoveImage removes a Docker image by ID or tag.
|
|
||||||
// It returns nil if the image was successfully removed or does not exist.
|
|
||||||
func (c *Client) RemoveImage(ctx context.Context, imageID string) error {
|
|
||||||
_, err := c.docker.ImageRemove(ctx, imageID, image.RemoveOptions{
|
|
||||||
Force: true,
|
|
||||||
PruneChildren: true,
|
|
||||||
})
|
|
||||||
if err != nil && !client.IsErrNotFound(err) {
|
|
||||||
return fmt.Errorf("failed to remove image %s: %w", imageID, err)
|
|
||||||
}
|
|
||||||
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func (c *Client) performBuild(
|
func (c *Client) performBuild(
|
||||||
ctx context.Context,
|
ctx context.Context,
|
||||||
opts BuildImageOptions,
|
opts BuildImageOptions,
|
||||||
@@ -754,6 +740,20 @@ func (c *Client) connect(ctx context.Context) error {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// RemoveImage removes a Docker image by ID or tag.
|
||||||
|
// It returns nil if the image was successfully removed or does not exist.
|
||||||
|
func (c *Client) RemoveImage(ctx context.Context, imageID string) error {
|
||||||
|
_, err := c.docker.ImageRemove(ctx, imageID, image.RemoveOptions{
|
||||||
|
Force: true,
|
||||||
|
PruneChildren: true,
|
||||||
|
})
|
||||||
|
if err != nil && !client.IsErrNotFound(err) {
|
||||||
|
return fmt.Errorf("failed to remove image %s: %w", imageID, err)
|
||||||
|
}
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
func (c *Client) close() error {
|
func (c *Client) close() error {
|
||||||
if c.docker != nil {
|
if c.docker != nil {
|
||||||
err := c.docker.Close()
|
err := c.docker.Close()
|
||||||
|
|||||||
@@ -8,6 +8,7 @@ import (
|
|||||||
"github.com/go-chi/chi/v5"
|
"github.com/go-chi/chi/v5"
|
||||||
|
|
||||||
"git.eeqj.de/sneak/upaas/internal/models"
|
"git.eeqj.de/sneak/upaas/internal/models"
|
||||||
|
"git.eeqj.de/sneak/upaas/internal/service/app"
|
||||||
)
|
)
|
||||||
|
|
||||||
// apiAppResponse is the JSON representation of an app.
|
// apiAppResponse is the JSON representation of an app.
|
||||||
@@ -73,13 +74,18 @@ func deploymentToAPI(d *models.Deployment) apiDeploymentResponse {
|
|||||||
// HandleAPILoginPOST returns a handler that authenticates via JSON credentials
|
// HandleAPILoginPOST returns a handler that authenticates via JSON credentials
|
||||||
// and sets a session cookie.
|
// and sets a session cookie.
|
||||||
func (h *Handlers) HandleAPILoginPOST() http.HandlerFunc {
|
func (h *Handlers) HandleAPILoginPOST() http.HandlerFunc {
|
||||||
|
type loginRequest struct {
|
||||||
|
Username string `json:"username"`
|
||||||
|
Password string `json:"password"`
|
||||||
|
}
|
||||||
|
|
||||||
type loginResponse struct {
|
type loginResponse struct {
|
||||||
UserID int64 `json:"userId"`
|
UserID int64 `json:"userId"`
|
||||||
Username string `json:"username"`
|
Username string `json:"username"`
|
||||||
}
|
}
|
||||||
|
|
||||||
return func(writer http.ResponseWriter, request *http.Request) {
|
return func(writer http.ResponseWriter, request *http.Request) {
|
||||||
var req map[string]string
|
var req loginRequest
|
||||||
|
|
||||||
decodeErr := json.NewDecoder(request.Body).Decode(&req)
|
decodeErr := json.NewDecoder(request.Body).Decode(&req)
|
||||||
if decodeErr != nil {
|
if decodeErr != nil {
|
||||||
@@ -90,10 +96,7 @@ func (h *Handlers) HandleAPILoginPOST() http.HandlerFunc {
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
username := req["username"]
|
if req.Username == "" || req.Password == "" {
|
||||||
credential := req["password"]
|
|
||||||
|
|
||||||
if username == "" || credential == "" {
|
|
||||||
h.respondJSON(writer, request,
|
h.respondJSON(writer, request,
|
||||||
map[string]string{"error": "username and password are required"},
|
map[string]string{"error": "username and password are required"},
|
||||||
http.StatusBadRequest)
|
http.StatusBadRequest)
|
||||||
@@ -101,7 +104,7 @@ func (h *Handlers) HandleAPILoginPOST() http.HandlerFunc {
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
user, authErr := h.auth.Authenticate(request.Context(), username, credential)
|
user, authErr := h.auth.Authenticate(request.Context(), req.Username, req.Password)
|
||||||
if authErr != nil {
|
if authErr != nil {
|
||||||
h.respondJSON(writer, request,
|
h.respondJSON(writer, request,
|
||||||
map[string]string{"error": "invalid credentials"},
|
map[string]string{"error": "invalid credentials"},
|
||||||
@@ -174,6 +177,106 @@ func (h *Handlers) HandleAPIGetApp() http.HandlerFunc {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// HandleAPICreateApp returns a handler that creates a new app.
|
||||||
|
func (h *Handlers) HandleAPICreateApp() http.HandlerFunc {
|
||||||
|
type createRequest struct {
|
||||||
|
Name string `json:"name"`
|
||||||
|
RepoURL string `json:"repoUrl"`
|
||||||
|
Branch string `json:"branch"`
|
||||||
|
DockerfilePath string `json:"dockerfilePath"`
|
||||||
|
DockerNetwork string `json:"dockerNetwork"`
|
||||||
|
NtfyTopic string `json:"ntfyTopic"`
|
||||||
|
SlackWebhook string `json:"slackWebhook"`
|
||||||
|
}
|
||||||
|
|
||||||
|
return func(writer http.ResponseWriter, request *http.Request) {
|
||||||
|
var req createRequest
|
||||||
|
|
||||||
|
decodeErr := json.NewDecoder(request.Body).Decode(&req)
|
||||||
|
if decodeErr != nil {
|
||||||
|
h.respondJSON(writer, request,
|
||||||
|
map[string]string{"error": "invalid JSON body"},
|
||||||
|
http.StatusBadRequest)
|
||||||
|
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
if req.Name == "" || req.RepoURL == "" {
|
||||||
|
h.respondJSON(writer, request,
|
||||||
|
map[string]string{"error": "name and repo_url are required"},
|
||||||
|
http.StatusBadRequest)
|
||||||
|
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
nameErr := validateAppName(req.Name)
|
||||||
|
if nameErr != nil {
|
||||||
|
h.respondJSON(writer, request,
|
||||||
|
map[string]string{"error": "invalid app name: " + nameErr.Error()},
|
||||||
|
http.StatusBadRequest)
|
||||||
|
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
createdApp, createErr := h.appService.CreateApp(request.Context(), app.CreateAppInput{
|
||||||
|
Name: req.Name,
|
||||||
|
RepoURL: req.RepoURL,
|
||||||
|
Branch: req.Branch,
|
||||||
|
DockerfilePath: req.DockerfilePath,
|
||||||
|
DockerNetwork: req.DockerNetwork,
|
||||||
|
NtfyTopic: req.NtfyTopic,
|
||||||
|
SlackWebhook: req.SlackWebhook,
|
||||||
|
})
|
||||||
|
if createErr != nil {
|
||||||
|
h.log.Error("api: failed to create app", "error", createErr)
|
||||||
|
h.respondJSON(writer, request,
|
||||||
|
map[string]string{"error": "failed to create app"},
|
||||||
|
http.StatusInternalServerError)
|
||||||
|
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
h.respondJSON(writer, request, appToAPI(createdApp), http.StatusCreated)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// HandleAPIDeleteApp returns a handler that deletes an app.
|
||||||
|
func (h *Handlers) HandleAPIDeleteApp() http.HandlerFunc {
|
||||||
|
return func(writer http.ResponseWriter, request *http.Request) {
|
||||||
|
appID := chi.URLParam(request, "id")
|
||||||
|
|
||||||
|
application, err := h.appService.GetApp(request.Context(), appID)
|
||||||
|
if err != nil {
|
||||||
|
h.respondJSON(writer, request,
|
||||||
|
map[string]string{"error": "internal server error"},
|
||||||
|
http.StatusInternalServerError)
|
||||||
|
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
if application == nil {
|
||||||
|
h.respondJSON(writer, request,
|
||||||
|
map[string]string{"error": "app not found"},
|
||||||
|
http.StatusNotFound)
|
||||||
|
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
deleteErr := h.appService.DeleteApp(request.Context(), application)
|
||||||
|
if deleteErr != nil {
|
||||||
|
h.log.Error("api: failed to delete app", "error", deleteErr)
|
||||||
|
h.respondJSON(writer, request,
|
||||||
|
map[string]string{"error": "failed to delete app"},
|
||||||
|
http.StatusInternalServerError)
|
||||||
|
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
h.respondJSON(writer, request,
|
||||||
|
map[string]string{"status": "deleted"}, http.StatusOK)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
// deploymentsPageLimit is the default number of deployments per page.
|
// deploymentsPageLimit is the default number of deployments per page.
|
||||||
const deploymentsPageLimit = 20
|
const deploymentsPageLimit = 20
|
||||||
|
|
||||||
@@ -220,6 +323,35 @@ func (h *Handlers) HandleAPIListDeployments() http.HandlerFunc {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// HandleAPITriggerDeploy returns a handler that triggers a deployment for an app.
|
||||||
|
func (h *Handlers) HandleAPITriggerDeploy() http.HandlerFunc {
|
||||||
|
return func(writer http.ResponseWriter, request *http.Request) {
|
||||||
|
appID := chi.URLParam(request, "id")
|
||||||
|
|
||||||
|
application, err := h.appService.GetApp(request.Context(), appID)
|
||||||
|
if err != nil || application == nil {
|
||||||
|
h.respondJSON(writer, request,
|
||||||
|
map[string]string{"error": "app not found"},
|
||||||
|
http.StatusNotFound)
|
||||||
|
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
deployErr := h.deploy.Deploy(request.Context(), application, nil, true)
|
||||||
|
if deployErr != nil {
|
||||||
|
h.log.Error("api: failed to trigger deploy", "error", deployErr)
|
||||||
|
h.respondJSON(writer, request,
|
||||||
|
map[string]string{"error": deployErr.Error()},
|
||||||
|
http.StatusConflict)
|
||||||
|
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
h.respondJSON(writer, request,
|
||||||
|
map[string]string{"status": "deploying"}, http.StatusAccepted)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
// HandleAPIWhoAmI returns a handler that shows the current authenticated user.
|
// HandleAPIWhoAmI returns a handler that shows the current authenticated user.
|
||||||
func (h *Handlers) HandleAPIWhoAmI() http.HandlerFunc {
|
func (h *Handlers) HandleAPIWhoAmI() http.HandlerFunc {
|
||||||
type whoAmIResponse struct {
|
type whoAmIResponse struct {
|
||||||
|
|||||||
@@ -10,8 +10,6 @@ import (
|
|||||||
"github.com/go-chi/chi/v5"
|
"github.com/go-chi/chi/v5"
|
||||||
"github.com/stretchr/testify/assert"
|
"github.com/stretchr/testify/assert"
|
||||||
"github.com/stretchr/testify/require"
|
"github.com/stretchr/testify/require"
|
||||||
|
|
||||||
"git.eeqj.de/sneak/upaas/internal/service/app"
|
|
||||||
)
|
)
|
||||||
|
|
||||||
// apiRouter builds a chi router with the API routes using session auth middleware.
|
// apiRouter builds a chi router with the API routes using session auth middleware.
|
||||||
@@ -25,7 +23,10 @@ func apiRouter(tc *testContext) http.Handler {
|
|||||||
apiR.Use(tc.middleware.APISessionAuth())
|
apiR.Use(tc.middleware.APISessionAuth())
|
||||||
apiR.Get("/whoami", tc.handlers.HandleAPIWhoAmI())
|
apiR.Get("/whoami", tc.handlers.HandleAPIWhoAmI())
|
||||||
apiR.Get("/apps", tc.handlers.HandleAPIListApps())
|
apiR.Get("/apps", tc.handlers.HandleAPIListApps())
|
||||||
|
apiR.Post("/apps", tc.handlers.HandleAPICreateApp())
|
||||||
apiR.Get("/apps/{id}", tc.handlers.HandleAPIGetApp())
|
apiR.Get("/apps/{id}", tc.handlers.HandleAPIGetApp())
|
||||||
|
apiR.Delete("/apps/{id}", tc.handlers.HandleAPIDeleteApp())
|
||||||
|
apiR.Post("/apps/{id}/deploy", tc.handlers.HandleAPITriggerDeploy())
|
||||||
apiR.Get("/apps/{id}/deployments", tc.handlers.HandleAPIListDeployments())
|
apiR.Get("/apps/{id}/deployments", tc.handlers.HandleAPIListDeployments())
|
||||||
})
|
})
|
||||||
})
|
})
|
||||||
@@ -61,16 +62,23 @@ func setupAPITest(t *testing.T) (*testContext, []*http.Cookie) {
|
|||||||
return tc, cookies
|
return tc, cookies
|
||||||
}
|
}
|
||||||
|
|
||||||
// apiGet makes an authenticated GET request using session cookies.
|
// apiRequest makes an authenticated API request using session cookies.
|
||||||
func apiGet(
|
func apiRequest(
|
||||||
t *testing.T,
|
t *testing.T,
|
||||||
tc *testContext,
|
tc *testContext,
|
||||||
cookies []*http.Cookie,
|
cookies []*http.Cookie,
|
||||||
path string,
|
method, path string,
|
||||||
|
body string,
|
||||||
) *httptest.ResponseRecorder {
|
) *httptest.ResponseRecorder {
|
||||||
t.Helper()
|
t.Helper()
|
||||||
|
|
||||||
req := httptest.NewRequest(http.MethodGet, path, nil)
|
var req *http.Request
|
||||||
|
if body != "" {
|
||||||
|
req = httptest.NewRequest(method, path, strings.NewReader(body))
|
||||||
|
req.Header.Set("Content-Type", "application/json")
|
||||||
|
} else {
|
||||||
|
req = httptest.NewRequest(method, path, nil)
|
||||||
|
}
|
||||||
|
|
||||||
for _, c := range cookies {
|
for _, c := range cookies {
|
||||||
req.AddCookie(c)
|
req.AddCookie(c)
|
||||||
@@ -167,7 +175,7 @@ func TestAPIWhoAmI(t *testing.T) {
|
|||||||
|
|
||||||
tc, cookies := setupAPITest(t)
|
tc, cookies := setupAPITest(t)
|
||||||
|
|
||||||
rr := apiGet(t, tc, cookies, "/api/v1/whoami")
|
rr := apiRequest(t, tc, cookies, http.MethodGet, "/api/v1/whoami", "")
|
||||||
assert.Equal(t, http.StatusOK, rr.Code)
|
assert.Equal(t, http.StatusOK, rr.Code)
|
||||||
|
|
||||||
var resp map[string]any
|
var resp map[string]any
|
||||||
@@ -180,7 +188,7 @@ func TestAPIListAppsEmpty(t *testing.T) {
|
|||||||
|
|
||||||
tc, cookies := setupAPITest(t)
|
tc, cookies := setupAPITest(t)
|
||||||
|
|
||||||
rr := apiGet(t, tc, cookies, "/api/v1/apps")
|
rr := apiRequest(t, tc, cookies, http.MethodGet, "/api/v1/apps", "")
|
||||||
assert.Equal(t, http.StatusOK, rr.Code)
|
assert.Equal(t, http.StatusOK, rr.Code)
|
||||||
|
|
||||||
var apps []any
|
var apps []any
|
||||||
@@ -188,23 +196,52 @@ func TestAPIListAppsEmpty(t *testing.T) {
|
|||||||
assert.Empty(t, apps)
|
assert.Empty(t, apps)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func TestAPICreateApp(t *testing.T) {
|
||||||
|
t.Parallel()
|
||||||
|
|
||||||
|
tc, cookies := setupAPITest(t)
|
||||||
|
|
||||||
|
body := `{"name":"test-app","repoUrl":"https://github.com/example/repo"}`
|
||||||
|
rr := apiRequest(t, tc, cookies, http.MethodPost, "/api/v1/apps", body)
|
||||||
|
assert.Equal(t, http.StatusCreated, rr.Code)
|
||||||
|
|
||||||
|
var app map[string]any
|
||||||
|
require.NoError(t, json.Unmarshal(rr.Body.Bytes(), &app))
|
||||||
|
assert.Equal(t, "test-app", app["name"])
|
||||||
|
assert.Equal(t, "pending", app["status"])
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestAPICreateAppValidation(t *testing.T) {
|
||||||
|
t.Parallel()
|
||||||
|
|
||||||
|
tc, cookies := setupAPITest(t)
|
||||||
|
|
||||||
|
body := `{"name":"","repoUrl":""}`
|
||||||
|
rr := apiRequest(t, tc, cookies, http.MethodPost, "/api/v1/apps", body)
|
||||||
|
assert.Equal(t, http.StatusBadRequest, rr.Code)
|
||||||
|
}
|
||||||
|
|
||||||
func TestAPIGetApp(t *testing.T) {
|
func TestAPIGetApp(t *testing.T) {
|
||||||
t.Parallel()
|
t.Parallel()
|
||||||
|
|
||||||
tc, cookies := setupAPITest(t)
|
tc, cookies := setupAPITest(t)
|
||||||
|
|
||||||
created, err := tc.appSvc.CreateApp(t.Context(), app.CreateAppInput{
|
body := `{"name":"my-app","repoUrl":"https://github.com/example/repo"}`
|
||||||
Name: "my-app",
|
rr := apiRequest(t, tc, cookies, http.MethodPost, "/api/v1/apps", body)
|
||||||
RepoURL: "https://github.com/example/repo",
|
require.Equal(t, http.StatusCreated, rr.Code)
|
||||||
})
|
|
||||||
require.NoError(t, err)
|
|
||||||
|
|
||||||
rr := apiGet(t, tc, cookies, "/api/v1/apps/"+created.ID)
|
var created map[string]any
|
||||||
|
require.NoError(t, json.Unmarshal(rr.Body.Bytes(), &created))
|
||||||
|
|
||||||
|
appID, ok := created["id"].(string)
|
||||||
|
require.True(t, ok)
|
||||||
|
|
||||||
|
rr = apiRequest(t, tc, cookies, http.MethodGet, "/api/v1/apps/"+appID, "")
|
||||||
assert.Equal(t, http.StatusOK, rr.Code)
|
assert.Equal(t, http.StatusOK, rr.Code)
|
||||||
|
|
||||||
var resp map[string]any
|
var app map[string]any
|
||||||
require.NoError(t, json.Unmarshal(rr.Body.Bytes(), &resp))
|
require.NoError(t, json.Unmarshal(rr.Body.Bytes(), &app))
|
||||||
assert.Equal(t, "my-app", resp["name"])
|
assert.Equal(t, "my-app", app["name"])
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestAPIGetAppNotFound(t *testing.T) {
|
func TestAPIGetAppNotFound(t *testing.T) {
|
||||||
@@ -212,7 +249,29 @@ func TestAPIGetAppNotFound(t *testing.T) {
|
|||||||
|
|
||||||
tc, cookies := setupAPITest(t)
|
tc, cookies := setupAPITest(t)
|
||||||
|
|
||||||
rr := apiGet(t, tc, cookies, "/api/v1/apps/nonexistent")
|
rr := apiRequest(t, tc, cookies, http.MethodGet, "/api/v1/apps/nonexistent", "")
|
||||||
|
assert.Equal(t, http.StatusNotFound, rr.Code)
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestAPIDeleteApp(t *testing.T) {
|
||||||
|
t.Parallel()
|
||||||
|
|
||||||
|
tc, cookies := setupAPITest(t)
|
||||||
|
|
||||||
|
body := `{"name":"delete-me","repoUrl":"https://github.com/example/repo"}`
|
||||||
|
rr := apiRequest(t, tc, cookies, http.MethodPost, "/api/v1/apps", body)
|
||||||
|
require.Equal(t, http.StatusCreated, rr.Code)
|
||||||
|
|
||||||
|
var created map[string]any
|
||||||
|
require.NoError(t, json.Unmarshal(rr.Body.Bytes(), &created))
|
||||||
|
|
||||||
|
appID, ok := created["id"].(string)
|
||||||
|
require.True(t, ok)
|
||||||
|
|
||||||
|
rr = apiRequest(t, tc, cookies, http.MethodDelete, "/api/v1/apps/"+appID, "")
|
||||||
|
assert.Equal(t, http.StatusOK, rr.Code)
|
||||||
|
|
||||||
|
rr = apiRequest(t, tc, cookies, http.MethodGet, "/api/v1/apps/"+appID, "")
|
||||||
assert.Equal(t, http.StatusNotFound, rr.Code)
|
assert.Equal(t, http.StatusNotFound, rr.Code)
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -221,13 +280,17 @@ func TestAPIListDeployments(t *testing.T) {
|
|||||||
|
|
||||||
tc, cookies := setupAPITest(t)
|
tc, cookies := setupAPITest(t)
|
||||||
|
|
||||||
created, err := tc.appSvc.CreateApp(t.Context(), app.CreateAppInput{
|
body := `{"name":"deploy-app","repoUrl":"https://github.com/example/repo"}`
|
||||||
Name: "deploy-app",
|
rr := apiRequest(t, tc, cookies, http.MethodPost, "/api/v1/apps", body)
|
||||||
RepoURL: "https://github.com/example/repo",
|
require.Equal(t, http.StatusCreated, rr.Code)
|
||||||
})
|
|
||||||
require.NoError(t, err)
|
|
||||||
|
|
||||||
rr := apiGet(t, tc, cookies, "/api/v1/apps/"+created.ID+"/deployments")
|
var created map[string]any
|
||||||
|
require.NoError(t, json.Unmarshal(rr.Body.Bytes(), &created))
|
||||||
|
|
||||||
|
appID, ok := created["id"].(string)
|
||||||
|
require.True(t, ok)
|
||||||
|
|
||||||
|
rr = apiRequest(t, tc, cookies, http.MethodGet, "/api/v1/apps/"+appID+"/deployments", "")
|
||||||
assert.Equal(t, http.StatusOK, rr.Code)
|
assert.Equal(t, http.StatusOK, rr.Code)
|
||||||
|
|
||||||
var deployments []any
|
var deployments []any
|
||||||
|
|||||||
@@ -77,14 +77,6 @@ func (h *Handlers) HandleAppCreate() http.HandlerFunc { //nolint:funlen // valid
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
repoURLErr := validateRepoURL(repoURL)
|
|
||||||
if repoURLErr != nil {
|
|
||||||
data["Error"] = "Invalid repository URL: " + repoURLErr.Error()
|
|
||||||
h.renderTemplate(writer, tmpl, "app_new.html", data)
|
|
||||||
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
if branch == "" {
|
if branch == "" {
|
||||||
branch = "main"
|
branch = "main"
|
||||||
}
|
}
|
||||||
@@ -233,17 +225,6 @@ func (h *Handlers) HandleAppUpdate() http.HandlerFunc { //nolint:funlen // valid
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
repoURLErr := validateRepoURL(request.FormValue("repo_url"))
|
|
||||||
if repoURLErr != nil {
|
|
||||||
data := h.addGlobals(map[string]any{
|
|
||||||
"App": application,
|
|
||||||
"Error": "Invalid repository URL: " + repoURLErr.Error(),
|
|
||||||
}, request)
|
|
||||||
_ = tmpl.ExecuteTemplate(writer, "app_edit.html", data)
|
|
||||||
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
application.Name = newName
|
application.Name = newName
|
||||||
application.RepoURL = request.FormValue("repo_url")
|
application.RepoURL = request.FormValue("repo_url")
|
||||||
application.Branch = request.FormValue("branch")
|
application.Branch = request.FormValue("branch")
|
||||||
@@ -518,7 +499,7 @@ func (h *Handlers) HandleAppLogs() http.HandlerFunc {
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
_, _ = writer.Write([]byte(SanitizeLogs(logs))) // #nosec G705 -- logs sanitized, Content-Type is text/plain
|
_, _ = writer.Write([]byte(logs))
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -553,7 +534,7 @@ func (h *Handlers) HandleDeploymentLogsAPI() http.HandlerFunc {
|
|||||||
|
|
||||||
logs := ""
|
logs := ""
|
||||||
if deployment.Logs.Valid {
|
if deployment.Logs.Valid {
|
||||||
logs = SanitizeLogs(deployment.Logs.String)
|
logs = deployment.Logs.String
|
||||||
}
|
}
|
||||||
|
|
||||||
response := map[string]any{
|
response := map[string]any{
|
||||||
@@ -600,8 +581,8 @@ func (h *Handlers) HandleDeploymentLogDownload() http.HandlerFunc {
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
// Check if file exists — logPath is constructed internally, not from user input
|
// Check if file exists
|
||||||
_, err := os.Stat(logPath) // #nosec G703 -- path from internal GetLogFilePath, not user input
|
_, err := os.Stat(logPath)
|
||||||
if os.IsNotExist(err) {
|
if os.IsNotExist(err) {
|
||||||
http.NotFound(writer, request)
|
http.NotFound(writer, request)
|
||||||
|
|
||||||
@@ -680,7 +661,7 @@ func (h *Handlers) HandleContainerLogsAPI() http.HandlerFunc {
|
|||||||
}
|
}
|
||||||
|
|
||||||
response := map[string]any{
|
response := map[string]any{
|
||||||
"logs": SanitizeLogs(logs),
|
"logs": logs,
|
||||||
"status": status,
|
"status": status,
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -916,7 +897,7 @@ func (h *Handlers) HandleEnvVarAdd() http.HandlerFunc {
|
|||||||
func (h *Handlers) HandleEnvVarDelete() http.HandlerFunc {
|
func (h *Handlers) HandleEnvVarDelete() http.HandlerFunc {
|
||||||
return func(writer http.ResponseWriter, request *http.Request) {
|
return func(writer http.ResponseWriter, request *http.Request) {
|
||||||
appID := chi.URLParam(request, "id")
|
appID := chi.URLParam(request, "id")
|
||||||
envVarIDStr := chi.URLParam(request, "varID")
|
envVarIDStr := chi.URLParam(request, "envID")
|
||||||
|
|
||||||
envVarID, parseErr := strconv.ParseInt(envVarIDStr, 10, 64)
|
envVarID, parseErr := strconv.ParseInt(envVarIDStr, 10, 64)
|
||||||
if parseErr != nil {
|
if parseErr != nil {
|
||||||
@@ -1022,14 +1003,6 @@ func (h *Handlers) HandleVolumeAdd() http.HandlerFunc {
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
pathErr := validateVolumePaths(hostPath, containerPath)
|
|
||||||
if pathErr != nil {
|
|
||||||
h.log.Error("invalid volume path", "error", pathErr)
|
|
||||||
http.Redirect(writer, request, "/apps/"+application.ID, http.StatusSeeOther)
|
|
||||||
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
volume := models.NewVolume(h.db)
|
volume := models.NewVolume(h.db)
|
||||||
volume.AppID = application.ID
|
volume.AppID = application.ID
|
||||||
volume.HostPath = hostPath
|
volume.HostPath = hostPath
|
||||||
|
|||||||
@@ -1,6 +0,0 @@
|
|||||||
package handlers
|
|
||||||
|
|
||||||
// ValidateRepoURLForTest exports validateRepoURL for testing.
|
|
||||||
func ValidateRepoURLForTest(repoURL string) error {
|
|
||||||
return validateRepoURL(repoURL)
|
|
||||||
}
|
|
||||||
@@ -564,7 +564,7 @@ func TestDeleteEnvVarOwnershipVerification(t *testing.T) { //nolint:dupl // inte
|
|||||||
return "/apps/" + appID + "/env/" + strconv.FormatInt(resourceID, 10) + "/delete"
|
return "/apps/" + appID + "/env/" + strconv.FormatInt(resourceID, 10) + "/delete"
|
||||||
},
|
},
|
||||||
chiParams: func(appID string, resourceID int64) map[string]string {
|
chiParams: func(appID string, resourceID int64) map[string]string {
|
||||||
return map[string]string{"id": appID, "varID": strconv.FormatInt(resourceID, 10)}
|
return map[string]string{"id": appID, "envID": strconv.FormatInt(resourceID, 10)}
|
||||||
},
|
},
|
||||||
handler: func(h *handlers.Handlers) http.HandlerFunc { return h.HandleEnvVarDelete() },
|
handler: func(h *handlers.Handlers) http.HandlerFunc { return h.HandleEnvVarDelete() },
|
||||||
verifyFn: func(t *testing.T, tc *testContext, resourceID int64) {
|
verifyFn: func(t *testing.T, tc *testContext, resourceID int64) {
|
||||||
@@ -695,153 +695,6 @@ func TestDeletePortOwnershipVerification(t *testing.T) {
|
|||||||
assert.NotNil(t, found, "port should still exist after IDOR attempt")
|
assert.NotNil(t, found, "port should still exist after IDOR attempt")
|
||||||
}
|
}
|
||||||
|
|
||||||
// TestHandleEnvVarDeleteUsesCorrectRouteParam verifies that HandleEnvVarDelete
|
|
||||||
// reads the "varID" chi URL parameter (matching the route definition {varID}),
|
|
||||||
// not a mismatched name like "envID".
|
|
||||||
func TestHandleEnvVarDeleteUsesCorrectRouteParam(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
testCtx := setupTestHandlers(t)
|
|
||||||
|
|
||||||
createdApp := createTestApp(t, testCtx, "envdelete-param-app")
|
|
||||||
|
|
||||||
envVar := models.NewEnvVar(testCtx.database)
|
|
||||||
envVar.AppID = createdApp.ID
|
|
||||||
envVar.Key = "DELETE_ME"
|
|
||||||
envVar.Value = "gone"
|
|
||||||
require.NoError(t, envVar.Save(context.Background()))
|
|
||||||
|
|
||||||
// Use chi router with the real route pattern to test param name
|
|
||||||
r := chi.NewRouter()
|
|
||||||
r.Post("/apps/{id}/env-vars/{varID}/delete", testCtx.handlers.HandleEnvVarDelete())
|
|
||||||
|
|
||||||
request := httptest.NewRequest(
|
|
||||||
http.MethodPost,
|
|
||||||
"/apps/"+createdApp.ID+"/env-vars/"+strconv.FormatInt(envVar.ID, 10)+"/delete",
|
|
||||||
nil,
|
|
||||||
)
|
|
||||||
recorder := httptest.NewRecorder()
|
|
||||||
|
|
||||||
r.ServeHTTP(recorder, request)
|
|
||||||
|
|
||||||
assert.Equal(t, http.StatusSeeOther, recorder.Code)
|
|
||||||
|
|
||||||
// Verify the env var was actually deleted
|
|
||||||
found, findErr := models.FindEnvVar(context.Background(), testCtx.database, envVar.ID)
|
|
||||||
require.NoError(t, findErr)
|
|
||||||
assert.Nil(t, found, "env var should be deleted when using correct route param")
|
|
||||||
}
|
|
||||||
|
|
||||||
// TestHandleVolumeAddValidatesPaths verifies that HandleVolumeAdd validates
|
|
||||||
// host and container paths (same as HandleVolumeEdit).
|
|
||||||
func TestHandleVolumeAddValidatesPaths(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
testCtx := setupTestHandlers(t)
|
|
||||||
|
|
||||||
createdApp := createTestApp(t, testCtx, "volume-validate-app")
|
|
||||||
|
|
||||||
tests := []struct {
|
|
||||||
name string
|
|
||||||
hostPath string
|
|
||||||
containerPath string
|
|
||||||
shouldCreate bool
|
|
||||||
}{
|
|
||||||
{"relative host path rejected", "relative/path", "/container", false},
|
|
||||||
{"relative container path rejected", "/host", "relative/path", false},
|
|
||||||
{"unclean host path rejected", "/host/../etc", "/container", false},
|
|
||||||
{"valid paths accepted", "/host/data", "/container/data", true},
|
|
||||||
}
|
|
||||||
|
|
||||||
for _, tt := range tests {
|
|
||||||
t.Run(tt.name, func(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
form := url.Values{}
|
|
||||||
form.Set("host_path", tt.hostPath)
|
|
||||||
form.Set("container_path", tt.containerPath)
|
|
||||||
|
|
||||||
request := httptest.NewRequest(
|
|
||||||
http.MethodPost,
|
|
||||||
"/apps/"+createdApp.ID+"/volumes",
|
|
||||||
strings.NewReader(form.Encode()),
|
|
||||||
)
|
|
||||||
request.Header.Set("Content-Type", "application/x-www-form-urlencoded")
|
|
||||||
request = addChiURLParams(request, map[string]string{"id": createdApp.ID})
|
|
||||||
|
|
||||||
recorder := httptest.NewRecorder()
|
|
||||||
|
|
||||||
handler := testCtx.handlers.HandleVolumeAdd()
|
|
||||||
handler.ServeHTTP(recorder, request)
|
|
||||||
|
|
||||||
assert.Equal(t, http.StatusSeeOther, recorder.Code)
|
|
||||||
|
|
||||||
// Check if volume was created by listing volumes
|
|
||||||
volumes, _ := createdApp.GetVolumes(context.Background())
|
|
||||||
found := false
|
|
||||||
|
|
||||||
for _, v := range volumes {
|
|
||||||
if v.HostPath == tt.hostPath && v.ContainerPath == tt.containerPath {
|
|
||||||
found = true
|
|
||||||
// Clean up for isolation
|
|
||||||
_ = v.Delete(context.Background())
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
if tt.shouldCreate {
|
|
||||||
assert.True(t, found, "volume should be created for valid paths")
|
|
||||||
} else {
|
|
||||||
assert.False(t, found, "volume should NOT be created for invalid paths")
|
|
||||||
}
|
|
||||||
})
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// TestSetupRequiredExemptsHealthAndStaticAndAPI verifies that the SetupRequired
|
|
||||||
// middleware allows /health, /s/*, and /api/* paths through even when setup is required.
|
|
||||||
func TestSetupRequiredExemptsHealthAndStaticAndAPI(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
testCtx := setupTestHandlers(t)
|
|
||||||
|
|
||||||
// No user created, so setup IS required
|
|
||||||
mw := testCtx.middleware.SetupRequired()
|
|
||||||
|
|
||||||
okHandler := http.HandlerFunc(func(w http.ResponseWriter, _ *http.Request) {
|
|
||||||
w.WriteHeader(http.StatusOK)
|
|
||||||
_, _ = w.Write([]byte("OK"))
|
|
||||||
})
|
|
||||||
|
|
||||||
wrapped := mw(okHandler)
|
|
||||||
|
|
||||||
exemptPaths := []string{"/health", "/s/style.css", "/s/js/app.js", "/api/v1/apps", "/api/v1/login"}
|
|
||||||
|
|
||||||
for _, path := range exemptPaths {
|
|
||||||
t.Run(path, func(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
req := httptest.NewRequest(http.MethodGet, path, nil)
|
|
||||||
rr := httptest.NewRecorder()
|
|
||||||
wrapped.ServeHTTP(rr, req)
|
|
||||||
|
|
||||||
assert.Equal(t, http.StatusOK, rr.Code,
|
|
||||||
"path %s should be exempt from setup redirect", path)
|
|
||||||
})
|
|
||||||
}
|
|
||||||
|
|
||||||
// Non-exempt path should redirect to /setup
|
|
||||||
t.Run("non-exempt redirects", func(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
req := httptest.NewRequest(http.MethodGet, "/", nil)
|
|
||||||
rr := httptest.NewRecorder()
|
|
||||||
wrapped.ServeHTTP(rr, req)
|
|
||||||
|
|
||||||
assert.Equal(t, http.StatusSeeOther, rr.Code)
|
|
||||||
assert.Equal(t, "/setup", rr.Header().Get("Location"))
|
|
||||||
})
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestHandleCancelDeployRedirects(t *testing.T) {
|
func TestHandleCancelDeployRedirects(t *testing.T) {
|
||||||
t.Parallel()
|
t.Parallel()
|
||||||
|
|
||||||
|
|||||||
@@ -1,77 +0,0 @@
|
|||||||
package handlers
|
|
||||||
|
|
||||||
import (
|
|
||||||
"errors"
|
|
||||||
"net/url"
|
|
||||||
"regexp"
|
|
||||||
"strings"
|
|
||||||
)
|
|
||||||
|
|
||||||
// Repo URL validation errors.
|
|
||||||
var (
|
|
||||||
errRepoURLEmpty = errors.New("repository URL must not be empty")
|
|
||||||
errRepoURLScheme = errors.New("file:// URLs are not allowed for security reasons")
|
|
||||||
errRepoURLInvalid = errors.New("repository URL must use https://, http://, ssh://, git://, or git@host:path format")
|
|
||||||
errRepoURLNoHost = errors.New("repository URL must include a host")
|
|
||||||
errRepoURLNoPath = errors.New("repository URL must include a path")
|
|
||||||
)
|
|
||||||
|
|
||||||
// scpLikeRepoRe matches SCP-like git URLs: git@host:path (e.g. git@github.com:user/repo.git).
|
|
||||||
// Only the "git" user is allowed, as that is the standard for SSH deploy keys.
|
|
||||||
var scpLikeRepoRe = regexp.MustCompile(`^git@[a-zA-Z0-9._-]+:.+$`)
|
|
||||||
|
|
||||||
// allowedRepoSchemes lists the URL schemes accepted for repository URLs.
|
|
||||||
//
|
|
||||||
//nolint:gochecknoglobals // package-level constant map parsed once
|
|
||||||
var allowedRepoSchemes = map[string]bool{
|
|
||||||
"https": true,
|
|
||||||
"http": true,
|
|
||||||
"ssh": true,
|
|
||||||
"git": true,
|
|
||||||
}
|
|
||||||
|
|
||||||
// validateRepoURL checks that the given repository URL is valid and uses an allowed scheme.
|
|
||||||
func validateRepoURL(repoURL string) error {
|
|
||||||
if strings.TrimSpace(repoURL) == "" {
|
|
||||||
return errRepoURLEmpty
|
|
||||||
}
|
|
||||||
|
|
||||||
// Reject path traversal in any URL format
|
|
||||||
if strings.Contains(repoURL, "..") {
|
|
||||||
return errRepoURLInvalid
|
|
||||||
}
|
|
||||||
|
|
||||||
// Check for SCP-like git URLs first (git@host:path)
|
|
||||||
if scpLikeRepoRe.MatchString(repoURL) {
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// Reject file:// explicitly
|
|
||||||
if strings.HasPrefix(strings.ToLower(repoURL), "file://") {
|
|
||||||
return errRepoURLScheme
|
|
||||||
}
|
|
||||||
|
|
||||||
return validateParsedRepoURL(repoURL)
|
|
||||||
}
|
|
||||||
|
|
||||||
// validateParsedRepoURL validates a standard URL-format repository URL.
|
|
||||||
func validateParsedRepoURL(repoURL string) error {
|
|
||||||
parsed, err := url.Parse(repoURL)
|
|
||||||
if err != nil {
|
|
||||||
return errRepoURLInvalid
|
|
||||||
}
|
|
||||||
|
|
||||||
if !allowedRepoSchemes[strings.ToLower(parsed.Scheme)] {
|
|
||||||
return errRepoURLInvalid
|
|
||||||
}
|
|
||||||
|
|
||||||
if parsed.Host == "" {
|
|
||||||
return errRepoURLNoHost
|
|
||||||
}
|
|
||||||
|
|
||||||
if parsed.Path == "" || parsed.Path == "/" {
|
|
||||||
return errRepoURLNoPath
|
|
||||||
}
|
|
||||||
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
@@ -1,60 +0,0 @@
|
|||||||
package handlers_test
|
|
||||||
|
|
||||||
import (
|
|
||||||
"testing"
|
|
||||||
|
|
||||||
"git.eeqj.de/sneak/upaas/internal/handlers"
|
|
||||||
)
|
|
||||||
|
|
||||||
func TestValidateRepoURL(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
tests := []struct {
|
|
||||||
name string
|
|
||||||
url string
|
|
||||||
wantErr bool
|
|
||||||
}{
|
|
||||||
// Valid URLs
|
|
||||||
{name: "https URL", url: "https://github.com/user/repo.git", wantErr: false},
|
|
||||||
{name: "http URL", url: "http://github.com/user/repo.git", wantErr: false},
|
|
||||||
{name: "ssh URL", url: "ssh://git@github.com/user/repo.git", wantErr: false},
|
|
||||||
{name: "git URL", url: "git://github.com/user/repo.git", wantErr: false},
|
|
||||||
{name: "SCP-like URL", url: "git@github.com:user/repo.git", wantErr: false},
|
|
||||||
{name: "SCP-like with dots", url: "git@git.example.com:org/repo.git", wantErr: false},
|
|
||||||
{name: "https without .git", url: "https://github.com/user/repo", wantErr: false},
|
|
||||||
{name: "https with port", url: "https://git.example.com:8443/user/repo.git", wantErr: false},
|
|
||||||
|
|
||||||
// Invalid URLs
|
|
||||||
{name: "empty string", url: "", wantErr: true},
|
|
||||||
{name: "whitespace only", url: " ", wantErr: true},
|
|
||||||
{name: "file URL", url: "file:///etc/passwd", wantErr: true},
|
|
||||||
{name: "file URL uppercase", url: "FILE:///etc/passwd", wantErr: true},
|
|
||||||
{name: "bare path", url: "/some/local/path", wantErr: true},
|
|
||||||
{name: "relative path", url: "../repo", wantErr: true},
|
|
||||||
{name: "just a word", url: "notaurl", wantErr: true},
|
|
||||||
{name: "ftp URL", url: "ftp://example.com/repo.git", wantErr: true},
|
|
||||||
{name: "no host https", url: "https:///path", wantErr: true},
|
|
||||||
{name: "no path https", url: "https://github.com", wantErr: true},
|
|
||||||
{name: "no path https trailing slash", url: "https://github.com/", wantErr: true},
|
|
||||||
{name: "SCP-like non-git user", url: "root@github.com:user/repo.git", wantErr: true},
|
|
||||||
{name: "SCP-like arbitrary user", url: "admin@github.com:user/repo.git", wantErr: true},
|
|
||||||
{name: "path traversal SCP", url: "git@github.com:../../etc/passwd", wantErr: true},
|
|
||||||
{name: "path traversal https", url: "https://github.com/user/../../../etc/passwd", wantErr: true},
|
|
||||||
{name: "path traversal in middle", url: "https://github.com/user/repo/../secret", wantErr: true},
|
|
||||||
}
|
|
||||||
|
|
||||||
for _, tc := range tests {
|
|
||||||
t.Run(tc.name, func(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
err := handlers.ValidateRepoURLForTest(tc.url)
|
|
||||||
if tc.wantErr && err == nil {
|
|
||||||
t.Errorf("ValidateRepoURLForTest(%q) = nil, want error", tc.url)
|
|
||||||
}
|
|
||||||
|
|
||||||
if !tc.wantErr && err != nil {
|
|
||||||
t.Errorf("ValidateRepoURLForTest(%q) = %v, want nil", tc.url, err)
|
|
||||||
}
|
|
||||||
})
|
|
||||||
}
|
|
||||||
}
|
|
||||||
@@ -1,30 +0,0 @@
|
|||||||
package handlers
|
|
||||||
|
|
||||||
import (
|
|
||||||
"regexp"
|
|
||||||
"strings"
|
|
||||||
)
|
|
||||||
|
|
||||||
// ansiEscapePattern matches ANSI escape sequences (CSI, OSC, and single-character escapes).
|
|
||||||
var ansiEscapePattern = regexp.MustCompile(`(\x1b\[[0-9;]*[a-zA-Z]|\x1b\][^\x07]*\x07|\x1b[^[\]])`)
|
|
||||||
|
|
||||||
// SanitizeLogs strips ANSI escape sequences and non-printable control characters
|
|
||||||
// from container log output. Newlines (\n), carriage returns (\r), and tabs (\t)
|
|
||||||
// are preserved. This ensures that attacker-controlled container output cannot
|
|
||||||
// inject terminal escape sequences or other dangerous control characters.
|
|
||||||
func SanitizeLogs(input string) string {
|
|
||||||
// Strip ANSI escape sequences
|
|
||||||
result := ansiEscapePattern.ReplaceAllString(input, "")
|
|
||||||
|
|
||||||
// Strip remaining non-printable characters (keep \n, \r, \t)
|
|
||||||
var b strings.Builder
|
|
||||||
b.Grow(len(result))
|
|
||||||
|
|
||||||
for _, r := range result {
|
|
||||||
if r == '\n' || r == '\r' || r == '\t' || r >= ' ' {
|
|
||||||
b.WriteRune(r)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
return b.String()
|
|
||||||
}
|
|
||||||
@@ -1,84 +0,0 @@
|
|||||||
package handlers_test
|
|
||||||
|
|
||||||
import (
|
|
||||||
"testing"
|
|
||||||
|
|
||||||
"git.eeqj.de/sneak/upaas/internal/handlers"
|
|
||||||
)
|
|
||||||
|
|
||||||
func TestSanitizeLogs(t *testing.T) { //nolint:funlen // table-driven tests
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
tests := []struct {
|
|
||||||
name string
|
|
||||||
input string
|
|
||||||
expected string
|
|
||||||
}{
|
|
||||||
{
|
|
||||||
name: "plain text unchanged",
|
|
||||||
input: "hello world\n",
|
|
||||||
expected: "hello world\n",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "strips ANSI color codes",
|
|
||||||
input: "\x1b[31mERROR\x1b[0m: something failed\n",
|
|
||||||
expected: "ERROR: something failed\n",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "strips OSC sequences",
|
|
||||||
input: "\x1b]0;window title\x07normal text\n",
|
|
||||||
expected: "normal text\n",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "strips null bytes",
|
|
||||||
input: "hello\x00world\n",
|
|
||||||
expected: "helloworld\n",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "strips bell characters",
|
|
||||||
input: "alert\x07here\n",
|
|
||||||
expected: "alerthere\n",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "preserves tabs",
|
|
||||||
input: "field1\tfield2\tfield3\n",
|
|
||||||
expected: "field1\tfield2\tfield3\n",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "preserves carriage returns",
|
|
||||||
input: "line1\r\nline2\r\n",
|
|
||||||
expected: "line1\r\nline2\r\n",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "strips mixed escape sequences",
|
|
||||||
input: "\x1b[32m2024-01-01\x1b[0m \x1b[1mINFO\x1b[0m starting\x00\n",
|
|
||||||
expected: "2024-01-01 INFO starting\n",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "empty string",
|
|
||||||
input: "",
|
|
||||||
expected: "",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "only control characters",
|
|
||||||
input: "\x00\x01\x02\x03",
|
|
||||||
expected: "",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "cursor movement sequences stripped",
|
|
||||||
input: "\x1b[2J\x1b[H\x1b[3Atext\n",
|
|
||||||
expected: "text\n",
|
|
||||||
},
|
|
||||||
}
|
|
||||||
|
|
||||||
for _, tt := range tests {
|
|
||||||
t.Run(tt.name, func(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
got := handlers.SanitizeLogs(tt.input)
|
|
||||||
if got != tt.expected {
|
|
||||||
t.Errorf("SanitizeLogs(%q) = %q, want %q", tt.input, got, tt.expected)
|
|
||||||
}
|
|
||||||
})
|
|
||||||
}
|
|
||||||
}
|
|
||||||
@@ -1,11 +0,0 @@
|
|||||||
package logger
|
|
||||||
|
|
||||||
import "log/slog"
|
|
||||||
|
|
||||||
// NewForTest creates a Logger wrapping the given slog.Logger, for use in tests.
|
|
||||||
func NewForTest(log *slog.Logger) *Logger {
|
|
||||||
return &Logger{
|
|
||||||
log: log,
|
|
||||||
level: new(slog.LevelVar),
|
|
||||||
}
|
|
||||||
}
|
|
||||||
@@ -411,14 +411,8 @@ func (m *Middleware) SetupRequired() func(http.Handler) http.Handler {
|
|||||||
}
|
}
|
||||||
|
|
||||||
if setupRequired {
|
if setupRequired {
|
||||||
path := request.URL.Path
|
// Allow access to setup page
|
||||||
|
if request.URL.Path == "/setup" {
|
||||||
// Allow access to setup page, health endpoint, static
|
|
||||||
// assets, and API routes even before setup is complete.
|
|
||||||
if path == "/setup" ||
|
|
||||||
path == "/health" ||
|
|
||||||
strings.HasPrefix(path, "/s/") ||
|
|
||||||
strings.HasPrefix(path, "/api/") {
|
|
||||||
next.ServeHTTP(writer, request)
|
next.ServeHTTP(writer, request)
|
||||||
|
|
||||||
return
|
return
|
||||||
|
|||||||
@@ -114,7 +114,10 @@ func (s *Server) SetupRoutes() {
|
|||||||
r.Get("/whoami", s.handlers.HandleAPIWhoAmI())
|
r.Get("/whoami", s.handlers.HandleAPIWhoAmI())
|
||||||
|
|
||||||
r.Get("/apps", s.handlers.HandleAPIListApps())
|
r.Get("/apps", s.handlers.HandleAPIListApps())
|
||||||
|
r.Post("/apps", s.handlers.HandleAPICreateApp())
|
||||||
r.Get("/apps/{id}", s.handlers.HandleAPIGetApp())
|
r.Get("/apps/{id}", s.handlers.HandleAPIGetApp())
|
||||||
|
r.Delete("/apps/{id}", s.handlers.HandleAPIDeleteApp())
|
||||||
|
r.Post("/apps/{id}/deploy", s.handlers.HandleAPITriggerDeploy())
|
||||||
r.Get("/apps/{id}/deployments", s.handlers.HandleAPIListDeployments())
|
r.Get("/apps/{id}/deployments", s.handlers.HandleAPIListDeployments())
|
||||||
})
|
})
|
||||||
})
|
})
|
||||||
|
|||||||
@@ -417,13 +417,15 @@ func (svc *Service) executeRollback(
|
|||||||
|
|
||||||
svc.removeOldContainer(ctx, app, deployment)
|
svc.removeOldContainer(ctx, app, deployment)
|
||||||
|
|
||||||
rollbackOpts, err := svc.buildContainerOptions(ctx, app, previousImageID)
|
rollbackOpts, err := svc.buildContainerOptions(ctx, app, deployment.ID)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
svc.failDeployment(bgCtx, app, deployment, err)
|
svc.failDeployment(bgCtx, app, deployment, err)
|
||||||
|
|
||||||
return fmt.Errorf("failed to build container options: %w", err)
|
return fmt.Errorf("failed to build container options: %w", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
rollbackOpts.Image = previousImageID
|
||||||
|
|
||||||
containerID, err := svc.docker.CreateContainer(ctx, rollbackOpts)
|
containerID, err := svc.docker.CreateContainer(ctx, rollbackOpts)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
svc.failDeployment(bgCtx, app, deployment, fmt.Errorf("failed to create rollback container: %w", err))
|
svc.failDeployment(bgCtx, app, deployment, fmt.Errorf("failed to create rollback container: %w", err))
|
||||||
@@ -724,7 +726,6 @@ func (svc *Service) cleanupCancelledDeploy(
|
|||||||
} else {
|
} else {
|
||||||
svc.log.Info("cleaned up build dir from cancelled deploy",
|
svc.log.Info("cleaned up build dir from cancelled deploy",
|
||||||
"app", app.Name, "path", dirPath)
|
"app", app.Name, "path", dirPath)
|
||||||
|
|
||||||
_ = deployment.AppendLog(ctx, "Cleaned up build directory")
|
_ = deployment.AppendLog(ctx, "Cleaned up build directory")
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -1016,9 +1017,9 @@ func (svc *Service) createAndStartContainer(
|
|||||||
ctx context.Context,
|
ctx context.Context,
|
||||||
app *models.App,
|
app *models.App,
|
||||||
deployment *models.Deployment,
|
deployment *models.Deployment,
|
||||||
imageID string,
|
_ string,
|
||||||
) (string, error) {
|
) (string, error) {
|
||||||
containerOpts, err := svc.buildContainerOptions(ctx, app, imageID)
|
containerOpts, err := svc.buildContainerOptions(ctx, app, deployment.ID)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
svc.failDeployment(ctx, app, deployment, err)
|
svc.failDeployment(ctx, app, deployment, err)
|
||||||
|
|
||||||
@@ -1062,7 +1063,7 @@ func (svc *Service) createAndStartContainer(
|
|||||||
func (svc *Service) buildContainerOptions(
|
func (svc *Service) buildContainerOptions(
|
||||||
ctx context.Context,
|
ctx context.Context,
|
||||||
app *models.App,
|
app *models.App,
|
||||||
imageID string,
|
deploymentID int64,
|
||||||
) (docker.CreateContainerOptions, error) {
|
) (docker.CreateContainerOptions, error) {
|
||||||
envVars, err := app.GetEnvVars(ctx)
|
envVars, err := app.GetEnvVars(ctx)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@@ -1096,7 +1097,7 @@ func (svc *Service) buildContainerOptions(
|
|||||||
|
|
||||||
return docker.CreateContainerOptions{
|
return docker.CreateContainerOptions{
|
||||||
Name: "upaas-" + app.Name,
|
Name: "upaas-" + app.Name,
|
||||||
Image: imageID,
|
Image: fmt.Sprintf("upaas-%s:%d", app.Name, deploymentID),
|
||||||
Env: envMap,
|
Env: envMap,
|
||||||
Labels: buildLabelMap(app, labels),
|
Labels: buildLabelMap(app, labels),
|
||||||
Volumes: buildVolumeMounts(volumes),
|
Volumes: buildVolumeMounts(volumes),
|
||||||
|
|||||||
@@ -32,7 +32,7 @@ func TestCleanupCancelledDeploy_RemovesBuildDir(t *testing.T) {
|
|||||||
require.NoError(t, os.MkdirAll(deployDir, 0o750))
|
require.NoError(t, os.MkdirAll(deployDir, 0o750))
|
||||||
|
|
||||||
// Create a file inside to verify full removal
|
// Create a file inside to verify full removal
|
||||||
require.NoError(t, os.WriteFile(filepath.Join(deployDir, "work"), []byte("test"), 0o600))
|
require.NoError(t, os.WriteFile(filepath.Join(deployDir, "work"), []byte("test"), 0o640))
|
||||||
|
|
||||||
// Also create a dir for a different deployment (should NOT be removed)
|
// Also create a dir for a different deployment (should NOT be removed)
|
||||||
otherDir := filepath.Join(buildDir, "99-xyz789")
|
otherDir := filepath.Join(buildDir, "99-xyz789")
|
||||||
|
|||||||
@@ -1,44 +0,0 @@
|
|||||||
package deploy_test
|
|
||||||
|
|
||||||
import (
|
|
||||||
"context"
|
|
||||||
"log/slog"
|
|
||||||
"os"
|
|
||||||
"testing"
|
|
||||||
|
|
||||||
"git.eeqj.de/sneak/upaas/internal/database"
|
|
||||||
"git.eeqj.de/sneak/upaas/internal/models"
|
|
||||||
"git.eeqj.de/sneak/upaas/internal/service/deploy"
|
|
||||||
)
|
|
||||||
|
|
||||||
func TestBuildContainerOptionsUsesImageID(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
db := database.NewTestDatabase(t)
|
|
||||||
|
|
||||||
app := models.NewApp(db)
|
|
||||||
app.Name = "myapp"
|
|
||||||
|
|
||||||
err := app.Save(context.Background())
|
|
||||||
if err != nil {
|
|
||||||
t.Fatalf("failed to save app: %v", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
log := slog.New(slog.NewTextHandler(os.Stderr, nil))
|
|
||||||
svc := deploy.NewTestService(log)
|
|
||||||
|
|
||||||
const expectedImageID = "sha256:abc123def456"
|
|
||||||
|
|
||||||
opts, err := svc.BuildContainerOptionsExported(context.Background(), app, expectedImageID)
|
|
||||||
if err != nil {
|
|
||||||
t.Fatalf("buildContainerOptions returned error: %v", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
if opts.Image != expectedImageID {
|
|
||||||
t.Errorf("expected Image=%q, got %q", expectedImageID, opts.Image)
|
|
||||||
}
|
|
||||||
|
|
||||||
if opts.Name != "upaas-myapp" {
|
|
||||||
t.Errorf("expected Name=%q, got %q", "upaas-myapp", opts.Name)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
@@ -10,7 +10,6 @@ import (
|
|||||||
|
|
||||||
"git.eeqj.de/sneak/upaas/internal/config"
|
"git.eeqj.de/sneak/upaas/internal/config"
|
||||||
"git.eeqj.de/sneak/upaas/internal/docker"
|
"git.eeqj.de/sneak/upaas/internal/docker"
|
||||||
"git.eeqj.de/sneak/upaas/internal/models"
|
|
||||||
)
|
)
|
||||||
|
|
||||||
// NewTestService creates a Service with minimal dependencies for testing.
|
// NewTestService creates a Service with minimal dependencies for testing.
|
||||||
@@ -53,10 +52,10 @@ func NewTestServiceWithConfig(log *slog.Logger, cfg *config.Config, dockerClient
|
|||||||
// cleanupCancelledDeploy for testing. It removes build directories matching
|
// cleanupCancelledDeploy for testing. It removes build directories matching
|
||||||
// the deployment ID prefix.
|
// the deployment ID prefix.
|
||||||
func (svc *Service) CleanupCancelledDeploy(
|
func (svc *Service) CleanupCancelledDeploy(
|
||||||
_ context.Context,
|
ctx context.Context,
|
||||||
appName string,
|
appName string,
|
||||||
deploymentID int64,
|
deploymentID int64,
|
||||||
_ string,
|
imageID string,
|
||||||
) {
|
) {
|
||||||
// We can't create real models.App/Deployment in tests easily,
|
// We can't create real models.App/Deployment in tests easily,
|
||||||
// so we test the build dir cleanup portion directly.
|
// so we test the build dir cleanup portion directly.
|
||||||
@@ -81,12 +80,3 @@ func (svc *Service) CleanupCancelledDeploy(
|
|||||||
func (svc *Service) GetBuildDirExported(appName string) string {
|
func (svc *Service) GetBuildDirExported(appName string) string {
|
||||||
return svc.GetBuildDir(appName)
|
return svc.GetBuildDir(appName)
|
||||||
}
|
}
|
||||||
|
|
||||||
// BuildContainerOptionsExported exposes buildContainerOptions for testing.
|
|
||||||
func (svc *Service) BuildContainerOptionsExported(
|
|
||||||
ctx context.Context,
|
|
||||||
app *models.App,
|
|
||||||
imageID string,
|
|
||||||
) (docker.CreateContainerOptions, error) {
|
|
||||||
return svc.buildContainerOptions(ctx, app, imageID)
|
|
||||||
}
|
|
||||||
|
|||||||
@@ -10,7 +10,6 @@ import (
|
|||||||
"fmt"
|
"fmt"
|
||||||
"log/slog"
|
"log/slog"
|
||||||
"net/http"
|
"net/http"
|
||||||
"net/url"
|
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"go.uber.org/fx"
|
"go.uber.org/fx"
|
||||||
@@ -248,15 +247,10 @@ func (svc *Service) sendNtfy(
|
|||||||
) error {
|
) error {
|
||||||
svc.log.Debug("sending ntfy notification", "topic", topic, "title", title)
|
svc.log.Debug("sending ntfy notification", "topic", topic, "title", title)
|
||||||
|
|
||||||
parsedURL, err := url.ParseRequestURI(topic)
|
|
||||||
if err != nil {
|
|
||||||
return fmt.Errorf("invalid ntfy topic URL: %w", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
request, err := http.NewRequestWithContext(
|
request, err := http.NewRequestWithContext(
|
||||||
ctx,
|
ctx,
|
||||||
http.MethodPost,
|
http.MethodPost,
|
||||||
parsedURL.String(),
|
topic,
|
||||||
bytes.NewBufferString(message),
|
bytes.NewBufferString(message),
|
||||||
)
|
)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@@ -266,7 +260,7 @@ func (svc *Service) sendNtfy(
|
|||||||
request.Header.Set("Title", title)
|
request.Header.Set("Title", title)
|
||||||
request.Header.Set("Priority", svc.ntfyPriority(priority))
|
request.Header.Set("Priority", svc.ntfyPriority(priority))
|
||||||
|
|
||||||
resp, err := svc.client.Do(request) // #nosec G704 -- URL from validated config, not user input
|
resp, err := svc.client.Do(request)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return fmt.Errorf("failed to send ntfy request: %w", err)
|
return fmt.Errorf("failed to send ntfy request: %w", err)
|
||||||
}
|
}
|
||||||
@@ -346,15 +340,10 @@ func (svc *Service) sendSlack(
|
|||||||
return fmt.Errorf("failed to marshal slack payload: %w", err)
|
return fmt.Errorf("failed to marshal slack payload: %w", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
parsedWebhookURL, err := url.ParseRequestURI(webhookURL)
|
|
||||||
if err != nil {
|
|
||||||
return fmt.Errorf("invalid slack webhook URL: %w", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
request, err := http.NewRequestWithContext(
|
request, err := http.NewRequestWithContext(
|
||||||
ctx,
|
ctx,
|
||||||
http.MethodPost,
|
http.MethodPost,
|
||||||
parsedWebhookURL.String(),
|
webhookURL,
|
||||||
bytes.NewBuffer(body),
|
bytes.NewBuffer(body),
|
||||||
)
|
)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@@ -363,7 +352,7 @@ func (svc *Service) sendSlack(
|
|||||||
|
|
||||||
request.Header.Set("Content-Type", "application/json")
|
request.Header.Set("Content-Type", "application/json")
|
||||||
|
|
||||||
resp, err := svc.client.Do(request) // #nosec G704 -- URL from validated config, not user input
|
resp, err := svc.client.Do(request)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return fmt.Errorf("failed to send slack request: %w", err)
|
return fmt.Errorf("failed to send slack request: %w", err)
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -12,7 +12,7 @@ import (
|
|||||||
|
|
||||||
// KeyPair contains an SSH key pair.
|
// KeyPair contains an SSH key pair.
|
||||||
type KeyPair struct {
|
type KeyPair struct {
|
||||||
PrivateKey string `json:"-"`
|
PrivateKey string
|
||||||
PublicKey string
|
PublicKey string
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -1,194 +0,0 @@
|
|||||||
/**
|
|
||||||
* upaas - App Detail Page Component
|
|
||||||
*
|
|
||||||
* Handles the single-app view: status polling, container logs,
|
|
||||||
* build logs, and recent deployments list.
|
|
||||||
*/
|
|
||||||
|
|
||||||
document.addEventListener("alpine:init", () => {
|
|
||||||
Alpine.data("appDetail", (config) => ({
|
|
||||||
appId: config.appId,
|
|
||||||
currentDeploymentId: config.initialDeploymentId,
|
|
||||||
appStatus: config.initialStatus || "unknown",
|
|
||||||
containerLogs: "Loading container logs...",
|
|
||||||
containerStatus: "unknown",
|
|
||||||
buildLogs: config.initialDeploymentId
|
|
||||||
? "Loading build logs..."
|
|
||||||
: "No deployments yet",
|
|
||||||
buildStatus: config.initialBuildStatus || "unknown",
|
|
||||||
showBuildLogs: !!config.initialDeploymentId,
|
|
||||||
deploying: false,
|
|
||||||
deployments: [],
|
|
||||||
// Track whether user wants auto-scroll (per log pane)
|
|
||||||
_containerAutoScroll: true,
|
|
||||||
_buildAutoScroll: true,
|
|
||||||
_pollTimer: null,
|
|
||||||
|
|
||||||
init() {
|
|
||||||
this.deploying = Alpine.store("utils").isDeploying(this.appStatus);
|
|
||||||
this.fetchAll();
|
|
||||||
this._schedulePoll();
|
|
||||||
|
|
||||||
// Set up scroll listeners after DOM is ready
|
|
||||||
this.$nextTick(() => {
|
|
||||||
this._initScrollTracking(this.$refs.containerLogsWrapper, '_containerAutoScroll');
|
|
||||||
this._initScrollTracking(this.$refs.buildLogsWrapper, '_buildAutoScroll');
|
|
||||||
});
|
|
||||||
},
|
|
||||||
|
|
||||||
_schedulePoll() {
|
|
||||||
if (this._pollTimer) clearTimeout(this._pollTimer);
|
|
||||||
const interval = Alpine.store("utils").isDeploying(this.appStatus) ? 1000 : 10000;
|
|
||||||
this._pollTimer = setTimeout(() => {
|
|
||||||
this.fetchAll();
|
|
||||||
this._schedulePoll();
|
|
||||||
}, interval);
|
|
||||||
},
|
|
||||||
|
|
||||||
_initScrollTracking(el, flag) {
|
|
||||||
if (!el) return;
|
|
||||||
el.addEventListener('scroll', () => {
|
|
||||||
this[flag] = Alpine.store("utils").isScrolledToBottom(el);
|
|
||||||
}, { passive: true });
|
|
||||||
},
|
|
||||||
|
|
||||||
fetchAll() {
|
|
||||||
this.fetchAppStatus();
|
|
||||||
// Only fetch logs when the respective pane is visible
|
|
||||||
if (this.$refs.containerLogsWrapper && this._isElementVisible(this.$refs.containerLogsWrapper)) {
|
|
||||||
this.fetchContainerLogs();
|
|
||||||
}
|
|
||||||
if (this.showBuildLogs && this.$refs.buildLogsWrapper && this._isElementVisible(this.$refs.buildLogsWrapper)) {
|
|
||||||
this.fetchBuildLogs();
|
|
||||||
}
|
|
||||||
this.fetchRecentDeployments();
|
|
||||||
},
|
|
||||||
|
|
||||||
_isElementVisible(el) {
|
|
||||||
if (!el) return false;
|
|
||||||
// Check if element is in viewport (roughly)
|
|
||||||
const rect = el.getBoundingClientRect();
|
|
||||||
return rect.bottom > 0 && rect.top < window.innerHeight;
|
|
||||||
},
|
|
||||||
|
|
||||||
async fetchAppStatus() {
|
|
||||||
try {
|
|
||||||
const res = await fetch(`/apps/${this.appId}/status`);
|
|
||||||
const data = await res.json();
|
|
||||||
const wasDeploying = this.deploying;
|
|
||||||
this.appStatus = data.status;
|
|
||||||
this.deploying = Alpine.store("utils").isDeploying(data.status);
|
|
||||||
|
|
||||||
// Re-schedule polling when deployment state changes
|
|
||||||
if (this.deploying !== wasDeploying) {
|
|
||||||
this._schedulePoll();
|
|
||||||
}
|
|
||||||
|
|
||||||
if (
|
|
||||||
data.latestDeploymentID &&
|
|
||||||
data.latestDeploymentID !== this.currentDeploymentId
|
|
||||||
) {
|
|
||||||
this.currentDeploymentId = data.latestDeploymentID;
|
|
||||||
this.showBuildLogs = true;
|
|
||||||
this.fetchBuildLogs();
|
|
||||||
}
|
|
||||||
} catch (err) {
|
|
||||||
console.error("Status fetch error:", err);
|
|
||||||
}
|
|
||||||
},
|
|
||||||
|
|
||||||
async fetchContainerLogs() {
|
|
||||||
try {
|
|
||||||
const res = await fetch(`/apps/${this.appId}/container-logs`);
|
|
||||||
const data = await res.json();
|
|
||||||
const newLogs = data.logs || "No logs available";
|
|
||||||
const changed = newLogs !== this.containerLogs;
|
|
||||||
this.containerLogs = newLogs;
|
|
||||||
this.containerStatus = data.status;
|
|
||||||
if (changed && this._containerAutoScroll) {
|
|
||||||
this.$nextTick(() => {
|
|
||||||
Alpine.store("utils").scrollToBottom(this.$refs.containerLogsWrapper);
|
|
||||||
});
|
|
||||||
}
|
|
||||||
} catch (err) {
|
|
||||||
this.containerLogs = "Failed to fetch logs";
|
|
||||||
}
|
|
||||||
},
|
|
||||||
|
|
||||||
async fetchBuildLogs() {
|
|
||||||
if (!this.currentDeploymentId) return;
|
|
||||||
try {
|
|
||||||
const res = await fetch(
|
|
||||||
`/apps/${this.appId}/deployments/${this.currentDeploymentId}/logs`,
|
|
||||||
);
|
|
||||||
const data = await res.json();
|
|
||||||
const newLogs = data.logs || "No build logs available";
|
|
||||||
const changed = newLogs !== this.buildLogs;
|
|
||||||
this.buildLogs = newLogs;
|
|
||||||
this.buildStatus = data.status;
|
|
||||||
if (changed && this._buildAutoScroll) {
|
|
||||||
this.$nextTick(() => {
|
|
||||||
Alpine.store("utils").scrollToBottom(this.$refs.buildLogsWrapper);
|
|
||||||
});
|
|
||||||
}
|
|
||||||
} catch (err) {
|
|
||||||
this.buildLogs = "Failed to fetch logs";
|
|
||||||
}
|
|
||||||
},
|
|
||||||
|
|
||||||
async fetchRecentDeployments() {
|
|
||||||
try {
|
|
||||||
const res = await fetch(`/apps/${this.appId}/recent-deployments`);
|
|
||||||
const data = await res.json();
|
|
||||||
this.deployments = data.deployments || [];
|
|
||||||
} catch (err) {
|
|
||||||
console.error("Deployments fetch error:", err);
|
|
||||||
}
|
|
||||||
},
|
|
||||||
|
|
||||||
submitDeploy() {
|
|
||||||
this.deploying = true;
|
|
||||||
},
|
|
||||||
|
|
||||||
get statusBadgeClass() {
|
|
||||||
return Alpine.store("utils").statusBadgeClass(this.appStatus);
|
|
||||||
},
|
|
||||||
|
|
||||||
get statusLabel() {
|
|
||||||
return Alpine.store("utils").statusLabel(this.appStatus);
|
|
||||||
},
|
|
||||||
|
|
||||||
get containerStatusBadgeClass() {
|
|
||||||
return (
|
|
||||||
Alpine.store("utils").statusBadgeClass(this.containerStatus) +
|
|
||||||
" text-xs"
|
|
||||||
);
|
|
||||||
},
|
|
||||||
|
|
||||||
get containerStatusLabel() {
|
|
||||||
return Alpine.store("utils").statusLabel(this.containerStatus);
|
|
||||||
},
|
|
||||||
|
|
||||||
get buildStatusBadgeClass() {
|
|
||||||
return (
|
|
||||||
Alpine.store("utils").statusBadgeClass(this.buildStatus) + " text-xs"
|
|
||||||
);
|
|
||||||
},
|
|
||||||
|
|
||||||
get buildStatusLabel() {
|
|
||||||
return Alpine.store("utils").statusLabel(this.buildStatus);
|
|
||||||
},
|
|
||||||
|
|
||||||
deploymentStatusClass(status) {
|
|
||||||
return Alpine.store("utils").statusBadgeClass(status);
|
|
||||||
},
|
|
||||||
|
|
||||||
deploymentStatusLabel(status) {
|
|
||||||
return Alpine.store("utils").statusLabel(status);
|
|
||||||
},
|
|
||||||
|
|
||||||
formatTime(isoTime) {
|
|
||||||
return Alpine.store("utils").formatRelativeTime(isoTime);
|
|
||||||
},
|
|
||||||
}));
|
|
||||||
});
|
|
||||||
581
static/js/app.js
Normal file
581
static/js/app.js
Normal file
@@ -0,0 +1,581 @@
|
|||||||
|
/**
|
||||||
|
* upaas - Frontend JavaScript with Alpine.js
|
||||||
|
*/
|
||||||
|
|
||||||
|
document.addEventListener("alpine:init", () => {
|
||||||
|
// ============================================
|
||||||
|
// Global Utilities Store
|
||||||
|
// ============================================
|
||||||
|
Alpine.store("utils", {
|
||||||
|
/**
|
||||||
|
* Format a date string as relative time (e.g., "5 minutes ago")
|
||||||
|
*/
|
||||||
|
formatRelativeTime(dateStr) {
|
||||||
|
if (!dateStr) return "";
|
||||||
|
const date = new Date(dateStr);
|
||||||
|
const now = new Date();
|
||||||
|
const diffMs = now - date;
|
||||||
|
const diffSec = Math.floor(diffMs / 1000);
|
||||||
|
const diffMin = Math.floor(diffSec / 60);
|
||||||
|
const diffHour = Math.floor(diffMin / 60);
|
||||||
|
const diffDay = Math.floor(diffHour / 24);
|
||||||
|
|
||||||
|
if (diffSec < 60) return "just now";
|
||||||
|
if (diffMin < 60)
|
||||||
|
return diffMin + (diffMin === 1 ? " minute ago" : " minutes ago");
|
||||||
|
if (diffHour < 24)
|
||||||
|
return diffHour + (diffHour === 1 ? " hour ago" : " hours ago");
|
||||||
|
if (diffDay < 7)
|
||||||
|
return diffDay + (diffDay === 1 ? " day ago" : " days ago");
|
||||||
|
return date.toLocaleDateString();
|
||||||
|
},
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Get the badge class for a given status
|
||||||
|
*/
|
||||||
|
statusBadgeClass(status) {
|
||||||
|
if (status === "running" || status === "success") return "badge-success";
|
||||||
|
if (status === "building" || status === "deploying")
|
||||||
|
return "badge-warning";
|
||||||
|
if (status === "failed" || status === "error") return "badge-error";
|
||||||
|
return "badge-neutral";
|
||||||
|
},
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Format status for display (capitalize first letter)
|
||||||
|
*/
|
||||||
|
statusLabel(status) {
|
||||||
|
if (!status) return "";
|
||||||
|
return status.charAt(0).toUpperCase() + status.slice(1);
|
||||||
|
},
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Check if status indicates active deployment
|
||||||
|
*/
|
||||||
|
isDeploying(status) {
|
||||||
|
return status === "building" || status === "deploying";
|
||||||
|
},
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Scroll an element to the bottom
|
||||||
|
*/
|
||||||
|
scrollToBottom(el) {
|
||||||
|
if (el) {
|
||||||
|
requestAnimationFrame(() => {
|
||||||
|
el.scrollTop = el.scrollHeight;
|
||||||
|
});
|
||||||
|
}
|
||||||
|
},
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Check if a scrollable element is at (or near) the bottom.
|
||||||
|
* Tolerance of 30px accounts for rounding and partial lines.
|
||||||
|
*/
|
||||||
|
isScrolledToBottom(el, tolerance = 30) {
|
||||||
|
if (!el) return true;
|
||||||
|
return el.scrollHeight - el.scrollTop - el.clientHeight <= tolerance;
|
||||||
|
},
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Copy text to clipboard
|
||||||
|
*/
|
||||||
|
async copyToClipboard(text, button) {
|
||||||
|
try {
|
||||||
|
await navigator.clipboard.writeText(text);
|
||||||
|
return true;
|
||||||
|
} catch (err) {
|
||||||
|
// Fallback for older browsers
|
||||||
|
const textArea = document.createElement("textarea");
|
||||||
|
textArea.value = text;
|
||||||
|
textArea.style.position = "fixed";
|
||||||
|
textArea.style.left = "-9999px";
|
||||||
|
document.body.appendChild(textArea);
|
||||||
|
textArea.select();
|
||||||
|
try {
|
||||||
|
document.execCommand("copy");
|
||||||
|
document.body.removeChild(textArea);
|
||||||
|
return true;
|
||||||
|
} catch (e) {
|
||||||
|
document.body.removeChild(textArea);
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
},
|
||||||
|
});
|
||||||
|
|
||||||
|
// ============================================
|
||||||
|
// Copy Button Component
|
||||||
|
// ============================================
|
||||||
|
Alpine.data("copyButton", (targetId) => ({
|
||||||
|
copied: false,
|
||||||
|
async copy() {
|
||||||
|
const target = document.getElementById(targetId);
|
||||||
|
if (!target) return;
|
||||||
|
const text = target.textContent || target.value;
|
||||||
|
const success = await Alpine.store("utils").copyToClipboard(text);
|
||||||
|
if (success) {
|
||||||
|
this.copied = true;
|
||||||
|
setTimeout(() => {
|
||||||
|
this.copied = false;
|
||||||
|
}, 2000);
|
||||||
|
}
|
||||||
|
},
|
||||||
|
}));
|
||||||
|
|
||||||
|
// ============================================
|
||||||
|
// Confirm Action Component
|
||||||
|
// ============================================
|
||||||
|
Alpine.data("confirmAction", (message) => ({
|
||||||
|
confirm(event) {
|
||||||
|
if (!window.confirm(message)) {
|
||||||
|
event.preventDefault();
|
||||||
|
}
|
||||||
|
},
|
||||||
|
}));
|
||||||
|
|
||||||
|
// ============================================
|
||||||
|
// Auto-dismiss Alert Component
|
||||||
|
// ============================================
|
||||||
|
Alpine.data("autoDismiss", (delay = 5000) => ({
|
||||||
|
show: true,
|
||||||
|
init() {
|
||||||
|
setTimeout(() => {
|
||||||
|
this.dismiss();
|
||||||
|
}, delay);
|
||||||
|
},
|
||||||
|
dismiss() {
|
||||||
|
this.show = false;
|
||||||
|
setTimeout(() => {
|
||||||
|
this.$el.remove();
|
||||||
|
}, 300);
|
||||||
|
},
|
||||||
|
}));
|
||||||
|
|
||||||
|
// ============================================
|
||||||
|
// Relative Time Component
|
||||||
|
// ============================================
|
||||||
|
Alpine.data("relativeTime", (isoTime) => ({
|
||||||
|
display: "",
|
||||||
|
init() {
|
||||||
|
this.update();
|
||||||
|
// Update every minute
|
||||||
|
setInterval(() => this.update(), 60000);
|
||||||
|
},
|
||||||
|
update() {
|
||||||
|
this.display = Alpine.store("utils").formatRelativeTime(isoTime);
|
||||||
|
},
|
||||||
|
}));
|
||||||
|
|
||||||
|
// ============================================
|
||||||
|
// App Detail Page Component
|
||||||
|
// ============================================
|
||||||
|
Alpine.data("appDetail", (config) => ({
|
||||||
|
appId: config.appId,
|
||||||
|
currentDeploymentId: config.initialDeploymentId,
|
||||||
|
appStatus: config.initialStatus || "unknown",
|
||||||
|
containerLogs: "Loading container logs...",
|
||||||
|
containerStatus: "unknown",
|
||||||
|
buildLogs: config.initialDeploymentId
|
||||||
|
? "Loading build logs..."
|
||||||
|
: "No deployments yet",
|
||||||
|
buildStatus: config.initialBuildStatus || "unknown",
|
||||||
|
showBuildLogs: !!config.initialDeploymentId,
|
||||||
|
deploying: false,
|
||||||
|
deployments: [],
|
||||||
|
// Track whether user wants auto-scroll (per log pane)
|
||||||
|
_containerAutoScroll: true,
|
||||||
|
_buildAutoScroll: true,
|
||||||
|
_pollTimer: null,
|
||||||
|
|
||||||
|
init() {
|
||||||
|
this.deploying = Alpine.store("utils").isDeploying(this.appStatus);
|
||||||
|
this.fetchAll();
|
||||||
|
this._schedulePoll();
|
||||||
|
|
||||||
|
// Set up scroll listeners after DOM is ready
|
||||||
|
this.$nextTick(() => {
|
||||||
|
this._initScrollTracking(this.$refs.containerLogsWrapper, '_containerAutoScroll');
|
||||||
|
this._initScrollTracking(this.$refs.buildLogsWrapper, '_buildAutoScroll');
|
||||||
|
});
|
||||||
|
},
|
||||||
|
|
||||||
|
_schedulePoll() {
|
||||||
|
if (this._pollTimer) clearTimeout(this._pollTimer);
|
||||||
|
const interval = Alpine.store("utils").isDeploying(this.appStatus) ? 1000 : 10000;
|
||||||
|
this._pollTimer = setTimeout(() => {
|
||||||
|
this.fetchAll();
|
||||||
|
this._schedulePoll();
|
||||||
|
}, interval);
|
||||||
|
},
|
||||||
|
|
||||||
|
_initScrollTracking(el, flag) {
|
||||||
|
if (!el) return;
|
||||||
|
el.addEventListener('scroll', () => {
|
||||||
|
this[flag] = Alpine.store("utils").isScrolledToBottom(el);
|
||||||
|
}, { passive: true });
|
||||||
|
},
|
||||||
|
|
||||||
|
fetchAll() {
|
||||||
|
this.fetchAppStatus();
|
||||||
|
// Only fetch logs when the respective pane is visible
|
||||||
|
if (this.$refs.containerLogsWrapper && this._isElementVisible(this.$refs.containerLogsWrapper)) {
|
||||||
|
this.fetchContainerLogs();
|
||||||
|
}
|
||||||
|
if (this.showBuildLogs && this.$refs.buildLogsWrapper && this._isElementVisible(this.$refs.buildLogsWrapper)) {
|
||||||
|
this.fetchBuildLogs();
|
||||||
|
}
|
||||||
|
this.fetchRecentDeployments();
|
||||||
|
},
|
||||||
|
|
||||||
|
_isElementVisible(el) {
|
||||||
|
if (!el) return false;
|
||||||
|
// Check if element is in viewport (roughly)
|
||||||
|
const rect = el.getBoundingClientRect();
|
||||||
|
return rect.bottom > 0 && rect.top < window.innerHeight;
|
||||||
|
},
|
||||||
|
|
||||||
|
async fetchAppStatus() {
|
||||||
|
try {
|
||||||
|
const res = await fetch(`/apps/${this.appId}/status`);
|
||||||
|
const data = await res.json();
|
||||||
|
const wasDeploying = this.deploying;
|
||||||
|
this.appStatus = data.status;
|
||||||
|
this.deploying = Alpine.store("utils").isDeploying(data.status);
|
||||||
|
|
||||||
|
// Re-schedule polling when deployment state changes
|
||||||
|
if (this.deploying !== wasDeploying) {
|
||||||
|
this._schedulePoll();
|
||||||
|
}
|
||||||
|
|
||||||
|
if (
|
||||||
|
data.latestDeploymentID &&
|
||||||
|
data.latestDeploymentID !== this.currentDeploymentId
|
||||||
|
) {
|
||||||
|
this.currentDeploymentId = data.latestDeploymentID;
|
||||||
|
this.showBuildLogs = true;
|
||||||
|
this.fetchBuildLogs();
|
||||||
|
}
|
||||||
|
} catch (err) {
|
||||||
|
console.error("Status fetch error:", err);
|
||||||
|
}
|
||||||
|
},
|
||||||
|
|
||||||
|
async fetchContainerLogs() {
|
||||||
|
try {
|
||||||
|
const res = await fetch(`/apps/${this.appId}/container-logs`);
|
||||||
|
const data = await res.json();
|
||||||
|
const newLogs = data.logs || "No logs available";
|
||||||
|
const changed = newLogs !== this.containerLogs;
|
||||||
|
this.containerLogs = newLogs;
|
||||||
|
this.containerStatus = data.status;
|
||||||
|
if (changed && this._containerAutoScroll) {
|
||||||
|
this.$nextTick(() => {
|
||||||
|
Alpine.store("utils").scrollToBottom(this.$refs.containerLogsWrapper);
|
||||||
|
});
|
||||||
|
}
|
||||||
|
} catch (err) {
|
||||||
|
this.containerLogs = "Failed to fetch logs";
|
||||||
|
}
|
||||||
|
},
|
||||||
|
|
||||||
|
async fetchBuildLogs() {
|
||||||
|
if (!this.currentDeploymentId) return;
|
||||||
|
try {
|
||||||
|
const res = await fetch(
|
||||||
|
`/apps/${this.appId}/deployments/${this.currentDeploymentId}/logs`,
|
||||||
|
);
|
||||||
|
const data = await res.json();
|
||||||
|
const newLogs = data.logs || "No build logs available";
|
||||||
|
const changed = newLogs !== this.buildLogs;
|
||||||
|
this.buildLogs = newLogs;
|
||||||
|
this.buildStatus = data.status;
|
||||||
|
if (changed && this._buildAutoScroll) {
|
||||||
|
this.$nextTick(() => {
|
||||||
|
Alpine.store("utils").scrollToBottom(this.$refs.buildLogsWrapper);
|
||||||
|
});
|
||||||
|
}
|
||||||
|
} catch (err) {
|
||||||
|
this.buildLogs = "Failed to fetch logs";
|
||||||
|
}
|
||||||
|
},
|
||||||
|
|
||||||
|
async fetchRecentDeployments() {
|
||||||
|
try {
|
||||||
|
const res = await fetch(`/apps/${this.appId}/recent-deployments`);
|
||||||
|
const data = await res.json();
|
||||||
|
this.deployments = data.deployments || [];
|
||||||
|
} catch (err) {
|
||||||
|
console.error("Deployments fetch error:", err);
|
||||||
|
}
|
||||||
|
},
|
||||||
|
|
||||||
|
submitDeploy() {
|
||||||
|
this.deploying = true;
|
||||||
|
},
|
||||||
|
|
||||||
|
get statusBadgeClass() {
|
||||||
|
return Alpine.store("utils").statusBadgeClass(this.appStatus);
|
||||||
|
},
|
||||||
|
|
||||||
|
get statusLabel() {
|
||||||
|
return Alpine.store("utils").statusLabel(this.appStatus);
|
||||||
|
},
|
||||||
|
|
||||||
|
get containerStatusBadgeClass() {
|
||||||
|
return (
|
||||||
|
Alpine.store("utils").statusBadgeClass(this.containerStatus) +
|
||||||
|
" text-xs"
|
||||||
|
);
|
||||||
|
},
|
||||||
|
|
||||||
|
get containerStatusLabel() {
|
||||||
|
return Alpine.store("utils").statusLabel(this.containerStatus);
|
||||||
|
},
|
||||||
|
|
||||||
|
get buildStatusBadgeClass() {
|
||||||
|
return (
|
||||||
|
Alpine.store("utils").statusBadgeClass(this.buildStatus) + " text-xs"
|
||||||
|
);
|
||||||
|
},
|
||||||
|
|
||||||
|
get buildStatusLabel() {
|
||||||
|
return Alpine.store("utils").statusLabel(this.buildStatus);
|
||||||
|
},
|
||||||
|
|
||||||
|
deploymentStatusClass(status) {
|
||||||
|
return Alpine.store("utils").statusBadgeClass(status);
|
||||||
|
},
|
||||||
|
|
||||||
|
deploymentStatusLabel(status) {
|
||||||
|
return Alpine.store("utils").statusLabel(status);
|
||||||
|
},
|
||||||
|
|
||||||
|
formatTime(isoTime) {
|
||||||
|
return Alpine.store("utils").formatRelativeTime(isoTime);
|
||||||
|
},
|
||||||
|
}));
|
||||||
|
|
||||||
|
// ============================================
|
||||||
|
// Deployment Card Component (for individual deployment cards)
|
||||||
|
// ============================================
|
||||||
|
Alpine.data("deploymentCard", (config) => ({
|
||||||
|
appId: config.appId,
|
||||||
|
deploymentId: config.deploymentId,
|
||||||
|
logs: "",
|
||||||
|
status: config.status || "",
|
||||||
|
pollInterval: null,
|
||||||
|
_autoScroll: true,
|
||||||
|
|
||||||
|
init() {
|
||||||
|
// Read initial logs from script tag (avoids escaping issues)
|
||||||
|
const initialLogsEl = this.$el.querySelector(".initial-logs");
|
||||||
|
this.logs = initialLogsEl?.textContent || "Loading...";
|
||||||
|
|
||||||
|
// Set up scroll tracking
|
||||||
|
this.$nextTick(() => {
|
||||||
|
const wrapper = this.$refs.logsWrapper;
|
||||||
|
if (wrapper) {
|
||||||
|
wrapper.addEventListener('scroll', () => {
|
||||||
|
this._autoScroll = Alpine.store("utils").isScrolledToBottom(wrapper);
|
||||||
|
}, { passive: true });
|
||||||
|
}
|
||||||
|
});
|
||||||
|
|
||||||
|
// Only poll if deployment is in progress
|
||||||
|
if (Alpine.store("utils").isDeploying(this.status)) {
|
||||||
|
this.fetchLogs();
|
||||||
|
this.pollInterval = setInterval(() => this.fetchLogs(), 1000);
|
||||||
|
}
|
||||||
|
},
|
||||||
|
|
||||||
|
destroy() {
|
||||||
|
if (this.pollInterval) {
|
||||||
|
clearInterval(this.pollInterval);
|
||||||
|
}
|
||||||
|
},
|
||||||
|
|
||||||
|
async fetchLogs() {
|
||||||
|
try {
|
||||||
|
const res = await fetch(
|
||||||
|
`/apps/${this.appId}/deployments/${this.deploymentId}/logs`,
|
||||||
|
);
|
||||||
|
const data = await res.json();
|
||||||
|
const newLogs = data.logs || "No logs available";
|
||||||
|
const logsChanged = newLogs !== this.logs;
|
||||||
|
this.logs = newLogs;
|
||||||
|
this.status = data.status;
|
||||||
|
|
||||||
|
// Scroll to bottom only when content changes AND user hasn't scrolled up
|
||||||
|
if (logsChanged && this._autoScroll) {
|
||||||
|
this.$nextTick(() => {
|
||||||
|
Alpine.store("utils").scrollToBottom(this.$refs.logsWrapper);
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
// Stop polling if deployment is done
|
||||||
|
if (!Alpine.store("utils").isDeploying(data.status)) {
|
||||||
|
if (this.pollInterval) {
|
||||||
|
clearInterval(this.pollInterval);
|
||||||
|
this.pollInterval = null;
|
||||||
|
}
|
||||||
|
// Reload page to show final state with duration etc
|
||||||
|
window.location.reload();
|
||||||
|
}
|
||||||
|
} catch (err) {
|
||||||
|
console.error("Logs fetch error:", err);
|
||||||
|
}
|
||||||
|
},
|
||||||
|
|
||||||
|
get statusBadgeClass() {
|
||||||
|
return Alpine.store("utils").statusBadgeClass(this.status);
|
||||||
|
},
|
||||||
|
|
||||||
|
get statusLabel() {
|
||||||
|
return Alpine.store("utils").statusLabel(this.status);
|
||||||
|
},
|
||||||
|
}));
|
||||||
|
|
||||||
|
// ============================================
|
||||||
|
// Deployments History Page Component
|
||||||
|
// ============================================
|
||||||
|
Alpine.data("deploymentsPage", (config) => ({
|
||||||
|
appId: config.appId,
|
||||||
|
currentDeploymentId: null,
|
||||||
|
isDeploying: false,
|
||||||
|
|
||||||
|
init() {
|
||||||
|
// Check for in-progress deployments on page load
|
||||||
|
const inProgressCard = document.querySelector(
|
||||||
|
'[data-status="building"], [data-status="deploying"]',
|
||||||
|
);
|
||||||
|
if (inProgressCard) {
|
||||||
|
this.currentDeploymentId = parseInt(
|
||||||
|
inProgressCard.getAttribute("data-deployment-id"),
|
||||||
|
10,
|
||||||
|
);
|
||||||
|
this.isDeploying = true;
|
||||||
|
}
|
||||||
|
|
||||||
|
this.fetchAppStatus();
|
||||||
|
this._scheduleStatusPoll();
|
||||||
|
},
|
||||||
|
|
||||||
|
_statusPollTimer: null,
|
||||||
|
|
||||||
|
_scheduleStatusPoll() {
|
||||||
|
if (this._statusPollTimer) clearTimeout(this._statusPollTimer);
|
||||||
|
const interval = this.isDeploying ? 1000 : 10000;
|
||||||
|
this._statusPollTimer = setTimeout(() => {
|
||||||
|
this.fetchAppStatus();
|
||||||
|
this._scheduleStatusPoll();
|
||||||
|
}, interval);
|
||||||
|
},
|
||||||
|
|
||||||
|
async fetchAppStatus() {
|
||||||
|
try {
|
||||||
|
const res = await fetch(`/apps/${this.appId}/status`);
|
||||||
|
const data = await res.json();
|
||||||
|
// Use deployment status, not app status - it's more reliable during transitions
|
||||||
|
const deploying = Alpine.store("utils").isDeploying(
|
||||||
|
data.latestDeploymentStatus,
|
||||||
|
);
|
||||||
|
|
||||||
|
// Detect new deployment
|
||||||
|
if (
|
||||||
|
data.latestDeploymentID &&
|
||||||
|
data.latestDeploymentID !== this.currentDeploymentId
|
||||||
|
) {
|
||||||
|
// Check if we have a card for this deployment
|
||||||
|
const hasCard = document.querySelector(
|
||||||
|
`[data-deployment-id="${data.latestDeploymentID}"]`,
|
||||||
|
);
|
||||||
|
|
||||||
|
if (deploying && !hasCard) {
|
||||||
|
// New deployment started but no card exists - reload to show it
|
||||||
|
window.location.reload();
|
||||||
|
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
this.currentDeploymentId = data.latestDeploymentID;
|
||||||
|
}
|
||||||
|
|
||||||
|
// Update deploying state based on latest deployment status
|
||||||
|
if (deploying && !this.isDeploying) {
|
||||||
|
this.isDeploying = true;
|
||||||
|
this._scheduleStatusPoll(); // Switch to fast polling
|
||||||
|
} else if (!deploying && this.isDeploying) {
|
||||||
|
// Deployment finished - reload to show final state
|
||||||
|
this.isDeploying = false;
|
||||||
|
window.location.reload();
|
||||||
|
}
|
||||||
|
} catch (err) {
|
||||||
|
console.error("Status fetch error:", err);
|
||||||
|
}
|
||||||
|
},
|
||||||
|
|
||||||
|
submitDeploy() {
|
||||||
|
this.isDeploying = true;
|
||||||
|
},
|
||||||
|
|
||||||
|
formatTime(isoTime) {
|
||||||
|
return Alpine.store("utils").formatRelativeTime(isoTime);
|
||||||
|
},
|
||||||
|
}));
|
||||||
|
|
||||||
|
// ============================================
|
||||||
|
// Dashboard Page - Relative Time Updates
|
||||||
|
// ============================================
|
||||||
|
Alpine.data("dashboard", () => ({
|
||||||
|
init() {
|
||||||
|
// Update relative times every minute
|
||||||
|
setInterval(() => {
|
||||||
|
this.$el.querySelectorAll("[data-time]").forEach((el) => {
|
||||||
|
const time = el.getAttribute("data-time");
|
||||||
|
if (time) {
|
||||||
|
el.textContent = Alpine.store("utils").formatRelativeTime(time);
|
||||||
|
}
|
||||||
|
});
|
||||||
|
}, 60000);
|
||||||
|
},
|
||||||
|
}));
|
||||||
|
});
|
||||||
|
|
||||||
|
// ============================================
|
||||||
|
// Legacy support - expose utilities globally
|
||||||
|
// ============================================
|
||||||
|
window.upaas = {
|
||||||
|
// These are kept for backwards compatibility but templates should use Alpine.js
|
||||||
|
formatRelativeTime(dateStr) {
|
||||||
|
if (!dateStr) return "";
|
||||||
|
const date = new Date(dateStr);
|
||||||
|
const now = new Date();
|
||||||
|
const diffMs = now - date;
|
||||||
|
const diffSec = Math.floor(diffMs / 1000);
|
||||||
|
const diffMin = Math.floor(diffSec / 60);
|
||||||
|
const diffHour = Math.floor(diffMin / 60);
|
||||||
|
const diffDay = Math.floor(diffHour / 24);
|
||||||
|
|
||||||
|
if (diffSec < 60) return "just now";
|
||||||
|
if (diffMin < 60)
|
||||||
|
return diffMin + (diffMin === 1 ? " minute ago" : " minutes ago");
|
||||||
|
if (diffHour < 24)
|
||||||
|
return diffHour + (diffHour === 1 ? " hour ago" : " hours ago");
|
||||||
|
if (diffDay < 7)
|
||||||
|
return diffDay + (diffDay === 1 ? " day ago" : " days ago");
|
||||||
|
return date.toLocaleDateString();
|
||||||
|
},
|
||||||
|
// Placeholder functions - templates should migrate to Alpine.js
|
||||||
|
initAppDetailPage() {},
|
||||||
|
initDeploymentsPage() {},
|
||||||
|
};
|
||||||
|
|
||||||
|
// Update relative times on page load for non-Alpine elements
|
||||||
|
document.addEventListener("DOMContentLoaded", () => {
|
||||||
|
document.querySelectorAll(".relative-time[data-time]").forEach((el) => {
|
||||||
|
const time = el.getAttribute("data-time");
|
||||||
|
if (time) {
|
||||||
|
el.textContent = window.upaas.formatRelativeTime(time);
|
||||||
|
}
|
||||||
|
});
|
||||||
|
});
|
||||||
@@ -1,71 +0,0 @@
|
|||||||
/**
|
|
||||||
* upaas - Reusable Alpine.js Components
|
|
||||||
*
|
|
||||||
* Small, self-contained components: copy button, confirm dialog,
|
|
||||||
* auto-dismiss alerts, and relative time display.
|
|
||||||
*/
|
|
||||||
|
|
||||||
document.addEventListener("alpine:init", () => {
|
|
||||||
// ============================================
|
|
||||||
// Copy Button Component
|
|
||||||
// ============================================
|
|
||||||
Alpine.data("copyButton", (targetId) => ({
|
|
||||||
copied: false,
|
|
||||||
async copy() {
|
|
||||||
const target = document.getElementById(targetId);
|
|
||||||
if (!target) return;
|
|
||||||
const text = target.textContent || target.value;
|
|
||||||
const success = await Alpine.store("utils").copyToClipboard(text);
|
|
||||||
if (success) {
|
|
||||||
this.copied = true;
|
|
||||||
setTimeout(() => {
|
|
||||||
this.copied = false;
|
|
||||||
}, 2000);
|
|
||||||
}
|
|
||||||
},
|
|
||||||
}));
|
|
||||||
|
|
||||||
// ============================================
|
|
||||||
// Confirm Action Component
|
|
||||||
// ============================================
|
|
||||||
Alpine.data("confirmAction", (message) => ({
|
|
||||||
confirm(event) {
|
|
||||||
if (!window.confirm(message)) {
|
|
||||||
event.preventDefault();
|
|
||||||
}
|
|
||||||
},
|
|
||||||
}));
|
|
||||||
|
|
||||||
// ============================================
|
|
||||||
// Auto-dismiss Alert Component
|
|
||||||
// ============================================
|
|
||||||
Alpine.data("autoDismiss", (delay = 5000) => ({
|
|
||||||
show: true,
|
|
||||||
init() {
|
|
||||||
setTimeout(() => {
|
|
||||||
this.dismiss();
|
|
||||||
}, delay);
|
|
||||||
},
|
|
||||||
dismiss() {
|
|
||||||
this.show = false;
|
|
||||||
setTimeout(() => {
|
|
||||||
this.$el.remove();
|
|
||||||
}, 300);
|
|
||||||
},
|
|
||||||
}));
|
|
||||||
|
|
||||||
// ============================================
|
|
||||||
// Relative Time Component
|
|
||||||
// ============================================
|
|
||||||
Alpine.data("relativeTime", (isoTime) => ({
|
|
||||||
display: "",
|
|
||||||
init() {
|
|
||||||
this.update();
|
|
||||||
// Update every minute
|
|
||||||
setInterval(() => this.update(), 60000);
|
|
||||||
},
|
|
||||||
update() {
|
|
||||||
this.display = Alpine.store("utils").formatRelativeTime(isoTime);
|
|
||||||
},
|
|
||||||
}));
|
|
||||||
});
|
|
||||||
@@ -1,21 +0,0 @@
|
|||||||
/**
|
|
||||||
* upaas - Dashboard Page Component
|
|
||||||
*
|
|
||||||
* Periodically updates relative timestamps on the main dashboard.
|
|
||||||
*/
|
|
||||||
|
|
||||||
document.addEventListener("alpine:init", () => {
|
|
||||||
Alpine.data("dashboard", () => ({
|
|
||||||
init() {
|
|
||||||
// Update relative times every minute
|
|
||||||
setInterval(() => {
|
|
||||||
this.$el.querySelectorAll("[data-time]").forEach((el) => {
|
|
||||||
const time = el.getAttribute("data-time");
|
|
||||||
if (time) {
|
|
||||||
el.textContent = Alpine.store("utils").formatRelativeTime(time);
|
|
||||||
}
|
|
||||||
});
|
|
||||||
}, 60000);
|
|
||||||
},
|
|
||||||
}));
|
|
||||||
});
|
|
||||||
@@ -1,176 +0,0 @@
|
|||||||
/**
|
|
||||||
* upaas - Deployment Components
|
|
||||||
*
|
|
||||||
* Deployment card (individual deployment log viewer) and
|
|
||||||
* deployments history page (list of all deployments).
|
|
||||||
*/
|
|
||||||
|
|
||||||
document.addEventListener("alpine:init", () => {
|
|
||||||
// ============================================
|
|
||||||
// Deployment Card Component (for individual deployment cards)
|
|
||||||
// ============================================
|
|
||||||
Alpine.data("deploymentCard", (config) => ({
|
|
||||||
appId: config.appId,
|
|
||||||
deploymentId: config.deploymentId,
|
|
||||||
logs: "",
|
|
||||||
status: config.status || "",
|
|
||||||
pollInterval: null,
|
|
||||||
_autoScroll: true,
|
|
||||||
|
|
||||||
init() {
|
|
||||||
// Read initial logs from script tag (avoids escaping issues)
|
|
||||||
const initialLogsEl = this.$el.querySelector(".initial-logs");
|
|
||||||
this.logs = initialLogsEl?.dataset.logs || "Loading...";
|
|
||||||
|
|
||||||
// Set up scroll tracking
|
|
||||||
this.$nextTick(() => {
|
|
||||||
const wrapper = this.$refs.logsWrapper;
|
|
||||||
if (wrapper) {
|
|
||||||
wrapper.addEventListener('scroll', () => {
|
|
||||||
this._autoScroll = Alpine.store("utils").isScrolledToBottom(wrapper);
|
|
||||||
}, { passive: true });
|
|
||||||
}
|
|
||||||
});
|
|
||||||
|
|
||||||
// Only poll if deployment is in progress
|
|
||||||
if (Alpine.store("utils").isDeploying(this.status)) {
|
|
||||||
this.fetchLogs();
|
|
||||||
this.pollInterval = setInterval(() => this.fetchLogs(), 1000);
|
|
||||||
}
|
|
||||||
},
|
|
||||||
|
|
||||||
destroy() {
|
|
||||||
if (this.pollInterval) {
|
|
||||||
clearInterval(this.pollInterval);
|
|
||||||
}
|
|
||||||
},
|
|
||||||
|
|
||||||
async fetchLogs() {
|
|
||||||
try {
|
|
||||||
const res = await fetch(
|
|
||||||
`/apps/${this.appId}/deployments/${this.deploymentId}/logs`,
|
|
||||||
);
|
|
||||||
const data = await res.json();
|
|
||||||
const newLogs = data.logs || "No logs available";
|
|
||||||
const logsChanged = newLogs !== this.logs;
|
|
||||||
this.logs = newLogs;
|
|
||||||
this.status = data.status;
|
|
||||||
|
|
||||||
// Scroll to bottom only when content changes AND user hasn't scrolled up
|
|
||||||
if (logsChanged && this._autoScroll) {
|
|
||||||
this.$nextTick(() => {
|
|
||||||
Alpine.store("utils").scrollToBottom(this.$refs.logsWrapper);
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
// Stop polling if deployment is done
|
|
||||||
if (!Alpine.store("utils").isDeploying(data.status)) {
|
|
||||||
if (this.pollInterval) {
|
|
||||||
clearInterval(this.pollInterval);
|
|
||||||
this.pollInterval = null;
|
|
||||||
}
|
|
||||||
// Reload page to show final state with duration etc
|
|
||||||
window.location.reload();
|
|
||||||
}
|
|
||||||
} catch (err) {
|
|
||||||
console.error("Logs fetch error:", err);
|
|
||||||
}
|
|
||||||
},
|
|
||||||
|
|
||||||
get statusBadgeClass() {
|
|
||||||
return Alpine.store("utils").statusBadgeClass(this.status);
|
|
||||||
},
|
|
||||||
|
|
||||||
get statusLabel() {
|
|
||||||
return Alpine.store("utils").statusLabel(this.status);
|
|
||||||
},
|
|
||||||
}));
|
|
||||||
|
|
||||||
// ============================================
|
|
||||||
// Deployments History Page Component
|
|
||||||
// ============================================
|
|
||||||
Alpine.data("deploymentsPage", (config) => ({
|
|
||||||
appId: config.appId,
|
|
||||||
currentDeploymentId: null,
|
|
||||||
isDeploying: false,
|
|
||||||
|
|
||||||
init() {
|
|
||||||
// Check for in-progress deployments on page load
|
|
||||||
const inProgressCard = document.querySelector(
|
|
||||||
'[data-status="building"], [data-status="deploying"]',
|
|
||||||
);
|
|
||||||
if (inProgressCard) {
|
|
||||||
this.currentDeploymentId = parseInt(
|
|
||||||
inProgressCard.getAttribute("data-deployment-id"),
|
|
||||||
10,
|
|
||||||
);
|
|
||||||
this.isDeploying = true;
|
|
||||||
}
|
|
||||||
|
|
||||||
this.fetchAppStatus();
|
|
||||||
this._scheduleStatusPoll();
|
|
||||||
},
|
|
||||||
|
|
||||||
_statusPollTimer: null,
|
|
||||||
|
|
||||||
_scheduleStatusPoll() {
|
|
||||||
if (this._statusPollTimer) clearTimeout(this._statusPollTimer);
|
|
||||||
const interval = this.isDeploying ? 1000 : 10000;
|
|
||||||
this._statusPollTimer = setTimeout(() => {
|
|
||||||
this.fetchAppStatus();
|
|
||||||
this._scheduleStatusPoll();
|
|
||||||
}, interval);
|
|
||||||
},
|
|
||||||
|
|
||||||
async fetchAppStatus() {
|
|
||||||
try {
|
|
||||||
const res = await fetch(`/apps/${this.appId}/status`);
|
|
||||||
const data = await res.json();
|
|
||||||
// Use deployment status, not app status - it's more reliable during transitions
|
|
||||||
const deploying = Alpine.store("utils").isDeploying(
|
|
||||||
data.latestDeploymentStatus,
|
|
||||||
);
|
|
||||||
|
|
||||||
// Detect new deployment
|
|
||||||
if (
|
|
||||||
data.latestDeploymentID &&
|
|
||||||
data.latestDeploymentID !== this.currentDeploymentId
|
|
||||||
) {
|
|
||||||
// Check if we have a card for this deployment
|
|
||||||
const hasCard = document.querySelector(
|
|
||||||
`[data-deployment-id="${data.latestDeploymentID}"]`,
|
|
||||||
);
|
|
||||||
|
|
||||||
if (deploying && !hasCard) {
|
|
||||||
// New deployment started but no card exists - reload to show it
|
|
||||||
window.location.reload();
|
|
||||||
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
|
|
||||||
this.currentDeploymentId = data.latestDeploymentID;
|
|
||||||
}
|
|
||||||
|
|
||||||
// Update deploying state based on latest deployment status
|
|
||||||
if (deploying && !this.isDeploying) {
|
|
||||||
this.isDeploying = true;
|
|
||||||
this._scheduleStatusPoll(); // Switch to fast polling
|
|
||||||
} else if (!deploying && this.isDeploying) {
|
|
||||||
// Deployment finished - reload to show final state
|
|
||||||
this.isDeploying = false;
|
|
||||||
window.location.reload();
|
|
||||||
}
|
|
||||||
} catch (err) {
|
|
||||||
console.error("Status fetch error:", err);
|
|
||||||
}
|
|
||||||
},
|
|
||||||
|
|
||||||
submitDeploy() {
|
|
||||||
this.isDeploying = true;
|
|
||||||
},
|
|
||||||
|
|
||||||
formatTime(isoTime) {
|
|
||||||
return Alpine.store("utils").formatRelativeTime(isoTime);
|
|
||||||
},
|
|
||||||
}));
|
|
||||||
});
|
|
||||||
@@ -1,143 +0,0 @@
|
|||||||
/**
|
|
||||||
* upaas - Global Utilities Store
|
|
||||||
*
|
|
||||||
* Shared formatting, status helpers, and clipboard utilities used across all pages.
|
|
||||||
*/
|
|
||||||
|
|
||||||
document.addEventListener("alpine:init", () => {
|
|
||||||
Alpine.store("utils", {
|
|
||||||
/**
|
|
||||||
* Format a date string as relative time (e.g., "5 minutes ago")
|
|
||||||
*/
|
|
||||||
formatRelativeTime(dateStr) {
|
|
||||||
if (!dateStr) return "";
|
|
||||||
const date = new Date(dateStr);
|
|
||||||
const now = new Date();
|
|
||||||
const diffMs = now - date;
|
|
||||||
const diffSec = Math.floor(diffMs / 1000);
|
|
||||||
const diffMin = Math.floor(diffSec / 60);
|
|
||||||
const diffHour = Math.floor(diffMin / 60);
|
|
||||||
const diffDay = Math.floor(diffHour / 24);
|
|
||||||
|
|
||||||
if (diffSec < 60) return "just now";
|
|
||||||
if (diffMin < 60)
|
|
||||||
return diffMin + (diffMin === 1 ? " minute ago" : " minutes ago");
|
|
||||||
if (diffHour < 24)
|
|
||||||
return diffHour + (diffHour === 1 ? " hour ago" : " hours ago");
|
|
||||||
if (diffDay < 7)
|
|
||||||
return diffDay + (diffDay === 1 ? " day ago" : " days ago");
|
|
||||||
return date.toLocaleDateString();
|
|
||||||
},
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Get the badge class for a given status
|
|
||||||
*/
|
|
||||||
statusBadgeClass(status) {
|
|
||||||
if (status === "running" || status === "success") return "badge-success";
|
|
||||||
if (status === "building" || status === "deploying")
|
|
||||||
return "badge-warning";
|
|
||||||
if (status === "failed" || status === "error") return "badge-error";
|
|
||||||
return "badge-neutral";
|
|
||||||
},
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Format status for display (capitalize first letter)
|
|
||||||
*/
|
|
||||||
statusLabel(status) {
|
|
||||||
if (!status) return "";
|
|
||||||
return status.charAt(0).toUpperCase() + status.slice(1);
|
|
||||||
},
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Check if status indicates active deployment
|
|
||||||
*/
|
|
||||||
isDeploying(status) {
|
|
||||||
return status === "building" || status === "deploying";
|
|
||||||
},
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Scroll an element to the bottom
|
|
||||||
*/
|
|
||||||
scrollToBottom(el) {
|
|
||||||
if (el) {
|
|
||||||
requestAnimationFrame(() => {
|
|
||||||
el.scrollTop = el.scrollHeight;
|
|
||||||
});
|
|
||||||
}
|
|
||||||
},
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Check if a scrollable element is at (or near) the bottom.
|
|
||||||
* Tolerance of 30px accounts for rounding and partial lines.
|
|
||||||
*/
|
|
||||||
isScrolledToBottom(el, tolerance = 30) {
|
|
||||||
if (!el) return true;
|
|
||||||
return el.scrollHeight - el.scrollTop - el.clientHeight <= tolerance;
|
|
||||||
},
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Copy text to clipboard
|
|
||||||
*/
|
|
||||||
async copyToClipboard(text, button) {
|
|
||||||
try {
|
|
||||||
await navigator.clipboard.writeText(text);
|
|
||||||
return true;
|
|
||||||
} catch (err) {
|
|
||||||
// Fallback for older browsers
|
|
||||||
const textArea = document.createElement("textarea");
|
|
||||||
textArea.value = text;
|
|
||||||
textArea.style.position = "fixed";
|
|
||||||
textArea.style.left = "-9999px";
|
|
||||||
document.body.appendChild(textArea);
|
|
||||||
textArea.select();
|
|
||||||
try {
|
|
||||||
document.execCommand("copy");
|
|
||||||
document.body.removeChild(textArea);
|
|
||||||
return true;
|
|
||||||
} catch (e) {
|
|
||||||
document.body.removeChild(textArea);
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
},
|
|
||||||
});
|
|
||||||
});
|
|
||||||
|
|
||||||
// ============================================
|
|
||||||
// Legacy support - expose utilities globally
|
|
||||||
// ============================================
|
|
||||||
window.upaas = {
|
|
||||||
// These are kept for backwards compatibility but templates should use Alpine.js
|
|
||||||
formatRelativeTime(dateStr) {
|
|
||||||
if (!dateStr) return "";
|
|
||||||
const date = new Date(dateStr);
|
|
||||||
const now = new Date();
|
|
||||||
const diffMs = now - date;
|
|
||||||
const diffSec = Math.floor(diffMs / 1000);
|
|
||||||
const diffMin = Math.floor(diffSec / 60);
|
|
||||||
const diffHour = Math.floor(diffMin / 60);
|
|
||||||
const diffDay = Math.floor(diffHour / 24);
|
|
||||||
|
|
||||||
if (diffSec < 60) return "just now";
|
|
||||||
if (diffMin < 60)
|
|
||||||
return diffMin + (diffMin === 1 ? " minute ago" : " minutes ago");
|
|
||||||
if (diffHour < 24)
|
|
||||||
return diffHour + (diffHour === 1 ? " hour ago" : " hours ago");
|
|
||||||
if (diffDay < 7)
|
|
||||||
return diffDay + (diffDay === 1 ? " day ago" : " days ago");
|
|
||||||
return date.toLocaleDateString();
|
|
||||||
},
|
|
||||||
// Placeholder functions - templates should migrate to Alpine.js
|
|
||||||
initAppDetailPage() {},
|
|
||||||
initDeploymentsPage() {},
|
|
||||||
};
|
|
||||||
|
|
||||||
// Update relative times on page load for non-Alpine elements
|
|
||||||
document.addEventListener("DOMContentLoaded", () => {
|
|
||||||
document.querySelectorAll(".relative-time[data-time]").forEach((el) => {
|
|
||||||
const time = el.getAttribute("data-time");
|
|
||||||
if (time) {
|
|
||||||
el.textContent = window.upaas.formatRelativeTime(time);
|
|
||||||
}
|
|
||||||
});
|
|
||||||
});
|
|
||||||
@@ -15,11 +15,7 @@
|
|||||||
</div>
|
</div>
|
||||||
{{template "footer" .}}
|
{{template "footer" .}}
|
||||||
<script defer src="/s/js/alpine.min.js"></script>
|
<script defer src="/s/js/alpine.min.js"></script>
|
||||||
<script src="/s/js/utils.js"></script>
|
<script src="/s/js/app.js"></script>
|
||||||
<script src="/s/js/components.js"></script>
|
|
||||||
<script src="/s/js/app-detail.js"></script>
|
|
||||||
<script src="/s/js/deployment.js"></script>
|
|
||||||
<script src="/s/js/dashboard.js"></script>
|
|
||||||
</body>
|
</body>
|
||||||
</html>
|
</html>
|
||||||
{{end}}
|
{{end}}
|
||||||
|
|||||||
@@ -98,7 +98,7 @@
|
|||||||
title="Scroll to bottom"
|
title="Scroll to bottom"
|
||||||
>↓ Follow</button>
|
>↓ Follow</button>
|
||||||
</div>
|
</div>
|
||||||
{{if .Logs.Valid}}<div hidden class="initial-logs" data-logs="{{.Logs.String}}"></div>{{end}}
|
{{if .Logs.Valid}}<script type="text/plain" class="initial-logs">{{.Logs.String}}</script>{{end}}
|
||||||
</div>
|
</div>
|
||||||
{{end}}
|
{{end}}
|
||||||
</div>
|
</div>
|
||||||
|
|||||||
Reference in New Issue
Block a user