Compare commits
1 Commits
d36e511032
...
remove-whi
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
e241b99d22 |
@@ -96,11 +96,10 @@ expiration 1704067200:
|
|||||||
4. URL:
|
4. URL:
|
||||||
`/v1/image/cdn.example.com/photos/cat.jpg/800x600.webp?sig=<base64url>&exp=1704067200`
|
`/v1/image/cdn.example.com/photos/cat.jpg/800x600.webp?sig=<base64url>&exp=1704067200`
|
||||||
|
|
||||||
**Whitelist patterns:**
|
**Whitelist entries** are exact host matches only (e.g.
|
||||||
|
`cdn.example.com`). Suffix/wildcard matching is not supported —
|
||||||
- **Exact match**: `cdn.example.com` — matches only that host
|
signatures are per-URL, so each allowed host must be listed
|
||||||
- **Suffix match**: `.example.com` — matches `cdn.example.com`,
|
explicitly.
|
||||||
`images.example.com`, and `example.com`
|
|
||||||
|
|
||||||
### Configuration
|
### Configuration
|
||||||
|
|
||||||
|
|||||||
@@ -13,8 +13,7 @@ state_dir: ./data
|
|||||||
# Generate with: openssl rand -base64 32
|
# Generate with: openssl rand -base64 32
|
||||||
signing_key: "CHANGE_ME_generate_with_openssl_rand_base64_32"
|
signing_key: "CHANGE_ME_generate_with_openssl_rand_base64_32"
|
||||||
|
|
||||||
# Hosts that don't require signatures
|
# Hosts that don't require signatures (exact match only)
|
||||||
# Use "." prefix for wildcard subdomain matching (e.g., ".example.com" matches "cdn.example.com")
|
|
||||||
whitelist_hosts:
|
whitelist_hosts:
|
||||||
- s3.sneak.cloud
|
- s3.sneak.cloud
|
||||||
- static.sneak.cloud
|
- static.sneak.cloud
|
||||||
|
|||||||
@@ -35,41 +35,6 @@ type Database struct {
|
|||||||
config *config.Config
|
config *config.Config
|
||||||
}
|
}
|
||||||
|
|
||||||
// ParseMigrationVersion extracts the numeric version prefix from a migration
|
|
||||||
// filename. Filenames must follow the pattern "<version>.sql" or
|
|
||||||
// "<version>_<description>.sql", where version is a zero-padded numeric
|
|
||||||
// string (e.g. "001", "002"). Returns the version string and an error if
|
|
||||||
// the filename does not match the expected pattern.
|
|
||||||
func ParseMigrationVersion(filename string) (string, error) {
|
|
||||||
name := strings.TrimSuffix(filename, filepath.Ext(filename))
|
|
||||||
if name == "" {
|
|
||||||
return "", fmt.Errorf("invalid migration filename %q: empty name", filename)
|
|
||||||
}
|
|
||||||
|
|
||||||
// Split on underscore to separate version from description.
|
|
||||||
// If there's no underscore, the entire stem is the version.
|
|
||||||
version := name
|
|
||||||
if idx := strings.IndexByte(name, '_'); idx >= 0 {
|
|
||||||
version = name[:idx]
|
|
||||||
}
|
|
||||||
|
|
||||||
if version == "" {
|
|
||||||
return "", fmt.Errorf("invalid migration filename %q: empty version prefix", filename)
|
|
||||||
}
|
|
||||||
|
|
||||||
// Validate the version is purely numeric.
|
|
||||||
for _, ch := range version {
|
|
||||||
if ch < '0' || ch > '9' {
|
|
||||||
return "", fmt.Errorf(
|
|
||||||
"invalid migration filename %q: version %q contains non-numeric character %q",
|
|
||||||
filename, version, string(ch),
|
|
||||||
)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
return version, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// New creates a new Database instance.
|
// New creates a new Database instance.
|
||||||
func New(lc fx.Lifecycle, params Params) (*Database, error) {
|
func New(lc fx.Lifecycle, params Params) (*Database, error) {
|
||||||
s := &Database{
|
s := &Database{
|
||||||
@@ -119,33 +84,96 @@ func (s *Database) connect(ctx context.Context) error {
|
|||||||
s.db = db
|
s.db = db
|
||||||
s.log.Info("database connected")
|
s.log.Info("database connected")
|
||||||
|
|
||||||
return ApplyMigrations(ctx, s.db, s.log)
|
return s.runMigrations(ctx)
|
||||||
}
|
}
|
||||||
|
|
||||||
// collectMigrations reads the embedded schema directory and returns
|
func (s *Database) runMigrations(ctx context.Context) error {
|
||||||
// migration filenames sorted lexicographically.
|
// Create migrations tracking table
|
||||||
func collectMigrations() ([]string, error) {
|
_, err := s.db.ExecContext(ctx, `
|
||||||
entries, err := schemaFS.ReadDir("schema")
|
CREATE TABLE IF NOT EXISTS schema_migrations (
|
||||||
|
version TEXT PRIMARY KEY,
|
||||||
|
applied_at DATETIME DEFAULT CURRENT_TIMESTAMP
|
||||||
|
)
|
||||||
|
`)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, fmt.Errorf("failed to read schema directory: %w", err)
|
return fmt.Errorf("failed to create migrations table: %w", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
var migrations []string
|
// Get list of migration files
|
||||||
|
entries, err := schemaFS.ReadDir("schema")
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Errorf("failed to read schema directory: %w", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Sort migration files by name (001.sql, 002.sql, etc.)
|
||||||
|
var migrations []string
|
||||||
for _, entry := range entries {
|
for _, entry := range entries {
|
||||||
if !entry.IsDir() && strings.HasSuffix(entry.Name(), ".sql") {
|
if !entry.IsDir() && strings.HasSuffix(entry.Name(), ".sql") {
|
||||||
migrations = append(migrations, entry.Name())
|
migrations = append(migrations, entry.Name())
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
sort.Strings(migrations)
|
sort.Strings(migrations)
|
||||||
|
|
||||||
return migrations, nil
|
// Apply each migration that hasn't been applied yet
|
||||||
|
for _, migration := range migrations {
|
||||||
|
version := strings.TrimSuffix(migration, filepath.Ext(migration))
|
||||||
|
|
||||||
|
// Check if already applied
|
||||||
|
var count int
|
||||||
|
err := s.db.QueryRowContext(ctx,
|
||||||
|
"SELECT COUNT(*) FROM schema_migrations WHERE version = ?",
|
||||||
|
version,
|
||||||
|
).Scan(&count)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Errorf("failed to check migration status: %w", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
if count > 0 {
|
||||||
|
s.log.Debug("migration already applied", "version", version)
|
||||||
|
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
|
// Read and apply migration
|
||||||
|
content, err := schemaFS.ReadFile(filepath.Join("schema", migration))
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Errorf("failed to read migration %s: %w", migration, err)
|
||||||
|
}
|
||||||
|
|
||||||
|
s.log.Info("applying migration", "version", version)
|
||||||
|
|
||||||
|
_, err = s.db.ExecContext(ctx, string(content))
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Errorf("failed to apply migration %s: %w", migration, err)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Record migration as applied
|
||||||
|
_, err = s.db.ExecContext(ctx,
|
||||||
|
"INSERT INTO schema_migrations (version) VALUES (?)",
|
||||||
|
version,
|
||||||
|
)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Errorf("failed to record migration %s: %w", migration, err)
|
||||||
|
}
|
||||||
|
|
||||||
|
s.log.Info("migration applied successfully", "version", version)
|
||||||
|
}
|
||||||
|
|
||||||
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// ensureMigrationsTable creates the schema_migrations tracking table if
|
// DB returns the underlying sql.DB.
|
||||||
// it does not already exist.
|
func (s *Database) DB() *sql.DB {
|
||||||
func ensureMigrationsTable(ctx context.Context, db *sql.DB) error {
|
return s.db
|
||||||
|
}
|
||||||
|
|
||||||
|
// ApplyMigrations applies all migrations to the given database.
|
||||||
|
// This is useful for testing where you want to use the real schema
|
||||||
|
// without the full fx lifecycle.
|
||||||
|
func ApplyMigrations(db *sql.DB) error {
|
||||||
|
ctx := context.Background()
|
||||||
|
|
||||||
|
// Create migrations tracking table
|
||||||
_, err := db.ExecContext(ctx, `
|
_, err := db.ExecContext(ctx, `
|
||||||
CREATE TABLE IF NOT EXISTS schema_migrations (
|
CREATE TABLE IF NOT EXISTS schema_migrations (
|
||||||
version TEXT PRIMARY KEY,
|
version TEXT PRIMARY KEY,
|
||||||
@@ -156,32 +184,27 @@ func ensureMigrationsTable(ctx context.Context, db *sql.DB) error {
|
|||||||
return fmt.Errorf("failed to create migrations table: %w", err)
|
return fmt.Errorf("failed to create migrations table: %w", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
return nil
|
// Get list of migration files
|
||||||
}
|
entries, err := schemaFS.ReadDir("schema")
|
||||||
|
|
||||||
// ApplyMigrations applies all pending migrations to db. An optional logger
|
|
||||||
// may be provided for informational output; pass nil for silent operation.
|
|
||||||
// This is exported so tests can apply the real schema without the full fx
|
|
||||||
// lifecycle.
|
|
||||||
func ApplyMigrations(ctx context.Context, db *sql.DB, log *slog.Logger) error {
|
|
||||||
if err := ensureMigrationsTable(ctx, db); err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
|
|
||||||
migrations, err := collectMigrations()
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return fmt.Errorf("failed to read schema directory: %w", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
for _, migration := range migrations {
|
// Sort migration files by name (001.sql, 002.sql, etc.)
|
||||||
version, parseErr := ParseMigrationVersion(migration)
|
var migrations []string
|
||||||
if parseErr != nil {
|
for _, entry := range entries {
|
||||||
return parseErr
|
if !entry.IsDir() && strings.HasSuffix(entry.Name(), ".sql") {
|
||||||
|
migrations = append(migrations, entry.Name())
|
||||||
}
|
}
|
||||||
|
}
|
||||||
|
sort.Strings(migrations)
|
||||||
|
|
||||||
// Check if already applied.
|
// Apply each migration that hasn't been applied yet
|
||||||
|
for _, migration := range migrations {
|
||||||
|
version := strings.TrimSuffix(migration, filepath.Ext(migration))
|
||||||
|
|
||||||
|
// Check if already applied
|
||||||
var count int
|
var count int
|
||||||
|
|
||||||
err := db.QueryRowContext(ctx,
|
err := db.QueryRowContext(ctx,
|
||||||
"SELECT COUNT(*) FROM schema_migrations WHERE version = ?",
|
"SELECT COUNT(*) FROM schema_migrations WHERE version = ?",
|
||||||
version,
|
version,
|
||||||
@@ -191,46 +214,29 @@ func ApplyMigrations(ctx context.Context, db *sql.DB, log *slog.Logger) error {
|
|||||||
}
|
}
|
||||||
|
|
||||||
if count > 0 {
|
if count > 0 {
|
||||||
if log != nil {
|
|
||||||
log.Debug("migration already applied", "version", version)
|
|
||||||
}
|
|
||||||
|
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
|
|
||||||
// Read and apply migration.
|
// Read and apply migration
|
||||||
content, readErr := schemaFS.ReadFile(filepath.Join("schema", migration))
|
content, err := schemaFS.ReadFile(filepath.Join("schema", migration))
|
||||||
if readErr != nil {
|
if err != nil {
|
||||||
return fmt.Errorf("failed to read migration %s: %w", migration, readErr)
|
return fmt.Errorf("failed to read migration %s: %w", migration, err)
|
||||||
}
|
}
|
||||||
|
|
||||||
if log != nil {
|
_, err = db.ExecContext(ctx, string(content))
|
||||||
log.Info("applying migration", "version", version)
|
if err != nil {
|
||||||
|
return fmt.Errorf("failed to apply migration %s: %w", migration, err)
|
||||||
}
|
}
|
||||||
|
|
||||||
_, execErr := db.ExecContext(ctx, string(content))
|
// Record migration as applied
|
||||||
if execErr != nil {
|
_, err = db.ExecContext(ctx,
|
||||||
return fmt.Errorf("failed to apply migration %s: %w", migration, execErr)
|
|
||||||
}
|
|
||||||
|
|
||||||
// Record migration as applied.
|
|
||||||
_, recErr := db.ExecContext(ctx,
|
|
||||||
"INSERT INTO schema_migrations (version) VALUES (?)",
|
"INSERT INTO schema_migrations (version) VALUES (?)",
|
||||||
version,
|
version,
|
||||||
)
|
)
|
||||||
if recErr != nil {
|
if err != nil {
|
||||||
return fmt.Errorf("failed to record migration %s: %w", migration, recErr)
|
return fmt.Errorf("failed to record migration %s: %w", migration, err)
|
||||||
}
|
|
||||||
|
|
||||||
if log != nil {
|
|
||||||
log.Info("migration applied successfully", "version", version)
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// DB returns the underlying sql.DB.
|
|
||||||
func (s *Database) DB() *sql.DB {
|
|
||||||
return s.db
|
|
||||||
}
|
|
||||||
|
|||||||
@@ -1,155 +0,0 @@
|
|||||||
package database
|
|
||||||
|
|
||||||
import (
|
|
||||||
"context"
|
|
||||||
"database/sql"
|
|
||||||
"testing"
|
|
||||||
|
|
||||||
_ "modernc.org/sqlite" // SQLite driver registration
|
|
||||||
)
|
|
||||||
|
|
||||||
func TestParseMigrationVersion(t *testing.T) {
|
|
||||||
tests := []struct {
|
|
||||||
name string
|
|
||||||
filename string
|
|
||||||
want string
|
|
||||||
wantErr bool
|
|
||||||
}{
|
|
||||||
{
|
|
||||||
name: "version only",
|
|
||||||
filename: "001.sql",
|
|
||||||
want: "001",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "version with description",
|
|
||||||
filename: "001_initial_schema.sql",
|
|
||||||
want: "001",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "multi-digit version",
|
|
||||||
filename: "042_add_indexes.sql",
|
|
||||||
want: "042",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "long version number",
|
|
||||||
filename: "00001_long_prefix.sql",
|
|
||||||
want: "00001",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "description with multiple underscores",
|
|
||||||
filename: "003_add_user_auth_tables.sql",
|
|
||||||
want: "003",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "empty filename",
|
|
||||||
filename: ".sql",
|
|
||||||
wantErr: true,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "leading underscore",
|
|
||||||
filename: "_description.sql",
|
|
||||||
wantErr: true,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "non-numeric version",
|
|
||||||
filename: "abc_migration.sql",
|
|
||||||
wantErr: true,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "mixed alphanumeric version",
|
|
||||||
filename: "001a_migration.sql",
|
|
||||||
wantErr: true,
|
|
||||||
},
|
|
||||||
}
|
|
||||||
|
|
||||||
for _, tt := range tests {
|
|
||||||
t.Run(tt.name, func(t *testing.T) {
|
|
||||||
got, err := ParseMigrationVersion(tt.filename)
|
|
||||||
if tt.wantErr {
|
|
||||||
if err == nil {
|
|
||||||
t.Errorf("ParseMigrationVersion(%q) expected error, got %q", tt.filename, got)
|
|
||||||
}
|
|
||||||
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
if err != nil {
|
|
||||||
t.Errorf("ParseMigrationVersion(%q) unexpected error: %v", tt.filename, err)
|
|
||||||
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
if got != tt.want {
|
|
||||||
t.Errorf("ParseMigrationVersion(%q) = %q, want %q", tt.filename, got, tt.want)
|
|
||||||
}
|
|
||||||
})
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestApplyMigrations(t *testing.T) {
|
|
||||||
db, err := sql.Open("sqlite", ":memory:")
|
|
||||||
if err != nil {
|
|
||||||
t.Fatalf("failed to open in-memory database: %v", err)
|
|
||||||
}
|
|
||||||
defer db.Close()
|
|
||||||
|
|
||||||
// Apply migrations should succeed.
|
|
||||||
if err := ApplyMigrations(context.Background(), db, nil); err != nil {
|
|
||||||
t.Fatalf("ApplyMigrations failed: %v", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
// Verify the schema_migrations table recorded the version.
|
|
||||||
var version string
|
|
||||||
|
|
||||||
err = db.QueryRowContext(context.Background(),
|
|
||||||
"SELECT version FROM schema_migrations LIMIT 1",
|
|
||||||
).Scan(&version)
|
|
||||||
if err != nil {
|
|
||||||
t.Fatalf("failed to query schema_migrations: %v", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
if version != "001" {
|
|
||||||
t.Errorf("expected version %q, got %q", "001", version)
|
|
||||||
}
|
|
||||||
|
|
||||||
// Verify a table from the migration exists (source_content).
|
|
||||||
var tableName string
|
|
||||||
|
|
||||||
err = db.QueryRowContext(context.Background(),
|
|
||||||
"SELECT name FROM sqlite_master WHERE type='table' AND name='source_content'",
|
|
||||||
).Scan(&tableName)
|
|
||||||
if err != nil {
|
|
||||||
t.Fatalf("expected source_content table to exist: %v", err)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestApplyMigrationsIdempotent(t *testing.T) {
|
|
||||||
db, err := sql.Open("sqlite", ":memory:")
|
|
||||||
if err != nil {
|
|
||||||
t.Fatalf("failed to open in-memory database: %v", err)
|
|
||||||
}
|
|
||||||
defer db.Close()
|
|
||||||
|
|
||||||
// Apply twice should succeed (idempotent).
|
|
||||||
if err := ApplyMigrations(context.Background(), db, nil); err != nil {
|
|
||||||
t.Fatalf("first ApplyMigrations failed: %v", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
if err := ApplyMigrations(context.Background(), db, nil); err != nil {
|
|
||||||
t.Fatalf("second ApplyMigrations failed: %v", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
// Should still have exactly one migration recorded.
|
|
||||||
var count int
|
|
||||||
|
|
||||||
err = db.QueryRowContext(context.Background(),
|
|
||||||
"SELECT COUNT(*) FROM schema_migrations",
|
|
||||||
).Scan(&count)
|
|
||||||
if err != nil {
|
|
||||||
t.Fatalf("failed to count schema_migrations: %v", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
if count != 1 {
|
|
||||||
t.Errorf("expected 1 migration record, got %d", count)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
@@ -82,7 +82,7 @@ func setupTestDB(t *testing.T) *sql.DB {
|
|||||||
t.Fatalf("failed to open test db: %v", err)
|
t.Fatalf("failed to open test db: %v", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
if err := database.ApplyMigrations(context.Background(), db, nil); err != nil {
|
if err := database.ApplyMigrations(db); err != nil {
|
||||||
t.Fatalf("failed to apply migrations: %v", err)
|
t.Fatalf("failed to apply migrations: %v", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -26,45 +26,20 @@ func initVips() {
|
|||||||
// Images larger than this are rejected to prevent DoS via decompression bombs.
|
// Images larger than this are rejected to prevent DoS via decompression bombs.
|
||||||
const MaxInputDimension = 8192
|
const MaxInputDimension = 8192
|
||||||
|
|
||||||
// DefaultMaxInputBytes is the default maximum input size in bytes (50 MiB).
|
|
||||||
// This matches the default upstream fetcher limit.
|
|
||||||
const DefaultMaxInputBytes = 50 << 20
|
|
||||||
|
|
||||||
// ErrInputTooLarge is returned when input image dimensions exceed MaxInputDimension.
|
// ErrInputTooLarge is returned when input image dimensions exceed MaxInputDimension.
|
||||||
var ErrInputTooLarge = errors.New("input image dimensions exceed maximum")
|
var ErrInputTooLarge = errors.New("input image dimensions exceed maximum")
|
||||||
|
|
||||||
// ErrInputDataTooLarge is returned when the raw input data exceeds the configured byte limit.
|
|
||||||
var ErrInputDataTooLarge = errors.New("input data exceeds maximum allowed size")
|
|
||||||
|
|
||||||
// ErrUnsupportedOutputFormat is returned when the requested output format is not supported.
|
// ErrUnsupportedOutputFormat is returned when the requested output format is not supported.
|
||||||
var ErrUnsupportedOutputFormat = errors.New("unsupported output format")
|
var ErrUnsupportedOutputFormat = errors.New("unsupported output format")
|
||||||
|
|
||||||
// ImageProcessor implements the Processor interface using libvips via govips.
|
// ImageProcessor implements the Processor interface using libvips via govips.
|
||||||
type ImageProcessor struct {
|
type ImageProcessor struct{}
|
||||||
maxInputBytes int64
|
|
||||||
}
|
|
||||||
|
|
||||||
// Params holds configuration for creating an ImageProcessor.
|
// NewImageProcessor creates a new image processor.
|
||||||
// Zero values use sensible defaults (MaxInputBytes defaults to DefaultMaxInputBytes).
|
func NewImageProcessor() *ImageProcessor {
|
||||||
type Params struct {
|
|
||||||
// MaxInputBytes is the maximum allowed input size in bytes.
|
|
||||||
// If <= 0, DefaultMaxInputBytes is used.
|
|
||||||
MaxInputBytes int64
|
|
||||||
}
|
|
||||||
|
|
||||||
// New creates a new image processor with the given parameters.
|
|
||||||
// A zero-value Params{} uses sensible defaults.
|
|
||||||
func New(params Params) *ImageProcessor {
|
|
||||||
initVips()
|
initVips()
|
||||||
|
|
||||||
maxInputBytes := params.MaxInputBytes
|
return &ImageProcessor{}
|
||||||
if maxInputBytes <= 0 {
|
|
||||||
maxInputBytes = DefaultMaxInputBytes
|
|
||||||
}
|
|
||||||
|
|
||||||
return &ImageProcessor{
|
|
||||||
maxInputBytes: maxInputBytes,
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// Process transforms an image according to the request.
|
// Process transforms an image according to the request.
|
||||||
@@ -73,20 +48,12 @@ func (p *ImageProcessor) Process(
|
|||||||
input io.Reader,
|
input io.Reader,
|
||||||
req *ImageRequest,
|
req *ImageRequest,
|
||||||
) (*ProcessResult, error) {
|
) (*ProcessResult, error) {
|
||||||
// Read input with a size limit to prevent unbounded memory consumption.
|
// Read input
|
||||||
// We read at most maxInputBytes+1 so we can detect if the input exceeds
|
data, err := io.ReadAll(input)
|
||||||
// the limit without consuming additional memory.
|
|
||||||
limited := io.LimitReader(input, p.maxInputBytes+1)
|
|
||||||
|
|
||||||
data, err := io.ReadAll(limited)
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, fmt.Errorf("failed to read input: %w", err)
|
return nil, fmt.Errorf("failed to read input: %w", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
if int64(len(data)) > p.maxInputBytes {
|
|
||||||
return nil, ErrInputDataTooLarge
|
|
||||||
}
|
|
||||||
|
|
||||||
// Decode image
|
// Decode image
|
||||||
img, err := vips.NewImageFromBuffer(data)
|
img, err := vips.NewImageFromBuffer(data)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
|||||||
@@ -71,7 +71,7 @@ func createTestPNG(t *testing.T, width, height int) []byte {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func TestImageProcessor_ResizeJPEG(t *testing.T) {
|
func TestImageProcessor_ResizeJPEG(t *testing.T) {
|
||||||
proc := New(Params{})
|
proc := NewImageProcessor()
|
||||||
ctx := context.Background()
|
ctx := context.Background()
|
||||||
|
|
||||||
input := createTestJPEG(t, 800, 600)
|
input := createTestJPEG(t, 800, 600)
|
||||||
@@ -118,7 +118,7 @@ func TestImageProcessor_ResizeJPEG(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func TestImageProcessor_ConvertToPNG(t *testing.T) {
|
func TestImageProcessor_ConvertToPNG(t *testing.T) {
|
||||||
proc := New(Params{})
|
proc := NewImageProcessor()
|
||||||
ctx := context.Background()
|
ctx := context.Background()
|
||||||
|
|
||||||
input := createTestJPEG(t, 200, 150)
|
input := createTestJPEG(t, 200, 150)
|
||||||
@@ -151,7 +151,7 @@ func TestImageProcessor_ConvertToPNG(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func TestImageProcessor_OriginalSize(t *testing.T) {
|
func TestImageProcessor_OriginalSize(t *testing.T) {
|
||||||
proc := New(Params{})
|
proc := NewImageProcessor()
|
||||||
ctx := context.Background()
|
ctx := context.Background()
|
||||||
|
|
||||||
input := createTestJPEG(t, 640, 480)
|
input := createTestJPEG(t, 640, 480)
|
||||||
@@ -179,7 +179,7 @@ func TestImageProcessor_OriginalSize(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func TestImageProcessor_FitContain(t *testing.T) {
|
func TestImageProcessor_FitContain(t *testing.T) {
|
||||||
proc := New(Params{})
|
proc := NewImageProcessor()
|
||||||
ctx := context.Background()
|
ctx := context.Background()
|
||||||
|
|
||||||
// 800x400 image (2:1 aspect) into 400x400 box with contain
|
// 800x400 image (2:1 aspect) into 400x400 box with contain
|
||||||
@@ -206,7 +206,7 @@ func TestImageProcessor_FitContain(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func TestImageProcessor_ProportionalScale_WidthOnly(t *testing.T) {
|
func TestImageProcessor_ProportionalScale_WidthOnly(t *testing.T) {
|
||||||
proc := New(Params{})
|
proc := NewImageProcessor()
|
||||||
ctx := context.Background()
|
ctx := context.Background()
|
||||||
|
|
||||||
// 800x600 image, request width=400 height=0
|
// 800x600 image, request width=400 height=0
|
||||||
@@ -236,7 +236,7 @@ func TestImageProcessor_ProportionalScale_WidthOnly(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func TestImageProcessor_ProportionalScale_HeightOnly(t *testing.T) {
|
func TestImageProcessor_ProportionalScale_HeightOnly(t *testing.T) {
|
||||||
proc := New(Params{})
|
proc := NewImageProcessor()
|
||||||
ctx := context.Background()
|
ctx := context.Background()
|
||||||
|
|
||||||
// 800x600 image, request width=0 height=300
|
// 800x600 image, request width=0 height=300
|
||||||
@@ -266,7 +266,7 @@ func TestImageProcessor_ProportionalScale_HeightOnly(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func TestImageProcessor_ProcessPNG(t *testing.T) {
|
func TestImageProcessor_ProcessPNG(t *testing.T) {
|
||||||
proc := New(Params{})
|
proc := NewImageProcessor()
|
||||||
ctx := context.Background()
|
ctx := context.Background()
|
||||||
|
|
||||||
input := createTestPNG(t, 400, 300)
|
input := createTestPNG(t, 400, 300)
|
||||||
@@ -298,7 +298,7 @@ func TestImageProcessor_ImplementsInterface(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func TestImageProcessor_SupportedFormats(t *testing.T) {
|
func TestImageProcessor_SupportedFormats(t *testing.T) {
|
||||||
proc := New(Params{})
|
proc := NewImageProcessor()
|
||||||
|
|
||||||
inputFormats := proc.SupportedInputFormats()
|
inputFormats := proc.SupportedInputFormats()
|
||||||
if len(inputFormats) == 0 {
|
if len(inputFormats) == 0 {
|
||||||
@@ -312,7 +312,7 @@ func TestImageProcessor_SupportedFormats(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func TestImageProcessor_RejectsOversizedInput(t *testing.T) {
|
func TestImageProcessor_RejectsOversizedInput(t *testing.T) {
|
||||||
proc := New(Params{})
|
proc := NewImageProcessor()
|
||||||
ctx := context.Background()
|
ctx := context.Background()
|
||||||
|
|
||||||
// Create an image that exceeds MaxInputDimension (e.g., 10000x100)
|
// Create an image that exceeds MaxInputDimension (e.g., 10000x100)
|
||||||
@@ -337,7 +337,7 @@ func TestImageProcessor_RejectsOversizedInput(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func TestImageProcessor_RejectsOversizedInputHeight(t *testing.T) {
|
func TestImageProcessor_RejectsOversizedInputHeight(t *testing.T) {
|
||||||
proc := New(Params{})
|
proc := NewImageProcessor()
|
||||||
ctx := context.Background()
|
ctx := context.Background()
|
||||||
|
|
||||||
// Create an image with oversized height
|
// Create an image with oversized height
|
||||||
@@ -361,7 +361,7 @@ func TestImageProcessor_RejectsOversizedInputHeight(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func TestImageProcessor_AcceptsMaxDimensionInput(t *testing.T) {
|
func TestImageProcessor_AcceptsMaxDimensionInput(t *testing.T) {
|
||||||
proc := New(Params{})
|
proc := NewImageProcessor()
|
||||||
ctx := context.Background()
|
ctx := context.Background()
|
||||||
|
|
||||||
// Create an image at exactly MaxInputDimension - should be accepted
|
// Create an image at exactly MaxInputDimension - should be accepted
|
||||||
@@ -383,7 +383,7 @@ func TestImageProcessor_AcceptsMaxDimensionInput(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func TestImageProcessor_EncodeWebP(t *testing.T) {
|
func TestImageProcessor_EncodeWebP(t *testing.T) {
|
||||||
proc := New(Params{})
|
proc := NewImageProcessor()
|
||||||
ctx := context.Background()
|
ctx := context.Background()
|
||||||
|
|
||||||
input := createTestJPEG(t, 200, 150)
|
input := createTestJPEG(t, 200, 150)
|
||||||
@@ -426,7 +426,7 @@ func TestImageProcessor_EncodeWebP(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func TestImageProcessor_DecodeAVIF(t *testing.T) {
|
func TestImageProcessor_DecodeAVIF(t *testing.T) {
|
||||||
proc := New(Params{})
|
proc := NewImageProcessor()
|
||||||
ctx := context.Background()
|
ctx := context.Background()
|
||||||
|
|
||||||
// Load test AVIF file
|
// Load test AVIF file
|
||||||
@@ -465,73 +465,8 @@ func TestImageProcessor_DecodeAVIF(t *testing.T) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestImageProcessor_RejectsOversizedInputData(t *testing.T) {
|
|
||||||
// Create a processor with a very small byte limit
|
|
||||||
const limit = 1024
|
|
||||||
proc := New(Params{MaxInputBytes: limit})
|
|
||||||
ctx := context.Background()
|
|
||||||
|
|
||||||
// Create a valid JPEG that exceeds the byte limit
|
|
||||||
input := createTestJPEG(t, 800, 600) // will be well over 1 KiB
|
|
||||||
if int64(len(input)) <= limit {
|
|
||||||
t.Fatalf("test JPEG must exceed %d bytes, got %d", limit, len(input))
|
|
||||||
}
|
|
||||||
|
|
||||||
req := &ImageRequest{
|
|
||||||
Size: Size{Width: 100, Height: 75},
|
|
||||||
Format: FormatJPEG,
|
|
||||||
Quality: 85,
|
|
||||||
FitMode: FitCover,
|
|
||||||
}
|
|
||||||
|
|
||||||
_, err := proc.Process(ctx, bytes.NewReader(input), req)
|
|
||||||
if err == nil {
|
|
||||||
t.Fatal("Process() should reject input exceeding maxInputBytes")
|
|
||||||
}
|
|
||||||
|
|
||||||
if err != ErrInputDataTooLarge {
|
|
||||||
t.Errorf("Process() error = %v, want ErrInputDataTooLarge", err)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestImageProcessor_AcceptsInputWithinLimit(t *testing.T) {
|
|
||||||
// Create a small image and set limit well above its size
|
|
||||||
input := createTestJPEG(t, 10, 10)
|
|
||||||
limit := int64(len(input)) * 10 // 10× headroom
|
|
||||||
|
|
||||||
proc := New(Params{MaxInputBytes: limit})
|
|
||||||
ctx := context.Background()
|
|
||||||
|
|
||||||
req := &ImageRequest{
|
|
||||||
Size: Size{Width: 10, Height: 10},
|
|
||||||
Format: FormatJPEG,
|
|
||||||
Quality: 85,
|
|
||||||
FitMode: FitCover,
|
|
||||||
}
|
|
||||||
|
|
||||||
result, err := proc.Process(ctx, bytes.NewReader(input), req)
|
|
||||||
if err != nil {
|
|
||||||
t.Fatalf("Process() error = %v, want nil", err)
|
|
||||||
}
|
|
||||||
defer result.Content.Close()
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestImageProcessor_DefaultMaxInputBytes(t *testing.T) {
|
|
||||||
// Passing 0 should use the default
|
|
||||||
proc := New(Params{})
|
|
||||||
if proc.maxInputBytes != DefaultMaxInputBytes {
|
|
||||||
t.Errorf("maxInputBytes = %d, want %d", proc.maxInputBytes, DefaultMaxInputBytes)
|
|
||||||
}
|
|
||||||
|
|
||||||
// Passing negative should also use the default
|
|
||||||
proc = New(Params{MaxInputBytes: -1})
|
|
||||||
if proc.maxInputBytes != DefaultMaxInputBytes {
|
|
||||||
t.Errorf("maxInputBytes = %d, want %d", proc.maxInputBytes, DefaultMaxInputBytes)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestImageProcessor_EncodeAVIF(t *testing.T) {
|
func TestImageProcessor_EncodeAVIF(t *testing.T) {
|
||||||
proc := New(Params{})
|
proc := NewImageProcessor()
|
||||||
ctx := context.Background()
|
ctx := context.Background()
|
||||||
|
|
||||||
input := createTestJPEG(t, 200, 150)
|
input := createTestJPEG(t, 200, 150)
|
||||||
|
|||||||
@@ -15,14 +15,13 @@ import (
|
|||||||
|
|
||||||
// Service implements the ImageCache interface, orchestrating cache, fetcher, and processor.
|
// Service implements the ImageCache interface, orchestrating cache, fetcher, and processor.
|
||||||
type Service struct {
|
type Service struct {
|
||||||
cache *Cache
|
cache *Cache
|
||||||
fetcher Fetcher
|
fetcher Fetcher
|
||||||
processor Processor
|
processor Processor
|
||||||
signer *Signer
|
signer *Signer
|
||||||
whitelist *HostWhitelist
|
whitelist *HostWhitelist
|
||||||
log *slog.Logger
|
log *slog.Logger
|
||||||
allowHTTP bool
|
allowHTTP bool
|
||||||
maxResponseSize int64
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// ServiceConfig holds configuration for the image service.
|
// ServiceConfig holds configuration for the image service.
|
||||||
@@ -51,17 +50,15 @@ func NewService(cfg *ServiceConfig) (*Service, error) {
|
|||||||
return nil, errors.New("signing key is required")
|
return nil, errors.New("signing key is required")
|
||||||
}
|
}
|
||||||
|
|
||||||
// Resolve fetcher config for defaults
|
|
||||||
fetcherCfg := cfg.FetcherConfig
|
|
||||||
if fetcherCfg == nil {
|
|
||||||
fetcherCfg = DefaultFetcherConfig()
|
|
||||||
}
|
|
||||||
|
|
||||||
// Use custom fetcher if provided, otherwise create HTTP fetcher
|
// Use custom fetcher if provided, otherwise create HTTP fetcher
|
||||||
var fetcher Fetcher
|
var fetcher Fetcher
|
||||||
if cfg.Fetcher != nil {
|
if cfg.Fetcher != nil {
|
||||||
fetcher = cfg.Fetcher
|
fetcher = cfg.Fetcher
|
||||||
} else {
|
} else {
|
||||||
|
fetcherCfg := cfg.FetcherConfig
|
||||||
|
if fetcherCfg == nil {
|
||||||
|
fetcherCfg = DefaultFetcherConfig()
|
||||||
|
}
|
||||||
fetcher = NewHTTPFetcher(fetcherCfg)
|
fetcher = NewHTTPFetcher(fetcherCfg)
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -77,17 +74,14 @@ func NewService(cfg *ServiceConfig) (*Service, error) {
|
|||||||
allowHTTP = cfg.FetcherConfig.AllowHTTP
|
allowHTTP = cfg.FetcherConfig.AllowHTTP
|
||||||
}
|
}
|
||||||
|
|
||||||
maxResponseSize := fetcherCfg.MaxResponseSize
|
|
||||||
|
|
||||||
return &Service{
|
return &Service{
|
||||||
cache: cfg.Cache,
|
cache: cfg.Cache,
|
||||||
fetcher: fetcher,
|
fetcher: fetcher,
|
||||||
processor: New(Params{MaxInputBytes: maxResponseSize}),
|
processor: NewImageProcessor(),
|
||||||
signer: signer,
|
signer: signer,
|
||||||
whitelist: NewHostWhitelist(cfg.Whitelist),
|
whitelist: NewHostWhitelist(cfg.Whitelist),
|
||||||
log: log,
|
log: log,
|
||||||
allowHTTP: allowHTTP,
|
allowHTTP: allowHTTP,
|
||||||
maxResponseSize: maxResponseSize,
|
|
||||||
}, nil
|
}, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -152,40 +146,6 @@ func (s *Service) Get(ctx context.Context, req *ImageRequest) (*ImageResponse, e
|
|||||||
return response, nil
|
return response, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// loadCachedSource attempts to load source content from cache, returning nil
|
|
||||||
// if the cached data is unavailable or exceeds maxResponseSize.
|
|
||||||
func (s *Service) loadCachedSource(contentHash ContentHash) []byte {
|
|
||||||
reader, err := s.cache.GetSourceContent(contentHash)
|
|
||||||
if err != nil {
|
|
||||||
s.log.Warn("failed to load cached source, fetching", "error", err)
|
|
||||||
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// Bound the read to maxResponseSize to prevent unbounded memory use
|
|
||||||
// from unexpectedly large cached files.
|
|
||||||
limited := io.LimitReader(reader, s.maxResponseSize+1)
|
|
||||||
data, err := io.ReadAll(limited)
|
|
||||||
_ = reader.Close()
|
|
||||||
|
|
||||||
if err != nil {
|
|
||||||
s.log.Warn("failed to read cached source, fetching", "error", err)
|
|
||||||
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
if int64(len(data)) > s.maxResponseSize {
|
|
||||||
s.log.Warn("cached source exceeds max response size, discarding",
|
|
||||||
"hash", contentHash,
|
|
||||||
"max_bytes", s.maxResponseSize,
|
|
||||||
)
|
|
||||||
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
return data
|
|
||||||
}
|
|
||||||
|
|
||||||
// processFromSourceOrFetch processes an image, using cached source content if available.
|
// processFromSourceOrFetch processes an image, using cached source content if available.
|
||||||
func (s *Service) processFromSourceOrFetch(
|
func (s *Service) processFromSourceOrFetch(
|
||||||
ctx context.Context,
|
ctx context.Context,
|
||||||
@@ -202,8 +162,22 @@ func (s *Service) processFromSourceOrFetch(
|
|||||||
var fetchBytes int64
|
var fetchBytes int64
|
||||||
|
|
||||||
if contentHash != "" {
|
if contentHash != "" {
|
||||||
|
// We have cached source - load it
|
||||||
s.log.Debug("using cached source", "hash", contentHash)
|
s.log.Debug("using cached source", "hash", contentHash)
|
||||||
sourceData = s.loadCachedSource(contentHash)
|
|
||||||
|
reader, err := s.cache.GetSourceContent(contentHash)
|
||||||
|
if err != nil {
|
||||||
|
s.log.Warn("failed to load cached source, fetching", "error", err)
|
||||||
|
// Fall through to fetch
|
||||||
|
} else {
|
||||||
|
sourceData, err = io.ReadAll(reader)
|
||||||
|
_ = reader.Close()
|
||||||
|
|
||||||
|
if err != nil {
|
||||||
|
s.log.Warn("failed to read cached source, fetching", "error", err)
|
||||||
|
// Fall through to fetch
|
||||||
|
}
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Fetch from upstream if we don't have source data or it's empty
|
// Fetch from upstream if we don't have source data or it's empty
|
||||||
|
|||||||
@@ -16,7 +16,7 @@ func setupStatsTestDB(t *testing.T) *sql.DB {
|
|||||||
if err != nil {
|
if err != nil {
|
||||||
t.Fatal(err)
|
t.Fatal(err)
|
||||||
}
|
}
|
||||||
if err := database.ApplyMigrations(context.Background(), db, nil); err != nil {
|
if err := database.ApplyMigrations(db); err != nil {
|
||||||
t.Fatal(err)
|
t.Fatal(err)
|
||||||
}
|
}
|
||||||
t.Cleanup(func() { db.Close() })
|
t.Cleanup(func() { db.Close() })
|
||||||
|
|||||||
@@ -2,7 +2,6 @@ package imgcache
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"bytes"
|
"bytes"
|
||||||
"context"
|
|
||||||
"database/sql"
|
"database/sql"
|
||||||
"image"
|
"image"
|
||||||
"image/color"
|
"image/color"
|
||||||
@@ -194,7 +193,7 @@ func setupServiceTestDB(t *testing.T) *sql.DB {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// Use the real production schema via migrations
|
// Use the real production schema via migrations
|
||||||
if err := database.ApplyMigrations(context.Background(), db, nil); err != nil {
|
if err := database.ApplyMigrations(db); err != nil {
|
||||||
t.Fatalf("failed to apply migrations: %v", err)
|
t.Fatalf("failed to apply migrations: %v", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -5,23 +5,20 @@ import (
|
|||||||
"strings"
|
"strings"
|
||||||
)
|
)
|
||||||
|
|
||||||
// HostWhitelist implements the Whitelist interface for checking allowed source hosts.
|
// HostWhitelist checks whether a source host is allowed without a signature.
|
||||||
|
// Only exact host matches are supported. Signatures are per-URL, so
|
||||||
|
// wildcard/suffix matching is intentionally not provided.
|
||||||
type HostWhitelist struct {
|
type HostWhitelist struct {
|
||||||
// exactHosts contains hosts that must match exactly (e.g., "cdn.example.com")
|
// exactHosts contains hosts that must match exactly (e.g., "cdn.example.com")
|
||||||
exactHosts map[string]struct{}
|
exactHosts map[string]struct{}
|
||||||
// suffixHosts contains domain suffixes to match (e.g., ".example.com" matches "cdn.example.com")
|
|
||||||
suffixHosts []string
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// NewHostWhitelist creates a whitelist from a list of host patterns.
|
// NewHostWhitelist creates a whitelist from a list of hostnames.
|
||||||
// Patterns starting with "." are treated as suffix matches.
|
// Each entry is treated as an exact host match. Leading dots are
|
||||||
// Examples:
|
// stripped so that legacy ".example.com" entries become "example.com".
|
||||||
// - "cdn.example.com" - exact match only
|
|
||||||
// - ".example.com" - matches cdn.example.com, images.example.com, etc.
|
|
||||||
func NewHostWhitelist(patterns []string) *HostWhitelist {
|
func NewHostWhitelist(patterns []string) *HostWhitelist {
|
||||||
w := &HostWhitelist{
|
w := &HostWhitelist{
|
||||||
exactHosts: make(map[string]struct{}),
|
exactHosts: make(map[string]struct{}),
|
||||||
suffixHosts: make([]string, 0),
|
|
||||||
}
|
}
|
||||||
|
|
||||||
for _, pattern := range patterns {
|
for _, pattern := range patterns {
|
||||||
@@ -30,9 +27,11 @@ func NewHostWhitelist(patterns []string) *HostWhitelist {
|
|||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
|
|
||||||
if strings.HasPrefix(pattern, ".") {
|
// Strip leading dot — suffix matching is no longer supported;
|
||||||
w.suffixHosts = append(w.suffixHosts, pattern)
|
// ".example.com" is normalised to "example.com" as an exact entry.
|
||||||
} else {
|
pattern = strings.TrimPrefix(pattern, ".")
|
||||||
|
|
||||||
|
if pattern != "" {
|
||||||
w.exactHosts[pattern] = struct{}{}
|
w.exactHosts[pattern] = struct{}{}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -40,7 +39,7 @@ func NewHostWhitelist(patterns []string) *HostWhitelist {
|
|||||||
return w
|
return w
|
||||||
}
|
}
|
||||||
|
|
||||||
// IsWhitelisted checks if a URL's host is in the whitelist.
|
// IsWhitelisted checks if a URL's host is in the whitelist (exact match only).
|
||||||
func (w *HostWhitelist) IsWhitelisted(u *url.URL) bool {
|
func (w *HostWhitelist) IsWhitelisted(u *url.URL) bool {
|
||||||
if u == nil {
|
if u == nil {
|
||||||
return false
|
return false
|
||||||
@@ -51,32 +50,17 @@ func (w *HostWhitelist) IsWhitelisted(u *url.URL) bool {
|
|||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
|
|
||||||
// Check exact match
|
_, ok := w.exactHosts[host]
|
||||||
if _, ok := w.exactHosts[host]; ok {
|
|
||||||
return true
|
|
||||||
}
|
|
||||||
|
|
||||||
// Check suffix match
|
return ok
|
||||||
for _, suffix := range w.suffixHosts {
|
|
||||||
if strings.HasSuffix(host, suffix) {
|
|
||||||
return true
|
|
||||||
}
|
|
||||||
// Also match if host equals the suffix without the leading dot
|
|
||||||
// e.g., pattern ".example.com" should match "example.com"
|
|
||||||
if host == strings.TrimPrefix(suffix, ".") {
|
|
||||||
return true
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
return false
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// IsEmpty returns true if the whitelist has no entries.
|
// IsEmpty returns true if the whitelist has no entries.
|
||||||
func (w *HostWhitelist) IsEmpty() bool {
|
func (w *HostWhitelist) IsEmpty() bool {
|
||||||
return len(w.exactHosts) == 0 && len(w.suffixHosts) == 0
|
return len(w.exactHosts) == 0
|
||||||
}
|
}
|
||||||
|
|
||||||
// Count returns the total number of whitelist entries.
|
// Count returns the total number of whitelist entries.
|
||||||
func (w *HostWhitelist) Count() int {
|
func (w *HostWhitelist) Count() int {
|
||||||
return len(w.exactHosts) + len(w.suffixHosts)
|
return len(w.exactHosts)
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -31,41 +31,41 @@ func TestHostWhitelist_IsWhitelisted(t *testing.T) {
|
|||||||
want: false,
|
want: false,
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "suffix match",
|
name: "no suffix matching for subdomains",
|
||||||
patterns: []string{".example.com"},
|
patterns: []string{"example.com"},
|
||||||
testURL: "https://cdn.example.com/image.jpg",
|
testURL: "https://cdn.example.com/image.jpg",
|
||||||
want: true,
|
want: false,
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "suffix match deep subdomain",
|
name: "leading dot stripped to exact match",
|
||||||
patterns: []string{".example.com"},
|
|
||||||
testURL: "https://cdn.images.example.com/image.jpg",
|
|
||||||
want: true,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "suffix match apex domain",
|
|
||||||
patterns: []string{".example.com"},
|
patterns: []string{".example.com"},
|
||||||
testURL: "https://example.com/image.jpg",
|
testURL: "https://example.com/image.jpg",
|
||||||
want: true,
|
want: true,
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "suffix match not found",
|
name: "leading dot does not enable suffix matching",
|
||||||
patterns: []string{".example.com"},
|
patterns: []string{".example.com"},
|
||||||
testURL: "https://notexample.com/image.jpg",
|
testURL: "https://cdn.example.com/image.jpg",
|
||||||
want: false,
|
want: false,
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "suffix match partial not allowed",
|
name: "leading dot does not match deep subdomain",
|
||||||
patterns: []string{".example.com"},
|
patterns: []string{".example.com"},
|
||||||
testURL: "https://fakeexample.com/image.jpg",
|
testURL: "https://cdn.images.example.com/image.jpg",
|
||||||
want: false,
|
want: false,
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "multiple patterns",
|
name: "multiple patterns exact only",
|
||||||
patterns: []string{"cdn.example.com", ".images.org", "static.test.net"},
|
patterns: []string{"cdn.example.com", "photos.images.org", "static.test.net"},
|
||||||
testURL: "https://photos.images.org/image.jpg",
|
testURL: "https://photos.images.org/image.jpg",
|
||||||
want: true,
|
want: true,
|
||||||
},
|
},
|
||||||
|
{
|
||||||
|
name: "multiple patterns no suffix leak",
|
||||||
|
patterns: []string{"cdn.example.com", "images.org"},
|
||||||
|
testURL: "https://photos.images.org/image.jpg",
|
||||||
|
want: false,
|
||||||
|
},
|
||||||
{
|
{
|
||||||
name: "empty whitelist",
|
name: "empty whitelist",
|
||||||
patterns: []string{},
|
patterns: []string{},
|
||||||
@@ -86,7 +86,7 @@ func TestHostWhitelist_IsWhitelisted(t *testing.T) {
|
|||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "whitespace in patterns",
|
name: "whitespace in patterns",
|
||||||
patterns: []string{" cdn.example.com ", " .other.com "},
|
patterns: []string{" cdn.example.com ", " other.com "},
|
||||||
testURL: "https://cdn.example.com/image.jpg",
|
testURL: "https://cdn.example.com/image.jpg",
|
||||||
want: true,
|
want: true,
|
||||||
},
|
},
|
||||||
@@ -139,6 +139,11 @@ func TestHostWhitelist_IsEmpty(t *testing.T) {
|
|||||||
patterns: []string{"example.com"},
|
patterns: []string{"example.com"},
|
||||||
want: false,
|
want: false,
|
||||||
},
|
},
|
||||||
|
{
|
||||||
|
name: "leading dot normalised to entry",
|
||||||
|
patterns: []string{".example.com"},
|
||||||
|
want: false,
|
||||||
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
for _, tt := range tests {
|
for _, tt := range tests {
|
||||||
@@ -168,14 +173,14 @@ func TestHostWhitelist_Count(t *testing.T) {
|
|||||||
want: 3,
|
want: 3,
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "suffix hosts only",
|
name: "leading dots normalised to exact",
|
||||||
patterns: []string{".a.com", ".b.com"},
|
patterns: []string{".a.com", ".b.com"},
|
||||||
want: 2,
|
want: 2,
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "mixed",
|
name: "mixed deduplication",
|
||||||
patterns: []string{"exact.com", ".suffix.com"},
|
patterns: []string{"example.com", ".example.com"},
|
||||||
want: 2,
|
want: 1,
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
Reference in New Issue
Block a user