Add SQL query logging and performance improvements
- Implement comprehensive SQL query logging for queries over 10ms - Add logging wrapper methods for all database operations - Replace timing code in GetStats with simple info log messages - Add missing database indexes for better query performance: - idx_live_routes_lookup for common prefix/origin/peer lookups - idx_live_routes_withdraw for withdrawal updates - idx_prefixes_prefix for prefix lookups - idx_asn_peerings_lookup for peering relationship queries - Increase SQLite cache size to 512MB - Add performance-oriented SQLite pragmas - Extract HTML templates to separate files using go:embed - Add JSON response middleware with @meta field (like bgpview.io API) - Fix concurrent map write errors in HTTP handlers - Add request timeout handling with proper JSON error responses These changes significantly improve database query performance and provide visibility into slow queries for debugging purposes.
This commit is contained in:
@@ -88,6 +88,7 @@ CREATE INDEX IF NOT EXISTS idx_announcements_prefix_id ON announcements(prefix_i
|
||||
CREATE INDEX IF NOT EXISTS idx_announcements_asn_id ON announcements(asn_id);
|
||||
CREATE INDEX IF NOT EXISTS idx_asn_peerings_from_asn ON asn_peerings(from_asn_id);
|
||||
CREATE INDEX IF NOT EXISTS idx_asn_peerings_to_asn ON asn_peerings(to_asn_id);
|
||||
CREATE INDEX IF NOT EXISTS idx_asn_peerings_lookup ON asn_peerings(from_asn_id, to_asn_id);
|
||||
|
||||
-- Indexes for live routes table
|
||||
CREATE INDEX IF NOT EXISTS idx_live_routes_active
|
||||
@@ -102,6 +103,19 @@ CREATE INDEX IF NOT EXISTS idx_live_routes_prefix
|
||||
ON live_routes(prefix_id)
|
||||
WHERE withdrawn_at IS NULL;
|
||||
|
||||
-- Critical index for the most common query pattern
|
||||
CREATE INDEX IF NOT EXISTS idx_live_routes_lookup
|
||||
ON live_routes(prefix_id, origin_asn_id, peer_asn)
|
||||
WHERE withdrawn_at IS NULL;
|
||||
|
||||
-- Index for withdrawal updates by prefix and peer
|
||||
CREATE INDEX IF NOT EXISTS idx_live_routes_withdraw
|
||||
ON live_routes(prefix_id, peer_asn)
|
||||
WHERE withdrawn_at IS NULL;
|
||||
|
||||
-- Additional indexes for prefixes table
|
||||
CREATE INDEX IF NOT EXISTS idx_prefixes_prefix ON prefixes(prefix);
|
||||
|
||||
-- Indexes for bgp_peers table
|
||||
CREATE INDEX IF NOT EXISTS idx_bgp_peers_asn ON bgp_peers(peer_asn);
|
||||
CREATE INDEX IF NOT EXISTS idx_bgp_peers_last_seen ON bgp_peers(last_seen);
|
||||
@@ -209,7 +223,23 @@ func NewWithConfig(config Config, logger *slog.Logger) (*Database, error) {
|
||||
|
||||
// Initialize creates the database schema if it doesn't exist.
|
||||
func (d *Database) Initialize() error {
|
||||
_, err := d.db.Exec(dbSchema)
|
||||
// Set SQLite pragmas for better performance
|
||||
pragmas := []string{
|
||||
"PRAGMA journal_mode=WAL", // Already set in connection string
|
||||
"PRAGMA synchronous=NORMAL", // Faster than FULL, still safe
|
||||
"PRAGMA cache_size=-524288", // 512MB cache
|
||||
"PRAGMA temp_store=MEMORY", // Use memory for temp tables
|
||||
"PRAGMA mmap_size=268435456", // 256MB memory-mapped I/O
|
||||
"PRAGMA optimize", // Run optimizer
|
||||
}
|
||||
|
||||
for _, pragma := range pragmas {
|
||||
if err := d.exec(pragma); err != nil {
|
||||
d.logger.Warn("Failed to set pragma", "pragma", pragma, "error", err)
|
||||
}
|
||||
}
|
||||
|
||||
err := d.exec(dbSchema)
|
||||
|
||||
return err
|
||||
}
|
||||
@@ -219,9 +249,19 @@ func (d *Database) Close() error {
|
||||
return d.db.Close()
|
||||
}
|
||||
|
||||
// beginTx starts a new transaction with logging
|
||||
func (d *Database) beginTx() (*loggingTx, error) {
|
||||
tx, err := d.db.Begin()
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
|
||||
return &loggingTx{Tx: tx, logger: d.logger}, nil
|
||||
}
|
||||
|
||||
// GetOrCreateASN retrieves an existing ASN or creates a new one if it doesn't exist.
|
||||
func (d *Database) GetOrCreateASN(number int, timestamp time.Time) (*ASN, error) {
|
||||
tx, err := d.db.Begin()
|
||||
tx, err := d.beginTx()
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
@@ -282,7 +322,7 @@ func (d *Database) GetOrCreateASN(number int, timestamp time.Time) (*ASN, error)
|
||||
|
||||
// GetOrCreatePrefix retrieves an existing prefix or creates a new one if it doesn't exist.
|
||||
func (d *Database) GetOrCreatePrefix(prefix string, timestamp time.Time) (*Prefix, error) {
|
||||
tx, err := d.db.Begin()
|
||||
tx, err := d.beginTx()
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
@@ -344,7 +384,7 @@ func (d *Database) GetOrCreatePrefix(prefix string, timestamp time.Time) (*Prefi
|
||||
|
||||
// RecordAnnouncement inserts a new BGP announcement or withdrawal into the database.
|
||||
func (d *Database) RecordAnnouncement(announcement *Announcement) error {
|
||||
_, err := d.db.Exec(`
|
||||
err := d.exec(`
|
||||
INSERT INTO announcements (id, prefix_id, asn_id, origin_asn_id, path, next_hop, timestamp, is_withdrawal)
|
||||
VALUES (?, ?, ?, ?, ?, ?, ?, ?)`,
|
||||
announcement.ID.String(), announcement.PrefixID.String(),
|
||||
@@ -356,7 +396,7 @@ func (d *Database) RecordAnnouncement(announcement *Announcement) error {
|
||||
|
||||
// RecordPeering records a peering relationship between two ASNs.
|
||||
func (d *Database) RecordPeering(fromASNID, toASNID string, timestamp time.Time) error {
|
||||
tx, err := d.db.Begin()
|
||||
tx, err := d.beginTx()
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
@@ -408,7 +448,7 @@ func (d *Database) UpdateLiveRoute(
|
||||
) error {
|
||||
// Check if route already exists
|
||||
var routeID sql.NullString
|
||||
err := d.db.QueryRow(`
|
||||
err := d.queryRow(`
|
||||
SELECT id FROM live_routes
|
||||
WHERE prefix_id = ? AND origin_asn_id = ? AND peer_asn = ? AND withdrawn_at IS NULL`,
|
||||
prefixID.String(), originASNID.String(), peerASN).Scan(&routeID)
|
||||
@@ -419,14 +459,14 @@ func (d *Database) UpdateLiveRoute(
|
||||
|
||||
if routeID.Valid {
|
||||
// Route exists and is active, update it
|
||||
_, err = d.db.Exec(`
|
||||
err = d.exec(`
|
||||
UPDATE live_routes
|
||||
SET next_hop = ?, announced_at = ?
|
||||
WHERE id = ?`,
|
||||
nextHop, timestamp, routeID.String)
|
||||
} else {
|
||||
// Either new route or re-announcement of withdrawn route
|
||||
_, err = d.db.Exec(`
|
||||
err = d.exec(`
|
||||
INSERT OR REPLACE INTO live_routes
|
||||
(id, prefix_id, origin_asn_id, peer_asn, next_hop, announced_at, withdrawn_at)
|
||||
VALUES (?, ?, ?, ?, ?, ?, NULL)`,
|
||||
@@ -439,7 +479,7 @@ func (d *Database) UpdateLiveRoute(
|
||||
|
||||
// WithdrawLiveRoute marks a route as withdrawn in the live routing table
|
||||
func (d *Database) WithdrawLiveRoute(prefixID uuid.UUID, peerASN int, timestamp time.Time) error {
|
||||
_, err := d.db.Exec(`
|
||||
err := d.exec(`
|
||||
UPDATE live_routes
|
||||
SET withdrawn_at = ?
|
||||
WHERE prefix_id = ? AND peer_asn = ? AND withdrawn_at IS NULL`,
|
||||
@@ -450,7 +490,7 @@ func (d *Database) WithdrawLiveRoute(prefixID uuid.UUID, peerASN int, timestamp
|
||||
|
||||
// GetActiveLiveRoutes returns all currently active routes (not withdrawn)
|
||||
func (d *Database) GetActiveLiveRoutes() ([]LiveRoute, error) {
|
||||
rows, err := d.db.Query(`
|
||||
rows, err := d.query(`
|
||||
SELECT id, prefix_id, origin_asn_id, peer_asn, next_hop, announced_at
|
||||
FROM live_routes
|
||||
WHERE withdrawn_at IS NULL
|
||||
@@ -484,7 +524,7 @@ func (d *Database) GetActiveLiveRoutes() ([]LiveRoute, error) {
|
||||
|
||||
// UpdatePeer updates or creates a BGP peer record
|
||||
func (d *Database) UpdatePeer(peerIP string, peerASN int, messageType string, timestamp time.Time) error {
|
||||
tx, err := d.db.Begin()
|
||||
tx, err := d.beginTx()
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
@@ -533,41 +573,49 @@ func (d *Database) GetStats() (Stats, error) {
|
||||
var stats Stats
|
||||
|
||||
// Count ASNs
|
||||
err := d.db.QueryRow("SELECT COUNT(*) FROM asns").Scan(&stats.ASNs)
|
||||
d.logger.Info("Counting ASNs")
|
||||
err := d.queryRow("SELECT COUNT(*) FROM asns").Scan(&stats.ASNs)
|
||||
if err != nil {
|
||||
return stats, err
|
||||
}
|
||||
|
||||
// Count prefixes
|
||||
err = d.db.QueryRow("SELECT COUNT(*) FROM prefixes").Scan(&stats.Prefixes)
|
||||
d.logger.Info("Counting prefixes")
|
||||
err = d.queryRow("SELECT COUNT(*) FROM prefixes").Scan(&stats.Prefixes)
|
||||
if err != nil {
|
||||
return stats, err
|
||||
}
|
||||
|
||||
// Count IPv4 and IPv6 prefixes
|
||||
d.logger.Info("Counting IPv4 prefixes")
|
||||
const ipVersionV4 = 4
|
||||
err = d.db.QueryRow("SELECT COUNT(*) FROM prefixes WHERE ip_version = ?", ipVersionV4).Scan(&stats.IPv4Prefixes)
|
||||
err = d.queryRow("SELECT COUNT(*) FROM prefixes WHERE ip_version = ?", ipVersionV4).Scan(&stats.IPv4Prefixes)
|
||||
if err != nil {
|
||||
return stats, err
|
||||
}
|
||||
|
||||
d.logger.Info("Counting IPv6 prefixes")
|
||||
const ipVersionV6 = 6
|
||||
err = d.db.QueryRow("SELECT COUNT(*) FROM prefixes WHERE ip_version = ?", ipVersionV6).Scan(&stats.IPv6Prefixes)
|
||||
err = d.queryRow("SELECT COUNT(*) FROM prefixes WHERE ip_version = ?", ipVersionV6).Scan(&stats.IPv6Prefixes)
|
||||
if err != nil {
|
||||
return stats, err
|
||||
}
|
||||
|
||||
// Count peerings
|
||||
err = d.db.QueryRow("SELECT COUNT(*) FROM asn_peerings").Scan(&stats.Peerings)
|
||||
d.logger.Info("Counting peerings")
|
||||
err = d.queryRow("SELECT COUNT(*) FROM asn_peerings").Scan(&stats.Peerings)
|
||||
if err != nil {
|
||||
return stats, err
|
||||
}
|
||||
|
||||
// Count live routes
|
||||
err = d.db.QueryRow("SELECT COUNT(*) FROM live_routes WHERE withdrawn_at IS NULL").Scan(&stats.LiveRoutes)
|
||||
d.logger.Info("Counting live routes")
|
||||
err = d.queryRow("SELECT COUNT(*) FROM live_routes WHERE withdrawn_at IS NULL").Scan(&stats.LiveRoutes)
|
||||
if err != nil {
|
||||
return stats, err
|
||||
}
|
||||
|
||||
d.logger.Info("Stats collection complete")
|
||||
|
||||
return stats, nil
|
||||
}
|
||||
|
||||
98
internal/database/slowquery.go
Normal file
98
internal/database/slowquery.go
Normal file
@@ -0,0 +1,98 @@
|
||||
package database
|
||||
|
||||
import (
|
||||
"context"
|
||||
"database/sql"
|
||||
"log/slog"
|
||||
"time"
|
||||
)
|
||||
|
||||
const slowQueryThreshold = 10 * time.Millisecond
|
||||
|
||||
// logSlowQuery logs queries that take longer than slowQueryThreshold
|
||||
func logSlowQuery(logger *slog.Logger, query string, start time.Time) {
|
||||
elapsed := time.Since(start)
|
||||
if elapsed > slowQueryThreshold {
|
||||
logger.Debug("Slow query", "query", query, "duration", elapsed)
|
||||
}
|
||||
}
|
||||
|
||||
// queryRow wraps QueryRow with slow query logging
|
||||
func (d *Database) queryRow(query string, args ...interface{}) *sql.Row {
|
||||
start := time.Now()
|
||||
defer logSlowQuery(d.logger, query, start)
|
||||
|
||||
return d.db.QueryRow(query, args...)
|
||||
}
|
||||
|
||||
// query wraps Query with slow query logging
|
||||
func (d *Database) query(query string, args ...interface{}) (*sql.Rows, error) {
|
||||
start := time.Now()
|
||||
defer logSlowQuery(d.logger, query, start)
|
||||
|
||||
return d.db.Query(query, args...)
|
||||
}
|
||||
|
||||
// exec wraps Exec with slow query logging
|
||||
func (d *Database) exec(query string, args ...interface{}) error {
|
||||
start := time.Now()
|
||||
defer logSlowQuery(d.logger, query, start)
|
||||
|
||||
_, err := d.db.Exec(query, args...)
|
||||
|
||||
return err
|
||||
}
|
||||
|
||||
// loggingTx wraps sql.Tx to log slow queries
|
||||
type loggingTx struct {
|
||||
*sql.Tx
|
||||
logger *slog.Logger
|
||||
}
|
||||
|
||||
// QueryRow wraps sql.Tx.QueryRow to log slow queries
|
||||
func (tx *loggingTx) QueryRow(query string, args ...interface{}) *sql.Row {
|
||||
start := time.Now()
|
||||
defer logSlowQuery(tx.logger, query, start)
|
||||
|
||||
return tx.Tx.QueryRow(query, args...)
|
||||
}
|
||||
|
||||
// Query wraps sql.Tx.Query to log slow queries
|
||||
func (tx *loggingTx) Query(query string, args ...interface{}) (*sql.Rows, error) {
|
||||
start := time.Now()
|
||||
defer logSlowQuery(tx.logger, query, start)
|
||||
|
||||
return tx.Tx.Query(query, args...)
|
||||
}
|
||||
|
||||
// QueryContext wraps sql.Tx.QueryContext to log slow queries
|
||||
func (tx *loggingTx) QueryContext(ctx context.Context, query string, args ...interface{}) (*sql.Rows, error) {
|
||||
start := time.Now()
|
||||
defer logSlowQuery(tx.logger, query, start)
|
||||
|
||||
return tx.Tx.QueryContext(ctx, query, args...)
|
||||
}
|
||||
|
||||
// QueryRowContext wraps sql.Tx.QueryRowContext to log slow queries
|
||||
func (tx *loggingTx) QueryRowContext(ctx context.Context, query string, args ...interface{}) *sql.Row {
|
||||
start := time.Now()
|
||||
defer logSlowQuery(tx.logger, query, start)
|
||||
|
||||
return tx.Tx.QueryRowContext(ctx, query, args...)
|
||||
}
|
||||
|
||||
// Exec wraps sql.Tx.Exec to log slow queries
|
||||
func (tx *loggingTx) Exec(query string, args ...interface{}) (sql.Result, error) {
|
||||
start := time.Now()
|
||||
defer logSlowQuery(tx.logger, query, start)
|
||||
|
||||
return tx.Tx.Exec(query, args...)
|
||||
}
|
||||
|
||||
// ExecContext wraps sql.Tx.ExecContext to log slow queries
|
||||
func (tx *loggingTx) ExecContext(ctx context.Context, query string, args ...interface{}) (sql.Result, error) {
|
||||
start := time.Now()
|
||||
defer logSlowQuery(tx.logger, query, start)
|
||||
|
||||
return tx.Tx.ExecContext(ctx, query, args...)
|
||||
}
|
||||
Reference in New Issue
Block a user