- Add pure Go SQLite driver (modernc.org/sqlite) to avoid CGO dependency - Implement database connection management with WAL mode - Add write mutex for serializing concurrent writes - Create schema for all tables matching DESIGN.md specifications - Implement repository pattern for all database entities: - Files, FileChunks, Chunks, Blobs, BlobChunks, ChunkFiles, Snapshots - Add transaction support with proper rollback handling - Add fatal error handling for database integrity issues - Add snapshot fields for tracking file sizes and compression ratios - Make index path configurable via VAULTIK_INDEX_PATH environment variable - Add comprehensive test coverage for all repositories - Add format check to Makefile to ensure code formatting
142 lines
2.7 KiB
Go
142 lines
2.7 KiB
Go
package database
|
|
|
|
import (
|
|
"context"
|
|
"database/sql"
|
|
"fmt"
|
|
)
|
|
|
|
type ChunkRepository struct {
|
|
db *DB
|
|
}
|
|
|
|
func NewChunkRepository(db *DB) *ChunkRepository {
|
|
return &ChunkRepository{db: db}
|
|
}
|
|
|
|
func (r *ChunkRepository) Create(ctx context.Context, tx *sql.Tx, chunk *Chunk) error {
|
|
query := `
|
|
INSERT INTO chunks (chunk_hash, sha256, size)
|
|
VALUES (?, ?, ?)
|
|
ON CONFLICT(chunk_hash) DO NOTHING
|
|
`
|
|
|
|
var err error
|
|
if tx != nil {
|
|
_, err = tx.ExecContext(ctx, query, chunk.ChunkHash, chunk.SHA256, chunk.Size)
|
|
} else {
|
|
_, err = r.db.ExecWithLock(ctx, query, chunk.ChunkHash, chunk.SHA256, chunk.Size)
|
|
}
|
|
|
|
if err != nil {
|
|
return fmt.Errorf("inserting chunk: %w", err)
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
func (r *ChunkRepository) GetByHash(ctx context.Context, hash string) (*Chunk, error) {
|
|
query := `
|
|
SELECT chunk_hash, sha256, size
|
|
FROM chunks
|
|
WHERE chunk_hash = ?
|
|
`
|
|
|
|
var chunk Chunk
|
|
|
|
err := r.db.conn.QueryRowContext(ctx, query, hash).Scan(
|
|
&chunk.ChunkHash,
|
|
&chunk.SHA256,
|
|
&chunk.Size,
|
|
)
|
|
|
|
if err == sql.ErrNoRows {
|
|
return nil, nil
|
|
}
|
|
if err != nil {
|
|
return nil, fmt.Errorf("querying chunk: %w", err)
|
|
}
|
|
|
|
return &chunk, nil
|
|
}
|
|
|
|
func (r *ChunkRepository) GetByHashes(ctx context.Context, hashes []string) ([]*Chunk, error) {
|
|
if len(hashes) == 0 {
|
|
return nil, nil
|
|
}
|
|
|
|
query := `
|
|
SELECT chunk_hash, sha256, size
|
|
FROM chunks
|
|
WHERE chunk_hash IN (`
|
|
|
|
args := make([]interface{}, len(hashes))
|
|
for i, hash := range hashes {
|
|
if i > 0 {
|
|
query += ", "
|
|
}
|
|
query += "?"
|
|
args[i] = hash
|
|
}
|
|
query += ") ORDER BY chunk_hash"
|
|
|
|
rows, err := r.db.conn.QueryContext(ctx, query, args...)
|
|
if err != nil {
|
|
return nil, fmt.Errorf("querying chunks: %w", err)
|
|
}
|
|
defer CloseRows(rows)
|
|
|
|
var chunks []*Chunk
|
|
for rows.Next() {
|
|
var chunk Chunk
|
|
|
|
err := rows.Scan(
|
|
&chunk.ChunkHash,
|
|
&chunk.SHA256,
|
|
&chunk.Size,
|
|
)
|
|
if err != nil {
|
|
return nil, fmt.Errorf("scanning chunk: %w", err)
|
|
}
|
|
|
|
chunks = append(chunks, &chunk)
|
|
}
|
|
|
|
return chunks, rows.Err()
|
|
}
|
|
|
|
func (r *ChunkRepository) ListUnpacked(ctx context.Context, limit int) ([]*Chunk, error) {
|
|
query := `
|
|
SELECT c.chunk_hash, c.sha256, c.size
|
|
FROM chunks c
|
|
LEFT JOIN blob_chunks bc ON c.chunk_hash = bc.chunk_hash
|
|
WHERE bc.chunk_hash IS NULL
|
|
ORDER BY c.chunk_hash
|
|
LIMIT ?
|
|
`
|
|
|
|
rows, err := r.db.conn.QueryContext(ctx, query, limit)
|
|
if err != nil {
|
|
return nil, fmt.Errorf("querying unpacked chunks: %w", err)
|
|
}
|
|
defer CloseRows(rows)
|
|
|
|
var chunks []*Chunk
|
|
for rows.Next() {
|
|
var chunk Chunk
|
|
|
|
err := rows.Scan(
|
|
&chunk.ChunkHash,
|
|
&chunk.SHA256,
|
|
&chunk.Size,
|
|
)
|
|
if err != nil {
|
|
return nil, fmt.Errorf("scanning chunk: %w", err)
|
|
}
|
|
|
|
chunks = append(chunks, &chunk)
|
|
}
|
|
|
|
return chunks, rows.Err()
|
|
}
|