fix: resolve rebase conflicts, fix errcheck issues, implement FetchAndDecryptBlob
This commit is contained in:
parent
9879668c31
commit
3f834f1c9c
@ -1,15 +1,12 @@
|
|||||||
package vaultik
|
package vaultik
|
||||||
|
|
||||||
// TODO: These are stub implementations for methods referenced but not yet
|
|
||||||
// implemented. They allow the package to compile for testing.
|
|
||||||
// Remove once the real implementations land.
|
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
"fmt"
|
"fmt"
|
||||||
"io"
|
"io"
|
||||||
|
|
||||||
"filippo.io/age"
|
"filippo.io/age"
|
||||||
|
"git.eeqj.de/sneak/vaultik/internal/blobgen"
|
||||||
)
|
)
|
||||||
|
|
||||||
// FetchAndDecryptBlobResult holds the result of fetching and decrypting a blob.
|
// FetchAndDecryptBlobResult holds the result of fetching and decrypting a blob.
|
||||||
@ -19,10 +16,40 @@ type FetchAndDecryptBlobResult struct {
|
|||||||
|
|
||||||
// FetchAndDecryptBlob downloads a blob, decrypts it, and returns the plaintext data.
|
// FetchAndDecryptBlob downloads a blob, decrypts it, and returns the plaintext data.
|
||||||
func (v *Vaultik) FetchAndDecryptBlob(ctx context.Context, blobHash string, expectedSize int64, identity age.Identity) (*FetchAndDecryptBlobResult, error) {
|
func (v *Vaultik) FetchAndDecryptBlob(ctx context.Context, blobHash string, expectedSize int64, identity age.Identity) (*FetchAndDecryptBlobResult, error) {
|
||||||
return nil, fmt.Errorf("FetchAndDecryptBlob not yet implemented")
|
rc, _, err := v.FetchBlob(ctx, blobHash, expectedSize)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
defer func() { _ = rc.Close() }()
|
||||||
|
|
||||||
|
reader, err := blobgen.NewReader(rc, identity)
|
||||||
|
if err != nil {
|
||||||
|
return nil, fmt.Errorf("creating blob reader: %w", err)
|
||||||
|
}
|
||||||
|
defer func() { _ = reader.Close() }()
|
||||||
|
|
||||||
|
data, err := io.ReadAll(reader)
|
||||||
|
if err != nil {
|
||||||
|
return nil, fmt.Errorf("reading blob data: %w", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
return &FetchAndDecryptBlobResult{Data: data}, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// FetchBlob downloads a blob and returns a reader for the encrypted data.
|
// FetchBlob downloads a blob and returns a reader for the encrypted data.
|
||||||
func (v *Vaultik) FetchBlob(ctx context.Context, blobHash string, expectedSize int64) (io.ReadCloser, int64, error) {
|
func (v *Vaultik) FetchBlob(ctx context.Context, blobHash string, expectedSize int64) (io.ReadCloser, int64, error) {
|
||||||
return nil, 0, fmt.Errorf("FetchBlob not yet implemented")
|
blobPath := fmt.Sprintf("blobs/%s/%s/%s", blobHash[:2], blobHash[2:4], blobHash)
|
||||||
|
|
||||||
|
rc, err := v.Storage.Get(ctx, blobPath)
|
||||||
|
if err != nil {
|
||||||
|
return nil, 0, fmt.Errorf("downloading blob %s: %w", blobHash[:16], err)
|
||||||
|
}
|
||||||
|
|
||||||
|
info, err := v.Storage.Stat(ctx, blobPath)
|
||||||
|
if err != nil {
|
||||||
|
_ = rc.Close()
|
||||||
|
return nil, 0, fmt.Errorf("stat blob %s: %w", blobHash[:16], err)
|
||||||
|
}
|
||||||
|
|
||||||
|
return rc, info.Size, nil
|
||||||
}
|
}
|
||||||
|
|||||||
@ -167,7 +167,7 @@ func (c *blobDiskCache) ReadAt(key string, offset, length int64) ([]byte, error)
|
|||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
defer f.Close()
|
defer func() { _ = f.Close() }()
|
||||||
|
|
||||||
buf := make([]byte, length)
|
buf := make([]byte, length)
|
||||||
if _, err := f.ReadAt(buf, offset); err != nil {
|
if _, err := f.ReadAt(buf, offset); err != nil {
|
||||||
|
|||||||
@ -12,7 +12,7 @@ func TestBlobDiskCache_BasicGetPut(t *testing.T) {
|
|||||||
if err != nil {
|
if err != nil {
|
||||||
t.Fatal(err)
|
t.Fatal(err)
|
||||||
}
|
}
|
||||||
defer cache.Close()
|
defer func() { _ = cache.Close() }()
|
||||||
|
|
||||||
data := []byte("hello world")
|
data := []byte("hello world")
|
||||||
if err := cache.Put("key1", data); err != nil {
|
if err := cache.Put("key1", data); err != nil {
|
||||||
@ -39,7 +39,7 @@ func TestBlobDiskCache_EvictionUnderPressure(t *testing.T) {
|
|||||||
if err != nil {
|
if err != nil {
|
||||||
t.Fatal(err)
|
t.Fatal(err)
|
||||||
}
|
}
|
||||||
defer cache.Close()
|
defer func() { _ = cache.Close() }()
|
||||||
|
|
||||||
for i := 0; i < 5; i++ {
|
for i := 0; i < 5; i++ {
|
||||||
data := make([]byte, 300)
|
data := make([]byte, 300)
|
||||||
@ -65,7 +65,7 @@ func TestBlobDiskCache_OversizedEntryRejected(t *testing.T) {
|
|||||||
if err != nil {
|
if err != nil {
|
||||||
t.Fatal(err)
|
t.Fatal(err)
|
||||||
}
|
}
|
||||||
defer cache.Close()
|
defer func() { _ = cache.Close() }()
|
||||||
|
|
||||||
data := make([]byte, 200)
|
data := make([]byte, 200)
|
||||||
if err := cache.Put("big", data); err != nil {
|
if err := cache.Put("big", data); err != nil {
|
||||||
@ -82,7 +82,7 @@ func TestBlobDiskCache_UpdateInPlace(t *testing.T) {
|
|||||||
if err != nil {
|
if err != nil {
|
||||||
t.Fatal(err)
|
t.Fatal(err)
|
||||||
}
|
}
|
||||||
defer cache.Close()
|
defer func() { _ = cache.Close() }()
|
||||||
|
|
||||||
if err := cache.Put("key1", []byte("v1")); err != nil {
|
if err := cache.Put("key1", []byte("v1")); err != nil {
|
||||||
t.Fatal(err)
|
t.Fatal(err)
|
||||||
@ -111,7 +111,7 @@ func TestBlobDiskCache_ReadAt(t *testing.T) {
|
|||||||
if err != nil {
|
if err != nil {
|
||||||
t.Fatal(err)
|
t.Fatal(err)
|
||||||
}
|
}
|
||||||
defer cache.Close()
|
defer func() { _ = cache.Close() }()
|
||||||
|
|
||||||
data := make([]byte, 1024)
|
data := make([]byte, 1024)
|
||||||
if _, err := rand.Read(data); err != nil {
|
if _, err := rand.Read(data); err != nil {
|
||||||
@ -159,7 +159,7 @@ func TestBlobDiskCache_LRUOrder(t *testing.T) {
|
|||||||
if err != nil {
|
if err != nil {
|
||||||
t.Fatal(err)
|
t.Fatal(err)
|
||||||
}
|
}
|
||||||
defer cache.Close()
|
defer func() { _ = cache.Close() }()
|
||||||
|
|
||||||
d := make([]byte, 100)
|
d := make([]byte, 100)
|
||||||
if err := cache.Put("a", d); err != nil {
|
if err := cache.Put("a", d); err != nil {
|
||||||
|
|||||||
@ -113,7 +113,7 @@ func (v *Vaultik) Restore(opts *RestoreOptions) error {
|
|||||||
if err != nil {
|
if err != nil {
|
||||||
return fmt.Errorf("creating blob cache: %w", err)
|
return fmt.Errorf("creating blob cache: %w", err)
|
||||||
}
|
}
|
||||||
defer blobCache.Close()
|
defer func() { _ = blobCache.Close() }()
|
||||||
|
|
||||||
for i, file := range files {
|
for i, file := range files {
|
||||||
if v.ctx.Err() != nil {
|
if v.ctx.Err() != nil {
|
||||||
@ -427,7 +427,9 @@ func (v *Vaultik) restoreRegularFile(
|
|||||||
if err != nil {
|
if err != nil {
|
||||||
return fmt.Errorf("downloading blob %s: %w", blobHashStr[:16], err)
|
return fmt.Errorf("downloading blob %s: %w", blobHashStr[:16], err)
|
||||||
}
|
}
|
||||||
if putErr := blobCache.Put(blobHashStr, blobData); putErr != nil { log.Debug("Failed to cache blob on disk", "hash", blobHashStr[:16], "error", putErr) }
|
if putErr := blobCache.Put(blobHashStr, blobData); putErr != nil {
|
||||||
|
log.Debug("Failed to cache blob on disk", "hash", blobHashStr[:16], "error", putErr)
|
||||||
|
}
|
||||||
result.BlobsDownloaded++
|
result.BlobsDownloaded++
|
||||||
result.BytesDownloaded += blob.CompressedSize
|
result.BytesDownloaded += blob.CompressedSize
|
||||||
}
|
}
|
||||||
|
|||||||
@ -851,7 +851,7 @@ func (v *Vaultik) RemoveSnapshot(snapshotID string, opts *RemoveOptions) (*Remov
|
|||||||
v.printfStdout("Remove snapshot '%s' from local database? [y/N] ", snapshotID)
|
v.printfStdout("Remove snapshot '%s' from local database? [y/N] ", snapshotID)
|
||||||
}
|
}
|
||||||
var confirm string
|
var confirm string
|
||||||
if err := v.scanlnStdin(&confirm); err != nil {
|
if _, err := v.scanStdin(&confirm); err != nil {
|
||||||
v.printlnStdout("Cancelled")
|
v.printlnStdout("Cancelled")
|
||||||
return result, nil
|
return result, nil
|
||||||
}
|
}
|
||||||
|
|||||||
Loading…
Reference in New Issue
Block a user