go-store/compact.go
Virgil bff79c31ca fix(store): add nil-safe guards
Add nil/closed checks across the store, scoped store, workspace, journal, event, and compact entry points so agent callers get wrapped errors instead of panics.

Co-Authored-By: Virgil <virgil@lethean.io>
2026-04-03 06:31:35 +00:00

214 lines
6.3 KiB
Go

package store
import (
"compress/gzip"
"io"
"time"
core "dappco.re/go/core"
"github.com/klauspost/compress/zstd"
)
var defaultArchiveOutputDirectory = ".core/archive"
// Usage example: `options := store.CompactOptions{Before: time.Now().Add(-90 * 24 * time.Hour), Output: "/tmp/archive", Format: "gzip"}`
type CompactOptions struct {
Before time.Time
Output string
Format string
}
type compactArchiveEntry struct {
entryID int64
bucketName string
measurementName string
fieldsJSON string
tagsJSON string
committedAtUnixMilli int64
}
// Usage example: `result := storeInstance.Compact(store.CompactOptions{Before: time.Now().Add(-30 * 24 * time.Hour), Output: "/tmp/archive", Format: "gzip"})`
func (storeInstance *Store) Compact(options CompactOptions) core.Result {
if err := storeInstance.ensureReady("store.Compact"); err != nil {
return core.Result{Value: err, OK: false}
}
if err := ensureJournalSchema(storeInstance.database); err != nil {
return core.Result{Value: core.E("store.Compact", "ensure journal schema", err), OK: false}
}
outputDirectory := options.Output
if outputDirectory == "" {
outputDirectory = defaultArchiveOutputDirectory
}
format := options.Format
if format == "" {
format = "gzip"
}
if format != "gzip" && format != "zstd" {
return core.Result{Value: core.E("store.Compact", core.Concat("unsupported archive format: ", format), nil), OK: false}
}
filesystem := (&core.Fs{}).NewUnrestricted()
if result := filesystem.EnsureDir(outputDirectory); !result.OK {
return core.Result{Value: core.E("store.Compact", "ensure archive directory", result.Value.(error)), OK: false}
}
rows, err := storeInstance.database.Query(
"SELECT entry_id, bucket_name, measurement, fields_json, tags_json, committed_at FROM "+journalEntriesTableName+" WHERE archived_at IS NULL AND committed_at < ? ORDER BY committed_at",
options.Before.UnixMilli(),
)
if err != nil {
return core.Result{Value: core.E("store.Compact", "query journal rows", err), OK: false}
}
defer rows.Close()
var archiveEntries []compactArchiveEntry
for rows.Next() {
var entry compactArchiveEntry
if err := rows.Scan(
&entry.entryID,
&entry.bucketName,
&entry.measurementName,
&entry.fieldsJSON,
&entry.tagsJSON,
&entry.committedAtUnixMilli,
); err != nil {
return core.Result{Value: core.E("store.Compact", "scan journal row", err), OK: false}
}
archiveEntries = append(archiveEntries, entry)
}
if err := rows.Err(); err != nil {
return core.Result{Value: core.E("store.Compact", "iterate journal rows", err), OK: false}
}
if len(archiveEntries) == 0 {
return core.Result{Value: "", OK: true}
}
outputPath := compactOutputPath(outputDirectory, format)
archiveFileResult := filesystem.Create(outputPath)
if !archiveFileResult.OK {
return core.Result{Value: core.E("store.Compact", "create archive file", archiveFileResult.Value.(error)), OK: false}
}
file, ok := archiveFileResult.Value.(io.WriteCloser)
if !ok {
return core.Result{Value: core.E("store.Compact", "archive file is not writable", nil), OK: false}
}
fileClosed := false
defer func() {
if !fileClosed {
_ = file.Close()
}
}()
writer, err := archiveWriter(file, format)
if err != nil {
return core.Result{Value: err, OK: false}
}
writeOK := false
defer func() {
if !writeOK {
_ = writer.Close()
}
}()
for _, entry := range archiveEntries {
lineMap, err := compactArchiveLine(entry)
if err != nil {
return core.Result{Value: err, OK: false}
}
lineJSON, err := jsonString(lineMap, "store.Compact", "marshal archive line")
if err != nil {
return core.Result{Value: err, OK: false}
}
if _, err := io.WriteString(writer, lineJSON+"\n"); err != nil {
return core.Result{Value: core.E("store.Compact", "write archive line", err), OK: false}
}
}
if err := writer.Close(); err != nil {
return core.Result{Value: core.E("store.Compact", "close archive writer", err), OK: false}
}
writeOK = true
if err := file.Close(); err != nil {
return core.Result{Value: core.E("store.Compact", "close archive file", err), OK: false}
}
fileClosed = true
transaction, err := storeInstance.database.Begin()
if err != nil {
return core.Result{Value: core.E("store.Compact", "begin archive transaction", err), OK: false}
}
committed := false
defer func() {
if !committed {
_ = transaction.Rollback()
}
}()
archivedAt := time.Now().UnixMilli()
for _, entry := range archiveEntries {
if _, err := transaction.Exec(
"UPDATE "+journalEntriesTableName+" SET archived_at = ? WHERE entry_id = ?",
archivedAt,
entry.entryID,
); err != nil {
return core.Result{Value: core.E("store.Compact", "mark journal row archived", err), OK: false}
}
}
if err := transaction.Commit(); err != nil {
return core.Result{Value: core.E("store.Compact", "commit archive transaction", err), OK: false}
}
committed = true
return core.Result{Value: outputPath, OK: true}
}
func compactArchiveLine(entry compactArchiveEntry) (map[string]any, error) {
fields := make(map[string]any)
fieldsResult := core.JSONUnmarshalString(entry.fieldsJSON, &fields)
if !fieldsResult.OK {
return nil, core.E("store.Compact", "unmarshal fields", fieldsResult.Value.(error))
}
tags := make(map[string]string)
tagsResult := core.JSONUnmarshalString(entry.tagsJSON, &tags)
if !tagsResult.OK {
return nil, core.E("store.Compact", "unmarshal tags", tagsResult.Value.(error))
}
return map[string]any{
"bucket": entry.bucketName,
"measurement": entry.measurementName,
"fields": fields,
"tags": tags,
"committed_at": entry.committedAtUnixMilli,
}, nil
}
func archiveWriter(writer io.Writer, format string) (io.WriteCloser, error) {
switch format {
case "gzip":
return gzip.NewWriter(writer), nil
case "zstd":
zstdWriter, err := zstd.NewWriter(writer)
if err != nil {
return nil, core.E("store.Compact", "create zstd writer", err)
}
return zstdWriter, nil
default:
return nil, core.E("store.Compact", core.Concat("unsupported archive format: ", format), nil)
}
}
func compactOutputPath(outputDirectory, format string) string {
extension := ".jsonl"
if format == "gzip" {
extension = ".jsonl.gz"
}
if format == "zstd" {
extension = ".jsonl.zst"
}
filename := core.Concat("journal-", time.Now().UTC().Format("20060102-150405"), extension)
return joinPath(outputDirectory, filename)
}