Align the watcher examples with the current API and sort recovered workspaces for predictable output. Co-Authored-By: Virgil <virgil@lethean.io>
12 KiB
| title | description |
|---|---|
| Architecture | Internal design of go-store -- storage layer, group/key model, TTL expiry, event system, namespace isolation, and concurrency model. |
Architecture
This document describes how go-store works internally. It covers the storage layer, the data model, TTL expiry, the event system, namespace isolation with quota enforcement, and the concurrency model.
Storage Layer
SQLite with WAL Mode
Every Store instance opens a single SQLite database and immediately applies two pragmas:
PRAGMA journal_mode=WAL;
PRAGMA busy_timeout=5000;
WAL (Write-Ahead Logging) mode allows concurrent readers to proceed without blocking writers. The busy_timeout of 5000 milliseconds causes the driver to wait and retry rather than immediately returning SQLITE_BUSY under write contention.
Single Connection Constraint
The database/sql package maintains a connection pool by default. SQLite pragmas are per-connection: if the pool hands out a second connection, that connection inherits none of the WAL or busy-timeout settings, causing SQLITE_BUSY errors under concurrent load.
go-store calls database.SetMaxOpenConns(1) to pin all access to a single connection. Since SQLite serialises writes at the file level regardless, this introduces no additional throughput penalty. It eliminates the BUSY errors by ensuring the pragma settings always apply.
Schema
CREATE TABLE IF NOT EXISTS entries (
group_name TEXT NOT NULL,
entry_key TEXT NOT NULL,
entry_value TEXT NOT NULL,
expires_at INTEGER,
PRIMARY KEY (group_name, entry_key)
)
The compound primary key (group_name, entry_key) enforces uniqueness per group-key pair and provides efficient indexed lookups. The expires_at column stores a Unix millisecond timestamp (nullable); a NULL value means the key never expires.
Schema migration. Databases created before the AX schema rename used a legacy key-value table. On New(), go-store migrates that legacy table into entries, preserving rows and copying the expiry data when present. Databases that already have entries but lack expires_at still receive an additive ALTER TABLE entries ADD COLUMN expires_at INTEGER migration; if the column already exists, SQLite returns a "duplicate column" error which is silently ignored.
Group/Key Model
Keys are addressed by a two-level path: (group, key). Groups act as logical namespaces within a single database. Groups are implicit -- they exist as a consequence of the keys they contain and are destroyed automatically when all their keys are deleted.
This model maps naturally to domain concepts:
group: "user:42:config" key: "colour"
group: "user:42:config" key: "language"
group: "session:abc" key: "token"
All read operations (Get, GetAll, Count, Render) are scoped to a single group. DeleteGroup atomically removes all keys in a group. CountAll and Groups operate across groups by prefix match.
UPSERT Semantics
All writes use INSERT ... ON CONFLICT(group_name, entry_key) DO UPDATE. This means:
- Inserting a new key creates it.
- Inserting an existing key overwrites its value and (for
Set) clears any TTL. - UPSERT never duplicates a key.
- The operation is idempotent with respect to row count.
Set clears expires_at on upsert by setting it to NULL. SetWithTTL refreshes the expiry timestamp on upsert.
TTL Expiry
Keys may be created with a time-to-live via SetWithTTL. Expiry is stored as a Unix millisecond timestamp in expires_at.
Expiry is enforced in three ways:
1. Lazy Deletion on Get
If a key is found but its expires_at is in the past, it is deleted synchronously before returning NotFoundError. This prevents stale values from being returned even if the background purge has not run yet.
2. Query-Time Filtering
All bulk operations (GetAll, All, Count, Render, CountAll, Groups, GroupsSeq) include (expires_at IS NULL OR expires_at > ?) in their WHERE clause. Expired keys are excluded from results without being deleted.
3. Background Purge Goroutine
New() launches a goroutine that calls PurgeExpired() every 60 seconds. This recovers disk space by physically removing expired rows. The goroutine is stopped cleanly by Close() via context.WithCancel and sync.WaitGroup.
PurgeExpired() is also available as a public method for applications that want manual control over purge timing.
String Splitting Helpers
Two convenience methods build on Get to return iterators over parts of a stored value:
GetSplit(group, key, separator)splits the value by a custom separator, returning aniter.Seq[string]viacore.Split.GetFields(group, key)splits the value by whitespace, returning aniter.Seq[string]via the package's internal field iterator.
core.Split keeps the package free of direct strings imports while preserving the same agent-facing API shape.
Both return NotFoundError if the key does not exist or has expired.
Template Rendering
Render(templateSource, group) is a convenience method that fetches all non-expired key-value pairs from a group and renders a Go text/template against them. The template data is a map[string]string keyed by the field name.
if err := storeInstance.Set("miner", "pool", "pool.lthn.io:3333"); err != nil {
return
}
if err := storeInstance.Set("miner", "wallet", "iz..."); err != nil {
return
}
renderedTemplate, err := storeInstance.Render(`{"pool":"{{ .pool }}","wallet":"{{ .wallet }}"}`, "miner")
if err != nil {
return
}
// renderedTemplate: {"pool":"pool.lthn.io:3333","wallet":"iz..."}
Template parse errors and execution errors are both returned as wrapped errors with context (e.g., store.Render: parse template: ... and store.Render: execute template: ...).
Missing template variables do not return an error by default -- Go's text/template renders them as <no value>. Applications requiring strict variable presence should validate data beforehand.
Event System
go-store provides two mechanisms for observing mutations: channel-based watchers and synchronous callbacks. Both are defined in events.go.
Event Model
type Event struct {
Type EventType
Group string
Key string
Value string
Timestamp time.Time
}
| EventType | String() | Key populated | Value populated |
|---|---|---|---|
EventSet |
"set" |
Yes | Yes |
EventDelete |
"delete" |
Yes | No |
EventDeleteGroup |
"delete_group" |
No (empty) | No |
Events are emitted synchronously after each successful database write inside the internal notify() method.
Watch/Unwatch
Watch(group) creates a buffered event channel (<-chan Event, capacity 16).
| group argument | Receives |
|---|---|
"mygroup" |
Mutations within that group, including DeleteGroup |
"*" |
Every mutation across the entire store |
Unwatch(group, events) removes the watcher from the registry and closes its channel. It is safe to call multiple times; subsequent calls are no-ops.
Backpressure. Event dispatch to a watcher channel is non-blocking: if the channel buffer is full, the event is dropped silently. This prevents a slow consumer from blocking a writer. Applications that cannot afford dropped events should drain the channel promptly or use OnChange callbacks instead.
events := storeInstance.Watch("config")
defer storeInstance.Unwatch("config", events)
for event := range events {
fmt.Println(event.Type, event.Group, event.Key, event.Value)
}
OnChange Callbacks
OnChange(callback func(Event)) registers a synchronous callback that fires on every mutation. The callback runs in the goroutine that performed the write. Returns an idempotent unregister function.
This is the designed integration point for consumers such as go-ws:
unregister := storeInstance.OnChange(func(event store.Event) {
hub.SendToChannel("store-events", event)
})
defer unregister()
go-store does not import go-ws. The dependency flows in one direction only: go-ws (or any consumer) imports go-store. Callbacks may safely register or unregister watchers and callbacks while handling an event. Dispatch snapshots the callback list before invoking it, so re-entrant subscription management does not deadlock. Offload any significant work to a separate goroutine if needed.
Internal Dispatch
The notify(event Event) method first acquires the watcher read-lock, iterates all watchers with non-blocking channel sends, then releases the lock. It then acquires the callback read-lock, snapshots the registered callbacks, releases the lock, and invokes each callback synchronously. This keeps watcher delivery non-blocking while allowing callbacks to manage subscriptions re-entrantly.
Watcher delivery is grouped by the registered group name. Wildcard "*" matches every mutation across the entire store.
Namespace Isolation (ScopedStore)
ScopedStore wraps a *Store and automatically prefixes all group names with namespace + ":". This prevents key collisions when multiple tenants share a single underlying database.
scopedStore, err := store.NewScoped(storeInstance, "tenant-42")
if err != nil {
return
}
if err := scopedStore.Set("config", "colour", "blue"); err != nil {
return
}
// Stored in underlying store as group="tenant-42:config", key="colour"
Namespace strings must match ^[a-zA-Z0-9-]+$. Invalid namespaces are rejected at construction time.
ScopedStore delegates all operations to the underlying Store after prefixing. Events emitted by scoped operations carry the full prefixed group name in Event.Group, enabling watchers on the underlying store to observe scoped mutations.
ScopedStore exposes the same read helpers as Store for Get, Set, SetWithTTL, Delete, DeleteGroup, GetAll, All, Count, CountAll, Groups, GroupsSeq, GetSplit, GetFields, Render, and PurgeExpired. Methods that return group names strip the namespace prefix before returning results. The Namespace() method returns the namespace string.
Quota Enforcement
NewScopedWithQuota(store, namespace, QuotaConfig) adds per-namespace limits. For example, store.QuotaConfig{MaxKeys: 100, MaxGroups: 10} caps a namespace at 100 keys and 10 groups:
type QuotaConfig struct {
MaxKeys int
MaxGroups int
}
Zero values mean unlimited. Before each Set or SetWithTTL, the scoped store:
- Checks whether the key already exists (upserts never consume quota).
- If the key is new, queries
CountAll(namespace + ":")and compares againstMaxKeys. - If the group is new (current count for that group is zero), queries
GroupsSeq(namespace + ":")and compares againstMaxGroups.
Exceeding a limit returns QuotaExceededError.
Concurrency Model
All SQLite access is serialised through a single connection (SetMaxOpenConns(1)). The store's event registry uses two separate sync.RWMutex instances: watchersLock for watcher registration and dispatch, and callbacksLock for callback registration and dispatch. These locks do not interact:
- Database writes acquire no application-level lock.
notify()acquireswatchersLock(read) after the database write completes, thencallbacksLock(read) to snapshot callbacks.Watch/UnwatchacquirewatchersLock(write) to modify watcher registrations.OnChangeacquirescallbacksLock(write) to modify callback registrations.
All operations are safe to call from multiple goroutines concurrently. The race detector is clean under the project's standard test suite (go test -race ./...).
File Layout
doc.go Package comment with concrete usage examples
store.go Core Store type, CRUD, TTL, background purge, iterators, rendering
events.go EventType, Event, Watcher, OnChange, notify
scope.go ScopedStore, QuotaConfig, namespace-local helper delegation, quota enforcement
store_test.go Tests: CRUD, TTL, concurrency, edge cases, persistence
events_test.go Tests: Watch, Unwatch, OnChange, event dispatch
scope_test.go Tests: namespace isolation, quota enforcement
coverage_test.go Tests: defensive error paths (scan errors, corruption)
bench_test.go Performance benchmarks