Skip to content

memoir.store package

memoir.store

Store layer for Memoir memory system.

This module contains storage implementations for different backends.

Submodules

memoir.store.prolly_adapter module

memoir.store.prolly_adapter

ProllyTree adapter implementing LangGraph's BaseStore interface. Provides high-performance semantic memory storage with versioning.

MemoryItem

Bases: BaseModel

Represents a memory item in the store.

Source code in src/memoir/store/prolly_adapter.py
class MemoryItem(BaseModel):
    """Represents a memory item in the store."""

    key: str = Field(description="Semantic taxonomy key")
    namespace: str = Field(description="User/agent namespace")
    content: Any = Field(description="Memory content")
    metadata: dict[str, Any] = Field(
        default_factory=dict, description="Additional metadata"
    )
    timestamp: float = Field(
        default_factory=time.time, description="Creation timestamp"
    )
    version: str | None = Field(default=None, description="Version/commit ID")
    confidence: float = Field(default=1.0, description="Classification confidence")

AggregatedMemory

Bases: BaseModel

Represents aggregated memories at a semantic path.

Source code in src/memoir/store/prolly_adapter.py
class AggregatedMemory(BaseModel):
    """Represents aggregated memories at a semantic path."""

    path: str = Field(description="Semantic taxonomy path")
    memories: list[dict[str, Any]] = Field(
        default_factory=list, description="List of memory entries at this path"
    )
    count: int = Field(default=0, description="Number of memories")
    first_timestamp: float = Field(
        default_factory=time.time, description="Timestamp of first memory"
    )
    last_timestamp: float = Field(
        default_factory=time.time, description="Timestamp of last memory"
    )
    last_updated: float = Field(
        default_factory=time.time, description="Last update timestamp"
    )

ProllyTreeStore

Bases: BaseStore

High-performance semantic memory store using ProllyTree. Implements LangGraph's BaseStore interface following the reference pattern.

Source code in src/memoir/store/prolly_adapter.py
 97
 98
 99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
class ProllyTreeStore(BaseStore):
    """
    High-performance semantic memory store using ProllyTree.
    Implements LangGraph's BaseStore interface following the reference pattern.
    """

    def __init__(
        self,
        path: str,
        enable_versioning: bool = True,
        auto_commit: bool = True,
        cache_size: int = 10000,
    ):
        """
        Initialize ProllyTree store.

        Storage layer is responsible only for storing and retrieving data.
        Classification is handled by higher layers (memory manager).

        ProllyTreeStore is strict: it opens an existing memoir store and
        refuses paths that aren't one yet. Use ``StoreService.create_store``
        (or ``memoir new``) to bootstrap a fresh store. Single init path =
        no surprise side-effects from running `memoir remember` in a random
        cwd.

        Args:
            path: Path to an existing memoir store (must contain a ``.git``
                directory when ``enable_versioning`` is True).
            enable_versioning: Whether to enable git-like versioning
            auto_commit: Whether to automatically commit on each put/delete operation
            cache_size: Size of internal caches
        """
        super().__init__()

        self.path = Path(path)
        self.path.mkdir(parents=True, exist_ok=True)

        if enable_versioning and not (self.path / ".git").exists():
            raise FileNotFoundError(
                f"Not a memoir store: {self.path} (no .git directory). "
                f"Create one with `memoir new <path>` first, or pass "
                f"-s/--store / set MEMOIR_STORE / cd into an existing store."
            )

        # Initialize ProllyTree
        if enable_versioning:
            # Create data subdirectory for VersionedKvStore
            data_dir = self.path / "data"
            data_dir.mkdir(exist_ok=True)
            # VersionedKvStore (prollytree Rust binding) uses cwd to locate the
            # enclosing git repository even when handed an absolute path —
            # which means callers in non-git cwds (e.g. /tmp, ~/.memoir) get
            # "Not in a git repository" errors. Construction needs a chdir;
            # so do per-operation calls (`.insert`/`.update`/`.commit`/`.get`).
            # We chdir here for the constructor, then wrap the tree in
            # _CwdLockedTree so every later method call also chdir's first.
            import os as _os

            _saved_cwd = _os.getcwd()
            try:
                _os.chdir(str(self.path))
                _raw_tree = VersionedKvStore(str(data_dir))
            finally:
                _os.chdir(_saved_cwd)
            self.tree = _CwdLockedTree(_raw_tree, self.path)
        else:
            # Memory mode doesn't touch git, so no cwd wrapper needed.
            self.tree = ProllyTree("memory")

        self.enable_versioning = enable_versioning
        self.auto_commit = auto_commit
        # Storage layer doesn't need taxonomy, classifier, or search engine
        # These are handled by higher layers

        # Performance tracking
        self._stats = {"reads": 0, "writes": 0, "searches": 0, "classifications": 0}

        # Key registry for memory mode (since ProllyTree doesn't have list_keys in memory mode)
        self._keys = set()

        # Populate key registry from existing data
        self._populate_key_registry()

        # Track aggregated memories to avoid redundant updates
        self._aggregation_cache = {}

    def _populate_key_registry(self):
        """Populate the key registry from existing data in the store."""
        try:
            if hasattr(self.tree, "scan"):
                # Use scan if available to iterate through all keys
                for key_bytes, _ in self.tree.scan():
                    key_str = key_bytes.decode("utf-8")
                    self._keys.add(key_str)
            elif hasattr(self.tree, "list_keys"):
                # Use list_keys if available
                for key_bytes in self.tree.list_keys():
                    key_str = key_bytes.decode("utf-8")
                    self._keys.add(key_str)
            else:
                # No way to enumerate keys, registry will be empty initially
                # Keys will be added as they are accessed via put()
                pass

            logger.info(f"Populated key registry with {len(self._keys)} existing keys")
        except Exception as e:
            logger.warning(f"Could not populate key registry: {e}")
            # Continue without existing keys - they'll be added as accessed

    def _encode_value(self, value: Any) -> bytes:
        """Encode any value to bytes for storage."""
        if isinstance(value, bytes):
            return value
        elif isinstance(value, str):
            return value.encode("utf-8")
        else:
            # Use JSON for complex objects
            json_str = json.dumps(value, default=str)
            return json_str.encode("utf-8")

    def _decode_value(self, data: bytes) -> Any:
        """Decode bytes from storage back to original type."""
        if not data:
            return None
        try:
            # Try to decode as JSON first
            json_str = data.decode("utf-8")
            return json.loads(json_str)
        except (json.JSONDecodeError, UnicodeDecodeError):
            # Return as string if not JSON
            try:
                return data.decode("utf-8")
            except UnicodeDecodeError:
                return data

    # BaseStore interface methods
    def batch(self, ops: list[tuple]) -> list[Any]:
        """Batch operations - required by BaseStore."""
        results = []
        for op in ops:
            if len(op) == 2:
                method, args = op
                result = getattr(self, method)(*args)
                results.append(result)
        return results

    def abatch(self, ops: list[tuple]) -> list[Any]:
        """Async batch operations - synchronous implementation."""
        return self.batch(ops)

    def search(
        self, namespace: tuple, *, filter: dict | None = None, limit: int = 10
    ) -> list[tuple]:
        """Search for items in a namespace."""
        self._stats["searches"] += 1
        prefix = ":".join(namespace) + ":"
        results = []

        try:
            # Use our key registry to find matching keys
            count = 0
            for full_key in self._keys:
                if count >= limit:
                    break

                if full_key.startswith(prefix):
                    key_bytes = full_key.encode("utf-8")
                    if self.enable_versioning:
                        value = self.tree.get(key_bytes)
                    else:
                        value = self.tree.find(key_bytes)
                    decoded_value = self._decode_value(value)

                    # Apply filter if provided
                    if filter and not all(
                        decoded_value.get(k) == v
                        for k, v in filter.items()
                        if isinstance(decoded_value, dict)
                    ):
                        continue

                    # Extract item key from full key
                    item_key = full_key[len(prefix) :]
                    results.append((namespace, item_key, decoded_value))
                    count += 1
        except Exception as e:
            logger.error(f"Error searching namespace {namespace}: {e}")

        return results

    def put(self, namespace: tuple, key: str, value: dict) -> None:
        """Store a value in a namespace."""
        self._stats["writes"] += 1
        full_key = ":".join(namespace) + ":" + key
        key_bytes = full_key.encode("utf-8")
        value_bytes = self._encode_value(value)

        try:
            if self.enable_versioning:
                # VersionedKvStore API - check if key exists using get
                existing = self.tree.get(key_bytes)
                if existing:
                    self.tree.update(key_bytes, value_bytes)
                else:
                    self.tree.insert(key_bytes, value_bytes)
                # Commit the change if auto_commit is enabled
                if self.auto_commit:
                    self.tree.commit(f"Store {key} in {':'.join(namespace)}")
            else:
                # ProllyTree API - check if key exists using find
                existing = self.tree.find(key_bytes)
                if existing:
                    self.tree.update(key_bytes, value_bytes)
                else:
                    self.tree.insert(key_bytes, value_bytes)

            # Track the key in our registry
            self._keys.add(full_key)

        except Exception as e:
            logger.error(f"Error storing {full_key}: {e}")
            raise

    def get(self, namespace: tuple, key: str) -> dict | None:
        """Retrieve a value from a namespace."""
        self._stats["reads"] += 1
        full_key = ":".join(namespace) + ":" + key
        key_bytes = full_key.encode("utf-8")

        try:
            if self.enable_versioning:
                # VersionedKvStore API
                data = self.tree.get(key_bytes)
            else:
                # ProllyTree API
                data = self.tree.find(key_bytes)
            return self._decode_value(data) if data else None
        except Exception as e:
            logger.error(f"Error getting key {full_key}: {e}")
            return None

    def delete(self, namespace: tuple, key: str) -> None:
        """Delete a key from a namespace."""
        full_key = ":".join(namespace) + ":" + key
        key_bytes = full_key.encode("utf-8")

        try:
            self.tree.delete(key_bytes)
            # Remove from key registry
            self._keys.discard(full_key)
            if self.enable_versioning and self.auto_commit:
                self.tree.commit(f"Delete {key} from {':'.join(namespace)}")
        except Exception as e:
            logger.error(f"Error deleting {full_key}: {e}")

    def commit(self, message: str = "Manual commit") -> str | None:
        """
        Manually commit pending changes to the versioned store.

        This is useful when auto_commit is disabled and you want to batch
        multiple operations before committing.

        Args:
            message: Commit message

        Returns:
            Commit hash if versioning is enabled, None otherwise
        """
        if not self.enable_versioning:
            logger.warning("Commit requested but versioning is not enabled")
            return None

        try:
            commit_hash = self.tree.commit(message)
            logger.debug(f"Manual commit successful: {message}")
            return commit_hash
        except Exception as e:
            logger.error(f"Error committing changes: {e}")
            raise

    def get_key_history(
        self, namespace: tuple, key: str, limit: int = 10
    ) -> list[dict]:
        """
        Get commit history for a specific key.

        Args:
            namespace: Namespace tuple
            key: Key to get history for
            limit: Maximum number of commits to return

        Returns:
            List of commit dictionaries with id, timestamp, message, author, committer
        """
        if not self.enable_versioning:
            return []

        full_key = ":".join(namespace) + ":" + key
        key_bytes = full_key.encode("utf-8")

        try:
            commits = self.tree.get_commits_for_key(key_bytes)
            # Limit results and return most recent first
            return commits[:limit]
        except Exception as e:
            logger.error(f"Error getting history for {full_key}: {e}")
            return []

    def get_key_at_commit(
        self, namespace: tuple, key: str, commit_id: str
    ) -> dict | None:
        """
        Get the value of a key at a specific commit.

        Note: Current implementation returns None since VersionedKvStore doesn't support
        direct commit checkout. This is a placeholder for future enhancement.

        Args:
            namespace: Namespace tuple
            key: Key to retrieve
            commit_id: Commit ID to retrieve from

        Returns:
            None (historical content retrieval not yet implemented)
        """
        if not self.enable_versioning:
            return None

        # TODO: Implement historical content retrieval when VersionedKvStore supports it
        # Current limitation: VersionedKvStore only supports branch checkout, not commit checkout
        logger.debug(
            f"Historical content retrieval not yet implemented for commit {commit_id[:8]}"
        )
        return None

    def create_time_snapshot(self, snapshot_name: str) -> bool:
        """
        Create a branch snapshot at the current point in time.

        When auto_commit=False, this will first commit any pending changes
        before creating the snapshot to ensure all recent changes are included.

        Args:
            snapshot_name: Name for the snapshot branch

        Returns:
            True if snapshot created successfully
        """
        if not self.enable_versioning:
            return False

        try:
            # If auto_commit is disabled, commit pending changes before snapshot
            if not self.auto_commit:
                commit_hash = self.commit(
                    f"Auto-commit before snapshot: {snapshot_name}"
                )
                if commit_hash:
                    logger.debug(
                        f"Auto-committed pending changes before snapshot: {commit_hash[:8]}"
                    )

            self.tree.create_branch(snapshot_name)
            logger.debug(f"Created time snapshot: {snapshot_name}")
            return True
        except Exception as e:
            logger.error(f"Failed to create snapshot {snapshot_name}: {e}")
            return False

    def get_state_at_snapshot(
        self, namespace: tuple, snapshot_name: str
    ) -> dict[str, Any]:
        """
        Get all keys in a namespace at a specific snapshot.

        Args:
            namespace: Namespace tuple
            snapshot_name: Name of the snapshot branch

        Returns:
            Dictionary of key -> value at that snapshot
        """
        if not self.enable_versioning:
            return {}

        try:
            # Save current branch
            current_branch = self.tree.current_branch()

            # Switch to snapshot
            self.tree.checkout(snapshot_name)

            # Get all keys in namespace
            state = {}
            namespace_prefix = ":".join(namespace) + ":"

            keys = self.tree.list_keys()
            for key in keys:
                key_str = key.decode("utf-8") if isinstance(key, bytes) else key
                if key_str.startswith(namespace_prefix):
                    # Get value
                    value = self.tree.get(
                        key if isinstance(key, bytes) else key.encode("utf-8")
                    )
                    if value:
                        # Extract the key without namespace prefix
                        short_key = key_str[len(namespace_prefix) :]
                        state[short_key] = self._decode_value(value)

            # Return to original branch
            self.tree.checkout(current_branch)

            return state

        except Exception as e:
            logger.error(f"Failed to get state at snapshot {snapshot_name}: {e}")
            # Try to return to original branch
            with contextlib.suppress(Exception):
                self.tree.checkout(current_branch)
            return {}

    # Enhanced methods for semantic memory functionality
    async def store_memory_async(
        self, namespace: str, content: Any, key: str
    ) -> MemoryItem:
        """
        Store a memory at the given semantic key.

        Note: Classification must be done by the caller (memory manager).
        Storage layer is responsible only for storing, not classifying.

        Args:
            namespace: User/agent namespace
            content: Memory content to store
            key: Semantic key where to store (REQUIRED - no classification here)

        Returns:
            MemoryItem with storage results
        """
        # Storage layer: just use the provided semantic key (no classification)
        semantic_key = key
        confidence = 1.0  # Confidence is determined by the caller (memory manager)

        # Use semantic key for aggregation
        storage_key = semantic_key

        # Create memory entry (not the full item)
        memory_entry = {
            "content": content,
            "confidence": confidence,
            "timestamp": time.time(),
            "metadata": {},
        }

        # Convert namespace to tuple format
        if ":" in namespace:
            namespace_parts = namespace.split(":")
            namespace_tuple = tuple(namespace_parts)
        else:
            namespace_tuple = (namespace,)

        # Get existing aggregated memory or create new one
        existing = self.get(namespace_tuple, storage_key)

        if existing and isinstance(existing, dict) and "memories" in existing:
            # Append to existing aggregated memory
            aggregated = AggregatedMemory(**existing)
            aggregated.memories.append(memory_entry)
            aggregated.count += 1
            aggregated.last_timestamp = memory_entry["timestamp"]
            aggregated.last_updated = time.time()
        else:
            # Create new aggregated memory
            aggregated = AggregatedMemory(
                path=semantic_key,
                memories=[memory_entry],
                count=1,
                first_timestamp=memory_entry["timestamp"],
                last_timestamp=memory_entry["timestamp"],
            )

        # Store the aggregated memory
        self.put(namespace_tuple, storage_key, aggregated.model_dump())

        # Create MemoryItem for return value (for compatibility)
        item = MemoryItem(
            key=semantic_key,
            namespace=namespace,
            content=content,
            confidence=confidence,
            timestamp=memory_entry["timestamp"],
        )

        if self.enable_versioning and hasattr(self.tree, "get_head"):
            item.version = self.tree.get_head()

        return item

    # Sync store_memory method removed - use store_memory_async for all operations
    # This eliminates the async/sync mismatch and fallback issues

    async def asearch(self, namespace: str, path_prefix: str) -> list[tuple[str, Any]]:
        """
        Async search for items with a given path prefix.
        Used by HierarchicalSearchEngine.

        Args:
            namespace: User namespace
            path_prefix: Path prefix to search for

        Returns:
            List of (semantic_key, data) tuples
        """
        # Use synchronous search with prefix
        results = []
        # Convert string namespace to tuple format
        # "memory:general" -> ("memory", "general")
        namespace_parts = namespace.split(":")
        namespace_tuple = tuple(namespace_parts)

        search_results = self.search(namespace_tuple, limit=100)

        for _, storage_key, data in search_results:
            semantic_key = storage_key

            # Check if semantic path matches prefix
            if semantic_key.startswith(path_prefix):
                # For aggregated memories, we return them as-is
                # The search engine will handle expanding them
                if isinstance(data, dict) and "memories" in data:
                    # This is an aggregated memory - return it
                    results.append((semantic_key, data))
                else:
                    # Legacy single memory format
                    results.append((semantic_key, data))

        return results

    async def retrieve_memories_async(
        self, namespace: str, query: str, limit: int = 10
    ) -> list[MemoryItem]:
        """
        Retrieve memories using semantic search (async version).

        Args:
            namespace: User/agent namespace
            query: Search query
            limit: Maximum number of results

        Returns:
            List of matching memory items
        """
        # Use the hierarchical search engine to find relevant memories
        search_results = await self.search_engine.search(query, namespace)

        # Convert search results to memory items with deduplication
        memories = []
        seen_content = set()

        for result in search_results:
            # The search result contains combined content from multiple items
            if result.combined_content:
                try:
                    # Split combined content back into individual memories
                    individual_contents = result.combined_content.split(" | ")
                    for content_text in individual_contents:
                        if content_text.strip():
                            # Create a memory item from the content
                            memory = MemoryItem(
                                key=result.path,
                                namespace=result.namespace,
                                content=content_text.strip(),
                                confidence=1.0,  # Default confidence
                                timestamp=time.time(),
                            )
                            # Deduplicate by content
                            content_hash = hash(memory.content)
                            if content_hash not in seen_content:
                                seen_content.add(content_hash)
                                memories.append(memory)
                                # Stop when we have enough unique results
                                if len(memories) >= limit:
                                    break
                    if len(memories) >= limit:
                        break
                except Exception as e:
                    logger.warning(f"Failed to parse memory item: {e}")

        return memories

    def retrieve_memories(
        self, namespace: str, query: str, limit: int = 10
    ) -> list[MemoryItem]:
        """
        Retrieve memories using semantic search (sync fallback).

        Note: This is a simple fallback. For proper semantic search,
        use retrieve_memories_async() which leverages the HierarchicalSearchEngine.

        Args:
            namespace: User/agent namespace
            query: Search query
            limit: Maximum number of results

        Returns:
            List of matching memory items
        """
        logger.warning(
            "Using fallback sync search. For better results, use retrieve_memories_async()"
        )

        # Simple fallback - just return all memories
        all_memories = []
        search_results = self.search((namespace,), limit=limit)

        for _, _key, data in search_results:
            if isinstance(data, dict):
                try:
                    memory = MemoryItem(**data)
                    all_memories.append(memory)
                except Exception as e:
                    logger.warning(f"Failed to parse memory item: {e}")

        return all_memories

    def get_statistics(self) -> dict[str, Any]:
        """Get store statistics."""
        stats = {
            "performance": self._stats.copy(),
            "total_keys": len(self._keys),
            "total_namespaces": len({key.split(":")[0] for key in self._keys}),
        }

        if self.enable_versioning and hasattr(self.tree, "get_head"):
            try:
                stats["versioning"] = {
                    "current_commit": self.tree.get_head(),
                }
                if hasattr(self.tree, "log"):
                    commits = self.tree.log()
                    stats["versioning"]["total_commits"] = len(commits)
            except Exception:
                pass

        return stats

    def export_namespace(self, namespace: str, output_path: str) -> None:
        """
        Export all memories from a namespace to JSON.

        Args:
            namespace: Namespace to export
            output_path: Path to save JSON file
        """
        memories = {}
        search_results = self.search((namespace,), limit=1000)

        for _, key, data in search_results:
            memories[key] = data

        with open(output_path, "w") as f:
            json.dump(
                {
                    "namespace": namespace,
                    "timestamp": time.time(),
                    "memories": memories,
                },
                f,
                indent=2,
            )

        logger.info(f"Exported {len(memories)} memories to {output_path}")

__init__

__init__(path: str, enable_versioning: bool = True, auto_commit: bool = True, cache_size: int = 10000)

Initialize ProllyTree store.

Storage layer is responsible only for storing and retrieving data. Classification is handled by higher layers (memory manager).

ProllyTreeStore is strict: it opens an existing memoir store and refuses paths that aren't one yet. Use StoreService.create_store (or memoir new) to bootstrap a fresh store. Single init path = no surprise side-effects from running memoir remember in a random cwd.

Parameters:

Name Type Description Default
path str

Path to an existing memoir store (must contain a .git directory when enable_versioning is True).

required
enable_versioning bool

Whether to enable git-like versioning

True
auto_commit bool

Whether to automatically commit on each put/delete operation

True
cache_size int

Size of internal caches

10000
Source code in src/memoir/store/prolly_adapter.py
def __init__(
    self,
    path: str,
    enable_versioning: bool = True,
    auto_commit: bool = True,
    cache_size: int = 10000,
):
    """
    Initialize ProllyTree store.

    Storage layer is responsible only for storing and retrieving data.
    Classification is handled by higher layers (memory manager).

    ProllyTreeStore is strict: it opens an existing memoir store and
    refuses paths that aren't one yet. Use ``StoreService.create_store``
    (or ``memoir new``) to bootstrap a fresh store. Single init path =
    no surprise side-effects from running `memoir remember` in a random
    cwd.

    Args:
        path: Path to an existing memoir store (must contain a ``.git``
            directory when ``enable_versioning`` is True).
        enable_versioning: Whether to enable git-like versioning
        auto_commit: Whether to automatically commit on each put/delete operation
        cache_size: Size of internal caches
    """
    super().__init__()

    self.path = Path(path)
    self.path.mkdir(parents=True, exist_ok=True)

    if enable_versioning and not (self.path / ".git").exists():
        raise FileNotFoundError(
            f"Not a memoir store: {self.path} (no .git directory). "
            f"Create one with `memoir new <path>` first, or pass "
            f"-s/--store / set MEMOIR_STORE / cd into an existing store."
        )

    # Initialize ProllyTree
    if enable_versioning:
        # Create data subdirectory for VersionedKvStore
        data_dir = self.path / "data"
        data_dir.mkdir(exist_ok=True)
        # VersionedKvStore (prollytree Rust binding) uses cwd to locate the
        # enclosing git repository even when handed an absolute path —
        # which means callers in non-git cwds (e.g. /tmp, ~/.memoir) get
        # "Not in a git repository" errors. Construction needs a chdir;
        # so do per-operation calls (`.insert`/`.update`/`.commit`/`.get`).
        # We chdir here for the constructor, then wrap the tree in
        # _CwdLockedTree so every later method call also chdir's first.
        import os as _os

        _saved_cwd = _os.getcwd()
        try:
            _os.chdir(str(self.path))
            _raw_tree = VersionedKvStore(str(data_dir))
        finally:
            _os.chdir(_saved_cwd)
        self.tree = _CwdLockedTree(_raw_tree, self.path)
    else:
        # Memory mode doesn't touch git, so no cwd wrapper needed.
        self.tree = ProllyTree("memory")

    self.enable_versioning = enable_versioning
    self.auto_commit = auto_commit
    # Storage layer doesn't need taxonomy, classifier, or search engine
    # These are handled by higher layers

    # Performance tracking
    self._stats = {"reads": 0, "writes": 0, "searches": 0, "classifications": 0}

    # Key registry for memory mode (since ProllyTree doesn't have list_keys in memory mode)
    self._keys = set()

    # Populate key registry from existing data
    self._populate_key_registry()

    # Track aggregated memories to avoid redundant updates
    self._aggregation_cache = {}

batch

batch(ops: list[tuple]) -> list[Any]

Batch operations - required by BaseStore.

Source code in src/memoir/store/prolly_adapter.py
def batch(self, ops: list[tuple]) -> list[Any]:
    """Batch operations - required by BaseStore."""
    results = []
    for op in ops:
        if len(op) == 2:
            method, args = op
            result = getattr(self, method)(*args)
            results.append(result)
    return results

abatch

abatch(ops: list[tuple]) -> list[Any]

Async batch operations - synchronous implementation.

Source code in src/memoir/store/prolly_adapter.py
def abatch(self, ops: list[tuple]) -> list[Any]:
    """Async batch operations - synchronous implementation."""
    return self.batch(ops)

search

search(namespace: tuple, *, filter: dict | None = None, limit: int = 10) -> list[tuple]

Search for items in a namespace.

Source code in src/memoir/store/prolly_adapter.py
def search(
    self, namespace: tuple, *, filter: dict | None = None, limit: int = 10
) -> list[tuple]:
    """Search for items in a namespace."""
    self._stats["searches"] += 1
    prefix = ":".join(namespace) + ":"
    results = []

    try:
        # Use our key registry to find matching keys
        count = 0
        for full_key in self._keys:
            if count >= limit:
                break

            if full_key.startswith(prefix):
                key_bytes = full_key.encode("utf-8")
                if self.enable_versioning:
                    value = self.tree.get(key_bytes)
                else:
                    value = self.tree.find(key_bytes)
                decoded_value = self._decode_value(value)

                # Apply filter if provided
                if filter and not all(
                    decoded_value.get(k) == v
                    for k, v in filter.items()
                    if isinstance(decoded_value, dict)
                ):
                    continue

                # Extract item key from full key
                item_key = full_key[len(prefix) :]
                results.append((namespace, item_key, decoded_value))
                count += 1
    except Exception as e:
        logger.error(f"Error searching namespace {namespace}: {e}")

    return results

put

put(namespace: tuple, key: str, value: dict) -> None

Store a value in a namespace.

Source code in src/memoir/store/prolly_adapter.py
def put(self, namespace: tuple, key: str, value: dict) -> None:
    """Store a value in a namespace."""
    self._stats["writes"] += 1
    full_key = ":".join(namespace) + ":" + key
    key_bytes = full_key.encode("utf-8")
    value_bytes = self._encode_value(value)

    try:
        if self.enable_versioning:
            # VersionedKvStore API - check if key exists using get
            existing = self.tree.get(key_bytes)
            if existing:
                self.tree.update(key_bytes, value_bytes)
            else:
                self.tree.insert(key_bytes, value_bytes)
            # Commit the change if auto_commit is enabled
            if self.auto_commit:
                self.tree.commit(f"Store {key} in {':'.join(namespace)}")
        else:
            # ProllyTree API - check if key exists using find
            existing = self.tree.find(key_bytes)
            if existing:
                self.tree.update(key_bytes, value_bytes)
            else:
                self.tree.insert(key_bytes, value_bytes)

        # Track the key in our registry
        self._keys.add(full_key)

    except Exception as e:
        logger.error(f"Error storing {full_key}: {e}")
        raise

get

get(namespace: tuple, key: str) -> dict | None

Retrieve a value from a namespace.

Source code in src/memoir/store/prolly_adapter.py
def get(self, namespace: tuple, key: str) -> dict | None:
    """Retrieve a value from a namespace."""
    self._stats["reads"] += 1
    full_key = ":".join(namespace) + ":" + key
    key_bytes = full_key.encode("utf-8")

    try:
        if self.enable_versioning:
            # VersionedKvStore API
            data = self.tree.get(key_bytes)
        else:
            # ProllyTree API
            data = self.tree.find(key_bytes)
        return self._decode_value(data) if data else None
    except Exception as e:
        logger.error(f"Error getting key {full_key}: {e}")
        return None

delete

delete(namespace: tuple, key: str) -> None

Delete a key from a namespace.

Source code in src/memoir/store/prolly_adapter.py
def delete(self, namespace: tuple, key: str) -> None:
    """Delete a key from a namespace."""
    full_key = ":".join(namespace) + ":" + key
    key_bytes = full_key.encode("utf-8")

    try:
        self.tree.delete(key_bytes)
        # Remove from key registry
        self._keys.discard(full_key)
        if self.enable_versioning and self.auto_commit:
            self.tree.commit(f"Delete {key} from {':'.join(namespace)}")
    except Exception as e:
        logger.error(f"Error deleting {full_key}: {e}")

commit

commit(message: str = 'Manual commit') -> str | None

Manually commit pending changes to the versioned store.

This is useful when auto_commit is disabled and you want to batch multiple operations before committing.

Parameters:

Name Type Description Default
message str

Commit message

'Manual commit'

Returns:

Type Description
str | None

Commit hash if versioning is enabled, None otherwise

Source code in src/memoir/store/prolly_adapter.py
def commit(self, message: str = "Manual commit") -> str | None:
    """
    Manually commit pending changes to the versioned store.

    This is useful when auto_commit is disabled and you want to batch
    multiple operations before committing.

    Args:
        message: Commit message

    Returns:
        Commit hash if versioning is enabled, None otherwise
    """
    if not self.enable_versioning:
        logger.warning("Commit requested but versioning is not enabled")
        return None

    try:
        commit_hash = self.tree.commit(message)
        logger.debug(f"Manual commit successful: {message}")
        return commit_hash
    except Exception as e:
        logger.error(f"Error committing changes: {e}")
        raise

get_key_history

get_key_history(namespace: tuple, key: str, limit: int = 10) -> list[dict]

Get commit history for a specific key.

Parameters:

Name Type Description Default
namespace tuple

Namespace tuple

required
key str

Key to get history for

required
limit int

Maximum number of commits to return

10

Returns:

Type Description
list[dict]

List of commit dictionaries with id, timestamp, message, author, committer

Source code in src/memoir/store/prolly_adapter.py
def get_key_history(
    self, namespace: tuple, key: str, limit: int = 10
) -> list[dict]:
    """
    Get commit history for a specific key.

    Args:
        namespace: Namespace tuple
        key: Key to get history for
        limit: Maximum number of commits to return

    Returns:
        List of commit dictionaries with id, timestamp, message, author, committer
    """
    if not self.enable_versioning:
        return []

    full_key = ":".join(namespace) + ":" + key
    key_bytes = full_key.encode("utf-8")

    try:
        commits = self.tree.get_commits_for_key(key_bytes)
        # Limit results and return most recent first
        return commits[:limit]
    except Exception as e:
        logger.error(f"Error getting history for {full_key}: {e}")
        return []

get_key_at_commit

get_key_at_commit(namespace: tuple, key: str, commit_id: str) -> dict | None

Get the value of a key at a specific commit.

Note: Current implementation returns None since VersionedKvStore doesn't support direct commit checkout. This is a placeholder for future enhancement.

Parameters:

Name Type Description Default
namespace tuple

Namespace tuple

required
key str

Key to retrieve

required
commit_id str

Commit ID to retrieve from

required

Returns:

Type Description
dict | None

None (historical content retrieval not yet implemented)

Source code in src/memoir/store/prolly_adapter.py
def get_key_at_commit(
    self, namespace: tuple, key: str, commit_id: str
) -> dict | None:
    """
    Get the value of a key at a specific commit.

    Note: Current implementation returns None since VersionedKvStore doesn't support
    direct commit checkout. This is a placeholder for future enhancement.

    Args:
        namespace: Namespace tuple
        key: Key to retrieve
        commit_id: Commit ID to retrieve from

    Returns:
        None (historical content retrieval not yet implemented)
    """
    if not self.enable_versioning:
        return None

    # TODO: Implement historical content retrieval when VersionedKvStore supports it
    # Current limitation: VersionedKvStore only supports branch checkout, not commit checkout
    logger.debug(
        f"Historical content retrieval not yet implemented for commit {commit_id[:8]}"
    )
    return None

create_time_snapshot

create_time_snapshot(snapshot_name: str) -> bool

Create a branch snapshot at the current point in time.

When auto_commit=False, this will first commit any pending changes before creating the snapshot to ensure all recent changes are included.

Parameters:

Name Type Description Default
snapshot_name str

Name for the snapshot branch

required

Returns:

Type Description
bool

True if snapshot created successfully

Source code in src/memoir/store/prolly_adapter.py
def create_time_snapshot(self, snapshot_name: str) -> bool:
    """
    Create a branch snapshot at the current point in time.

    When auto_commit=False, this will first commit any pending changes
    before creating the snapshot to ensure all recent changes are included.

    Args:
        snapshot_name: Name for the snapshot branch

    Returns:
        True if snapshot created successfully
    """
    if not self.enable_versioning:
        return False

    try:
        # If auto_commit is disabled, commit pending changes before snapshot
        if not self.auto_commit:
            commit_hash = self.commit(
                f"Auto-commit before snapshot: {snapshot_name}"
            )
            if commit_hash:
                logger.debug(
                    f"Auto-committed pending changes before snapshot: {commit_hash[:8]}"
                )

        self.tree.create_branch(snapshot_name)
        logger.debug(f"Created time snapshot: {snapshot_name}")
        return True
    except Exception as e:
        logger.error(f"Failed to create snapshot {snapshot_name}: {e}")
        return False

get_state_at_snapshot

get_state_at_snapshot(namespace: tuple, snapshot_name: str) -> dict[str, Any]

Get all keys in a namespace at a specific snapshot.

Parameters:

Name Type Description Default
namespace tuple

Namespace tuple

required
snapshot_name str

Name of the snapshot branch

required

Returns:

Type Description
dict[str, Any]

Dictionary of key -> value at that snapshot

Source code in src/memoir/store/prolly_adapter.py
def get_state_at_snapshot(
    self, namespace: tuple, snapshot_name: str
) -> dict[str, Any]:
    """
    Get all keys in a namespace at a specific snapshot.

    Args:
        namespace: Namespace tuple
        snapshot_name: Name of the snapshot branch

    Returns:
        Dictionary of key -> value at that snapshot
    """
    if not self.enable_versioning:
        return {}

    try:
        # Save current branch
        current_branch = self.tree.current_branch()

        # Switch to snapshot
        self.tree.checkout(snapshot_name)

        # Get all keys in namespace
        state = {}
        namespace_prefix = ":".join(namespace) + ":"

        keys = self.tree.list_keys()
        for key in keys:
            key_str = key.decode("utf-8") if isinstance(key, bytes) else key
            if key_str.startswith(namespace_prefix):
                # Get value
                value = self.tree.get(
                    key if isinstance(key, bytes) else key.encode("utf-8")
                )
                if value:
                    # Extract the key without namespace prefix
                    short_key = key_str[len(namespace_prefix) :]
                    state[short_key] = self._decode_value(value)

        # Return to original branch
        self.tree.checkout(current_branch)

        return state

    except Exception as e:
        logger.error(f"Failed to get state at snapshot {snapshot_name}: {e}")
        # Try to return to original branch
        with contextlib.suppress(Exception):
            self.tree.checkout(current_branch)
        return {}

store_memory_async async

store_memory_async(namespace: str, content: Any, key: str) -> MemoryItem

Store a memory at the given semantic key.

Note: Classification must be done by the caller (memory manager). Storage layer is responsible only for storing, not classifying.

Parameters:

Name Type Description Default
namespace str

User/agent namespace

required
content Any

Memory content to store

required
key str

Semantic key where to store (REQUIRED - no classification here)

required

Returns:

Type Description
MemoryItem

MemoryItem with storage results

Source code in src/memoir/store/prolly_adapter.py
async def store_memory_async(
    self, namespace: str, content: Any, key: str
) -> MemoryItem:
    """
    Store a memory at the given semantic key.

    Note: Classification must be done by the caller (memory manager).
    Storage layer is responsible only for storing, not classifying.

    Args:
        namespace: User/agent namespace
        content: Memory content to store
        key: Semantic key where to store (REQUIRED - no classification here)

    Returns:
        MemoryItem with storage results
    """
    # Storage layer: just use the provided semantic key (no classification)
    semantic_key = key
    confidence = 1.0  # Confidence is determined by the caller (memory manager)

    # Use semantic key for aggregation
    storage_key = semantic_key

    # Create memory entry (not the full item)
    memory_entry = {
        "content": content,
        "confidence": confidence,
        "timestamp": time.time(),
        "metadata": {},
    }

    # Convert namespace to tuple format
    if ":" in namespace:
        namespace_parts = namespace.split(":")
        namespace_tuple = tuple(namespace_parts)
    else:
        namespace_tuple = (namespace,)

    # Get existing aggregated memory or create new one
    existing = self.get(namespace_tuple, storage_key)

    if existing and isinstance(existing, dict) and "memories" in existing:
        # Append to existing aggregated memory
        aggregated = AggregatedMemory(**existing)
        aggregated.memories.append(memory_entry)
        aggregated.count += 1
        aggregated.last_timestamp = memory_entry["timestamp"]
        aggregated.last_updated = time.time()
    else:
        # Create new aggregated memory
        aggregated = AggregatedMemory(
            path=semantic_key,
            memories=[memory_entry],
            count=1,
            first_timestamp=memory_entry["timestamp"],
            last_timestamp=memory_entry["timestamp"],
        )

    # Store the aggregated memory
    self.put(namespace_tuple, storage_key, aggregated.model_dump())

    # Create MemoryItem for return value (for compatibility)
    item = MemoryItem(
        key=semantic_key,
        namespace=namespace,
        content=content,
        confidence=confidence,
        timestamp=memory_entry["timestamp"],
    )

    if self.enable_versioning and hasattr(self.tree, "get_head"):
        item.version = self.tree.get_head()

    return item

asearch async

asearch(namespace: str, path_prefix: str) -> list[tuple[str, Any]]

Async search for items with a given path prefix. Used by HierarchicalSearchEngine.

Parameters:

Name Type Description Default
namespace str

User namespace

required
path_prefix str

Path prefix to search for

required

Returns:

Type Description
list[tuple[str, Any]]

List of (semantic_key, data) tuples

Source code in src/memoir/store/prolly_adapter.py
async def asearch(self, namespace: str, path_prefix: str) -> list[tuple[str, Any]]:
    """
    Async search for items with a given path prefix.
    Used by HierarchicalSearchEngine.

    Args:
        namespace: User namespace
        path_prefix: Path prefix to search for

    Returns:
        List of (semantic_key, data) tuples
    """
    # Use synchronous search with prefix
    results = []
    # Convert string namespace to tuple format
    # "memory:general" -> ("memory", "general")
    namespace_parts = namespace.split(":")
    namespace_tuple = tuple(namespace_parts)

    search_results = self.search(namespace_tuple, limit=100)

    for _, storage_key, data in search_results:
        semantic_key = storage_key

        # Check if semantic path matches prefix
        if semantic_key.startswith(path_prefix):
            # For aggregated memories, we return them as-is
            # The search engine will handle expanding them
            if isinstance(data, dict) and "memories" in data:
                # This is an aggregated memory - return it
                results.append((semantic_key, data))
            else:
                # Legacy single memory format
                results.append((semantic_key, data))

    return results

retrieve_memories_async async

retrieve_memories_async(namespace: str, query: str, limit: int = 10) -> list[MemoryItem]

Retrieve memories using semantic search (async version).

Parameters:

Name Type Description Default
namespace str

User/agent namespace

required
query str

Search query

required
limit int

Maximum number of results

10

Returns:

Type Description
list[MemoryItem]

List of matching memory items

Source code in src/memoir/store/prolly_adapter.py
async def retrieve_memories_async(
    self, namespace: str, query: str, limit: int = 10
) -> list[MemoryItem]:
    """
    Retrieve memories using semantic search (async version).

    Args:
        namespace: User/agent namespace
        query: Search query
        limit: Maximum number of results

    Returns:
        List of matching memory items
    """
    # Use the hierarchical search engine to find relevant memories
    search_results = await self.search_engine.search(query, namespace)

    # Convert search results to memory items with deduplication
    memories = []
    seen_content = set()

    for result in search_results:
        # The search result contains combined content from multiple items
        if result.combined_content:
            try:
                # Split combined content back into individual memories
                individual_contents = result.combined_content.split(" | ")
                for content_text in individual_contents:
                    if content_text.strip():
                        # Create a memory item from the content
                        memory = MemoryItem(
                            key=result.path,
                            namespace=result.namespace,
                            content=content_text.strip(),
                            confidence=1.0,  # Default confidence
                            timestamp=time.time(),
                        )
                        # Deduplicate by content
                        content_hash = hash(memory.content)
                        if content_hash not in seen_content:
                            seen_content.add(content_hash)
                            memories.append(memory)
                            # Stop when we have enough unique results
                            if len(memories) >= limit:
                                break
                if len(memories) >= limit:
                    break
            except Exception as e:
                logger.warning(f"Failed to parse memory item: {e}")

    return memories

retrieve_memories

retrieve_memories(namespace: str, query: str, limit: int = 10) -> list[MemoryItem]

Retrieve memories using semantic search (sync fallback).

Note: This is a simple fallback. For proper semantic search, use retrieve_memories_async() which leverages the HierarchicalSearchEngine.

Parameters:

Name Type Description Default
namespace str

User/agent namespace

required
query str

Search query

required
limit int

Maximum number of results

10

Returns:

Type Description
list[MemoryItem]

List of matching memory items

Source code in src/memoir/store/prolly_adapter.py
def retrieve_memories(
    self, namespace: str, query: str, limit: int = 10
) -> list[MemoryItem]:
    """
    Retrieve memories using semantic search (sync fallback).

    Note: This is a simple fallback. For proper semantic search,
    use retrieve_memories_async() which leverages the HierarchicalSearchEngine.

    Args:
        namespace: User/agent namespace
        query: Search query
        limit: Maximum number of results

    Returns:
        List of matching memory items
    """
    logger.warning(
        "Using fallback sync search. For better results, use retrieve_memories_async()"
    )

    # Simple fallback - just return all memories
    all_memories = []
    search_results = self.search((namespace,), limit=limit)

    for _, _key, data in search_results:
        if isinstance(data, dict):
            try:
                memory = MemoryItem(**data)
                all_memories.append(memory)
            except Exception as e:
                logger.warning(f"Failed to parse memory item: {e}")

    return all_memories

get_statistics

get_statistics() -> dict[str, Any]

Get store statistics.

Source code in src/memoir/store/prolly_adapter.py
def get_statistics(self) -> dict[str, Any]:
    """Get store statistics."""
    stats = {
        "performance": self._stats.copy(),
        "total_keys": len(self._keys),
        "total_namespaces": len({key.split(":")[0] for key in self._keys}),
    }

    if self.enable_versioning and hasattr(self.tree, "get_head"):
        try:
            stats["versioning"] = {
                "current_commit": self.tree.get_head(),
            }
            if hasattr(self.tree, "log"):
                commits = self.tree.log()
                stats["versioning"]["total_commits"] = len(commits)
        except Exception:
            pass

    return stats

export_namespace

export_namespace(namespace: str, output_path: str) -> None

Export all memories from a namespace to JSON.

Parameters:

Name Type Description Default
namespace str

Namespace to export

required
output_path str

Path to save JSON file

required
Source code in src/memoir/store/prolly_adapter.py
def export_namespace(self, namespace: str, output_path: str) -> None:
    """
    Export all memories from a namespace to JSON.

    Args:
        namespace: Namespace to export
        output_path: Path to save JSON file
    """
    memories = {}
    search_results = self.search((namespace,), limit=1000)

    for _, key, data in search_results:
        memories[key] = data

    with open(output_path, "w") as f:
        json.dump(
            {
                "namespace": namespace,
                "timestamp": time.time(),
                "memories": memories,
            },
            f,
            indent=2,
        )

    logger.info(f"Exported {len(memories)} memories to {output_path}")