Skip to content

fix: prevent convo_miner from re-processing 0-chunk files on every run (#654)#732

Merged
bensig merged 2 commits intoMemPalace:developfrom
mvalentsev:fix/convo-miner-reprocess-sentinel
Apr 12, 2026
Merged

fix: prevent convo_miner from re-processing 0-chunk files on every run (#654)#732
bensig merged 2 commits intoMemPalace:developfrom
mvalentsev:fix/convo-miner-reprocess-sentinel

Conversation

@mvalentsev
Copy link
Copy Markdown
Contributor

Closes #654 (Bug 1 only).

mine_convos() has three early-exit paths (OSError during normalize, content below MIN_CHUNK_SIZE, zero chunks from chunk_exchanges) that continue without writing anything to ChromaDB. Since file_already_mined() checks for a document with a matching source_file metadata value, these files return False on every subsequent run and get re-read, re-normalized, and re-chunked -- forever.

With 50 such files in a directory, that is 50 wasted reads on every mine invocation.

Fix (2 files, +79/-0):

mempalace/convo_miner.py:

  • Add _register_file() helper that upserts a lightweight sentinel document (room="_registry", ingest_mode="registry") so file_already_mined() returns True on future runs
  • Call it at all three early-exit points, guarded by if not dry_run
  • Uses upsert() (not add()) so repeated runs are idempotent

tests/test_convo_miner.py:

  • test_mine_convos_does_not_reprocess_short_files -- verifies a too-short file gets a sentinel and is skipped on second run
  • test_mine_convos_does_not_reprocess_empty_chunk_files -- verifies a file with no exchange markers gets a sentinel and is skipped

Scope note: Bug 2 from the issue (drawers_added counter always 0) was already resolved upstream via the switch from collection.add() to collection.upsert(). This PR only addresses Bug 1, as @DevOPsJourneyman suggested in the issue thread -- a small focused follow-up separate from the batching logic in #629.

…emPalace#654)

mine_convos() has three early-exit paths (OSError, content too short,
zero chunks) that skip writing anything to ChromaDB. Since
file_already_mined() checks for the presence of a document with a
matching source_file, these files are re-read and re-processed on
every subsequent run.

Add _register_file() that upserts a lightweight sentinel document
(room="_registry", ingest_mode="registry") so file_already_mined()
returns True on future runs.

Note: Bug 2 from the issue (drawers_added counter always 0) was
already resolved upstream via the switch from collection.add() to
collection.upsert().
@mvalentsev mvalentsev marked this pull request as ready for review April 12, 2026 21:10
Copy link
Copy Markdown
Collaborator

@bensig bensig left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code review + security audit clean.

@bensig bensig merged commit 87e8baf into MemPalace:develop Apr 12, 2026
6 checks passed
jphein added a commit to jphein/mempalace that referenced this pull request Apr 12, 2026
Upstream merged MemPalace#682-684 (our splits), MemPalace#687 (dry-run None room),
MemPalace#695/MemPalace#708 (convo_miner full response), MemPalace#732 (0-chunk re-processing),
plus VitePress docs site. Conflicts:
- config.py: take upstream's [^\W_] regex (our MemPalace#683 merged version)
- miner.py: integrate upstream's early-return for tiny files, dedupe
  dry-run read path
- test_miner.py: keep our detect_room tests + upstream's dry-run test
- CONTRIBUTING.md: take upstream's org URL update

Co-Authored-By: Claude Opus 4.6 <[email protected]>
gnusam pushed a commit to gnusam/mempalace-pgsql that referenced this pull request Apr 25, 2026
… 0-chunk files

Three upstream fixes ported together because they're conceptually one
"convo_miner polish" pass on the same exchange-chunking path.

1. Remove ai_lines[:8] truncation (upstream d52d6c9, PR MemPalace#695). The
   _chunk_by_exchange path was silently dropping every line past line 8
   of the AI response, violating the verbatim-storage principle.

2. Split oversize exchanges across drawers (upstream 9b60c6e, PR MemPalace#708).
   Now that the full response is preserved, an exchange that exceeds
   CHUNK_SIZE (800 chars, aligned with miner.py) is split into
   consecutive drawers instead of a single oversized one. Adds
   CHUNK_SIZE module constant.

3. Register a no-embedding sentinel for files that produce zero chunks
   (upstream 87e8baf, PR MemPalace#732). mine_convos has three early-exit paths
   (OSError, content too short, zero chunks) that previously wrote
   nothing — file_already_mined() then returned False on the next run
   and the file was re-read every time.

Adapted MemPalace#3 for the PG backend: the upstream sentinel uses
collection.upsert() (ChromaDB API). This fork instead adds a
PalaceDB.register_empty_file() method that inserts a row directly with
embedding=NULL and metadata.ingest_mode='registry', so the sentinel is
free of embedding cost and invisible to vector search. file_already_mined()
already keys on source_file + source_mtime, so the existing path picks
up the sentinel without further changes.

Three behavioural tests added: full AI response preserved, oversize
exchange split across drawers, and the sentinel + file_already_mined
round trip.

Upstream:
  MemPalace@d52d6c9
  MemPalace@9b60c6e
  MemPalace@87e8baf

Co-authored-by: shafdev <[email protected]>
Co-authored-by: Sanjay Ramadugu <[email protected]>
Co-authored-by: Mikhail Valentsev <[email protected]>
Co-Authored-By: Claude Opus 4.7 (1M context) <[email protected]>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

bug: convo_miner re-processes files every run + drawers_added counter always 0

2 participants