If you have used a smartphone for a decade, there is a good chance you are sitting on tens of thousands of saved fragments you have never reread. Recipes screenshotted from a friend's WhatsApp. Career advice clipped from a thread. Quotes from a book that hit at the right time. Photos of your kids that you saved but cannot find again. Voice notes from a parent. A fragment of a poem that meant something on a particular morning.
You did not save these for the future. You saved them in the moment, because they were resonant in the moment, and then you put the phone in your pocket and the moment passed. The fragment is still there. So are forty thousand others. None of them are findable. The archive is a kind of slow-motion erasure: every fragment was meaningful when saved, and every fragment is now invisible because it sits in the same scrollable river as everything else.
This is not a personal problem. It is the default state of every smartphone-using adult on the planet. We have spent fifteen years collecting personal media, and most of us do not have a working tool for reading it back.
Why it has been unsolvable until recently
The reason this archive sits dormant is not laziness. It is that the volume crossed a threshold below which human curation would have worked and above which only machines can. Forty thousand screenshots is not "a folder you should organise this weekend." It is more text than an average novel and more visual material than a photographer's career portfolio, all of it unsorted, all of it untagged, all of it captured under wildly varying conditions — some of it text, some images, some bilingual, some near-duplicates of each other, some with character-encoding artefacts inherited from old phone software.
Ten years ago, the technology to make sense of this corpus did not exist. Neither image-recognition nor text-extraction nor semantic-similarity were good enough at consumer scale. Five years ago, the technology started existing but only as cloud APIs — which meant that "make my dormant archive useful" required uploading every photo, every screenshot, every voice note, to a third-party vendor. Most people, presented with that proposition, correctly declined. The archive stayed dormant.
Today, for the first time, the technology to make this archive useful exists, and it can run on a single machine in a household. That second clause is the entire point of this post.
The technical capability and the privacy capability arrived in the same window. Recognition models are good enough; embedding models can search across modalities; small open-weight models can be hosted locally. None of this required a cloud vendor or required exposing a decade of personal life to a stranger's logs.