Validiti Validiti
Validiti RAM

Hold 2.7× more data in the same memory.
Query it without ever decoding it.

Validiti RAM is an in-process memory layer that holds working sets in compressed form and runs token-presence queries against the compressed bytes directly. Same workload, fewer machines. Same RAM, more cache. Powered by the Pacta42 substrate.

2.7×
More cache
1,942×
Faster queries
356µs
GET latency
0
Disk fallback
↓ How it works

What Validiti RAM is

A way to use the Pacta42 substrate as an in-memory layer. Same physical RAM, more working data, faster queries.

01

The same working set fits in a fraction of physical RAM — because each entry is held as Pacta42 substrate bytes (compressed) instead of raw bytes.

02

Decoded data is still available on demand — any entry can be reconstructed byte-exact via a single in-process call.

03

Useful operations work without full decode — token-presence and phrase queries run against per-document position indexes built once at insert time. The compressed form is queryable, not just decodable.

Measured numbers

71 MB working set (18,180 docs at ~4 KB each) on a developer workstation. Single-threaded run. All numbers reproducible from the included demo script.

Metric Naive (raw bytes in dict) Validiti RAM Ratio
Total bytes held 71.01 MB 26.12 MB 2.72× smaller
GET latency (5,000 random ops) 0.23 µs/op 356 µs/op (slower than dict, faster than disk/network)
Token-presence query (single token) 5,216 ms · 5,826 hits 2.7 ms · 5,826 hits 1,942× faster · identical hits
Compound query (token A AND token B) 5,263 ms · 1,973 hits 2.9 ms · 1,973 hits 1,838× faster · identical hits

Both query methods return the identical hit set — verified by set equality. Pacta42 query phase is a frozenset membership test against pre-built per-document indexes. Every additional query against the same cache is essentially free.

What runs on Validiti RAM

Anywhere your working set is too big for RAM and too valuable to page to disk. Anywhere you query the same compressed data over and over.

Search & analytics indexes

Keep more of the index in memory

Search engines, analytics rollups, full-text indexes — the index is structurally repetitive and compresses well. Hold more index per node; fewer disk hits; faster query response.

2.7× more index in same RAM
In-memory caches

Drop-in capacity boost for Redis-shaped workloads

Same eviction policies, same key/value access pattern, same in-process latency profile when keys are frequently re-queried. 2-3× more capacity per node before you need to shard.

2-3× more entries per node
LLM serving

KV cache compression for higher concurrency

The KV cache is a memory hog at LLM serve time. Validiti RAM-compressed KV cache reduces per-request memory footprint and allows higher concurrency on the same GPU/CPU host. Pair with Validiti Accelerate for full-stack savings.

More concurrent requests per machine
Edge & embedded

Server-class working sets on constrained hardware

Devices with constrained RAM (8–32 GB) can hold working sets that would normally need a server. Move expensive lookups out of the cloud and onto the edge node.

Fits 30+ GB into 16 GB hardware
Database working sets

Buffer pools and query result caches with more headroom

Particularly useful for systems where the cached data is text-heavy (legal, regulatory, structured records). Same buffer pool eviction, same query result cache — just bigger before you start spilling to disk.

Lower disk-spill rate per query
Compliance & privacy scanning

Scan a corpus without decoding the text

"Find every doc that mentions term X" runs in milliseconds against millions of docs. The unencrypted text never appears in the process address space during the scan — that's a security property, not just a performance one.

Privacy-preserving by construction

Capacity projection

At the asymptotic 2.72× ratio (regulatory-text corpus), here's how much working data the same machine holds. 80%-of-RAM utilization for both columns to keep the comparison fair.

Machine RAM Naive holds Validiti RAM holds Extra
1 GB0.8 GB2.2 GB+1.4 GB
16 GB12.8 GB34.8 GB+22.0 GB
64 GB51.2 GB139.2 GB+88.0 GB
256 GB204.8 GB556.8 GB+352.0 GB
1 TB819.2 GB2.2 TB+1.4 TB

A 64 GB workstation holds 51 GB of working data naively; the same machine through Validiti RAM holds 139 GB — 2.7× more data accessible at memory speeds, with no disk fallback. Compression ratio is corpus-dependent: regulatory text 2.72×, templated logs 8–15×, diverse natural prose ~2×.

Where this replaces a Redis cluster

"I have more data than RAM" is solved today by sharding, paging to SSD, or standing up a Redis cluster. Each adds latency, complexity, and operational cost. Validiti RAM is a single-machine, in-process answer.

Honest fit-test — when Validiti RAM beats the alternative

  • Working set is text-heavy and structurally repetitive. Regulatory documents, logs, records, JSON payloads, schema-shaped data — substrate-friendly. Diverse natural prose still helps but less.
  • Working set is bigger than 6 MB. Below ~6 MB raw, the substrate library overhead dominates and naive is faster. Substrate-friendly workloads cross the break-even quickly.
  • Reads are not pure hot-path dict lookups. If every operation is a 0.2µs in-process dict access, Validiti RAM (356µs/op) is the wrong tool. If your alternative is hitting disk (~50µs/op SSD seek) or network (~500µs Redis round-trip), Validiti RAM is faster end-to-end.
  • You query the cache repeatedly. Per-document position indexes are built once at insert; every subsequent query against the same cache is a microsecond-class set-membership test. The query advantage compounds across re-queries.

Pricing

Foundation tier free for non-commercial. Per-machine licensing above.

Foundation

Free, forever, for personal / education / research / journalism

Single-machine deployment. Public Pacta42 foundation libraries (Wikipedia, Gutenberg, StackExchange). Generous non-commercial usage threshold. Upgrade to Commercial when your traffic crosses it.

$0 forever for non-commercial use
Commercial · per machine

$199 / machine / year

Single-server commercial license. Production foundation libraries. Custom curated libraries available. Email support.

Production deployment, single host
Fleet · volume

$99 / machine / year (10+ machines)

Volume pricing for fleets. Fleet management. Dedicated support. Custom integration. Co-publishing of vertical libraries.

Multi-host production fleets
Validiti Inherent

Bundled with any Validiti SKU purchase

Already running Validiti Accelerate, DMS, Maths, Titus, or Provenance? Validiti RAM is included in the substrate that ships with every Validiti product. Pacta42 is native; the in-RAM deployment shape is just configuration.

Included · no separate license

Honest limitations

A substrate-based memory layer isn't a universal answer. The places it wins are real; the places it doesn't are real too.

Where Validiti RAM is the wrong tool

  • Tiny working sets. Below ~6 MB of raw data, library overhead exceeds savings. This is a server / cache product, not a library for tiny apps.
  • Pure hot-path dict workloads. If every read is a 0.2µs in-process dict access on a small key set, Validiti RAM's 356µs/op decode is dominant. Use a normal dict.
  • Diverse natural prose. Compression ratio drops to ~2× on free-form text. Still wins on capacity, but the headline number applies to structured/regulatory/logs/records, not unconstrained prose.
  • Frequency queries. The current per-document index records which positions are present, not how often each appears. Frequency-of-term queries fall through to wire-body decode (still much faster than scanning raw text). Multiset index is a planned capability.
  • Index build is currently slow in Python. 59 sec to build the per-doc index for 18K docs. Rust port lands next; will be sub-second. Once built, queries are sub-millisecond.

Reserve a launch slot

We are not shipping a download yet. The reference demo is built and reproducible; the productized SKU is in finalization. When the .deb opens, you'll install in five minutes and your working set will gain capacity overnight.

Get on the launch list

Tell us your workload (cache shape, working-set size, current Redis/SSD spend). We'll size your savings before launch and reserve a slot.

Reserve your launch slot

Built-in guarantees

Every tier — including Foundation — ships the same protection envelope.

What every install gets, on day one

  • Sealed binary. Every numerical kernel ships as a stripped, signed native module. The package self-verifies its identity at first import; any mismatch refuses to load.
  • Machine-bound activation. One license, one machine. The license is bound at activation; moving to a new host requires deactivating the old one.
  • Runtime tamper detection. Validiti Titus watches the substrate library and the in-RAM cache continuously. Any tamper trips a defensive shutdown in milliseconds.
  • Same engine on every tier. Foundation, Commercial, Fleet, Inherent — identical code, identical protection. We don't downgrade defenses to save cost.
  • Bytewise-identical decode. Every entry reconstructs byte-for-byte against the original. Verified by set-equality on the demo's 18,180-document corpus.
  • U.S.-headquartered, U.S. only at launch. Single legal jurisdiction. Your hardware, your data, your call.