Skip to content

Verified Results

Case Studies

Four real tests that put SPECTR and standard tools side by side. Each was chosen because a method routinely used in data analysis fails to solve it. All numbers are exact results from re-runnable tests. Every claim maps to a test ID you can verify independently.

Transparency: Each case study shows exactly what was tested, what it was compared against, and what the result does and does not prove. Where claims have been retracted, they are documented honestly; none appear in the case studies above.
C-DISCOV-047

A hidden pattern that gzip is blind to

Both datasets look identical from the outside. Only one has structure.

SPECTR AUC: 0.9847. Baseline gzip AUC: 0.5000.

0.9847
SPECTR AUC95% CI: [0.954, 1.000]

Setup

We created two types of data: one with a hidden repeating pattern woven through it, and one that was randomly shuffled. Both had identical byte-level statistics — byte frequency, byte randomness score, everything. We then gave both to SPECTR and to gzip and asked: which dataset has the pattern? 40 pairs tested.

Why SPECTR detects this

SPECTR reads data at the level of individual bits, not bytes. It spotted the repeating pattern on first encounter and recognised it every time it reappeared, even though the pattern never aligned with byte edges. The structured dataset slowed down in discovery — it was recognising patterns rather than finding new ones — and that signal is what drives the result.

Why standard tools miss it

gzip works by finding repeated sequences of bytes. In this test, the pattern spans across byte edges and never appears as a byte-level repetition. gzip simply can't see it and scores no better than a coin flip.

MetricSPECTRBaseline
SPECTR accuracy (AUC)0.9847-
gzip accuracy (AUC)-0.5000
Confidence interval (SPECTR)[0.954, 1.000]-
Confidence interval (gzip)-[0.500, 0.500]
Byte randomness difference0.0000000.000000
Test pairs4040
Data length per sequence32,000 bits32,000 bits
Honest scopeThis result was tested for one specific type of hidden pattern (a 17-bit prime repeating unit, chosen because it cannot align to byte boundaries). Other types of sub-byte patterns have not been tested yet.
VerificationC-DISCOV-047

40 pairs generated from a fixed seed. The same configuration produces identical numbers on any machine. Runnable with the Python wheel.

C-DISCOV-045 / C-DISCOV-017

Same letters, different order — and standard tools can't tell the difference

Two DNA sequence types share every building block. Only the arrangement differs.

SPECTR AUC (Alu vs LINE-1): 0.9645. Baseline Standard method AUC†: ~0.5000.

0.9645
SPECTR AUC (Alu vs LINE-1)

Setup

We took 900 real DNA sequences from the human reference genome, covering five different types of genetic element. The headline test: can SPECTR distinguish Alu repeats from LINE-1 repeats? Both are common DNA repeat elements, and they share the exact same vocabulary of structural building blocks — none are unique to either type. Only the internal arrangement differs.

Why SPECTR detects this

SPECTR tracks not just which structural patterns appear in a sequence, but the order in which they appear and how they accumulate from start to finish. Alu and LINE-1 elements have different internal arrangements, so their structural profiles look different even when the vocabulary is identical.

Why standard tools miss it

We measured the vocabulary overlap between Alu and LINE-1: it was perfect. Every structural building block found in Alu was also found in LINE-1, and vice versa — 3,529 shared patterns, zero unique to either type. Any method that works from what appears (rather than how it's arranged) produces identical scores for both, making them indistinguishable. † This chance result is derived logically from complete vocabulary overlap, not measured directly.

MetricSPECTRBaseline
SPECTR accuracy, Alu vs LINE-1 (AUC)0.9645-
Standard method accuracy (AUC)†-~0.5000
Shared building blocks, Alu and LINE-13,529 of 3,529 (100%)3,529 of 3,529 (100%)
5-type classification — SPECTR0.6122-
5-type classification — k-mer baseline-0.7067
Total sequences tested900900
Honest scopeThis result covers the specific Alu vs LINE-1 comparison. For a broader task classifying all five element types at once, k-mer frequency methods outperform SPECTR — that result is documented honestly in the numbers below. Arrangement signal is most valuable when shared vocabulary would otherwise make two classes indistinguishable.
VerificationC-DISCOV-045 / C-DISCOV-017

900 real human genome sequences from GRCh38 reference assembly. Fixed extraction; results are deterministic across machines. Runnable with the Python wheel.

C-DISCOV-025

The standard measurement says they're identical. They're not.

When randomness scores agree but structure disagrees.

SPECTR AUC: 0.9861. Baseline Shannon entropy AUC: 0.5000.

0.9861
SPECTR AUC95% CI: [0.9832, 0.9888]

Setup

We generated 1,000 pairs of data sequences. In each pair, one sequence had an internal repeating pattern; the other was randomly shuffled. Both sequences in every pair were engineered to have the identical randomness score. We then asked SPECTR and Shannon entropy to identify which was which.

Why SPECTR detects this

SPECTR builds a picture of structure as it processes data from beginning to end. A sequence with a repeating pattern 'fills up' differently from a shuffled sequence — the structured one stops discovering new patterns earlier because it keeps recognising ones it has already seen. That shape difference is what SPECTR uses to tell them apart, even when the randomness scores are identical.

Why standard tools miss it

Shannon entropy is a single-number summary of how random a dataset appears. It counts how often each symbol appears, then collapses that into one score. It pays no attention to order. So two datasets with the same randomness score look completely identical to entropy, even if one has a clear internal structure and the other is pure noise.

MetricSPECTRBaseline
SPECTR accuracy (AUC)0.9861-
Shannon entropy accuracy (AUC)-0.5000
Confidence interval (SPECTR)[0.9832, 0.9888]-
Test pairs1,0001,000
Randomness score difference0.00 (identical by design)0.00 (identical by design)
Honest scopeThis specific test — structured vs shuffled sequences with matched randomness scores — is also solvable by gzip. It is Shannon entropy that fails here, not all compression methods. SPECTR's advantage over gzip on this type of task is that it produces an explicit, enumerable vocabulary of structural building blocks — gzip only produces a compressed output.
VerificationC-DISCOV-025

1,000 pairs generated from a fixed seed; confidence intervals from 10,000 resamples. Identical inputs produce the same result on any machine. Runnable with the Python wheel.

C-DISCOV-019

Spot the anomaly. Then reconstruct it exactly — from the same record.

Detection and byte-exact reconstruction from a single structural record.

1.0
Detection Rate
Shuffled
1.0
Detection Rate
Reversed
1.0
Detection Rate
Pair-swap
1.0
Reconstruction Accuracy
exact bytes recovered from the same record

Detection rate = fraction of anomalous windows correctly flagged (30 trials per method). Reconstruction accuracy = fraction of flagged windows recovered byte-exactly from the structural record alone.

1.0
SPECTR Detection rate (all methods)

Setup

We set up a stream of synthetic data and injected hidden anomalies — windows of data that had been shuffled, reversed, or rearranged. The anomalies were engineered to look statistically normal: they had the same randomness score as the surrounding data. We then asked SPECTR to find them, and to reconstruct them exactly once found. 30 trials for each of three injection methods.

Why SPECTR detects this

When SPECTR processes data, it stores the structural identity of every window it sees. When it flags a window as anomalous — because the rate of new patterns spikes unexpectedly — that window is already in the ledger. You can retrieve the exact original bytes from the same record that raised the flag. One structure. Two functions: detection and reconstruction.

Why standard tools miss it

Standard anomaly detectors based on randomness scores can't find these anomalies — the injected windows score the same as normal data. Even if they could detect them, detection and reconstruction are normally separate systems: one to flag, another to store the original bytes. SPECTR does both from a single structural record.

MetricSPECTRBaseline
Detection rate, shuffle anomalies (30 trials)1.0-
Detection rate, reversed anomalies (30 trials)1.0-
Detection rate, swap anomalies (30 trials)1.0-
Reconstruction accuracy (byte-exact)1.0N/A
Randomness score difference (anomaly vs background)< 0.001< 0.001
Needs separate storage systemNoYes
Honest scopeThis test uses synthetic data with carefully engineered anomalies. Real-world streaming data such as network traffic or sensor readings has not been validated. The capability has been demonstrated in a controlled setting, not production.
VerificationC-DISCOV-019

90 trials (30 per strategy) from a fixed seed. Anomaly injection is fully deterministic. Runnable with the Python wheel.

Evidence Scope

What has been tested and where the limits are

Every row below maps to a deterministic, re-runnable test. “Boundary” rows are not failures; they define the conditions under which SPECTR has no advantage over conventional methods. Honest limits matter as much as capabilities.

Boundary conditionConfigurationSPECTRBaselineMethodStatus
Pattern hidden between byte edges
C-DISCOV-047
17-bit repeating unit (prime length so it can't align to bytes), byte statistics identical across all pairs0.9847 AUC0.5000 AUCgzipPass
Identical randomness scores, different structure
C-DISCOV-025
1,000 pairs — one structured, one shuffled — randomness score identical in each pair by design0.9861 AUC0.5000 AUCShannon entropyPass
Same vocabulary, different arrangement (DNA repeats)
C-DISCOV-045
Real human genome sequences — Alu vs LINE-1 share every structural building block, zero unique to either0.9645 AUC0.5000 AUCVocabulary-based (k-mer)Pass
Anomaly detection + reconstruction from one record
C-DISCOV-019
3 anomaly injection methods, 30 trials each, anomalies match background randomness score1.000 detection + byte-exact reconstruction0 / 30 detectedEntropy-based detectorPass
Pattern that aligns to byte edges (gzip sees it too)
C-DISCOV-046 (retracted)
Shorter repeating units that happen to land exactly on byte boundaries~0.87 AUC~0.87 AUCgzipBoundary

When structure naturally aligns to byte boundaries, gzip detects it just as well. SPECTR's advantage only appears when patterns cross byte edges.

Binary files with large, uniform vocabularies
IDR boundary
Binary file classification — SPECTR vs simple byte-frequency counting0.650 accuracy0.769 accuracyByte-frequency countingBoundary

Arrangement signal matters most where vocabulary is small and repetition is meaningful. In binary files with hundreds of unique byte patterns, simple frequency counting already captures nearly all the signal.

Genuinely random data — no structure to find
C-DISCOV-044
Data at the theoretical randomness boundary with effectively uniform pattern distribution~0.5 discovery rate~0.5 discovery rateEntropyBoundary

Where there is no structure, there is nothing to detect. Both SPECTR and standard methods return chance-level performance. This is the expected and correct result.

All numbers are exact values from deterministic test runs. Retracted claims are documented with root causes; none appear in the case studies above.