Service
Training Data Forensics
Analysis to determine if specific data was used for AI model training.1
Analysis of Training Data
For copyright and privacy cases involving AI, establishing whether a model ingested specific data requires a defensible statistical workflow anchored in peer-reviewed techniques.2910
Training Data Verification
We deploy membership inference, loss landscape analysis, and gradient leakage tests to quantify whether a dataset influenced model parameters.3
How We Investigate Suspect Training Data
The workflow implements dataset inference as a formal hypothesis test: we engineer matched cohorts, extract high-dimensional probe signals from the model, learn a distribution-specific decision surface, and quantify attribution strength with controlled error bounds.5
1. Controlled Cohort Design
Input corpora are shingled, deduplicated, and stratified by metadata (timestamp, domain, format) before being partitioned into suspect and validation cohorts A/B. We verify distributional alignment via Jensen–Shannon distance and n-gram frequency profiles, ensuring the inference task is not confounded by temporal or topical drift.8
2. Multi-Signal Probing
We execute a probe suite spanning token-level log-likelihood deltas, Min-k% probability aggregates, perturbation Jacobians, zlib compression ratios, and reference-model KL divergences. Signals are sampled across checkpoints and temperature settings to factor out instruction-tuning artifacts and decoder variance.
3. Signal Calibration
Probe outputs are standardized, then an elastic-net logistic model is trained on the A-split holdout to learn probe weights specific to the distribution. We perform nested cross-validation and permutation importance to isolate signals with non-spurious predictive contribution.
4. Statistical Attribution
Calibrated scores for the B-split cohorts feed a Welch t-test with Benjamini–Hochberg correction across datasets. We return p-values, standardized effect sizes, and bootstrap confidence intervals to quantify attribution strength under explicit Type I error guarantees.
The Deliverable: A Technical Report
Our analysis culminates in a technical report detailing methodology, statistical findings, and data provenance narratives. The report is structured for courtroom use, with reproducible metrics, confidence statements, and expert interpretations.
Source Notes
- S-Square Research, "Litigation Impact Metrics 2023–2025." ↩
- Maini et al., "LLM Dataset Inference: Did you train on my dataset?" ICLR (2024). Source ↩
- Shokri et al., "Membership Inference Attacks Against Machine Learning Models," IEEE S&P (2017). Source ↩
- Carlini et al., "Extracting Training Data from Large Language Models," IEEE S&P (2021). Source ↩
- S-Square Research, "Discovery Effort Benchmarks," and methodology cross-check with Maini et al. (2024). ↩
- National Institute of Standards and Technology, "AI Risk Management Framework." ↩
- Google Search Central, "Introduction to robots.txt." ↩
- Andrei Z. Broder, "On the resemblance and containment of documents," ACM Digital Library. ↩
- Gao et al., "The Pile: An 800GB Dataset of Diverse Text for Language Modeling," arXiv (2021). ↩
- Schuhmann et al., "LAION-5B: An open large-scale dataset for training next generation image-text models," arXiv (2022). ↩
- National Institute of Standards and Technology, "Generative AI Risk Management Profile." ↩