Methodology Comparison Matrix
Systematic comparative analysis of six professional investigation methodologies informing forensic intelligence platform design.
Methodology Comparison Matrix for Forensic Intelligence
Document Version: 1.0 Date: January 2026 Purpose: Comprehensive comparative analysis of six professional investigation methodologies Relevance: Informs design decisions for Phronesis FCIP v6.0 and forensic intelligence platforms
1. Overview and Purpose
This matrix provides a systematic comparison of six established investigation methodologies:
- Police Investigations (UK College of Policing, HOLMES2, IOPC)
- Investigative Journalism (ICIJ, OCCRP, ProPublica)
- Legal eDiscovery (EDRM, TAR 2.0, Cochrane-influenced systematic approaches)
- Regulatory Investigations (GMC, HCPC, NMC, BPS professional standards)
- Intelligence Analysis (CIA, UK JIC, NATO, 66 SATs, ACH)
- Academic Research (PRISMA, Cochrane, Grounded Theory, Thematic Analysis)
How to Use This Matrix
For System Design:
- Identify requirements: Which methodology's standards apply to your use case?
- Select tools: Which methodology's tooling best matches your investigation scale?
- Design workflows: Which methodology's process structure fits your institutional context?
For Investigation Planning:
- Match methodology to investigation type using Decision Tree (Section 9)
- Understand evidence requirements, quality control standards, timeline expectations
- Anticipate resource needs (team size, tool costs, time to results)
For Quality Assurance:
- Benchmark your processes against established standards
- Identify gaps in current methodology
- Justify methodological choices to stakeholders
2. Evidence Standards Comparison
2.1 Documentary vs. Testimonial Evidence Priority
| Methodology | Documentary Evidence | Testimonial Evidence | Hierarchy |
|---|---|---|---|
| Police | π‘ Moderate priority - Contemporaneous records valued, but witness testimony central | π’ High priority - PEACE interviews, Cognitive Interview (46% recall increase) | Equal weight - Both critical, cross-verification required |
| Journalism | π’ Highest priority - Official docs systematically prioritized over testimonial | π‘ Moderate priority - Multiple independent testimonies required for corroboration | Documentary > Testimonial - Explicit hierarchy |
| Legal eDiscovery | π’ Highest priority - Hash-certified, metadata-preserved, chain of custody | π‘ Moderate - Depositions important but documentary evidence is core litigation evidence | Documentary > Testimonial - Legal admissibility priority |
| Regulatory | π’ High priority - Contemporaneous clinical/professional records strongest | π‘ Moderate - Witness statements important but subject to bias assessment | Documentary > Testimonial - Contemporary records trump later accounts |
| Intelligence | π’ High priority - SIGINT, IMINT, GEOINT preferred (hard to fake) | π‘ Variable - HUMINT critical but highest deception risk (Admiralty Code) | Multi-INT fusion - Contradictions require reconciliation |
| Academic | π’ High priority - Peer-reviewed sources, official records, archival materials | π‘ Moderate - Interview data important but requires methodological rigor (IRR testing) | Depends on research design - Qualitative research may prioritize interviews |
Key Insight: Only Journalism and Legal eDiscovery explicitly prioritize documentary over testimonial evidence. Police and Regulatory balance both. Intelligence fuses multiple source types with explicit reliability ratings.
2.2 Authentication Requirements
| Methodology | Authentication Standard | Chain of Custody | Digital Integrity | Admissibility Test |
|---|---|---|---|---|
| Police | FBI 5-step protocol: Legal collection, detailed description, accurate ID, proper packaging, chain logs | Mandatory - Every transfer logged, signed, timestamped | SHA-256/MD5 hashing, write-blocking, NIST standards | Court admissible - FRE 901 authentication |
| Journalism | 3-step verification: Primary source authentication, multiple corroborations, expert validation | Documented - Provenance tracked, not legally required | Metadata preservation, forensic validation for leaks | Ethical standard - Not legally admissible, reputational risk |
| Legal eDiscovery | FRE 902(13)/(14) - Self-authenticating with certification; SHA-256 hash certification | Automated audit trails - Every access logged, immutable | Dual hash: Collection hash + Production hash; metadata preservation mandatory | Court admissible - Designed for litigation |
| Regulatory | Balance of probabilities - Contemporary records authenticated by custodian testimony | Less formal - Not criminal standard, but provenance tracked | Electronic health records: NHS audit trails, access logs | Tribunal standard - Civil burden, hearsay admissible |
| Intelligence | Admiralty Code - Source reliability (A-F) + Information credibility (1-6) rated independently | Classification controls - Access logs, compartmentalization | Technical validation (SIGINT decryption, IMINT geo-registration) | Not court-focused - Supports operational/policy decisions |
| Academic | Peer review + citation - Source verification, archival confirmation, replication data | Audit trail - Research data management plans, version control | Dataset hashing, repository DOIs (Zenodo, OSF), FAIR principles | Scholarly standard - Not legal admissibility |
Key Takeaway: Legal eDiscovery has most mature digital authentication standards (hash certification, metadata preservation). Police follow similar standards for court use. Intelligence focuses on source reliability over legal admissibility. Journalism and Academic rely on ethical/scholarly standards, not legal requirements.
2.3 Burden and Standard of Proof
| Methodology | Standard of Proof | Burden | Presumption | Appeal/Review |
|---|---|---|---|---|
| Police | Criminal: Beyond reasonable doubt (~95%+ certainty) | Prosecution bears burden throughout | Innocence presumed | CPS review, court trial, IOPC oversight |
| Journalism | Editorial: Confidence in veracity (defensible against libel) | Publication bears burden | No presumption | Editorial review, legal vetting, fact-checking |
| Legal eDiscovery | Civil: Preponderance of evidence (51%+) or Reasonable defensibility (for discovery disputes) | Producing party must show reasonable effort | No presumption of misconduct | Court sanctions for spoliation/bad faith |
| Regulatory | Civil: Balance of probabilities (51%+) | Regulator bears burden | No presumption of impairment | Panel appeals, judicial review |
| Intelligence | Analytic confidence levels (High/Moderate/Low) + Probability estimates (WEP) | Analyst bears burden to justify confidence | No presumption | Peer review, Red Cell challenge, senior review |
| Academic | Scholarly standards - Peer review consensus, replication | Researcher bears burden for claims | Null hypothesis (no effect presumed) | Peer review, post-publication critique, replication |
Critical Distinction: Criminal (Police) requires highest certainty. Civil (Legal, Regulatory) uses balance of probabilities. Journalism and Intelligence use confidence-based frameworks. Academic uses statistical significance + peer consensus.
3. Timeline Construction and Temporal Analysis
3.1 Methods and Tools
| Methodology | Timeline Method | Evidence Linking | Temporal Contradiction Detection | Tools |
|---|---|---|---|---|
| Police | Chronologies (5WH framework) - What, Who, When, Where, Why, How structure | Every event linked to evidence - Witness statements, exhibits, CCTV timestamps | Gap analysis - Suspicious absences, timeline inconsistencies flag further inquiry | HOLMES2 Dynamic Reasoning Engine, MIR logs |
| Journalism | Document-driven timelines - Events extracted from authenticated documents, not assumptions | Bates-style referencing - Every claim cites specific document (Panama Papers: 11.5M docs) | Cross-document contradiction flagging - Automated + manual review for date conflicts | Apache Tika, Solr, Neo4j, Aleph (4B+ docs), Datashare |
| Legal eDiscovery | 8-step timeline process - Early construction, central repository, objective language, continuous updates | Bates numbers mandatory - Every event cites supporting documents with page/line references | Version tracking + Near-duplicate detection - Identify document evolution, flag inconsistent language changes | Relativity, Everlaw Storybuilder, CaseFleet, TimeMap |
| Regulatory | Three-stage temporal analysis - Facts proven β Standards breached β Current impairment (focus on present, not past) | Evidence bundles per allegation - Each finding tied to specific records, dates, witnesses | Temporal focus on patterns - One-off vs. repeated incidents, remediation timeline assessment | Case management systems (GMC, HCPC databases), Excel-based investigation logs |
| Intelligence | Chronologies as diagnostic SAT - Establish factual sequence, identify assumptions, detect deception | Multi-INT fusion timelines - SIGINT intercepts + IMINT timestamps + HUMINT accounts reconciled | F3EAD cycle exploitation - Each operation generates timeline leads for next cycle (hours, not weeks) | Intelligence databases, timeline analysis software, GEOINT platforms |
| Academic | Event history analysis - Qualitative timelines, process tracing, historical analysis | Citation to sources - Archival references, interview dates, document provenance | Thematic coding of temporal themes - Sequences, turning points, trajectories analyzed qualitatively | QDAS tools (NVivo timeline view, Atlas.ti network chronologies), qualitative timeline software |
Standout Practice: Legal eDiscovery's 8-step timeline process with mandatory Bates linking is gold standard for audit trails. Intelligence's F3EAD cycle demonstrates fastest iterative timeline construction (hours vs. weeks).
3.2 Temporal Contradiction Handling
| Methodology | Detection Approach | Resolution Process | Escalation |
|---|---|---|---|
| Police | Manual review + HOLMES2 automated flagging - Analysts review chronologies, system flags timeline gaps | Investigative action - Collect additional evidence, re-interview witnesses, challenge suspects | SIO review - Senior Investigating Officer assesses significance, adjusts strategy |
| Journalism | Multi-layered verification - Editor review, fact-checker validation, legal vetting before publication | Seek additional sources - Do not publish contradiction unless explained or acknowledged | Editorial decision - Publish with caveat ("conflicting accounts exist") or withhold until resolved |
| Legal eDiscovery | Near-duplicate detection + version diff analysis - Software flags temporal inconsistencies automatically | Discovery dispute - Present contradiction to opposing party, request explanation or admit inconsistency | Court resolution - Judge rules on admissibility, weight of contradictory evidence |
| Regulatory | Case examiner review - Dual decision-makers (professional + lay) assess contradiction's impact on findings | Registrant response - Provide opportunity to explain discrepancies before final hearing | Tribunal decision - Panel weighs credibility, may find some facts proved, others not |
| Intelligence | ACH matrix evaluation - Contradiction = diagnostic evidence (eliminates hypotheses) | Multi-INT reconciliation - Check if temporal difference due to source perspective, timing, or deception | Red Cell review - Alternative explanations for contradiction examined by contrarian team |
| Academic | Triangulation - Compare accounts across sources (documents, interviews, observations) | Reflexive analysis - Researcher acknowledges contradiction, explores meaning (not error to eliminate) | Peer review - Reviewers assess researcher's handling of contradictory data |
Key Difference: Intelligence treats contradictions as diagnostic (actively useful for hypothesis elimination). Police and Legal treat contradictions as investigative leads (triggers for additional evidence gathering). Academic treats contradictions as analytical richness (reveals complexity, not necessarily error).
4. Quality Control and Peer Review
4.1 Review Structures
| Methodology | Review Stages | Minimum Reviewers | Independence Requirement | Statistical Validation |
|---|---|---|---|---|
| Police | 3-tier: First-line supervision (daily) β Second-line (strategic) β Senior management (major incidents) | Minimum 2 - Action supervisor + case officer; Major incidents: 3+ (SIO, Deputy SIO, external review) | PIP-level independence - Reviewers must be higher PIP level than investigator | Not statistical - Professional judgment, case law precedent |
| Journalism | 4-stage: Reporter β Desk editor β Senior editor β Legal review (pre-publication) | Minimum 3 - Reporter + 2 editors; Investigative projects: 5+ (+ fact-checker, lawyer) | Editorial independence - Editors not involved in story development review final product | 3-source rule - Significant claims require 3 independent corroborations |
| Legal eDiscovery | Multi-phase QC: First-pass review β QC sample (5-10%) β Partner/Senior Counsel review β Court scrutiny | Minimum 2 - Reviewer + QC sampler; TAR: Statistical validation required | Independent QC - Different attorney from reviewer samples for quality | Yes - TAR validation: Precision/recall testing, elusion testing, confidence intervals |
| Regulatory | Dual decision-makers + tribunal: Case examiner pair (professional + lay) β ICP/Hearing panel (3-5 members) | Minimum 2 - Case examiners; Hearing: 3 (registrant member, lay member, legal chair) | Lay member independence - Non-professional perspective mandatory | Balance of probabilities - Not statistical, but explicit confidence assessment |
| Intelligence | 4-layer review: Peer review (tradecraft) β Management review (policy) β Red Cell (contrarian) β Customer feedback | Minimum 3 independent reviewers (research finding for reliable QC) | Red Cell independence - Protected from retaliation for contrarian views | ICD 203 standards - Confidence levels (High/Moderate/Low) required, no quantitative validation |
| Academic | 2-3 rounds: Peer review (2-3 reviewers, blinded) β Editorial decision β Post-publication critique | Minimum 2 peer reviewers - Anonymous, expert in field | Double-blind review - Reviewers and authors mutually anonymous (not always enforced) | Yes - IRR testing: Cohen's Kappa β₯0.60 acceptable, β₯0.70 preferred for qualitative coding |
Gold Standard: Intelligence's minimum 3 independent reviewers is research-backed. Academic's Cohen's Kappa β₯0.70 provides quantitative reliability measure. Legal's statistical TAR validation demonstrates defensible AI-assisted review.
4.2 Inter-Rater Reliability Standards
| Methodology | IRR Measure | Acceptable Threshold | Testing Phase | Remediation if Below Threshold |
|---|---|---|---|---|
| Police | Supervisor agreement - Qualitative assessment, not quantified | Professional consensus - No formal threshold | Ongoing supervision, spot checks | Retraining, reassignment, SIO intervention |
| Journalism | Editorial consensus - Qualitative, no statistical measure | Editorial judgment - No formal threshold | Pre-publication editorial review | Additional reporting, fact-checking, legal review |
| Legal eDiscovery | QC sampling: % agreement on responsiveness, privilege coding | β₯95% agreement for privilege; β₯90% for responsiveness | QC phase (5-10% sample after initial review) | Re-review batch, retrain reviewers, adjust guidelines |
| Regulatory | Case examiner agreement - Dual review, disagreements escalated | Consensus required - If disagreement, third examiner or ICP decides | Case examiner review stage | Additional evidence sought, legal advice, escalation to panel |
| Intelligence | Peer review consensus - Qualitative, no statistical measure | Tradecraft compliance - ICD 203 standards, no numerical threshold | Peer review before dissemination | Red Cell challenge, additional analysis, alternative hypothesis development |
| Academic | Cohen's Kappa - Statistical IRR for qualitative coding | β₯0.60 acceptable; β₯0.70 preferred; β₯0.80 excellent | Pilot coding (10-20% of corpus) | Refine codebook, retrain coders, re-pilot until threshold met |
Standout: Academic research uses quantitative IRR testing (Cohen's Kappa), providing objective reliability measure. Legal eDiscovery's β₯95% privilege agreement reflects high-stakes nature of attorney-client privilege. Intelligence's tradecraft review emphasizes methodological compliance over numerical thresholds.
5. Bias Mitigation Techniques
5.1 Structured Approaches
| Methodology | Primary Bias | Mitigation Technique | Timing | Effectiveness Evidence |
|---|---|---|---|---|
| Police | Confirmation bias - Seeking evidence confirming suspect guilt | CPIA reasonable lines of enquiry - Must pursue evidence towards AND away from suspect | Throughout investigation - Legal obligation from arrest to trial | Moderate - CPIA violations common in miscarriages of justice; structural requirement doesn't eliminate bias |
| Journalism | Confirmation bias - Story angle driving evidence selection | Hypothesis-based framework - Separate facts from assumptions, state testable hypotheses | Story planning stage - Before document review | Moderate-High - Editorial oversight + legal review catch bias; Panama Papers scale (370+ journalists) dilutes individual bias |
| Legal eDiscovery | Selection bias - Producing only favorable documents | TAR 2.0/CAL transparency - Algorithm ranks all docs, opposing party can challenge methodology | Review phase - Continuous active learning adjusts to reviewer feedback | High - Court-validated, statistically tested; opposing party oversight creates adversarial check |
| Regulatory | Professional loyalty bias - Regulators sympathetic to registrant profession | Dual decision-makers - Professional member + lay member required for balance | Case examiner + tribunal stages | Moderate-High - Lay perspective mitigates professional bias; "real prospect test" gatekeeping reduces frivolous cases |
| Intelligence | Confirmation bias, mirror imaging, groupthink | 66 Structured Analytic Techniques (SATs) - ACH, Devil's Advocacy, Red Cell, Pre-mortem | Analysis phase - Multiple SATs applied per assessment | Moderate - Research (Fisher 2008) found no empirical basis for ACH reducing bias; value is transparency, not debiasing |
| Academic | Confirmation bias, researcher expectations | Reflexivity + Audit trails - Researcher acknowledges positioning, documents all decisions | Throughout research - Memoing during data collection/analysis | Moderate - Transparency aids detection by reviewers, but doesn't eliminate bias; replication studies rare |
Critical Finding: Intelligence research (Fisher et al., 2008) challenges assumption that structured techniques eliminate biasβ"Biases cannot be eliminated by training aloneβonly mitigated through structure and tools." Value is in transparency (making reasoning auditable) rather than debiasing (eliminating cognitive distortions).
5.2 When Applied and Institutional Protection
| Methodology | Bias Mitigation Timing | Institutional Protection for Contrarian Views | Documentation Requirement |
|---|---|---|---|
| Police | Before charge decision - CPS review for prosecution bias; Pre-trial - Disclosure of unused material | IOPC independence - External oversight for serious cases; officers can report concerns | Policy File (Major Incidents) - SIO documents all strategic decisions and rationale |
| Journalism | Editorial review cycles - Multiple stages pre-publication | Editorial independence policies - Protect reporters from advertiser/owner pressure | Story memos - Reporters document sourcing, verification, editorial discussions |
| Legal eDiscovery | QC sampling throughout review - Statistical checks catch systematic bias | Court oversight - Judges sanction bad faith discovery conduct | Privilege logs, TAR validation reports - Methodology documented for court scrutiny |
| Regulatory | Case examiner stage - Before tribunal hearing | Lay member independence - Non-professionals cannot be professionally biased | Investigation reports - Evidence, rationale, alternative explanations documented |
| Intelligence | Before dissemination - Peer review, management review, Red Cell challenge | Red Cell cannot be penalized - Institutional protection for contrarian analysis | Analytic line - Assumptions, evidence, alternative hypotheses, confidence levels documented |
| Academic | Peer review pre-publication - Blinded reviewers challenge methodology, interpretations | Academic freedom - Tenure protects researchers from institutional pressure | Audit trails - Research data, coding decisions, reflexive memos, methodological rationale |
Institutional Protection Matters: Intelligence's Red Cell (established Sept 12, 2001) and Israeli Mahleket Bakara (post-Yom Kippur War 1973) demonstrate that organizational structure matters more than individual analyst training. Protecting dissenters from retaliation is critical for bias mitigation.
6. Scale Capabilities and Automation
6.1 Volume Handled and Automation Level
| Methodology | Typical Volume | Maximum Demonstrated Scale | Automation Level | Tool Maturity | Processing Time |
|---|---|---|---|---|---|
| Police | Volume crime: 10-100 docs per case; Major incidents: 1,000-10,000+ docs (HOLMES2) | Stephen Lawrence Inquiry: 100,000+ documents reviewed | Medium - HOLMES2 automated action tracking, timeline construction, pattern detection | High - HOLMES2 mature, cloud-based, 20+ years development | Volume: days-weeks; Major: months-years |
| Journalism | Investigative projects: 100,000-11.5M documents (Panama Papers: 2.6TB) | Pandora Papers: 11.9M documents, 2.9TB (600+ journalists, 150 orgs) | High - Apache Tika extraction, Solr search, Neo4j network analysis, Aleph cross-reference (4B+ docs) | High - ICIJ infrastructure battle-tested, OCCRP Aleph mature | Large projects: 12-18 months for 11M+ docs |
| Legal eDiscovery | Small: <10,000 docs; Medium: 10K-1M; Large: 1M-10M+ | Largest cases: 10M+ documents (antitrust, securities fraud) | Very High - TAR 2.0/CAL (40-60% review reduction), near-duplicate detection, email threading, privilege AI (60-80% time reduction) | Very High - Relativity, Everlaw, DISCO mature, court-validated | Small: days; Medium: weeks-months; Large: 3-12 months |
| Regulatory | Simple: 10-100 docs; Moderate: 100-1,000; Complex: 1,000-10,000 | GMC fitness-to-practice cases: rarely exceed 10,000 documents | Low - Case management systems track workflow, but analysis is manual | Medium - Regulator databases functional but not cutting-edge | Simple: 2-4 months; Complex: 12-24 months |
| Intelligence | Tactical (F3EAD): 100-1,000 docs per cycle; Strategic assessments: 1,000-100,000 sources | NSA PRISM: Billions of communications monitored (collection, not analysis) | Very High - SIGINT automated collection/processing, GEOINT automated change detection, OSINT AI scraping | High - National security budgets enable cutting-edge tech, but classified | F3EAD cycle: hours-days; Strategic: weeks-months |
| Academic | Small qualitative: 10-50 interviews/documents; Systematic reviews: 100-10,000 studies | Cochrane reviews: 10,000+ studies screened (PRISMA) | Medium - QDAS tools (NVivo, Atlas.ti, MAXQDA) for coding; Covidence/Rayyan for systematic review screening | High - QDAS tools mature, but manual coding predominates | Qualitative: 6-18 months; Systematic reviews: 12-24 months |
Scale Champions:
- Journalism handles largest volumes of heterogeneous documents (11.9M, 2.9TB Pandora Papers).
- Legal eDiscovery has most mature AI-assisted review (TAR 2.0 court-validated, 40-60% reduction).
- Intelligence achieves fastest turnaround (F3EAD cycle: hours) but often smaller analytical volumes (collection β analysis).
6.2 Cost-Benefit Analysis by Scale
| Investigation Scale | Recommended Methodology | Rationale | Typical Cost | Time to Results |
|---|---|---|---|---|
| Small (<1,000 docs) | Regulatory/Academic - Manual review feasible, structured coding provides rigor | Linear review reasonable; TAR/AI overkill for this scale | $5K-$20K - Investigator time, basic tools (Excel, Word) | 1-3 months |
| Medium (1K-10K docs) | Police/Journalism hybrid - HOLMES2-style case management + document-driven timelines | Systematic workflow essential; automation helpful but not mandatory | $20K-$100K - Case management system, dedicated team (2-5 investigators) | 3-6 months |
| Large (10K-100K docs) | Legal eDiscovery - TAR 2.0, email threading, near-duplicate detection mandatory | Manual review infeasible; AI-assisted review defensible and cost-effective | $100K-$500K - eDiscovery platform license, review team (5-20 attorneys), QC | 6-12 months |
| Very Large (100K-10M+ docs) | Journalism (Aleph/Datashare) or Legal (TAR 2.0) - Network analysis, entity extraction, AI prioritization | Only methodologies proven at this scale; academic/regulatory approaches collapse | $500K-$5M+ - Platform infrastructure, large team (20-100+), computational resources | 12-24 months |
Critical Threshold: 10,000 documents is inflection point where manual review becomes infeasible and AI-assisted review becomes cost-effective. Legal eDiscovery research: TAR 2.0 saves 40-60% review costs on cases >10K documents.
7. Legal Defensibility and Court Acceptance
7.1 Standards Bodies and Court Validation
| Methodology | Standards Body | Court Acceptance | Audit Trail Requirements | Expert Testimony Standard |
|---|---|---|---|---|
| Police | College of Policing (UK), IACP (International), FBI (US) | Criminal courts worldwide - PACE, CPIA compliance required for admissibility | Mandatory - Every action, decision, interview logged; Policy File for major incidents | FRE 702/Daubert (US), Ikarian Reefer (UK) - Expert must state methodology, limitations |
| Journalism | No legal standard - Ethical guidelines (SPJ Code, Reuters Trust Principles) but not court-validated | Not court evidence - Used for libel defense ("truth" or "public interest"), not direct admissibility | Ethical obligation - Source protection, fact-checking logs, editorial review notes | Not applicable - Journalists generally not expert witnesses (shield laws) |
| Legal eDiscovery | EDRM (145 countries), Sedona Conference (thought leadership), FRE/FRCP (US), CPR (UK) | Highest court acceptance - TAR 2.0 mandated over linear review (Hyles v. NYC 2016), FRE 502(b) safe harbor | Extensive - Hash certification, metadata preservation, privilege logs, production logs, TAR validation reports | FRE 702 + FRE 902(14) - Forensic experts certify data integrity, ESI self-authenticates with certification |
| Regulatory | GMC, HCPC, NMC, BPS (UK), NPDB, FSMB (US) - Professional regulators, not courts | Tribunal standard - Balance of probabilities, civil rules; judicial review possible but rare | Moderate - Investigation reports, case examiner rationale, hearing transcripts | Expert witnesses - Clinical/professional standards experts common, independence required |
| Intelligence | ODNI ICD 203 (US), JIC standards (UK), NATO AJP-2 (alliance) | Not court-focused - Supports policy/military decisions, rarely admissible in court (state secrets) | Classified audit trails - Analytic line, source protection, dissemination controls | Not applicable - Intelligence assessments not for courtroom use |
| Academic | APA, ASA, BPS (discipline-specific), COPE, ICMJE (publication ethics) | Expert testimony - Researchers as expert witnesses; research findings cited in amicus briefs | Scholarly standard - Research data repositories, IRB approvals, replication datasets (increasingly required) | Daubert/Frye - Scientific methodology must be peer-reviewed, testable, accepted in field |
Court-Ready Methodologies: Legal eDiscovery is designed for court admissibility (FRE 902(14), hash certification). Police evidence follows criminal court standards (PACE, CPIA). Regulatory uses civil tribunal standards (balance of probabilities). Intelligence and Journalism not court-focused (operational/editorial decisions). Academic indirectly used (expert witnesses, cited research).
7.2 Precedent Cases and Validation
| Methodology | Landmark Cases/Validation Events | Impact on Standards | Current Status (2026) |
|---|---|---|---|
| Police | Zubulake v. UBS Warburg (2004) - Duty to preserve when litigation anticipated; Stephen Lawrence Inquiry (1999) - Racism in Met Police, led to PACE reforms | CPIA 1996 obligations strengthened; IOPC established 2018 (replaced IPCC); Body-worn cameras mandated | Mature - Standards stable, incremental improvements (AI video analysis emerging) |
| Journalism | Pentagon Papers (1971) - First Amendment protects publication of classified docs; Panama Papers (2016) - 11.5M docs, no successful libel suits | Established "public interest" defense; validated massive leak journalism methodologies | Evolving - ICIJ model now standard for cross-border investigations; AI-assisted document review growing |
| Legal eDiscovery | Da Silva Moore v. Publicis (2012) - First TAR 1.0 approval; Rio Tinto v. Vale (2015) - TAR 2.0/CAL approved; Hyles v. NYC (2016) - TAR mandated over linear review | TAR now industry standard; AI-assisted review defensible; Proportionality amendment (FRCP 2015) | Mature and evolving - TAR 2.0 standard practice; Generative AI (GPT-4) entering eDiscovery (2023-2026) |
| Regulatory | GMC v. Meadow (2006) - Expert witness overreach; Bawa-Garba (2018) - Manslaughter conviction of doctor, led to "learning not blaming" culture shift | Expert evidence standards tightened; "Real prospect test" clarified; Apology legislation (2019) - saying sorry doesn't admit liability | Stable - 2025-2026 HCPC policy shift away from "last resort" language for striking-off; otherwise stable standards |
| Intelligence | 9/11 Commission (2004) - Intelligence failures led to ODNI creation, ICD 203; Iraq WMD (Butler Review 2004) - UK JIC reforms, Red Teaming institutionalized | ICD 203 tradecraft standards mandated (2015); Red Cell/Mahleket Bakara models adopted widely | Mature - Standards stable; AI/ML integration ongoing but classified; OSINT growing (80-90% of intel in some domains) |
| Academic | Reproducibility Crisis (2010s) - Psychology, social sciences replication failures; PRISMA 2020 - Updated reporting standards | Pre-registration mandated by journals; Open data requirements; IRR reporting standard | Evolving - PRISMA extensions (NMA, QES) in development; AI-assisted coding emerging but controversial |
Trend: All methodologies show increasing automation and AI integration (TAR 2.0 in Legal, OSINT in Intelligence, QDAS in Academic). Transparency and replicability are growing requirements across all domains (PRISMA, ICD 203, CPIA disclosure).
8. Strengths and Weaknesses Matrix
8.1 Comparative Assessment
| Dimension | Police | Journalism | Legal eDiscovery | Regulatory | Intelligence | Academic |
|---|---|---|---|---|---|---|
| Evidence Authentication | π’ Very Strong - FBI 5-step, write-blocking, hash certification | π‘ Moderate - Ethical rigor but not legal standard | π’ Very Strong - FRE 902, dual hash, metadata preservation | π‘ Moderate - Contemporary records strong, but hearsay admissible | π’ Strong - Technical validation (SIGINT, IMINT) but classified | π‘ Moderate - Peer review, citation, but not legal authentication |
| Timeline Construction | π’ Strong - HOLMES2 automated, 5WH framework | π’ Strong - Document-driven, Bates-style referencing | π’ Very Strong - 8-step process, mandatory evidence linking | π‘ Moderate - Manual construction, focus on current impairment | π’ Strong - F3EAD cycle (hours), multi-INT fusion | π‘ Moderate - Qualitative timelines, no standardized process |
| Bias Mitigation | π‘ Moderate - CPIA obligations, but violations common | π’ Strong - Editorial layers, 3-source rule, legal review | π’ Very Strong - TAR transparency, opposing party oversight | π’ Strong - Dual decision-makers (professional + lay) | π‘ Moderate - SATs transparency, but no empirical debiasing proof | π‘ Moderate - Reflexivity, audit trails, but individual researcher bias |
| Quality Control | π’ Strong - 3-tier review, IOPC oversight | π’ Strong - 4-stage editorial review | π’ Very Strong - QC sampling (β₯95% agreement), statistical validation | π’ Strong - Dual case examiners, tribunal review | π’ Very Strong - Minimum 3 reviewers, Red Cell challenge | π’ Strong - Peer review, Cohen's Kappa β₯0.70 |
| Scalability | π‘ Moderate - HOLMES2 handles 10K docs well, but manual analysis limits scale | π’ Very Strong - Proven at 11.9M docs (Pandora Papers) | π’ Very Strong - TAR 2.0 handles 10M+ docs, 40-60% reduction | π΄ Weak - Manual review limits to <10K docs typically | π’ Strong - F3EAD cycle handles large collection volumes, but analysis often smaller | π‘ Moderate - Systematic reviews handle 10K studies, but qualitative limited to <100 |
| Speed | π‘ Moderate - Volume crime: weeks; Major incidents: months-years | π΄ Slow - Large projects: 12-18 months | π‘ Moderate - Small: days; Large: 6-12 months | π΄ Slow - Simple: 2-4 months; Complex: 12-24 months | π’ Very Fast - F3EAD: hours-days (tactical); Strategic: weeks-months | π΄ Slow - Qualitative: 6-18 months; Systematic reviews: 12-24 months |
| Legal Defensibility | π’ Very Strong - Criminal court standards (beyond reasonable doubt) | π΄ Weak - Not court-focused, editorial/ethical standards | π’ Very Strong - Designed for court admissibility (FRE, FRCP) | π’ Strong - Civil tribunal standards (balance of probabilities) | π΄ Weak - Operational/policy focus, rarely court-admissible | π‘ Moderate - Expert testimony accepted, but not legal evidence standard |
| Cost Efficiency | π‘ Moderate - Public funding, but major incidents expensive (Β£millions) | π΄ Expensive - Large projects require consortia (Panama Papers: 370+ journalists) | π΄ Very Expensive - eDiscovery platforms + attorney review ($100K-$5M+) | π’ Strong - Public sector funding, smaller scale cases | π΄ Very Expensive - National security budgets, classified infrastructure | π’ Strong - Academic funding models, lower labor costs (grad students) |
| Transparency | π‘ Moderate - CPIA disclosure required, but Policy Files restricted | π’ Strong - Methodology published, sources protected but process transparent | π’ Very Strong - Privilege logs, TAR validation, production logs all auditable | π’ Strong - Investigation reports, tribunal transcripts public | π΄ Weak - Classified analytic lines, sources protected (but ICD 203 requires internal transparency) | π’ Very Strong - Peer review, replication data, audit trails, open access |
| Tool Maturity | π’ Strong - HOLMES2 mature (20+ years), cloud-based | π’ Strong - ICIJ/OCCRP tools battle-tested, Aleph 4B+ docs | π’ Very Strong - Relativity, Everlaw, DISCO mature, court-validated | π‘ Moderate - Case management functional but not cutting-edge | π’ Strong - National security budgets enable advanced tech (but classified) | π’ Strong - NVivo, Atlas.ti, MAXQDA mature, PRISMA/Cochrane frameworks established |
Legend:
- π’ = Strong capability (competitive advantage)
- π‘ = Moderate capability (adequate but not exceptional)
- π΄ = Weak capability (limitation or disadvantage)
8.2 Summary Scorecard
| Methodology | Overall Strengths | Critical Weaknesses | Best Use Case Match |
|---|---|---|---|
| Police | Evidence authentication, quality control, legal defensibility | Bias mitigation (CPIA violations common), speed (major incidents years), cost | Criminal investigations - When court admissibility and chain of custody critical |
| Journalism | Scalability (11M+ docs), transparency, editorial review | Speed (12-18 months), cost (consortia required), legal defensibility (not court-focused) | Large-scale leaks, public interest investigations - When volume is massive and publication goal |
| Legal eDiscovery | Scalability, automation (TAR 2.0), legal defensibility, quality control (statistical) | Cost (most expensive), tool complexity | Litigation, regulatory enforcement - When court admissibility and AI-assisted review needed |
| Regulatory | Bias mitigation (dual decision-makers), quality control, cost efficiency | Scalability (limited to <10K docs), speed (12-24 months), tool maturity | Professional misconduct, standards violations - When current impairment assessment required |
| Intelligence | Speed (F3EAD: hours), quality control (3+ reviewers, Red Cell), multi-INT fusion | Legal defensibility (not court-focused), transparency (classified), bias mitigation (no empirical proof) | Operational intelligence, time-sensitive analysis - When speed and multi-source fusion critical |
| Academic | Transparency (peer review, open data), quality control (IRR testing), tool maturity (QDAS) | Speed (6-24 months), scalability (qualitative <100), legal defensibility | Systematic reviews, qualitative research - When peer-reviewed rigor and replicability essential |
9. Decision Tree: Which Methodology for Which Investigation?
9.1 Primary Decision Factors
START: What is the investigation goal?
1. COURT ADMISSIBILITY REQUIRED?
ββ YES β Use POLICE or LEGAL eDISCOVERY
β ββ Criminal prosecution? β POLICE (beyond reasonable doubt)
β ββ Civil litigation? β LEGAL eDISCOVERY (balance of probabilities)
ββ NO β Continue to #2
2. DOCUMENT VOLUME?
ββ <1,000 docs β REGULATORY or ACADEMIC (manual review feasible)
ββ 1K-10K docs β POLICE or JOURNALISM (case management essential)
ββ 10K-100K docs β LEGAL eDISCOVERY (TAR 2.0 mandatory)
ββ >100K docs β JOURNALISM (Aleph/Datashare) or LEGAL (TAR)
3. TIME SENSITIVITY?
ββ Hours-days β INTELLIGENCE (F3EAD cycle)
ββ Weeks-months β POLICE or LEGAL eDISCOVERY
ββ Not time-sensitive β REGULATORY or ACADEMIC
4. BUDGET?
ββ <$20K β REGULATORY or ACADEMIC (public/grant funding)
ββ $20K-$100K β POLICE (public funding)
ββ $100K-$500K β LEGAL eDISCOVERY (mid-sized case)
ββ >$500K β JOURNALISM (consortium) or LEGAL (large case)
5. PROFESSIONAL STANDARDS ASSESSMENT?
ββ YES β REGULATORY (GMC, HCPC, NMC, BPS frameworks)
ββ NO β Continue to #6
6. PEER-REVIEWED RIGOR REQUIRED?
ββ YES β ACADEMIC (PRISMA, Cochrane, IRR testing)
ββ NO β Default to POLICE or JOURNALISM
9.2 Use Case Matching Table
| Investigation Type | Primary Methodology | Secondary (Hybrid) | Rationale |
|---|---|---|---|
| Criminal investigation | Police | + Legal eDiscovery (if digital evidence volume >10K) | Court admissibility, chain of custody, beyond reasonable doubt standard |
| Civil litigation discovery | Legal eDiscovery | + Police (if criminal crossover) | Court admissibility, TAR 2.0 efficiency, statistical validation |
| Regulatory enforcement (SEC, FDA, FCA) | Legal eDiscovery | + Regulatory (professional standards) | Document volume typically >10K, civil standard, court admissibility |
| Professional misconduct (medical, legal, psychological) | Regulatory | + Academic (systematic review for precedent) | Current impairment focus, professional standards, balance of probabilities |
| Institutional accountability (corruption, abuse) | Journalism | + Academic (qualitative analysis for patterns) | Public interest, transparency, large document volumes (leaks) |
| National security threat assessment | Intelligence | + Academic (open-source research) | Speed (F3EAD), multi-INT fusion, operational focus |
| Systematic evidence synthesis | Academic | + Journalism (if large document corpus) | Peer-reviewed rigor, PRISMA transparency, IRR testing |
| Internal corporate investigation | Legal eDiscovery | + Police (if criminal referral likely) | Document volume management, attorney-client privilege, audit trail |
| Historical analysis (truth commissions) | Academic | + Journalism (document-driven narratives) | Long timescales, qualitative depth, peer review |
| Complaint triage (initial assessment) | Regulatory | + Intelligence (ACH for complex cases) | Fast decision-making, balance of probabilities, real prospect test |
10. Integration Recommendations for Phronesis FCIP
10.1 Hybrid Methodology Framework
Phronesis FCIP should adopt a hybrid methodology integrating strengths across domains:
| Investigation Phase | Primary Methodology | Rationale | Phronesis Implementation |
|---|---|---|---|
| Document Assembly | Legal eDiscovery (EDRM) + Academic (PRISMA) | Systematic identification, transparent selection, deduplication | investigate.rs orchestration: PRISMA flow diagram, hash-based deduplication, metadata preservation |
| Timeline Construction | Legal eDiscovery (8-step) + Police (5WH) | Evidence linking mandatory, objective language, temporal contradiction detection | temporal.rs engine: Bates-style references, event-document linking, gap analysis, version tracking |
| Contradiction Detection | Intelligence (ACH) + Legal (near-duplicate detection) | Hypothesis testing, multi-document comparison, version diff analysis | S.A.M. contradiction engine: 8 types (TEMPORAL, EVIDENTIARY, MODALITY_SHIFT, etc.) with ACH integration |
| Bias Mitigation | Intelligence (SATs, Red Cell) + Regulatory (dual decision-makers) | Structured techniques, contrarian review, institutional protection for dissenters | bias.rs engine: 66 SATs catalog, Red Cell mode (alternative explanations), transparency dashboard |
| Quality Control | Academic (Cohen's Kappa β₯0.70) + Legal (QC sampling) | Statistical IRR, peer review, audit trails | qc.rs module: Sample findings for human review, calculate precision/recall, version control all changes |
| Source Reliability | Intelligence (Admiralty Code) + Journalism (evidence hierarchy) | Source rating (A-F) + information credibility (1-6), documentary > testimonial | evidence.rs: Source reliability ratings, information credibility assessment, provenance tracking |
| Automation | Legal (TAR 2.0/CAL) + Journalism (Aleph network analysis) | AI-assisted review (40-60% reduction), entity extraction, network graphs | AI prioritization: Continuous active learning ranks documents, entity relationship networks, concept clustering |
| Professional Standards | Regulatory (GMC/HCPC/NMC/BPS) | Standards mapping, current impairment assessment, contributory factors analysis | professional.rs engine: Code violation detection, severity assessment, remediation tracking |
| Reporting | Academic (Framework Method) + Intelligence (ICD 203) | Matrix-based analysis (case x theme), confidence levels (High/Moderate/Low), Words of Estimative Probability | Export module: Framework matrices, confidence assessments, PRISMA diagrams, audit package with hashes |
10.2 Recommended Tool Stack
| Function | Tool Category | Specific Recommendation | Rationale |
|---|---|---|---|
| Case Management | Police-inspired | HOLMES2-style architecture (cases, documents, actions, entities, timelines) | Proven at 10K+ docs, systematic workflow |
| Document Processing | Legal eDiscovery | De-duplication (hash-based), metadata extraction, OCR, email threading | EDRM standards, court-validated |
| AI Review | Legal eDiscovery | TAR 2.0/Continuous Active Learning | 40-60% review reduction, court-approved |
| Timeline Engine | Legal + Police | 8-step process with HOLMES2-style timeline view | Evidence linking mandatory, gap analysis |
| Network Analysis | Journalism | Neo4j-style entity relationship graphs | Proven at Panama Papers scale (11.5M docs) |
| Contradiction Detection | Intelligence + Legal | ACH matrix + near-duplicate detection + version diff | Hypothesis testing + automated flagging |
| Coding Framework | Academic | NVivo/Atlas.ti-inspired hierarchical coding with Framework matrices | Systematic qualitative analysis, IRR testing |
| Quality Control | Academic + Legal | Cohen's Kappa calculation + QC sampling dashboard | Statistical reliability + spot-check validation |
| Reporting | Academic | PRISMA flow diagram + Framework matrix export + confidence levels | Transparency, auditability, peer review-ready |
10.3 Feature Priority Matrix
| Feature | Priority | Methodology Source | Implementation Phase | Complexity |
|---|---|---|---|---|
| Hash certification (SHA-256) | π΄ Critical | Legal eDiscovery | Phase 1 (Core) | Low |
| Timeline with evidence linking | π΄ Critical | Legal + Police | Phase 1 (Core) | Medium |
| S.A.M. contradiction detection (8 types) | π΄ Critical | Intelligence + Legal | Phase 1 (Core) | High |
| Document de-duplication | π΄ Critical | Legal eDiscovery | Phase 1 (Core) | Low |
| Metadata preservation | π΄ Critical | Legal eDiscovery | Phase 1 (Core) | Low |
| Admiralty Code source ratings | π High | Intelligence | Phase 2 (Enhanced) | Low |
| ACH matrix implementation | π High | Intelligence | Phase 2 (Enhanced) | High |
| TAR 2.0/CAL prioritization | π High | Legal eDiscovery | Phase 2 (Enhanced) | Very High |
| Framework matrix export | π High | Academic | Phase 2 (Enhanced) | Medium |
| Entity extraction + network graphs | π‘ Medium | Journalism | Phase 3 (Advanced) | High |
| Near-duplicate detection | π‘ Medium | Legal eDiscovery | Phase 3 (Advanced) | Medium |
| Cohen's Kappa IRR calculation | π‘ Medium | Academic | Phase 3 (Advanced) | Low |
| Red Cell alternative analysis | π‘ Medium | Intelligence | Phase 3 (Advanced) | Medium |
| Professional standards mapping | π‘ Medium | Regulatory | Phase 2 (Enhanced) | Medium |
| PRISMA flow diagram generation | π’ Nice-to-have | Academic | Phase 4 (Polish) | Low |
| Words of Estimative Probability | π’ Nice-to-have | Intelligence | Phase 4 (Polish) | Low |
Legend:
- π΄ Critical = Core functionality, must-have for MVP
- π High = Important for professional use, implement soon
- π‘ Medium = Valuable enhancement, implement when resources permit
- π’ Nice-to-have = Refinement, polish, future roadmap
11. Cost and Resource Requirements
11.1 Typical Team Composition by Methodology
| Methodology | Small Investigation | Medium Investigation | Large Investigation | Specialized Roles |
|---|---|---|---|---|
| Police | 1-2 officers (volume crime) | 5-10 officers + SIO + analyst | 20-50+ officers + MIR team (SIO, Deputy, analysts, indexers, disclosure officer) | Forensic specialists, intelligence officers, family liaison |
| Journalism | 1-3 reporters + editor | 5-15 reporters + editors + fact-checker | 100-600+ reporters (Panama Papers: 370+) + editors + data analysts | Data journalists, digital forensics, legal advisors |
| Legal eDiscovery | 2-5 attorneys (review + QC) | 10-50 attorneys + project manager + IT support | 50-200+ attorneys + managed review teams + forensic experts | eDiscovery consultants, forensic examiners, privilege specialists |
| Regulatory | 1 case officer + 2 case examiners | 2-3 investigators + expert witness | 5-10 investigators + multiple experts + legal advisor | Clinical/professional experts, legal advisors, HR specialists |
| Intelligence | 3-5 analysts + manager (minimum 3 reviewers for QC) | 10-20 analysts + collection management + targeteers | 50-100+ analysts + collectors + all-source fusion + Red Cell | SIGINT analysts, GEOINT specialists, HUMINT handlers, targeteers |
| Academic | 1-2 researchers (PhD students) + supervisor | 3-5 researchers + PI + research assistants | 10-20 researchers (systematic review team) + librarian + statistician | Research assistants, coders, statisticians, methodologists |
Key Insight: Intelligence requires minimum 3 independent reviewers for reliable quality control (research-backed). Legal eDiscovery scales to largest teams (200+ attorneys for 10M+ doc cases). Academic most efficient for small-scale investigations (1-2 researchers).
11.2 Time to Results
| Methodology | Small (<1K docs) | Medium (1K-10K) | Large (10K-100K) | Very Large (>100K) |
|---|---|---|---|---|
| Police | 2-4 weeks (volume crime) | 2-6 months (serious crime) | 6-18 months (major incident) | 12-36 months (complex fraud) |
| Journalism | 1-3 months (single-story investigation) | 6-12 months (investigative series) | 12-18 months (cross-border project) | 12-24 months (Panama Papers: 12 months) |
| Legal eDiscovery | 1-2 weeks (TAR on small set) | 1-3 months (medium case) | 3-12 months (large litigation) | 12-24 months (10M+ docs, complex) |
| Regulatory | 2-4 months (simple case) | 6-12 months (moderate case) | 12-18 months (complex case) | Rare (regulatory cases rarely exceed 10K docs) |
| Intelligence | Hours-days (F3EAD tactical cycle) | Weeks (operational intelligence) | 1-6 months (strategic assessment) | Not applicable (collection β analysis) |
| Academic | 3-6 months (small qualitative study) | 6-12 months (moderate study) | 12-18 months (large qualitative/systematic review) | 18-36 months (very large systematic review) |
Fastest: Intelligence (F3EAD: hours-days for tactical). Slowest: Regulatory and Academic (12-36 months for complex). Most scalable: Legal eDiscovery (TAR 2.0 handles 10M+ docs in 12 months).
11.3 Tool Costs (2026 Estimates)
| Tool Category | Free/Open Source | Mid-Tier Commercial | Enterprise/Premium | Typical Use Case |
|---|---|---|---|---|
| Case Management | Excel, Google Sheets ($0) | Airtable, Notion ($10-20/user/month) | HOLMES2, Relativity ($50K-$500K/year) | Small: Free; Medium: Mid-tier; Large: Enterprise |
| Document Processing | Tika, Tesseract OCR ($0) | Nuix, Exterro ($50K-$200K/year) | Relativity Processing (included in platform) | Small: Free; Medium-Large: Commercial |
| eDiscovery Platform | None viable | Logikcull ($250-$10K/month) | Relativity, Everlaw, DISCO ($100K-$5M/year) | Small: Logikcull; Large: Relativity/Everlaw |
| QDAS (Qualitative) | QualCoder, Taguette ($0) | Dedoose ($10-30/month) | NVivo, Atlas.ti, MAXQDA ($600-$1,500/license) | Small: Free; Medium-Large: NVivo/Atlas |
| Network Analysis | Gephi, Cytoscape ($0) | NodeXL ($500/year) | Palantir, i2 Analyst's Notebook ($10K-$100K/year) | Small-Medium: Gephi; Large: Palantir (gov/corp only) |
| Systematic Review | Rayyan (free tier) ($0-$1K/year) | Covidence ($1K-$5K/year) | DistillerSR, Nested Knowledge ($5K-$20K/year) | Small: Rayyan; Medium-Large: Covidence |
| Timeline Software | Excel, Google Sheets ($0) | Aeon Timeline, TimelineJS ($50-$100) | CaseFleet, TimeMap ($500-$5K/year) | Small: Excel; Medium-Large: CaseFleet |
Cost-Effective Stack for <$5K:
- Case management: Airtable ($240/year)
- Document processing: Tika + Tesseract (free)
- QDAS: NVivo ($1,500 one-time)
- Network analysis: Gephi (free)
- Systematic review: Rayyan (free tier)
- Timeline: Excel/Google Sheets (free)
- Total: ~$2,000 (viable for small investigations)
Enterprise Stack for Large Investigations ($100K-$1M):
- eDiscovery platform: Relativity ($200K-$500K/year)
- Processing: Nuix ($100K/year)
- Network analysis: i2 Analyst's Notebook ($50K/year)
- Systematic review: Covidence ($5K/year)
- Total: $355K-$655K/year (large-scale, court-focused)
12. Hybrid Approach Recommendations
12.1 Common Hybrid Patterns
| Hybrid Combination | Use Case | Strengths Combined | Example |
|---|---|---|---|
| Police + Legal eDiscovery | Criminal investigation with large digital evidence volume (>10K docs) | Court admissibility (Police) + TAR 2.0 efficiency (Legal) | Fraud investigation: Police chain of custody + eDiscovery TAR review |
| Journalism + Academic | Investigative research projects, historical analysis | Large-scale document handling (Journalism) + peer-reviewed rigor (Academic) | Truth commission: Journalism document assembly + Academic qualitative analysis |
| Legal + Intelligence | Complex civil litigation requiring rapid prioritization | Court admissibility (Legal) + ACH hypothesis testing (Intelligence) | Securities fraud: Legal TAR + Intelligence ACH for multiple defendant theories |
| Regulatory + Academic | Professional standards violations with precedent research | Current impairment assessment (Regulatory) + systematic evidence synthesis (Academic) | Medical misconduct: Regulatory case + Academic systematic review of prior cases |
| Intelligence + Police | Counter-terrorism, organized crime | F3EAD speed (Intelligence) + Court admissibility (Police) | Counter-terror: Intelligence F3EAD targeting + Police arrest/prosecution evidence |
| Journalism + Legal | High-profile public interest litigation | Transparency (Journalism) + Legal defensibility (Legal) | Whistleblower case: Journalism leak analysis + Legal eDiscovery for court |
Most Common Hybrid: Police + Legal eDiscovery (digital evidence in criminal cases). Most Effective: Journalism + Academic (combines scale with rigor for research-focused investigations).
12.2 When to Combine Methodologies
Combine methodologies when:
-
Volume exceeds methodology's typical scale
- Example: Regulatory case with >10K documents β Add Legal eDiscovery TAR 2.0
-
Multiple goals require different standards
- Example: Investigation for both internal discipline (Regulatory) AND criminal prosecution (Police)
-
Speed and rigor both critical
- Example: Safeguarding investigation requiring F3EAD speed (Intelligence) + court admissibility (Police)
-
Cross-border or multi-jurisdictional
- Example: Journalism collaboration (ICIJ model) + Legal eDiscovery (discovery coordination)
-
Public interest + legal accountability
- Example: Journalism transparency (public reporting) + Legal defensibility (litigation)
Don't combine when:
- Single methodology already handles scale and requirements (avoid unnecessary complexity)
- Methodologies have conflicting standards (e.g., Journalism source protection vs. Police disclosure obligations)
- Budget insufficient for dual tooling
13. Conclusion and Key Takeaways
13.1 Methodology Selection Summary
If you need...
- Criminal court admissibility β Police (beyond reasonable doubt, PACE/CPIA compliance)
- Civil court admissibility β Legal eDiscovery (balance of probabilities, TAR validated)
- Large-scale document handling (>100K) β Journalism (Aleph 4B+ docs) or Legal (TAR 2.0)
- Fastest results (hours-days) β Intelligence (F3EAD cycle)
- Professional standards assessment β Regulatory (GMC/HCPC/NMC/BPS)
- Peer-reviewed rigor β Academic (PRISMA, Cochrane, IRR β₯0.70)
- Transparency and public accountability β Journalism (editorial review, publication) or Academic (open data)
- Cost-efficiency (<$20K) β Regulatory or Academic (public/grant funding)
- Bias mitigation with institutional protection β Intelligence (Red Cell, minimum 3 reviewers)
- Statistical validation of findings β Legal (TAR precision/recall) or Academic (Cohen's Kappa)
13.2 Critical Insights for Forensic Intelligence Platforms
-
No single methodology handles all requirements. Phronesis FCIP must adopt hybrid approach integrating:
- Legal eDiscovery (evidence authentication, timeline construction, TAR automation)
- Intelligence (bias mitigation via SATs, ACH, Red Cell, Admiralty Code)
- Academic (quality control via IRR testing, Framework Method, audit trails)
- Regulatory (professional standards, current impairment assessment)
-
10,000 documents is inflection point where manual review becomes infeasible and AI-assisted review (TAR 2.0) becomes cost-effective (40-60% reduction).
-
Transparency β Debiasing. Intelligence research (Fisher 2008) found no empirical basis for SATs eliminating bias. Value is in making reasoning auditable, not bias-free. Combine with peer review and contrarian challenge (Red Cell).
-
Minimum 3 independent reviewers required for reliable quality control (Intelligence finding, research-backed).
-
Evidence linking is non-negotiable. Every timeline event, contradiction, finding must cite supporting documents (Bates numbers, page/line references). Legal eDiscovery's 8-step timeline + Journalism's document-driven approach are gold standards.
-
Hash certification (SHA-256) + metadata preservation are baseline requirements for evidence integrity. Legal eDiscovery standards (FRE 902(14), dual hash) should be adopted universally.
-
Source reliability (Admiralty Code) independent of information credibility. Reliable source can report bad information (deceived); unreliable source can report true information (broken clock). Rate independently.
-
Words of Estimative Probability (WEP) should replace vague language ("likely" = 60-80%, not subjective interpretation). Separate probability (likelihood of event) from confidence (quality of evidence).
-
Court validation matters. Legal eDiscovery (TAR mandated in Hyles v. NYC 2016) and Police (PACE/CPIA compliance) have strongest legal defensibility. If court admissibility possible, design for it from start.
-
Tool maturity varies dramatically. Legal eDiscovery (Relativity, Everlaw), Journalism (Aleph), Academic (NVivo, Atlas.ti) have mature, battle-tested tools. Regulatory lags significantly (Excel-based case management common).
13.3 Implementation Priorities for Phronesis FCIP
Phase 1 (Core - MVP):
- Hash certification (SHA-256) for evidence integrity
- Timeline engine with mandatory evidence linking (Legal 8-step + Police 5WH)
- S.A.M. contradiction detection (8 types: TEMPORAL, EVIDENTIARY, MODALITY_SHIFT, etc.)
- Document de-duplication (hash-based, EDRM standard)
- Metadata preservation (Legal eDiscovery standard)
Phase 2 (Enhanced - Professional Use): 6. Admiralty Code source reliability ratings (Intelligence) 7. ACH matrix implementation (Intelligence hypothesis testing) 8. Professional standards mapping (Regulatory: GMC/HCPC/NMC/BPS) 9. Framework matrix export (Academic: case x theme analysis) 10. TAR 2.0/Continuous Active Learning prioritization (Legal eDiscovery)
Phase 3 (Advanced - Specialized Features): 11. Entity extraction + network graphs (Journalism: Neo4j-style) 12. Near-duplicate detection + version diff (Legal eDiscovery) 13. Cohen's Kappa IRR calculation (Academic quality control) 14. Red Cell alternative analysis mode (Intelligence contrarian review) 15. Quality control dashboard with precision/recall metrics
Phase 4 (Polish - Reporting Enhancements): 16. PRISMA flow diagram generation (Academic transparency) 17. Words of Estimative Probability templates (Intelligence reporting) 18. Confidence level assessment framework (High/Moderate/Low) 19. Audit package export (report + hashes + source docs) 20. Framework matrix visualization (Academic Framework Method)
13.4 Final Recommendation
For Phronesis FCIP: Adopt Legal eDiscovery as architectural foundation (evidence authentication, timeline construction, document management) with targeted enhancements from other methodologies:
- Intelligence β Bias mitigation (SATs, ACH, Red Cell), source reliability (Admiralty Code)
- Academic β Quality control (IRR testing, Framework Method), transparency (audit trails)
- Regulatory β Professional standards (GMC/HCPC/NMC/BPS), current impairment assessment
- Journalism β Network analysis (entity relationships), large-scale document handling
- Police β Investigative workflow (5WH framework), chain of custody
This hybrid approach provides:
- Legal defensibility (FRE 902(14), hash certification, metadata preservation)
- Scalability (TAR 2.0 handles 10M+ docs)
- Quality control (IRR testing, peer review, Red Cell challenge)
- Transparency (audit trails, PRISMA diagrams, Framework matrices)
- Professional standards compliance (Regulatory frameworks integrated)
Result: A forensic intelligence platform that combines court-validated rigor (Legal) with peer-reviewed quality (Academic) and operational speed (Intelligence), suitable for investigations from 100 documents to 10 million.
Document Version: 1.0 Last Updated: January 2026 Next Review: Annually or upon significant methodology updates (e.g., PRISMA extensions, new TAR court cases) Maintained By: Phronesis FCIP Research Team Purpose: Living reference document for methodology selection and system design decisions
Sources and References
All findings in this matrix are synthesized from the six comprehensive methodology research documents:
- Police Investigation Workflows - College of Policing, HOLMES2, PEACE, CPIA
- Investigative Journalism Methods - ICIJ, OCCRP, Panama Papers, Aleph
- Legal eDiscovery Workflows - EDRM, TAR 2.0, Cochrane-influenced approaches
- Regulatory Investigations - GMC, HCPC, NMC, BPS professional standards
- Intelligence Analysis Methods - CIA, UK JIC, 66 SATs, ACH, ICD 203
- Academic Research Methods - PRISMA 2020, Cochrane v6.5, Grounded Theory, QDAS
See individual methodology documents for full citations and source materials.