Predictive History Audit / Systematic Content Analysis

Methodology

How we analyze each lecture, what we measure, and why. Transparency is the point.

Process

Each lecture in the Predictive History corpus is analyzed through a systematic, repeatable pipeline:

  1. Transcript extraction — Auto-generated English subtitles downloaded via yt-dlp in SRT format with timestamps.
  2. AI analysis — The full transcript is provided to Claude Opus 4.6 with the analysis schema and scoring rubric. The model reads the entire transcript and produces a structured JSON assessment.
  3. Screencap extraction — For every timestamped item (quotes, rhetoric, predictions), a video frame is captured automatically via ffmpeg.
  4. Publication — JSON is rendered to static HTML. No editorial intervention beyond the rubric design.

Prediction Tracking & Calibration

Every falsifiable prediction and claim extracted from the corpus carries a status — confirmed, partially_confirmed, disconfirmed, untested, or unfalsifiable — re‑evaluated against real‑world events. The reference document that drives scoring is briefing-data.json, accompanied by a human‑readable calibration reference and a long‑form geopolitical briefing. None of these files are hand‑edited; they are machine‑maintained by Claude Opus 4.6 under two scheduled runs.

Update cadence

Script Schedule Scope
daily-briefing-update.sh 06:00 daily Prepends a new entry to daily_entries, refreshes theatre summaries, metrics, and the ground‑invasion tracker. Also edits the calibration reference and the geopolitical briefing.
monthly-scoring-run.sh 07:00 on the 1st Same daily refresh, plus walks every analysis file and rewrites thesis.predictions[*].status / status_note against current events.

How subject matter enters the briefing

The daily script's prompt hard‑codes a tracking list: Iran / Strait of Hormuz, dedicated ground‑invasion indicators (troop movements, Kharg Island, amphibious assets, draft signals, logistics staging), Gulf states, the Russia–Ukraine war, Latin America (Venezuela / Cuba / Colombia), US–China and Taiwan, the Korean peninsula, and countries exposed to a Hormuz blockade (Japan, South Korea, India, Pakistan, Bangladesh, Europe). For each theatre Claude runs web searches for the day's developments and decides what's material.

Subject matter broadens three ways:

  1. A new daily entry. Each record carries its own date / tags / summary / details / prediction_impact / source — additive, no schema change.
  2. A new theatre. The prompt explicitly allows Claude to add new theatre sections when a story grows large enough to warrant separate tracking.
  3. A new metric. The metrics array is free‑form; entries are added or retired as the situation evolves (e.g. Ships Trapped in Gulf, Qatar Gas Production were added after the Gulf crisis escalated).

The prediction_impact field on each daily entry is the bridge to the predictions tables — it's the claim‑to‑evidence linkage the monthly run consumes when re‑scoring statuses.

Sources used to monitor falsifiable claims

Sources are cited per entry in the source field of each daily record, not a fixed global list. Recurring outlets across recent entries:

The baseline anchor set used for the calibration reference is narrower — Al Jazeera, BBC, Reuters, NPR, CNN, Washington Post, CNBC, plus institutional research from CRS, CSIS, Atlantic Council, Chatham House, Carnegie Endowment, and the Arms Control Association — but the tooling does not enforce an allow‑list.

What the tooling does not do

Scoring Rubric

Each lecture is evaluated on seven axes, scored 1–5. Higher is better. Click any score on a lecture page to read the full justification.

Axis What it measures 1 means 5 means
Historical Accuracy Are facts, dates, events, and causal claims correct per mainstream historiography? Major errors Solid throughout
Argumentative Rigor Is reasoning logically sound? Are conclusions supported by the evidence presented? Fallacious Rigorous
Framing & Selectivity Is the evidence cherry-picked? What's included versus excluded? Highly selective Balanced
Perspective Diversity Are competing interpretations acknowledged? Counterarguments engaged? Single narrative Multiple perspectives
Normative Loading How much moral judgment is embedded? Is the audience told what to think? Heavily prescriptive Descriptive / analytical
Determinism vs. Contingency Is history presented as inevitable, or does the lecture acknowledge accident and agency? Fully deterministic Balanced contingency
Civilizational Framing How are different civilizations characterized? Consistent hero/villain casting? Strongly biased Even-handed

Epistemic Asymmetries We Try to Name

An audit project has defaults. Ours are these: we weight documented events over circumstantial ones, mainstream reporting over adversarial reporting, and confirmed facts over alleged ones. These are reasonable defaults. They are also defaults with known failure modes, and an honest audit should name them instead of pretending they don't exist.

The October Surprise trap

In 1980, allegations that Reagan campaign figures negotiated with the Islamic Republic to delay the US embassy hostage release until after the November election were dismissed as conspiracy theory by most of the press. The allegations were partially corroborated across the 1990s — first by Gary Sick's 1991 book, then by a 1992 Senate investigation that found “credible evidence” of back‑channel contact, and again in March 2023 when Ben Barnes told the New York Times on the record that he had accompanied former Texas Governor John Connally on a 1980 Middle East trip delivering exactly that message. The category of the allegation shifted from conspiracy theory to plausible, partially supported, incompletely documented.

The lesson for an audit project: a posture that reflexively codes unverified → wrong will systematically undercount exactly the kind of event that takes decades to settle. For a project auditing a speaker whose recurring theme is cover‑ups and hidden strategic logic, that asymmetry is not neutral — it silently biases the audit against the speaker's category of claim while taking the official rebuttal at face value. A reader who senses that tilt will lose trust in the audit, not in the speaker.

What we do about it

We do not validate rumors. We also do not silently dismiss them. The schema and site carry four specific affordances to keep the middle ground honest:

The limits of the limits

These affordances still privilege open‑source evidence. We cannot, from public reporting, prove or disprove a genuine cover‑up; the best we can do is refuse to pretend an unresolvable question is resolved. If decades from now a declassification moves a current Open Question into confirmed or disconfirmed, the audit trail should show that the question was named honestly while it was open. That is the standard we hold ourselves to.

Limitations & Caveats

Reproducibility

Every claim in every report is traceable to a timestamp in a public YouTube video. The analysis schema is published — anyone can feed the same transcripts and rubric to an LLM and compare results.

Fair Use Notice

This site constitutes fair use under Section 107 of the U.S. Copyright Act (17 U.S.C. § 107) and analogous provisions in other jurisdictions. All original lecture content remains the property of its creator(s).

The four statutory factors:

  1. Purpose and character of the use. This project is transformative — it subjects publicly available lectures to systematic critical analysis, commentary, and research. No original content is reproduced for entertainment or as a substitute for the source material.
  2. Nature of the copyrighted work. The source lectures are published on YouTube, a public platform, and address matters of public interest (geopolitics, history, current affairs).
  3. Amount and substantiality. Only brief, timestamped quotations are reproduced to support specific analytical points. Individual video frames are used as contextual reference. Neither substitutes for viewing the original lectures.
  4. Effect on the market. This analysis does not compete with the original lectures and is likely to drive additional viewership to the source channel. No content is monetized.

All quoted material and screencaps are used solely for purposes of criticism, commentary, and scholarly analysis. Each item links or refers to the original public source.