Flowstate AI company logo - AI Video Intelligence Platform

How Attention Economics Quantified Editorial Bias Using AI

Case Study

5min Read

Mar 6, 2026

Attention Economics used Flowstate to analyze 10+ hours of live news daily, quantify bias, and produce auditable reports with timestamps.

iFit Case Study: Scaling video metadata tagging and content discovery with Flowstate AI.

Attention Economics partnered with Flowstate to monitor and analyze a live Eastern European regional news channel, transforming subjective perceptions of political bias into structured, defensible data. Over a multi-week engagement, Flowstate enabled automated ingestion, transcription, contextual analysis, and query-based reporting across daily news programming, reducing manual monitoring effort while increasing analytical depth and traceability.

About Attention Economics

Attention Economics is a media advisory and analytics firm specializing in editorial strategy, governance assessments, and broadcast intelligence.

When engaged to evaluate the editorial positioning of a regional Eastern European news channel, the central question was not whether stakeholders felt bias. It was whether bias could be measured consistently, explained clearly, and verified directly in the source footage.

The team needed a repeatable way to quantify:

  • Government vs. opposition representation

  • How protests and contentious events were framed

  • Topic distribution across the news cycle

  • Sentiment patterns over time

  • Evidence that could withstand executive or regulatory scrutiny

The Challenge: Turning Live News Into Defensible Data

1) Monitoring high-volume daily programming

The channel broadcast 10+ hours of daily editorial content, including studio bulletins, live field reporting, political interviews, panel discussions, and breaking news segments. Traditional monitoring required continuous viewing, transcription, and manual logging across shifts.

2) Moving from perception to proof

Anecdotal assessments were not enough. Attention Economics needed:

  • Quantitative evidence

  • A reproducible methodology

  • Aggregation across multiple days

  • Traceable references back to original video segments

3) Language and contextual complexity

Programming was produced in a local Eastern European language and included political nuance, rhetorical framing, and region-specific context. On-screen graphics mixed Latin and Cyrillic scripts. The channel also blended imported syndicated programming with local editorial content, requiring the system to distinguish editorial segments from non-editorial imports while maintaining temporal accuracy.

4) Time-sensitive executive reporting

Attention Economics was required to deliver daily monitoring updates, weekly executive summaries, and a final aggregate report suitable for leadership presentation. The workflow needed to operate continuously, not retroactively.


The Solution: AI-Powered Live Broadcast Monitoring

Attention Economics implemented Flowstate’s Live Watcher workflow to convert broadcast television into structured, searchable intelligence.

1) Live stream ingestion and structured segmentation

Flowstate ingested the channel’s live stream and:

  • Captured programming in structured time segments

  • Preserved full video archives

  • Generated timestamped transcripts

  • Produced contextual summaries per segment

This segment-based processing preserved traceability while scaling across long daily broadcasts.

2) Structured political and event classification

Custom watcher instructions were configured to analyze and label:

  • Political sentiment (pro-government, anti-government, neutral)

  • Named political actors and affiliations

  • Protest coverage frequency

  • Airtime allocation between government and opposition voices

  • Topic clustering (corruption, protests, energy policy, elections, and more)

Each segment produced structured outputs designed for aggregation and reporting.

3) Natural language query over broadcast data

Rather than manually scanning hours of content, analysts could query the system directly:

  • How many protest events were covered over the past two weeks?

  • How often did government officials appear compared to opposition representatives?

  • What percentage of segments leaned anti-government?

  • Which topics were covered most frequently, and with what sentiment?

Flowstate returned results with timestamp references so analysts could quickly verify findings against the archived footage.

4) Human-in-the-loop validation

Given the sensitivity of political analysis, Attention Economics layered validation on top of AI outputs:

  • Native language experts reviewed selected segments

  • High-impact or contentious classifications were cross-checked

  • Instructions were refined iteratively to improve consistency

This hybrid workflow combined automation speed with editorial rigor suited for governance reporting.


Results

Over the course of the engagement, Attention Economics transformed live broadcast into structured political intelligence.

>Broadcast quantified at scale

  • 10+ hours of daily programming analyzed

  • Hundreds of broadcast segments processed

  • Multi-week archive indexed and searchable

  • Political actors identified and tabulated automatically

>From manual review to structured reporting

Before Flowstate:

  • Monitoring required full-day human viewing

  • Bias assessments were subjective and hard to reproduce

  • Protest coverage required manual logging

  • Cross-day comparisons were slow and labor-intensive

After Flowstate:

  • Daily summaries were auto-generated

  • Government vs. opposition appearances were tabulated

  • Protest frequency was quantified across weeks

  • Weekly executive reports were assembled from structured outputs

>Operational efficiency gains

  • 70 to 80 percent reduction in manual monitoring effort

  • Reports generated in hours instead of days

  • Faster iteration on stakeholder questions

  • Immediate access to archived segments for validation

Most importantly, Attention Economics could deliver findings supported by:

  • Structured datasets

  • Timestamp references

  • Archived video evidence

  • A reproducible methodology

Example Analytical Outputs Enabled

  • Total protest coverage across multiple weeks

  • Minutes of airtime allocated to government vs. opposition voices

  • Frequency of anti-government framing in headlines

  • Distribution of political topics by sentiment

  • Comparison of live segments vs. recorded bulletins

All generated without building custom dashboards, using structured extraction plus query workflows.


Workflow Overview

  1. Live capture: Flowstate ingests broadcast streams in structured intervals.

  2. Automated transcription and context extraction: Speech is transcribed and enriched with contextual analysis.

  3. Political and event classification: Segments are labeled for sentiment, actor presence, and event type.

  4. Query and aggregate reporting: Analysts query the system across days or weeks to produce rollups and executive summaries.

  5. Human validation: Reviewers validate sensitive findings and refine watcher instructions.

Strategic Impact

This engagement demonstrated that:

  • Live news can be converted into structured, searchable data.

  • Editorial bias analysis can be quantified rather than debated.

  • Broadcast oversight can scale without increasing headcount.

  • Executive reporting can be supported by verifiable, timestamped video evidence.

What began as a channel-level assessment evolved into a blueprint for scalable media monitoring, regulatory analysis, and editorial governance.


How Flowstate Supports Live News Monitoring and Governance

Flowstate turns live news into structured, searchable intelligence that teams can use for monitoring, reporting, and oversight.

Instead of relying on manual viewing and subjective notes, Flowstate continuously ingests live streams and converts broadcasts into timestamped transcripts, segment-level summaries, and structured signals. Video is analyzed across speech, visuals, on-screen text, and temporal context, creating an auditable record of what aired, when it aired, and how it was framed.

Flowstate enables teams to:

  • Monitor 10+ hours of daily programming without continuous manual viewing

  • Quantify government vs. opposition representation and airtime allocation

  • Track coverage of protests and contentious events over days or weeks

  • Measure framing and sentiment patterns at the segment level

  • Query the archive using natural language and retrieve exact timestamps for verification

  • Produce executive-ready summaries and rollups on a recurring cadence

Because outputs are time-coded and traceable back to the source footage, findings can be validated quickly by internal reviewers or third parties. This makes the workflow suitable for governance, compliance, and advisory reporting where conclusions must be defensible, reproducible, and supported by evidence.

Analysts can automate baseline monitoring and aggregation, then apply human review where it matters most, such as sensitive classifications, edge cases, and final interpretations.

Looking Ahead: Broadcast as Queryable Evidence

Live news is one of the least structured and most high impact content environments. As scrutiny increases around bias, misinformation, and trust, teams need a way to turn what aired into evidence, not opinion.

Flowstate supports that by converting live broadcasts into a time-coded, searchable record. It captures transcripts, segment summaries, and structured signals that can be queried across days or weeks, then verified directly against the underlying footage.

About the Author

Sahil Shah

Founder & CEO, Flowstate

Sahil Shah is the Founder and CEO of Flowstate. Prior to founding the company, he spent nearly a decade at Waymo and Apple building large-scale video AI and computer vision systems, and has over 15 years of experience bringing frontier video technologies from research into production environments.

About the Author

Sahil Shah

Founder & CEO, Flowstate

Sahil Shah is the Founder and CEO of Flowstate. Prior to founding the company, he spent nearly a decade at Waymo and Apple building large-scale video AI and computer vision systems, and has over 15 years of experience bringing frontier video technologies from research into production environments.

Testimonials

Read Similar Case Studies

Case Study

iFit Case Study: Scaling video metadata tagging and content discovery with Flowstate AI.

How iFit scaled video metadata tagging with Flowstate

Case Study

South Park Commons Case Study: Automating social media clip generation from long-form video.

How South Park Commons Scaled Social Content from Long-Form Video

Experience FlowState in action

Explore Enterprise-Grade Video Intelligence Built for Scale and Security.

Experience FlowState in action

Explore Enterprise-Grade Video Intelligence Built for Scale and Security.

Experience FlowState in action

Explore Enterprise-Grade Video Intelligence Built for Scale and Security.