Back to all projects
Ana: AI News Aggregator hero image
"What if a trustworthy AI could help you see media bias and understand the full spectrum of news perspectives?"

Ana: AI News Aggregator

Empowering users with balanced news perspectives and highlighting media bias

Role
UX Researcher & Designer
Timeline
12 weeks (Spring 2025)
Team
2 UX Designers, 2 UX Researchers
Chapter 01

Overview

In a media environment shaped by disinformation and polarization, Ana redefines news engagement through explainable AI and media literacy.

Final Solution

Ana: An AI-powered news aggregator that detects bias, educates users, and breaks information bubbles through transparent analysis and interactive exploration.

Article Bias Analysis

Comprehensive bias analysis workflow for any article or link

Start on Homepage

User enters any article link or text to be analyzed for bias

Submit for Analysis

Click the arrow to send request to Ana's analysis engine

Comprehensive Bias Report

Receive detailed analysis with main takeaways and insights

Political Bias Analysis

View bias highlighted with color coding throughout the content

Factual Reporting & Source Credibility

Assess accuracy and reliability of the source

Deeper Insights & Sharing

Explore "The Bigger Picture" and "About the Source" sections, share reports with friends

News For You Page

Personalized news discovery with bias transparency

Start with Empty State

Clean interface with search bar for topic discovery

Search or Browse Topics

Enter any topic or choose from personalized recommendations

Preference-Based Results

See articles curated based on onboarding preferences

Browse by Bias Perspective

Choose to explore articles with specific political biases

Detailed Bias Reports

Access comprehensive bias analysis for any article of interest

Introduction

Our team developed Ana, an AI-powered news aggregator that detects and flags bias in news articles. Biased sources often dominate the information landscape due to speed and reach, leaving users to manually discern bias—a cognitively demanding task, especially for younger audiences. Ana addresses this through automated, explainable bias detection, educational tools, and interactive features that empower users to access balanced information and break out of information bubbles.

The Problem

Gen Z users (18-28) are trapped in algorithmic information bubbles. Biased news dominates due to speed and reach, while users struggle with manually discerning subtle framing, selective facts, and linguistic patterns that shape information perception.

Our Approach

Competitive Analysis

We analyzed Ground News, AllSides, and TIMINO. While competitors provide varied perspectives, Ana differentiates through micro-lessons in media literacy, user-reporting mechanisms, and a bias slider for exploring the ideological spectrum—embedded in a Gen Z-optimized mobile interface.

Literature Review

Synthesizing 63 studies revealed that media bias is often subtle and context-dependent, with current AI systems still in early development. We identified 17 distinct bias forms—from spin and ideology to gatekeeping bias—which shaped Ana's bias taxonomy and training examples.

Psychological frameworks for trustworthy AI emphasize transparency, context-awareness, and accountability, influencing Ana's visual explanations, bias-source links, and user-controlled feedback loops.

Design Principles

Three core principles—transparency, education, and empowerment—guide Ana from onboarding tutorials explaining bias detection to bias sliders inviting exploration of contrasting narratives. Ana is both a news filter and literacy tool designed to meet users' cognitive, emotional, and social needs.

Transparency

Clear explanations of how bias is detected and analyzed

Education

Embedded media literacy tools and micro-lessons

Empowerment

Interactive features for exploring diverse perspectives

Goals

Research validated user concerns about mainstream and AI-generated news, revealing deep skepticism and misinformation fears. Our refined goals focused on:

Reading Habits & Preferences

Understand user news consumption patterns and topic preferences, confirming interest in political, social, and environmental news.

AI Prompt Development

Refined scope to iterate on focused prompts for one model (ChatGPT) based on research cautioning against uncontrolled outputs.

User vs. AI Bias Perception

Revealed users and AI often differ in bias interpretation, particularly with subtle framing, informing explainable bias visualizations.

Media Literacy Enhancement

Added micro-lessons and onboarding tutorials to enhance literacy without overwhelming users due to strong demand for contextual learning.

By combining evidence-based design with human-centered AI, Ana redefines how users engage with the news. It is not just a content delivery system but a cognitive aid built to detect bias, illuminate nuance, and cultivate critical thinking in a media landscape that urgently demands it.

Chapter 02

Research Methods

A multi-stage, mixed-methods research approach grounding every design decision in evidence.

Our multi-stage, mixed-methods approach combined qualitative exploration with quantitative validation, followed by systematic AI development and iterative user testing.

Competitive Analysis

Analysis of Ground News, AllSides, and TIMINO revealed opportunities for differentiation through AI integration and educational features:

Ground News

Offers tools to track media bias and compare coverage across sources. Features a bias checker, bias distribution visuals, and AI-generated summaries. Available via desktop, mobile app, website, and Chrome extension.

Strengths: Strong multi-platform presence, visual bias distribution

Gaps: Limited educational content

AllSides

Emphasizes presenting diverse political and social perspectives. Includes a media bias vote and human-written articles to highlight contrasting viewpoints. Available on desktop and mobile.

Strengths: Community-driven bias assessment, diverse perspectives

Gaps: Lacks AI summaries and in-depth bias comparison tools

TIMINO

Provides bias insights through a checker and comparison features. Relies on human-written content but does not support AI summaries or article reporting.

Strengths: Focused bias detection tools

Gaps: No AI summaries, limited interactive features

Key Opportunity Identified:

Ana integrates micro-lessons in media literacy, user-reporting mechanisms for misinformation, and a bias slider for exploring articles across the ideological spectrum—features not comprehensively integrated in any existing platform.

Literature Review Findings

Key Human Tendencies in Information Consumption

Naive Realism

Leads individuals to view their interpretations as objective truths, causing them to dismiss opposing perspectives as biased or misinformed. This hinders balanced news consumption.

Confirmation Bias

Drives people to seek information that aligns with their existing beliefs. In digital spaces, this is amplified by algorithm-driven "information bubbles" which restrict exposure to diverse viewpoints.

Media Bias Complexity

We identified 17 types of media bias—spin, ideology, omission, placement, labeling, source selection—detectable by experts but largely undetectable by current AI systems in early development stages.

This gap led us to explore prompt engineering and trust-building between humans and AI content.

Building Trustworthy AI

Research on trustworthy AI identified three critical components for building user trust:

The Trustor (User)

User's background—age, education, job, culture, and experience with AI—impacts their relationship with AI systems.

The Trustee (AI System)

An AI system that is accountable, focuses on privacy protection with proper safeguards, and is transparent increases user trust.

Interactive Context

The context of AI impacts perceived trust based on how well it is viewed by media and society.

User Research

User Interviews with Card Sorting

Nine semi-structured interviews explored political views, bias perceptions, and news habits. Participants completed card sorting for concepts like "source credibility" and "political bias," informing Ana's information architecture. Results organized topics into: Core Evaluation, Contextual Analysis, and Explore More.

Qualtrics Survey (n=43)

Comprehensive survey analyzed using R revealed key insights from US residents aged 20-34:

Demographics & Usage
  • • Total participants: 43 (US residents)
  • • Age range: 20-34 (avg: 26.7)
  • • Political leaning: 68% left-leaning
  • • AI news aggregator usage: 82% never used
Key Insights
  • Market education potential: High unfamiliarity with AI news tools
  • Full story priority: Users want depth and multiple perspectives
  • Political perspective impact: System must account for users' own biases
  • Generalization limits: Young, politically homogeneous sample

Limitation: The sample's left-leaning political composition (68%) and modest size (43 responses) may have impacted perceptions, with participants potentially having difficulty recognizing bias aligned with their own views.

Data Analysis

Affinity Diagrams

Each team member coded interviews using inductive coding method, with findings aggregated into a comprehensive affinity diagram in FigJam to identify key themes.

User Personas

Insights distilled into three distinct user personas: "Laura Left" and "Alex Neutral," capturing diverse goals, behaviors, and pain points.

Laura Left - User Persona
Alex Neutral - User Persona
User Journey Map

Visualized typical user path from discovering Ana to building trust, highlighting "Evaluation & Trust Building" as the most critical stage.

User Journey Map - Ana's Complete User Experience

The User's Dilemma

Interviews revealed engaged yet weary news consumers describing misinformation as a "dangerous virus." Users felt overwhelmed by "misinformation and exaggeration" requiring mental health breaks from news.

This skepticism extends to both content and platforms. Users perform laborious verification rituals: cross-referencing sources, checking comments for dissent, seeking journalistic distance. AI enters as both potential "Socratic method machine" and feared source of additional bias.

Survey Insights: Bias Detection Challenges

Our headline analysis revealed significant "blind spots" in how users perceive bias, validating Ana's core premise:

Subtle Framing

46.2% misidentified a right-center headline ("Trump needs unity...") as "Least Biased," showing users miss bias embedded in seemingly neutral language.

Subject vs. Perspective

66.7% misidentified a left-leaning headline critical of GOP as right-leaning, confusing the article's subject with its critical perspective.

Overt Bias

51.3% accurately identified headlines with overt, emotionally charged language, showing better recognition of obvious bias.

Creating Custom Research-Based GPT for Ana

We developed a rule-based analytical framework ensuring Ana's analysis is transparent and defensible, grounded in information science and media analysis rather than arbitrary "black box" outputs.

Source Credibility Assessment

We adapted the "5 Ws" method, a standard framework used by academic institutions like the University of Washington Libraries to evaluate source reliability.

  • • Author credentials assessment
  • • Publisher reputation evaluation
  • • Content purpose analysis
  • • Sourcing practices review
Political Bias Scoring

We operationalized the methodology of Media Bias/Fact Check (MBFC), adopting their weighted scoring system across multiple dimensions.

  • • Economic systems evaluation
  • • Social values assessment
  • • Editorial bias patterns
  • • Seven-point spectrum placement

Bias Check Validation & User-AI Perception Gap

To validate Ana's bias detection accuracy without introducing system bias, we examined the user-AI perception gap through focused testing.

Media Bias Detection Sprint

In-class "Media Bias Detection Sprint" had classmates manually analyze headlines for bias types (spin, statement bias) and compare findings to Ana's AI output.

Human Analysis Patterns

Participants were often swayed by emotional language (e.g., the word "bloodbath"), focusing on surface-level emotional triggers rather than systematic bias patterns.

AI Analysis Strengths

The AI could more systematically identify and categorize bias types, providing consistent methodology across different content samples.

Outcome: This experiment reinforced the necessity of a methodical, rule-based tool to augment human intuition and successfully validated Goal 4 (Compare user vs. AI perceptions of bias).

Chapter 03

Iterative Design & Testing

From concept to prototype: How user feedback and peer evaluation shaped Ana's evolution through structured design iterations.

Our design process moved from low-fidelity concepts to refined mid-fidelity prototype through structured, iterative feedback and testing cycles.

Early Stages & User Flow Mapping

We created a user flow chart of various tasks and functionality to define Ana's MVP, grounding our decision-making and identifying the most important aspects to design and test.

User Flow Chart - Ana's Core Functionality

The most important flows identified included:

1
Article Bias Analysis
2
News For You
3
Account Creation
4
Discussions
5
Methodology

Lo-Fi Prototyping

Ana's primary function is providing article political bias information. We kicked off design with collaborative sketches exploring screen concepts. Figma-proficient teammates then created wireframes, mapping basic user flow and information architecture.

Lo-Fi Home Screen

Lo-Fi Home Screen

Initial Layout Concept

Lo-Fi Bias Analysis

Lo-Fi Bias Analysis

Core Feature Wireframe

Lo-Fi News For You

Lo-Fi News For You

Article Discovery Flow

Peer Evaluation: RFP Exchange with "Her Team"

We conducted an RFP exchange with "Her Team" for expert cognitive walkthrough of our lo-fi prototype. Their evaluation provided external validation and critique that informed design changes.

Strengths Identified

  • Core concept was strong and compelling
  • Onboarding flow was "smooth and intuitive"
  • Successfully provided detailed bias information

Critical Weaknesses

  • Inconsistent design and unclear terminology
  • Non-functional navigation elements ("Core Evaluation")
  • Meaningless visual indicators without context
  • "Home" and "News for You" pages were conflated

Impact: This peer feedback was instrumental in validating our decision to pivot to a mid-fidelity design using a standardized iOS component library, which would resolve the inconsistencies in navigation and visual language that were undermining the user experience.

Design Pivot to Mid-Fi

A major hurdle was debating micro design decisions easily solved with a component library. We selected iOS components for their mobile popularity, developing a mid-fidelity Figma prototype that resolved lo-fi inconsistencies.

Login Page

Login Page

User Authentication

Home Page

Home Page

Main Dashboard

Bias Analysis Report

Bias Analysis Report

Detailed Analysis

Methodology Page

Methodology Page

AI Explanation

News For You - Browse

News For You - Browse

Article Discovery

News For You - Filters

News For You - Filters

Bias Filtering

Discussions Page

Discussions

Community Features

Methodology Keywords

Methodology Details

Keyword Analysis

Early Branding

Though low priority, we used color in the mid-fi prototype for hierarchy and to prototype the final design aesthetic.

Trustworthy White

Heavy usage of white for trustworthy look/feel, giving space for political expression

Yellow Highlighter

Represents the act of highlighting key information

Purple Interactables

Anomaly color compared to iOS blue to implicate political favorability

Bias Slider Scale: From Research to Design

The bias slider scale emerged from our literature review of Media Bias Fact Check methodology, where political bias was represented on a linear scale. Through iterative design exploration, we evolved from conceptual sketches to a refined interactive component.

The design process involved exploring various metaphors—from audio tuners and compass designs to speedometer interfaces—before settling on a clean slider that users could intuitively understand and interact with.

Design Evolution

Figma Inspiration for Slider

Figma Inspiration

Visual representation idea source

Original Bias Scale from Literature

Literature Foundation

Media Bias Fact Check methodology scale

X-Y Axis Slider Sketch

X-Y Axis Exploration

Two-dimensional bias mapping concept

Speedometer Design Sketch

Speedometer Concept

Circular bias indicator exploration

Final Bias Slider Component

Final Slider Component

Clean, intuitive bias scale interface - primary design solution

Design Insight: The evolution from complex metaphors to a simple slider demonstrated that users preferred familiar interface patterns over novel visualizations when dealing with sensitive political content. The final design prioritized clarity and trustworthiness over visual novelty.

RITE Method: Rapid Iterative Testing & Evaluation

We tested the mid-fidelity prototype using RITE (Rapid Iterative Testing and Evaluation) method—two usability testing rounds with targeted design changes after each session.

This agile approach quickly identified and fixed usability issues, ensuring the final prototype directly responded to user needs while working under pressure.

Design Strengths Confirmed

  • Aesthetic and Tone:

    Users responded positively to the soft, authoritative color scheme, which established credibility

  • Clarity:

    Users found the general layout and typography intuitive

Areas for Improvement

  • Information Hierarchy:

    Users needed to see "Main Takeaways" immediately, but initial design buried this content below the scroll

  • Interactive Clarity:

    Key interactive elements lacked strong visual cues to show they were clickable

  • Screen Real Estate:

    Mobile layout used space inefficiently with oversized elements

  • Sharing Options:

    Users wanted to share specific sections, not just the entire article

Iterative Design Changes

This feedback directly informed our design pivot. The redesign incorporated the following improvements:

  • Elevated "Main Takeaways" for immediate visibility
  • Collapsed summary to prioritize key points
  • Added visual borders to clarify button interactivity
  • Implemented prominent flag and share icons

This iterative loop demonstrates how we refined Ana from a strong concept into a functional, intuitive prototype.

Chapter 04

Reflection & Next Steps

Looking back on our journey and forward to Ana's potential: What we learned, what we'd do differently, and where we go from here.

Ana represents a promising step toward news tools built on trust and media literacy, combining AI's analytical power with user-centered design to empower critical news understanding.

Project Reflection

What Went Well

  • User-Centered Research: Our mixed-methods approach successfully uncovered deep insights about user perception gaps and bias blind spots
  • Iterative Design Process: The RITE method and RFP feedback led to significant usability improvements
  • AI Framework Development: Successfully created a rule-based, transparent bias detection system
  • Educational Integration: Seamlessly combined bias detection with media literacy learning

What Could Be Improved

  • Demographic Diversity: Our sample was heavily left-leaning (68%) and limited in age range, potentially skewing insights
  • AI Explainability: Current system struggles with nuanced explanations for complex bias cases
  • Long-term Validation: Limited testing period for measuring sustained behavior change
  • Technical Scalability: Resource-intensive analysis may create latency issues at scale

Project Limitations

While Ana shows promise, several limitations emerged during our research and development process that future iterations should address:

Sample Bias

Our participant pool's political homogeneity and limited age range (20-34) may not represent broader user needs, particularly among conservative users or older demographics who consume news differently.

AI Complexity

The subject/perspective confusion we observed highlights AI's current limitations in understanding nuanced bias, particularly when article subjects differ from editorial perspectives.

Scale Validation

While prototype testing showed promise, real-world deployment with thousands of users and diverse news sources would likely reveal additional challenges and edge cases.

Next Steps

Looking forward, Ana's next steps involve expanding capabilities and validating effectiveness with broader audiences. Future work should focus on three key areas:

Enhancing AI Explainability

Develop AI to provide more nuanced explanations for bias ratings, particularly for complex cases like subject/perspective confusion. This includes:

  • Advanced prompt engineering for contextual analysis
  • Multi-layered bias detection beyond left/right paradigms
  • Visual explanations showing AI reasoning process

Broader Demographic Testing

Conduct studies with more politically and demographically diverse users to ensure Ana's effectiveness for all users. This expansion should include:

  • Conservative and moderate political perspectives
  • Broader age ranges (35-65+) and cultural backgrounds
  • International users with different media landscapes

Building Interactivity

Implement user-requested flagging and feedback mechanisms to create a "human-in-the-loop" system that refines AI assessments over time, building trust and engagement:

  • User flagging system for bias detection accuracy
  • Community-driven bias validation and appeals process
  • Continuous learning algorithms that adapt to user feedback

Final Reflection

Ana is more than a bias detection tool—it's an educational platform helping users develop critical thinking for all media consumption. In today's fragmented information landscape, we need technologies that help us think better, not just tell us what to think.

This project demonstrates that combining rigorous user research with thoughtful AI development can create solutions addressing real user needs while building essential media literacy skills.

— The Fantastic Four Team

Like my work? Explore more!

Discover more UX design projects that showcase my approach to solving complex problems through research-driven design.

Designed in Figma, developed in VSCode

Last Updated Aug 2025