Flipside AI Workflow Breakdown

What happens when you run a Flipside AI Workflow? Let's understand the data used, architecture followed, sequences implemented, pitfalls avoided, and analysis performed under the hood.

CEFI Bridge Intelligence - Technical Data Flow

Business Requirement

Track institutional capital movement across blockchains to understand which centralized entities (funds, known institutional traders, exchanges, custodians, etc) are moving liquidity between chains, through which bridge platforms, and in what volumes.

The Data Architecture

Primary Dataset: crosschain.core.dim_labels

Primary Dataset: crosschain.defi.ez_bridge_activity

The Sequence

Step 1: Build the Entity Filter (Subquery)

This creates a focused filter of institutional operational addresses. We're not looking for every Coinbase user - we're looking for Coinbase's own infrastructure addresses that move institutional liquidity.

Step 2: Join Against Bridge Activity (Main Query)

Critical join logic: We join on BOTH source and destination addresses because institutional activity can occur in either direction:

The (src.address IS NOT NULL OR dest.address IS NOT NULL) ensures we capture transactions where institutions are on either side.

Step 3: Fraud Pattern Exclusion

Filter: NOT (source_project_name IS NULL AND destination_address_type = 'hot_wallet')

This removes suspicious patterns where unknown projects are sending directly to hot wallets - likely scam tokens or wash trading attempts.

Step 4: Aggregation Structure

Group by: month, bridge_platform, source_chain, destination_chain, source_project_name, source_address_type, destination_project_name, destination_address_type, token_symbol

This granular grouping preserves analytical flexibility - you can roll up to entity-level summaries or drill down to specific chain corridors and token movements.

Computed metrics per group:

Why This Order Matters

  1. Entity identification first: You must know WHO before you can filter WHAT. Building the labeled address filter upfront prevents scanning billions of irrelevant bridge transactions.
  2. Dual-sided join: Institutions appear on both sides of bridge transactions. A single-sided join would miss half the activity.
  3. Time-scoped retrieval: Starting from 2024-01-01 balances data freshness with query performance. Bridge activity compounds quickly - earlier dates add diminishing analytical value.
  4. Verified tokens only: Filters out spam tokens and scam projects that would otherwise inflate volume metrics.
  5. Monthly granularity: Strikes balance between temporal precision and data volume. Daily would create sparse results for many entities; quarterly would obscure short-term shifts.

Data Pitfalls Avoided

Result Structure

The query returns a denormalized fact table where each row represents:

This structure supports immediate analysis without additional joins, enabling rapid calculation of entity rankings, chain corridor analysis, bridge platform market share, and temporal trend detection.