What is Data Aggregation?

Definition

Data aggregation is the process of collecting and combining data from multiple disparate sources into a unified dataset, enabling comprehensive analysis and more complete records.

Key Takeaways

  • Combines data from multiple independent sources into unified records
  • No single provider has complete coverage, making aggregation essential
  • Requires entity resolution, normalization, and conflict resolution
  • Confidence scoring helps determine which aggregated values to trust
  • Common patterns include merge, append, deduplicate, and normalize-then-combine
  • Aggregation collects data into a unified view; integration connects systems for real-time flow
  • Real-time aggregation runs inline (1-2 seconds); batch aggregation runs scheduled and is cheaper
  • Cleanlist data: B2B teams that mix both modes see 31% higher first-touch reply rates

Data aggregation is the process of collecting, combining, and summarizing data from multiple sources into a single, unified dataset. The data aggregation meaning is consistent across industries — transform fragmented data into a comprehensive view. In B2B sales and marketing, this means pulling contact and company data from CRMs, enrichment providers, social networks, and public records, then merging them into one complete record per person or company. There are three main types: temporal aggregation (across time periods), spatial aggregation (by geography), and record-level aggregation (merging attributes from multiple sources into one entity). Based on Cleanlist's analysis of 2.1 million B2B records processed through waterfall enrichment, multi-source aggregation improves data coverage by 47% compared to single-provider approaches, with email accuracy reaching 98% when three or more sources are combined.

What is data aggregation?

Data aggregation is the process of collecting, combining, and summarizing data from multiple sources into a single, unified dataset. In B2B contexts, this means pulling contact and company information from providers like LinkedIn, ZoomInfo, and public records, then merging them into one clean record per person or company. The data aggregation definition applies across industries and contexts, but in B2B sales and marketing, it specifically refers to gathering information about companies and contacts from multiple independent sources and merging them into comprehensive records. Rather than relying on one data provider or one internal system, aggregation pulls relevant data points from CRMs, marketing tools, web scraping sources, public filings, social networks, data vendors, and proprietary databases, then merges them into unified records.

What are common data aggregation examples?

To make the concept concrete, here are four common data aggregation scenarios in B2B operations: (1) Prospect profile aggregation — A sales team building prospect profiles aggregates LinkedIn profile data with CRM activity history, marketing automation engagement scores, and third-party enrichment data from vendors like ZoomInfo or Cognism. The result is a single record that captures firmographics, contact details, behavioral signals, and technographic attributes in one place. (2) Market research aggregation — A product team aggregates data from G2 reviews, Gartner reports, customer survey responses, and competitive intelligence tools to build a comprehensive view of market positioning and feature gaps. (3) Pipeline reporting aggregation — A revenue operations team aggregates data from Salesforce (deal stages), Outreach (email sequences), Gong (call recordings), and Stripe (revenue) to build an accurate pipeline report that no single system could produce alone. (4) Multi-provider enrichment aggregation — An enrichment platform like Cleanlist aggregates contact and company data from 15+ data providers, selecting the best value for each field based on confidence scoring and recency. This last example is the most common form of data aggregation in B2B data enrichment.

What are the main data aggregation methods?

Teams use several approaches depending on the data volume, source variety, and accuracy requirements. Manual aggregation involves exporting data from multiple systems into spreadsheets and merging them using VLOOKUP, INDEX/MATCH, or similar formulas — this works for small one-time projects but does not scale. ETL pipeline aggregation uses tools like dbt, Fivetran, or Airbyte to extract data from multiple sources, transform it into a consistent schema, and load it into a data warehouse where it can be queried holistically. API-based aggregation queries multiple data sources programmatically in real time or near-real-time, combining responses into unified records before delivering them to downstream systems. Reverse ETL aggregation pushes already-aggregated data from a warehouse back into operational tools like CRMs and marketing platforms. For most B2B teams, the practical choice is between manual spreadsheet work (free but slow and error-prone) and automated platforms that handle aggregation as part of a broader enrichment or data management workflow.

Top data aggregation tools compared

ToolBest ForApproachPricing
[Cleanlist](/product/waterfall-enrichment)B2B contact and company aggregationWaterfall enrichment across 15+ providers with confidence scoringCredit-based, pay per record
FivetranData warehouse ingestion300+ pre-built ETL connectorsUsage-based per monthly active row
AirbyteOpen-source data pipelinesSelf-hosted or cloud ETL with 350+ connectorsFree (open-source) or cloud pricing
dbtData transformationSQL-based transform layer in the warehouseFree (Core) or Team/Enterprise
HightouchReverse ETLPush aggregated warehouse data to operational toolsPer-destination pricing
SegmentCustomer data platformReal-time event tracking and routingVolume-based pricing
Apache SparkLarge-scale batch processingDistributed computing for terabyte-scale aggregationOpen-source, infrastructure costs

For B2B sales and marketing teams, the most practical choice is a platform like Cleanlist that handles multi-provider aggregation, normalization, and conflict resolution automatically — eliminating the need to build and maintain custom ETL pipelines. Engineering teams building internal data infrastructure typically combine Fivetran or Airbyte (extraction), dbt (transformation and aggregation logic), and Hightouch (reverse ETL delivery) into a modern data stack.

The rationale for data aggregation is coverage. No single data source has complete information about every company and contact in your addressable market. Provider A might have strong coverage of US-based tech companies but limited data on European manufacturers. Provider B might excel at direct dial phone numbers but lack technographic information. By aggregating data from both, you build a more complete picture than either could provide alone. This principle scales across any number of sources and data types. In B2B sales specifically, aggregation is how teams construct complete prospect profiles from fragmented signals — combining a contact's job title from LinkedIn, their verified email from an enrichment vendor, their company's revenue data from a firmographic database, and their recent content engagement from a marketing automation platform.

The technical challenges of data aggregation are significant. Different sources use different formats, naming conventions, and identifiers. Company names appear in variations — "International Business Machines," "IBM," and "IBM Corporation" must all be recognized as the same entity. Job titles vary wildly — "VP of Marketing," "Vice President, Marketing," and "Marketing VP" represent the same role. Addresses follow different formatting standards across countries. Effective aggregation requires robust entity resolution, data normalization, and conflict resolution rules that determine which source to trust when values disagree. Four common aggregation patterns address these challenges. Merge combines overlapping records into a single golden record by matching on shared identifiers like email or domain. Append adds new fields from a secondary source to existing records without overwriting. Deduplicate identifies and collapses duplicate entries created when the same entity appears across multiple sources. Normalize-then-combine standardizes field formats (date formats, address structures, title conventions) before merging, which reduces downstream conflicts.

It is worth distinguishing data aggregation from data integration, since the two terms are frequently confused. Data aggregation is the process of collecting data from multiple sources and combining it into a unified dataset — typically a batch or periodic operation that produces a consolidated view. Data integration, by contrast, focuses on connecting systems so data flows between them continuously and in real time. Integration ensures your CRM, marketing platform, and data warehouse stay synchronized as records change. Aggregation produces a snapshot — a compiled dataset drawn from many inputs at a point in time. In practice, most B2B data operations use both: integration keeps systems connected, and aggregation builds the comprehensive records that sales and marketing teams work from.

What does data aggregation mean in practice?

The data aggregation meaning becomes concrete when you look at the numbers. Cleanlist's internal benchmarks from processing 2.1 million B2B records show that single-provider enrichment achieves an average 52% field coverage rate across email, phone, title, and company attributes. Adding a second provider raises coverage to 71%. By the time five providers are aggregated through a waterfall, coverage reaches 94% — with email verification rates hitting 98% and direct dial accuracy at 85%. These numbers illustrate why aggregation matters: no single source is sufficient, but the marginal gain from each additional source follows a predictable curve with diminishing returns after 5-7 providers.

Beyond simple merging, intelligent aggregation adds a confidence layer. When three out of four sources agree that a contact's title is "Director of Sales," that value gets a higher confidence score than a title reported by only one source. This confidence-based approach lets downstream systems make better decisions about which data points to trust and display. It also highlights records where sources strongly disagree, flagging them for review. Types of aggregation also vary by dimension: temporal aggregation rolls up data across time periods (quarterly revenue, monthly engagement trends), spatial aggregation groups data by geography (regional pipeline, country-level coverage), and record-level aggregation — the most relevant for B2B — merges attributes from multiple sources into a single contact or company record.

In the modern data stack, data aggregation sits at a critical junction between ETL (extract, transform, load) and reverse ETL workflows. Traditional ETL pipelines extract raw data from operational systems, transform it into a consistent schema, and load it into a data warehouse — aggregation happens during the transform step. Reverse ETL then pushes aggregated, enriched records back into operational tools like CRMs and marketing platforms, closing the loop. For B2B teams, this means prospect data can be aggregated in a warehouse from multiple enrichment providers and then synced back to Salesforce or HubSpot as complete, ready-to-use records.

How does data aggregation work in databases?

In relational databases, data aggregation refers to operations that compute summary statistics across groups of rows. SQL provides built-in aggregate functions — COUNT, SUM, AVG, MIN, and MAX — that collapse multiple rows into a single result. The GROUP BY clause is the primary mechanism for database aggregation: SELECT department, COUNT(*) AS headcount, AVG(salary) AS avg_salary FROM employees GROUP BY department returns one row per department with the employee count and average salary. Window functions extend aggregation by computing values across a set of rows related to the current row without collapsing them: SELECT name, salary, AVG(salary) OVER (PARTITION BY department) AS dept_avg FROM employees returns every row but adds the department average alongside each individual salary. For analytical workloads, OLAP (Online Analytical Processing) cubes provide multidimensional aggregation using operations like roll-up (aggregating from day to month to quarter), drill-down (decomposing from quarter to month to day), slice (filtering one dimension), and dice (filtering multiple dimensions). Dimensional modeling — the star schema and snowflake schema patterns popularized by Ralph Kimball — organizes data for efficient aggregation by separating measurable facts (revenue, quantity, duration) from descriptive dimensions (customer, product, time, geography). In B2B data operations, database aggregation is commonly used for pipeline reporting (aggregating deal values by stage, rep, or quarter), engagement analysis (aggregating email metrics by campaign, segment, or time period), and coverage reporting (aggregating enrichment match rates by provider or data type).

What is the difference between aggregation, integration, and enrichment?

These three terms describe related but distinct processes in the data pipeline, and confusing them leads to miscommunication between teams.

ConceptDefinitionOperationTypical CadenceExample
Data AggregationCollecting and combining data from multiple sources into a unified datasetMerge, combine, summarizeBatch or periodicCombining contact data from LinkedIn, CRM, and enrichment providers into one record
Data IntegrationConnecting systems so data flows between them continuouslySync, replicate, streamReal-time or near-real-timeBidirectional sync between Salesforce and HubSpot
Data EnrichmentEnhancing existing records with additional attributes from external sourcesAppend, enhance, scoreOn-demand or scheduledAdding phone number, revenue, and tech stack to a lead record that only has name and email

When do you use each? Aggregation is the collection step — you use it when you need a comprehensive view compiled from multiple inputs. Integration is the plumbing — you use it to keep systems synchronized as records change. Enrichment is the enhancement — you use it to make incomplete records actionable. In practice, most B2B data operations use all three: integration keeps CRM and marketing platforms in sync, enrichment fills gaps and refreshes stale fields, and aggregation builds the comprehensive records that sales and marketing teams work from.

How do you aggregate B2B data step by step?

Follow these six steps to aggregate data from multiple sources into clean, unified records:

  1. 1.Identify data sources. List every system and provider that holds relevant data: CRM, marketing automation, enrichment providers, web scraping tools, public records, social networks, and spreadsheets. For each source, document what fields it provides, how frequently data is updated, and any API or export limitations.
  1. 1.Map fields across sources. Create a field mapping table that aligns equivalent fields across sources. Provider A's "job_title" maps to Provider B's "position" and your CRM's "Title." Decide on canonical field names and data types that will serve as the output schema.
  1. 1.Normalize formats. Before merging, standardize the raw data: convert phone numbers to E.164 format, normalize job titles to a canonical taxonomy, resolve company name variations ("IBM" vs "International Business Machines Corp"), and ensure consistent date formats and currency conventions.
  1. 1.Resolve entities. Use entity resolution (also called record matching or identity resolution) to determine which records across sources refer to the same person or company. Match on high-confidence identifiers first (email, domain), then fall back to fuzzy matching (name + company similarity scoring using algorithms like Jaro-Winkler distance).
  1. 1.Apply confidence scoring. When multiple sources provide different values for the same field, use confidence-based resolution. Weight sources by their historical accuracy for each field type, prefer more recent data, and apply consensus logic — when three out of four sources agree on a value, that consensus increases confidence. Flag records where sources strongly disagree for manual review.
  1. 1.Validate output. Run quality checks on the aggregated dataset: verify email deliverability, check for remaining duplicates, confirm that required fields are populated, and spot-check a sample of records against original sources. Document the aggregation rules and lineage so the process is repeatable and auditable.

Cleanlist implements data aggregation as a core part of its waterfall enrichment process. When a record is processed, the platform queries multiple data providers and aggregates their responses into a single enriched profile. Normalization rules standardize the output format, conflict resolution logic selects the best value for each field, and confidence scoring indicates the reliability of each data point. This automated aggregation replaces the manual process of querying multiple tools and spreadsheet-merging results that many teams still rely on. Teams can get started with Cleanlist's free tier of 30 credits to see how automated aggregation compares to their current manual workflows.

Real-time vs batch data aggregation: which one do you need?

The two architectures solve different problems and most teams underestimate the trade-off. Batch aggregation runs on a schedule (hourly, daily, weekly) and consolidates a fixed set of records in one pass. It is cheaper, more predictable, and easier to debug — most CRM enrichment refresh jobs run as nightly batches. Real-time aggregation runs synchronously: the moment a record is created or queried, the system fans out requests to every source and merges responses inline. It is more expensive per record but essential when staleness is unacceptable — for example, in form-fill personalization, fraud scoring, or live ABM workflows.

DimensionBatch aggregationReal-time aggregation
LatencyMinutes to hours< 2 seconds typical
Cost per recordLower (bulk discounts)Higher (per-call pricing)
Best forCRM hygiene, list enrichment, reportingForm-fills, lead routing, fraud signals
Failure handlingEasy to retryMust degrade gracefully
Implementation effortLowModerate to high

In Cleanlist's data, B2B teams that mix both — real-time on inbound web leads, batch on the long tail of CRM records — see 31% higher first-touch reply rates than teams using only one mode. The general rule: aggregate in real time wherever the data point will be acted on within minutes; aggregate in batch wherever the data exists for reporting, segmentation, or scheduled outreach.

Data aggregation across industries

Industry context shapes which sources matter and how aggregation rules differ. Four common patterns:

Banking and financial services: Aggregate transaction data from card networks, ACH systems, and core banking platforms to build a unified customer ledger. Reconciliation rules favor the source-of-record (the bank's general ledger) over derived sources. AML and KYC workflows aggregate identity data from credit bureaus, sanctions lists, and government registries, with strict audit trails on every merge decision. The Federal Reserve's 2024 payments study estimates banks aggregate data from 12-30 distinct internal and external systems per customer.

Healthcare: Aggregate patient data from EHRs (Epic, Cerner), lab systems (LabCorp, Quest), imaging archives, and claims databases into a longitudinal patient record. HIPAA forces aggregation rules to track provenance and consent at the field level. Health Information Exchanges (HIEs) are explicitly aggregation networks. Match rates are notoriously hard — patient name + DOB + last-four-SSN matching averages 85-90% accuracy without probabilistic methods.

E-commerce and retail: Aggregate order, payment, loyalty, and web behavior data into a single customer profile that powers personalization. Modern stacks centralize this in a Customer Data Platform (CDP) that aggregates from Shopify or Salesforce Commerce Cloud, Stripe or Adyen, segment events, and email engagement. Identity stitching across devices is the core technical challenge.

B2B sales and marketing (the Cleanlist focus): Aggregate contact and company data from CRMs (Salesforce, HubSpot), enrichment providers (ZoomInfo, Apollo, Lusha), public records (LinkedIn, Crunchbase), and intent signals (G2, Bombora) into a complete prospect record. Match rates are the metric that matters — single-source enrichment caps at 60-70%, while waterfall aggregation across 5+ providers sustains 90%+ across most ICPs. This is the pattern Cleanlist automates end-to-end.

Common data aggregation mistakes (and how to avoid them)

Teams new to multi-source aggregation tend to make four predictable errors. Treating all sources as equally trustworthy — instead, weight sources by historical accuracy per field type (Provider A might be best for titles, Provider B for phones), and prefer recent data over old. Skipping normalization before merge — merging "VP of Sales" and "Vice President, Sales" into two separate records inflates duplicate counts and tanks downstream match rates. Overweighting recency — recency matters for phone numbers and titles but not for company founding year or industry classification, which should prefer authoritative sources. No audit trail — without lineage on every aggregated field, debugging conflicts becomes impossible at scale; every aggregated record should carry the provider attribution and confidence score for each field.

For a comprehensive walkthrough with examples, SQL functions, and conflict resolution frameworks, see the complete data aggregation guide.

Data aggregation is what transforms scattered touchpoints into a complete picture of your prospect. The challenge isn't collecting data — it's merging records from 10+ sources without creating duplicates or conflicts.

VP
Victor Paraschiv
Co-Founder, Cleanlist AI

References & Sources

  1. [1]
  2. [2]
    Customer Data Platform WaveForrester(2025)
  3. [3]
  4. [4]
  5. [5]
    Data-Driven Decision Making in OrganizationsHarvard Business Review(2024)

Frequently Asked Questions

What is the difference between data aggregation and data enrichment?

+

Data aggregation is the process of collecting and combining raw data from multiple sources into a single dataset. Data enrichment is the process of enhancing existing records with additional information. Aggregation is often a step within the enrichment process - to enrich a contact record, you might aggregate data from several providers, then select and append the best values. Think of aggregation as the collection step and enrichment as the enhancement outcome.

How do you resolve conflicts when aggregating B2B data?

+

Conflict resolution typically uses a combination of source reliability rankings, recency weighting, and consensus logic. Sources are ranked by historical accuracy for each data type - one provider might be more reliable for job titles while another is better for revenue data. More recent data generally wins over older data. When multiple sources agree on a value, that consensus increases confidence. The best platforms automate this logic rather than requiring manual decisions.

How many data sources should I aggregate for B2B records?

+

For most B2B use cases, aggregating 3-5 data sources provides the optimal balance of coverage and complexity. Beyond 5 sources, the incremental data improvement diminishes while the normalization and conflict resolution challenges increase. The specific number depends on your data needs - email enrichment may need fewer sources than firmographic enrichment. Cleanlist's waterfall approach queries 10+ providers but handles all aggregation complexity automatically.

What is data aggregation with example?

+

Data aggregation is the process of collecting data from multiple sources and combining it into a single dataset. For example, a B2B sales team might aggregate a prospect's job title from LinkedIn, their verified email from an enrichment vendor, their company's revenue from a firmographic database, and their engagement history from a marketing automation platform. The result is one comprehensive prospect record instead of four fragmented data points across different tools.

What are the types of data aggregation?

+

The main types are temporal aggregation (rolling up data across time periods like monthly or quarterly), spatial aggregation (grouping data by geographic region or location), and record-level aggregation (merging attributes from multiple sources into a single entity record). In B2B contexts, record-level aggregation is the most common — combining contact and company data from CRMs, enrichment providers, and marketing tools into unified profiles.

What is the difference between data aggregation and data integration?

+

Data aggregation collects and combines data from multiple sources into a unified dataset, typically as a batch or periodic operation. Data integration connects systems so data flows between them continuously and in real time. Aggregation produces a consolidated snapshot; integration maintains ongoing synchronization. Most B2B data operations use both — integration keeps CRM and marketing platforms in sync, while aggregation builds the comprehensive prospect records teams work from.

Why is data aggregation important in B2B?

+

No single data source has complete coverage of every company and contact in a B2B addressable market. Aggregation solves this by combining data from multiple providers and systems to build more complete prospect profiles. This improves email deliverability (verified addresses from multiple sources), increases connect rates (accurate phone numbers), and gives sales reps better context before outreach. Cleanlist automates this through waterfall enrichment, querying multiple providers and aggregating responses into a single enriched record.

What tools are used for data aggregation?

+

Data aggregation tools range from general-purpose ETL platforms like Fivetran and Airbyte to specialized B2B data tools. For sales and marketing teams, enrichment platforms like Cleanlist aggregate data from multiple providers automatically through waterfall queries. Data warehouses such as Snowflake and BigQuery serve as central aggregation layers, while reverse ETL tools like Hightouch and Census push aggregated data back into operational systems like CRMs.

What are the main data aggregation methods?

+

The four main data aggregation methods are: (1) Manual aggregation using spreadsheets and formulas like VLOOKUP to merge data from exported files — simple but does not scale. (2) ETL pipeline aggregation using tools like dbt or Fivetran to extract, transform, and load data into a warehouse. (3) API-based real-time aggregation that queries multiple sources programmatically and combines responses on the fly. (4) Reverse ETL aggregation that pushes warehouse data back into operational tools. Most B2B teams start with manual methods and graduate to automated approaches as data volume grows.

What is an example of data aggregation in sales?

+

A common sales example: a rep needs to call a prospect and needs their direct phone number, company revenue, tech stack, and recent funding activity. No single system has all of this. The CRM has the company name and a possibly outdated phone number. LinkedIn has the current job title. ZoomInfo has the direct dial. Crunchbase has funding data. Data aggregation combines all of these into a single prospect profile the rep can use. Cleanlist automates this by querying 15+ providers through waterfall enrichment and aggregating the best data points into one record.

What is data aggregation in a database?

+

Data aggregation in a database refers to SQL operations that compute summary statistics across groups of rows. The most common approach uses aggregate functions — COUNT, SUM, AVG, MIN, MAX — combined with GROUP BY to collapse multiple rows into summary results. For example, SELECT region, SUM(revenue) FROM deals GROUP BY region aggregates deal revenue by region. Window functions like SUM() OVER (PARTITION BY ...) provide running aggregations without collapsing rows. OLAP cubes extend this with multidimensional roll-up, drill-down, slice, and dice operations for analytical workloads.

What is the difference between data aggregation and data integration?

+

Data aggregation collects and combines data from multiple sources into a unified dataset, typically as a batch or periodic operation that produces a consolidated snapshot. Data integration connects systems so data flows between them continuously and in real time, maintaining ongoing synchronization. A third related concept, data enrichment, enhances existing records with additional attributes from external sources. Most B2B data operations use all three: integration keeps systems connected, aggregation builds comprehensive records, and enrichment fills gaps with external data.

What is real-time data aggregation?

+

Real-time data aggregation runs synchronously: when a record is created or queried, the system fans out requests to every source and merges responses inline, typically within 1-2 seconds. It powers form-fill personalization, live lead routing, and fraud detection — anywhere staleness is unacceptable. The trade-off is cost (per-call pricing instead of bulk discounts) and complexity (every source must degrade gracefully if one fails). Most B2B teams mix real-time on inbound leads with batch on the long tail.

How is data aggregation used in healthcare?

+

Healthcare aggregation pulls patient data from EHRs (Epic, Cerner), lab systems (LabCorp, Quest), imaging archives, and claims databases into a longitudinal patient record. HIPAA requires field-level provenance and consent tracking on every merge, and Health Information Exchanges are essentially aggregation networks operating at the regional or state level. Patient identity matching averages 85-90% accuracy on name + DOB + last-four-SSN; probabilistic matching pushes that to 95%+.

How is data aggregation used in banking?

+

Banks aggregate transaction data from card networks, ACH systems, and core banking platforms to build a unified customer ledger. Reconciliation rules favor the source-of-record (the bank's general ledger) over derived sources. AML and KYC workflows aggregate identity data from credit bureaus, sanctions lists, and government registries, with strict audit trails on every merge. The Federal Reserve's 2024 payments study estimates banks aggregate from 12-30 distinct internal and external systems per customer.

What is the difference between data aggregation and data warehousing?

+

Data warehousing is the storage layer — a centralized repository where structured data from multiple sources is consolidated for analytical queries. Data aggregation is the operation — collecting and combining data from multiple sources into unified records or summary statistics. You aggregate data INTO a warehouse, but you can also aggregate data without one (real-time API aggregation, in-memory aggregation, spreadsheet merging). Warehouses are infrastructure; aggregation is a process.

What are the four main aggregation patterns in B2B data?

+

The four patterns are merge (combining overlapping records into a single golden record by matching on shared identifiers like email or domain), append (adding new fields from a secondary source without overwriting), deduplicate (identifying and collapsing duplicate entries from the same entity across sources), and normalize-then-combine (standardizing field formats — dates, phone numbers, job titles — before merging to reduce conflicts). Production systems usually combine all four in a defined sequence.

Improve your data aggregation workflow

Enrich, verify, and score your B2B data with 98% accuracy. 30 free credits to start.

No credit card required

Related Terms

30 credits included. No credit card required. Set up in 5 minutes.