Data ingestion is “the overall process of collecting, transferring, and loading data from one or multiple sources so that it may be analyzed immediately or stored in a database for later use” (Import.io). Data may be entered “into a database, data warehouse, data repository or application” (Keren Blau).

“Data can be streamed in real time or ingested in batches. When data is ingested in real time, each data item is imported as it is emitted by the source. When data is ingested in batches, data items are imported in discrete chunks at periodic intervals of time.

“[I]t can be challenging for businesses to ingest data at a reasonable speed. … When data ingestion is automated, the software used to carry out the process may also include data preparation features to structure and organize data so it can be analyzed on the fly or at a later time by business intelligence (BI) and business analytics (BA) programs” (TechTarget).

“Data ingestion tools provide a framework that allows companies to collect, import, load, transfer, integrate, and process data from a wide range of data sources” (PAT Research).

More from the data glossary

A definitive guide to data definitions and trends, from the team at Stitch.

Give Stitch a try, on us

Stitch streams all of your data directly to your analytics warehouse.

Set up in minutes Unlimited data volume during trial 5 million rows of data free, forever