Help the communities most affected by the California wildfires in only a few minutes. We'll donate $10 for every review you submit.

Best Big Data Integration Platform

Big data integration platforms help facilitate and analyze big data integrations across cloud applications. They will typically facilitate the intgration between big data processing solutions, applications and databases. Big data integration platforms usually require big data to have been processed prior to integration, but they facilitate the use of big data sets and insights. Companies use these to manage and store big data clusters and use them within cloud applications. They can help simplify the management of enormous amounts of data collected from IoT endpoints, applications, and communications. Some big data integration tools provide stream analytics capabilities, but provide more functionality for data management.

To qualify for inclusion in the Big Data Integration category, a product must:

  • Integrate big data processing data to external sources
  • Ingest and distribute large sets of homogenous and heterogenous data
  • Create a structured pipeline for big data management processes
G2 Crowd Grid® for Big Data Integration Platform
High Performers
Momentum Leaders
Momentum Score
Market Presence
Star Rating

Big Data Integration Platform reviews by real, verified users. Find unbiased ratings on user satisfaction, features, and price based on the most reviews available anywhere.

Compare Big Data Integration Platform
Results: 24
    G2 Crowd takes pride in showing unbiased ratings on user satisfaction. G2 Crowd does not allow for paid placement in any of our ratings.
    Sort By:

    Omni-Gen Master Data Management (MDM) Edition provides a single platform for generating applications that combine data integration, data quality, and master data management – in a fraction of the time such projects used to require. The benefits are huge – typical project times can be reduced from a year-and-a-half to six months or less. (NOTE: Other iWay components, such as iWay Service Manager, are included as part of the Omni-Gen platform.)

    SnapLogic is the leader in self-service integration. The company’s Enterprise Integration Cloud makes it fast and easy to connect applications, data, APIs, and things. Hundreds of Global 2000 customers — including Adobe, AstraZeneca, Box, GameStop, Verizon, and Wendy’s — rely on SnapLogic to automate business processes, accelerate analytics, and drive digital transformation. SnapLogic was founded by data industry veteran Gaurav Dhillon and is backed by blue-chip investors including Andreessen Horowitz, Capital One, Ignition Partners, Microsoft, Triangle Peak Partners, and Vitruvian Partners. For a free trial, please visit

    Talend simplifies big data integration with graphical tools and wizards that generate native code so you can start working with Apache Hadoop, Apache Spark, Spark Streaming and NoSQL databases today. Talend Big Data Integration platform delivers high-scale, in-memory fast data processing, as part of the Talend Data Fabric solution, so your enterprise can turn more and more data into real-time decisions.

    Azure Data Factory (ADF) is a service designed to allow developers to integrate disparate data sources. It provides access to on-premises data in SQL Server and cloud data in Azure Storage (Blob and Tables) and Azure SQL Database.

    Apache NiFi is a software project designed to enable the automation of data flow between systems.

    Data Virtuality provides data integration solutions that help companies to easily connect and manage their data from multiple data sources such as APIs, databases and flat files. The revolutionary single source of data truth platform combines data virtualization and automated ETL. In this way not only is data management simplified but data integration efforts are significantly reduced – by up to 80%. Imprint:

    IBM® InfoSphere® DataStage® is a leading ETL platform that integrates data across multiple enterprise systems. It leverages a high performance parallel framework, available on-premises or in the cloud. The scalable platform provides extended metadata management and enterprise connectivity. It integrates heterogeneous data, including big data at rest (Hadoop-based) or big data in motion (stream-based), on both distributed and mainframe platforms. It supports IBM Db2® Z and Db2 for z/OS®, applies workload and business rules, and integrates real-time data in an easy to deploy, scalable platform. Learn More:

    Apache Gobblin is a distributed data integration framework designed to simplify common aspects of big data integration such as data ingestion, replication, organization and lifecycle management for both streaming and batch data ecosystems.

    Apache Sqoop is a tool designed for efficiently transferring bulk data between Apache Hadoop and structured datastores such as relational databases.

    Denodo provides performance and unified access to the broadest range of enterprise, Big Data, cloud and unstructured sources.

    Oracle Event Hub Cloud Service delivers the power of Kafka as a managed streaming data platform integrated into the Oracle Cloud ecosystem. Create Topics and start streaming or manage and deploy your own Dedicated Kafka Cluster with Elastic Scalability.

    Attunity CloudBeam offers accelerated file transfer to cloud storage such as AWS S3

    Reduce costs and improve service delivery with faster Hadoop implementations

    Connect data gathered from Akamai DNS infrastructure with open big data systems or purpose-built applications.

    XenonStack is a software company that specializes in product development and providing DevOps, big data integration, real time analytics and data science solutions.

    HVR is designed to move large volumes of data fast and efficiently in complex environments for real-time updates.

    Lore IO is a data management platform provider that unifies on-demand, real-time business knowledge.

    Nexla monitors, adapts, and securely moves data between companies so you can focus on the real work

    Leverage the power and scale of Hadoop and NoSQL databases using the leading free open source ETL tool for big data. Simply drag, drop, and configure pre-built components, generate native code, and deploy to Hadoop for simple EDW offloading and ingesting data into your data lake.

    DataVirtuality Pipes is an agile and scalable cloud data integration solution to empower analytic tools with data that matters. With Pipes, you can integrate data from 150+ databases and API’s to any data warehouse in 5 minutes. There is no coding or maintenance of API’s required. It scales with your company and Business Intelligence needs and is fully upgradeable to DataVirtuality’s Logical Data Warehouse.

    Striim platform is an end-to-end streaming data integration and operational intelligence solution designed to enable continuous query and processing and streaming analytics.

    Talend Data Fabric is a unified platform that enables you to manage all your enterprise data within a single environment. Leverage all the cloud has to offer to manage your entire data lifecycle – from connecting the broadest set of data sources and platforms to intuitive self-service data access.

    Vortex integrates, normalizes, cleanses and protects data at the speed it is being generated without slowing it down. The sources of data in large enterprises and broadband environments is growing exponentially. In order for any enterprise to make use of these data it is necessary to integrate, translate and protect the data. It is your enterprise data hub–your one-stop shop for converting raw data into useful information.

    Provides fast and easy adoption of industry standards for B2B integration. An add-on tool for webMethods Trading Networks. Read more