Build Data iPaaS Applications with Wizards Using Striim

Now that you have a high-level overview of the Striim platform, let’s discuss how you can build data iPaaS applications with Striim.

You can deploy the entire platform in the cloud either by bringing your own license or as a metered iPaaS service. This gives you everything – it gives you all the sources, all the targets, and all the capabilities of the platform. There are also specific versions that you can deploy for particular solutions. So, for example, if you have on-premises Oracle databases and you want to push that data, as it’s changing, say to Azure SQL Data Warehouse, you can use that specific solution. You can still work with all of the sources, but you’re limited to delivering the data into Azure SQL Data Warehouse. There are dozens of specific cloud service solutions. They also are metered; they run as iPaaS in the cloud.

There are also a lot of different flavors of iPaaS. People usually bring up the multi-tenant type of iPaaS where the vendor hosts the service for you, allowing you to login and have access within an environment to be able to build data flows, etc. Striim chose not to go that route because customers are not typically that happy with the notion of being in a joint, multi-tenant environment where they are worried about data security and being guaranteed use of resources so that their applications will run at the right speed, etc.

Instead, Striim went with the ability to purchase the platform on Azure, Google Cloud, or Amazon as a metered service. With this approach, it’s running in your cloud environments, so you control the security, data, and everything else. Customers are more comfortable with this than the notion of a multi-tenant solution for iPaaS. As you can see in this video, we have metered iPaaS solutions for data in the marketplace for all three major cloud environments – Azure, AWS, and Google Cloud.

When you are working with the platform, on-premises or in the cloud, you interact with it through our intuitive web-based UI. This provides access to existing applications, as well as being able to import and create new applications.

You can start by building or importing applications, so, for example, if you’ve already built something in development, you can import it into production. If you are starting from scratch, you begin with an empty application and drag and drop components into the flow designer. But the easier way to get going is through the wizards which provide a large number of application templates. A lot of users start with a template because it enables you to rapidly build simple data flows, and check everything is correct as you go along.

For example, if you wanted to read from a MySQL database on-premises and deliver into Azure Cosmos DB, you could name the application, “MySQLtoCosmos,” and put it in a namespace. Namespaces keep things separate, and the way our security works, you can lock things down so that only certain people have access to certain namespaces. You can do much finer-grain things than that. You can give users access to the data that’s produced as the end result of the data pipeline, but not the raw data because that may have personally identifiable information in it. In our example, we will filter all that out before we push it into the cloud.

So you create a new namespace and save it. And then you can actually build data iPaaS applications, letting the wizards walk you through setting up the connection. Once all properties are configured, it will test everything to make sure that the connection is correct. This is an important step. One of the reasons Striim introduced its many wizards and templates was to make the development process as easy, intuitive, and fast as possible.

So in these steps, we check to make sure that not only does the connection to the database work, but also that connection has the right privileges, and that change data capture (CDC) is turned on. CDC collects all the inserts, updates, and deletes as they happen in a database (this is enabled at the database level). It also checks that you can get to the database metadata so you can actually see what tables and columns there are. If any of these steps don’t work, then the wizards will tell you what to do. Basically the instructions in the manual are mirrored by steps in the wizards so people know exactly what to do. In certain cases, the wizards can even do it for you. Once the connection is verified, you get to choose your data and go on to the next step. And then finally you’ll configure your target.

To learn more about how to build data iPaaS applications with Striim, read our Striim Platform Overview data sheet, set up a quick demo with a Striim technologist, or provision the Striim platform as an iPaaS solution on Microsoft Azure, Google Cloud Platform, or Amazon Web Services.

If you missed it or would like to catch up on this iPaaS blog series, please read part 1, “The Striim Platform as a Data Integration Platform as a Service.”

 

What is Streaming SQL?esdfgv

 

 

Streaming SQL has become essential to real-world, real-time data processing solutions. But before examining what it is and how it works, we need to take a brief look back.

With the continuous and staggering growth of data volumes over the years, and the rising demands for analysis of data, Structured Query Language, or SQL, has become an essential component of data management and business analytics.What is Streaming SQL?

Because databases store data before it’s available for querying, however, this data is invariably old by the time it’s queried. Today, many organizations need to analyze data in real-time, which requires the data to be streamed. As a result of this shift, there’s a need for a new version of SQL that supports stream processing.

Enter Streaming SQL. Streaming SQL is similar to the older version of SQL, but it differs in how it addresses stored and real-time data. Streaming SQL platforms are continuously receiving flows of data. It’s this continuous nature of streaming that gives the technology its true value compared with traditional SQL solutions.

A key part of streaming SQL are windows and event tables, which trigger actions when any kind of change occurs with the data. When a window is updated, aggregate queries recalculate, and this provides results such as sums over micro-batches.

Streaming systems allow organizations to input huge volumes of data—including reference, context, or historical data—into event tables from files, databases, and various other sources. These tools enable users to write SQL-like queries for streaming data without the need to write code.

With Streaming SQL, queries are often highly complex, using case statements and pattern-matching syntax. These solutions make it easy for organizations to ingest, process, and deliver real-time data across a variety of environments—whether they are in the cloud or on-premises.

This helps enterprises quickly adopt a modern data architecture, creating streaming data pipelines to public cloud environments such as Microsoft Azure, Amazon Web Services, and Google Cloud Platform, as well as to Kafka, Hadoop, NoSQL, and relational databases.

It’s important to realize that Streaming SQL is not something that should be used to run on all data, such as massive databases with a billion rows. That’s not what it’s designed for. It’s better suited for working on smaller subsets of data, when there is a need to get quick results and immediately identify value in new data that’s being created.

One of the strengths of Streaming SQL comes from its ability to transform, filter, aggregate, and enrich data. It has the ability to combine all these functions together to enable organizations to get maximum value from the data constantly streaming into their systems.

To learn more about the power of streaming SQL, visit Striim Platform Overview product page, schedule a demo with a Striim technologist, or download a free trial of the platform and try it for yourself!

 

Striim Announces Strategic Partnership with Snowflake to Drive Cloud-Based Data-Driven Analytics

We are excited to announce that we’ve entered into a strategic partnership with Snowflake, the data warehouse built for the cloud, in which Striim will be used to move real-time data into Snowflake. Through this strategic partnership, Snowflake users will be empowered to gain fast insights from their cloud-based analytics.

Enterprise companies are quickly adopting Snowflake because its architecture is built from the ground up for the cloud. Snowflake offers speed, scalability, and cost-effectiveness, along with zero management. In order to attain fast analytics, you need access to real-time data, and that’s where Striim comes in. Striim is leveraging its vast real-time data integration capabilities to enable Snowflake users to collect and move data from a variety of sources into their environment to accelerate their data-driven analytics.

Striim uses low-impact change data capture (CDC) to move data from existing on-prem databases, including SQL Server, Oracle, MongoDB, HPE NonStop, PostgreSQL, MySQL and Amazon RDS. Striim can also help you migrate data warehouses such as Teradata, Netezza, Amazon Redshift, and Oracle Exadata. Additionally, Striim can collect from messaging systems, Hadoop, log files, sensors, and security devices and other systems. Striim also has analytical capabilities to monitor and measure transaction lag and alert when SLAs are not met.

Through CDC, Striim can handle large volumes of enterprise data securely and reliably. Along with its CDC capabilities, Striim adds further value through in-flight processing, transformations, and denormalization to further assist Snowflake users in providing quicker analysis by continuously delivering data to Snowflake in the right format, and with added context.

Striim has a number of use cases with customers using the solution for both online migrations and continuous integration to Snowflake.

For example, a company offering HR and well-being solutions, is a joint customer that was searching for a low-latency streaming integration solution that was scalable and also offered a secure data warehouse with analytical options. This organization’s  goal was to enable employees to instantly query their personal information, as well as allow employers to identify trends and patterns from the data.

With Striim + Snowflake, this business has been delivering real-time data and analytics using CDC from Oracle to Azure for streamlined operations. The partnership between Striim and Snowflake has dramatically enhanced the company’s operationsoperations, enabling them to make faster, smarter decisions based on their real-time data.

To learn more about the Striim-Snowflake solution and Striim’s partnership with Snowflake, please read our press release, visit our Striim for Snowflake product page, or set up a quick demo with a Striim technologist.

Kafka to HDFS

The real-time integration of messaging data from Kafka to HDFS augments transactional data for richer context. This allows organizations to gain optimal value from their analytics solutions and achieve a deeper understanding of operations – essential to establishing and sustaining competitive advantage.

To truly leverage the high volumes of data residing in Kafka stores, companies need to be able move it, process it, and deliver it to a variety of on-premises and cloud systems with sub-second latency. It also needs to be integrated with operational data from a wide variety of sources.

Traditional batch-based solutions are not designed for situations where data is time-sensitive – they are simply too slow. To allow organizations use their data to enhance operations, tailor services, and improve customer experiences, data delivery from Kafka to HDFS systems needs to be scalable and in real time.

Continuously Deliver Data

With Striim, companies can continuously deliver data in real time from Kafka to HDFS, as well as to a wide range of targets including Hadoop and cloud environments. Depending on the requirements of the organization, all the Kafka data can be written to a number of different targets simultaneously. In use cases where not all the data is required, data can be matched to specific criteria to deliver a highly relevant subset of data to the target.

Striim can create data flows to deliver the data from Kafka to HDFS in milliseconds, “as-is.” However, depending on how the data is going to be utilized, the user may require the data to be processed, prepared, and delivered in the right format. Striim supports continuous queries to filter, transform, aggregate, enrich, and analyze the data in-flight before delivering it with sub-second latency.

Analyze Data In-Flight

By analyzing the data in-flight, Kafka users can capture time-sensitive information as the data is flowing through the data stream. Striim pushes insights and alerts to interactive dashboards highlighting real-time data and the results of pattern matching, correlation, outlier detection, predictive analytics, and further enables drill-down and in-page filtering.

Learn more about integrating and processing Kafka to HDFS in real-time, please visit our Kafka integration page.

Our experts can show you how to get maximum value from your analytics solutions using Striim for real-time data integration from Kafka to HDFS. Please contact us to schedule a demo.

Oracle CDC to Postgres

Real-Time Data Movement with Oracle CDC to Postgres

As an open source alternative, Postgres offers a lower total cost of ownership and the ability to store structured and unstructured data. Real-time movement of transactional data using Oracle CDC to Postgres is essential to creating a rich and up-to-date view of operations and improving
customer experiences.

Oracle CDC to Postgres

IDC projects that by the year 2025, 80% of all data will be unstructured. Emails and social media posts are good examples of unstructured data. The ability to integrate unstructured, semi-structured and structured data from transactional databases into the enterprise is vital for timely and relevant analysis. To get a deep understanding from all the data an organization captures and records and to get the most value from it, it must be in the right place and in the right format – in real time.

Continuous movement of transactional data using Oracle CDC to Postgres ensures the organization is utilizing the real-time information from on-prem transactional databases and other data stores that is needed to make decisions that optimize user experience and drive higher revenue.

Moving data from enterprise databases to Postgres using traditional ETL processes introduces latency. Delays incurred while the data is being migrated or updated results in an out-of-date picture of the business, and limits the extent to which decisions can have any significant impact. Organizations also face a series of challenges managing storage and accessing the actual data that can produce real value to the organization if they move all the data as is.

How Striim Simplifies Oracle CDC to Postgres

Striim enables organizations to generate real value from the transactional data residing in their existing Oracle databases. Using non-intrusive change data capture (CDC), Striim enables continuous data ingestion from Oracle to Postgres with sub-second latency. Users can easily set up ingestion via Striim’s pre-configured CDC wizards, and drag-and-drop UI.

Moving and processing data in-flight, Striim filters data that is not required and delivers what is important to Postgres – in real time. The data can also be transformed and enriched so it is delivered in the format required. Oracle CDC to Postgres allows organizations gain access to critical insights sooner and make more informed operational decisions faster.

Once the real-time data pipelines are built and the initial data load using Oracle CDC to Postgres has been performed, continuous updating with every new database transaction ensures that analytics applications have the most up-to-date information. Built-in monitoring continuously compares the source and target, validating database consistency and providing assurance that the replicated environment is completely up-to-date with the on-prem Oracle instance.

For more information on real-time data integration and processing using Striim’s Oracle CDC to Postgres solution, please visit our Change Data Capture page.

To see first-hand how easy it is to move data to Postgres using Striim’s Oracle CDC to Postgres functionality, please schedule a demo with one of our technologists.

Striim Announces Real-Time Data Migration to Google Cloud Spanner

Google Cloud Marketplace

The Striim team has been working closely with Google to deliver an enterprise-grade solution for online data migration to Google Cloud Spanner. We’re happy to announce that it is available in the Google Cloud Marketplace. This PaaS solution facilitates the initial load of data (with exactly once processing and delivery validation), as well as the ongoing, continuous movement of data to Cloud Spanner.Real-Time Migration to Google Cloud Spanner

The real-time data pipelines enabled by Striim from both on-prem and cloud sources are scalable, reliable and high-performance. Cloud Spanner users can further leverage change data capture to replicate data in transactional databases to Cloud Spanner without impacting the source database, or interrupting operations.

Google Cloud Spanner is a cloud-based database system that is ACID compliant, horizontally scalable, and global. Spanner is the database that underlies much of Google’s own data collection, and it has been designed to offer the consistency of a relational database with the scale and performance of a non-relational database.

Migration to Google Cloud Spanner requires a low-latency, low-risk solution to feed mission-critical applications. Striim offers an easy-to-use solution to move data in real time from Oracle, SQL Server, PostgreSQL, MySQL, and HPE NonStop to Cloud Spanner while ensuring zero downtime and zero data loss. Striim is also used for real-time data migration from Kafka, Hadoop, log files, sensors, and NoSQL databases to Cloud Spanner.

While the data is streaming, Striim enables in-flight processing and transformation of the data to maximize usability of the data the instant it lands in Cloud Spanner.

Learn More

To learn more about Striim’s Real-Time Migration to Google Cloud Spanner, read the related press release or provision Striim’s Real-Time Data Integration to Cloud Spanner in the Google Cloud Marketplace.

Real-Time Database CDC to Cloudera

 

As Cloudera increasingly invests in its Enterprise Data Cloud, the ability move data via change data capture or CDC to Cloudera has never been more important. Database CDC to Cloudera helps Cloudera users gain more operational value from their analytics solutions by loading critical database transactions in real time.CDC to Cloudera

The timely ingestion of large volumes of data to Cloudera is imperative to realizing the true operational value of the platform. The explosion in the amount of data generated and the variety of data formats residing in traditional relational databases and data warehouses requires an ingestion process that is real-time and scalable.

Traditional methods or batch ETL uploads fall short in today’s business timeframes. Latency renders operational and transactional data obsolete and unable to provide Cloudera solutions with the real-time data required for operational intelligence and reporting. The negative performance impact of batch processing on transactional databases is also a major reason to move only the changed data in a continuous fashion.

To address the concerns mentioned above, there is a solution to ingest changed data in real time from databases: CDC to Cloudera from Striim. This enterprise-grade streaming data integration solution for Cloudera supports high-volume environments and allows users to move real-time data from a wide variety of sources without impacting source systems.

By moving only change data – continuously and with essential scalability – Cloudera users can rely on the Striim platform for the delivery of data. Data can be loaded as-is, or with a variety of processing, transformations or enrichments applied, all with sub-second latency and in the right format to support specific use cases.

A one-time initial load with continuous change updates ensures up-to-the-second data delivery to Cloudera to support operational decision making. Striim also offers real-time pipeline monitoring with alerting, which is particularly important in the context of mission-critical solutions.

Striim currently offers low-impact, log-based CDC to Cloudera from the following data sources: Oracle, Microsoft SQL Server, MySQL, PostgreSQL, HPE NonStop SQL/MX, HPE NonStop SQL/MP, HPE NonStop Enscribe, MongoDB, and MariaDB. All of these databases can be accessed via Striim’s easy-to-use Wizards and drag-and-drop UI, speeding delivery of CDC to Cloudera solutions. In addition, Striim offers pre-built starter integration applications, such as PostgreSQL CDC to Kafka, that can be leveraged to significantly reduce development efforts of any CDC-based application.

If you’d like a brief walk-through of Striim’s CDC to Cloudera offering, please schedule a demo.

What is iPaaS for Data?

Organizations can leverage a wide variety of cloud-based services today, and one of the fastest growing offerings is integration platform as a service. But what is iPaaS?

There are two major categories of iPaaS solutions available, focusing on application integration and data integration. Application integration works at the API level, typically involves relatively low volumes of messages, and enables multiple SaaS applications to be woven together.What is iPaaS for Data?

Integration platform as a service for data enables organizations to develop, execute, monitor, and govern integration across disparate data sources and targets, both on-premises and in the cloud, with processing and enrichment of the data as its streaming.

Within the scope of iPaaS for data there are older batch offerings, and more modern real-time streaming solutions. The latter are better suited to the on-demand and continuous way organizations are utilizing cloud resources.

Streaming data iPaaS solutions facilitate integration through intuitive UIs, by providing pre-configured connectors, automated operators, wizards and visualization tools to facilitate creation of data pipelines for real-time integration. With the iPaaS model, companies can develop and deploy the integrations they need without having to install or manage additional hardware or middleware, or acquire specific skills related to data integration. This can result in significant cost savings and accelerated deployment.

This is particularly useful as enterprise-scale cloud adoption becomes more prevalent, and organizations are required to integrate on-premises data and cloud data in real time to serve the company’s analytics and operational needs.

Factors such as increasing awareness of the benefits of iPaaS among enterprises – including reduced cost of ownership and operational optimization – are fueling the growth of the market worldwide.

For example, a report by Markets and Markets notes that the Integration Platform as a Service market is estimated to grow from $528 million in 2016 to nearly $3 billion by 2021, at a compound annual growth rate (CAGR) of 42% during the forecast period.

“The iPaaS market is booming as enterprises [embrace] hybrid and multi-cloud strategies to reduce cost and optimize workload performance” across on-premises and cloud infrastructure, the report says. Organizations around the world are adopting iPaaS and considering the deployment model an important enabler for their future, the study says.

Research firm Gartner, Inc. notes that the enterprise iPaaS market is an increasingly attractive space due to the need for users to integrate multi-cloud data and applications, with various on-premises assets. The firm expects the market to continue to achieve high growth rates over the next several years.

By 2021, enterprise iPaaS will be the largest market segment in application middleware, Gartner says, potentially consuming the traditional software delivery model along the way.

“iPaaS is a key building block for creating platforms that disrupt traditional integration markets, due to a faster time-to-value proposition,” Gartner states.

The Striim platform can be deployed on-premises, but is also available as an iPaaS solution on Microsoft Azure, Google Cloud Platform, and Amazon Web Services. This solution can integrate with on-premise data through a secure agent installation. For more information, we invite you to schedule a demo with one of our lead technologists, or download the Striim platform.

Oracle Change Data Capture: Methods, Benefits, Challenges

If there’s one thing today’s economy values, it’s speed. To enable faster decisions, businesses are rapidly moving data to the cloud, building powerful AI-driven applications, and increasingly relying on operational analytics. These initiatives all depend on one thing: a constant, reliable stream of real-time data.

But many organizations struggle to deliver real-time data; their data strategies are stuck in the past. Traditional data movement, built on slow, scheduled batch jobs (ETL), simply can’t keep up with the industry’s need for speed. This legacy approach creates data latency, leaving decision-makers with stale information and preventing applications from responding to events as they happen.

Sound familiar? Perhaps you already know the consequences of stale data. When you can’t get data when you need it, you risk missing key opportunities, creating inefficiencies, and widening the gap between data and its potential value.

This is where Oracle Change Data Capture (CDC) comes in. CDC offers a powerful and efficient way to capture every insert, update, and delete from your critical Oracle databases in real time. When implemented correctly, it can become the engine for modern, event-driven data architectures. But without the right strategy and tools, navigating the complexities of Oracle CDC can be challenging.

This guide will provide a clear roadmap to mastering Oracle CDC. We’ll explore what it is, how it works, and how to choose the right approach for your business—transforming your data infrastructure from a slow-moving liability into a real-time strategic asset.

What is Oracle Change Data Capture?

Oracle Change Data Capture (CDC) is a technology designed to identify and capture changes made to data in an Oracle database. It can capture DML (INSERT, UPDATE, and DELETE), DDL (CREATE, ALTER, DROP, and TRUNCATE) changes in your database the moment they occur. Think of it as a surveillance system for your data, noting every single modification in real time. This is about building infrastructure that can understand and react to new events. By tracking changes as they happen, CDC provides a continuous stream of change events that form the foundation of a responsive data strategy. This capability is essential for businesses that need to power streaming analytics, execute seamless cloud migrations with zero downtime, and build sophisticated, event-driven AI applications that rely on the freshest data possible.

Common Use Cases for Oracle Change Data Capture

At its best, Oracle CDC doesn’t just move data; it enables better outcomes. By providing a real-time stream of changes, CDC unlocks new capabilities for companies of all sizes, from agile startups to large enterprises across finance, retail, manufacturing, and more.

Cloud Migration and Adoption

For any company moving its Oracle workloads to the cloud, minimizing downtime is critical. Oracle CDC facilitates zero-downtime migrations by continuously syncing the on-premises source database with the new cloud target. This allows for a phased, low-risk cutover, ensuring business operations are never disrupted.

Streaming Data Pipelines for Analytics and AI

Advanced analytics and AI applications thrive on fresh data. CDC is the engine that feeds real-time data from Oracle databases into cloud data warehouses like Snowflake, Google BigQuery, and Databricks, or into streaming platforms like Apache Kafka. This allows data science teams to build dashboards with up-to-the-second accuracy and train machine learning models on the most current dataset available.

Offloading Operational Reporting and Upstream Analytics

Running heavy analytical queries against a live production (OLTP) database can degrade its performance, impacting core business applications. CDC allows companies to replicate transactional data to a secondary database or another backup storage option in real time. This offloads the reporting workload, ensuring that intensive analytics don’t slow down critical operational systems.

Event-Driven Application Development and Platform Modernization

In event-driven architecture, services communicate by reacting to events as they happen. Oracle CDC turns database changes into a stream of events. For example, a new entry in an orders table can trigger a notification to the shipping department, update inventory levels, and alert the customer, all in real time. This is invaluable for industries like e-commerce and logistics that need to automate complex workflows.

Disaster Recovery and High Availability

For mission-critical systems, maintaining a real-time, up-to-date replica of a production database is essential for disaster recovery. Oracle CDC ensures that a standby database is always in sync with the primary system. In the event of an outage, the business can failover to the replica with minimal data loss and disruption.

Data Synchronization Across Systems

Enterprises often have multiple systems that need a consistent view of the same data. Whether it’s keeping a CRM and an ERP system in sync or ensuring data consistency across geographically distributed databases, CDC is a reliable solution for real-time data synchronization, eliminating data silos and inconsistencies before they spring up.

Regulatory Compliance and Audit Readiness

For industries with strict regulatory requirements, like finance and healthcare, maintaining a detailed audit trail of all data changes is non-negotiable. Oracle CDC provides an immutable, chronological log of every insert, update, and delete. This creates a reliable audit history that can be used to ensure compliance and simplify audit processes.

AI Enablement

When it comes to getting AI-ready, enterprises need the freshest data available to fuel AI models with relevant insights. Real-time CDC ensures AI applications get the most up-to-date insights to power RAG engines with continuous, accurate updates. The result: faster, smarter, more responsive AI outputs based on relevant business contexts.

How Oracle Change Data Capture Works

Unlike systems that repeatedly poll tables for changes—an approach that is both inefficient and resource-intensive—Change Data Capture (CDC) taps directly into Oracle’s internal mechanisms. The most robust and performant CDC methods leverage Oracle’s transaction logs to capture changes with minimal impact on the source system. At the core of this process are Oracle redo logs. Every data-modifying transaction—whether an insert, update, or delete—is first recorded in a redo log file. This built-in mechanism ensures data integrity and supports recovery in the event of a system failure. Once redo logs reach capacity, they are archived into archive logs for persistence and historical tracking. Log-based CDC tools like Striim connect to the database and “mine” these redo and archive logs in a non-intrusive way. Striim offers two Oracle CDC adapters:

  • LogMiner-based Oracle Reader – Uses an Oracle LogMiner session to scan and capture server-side changes.
  • OJet Adapter – A high-performance, API-driven solution designed for large-scale, real-time data capture.

Both approaches are highly efficient and have minimal overhead, preserving the performance and stability of the source database. Learn more about Striim’s Oracle CDC adapters here.

Simple Oracle CDC Flow:

  1. Transaction Occurs: An application performs an INSERT, UPDATE, or DELETE or a DDL change on an Oracle database table.
  2. Log Write: Oracle writes the change to its redo log.
  3. CDC Capture: A CDC tool (like Striim) reads the change from the redo log in real time.
  4. Stream Processing (Optional): The data can be transformed, filtered, or enriched in-flight.
  5. Data Delivery: The processed data is delivered to the target (e.g., Snowflake, Kafka, BigQuery).

Methods of Implementing CDC in Oracle

There are multiple ways to implement CDC in Oracle, each with its own trade-offs in performance, complexity, and cost. There’s no one “correct” method to choose—it comes down to selecting the approach that best matches the needs of your data management strategy and business goals.



Log-Based CDC

Reads changes directly from Oracle redo/archive logs. The gold standard for high-performance, low-latency pipelines where source performance is critical.

Impact:
Very Low
Complexity:
Moderate to High
Cost:
Variable

Trigger-Based CDC

Uses database triggers on each table to write changes to audit tables. Best for low-volume tables or when log access is restricted.

Impact:
High
Complexity:
Low to High
Cost:
High (Performance)

Oracle GoldenGate

Oracle’s proprietary log-reading replication software. Ideal for enterprise Oracle-to-Oracle replication with a large budget.

Impact:
Low
Complexity:
High
Cost:
Very High

Oracle Native CDC

Deprecated

A built-in feature in older Oracle versions using triggers and system objects. It is no longer supported and should not be used for new projects.

Impact:
Moderate to High
Complexity:
High
Cost:
N/A

Log-Based Oracle API CDC

The gold standard for high-performance Oracle CDC leverages Oracle’s native APIs to capture changes directly from Logical Change Records (LCRs)—Oracle’s internal representation of both DML (INSERT, UPDATE, DELETE) and DDL (CREATE, ALTER, DROP) operations. These records are derived from the database’s redo logs, offering a highly accurate, low-latency stream of transactional and structural changes. Because this method uses the same internal mechanisms Oracle relies on for replication and recovery, it ensures minimal performance impact on the source system. However, interacting directly with LCRs and Oracle’s APIs can be complex and requires advanced database knowledge. Striim simplify this by providing a fully managed, Oracle-integrated CDC solution that captures both data and schema changes in real time—without the need for extensive manual configuration.

Trigger-Based CDC

This approach involves placing database triggers on each source table. When a row is inserted, updated, or deleted, the trigger fires and copies the change into a separate “shadow” or audit table. While conceptually simple, this method adds significant overhead to the production database, as every transaction now requires an additional write operation. This can slow down applications and become a major performance bottleneck, especially in high-throughput environments. It’s also difficult to maintain as the number of tables grows.

Oracle GoldenGate

Oracle GoldenGate is a premium, feature-rich data replication solution known for its deep integration with the Oracle database and its ability to support high-volume, low-latency replication. While it excels in large-scale, mission-critical environments—particularly for Oracle-to-Oracle replication—its complexity and high licensing costs can be a barrier for many organizations. Striim offers a unique advantage by allowing customers to leverage existing GoldenGate trail files without requiring a full GoldenGate deployment. This capability enables organizations to preserve their investment in GoldenGate infrastructure while using Striim’s modern, flexible platform for real-time data integration, transformation, and delivery. Striim is one of the few solutions on the market that can read GoldenGate trail files directly, providing a cost-effective and simplified alternative for operationalizing data across diverse targets like Snowflake, BigQuery, Kafka, and more.

Oracle Native LogMiner

Oracle previously offered a built-in feature called Continuous Mine Mode to support Change Data Capture (CDC) in earlier versions of its database. However, this mode was complex, less performant than modern alternatives, and has been deprecated starting with Oracle 19c. While CONTINUOUS_MINE is no longer supported, LogMiner remains fully functional and officially supported by Oracle. LogMiner traditionally reads redo and archived redo logs to extract transactional changes, enabling real-time CDC. However, with the deprecation of Continuous Mine Mode, organizations have sought more efficient and forward-compatible solutions. To meet this need, Striim introduced Active Log Mining Mode (ALM)—a high-performance, real-time CDC capability built for Oracle 19c and beyond. ALM enables Striim to efficiently mine redo and archive logs without relying on deprecated features, ensuring low-latency, uninterrupted CDC across supported Oracle versions. For organizations seeking a future-proof CDC solution, Striim also offers Oracle OJET—an API-based integration that reads Logical Change Records (LCRs) directly from Oracle. OJET is Oracle’s strategic path forward for CDC, providing robust, enterprise-grade replication with long-term compatibility and official support.

Choosing the Right Oracle CDC Approach

To choose the right CDC method, you’ll need to align your technical strategy with your business goals, budget, and scalability needs.  Striim has developed two CDC adapters for integrating data from Oracle. The first one is an Oracle Reader that captures CDC data using the LogMiner session on the server side. The second is the OJet adapter that uses a high-performing logmining API and offers the best performance for high-scale workloads.  To learn more, check out this performance study which demonstrates the advantages of each adapter option.

The Benefits of Using Oracle CDC

When implemented with a clear strategy, Oracle CDC offers transformational benefits that go far beyond simple data replication. It empowers organizations to:

  • Enable real-time operational visibility for faster decision-making. By streaming every transaction, CDC provides an up-to-the-second view of business operations. This allows leaders to monitor KPIs, detect anomalies, and react to market changes instantly, rather than waiting for end-of-day reports.
  • Support phased and zero-downtime cloud migrations. CDC de-risks one of the most challenging aspects of cloud adoption: data downtime. By keeping on-premises and cloud databases perfectly in sync, businesses can migrate at their own pace without service interruptions, ensuring a smooth and seamless transition.
  • Streamline data ingestion for analytics, AI, and customer personalization. Feeding fresh, granular data to analytical systems is crucial for competitive advantage. CDC provides a continuous, low-latency stream of data that powers everything from dynamic pricing models and fraud detection algorithms to hyper-personalized customer experiences.

Challenges and Limitations of Change Data Capture

While Oracle CDC is a powerful way to get fresh data into downstream tools and systems, a poorly planned implementation can be risky and hugely costly. Without the right platform and strategy, data teams can run into several major challenges.

Performance Overhead on Source Systems

The Challenge: Trigger-based CDC or inefficient log-mining can place a heavy burden on production OLTP systems, slowing down the applications that the business depends on. This is especially damaging for startups and scaling companies with resource-constrained databases.

How Striim Helps: Striim uses a highly optimized, agentless, log-based CDC method on the source database, ensuring production workloads are not compromised. Striim also supports reading from Oracle ADG (Active Data Guard) or other downstream databases to minimize impact on the primary database.

Complexity of Managing Schema Changes

The Challenge: When the structure of a source table changes (e.g., a new column is added), it’s known as schema drift. These DDL changes can easily break data pipelines, forcing teams to manually intervene to resynchronize systems. This is a common struggle for mid-size and enterprise teams managing evolving applications.

How Striim Helps: Striim offers built-in, automated schema migration services and schema evolution capabilities that automatically detect and propagate schema changes from data source to target, ensuring pipelines remain resilient and data stays in sync without manual effort.

High Licensing and Operational Costs

The Challenge: Native Oracle solutions like GoldenGate come with a hefty price tag, adding a significant licensing burden to any project. This can be a major roadblock for enterprises looking to control the costs of their initiatives.

How Striim Helps: Striim provides a cost-effective solution with scalable pricing and cloud-native architecture, reducing the total cost of ownership (TCO) for real-time data integration.

Lack of Real-Time Observability and Alerting

The Challenge: Many traditional CDC solutions are “black boxes.” Teams often don’t know a pipeline has failed until a downstream report is broken or a user complains about stale data. This is particularly painful for lean IT teams and cloud-first startups that can’t afford to spend hours troubleshooting.

How Striim Helps: Striim provides comprehensive, real-time monitoring dashboards, data validation, and proactive alerting. This gives teams end-to-end observability into their data pipelines, allowing them to identify and resolve issues before they impact the business.

Real-Time AI Model Enablement on Live Enterprise Data Streams

The Challenge: Businesses struggle to apply AI in real time because traditional methods rely on batch processing and siloed systems, causing delays in detecting sensitive data, anomalies, or insights. Integrating AI directly into live data streams to enable instant action remains a complex problem.

How Striim Helps: Striim offers highly performant AI agents that embed advanced AI capabilities directly into streaming pipelines, enabling real-time intelligence and automation:

    • Sherlock AI: Uses large language models to classify and tag sensitive fields on-the-fly.
    • Sentinel AI: Detects and protects sensitive data in real time within streaming applications.
    • Euclid: Enables semantic search and categorization through vector embeddings for deeper analysis.
    • Foreseer: Provides real-time anomaly detection and time series forecasting for predictive monitoring.

By integrating these AI agents seamlessly, Striim empowers organizations to operationalize AI-driven insights instantly, improve data privacy, detect risks early, and make faster, smarter decisions.

Simplify Oracle Change Data Capture With Striim

When it comes to moving data from Oracle systems, Oracle CDC is a trusted approach—but building and managing reliable, scalable pipelines without the right platform is complex, risky, and costly. Manual infrastructure and legacy tools often introduce delays and budget overruns, putting projects at risk before they even start. Striim streamlines Oracle CDC with a comprehensive, agentless platform designed for high-throughput, real-time data integration. Optimized for modern cloud environments, Striim enables you to:

  • Deliver data with sub-second latency using best-in-class, log-based CDC.
  • Process and transform data on the fly through a powerful SQL-based streaming analytics engine.
  • Achieve enterprise-grade observability with real-time monitoring, alerting, and data validation.
  • Securely connect to any cloud platform with extensive, pre-built, scalable integrations.

With Striim, Oracle CDC becomes simpler, faster, and more reliable—empowering your data initiatives to succeed from day one. Ready to learn more? Here’s a few ways to dive in with Striim:

Stop wrestling with brittle pipelines and start building the future of your data infrastructure.
Book a Demo with a Striim Expert or Start Your Free Trial Today

Striim Recognized on FORTUNE’s “2019 Best Workplaces in the Bay Area” List

We are excited to announce that Striim has been recognized as a “Best Workplace” on FORTUNE’s “2019 Best Workplaces in the Bay Area” list.

Striim was selected based on a survey that was created, launched, and evaluated by Great Place to Work, a global people analytics and consulting firm.

The rankings took into account more than 30,000 surveys by employees across the Bay Area, designed to evaluate more than 60 elements of an employee’s job and work environment, including trust in leadership, camaraderie in a team setting, and respect among colleagues. Employee perks and benefits were also factored into the rankings.

This Best Workplaces in the Bay Area recognition is very important to Striim because the rankings were completely driven by employee feedback that Great Place to Work collected and evaluated. Additionally, given the fierce competition of not only attracting, but also retaining the best talent in the Bay Area, having our employees thrive in a culture that the Striim team Striim works so hard to foster is extremely rewarding and indicative that we’re on the right track for employee satisfaction.

Striim scored high across the board in many categories including Justice (100 %), Camaraderie (98%), Integrity (96%), Credibility (96%), and Innovation (96%), just to name a few.

Additionally, according to the survey, the overall Striim employee experience was rated 96%. Other great indications that our employees noted include:

  • “Managers avoid playing favorites.” – 100%
  • “I can be myself around here.” – 100%
  • “When you join the company, you re made to feel welcome.” – 100%
  • “Management is approachable, easy to talk with.” – 98%
  • “People here are given a lot of responsibility.” – 98%

Learn more about what employees had to say about Striim, as well as further information on the company, by reading the full Great Place to Work review.

To learn more about why Striim was included on FORTUNE’s Best Workplaces in the Bay Area, as well as to see the full list of winners, please read our press release, “Striim Named One of the 2019 Best Workplaces in the Bay Area by FORTUNE and Great Place to Work.”

Back to top