5 Reasons Low Code Developers Should Join Striim

Low code development is a powerful tool for businesses looking to streamline their processes and improve efficiency. Striim is a low-code platform that provides users with a variety of benefits, including the ability to quickly and efficiently process and analyze data in real time. By joining the Striim community, low-code users can take advantage of the following benefits:

  1. Real-time analytics: One of the key benefits of using Striim is the ability to process and analyze data in real time. This means businesses can gain insights into their operations and make more informed decisions without waiting for data to be processed and analyzed.
  2. Single source of truth with easy integration: Striim is designed to be easy to integrate with a wide range of data sources and systems, including databases, data lakes, and cloud services. This means that businesses can easily connect all of their data sources and gain a complete view of their operations.
  3. High scalability: Striim is highly scalable, which means that it can easily handle large volumes of data. This is particularly useful for businesses experiencing rapid growth or needing to process large amounts of data in real-time.
  4. Community support: By joining the Striim community, users can take advantage of the support and knowledge of other Striim users. This can be particularly valuable for businesses new to low-code development or looking to improve their processes.
  5. Cost-effective: Low code development is a cost-effective solution for businesses. Striim is no exception. It provides businesses with a robust platform that enables them to streamline their operations and gain insights into their data without investing in expensive development resources.

In conclusion, Striim is a low-code platform that provides businesses with a powerful real-time tool for processing and analyzing data. By joining the Striim community, low-code users can take advantage of benefits such as real-time data processing, easy integration, high scalability, community support, and cost-effectiveness. These benefits can help businesses improve their operations and gain insights into their data, which can lead to increased efficiency and better decision-making.

Join the discussion: https://community.striim.com/

A Guide to Data Contracts

Companies need to analyze large volumes of datasets, leading to an increase in data producers and consumers within their IT infrastructures. These companies collect data from production applications and B2B SaaS tools (e.g., Mailchimp). This data makes its way into a data repository, like a data warehouse (e.g., Redshift), and is shown to users via a dashboard for decision-making.  

This entire data ecosystem can be wobbly at times due to a number of assumptions. The dashboard users may assume that data is being transformed the same way as when the service was initially launched. Similarly, an engineer might change something in the schema of the source system. Although it might not affect production, it might break something in other cases.

Data contracts tackle this uncertainty and end assumptions by creating a formal agreement. This agreement contains a schema that describes and documents data, which determines who can expose data from your service, who can consume your data, and how you can manage your data. 

What are data contracts?

A data contract is a formal agreement between the users of a source system and the data engineering team that is extracting data for a data pipeline. This data is loaded into a data repository — such as a data warehouse — where it can be transformed for end users. 

As per James Densmore in Data Pipelines Pocket Reference, the contract must include a number of things, such as:

  • What data are you extracting?
  • What method are you using to extract data (e.g., change data capture)?
  • At what frequency are you ingesting data? 
  • Who are the points of contact for the source system and ingestion?

You can write a data contract in a text document. However, it’s better to use a configuration file to standardize it. For example, if you are ingesting data from a table in Postgres, your data contract could look like the following in JSON format:

“{
  ingestion_jobid: "customers_postgres",
  source_host: "ABC_host.com",
  source_db: "bank",
  source_table: "customers",
  ingestion_type: "full",
  ingestion_frequency_minutes: "15",
  source_owner: "developmentteam@ABC.com",
  ingestion_owner: "datateam@ABC.com"
};”

How to implement data contracts

When your architecture becomes distributed or large enough, it’s increasingly difficult to track changes, and that’s where a data contract brings value to the table. 

When your applications access data from each other, it can cause high coupling, i.e., applications are highly interdependent on each other. If you make any changes in the data structure, such as dropping a table from the database, it can affect the applications that are ingesting or using data from it. Therefore, you need data contracts to implement versioning to track and handle these changes. 

To ensure your data contracts fulfill their purpose, you must:

  • Enforce data contracts at the data producer level. You need someone on the data producer side to manage data contracts. That’s because you don’t know how many target environments can be used to ingest data from your operational systems. Maybe, you first load data into a data warehouse and later go on to load data into a data lake. 
  • Cover schemas in data contracts. On a technical level, data contracts handle schemas of entities and events. They also prevent changes that are not backward-compatible, such as dropping a column. 
  • Cover semantics in data contracts. If you alter the underlying meaning of the data being generated, it should break the contract. For instance, if your entity has distance as a numeric field, and you start collecting distance in kilometers instead of miles, this alteration is a breaking change. This means that your contract should include metadata about your schema, which you can use to describe your data and add value constraints for certain fields (e.g., temperature).
  • Ensure data contracts don’t affect iteration speed for software developers. Provide developers with familiar tools to define and implement data contracts and add them to your CI/CD pipeline. Implementing data contracts can minimize tech debt, which can positively impact iteration speed. 

In their recent article, Chad Sanderson and Adrian Kreuziger shared an example of a CDC-based implementation of data contracts. According to them, a data contract implementation consists of the following components, as depicted below:

  1. Defining data contracts as code using open-source projects (e.g. Apache Avro) to serialize and deserialize structured data.
  2. Data contract enforcement using integration tests to verify that the data contract is correctly implemented, and ensuring schema compatibility so that changes in the producers won’t break downstream consumers. In their example, they use Docker compose to spin up a test instance of their database, a CDC pipeline (using Debezium), Kafka, and the Confluent Schema Registry.
  3. Data contract fulfillment using stream processing jobs (using KSQL, for example) to process CDC events and output a schema that matches the previously-defined data contract.
  4. Data contract monitoring to catch changes in the semantics of your data.
A data contract implementation,
A data contract implementation, from Chad Sanderson and Adrian Kreuziger’s “An Engineer’s Guide to Data Contracts – Part 1

Data contract use cases

Data contracts can be useful in different stages, such as production and development, by acting as a validation tool, as well as supporting your data assets like data catalogs to improve data quality. 

Assess how data behaves on the fly

During production, you can use data contracts as a data validation tool to see how data needs to behave in real time. For example, let’s say your application is collecting data for equipment in a manufacturing plant. Your data contract says that the pressure for your equipment should not exceed the specified limit. You can monitor the data in the table and send out a warning if the pressure is getting too high.  

Avoid breaking changes 

During software development, you can use data contracts to avoid breaking changes that can cause any of your components to fail since data contracts can validate the latest version of data. 

Improve discoverability and data understanding

Like data contracts, data catalogs accumulate and show various types of information about data assets. However, data catalogs only define the data, whereas data contracts define the data and specify how your data should look. Moreover, data catalogs are made for humans, whereas data contracts are made for computers. Data contracts can be used with data catalogs by acting as a reliable source of information for the latter to help people discover and understand data through additional context (e.g., tags). 

Striim helps you manage breaking changes

Striim Cloud enables you to launch fully-managed streaming Change Data Capture pipelines, greatly simplifying and streamlining data contract implementation and management. With Striim, you can easily define, enforce, fulfil, and monitor your data contracts, without having to wrangle with various open-source tools.

For example, using Striim, you can set parameters for Schema Evolution based on internal data contracts. This allows you to pass schema changes to data consumers on an independent table-specific basis. If your data contract is broken, you can use Striim to automate sending alerts on Slack.. Consider the workflow in this following diagram:

Data contracts with schema evolution parameters

You can use Striim to move data from a database (PostgreSQL) to a data warehouse (BigQuery). Striim is using Streaming SQL to filter tables in your PostgreSQL based on data contracts. If a schema change breaks your contract, Striim will stop the application and send you an alert through Slack, allowing your engineers to stop the changes in your source schema. If the schema change is in line with your contract, Striim will automatically propagate all the changes in your BigQuery. 

Learn more about how Striim helps you manage data contracts here.

 

Introducing the Striim Community and Discord Server

As a data architect, business intelligence professional, or Chief Technical Officer, you know how important it is to have access to real-time data streaming to make the most informed decisions for your organization. That’s where Striim comes in.

One of the biggest benefits of using Striim is the ability to easily integrate with a variety of data sources, including databases, message queues, data warehouses, sensors, and files. This allows you to collect and stream data from a wide range of sources, providing a comprehensive view of your organization’s data.

But, as a busy professional, you may be wondering how you can stay up-to-date on the latest developments and best practices in the world of data streaming. That’s where the Striim Community and Discord come in.

By joining the Striim Community, you’ll have access to a wealth of knowledge and resources from other professionals using Striim to stream data in real-time. You can ask questions, share your experiences, and learn from others facing similar challenges.

The Striim Discord server is another great resource for staying connected with the Striim community. Here, you can join discussions and participate in live chats with other Striim users. You can also access support from Striim experts and get answers to your technical questions.

In addition to the knowledge and support you’ll gain from the Striim Community and Discord, there are many other benefits to using Striim for data streaming. For example, Striim’s built-in machine-learning capabilities allow you to analyze data streams in real-time, providing valuable insights and helping you make more informed decisions.

Striim also offers a low-code development environment, making it easy for non-technical users to build and deploy data streaming applications. This can save your organization time and resources, allowing you to quickly and easily implement data streaming solutions.

Overall, using Striim for data streaming offers a wide range of benefits for data architects, business intelligence professionals, and Chief Technical Officers. By joining the Striim Community and accessing the resources on the Striim Discord server, you can stay up-to-date on the latest developments and best practices, gain valuable support and insights from other professionals, and leverage the powerful features of Striim to stream data in real-time. Click below to get your free invite code.

Discord Invite Code

Join the Discussion!

 

Striim Cloud on AWS: Unify your data with a fully managed change data capture and data streaming service

Businesses of all scales and industries have access to increasingly large amounts of data, which need to be harnessed effectively. According to an IDG Market Pulse survey, companies collect data from 400 sources on average. Companies that can’t process and analyze it to glean useful insights for their operations are falling behind.

Thousands of companies are centralizing their analytics and applications on the AWS ecosystem. However, fragmented data can slow down the delivery of great product experiences and internal operations.

We are excited to launch Striim Cloud on AWS: a real-time data integration and streaming platform that connects clouds, data and applications with unprecedented speed and simplicity.

With a serverless experience to build smart data pipelines in minutes, Striim Cloud on AWS helps you unify your data in real time with out-of-the box support for the following targets: 

  • AWS S3
  • AWS Databases on RDS and Aurora 
  • AWS Kinesis
  • AWS Redshift
  • AWS MSK
  • Snowflake
  • Databricks with Delta Lake on S3

along with over 100 additional connectors available at your fingertips as a fully managed service.

Striim Cloud runs natively on AWS services like EKS, VPC, EBS, Cloudwatch, and S3 enabling it to offer infinite large-scale, high performance, and reliable data streaming. 

How does Striim Cloud bring value to the AWS ecosystem?

Striim enables you to ingest and process real-time data from over one hundred streaming sources. This includes enterprise databases via Change Data Capture, transactional data, and AWS Cloud environments. When you run Striim on AWS, it lets you create real-time data pipelines for Redshift, S3, Kinesis, Databricks, Snowflake and RDS for enterprise workloads. 

Sources and targets

Striim supports more than 120+ sources and targets. It comes with pre-built data connectors that can automate your data movement from any source to AWS Redshift or S3 within a few minutes. 

With Striim, all your team needs to do is to hit a few clicks for configuration, and an automated pipeline will be created between your source and AWS targets. Some of the sources Striim supports include:

  • Databases: Oracle, Microsoft SQL Server, MySQL, PostgreSQL, etc.
  • Data Streams: Kafka, JMS, IBM MQ, Rabbit MQ, IoT data over MQTT
  • Data formats: JSON, XML, Parquet, Free Form Text, CSV, and XML
  • AWS targets: RDS for Oracle, RDS for MySQL, RDS for SQL Server, Amazon S3, Databricks via Delta Lake on S3, Snowflake, Redshift, and Kinesis
  • Additional targets: Over 100 additional connectors including custom Kafka endpoints with Striim’s full-blown schema registry support

Change data capture

Change data capture (CDC) is a process in ETL used to track changes to data in databases (e.g., insert, update, delete) and stream those changes to target systems like Redshift. However, CDC approaches like trigger-based CDC or timestamps can affect the performance of the source system. 

Striim supports the latest form of CDC — log-based CDC — which can reduce overhead on source systems by retrieving transaction logs from databases. It also helps move data continuously in real time in a non-intrusive manner. Learn about log-based CDC in detail here.                   

Streaming SQL

Standard SQL can only work with bounded data that are stored in a system. While dealing with streaming data in Redshift, you can’t use standard SQL because you are dealing with unbounded data, i.e., data that keep coming in. Striim provides a Streaming SQL engine that helps your data engineers and business analysts write SQL-style declarative queries over streaming data. These queries never stop running and can continuously produce outputs as streams.  

Data transformation and enrichment 

Data transformation and enrichment are critical steps to creating operational data products in the form of tables and materialized views with minimal cost and duplication of data. To organize these data into a compatible format for the target system, Striim helps you perform data transformation with Streaming SQL. This can include operations such as joining, cleaning, correlating, filtering, and enriching. For example, enriching helps you to add context to your data (e.g., by adding geographical information to customer data to understand their behavior). 

What makes Striim unique in this regard is that it not only supports data transformation for batch data, but it also supports in-flight transformations for real-time streams with a full blown Streaming SQL engine called Tungsten. 

Use case: How can an apparel business analyze data with Striim? 

Suppose there’s a hypothetical company, Acme Corporation, which sells apparel across the country. The management wants to make timely business decisions that can help them to increase sales and minimize the number of lost opportunities due to delays in decision-making. Some of the questions that can help them to make the right decisions include the following: 

  • Which product is trending at the moment?
  • Which store and location received the highest traffic last month?
  • What’s the inventory status across warehouses?

Currently, all store data is stored in a transaction database (Oracle). Imagine you’re Acme Corporation’s data architect. You can generate and visualize answers to the above questions by building a data pipeline in two steps:

  1. Use Striim Cloud Enterprise to stream data from Oracle to Amazon Redshift. 
  2. After data is loaded into Redshift, use Amazon QuickSight service to show data insights and create dashboards. 

Here’s how the flow will look: 

Striim Cloud AWS

In this blog, we will show you how you can configure and manage Striim Cloud Enterprise on AWS to create this pipeline for your apparel business within a few minutes. 

Sign up for Striim Cloud

Signing up for Striim Cloud Enterprise is simple: just visit striim.com, get a free trial and sign up for the AWS solution. Activate your account by following the instructions. 

Sign up for Striim on AWS

 

Once you are signed in, create a Striim Cloud service, which essentially runs in the background and creates a dedicated Kubernetes cluster (EKS service on AWS) to host your pipeline, as you can see in the picture below. 

 

Kubernetes cluster

 

Once the cluster is ready and before launching your service, configure secure connections using the secure SSH connection configuration, as seen below.

 

SSH connection

Create a pipeline for Oracle to Amazon Redshift

To create a pipeline, simply type Source: Oracle and target Amazon to see all the supported targets. In our example, we are selecting Amazon S3 as our target. This could be Amazon Redshift, Kinesis, etc.

S3 target

The wizard will help you walk through the simple process with source and target credentials. The service automatically validates the credentials, connects to the sources, and fetches the list of schemas and tables available on the sources for your selection, as shown below.

wizard 1

 

wizard 2

On the target side, enter Amazon Redshift Access Key and secret key with the appropriate S3 bucket name and Object names to write Oracle data into, as depicted in the image below.

New target

Follow the wizard to finish the configuration, which creates a data pipeline that collects historical data from the Oracle database and moves them to Amazon Redshift. For example, you can see the total number of sales across all branches during the last week.  

In the next step, you can create an Oracle CDC pipeline via Striim to stream real-time change data coming in from different stores into Oracle to Redshift. Now, you can see real-time store data.

Pipeline to Amazon RedShift
A data pipeline streaming data from the source (Oracle) to Amazon RedShift

A data pipeline streaming data from the source (Oracle) to Amazon RedShift

Your data engineers can use streaming SQL to join, filter, cleanse, and enrich data on the real-time data stream before it’s written to the target system (S3). A monitoring feature offers real-time stream views for further low-latency insights. 

Once data becomes available on Redshift, your data engineer can create dashboards and set up metrics for the relevant business use cases such as:

  • Current product trending
  • Store and location with the highest traffic last month
  • Inventory status dashboard across warehouses; quantity sold by apparel, historic graph vs. latest (last 24 hours)

Data like current trending products can be easily shared with management for real-time decision-making and the creation of business strategies.

For example, here’s a real-time view of the apparel trends by city:

Apparel trends by city

And below are insights on the overall business, where you can see the top-selling and bottom-selling locations. The management can use this information to try out new strategies to increase sales in the bottom-selling locations, such as by introducing discounts or running a more aggressive social media campaign in those locations.

Striim is available for other cloud environments, too

Like AWS, Striim Cloud is available on other leading cloud ecosystems like Google Cloud and Microsoft Azure. You can use Striim with Azure to move data between on-premises and cloud enterprise sources while using Azure analytics tools like Power BI and Synapse. Similarly, you can use Striim with Google Cloud to move real-time data to analytics systems, such as Google BigQuery, without putting any significant load on your data sources. 

Learn more about them here and here.

 

Real-Time Healthcare Analytics: How Leveraging It Improves Patient Care

On a Tuesday night, a nurse in the emergency department receives a real-time alert on her smartphone: the department will be overcrowded within 1.5 hours. This alert, powered by real-time healthcare analytics, projects bed occupancy and anticipated care needs, allowing the nurse to coordinate with transport, radiology, and lab teams to prepare for the surge.

Historically, data silos limited information access, but real-time analytics now makes healthcare processes more connected. By aggregating and analyzing data, these insights boost operational efficiency and enhance patient care. In this post, we’ll explore how leveraging real-time healthcare analytics ensures seamless patient care and a smoother workflow for your team.

Why Leverage Real-Time Healthcare Analytics? 

There are several compelling reasons why real-time healthcare analytics is essential for healthcare institutions. These include: 

To Analyze EHR Data and Improve Patient Care

An electronic health record (EHR) digitally stores patient information, such as medical history, prescriptions, lab results, and treatments. While EHRs collect and display data, they lack real-time analysis capabilities — a gap filled by real-time healthcare analytics.

With real-time analytics, medical professionals can instantly access insights and recommendations based on current EHR data. This system ingests relevant data points, like progress and nursing notes, identifies diagnostic patterns, detects minor condition changes, and prioritizes patients with deteriorating health, enabling swift and proactive care.

Leveraging real-time healthcare analytics is essential in early sepsis detection. According to the CDC, sepsis claims 350,000 adult lives annually in the U.S. Early detection is vital yet challenging due to symptom overlap with other conditions. However, real-time analytics combined with AI can improve sepsis detection rates by up to 32%, according to one report. 

The Medical University of South Carolina (MUSC)  uses this technology to monitor patient health continuously, drawing on EHR data and machine learning to classify signs of sepsis onset. This proactive approach enables timely intervention, potentially saving lives, due to real-time data

To Encourage People to Take a Proactive Approach to their Health

Another popular use case of real-time analytics in healthcare includes smartwatches and fitness trackers. Devices from the likes of Apple, Samsung, Fitbit, and others have exploded in popularity in recent years, enabling people to monitor their own health and adopt healthier habits. 

They help people walk more by tracking their daily step count via in-app challenges, calculate the calories they lose during workouts and sports activities, and monitor their daily caloric intake. These wearables collect data from their sensors and use real-time analytics to provide useful insights. 

While these devices are far from replacements for a doctor visit, they might alert the user to potential health risks. If someone notices their heart rate is often too high/too low, they may be more likely to visit their physician to check in. 

For instance, a 12-year-old girl was alerted by her Apple Watch that she had an unusually high heart rate, and promptly sought medical attention. She was taken to a healthcare facility where doctors found her suffering from a rare condition in children: a neuroendocrine tumor on her appendix.

To Manage the Spread of Disease 

Real-time analytics in healthcare can also help healthcare institutions and doctors identify trends in regards to the spread of illness. For instance, in 2020 during the Covid-19 pandemic, healthcare institutions leveraged real-time analytics to identify the growing disease. Healthcare organizations used machine learning algorithms fueled by data to analyze trends from 50 countries with the highest rates of Covid-19 and predict what would happen in the next several days. 

Healthcare providers also leveraged real-time analytics in healthcare to determine how fast the virus was spreading in real time and how it mutated under various conditions. For example, the EU launched a software in 2020, InferRead, that collected image data from a CT scanner to analyze whether lungs were damaged due to a COVID infection. This analysis was generated within a few seconds, allowing a doctor to study it and diagnose the patient quickly. 

Real-time analytics can also help to manage resources in the case of an outbreak. In the US, the Kinetica Active Analytics Platform was used to create a real-time analytics program for aggregating and tracking data. The purpose of this program was to aid emergency responders by collecting information on test kit quantities, personal protective equipment (PPE) availability, and hospital capacity. This allowed decision-makers to determine whether they could redirect patients to a hospital with capacity or set up alternative triage centers. Similarly, these insights also helped to distribute PPE to the locations where it was needed most, especially when a shortage made access more difficult. 

To Optimize Hospital Staff Allocation 

Healthcare institutions often face the critical challenge of maintaining optimal staffing levels. Leveraging real-time healthcare analytics can transform how hospitals predict staffing needs by analyzing historical data and identifying patterns in staffing operations. By continuously examining how nurses and other staff operated under varying circumstances, real-time analytics generates recommendations for each hour, considering potential unforeseen scenarios. This ensures that patients receive an appropriate level of care, minimizing resource gaps and elevating the standard of patient care.

Intel’s recent paper highlights how real-time healthcare analytics enables four hospitals to use data from diverse sources to forecast admissions accurately. By applying time series analysis — a statistical technique designed to identify patterns within admission records — these hospitals can predict patient arrivals hour by hour, optimizing preparation and resource allocation. 

Additionally, data insights from real-time analytics empower healthcare institutions to enhance job satisfaction and reduce turnover. By identifying the percentage of experienced staff open to emergency shifts or overtime with incentives, healthcare providers can better manage workloads and redistribute tasks to prevent burnout.

Improve Patient Care and Operational Efficiency with Striim 

For healthcare organizations aiming to optimize real-time healthcare analytics, Striim 5.0 offers a robust, secure solution. The platform not only ingests and analyzes high volumes of data in real-time but also introduces AI agents Sentinel and Sherlock to protect sensitive patient information. This feature automates authentication and connection processes, reducing overhead, enhancing data security, and ensuring compliance by masking personally identifiable information.

Discovery Health achieved a remarkable transformation with Striim, slashing data processing times from 24 hours to seconds. By replacing daily ETL processes with Striim’s Change Data Capture (CDC) technology, the organization seamlessly integrated disparate systems, eliminating delays and enabling faster, more responsive decisions. This innovation improved efficiency, reduced costs, and fostered personalized engagement by leveraging predictive analytics to encourage healthier member choices.

Backed by Oracle, Striim delivered unmatched reliability and scalability, utilizing advanced logical database replication expertise. The platform’s real-time insights empowered Discovery Health to promote wellness, enhance health outcomes, and streamline workflows. Through ongoing optimization, Discovery Health revolutionized its data infrastructure, driving informed decision-making and elevating customer experiences on a global scale.

Another healthcare organization that leverages Striim is Boston Children’s Hospital. In addition to enhancing patient outcomes, improving operational efficiency is critical to success in healthcare organizations. By consolidating data from multiple systems, including patient, billing, scheduling, clinical, and financial information, hospitals can streamline their operations and make faster, data-driven decisions.

Striim’s platform enables near real-time and batch-style processing of data from diverse sources like MS SQL Server, Google BigQuery, and Oracle, all feeding into a centralized Snowflake data warehouse. This seamless integration reduces the need for various scripts and disparate source systems, providing a single, cohesive view of the data pipelines. The hospital has not only saved time and money on support resources but has also significantly reduced the time it takes to deliver actionable insights to business users, a crucial factor in the fast-paced healthcare industry.

Ready to see for yourself how Striim can streamline operations and improve patient outcomes? Get started with a demo today.

How Striim Extends Azure Synapse Link

We recently announced that Striim is a participant in Microsoft’s Intelligent Data Platform partner ecosystem. We’re also excited to share that Striim extends Synapse Link to add support for additional source systems. 

There’s no question about the benefits of Azure Synapse. Whether it’s around on-demand usage, the ability to reduce high CapEx projects and increase cost savings, or enabling insight-driven decisions as fast as possible, Synapse can be an integral piece to your digital transformation journey. However, in order to make the most of Synapse and Power BI you need to reliably ingest data from disparate sources in real time. 

In order to do so, Azure introduced Synapse Link, a method of easily ingesting data from Cosmos DB, SQL Server 2022, SQL DB, and Dataverse. Synapse Link utilizes either the change feed or change tracking to support continuous replication from the source transactional system. Rather than relying on legacy ETL tools to ingest data into Synapse on a nightly basis, Synapse Link enables more real-time analytical workloads with a smaller performance impact on the source database. 

Outside of the sources included today with Synapse Link, Microsoft partnered with Striim to add support for real-time ingestion from Oracle and Salesforce to Synapse. Striim enables real-time Smart Data Pipelines into critical cloud services via log-based change data capture (CDC). CDC is the least intrusive method of reading from a source database, which reads from the underlying transaction logs rather than the database itself – empowering replications of high-value business-critical workloads to the cloud with minimal downtime and risk. 

Besides pure data replication use cases, one common pattern that we see is the requirement to pre-process data in flight before even landing on Synapse. This reduces the time to value, and gets the data in the right format ahead of time. Within Striim it’s incredibly easy to do so either with out-of-the-box transformations, SQL code, or even Java for the most flexibility. 

Whether you’re interested in replication or Smart Data Pipelines, to learn more please watch the free joint webinar: https://info.microsoft.com/ww-ondemand-unlock-insights-to-your-data-with-azure-synapse-link.html?lcid=en-us, or download our Oracle to Synapse or Salesforce to Synapse reference architectures.

If you have any questions please reach out to microsoft@striim.com, we’d be happy to discuss your specific use case in more detail.

Making Data Intelligent with Microsoft

Microsoft Cloud Architecture

I am excited to share with you that Striim is a proud participant in the Microsoft Intelligent Data Platform partner ecosystem as announced at Microsoft Ignite 2022. We have a history of working with Microsoft to help provide our mutual customers with access to enhanced data insights in real time, allowing them to make decisions the moment data is created.  Earlier this year, Striim announced an enhanced and growing relationship with Microsoft around our Striim Cloud software-as-a-service offering. 

More specifically, we said:

Microsoft customers can now leverage Striim Cloud on Microsoft Azure for continuous, streaming data integration from on-premises and cloud enterprise sources to Azure Synapse Analytics and Power BI, taking full advantage of the Microsoft Intelligent Data Platform

The Microsoft Intelligent Data Platform promise is for customers to accelerate innovation and get more from their data, to increase their agility as environments and regulations evolve, and to do it all on an open and governed platform. This is why Striim is so pleased to partner with Microsoft to further accelerate customers’ time to value.

At Striim we continue to offer customers the ability to take full advantage of the Microsoft Intelligent Data Platform. We provide Microsoft Azure enterprise customers immediate access to critical business data in real time. With continuous, streaming data integration from on-premises and cloud enterprise sources to Azure analytics tools like Synapse and Power BI, users have an unbeatable, data-driven experience with up-to-the-second operational visibility. 

We will continue to grow and enhance our relationship with Microsoft and the Intelligent Data Platform as we work toward the mutual goal of providing customers the ability to, adapt rapidly, add layers of intelligence to apps, generate predictive insights, and govern all your data—wherever it resides. As Faisal Mohamood, VP, Azure Data Integration at Microsoft says, “We look forward to continuing to grow our partnership with Striim to help our customers accelerate their data-driven digital transformation.”

Learn More: Microsoft Intelligent Data Platform Parter Ecosystem announcement

How Can Real-Time Customer Analytics Lead To More Optimized and Refined Customer Experiences?

Modern-day customers have higher expectations from the brands they interact with. They crave customer experiences that are more timely, targeted, and personalized to their needs. Brands can meet these expectations by integrating real-time analytics into their customer experience. According to a study from Harvard Business Review, 44% of organizations found the adoption of real-time customer analytics to increase their total number of customers and revenue. 

With data analytics, companies can deliver a real-time experience across all touchpoints that’s guided by data-driven insights. This way, they can have a competitive advantage that helps them to sell more, engage more effectively with customers, and boost customer retention. 

happy customer due to personalized customer experience

Create Instant Personalized Experiences 

A study found that nearly 80% of consumers are more likely to remain loyal to a company if it uses personalization. Real-time customer analytics pave the way for real-time personalization — the capability to show contextually relevant information to each prospect or customer at any touchpoint. 

This involves going through showing recommendations or offers based on customers’ historical data or their buyer persona. There are three types of customer data you can collect: 

  • First-party data: Data generated by your organization’s owned channels, such as websites or mobile apps.
  • Second-party data: Data collected from another company that collected data through its owned channels. Examples include customer feedback or surveys. 
  • Third-party data: Data bought and collected from large data aggregators that have collected data from other sources and don’t have any direct relationship with the data. You can collect this data from data management platforms (DMPs) or demand side platforms (DSPs). 

Note: Over the last few years, there have been a lot of privacy regulations that make third-party data hard to collect

You have to ingest real-time customer data from several sources (e.g., CRM platforms), analyze that data in a data warehouse or data lake to look for trends and patterns, and prescribe the most suitable response. Some examples include:

  • Geo-targeting: Tailor a message according to your visitor’s physical location. 
  • Weather: Adapt your offerings based on the weather in the visitor’s surroundings. 
  • Channel: Adjust messaging based on the channels used by visitors, such as paid search or organic search.

Build Smarter Loyalty Programs

A KPMG study found that 86% of loyal customers are likely to recommend a brand to their friends and family. Customer loyalty comes from an emotional connection between the brand and the customer. With real-time analytics, you can strengthen this bond. One way to do this is by redeeming loyalty rewards in real time. Loyalty programs generate large amounts of data that can be used to discover plenty of useful insights. 

Customers can receive loyalty coupons and rewards as soon as they buy something, whether they are standing next to a PoS system or completing the purchase online. Customers are notified of the rewards via email, in-app notifications, or SMS. 

customer loyalty program

Here’s how real-time customer loyalty analytics work: 

  1. Purchase history data is collected: As soon as a customer buys something, their data is sent to a repository like a data warehouse, where their purchase history is stored. 
  2. Customers are sorted into buyer categories: Machine learning models go through data to categorize customers into clusters for categories that fit each customer in real time. For instance, regular customers, occasional customers, and first-time buyers are assigned to different categories, with these categories further split based on demographic and behavioral data. 
  3. Assign the right loyalty reward for each category: Based on historical data, the system can assign the relevant loyalty strategy to each category, such as bonus points, free product, priority service, or any other type of reward. 

Similarly, you can also use real-time analytics to identify stores where loyalty programs fail to generate desired results. This gives you an opportunity to improve your marketing based on the demographic and behavioral data for a store or location. For example, you might have a store in an area with more young customers than your other branches, where your existing models might not work. In that case, you can experiment with different loyalty strategies to reward customers in real time and assess their engagement with the brand to enhance your model. 

Improve the Customer Journey

You can use customer journey analytics to analyze how customers feel across every touchpoint. This can help you to diagnose issues in real time, figure out a way to resolve them quickly, and prioritize improvements that have a bigger impact on your overall customer experience. 

customer journey analytics

Here’s how real-time customer journey analytics work: 

  1. You collect customer information via your relationship with your customers online and offline. This can include information from sources like email and CRMs. This data mainly informs you how your consumers are coming to your website, what type of content they search, and other ways they engage with your brand. 
  2. From the customer data you collected, you can use customer segmentation to create buyer personas and understand the preferences and expectations of your customers. 
  3. Analyze this data with advanced analytics tools. The purpose of this analysis is to identify which sources lead users to your website, how they navigate it, how they engage with your website, and what bottlenecks are disrupting them. Data can be analyzed via different techniques, such as time series analysis or Monte Carlo simulation. 
  4. Once you get the insights, the system also generates possible courses of action. 

The following reports can help you work on your customer journeys: 

  • Journey reports: Visualize customer touchpoints for each channel via graphs and charts. For example, if the company gets lower engagement on social media than its competitors, real-time analytics can suggest using a different strategy, such as chatbots, to make customer service operational 24/7. 
  • Attribution reports: Identify successful customers and trace actions that led them to convert or upgrade. For example, it can tell you how to revamp your site or optimize the payment or sign-up processes for better conversions by reducing the steps that lead to them. 
  • Cohort reports: Sum and show a customer’s activities over a specific period. For instance, if the usage of one of the company’s main features is continuously declining, the company can revamp it by working on the reported issues (e.g., UI issues). 

Start Small With Real-Time Customer Analytics Through Striim

The implementation and deployment of real-time analytics can take considerable effort, time, and resources. This can sometimes discourage companies from adopting it. Therefore, start small by focusing on a few use cases and prioritize them. For this purpose, Striim can be a strategic ally that can address your concerns by showing how you can build a real-time infrastructure that can improve your customer experience.  

Striim can help you:

  • Combine sources and channels in one place
  • Create event-driven experiences
  • Connect third-party data sources
  • Predict customer intent and behavior

Find out more here.

Striim Achieves SOC 2 Type II Certification

At Striim we recognize the essential role that our software plays in the data architecture of our customers. Our unified real-time data integration and stream processing platform and our fully managed SaaS data products in Striim Cloud, are the vital engines that drive the data for many mission critical applications. Our customers need to trust us, and our software, to be secure and available.

Nine months ago we announced our SOC 2 Type I certification. To further this trust, we are very excited to announce that Striim has now achieved SOC 2 Type II certification.

A SOC 2 assessment report provides detailed information and assurance about an organization’s security, confidentiality, availability, processing integrity, and/or privacy controls, based on their assurance of compliance with the American Institute of Certified Public Accountants (AICPA) Trust Services Principles and Criteria for Security. A SOC 2 report is often the primary document that the security departments of our customers will rely on to assess Striim’s ability to maintain adequate security, and reviewing such documents is itself often required by SOC 2 controls.

SOC 2 compliance comes in two forms: the SOC 2 Type I report which describes the design of the controls we have in place to meet relevant trust criteria at specific point in time; and a SOC 2 Type II report which details the operational effectiveness of those controls over a specified period of time. These reports are the results of audits performed by independent third parties, in our case Grant Thornton LLP.

We completed SOC 2 Type I last year, and successfully operated the controls for a period of nine months in order to become SOC 2 Type II certified. The Controls that the audit covers include Striim as a corporation, our on-premise Platform and the Striim Cloud managed SaaS offering. They cover infrastructure, software, devices, people, data, and our corporate and customer policies, procedures and processes.

To achieve this certification, we relied on the investments we made for SOC 2 Type I certification in defining processes, policies and procedures, as well as training and utilization of technologies. Continual internal audits ensured we were meeting our goals and not straying from the many controls we have in place. This required the continual efforts of a cross functional team including contributions from executive management, security, human resources, engineering, infrastructure and legal departments.

SOC 2 is not just a certification, and it is not something you do once just to gain a check mark. The annual audits require that the controls and processes around them are ingrained into the DNA of every Striimer, and the insight gained during the process is a stepping stone to other broader and industry specific certifications.

This is just the start of our journey, so stay tuned for further exciting updates. The SOC 2 Type II report is available on request for our customers and those in the process of evaluating Striim.

Back to top