Enabling the Real-Time Enterprise with Data Streaming

In virtually every industry, organizations want to unlock real-time intelligence from disconnected data sources in order to improve business agility and competitiveness. However, legacy batch upload architectures are too rigid to accommodate applications that require continuous data streams for adaptive decision making.

Although a growing number of organizations have placed their data warehouses in the cloud, on-demand storage and compute cycles do not address their data pipeline challenges.

Because batch processes only upload data once a day, twice a day, or even once an hour, their ability to support the data-to-data needs of warehouses and today's real-time BI applications is diminished.

For Further Reading:

The Benefits of Streaming Data Are Contagious

Modernizing a Data Warehouse with Real-Time Functions

Define Your Business Case for Streaming Analytics

Data Streaming: A Disruptive New Approach

To take its place, an innovative new architecture called data streaming has emerged. It was conceived to create real-time data pipelines for implementing data streaming applications powered by open source technologies such as Apache Kafka.

Data streaming creates secure pipelines that stream data in real time from various sources -- notably databases, applications, and APIs -- to cloud data warehouse platforms.

It enables organizations to connect any data source within minutes to Amazon Redshift, Google BigQuery, Snowflake Computing, and other cloud data warehouses.

These real-time pipelines can support dozens of the most popular data sources, including transactional databases (such as Oracle, Microsoft SQL Server, or MySQL), SaaS vendors (such as Salesforce and HubSpot), and direct event streams from a Kafka topic or directly from iOS, Android, or JavaScript.

Organizations in virtually every industry are moving to a real-time data model to be more agile and achieve a competitive edge through faster, better decision making. Data streaming ensures that customers can integrate all their disparate data silos with the cloud provider of their choice.

Using Amazon Redshift, Google BigQuery, Snowflake Computing, and others, enterprises can create a central warehouse for data available from back-end relational databases, online events and metrics, support services, and other internal and external sources. Without centralization, analytics are both piecemeal and siloed. This makes it difficult, if not impossible, to produce real-time intelligence.

Creating this centralized data warehouse is not without its challenges because this data is spread across multiple sources and different systems in different formats. Some of it's flat, some is relational, some is JSON. Instead of writing custom scripts to integrate it all, which is beyond the resources of most companies, data streaming technology can perform these tasks.

Data streaming relieves IT staff from the drudgery of data movement so IT can focus entirely on data analytics. Because data streaming technologies can support a comprehensive set of integrations, enterprises can easily stream and access all their data in the cloud data warehouse of their choice. Every bit of data -- no matter how big or small and regardless of the source or format -- can be moved to the cloud without errors or requiring a team of engineers to write scripts.

Data Streaming Checklist

Here are the key elements to consider when planning a data streaming project.

Flexible data integration: The ability to transport data in the format required to any data warehouse, regardless of whether the data is structured or semistructured, direct or customized, static or changing. This includes sources such as:

  • Transactional databases (e.g. Oracle, PostgreSQL)
  • Salesforce.com (account information, stage, ownership, etc.)
  • Website tracking (all Web event data)
  • Web servers (customer activity such as adding inputs and deploying new code in the code engine)
  • Back-end logs (internal platform events such as data being loaded to the output, new table created)
  • Monitoring systems (to capture system issues such as input connections and latency)

Schema import and schema inference: Expect step-by-step data preconfiguration tools that make it easy to map every field of structured or semistructured data to a table and control how data is loaded into the data warehouse.

Code engine: Data scientists and engineers should be able to write custom code to enrich and cleanse data, create alerts, implement sessionization, and detect anomalies. To eliminate lengthy, high-latency data preparation jobs, all changes should be executed in stream in real time so data reaches its intended destination.

Live monitoring: Real-time visibility into data streams for monitoring behavior, identifying potential discrepancies, and debugging data records saves considerable time and helps you avoid problems. Live monitoring also lets you track incoming throughput, latency, loading rates, and error rates and can generate Web and email alerts.

Pipeline transparency: A dashboard that provides continuous views of data in motion and notifications that allow users to view incoming events, monitor throughput and latency, and identify errors in real time.

Schema management: When data changes, a real-time response is needed to make sure no event is lost. The ability to manage these automatically or generate notifications so changes can be made on-demand is critical.

Subscribe to Upside

Q&A with Jill Dyché

Find out what's keeping teams up at night and get great advice on how to face common problems when it comes to analytic and data programs. From head-scratchers about analytics and data management to organizational issues and culture, we are talking about it all with Q&A with Jill Dyche.

View Recent Article

Submit Your Questions to Jill

Powered by TDWI. Advancing All Things Data
A Division of 1105 Media, Inc.