Axual Documentation

Welcome to the Axual documentation. Use the menu on the left to navigate through the various sections.

What’s new in Axual 2024.4?

In the winter release of Axual Platform, exciting new features are introduced,

  • Data Masking Feature designed to protect sensitive data by masking specific fields when browsing stream messages.

  • Kafka Streams Feature to improve the Kafka Streams Application journey

  • Consumer Group Reset Feature to control the starting point for message consumption for a specific consumer group within a topic

  • Many improvements and bug fixes improve the stability and usability of the platform and Self-Service.

Read more about the latest platform release in our Release Notes

Support

For urgent production issues, we are on standby 24/7 to answer your questions. Please use the standby number communicated to you.

For non-urgent requests, you’ll find our support options here.

Getting Started

As soon as you understand the basic concepts, and you want to get started building your application, creating a schema, or just want a guide to follow along to, please refer to the Getting started section.

Why Data Streaming is Crucial in Modern Systems Compared to Batch Processing

In today’s world, information is generated everywhere—by people, devices, and applications. From a bank transfer to a car approaching a garage entrance, to an energy meter reporting usage, data is constantly being created.

Picture showing things emitting events

But the real value of this data lies not just in its storage, but in its immediate use. For instance, you want your banking app to show a transaction as soon as it happens, or the garage door to open as you approach.

Traditional approaches like batch processing may be slow, processing data in intervals, which can delay actions.

This is where data streaming becomes essential.

Real-Time Data Utilization

In contrast to older batch processing methods, where data is collected and processed at set intervals, data streaming allows for real-time data processing.

Events are captured and made available instantly. Whether it’s detecting a fraudulent bank transfer or processing an energy bill as soon as the reading is captured, streaming data enables businesses to act when the event is happening, offering a significant advantage over delayed batch processing.

Reducing Redundant Processes

Another key benefit of data streaming is its efficiency in reusing data. In traditional systems, APIs or database queries might be repeatedly called to fetch the same information, leading to inefficiencies and delays. Streaming data ensures that the data is accessible to multiple consumers at the same time, allowing different applications to use it as soon as it’s generated without repeated requests. This asynchronous communication not only speeds up the process but also reduces the load on systems.

Decoupling Producers and Consumers

In a streaming architecture, data producers (like sensors or applications) generate events and push them to the platform, without needing to know who will use the data.

Similarly, consumers (applications or services) can subscribe to data streams they’re interested in and react to those events in real-time.

Picture showing the streaming platform

This decoupling of producers and consumers creates a more flexible system, where data can be used by multiple services without requiring tight integration, as is often necessary in traditional batch systems.

Speaking The Same Language: Schemas

Producers might not care about who is reading their messages; they need to speak in a language understood by the other party, the consumers.

In other words, the messages they produce to a data stream need to adhere to a specific schema, comparable to an API specification or a database table structure.

You might conclude that there isn’t a decoupling between producer and consumer strictly speaking because they need to agree on what schema to use for messages on a particular data stream. This is only partly true; producers are allowed to change schemas considering the backward compatibility of the schemas.

High Availability and Scalability

Data streaming platforms are designed to handle the high velocity and volume of real-time data, offering scalability and stability.

Unlike batch systems that might struggle with spikes in data or system failures, streaming platforms ensure that data is always available and that systems can continue to operate without disruption.

Support for Reactive Patterns

Streaming data naturally supports reactive patterns, where systems respond to events as they happen. This flexibility allows consumers to process events at their own pace, a contrast to the rigid structure of batch processing. As a result, streaming data can often replace traditional request-response communication, creating more efficient, event-driven systems.

In summary, while older approaches like batch processing have their place, data streaming is crucial for modern systems that require real-time, scalable, and flexible data processing. It allows organizations to act instantly on critical data, making their operations more responsive, efficient, and future-proof.