Central Log Monitoring Solution

SUMMARY

Client
Česká spořitelna
Tech stack
Apache Impala, Spark Streaming, Apache Kafka, Apache Hive, Apache Spark, Hadoop, Java

Česká spořitelna, the Czech arm of Erste Group, sought a tailored, centralised solution to monitor log data across its operational banking systems. We delivered a central log monitoring platform that gives security teams a single, compliant view of activity across the bank. Moving from overnight batches to near-real-time insight means incidents can be analysed within minutes—without changing analysts’ workflows. The platform scales with demand and keeps pace with evolving regulation.

Graphic promoting case study Central Log Monitoring Solution

Results

10,000
logs per second
Compliance
with new cyber security legislation
One platform
for real-time and batch data

Česká spořitelna, the Czech arm of the Erste Group, was looking for a custom solution for the central monitoring of log data from their operational banking systems. The central log collects and stores a massive amount of data – over a specific period of time – to analyse trends or record events from various banking systems plus network and IT environments.

Improvements to their data collection and storing processes had to be made to comply with the new cyber security regulations. But the bank also wanted their in-house security specialists to be able to analyse events as soon as possible after they are logged.

The bank’s central monitoring system was based on batch processing of system logs using a standard relational database. It wasn’t compliant with new legislation as it didn’t collect complete data from every system. Plus, the existing system couldn’t meet the new requirements to retain data for a minimum period of time.

Aside from the compliance issues, this system’s operational mode didn’t allow the use of analytical queries that would include log data from the current date of generation – meaning rapid analysis was impossible.

The solution needed to meet the following specifications:

  • Fulfil requirements of new cyber security legislation
  • Fast and stable data processing 
  • Able to analyse data including freshly logged events 
  • Scalable solution for future data volume increases

We developed a system that can process streams of logging data from all monitored banking systems in real time. Working with the bank’s security, IT, and operations departments, we tailored the solution to meet each of their specific needs.

Logging data is streamed from newly created central storage through Apache Kafka into a new, dedicated system. There, the data is being processed using Apache Spark and Spark Streaming for monitoring purposes. The whole solution can handle tens of thousands of log entries per second.

Thanks to the implementation of Apache Spark, using lambda architecture, we ensured a uniform handling of stream and batch processing of data. The same code is shared for both processing methods. There is no need to maintain two separate codebases, which makes development, deployment, and servicing much simpler and faster. The processed data is stored in Hive tables within a Hadoop cluster. 

Rapid analysis of security data

The unified data overview provides access for specialists in the security department for their analyses. The data is shown in the same format as previously, so no analytical process changes had to be implemented. However, it is now possible to access data within minutes of it being generated.