Kafka Security Fundamentals— The Rosetta Stone to your event-streaming infrastructure

Rick Osowski
13 min readApr 27, 2021

--

Getting Started

Apache Kafka project official logo

Understanding Apache Kafka security is much like discovering the Rosetta Stone to the entire event-streaming landscape. This is not meant to be hyperbole at all, since once you understand a few common core configurations, you have the ability to connect any Kafka-based, event-driven application to dozens of Kafka offerings with relatively few changes. The vast majority of these changes come in the form of Kafka broker endpoints, as well as authentication configuration information in the form of protocol selection and providing credentials in some fashion. This blog post will cover the six most important Kafka security configuration options from a Kafka client perspective (based on my team’s experiences so far) to allow you to become as flexible as a Swiss Army Knife when working with any Kafka offering.

As this is a deeper dive into Apache Kafka and the capabilities it provides, I will not be covering any introductory Kafka material here. You can head over to kafka.apache.org for all the foundational information necessary or check out my friend Whitney’s What is Kafka? video (embedded below) to learn everything you need to get started in under 10 minutes. Once you’ve read through everything here, if you still have questions or want to dig in further, the Security section of the official Kafka documentation covers everything in depth with a great deal of clarity — so that would be my first recommendation if you read through here to find that none of the common configurations below match your expected environment setups!

Before we jump in, there are two caveats to address on the scope of this blog series:

  1. There are many, many ways to configure Kafka security from a Kafka administrator point-of-view. The focus of this short blog series is to more effectively comprehend Kafka security as a Kafka client (producer or consumer) and not a Kafka administrator (operations / provider). For a deeper understanding of practical Kafka operational security from an administrator’s point of view (beyond the aforementioned Security docs above), you can reference Confluent’s Security Overview for a much deeper understanding on how on to securely deploy and maintain Kafka well into the future!
  2. There are two different patterns for authenticating with Kafka. The one we will address in this first post focuses on username and password-based authentication mechanisms. The other, which we will cover in a follow-up post, focuses on using SSL certificates for authentication, above and beyond their use in traditional encryption mechanisms. This latter pattern is traditionally referred to as mutual TLS-based authentication or mTLS, for short. If your current Kafka scenario doesn’t provide you with user-based credentials, don’t worry, we’ll cover that soon. The majority of this blog post will still hold true, with a few changes to the core six configuration properties covered below.

Without any further ado, let’s get started!

Kafka Settings

bootstrap.servers

This is the most important setting as it tells Kafka clients where to look to talk to your applications. Sometimes this will be a string that contains just a single hostname:port combination, while other times it will be a comma-separated list of hostname:port combinations. Depending upon the Kafka offering you are connecting to, as well as how and where it is deployed, you will either be providing a singular bootstrap address or the entire collection of Kafka brokers in the same string. The bootstrap endpoint is a functional intermediary that Kafka offerings provide to allow for simpler networking in cases of expected fluctuations in the number of Kafka brokers, as it is a single URL that will return all the active brokers in the cluster to your clients automatically.

security.protocol

This is the next most important security setting as it tells the clients how to communicate with the brokers. If you’ve ever had HTTPS issues while using a web browser, this is the setting that would give Kafka clients SSL issues while attempting to connect to Kafka brokers! If you are receiving errors in your application while attempting to connect to Kafka with references to a broker(-1), you will usually need to revisit this setting (i.e. you should be using SASL_SSL but you configured PLAINTEXT, etc.).

The valid values are:

  • PLAINTEXT (using PLAINTEXT transport layer & no authentication — default value)
  • SSL (using SSL transport layer & certificate-based authentication)
  • SASL_PLAINTEXT (using PLAINTEXT transport layer & SASL-based authentication)
  • SASL_SSL (using SSL transport layer & SASL-based authentication)

ssl.truststore.location

To allow for encryption of communication between Kafka brokers and clients, as specified by our security.protocol setting above, we need to provide our Kafka clients with the location of a trusted Certificate Authority-based certificate. This file is often provided by the Kafka administrator and is generally unique to the specific Kafka cluster deployment. It is outside the scope of this blog post to determine how you provide access to this file for your individual runtimes, but it must exist inside the runtime container somewhere.

For Java-based clients, this file needs to be in the JKS format. For Python or NodeJS-based clients, this file should be in either a PEM or P12 format. To extract a PEM-based certificate from a JKS-based truststore, you can use the following command:

ssl.truststore.password

When providing a JKS-based truststore for validation of encryption certificates via the ssl.truststore.location configuration property above, you will generally need to provide a password to access the truststore file. This should be provided by your Kafka administrator at the same time they provide you with the truststore (or at least provide you with a method to acquire the truststore password). Truststore passwords are not required for PEM-based truststores, only JKS-based truststores.

sasl.mechanism

SASL stands for Simple Authentication and Security Layer and focuses on decoupling authentication mechanisms from application protocols — something that the majority of the early Internet-era applications and middleware did a very poor job of. As such, there are a number of pluggable mechanisms that can be included here to determine authentication between Kafka clients and brokers once the necessary communication protocol has been established with our previous setting.

The most commonly encountered values here are:

  • PLAIN (cleartext passwords, although they will be encrypted across the wire per security.protocol settings above)
  • SCRAM-SHA-512 (modern Salted Challenge Response Authentication Mechanism)
  • GSSAPI (Kerberos-supported authentication and the default if not specified otherwise)

sasl.jaas.config

Java-based clients make use of the Java Authentication and Authorization Service or JAAS framework that extends the plug-ability of the SASL decoupling approach to the enterprise-secured Java programming language. This allows for flexible yet robust security configuration through JAAS configuration. All of this boils down to you needing to provide a string named sasl.jaas.config that contains three main components: (1) a class name, (2) a username, and (3) a password. The security components and plugins take it from there. This string is often provided by your Kafka administrator with all the components already filled in for you.

Some commonly encountered sasl.jaas.config strings are:

  • org.apache.kafka.common.security.plain.PlainLoginModule required username="{USERNAME}" password="{PASSWORD}";
  • org.apache.kafka.common.security.scram.ScramLoginModule required username="{USERNAME}" password="{PASSWORD}";

Common configuration options for the most popular offerings

Strimzi / Red Hat AMQ Streams / IBM Event Streams

Strimzi is an open-source project which provides a Kubernetes Operator-focused way of deploying Apache Kafka clusters flexibly, repeatedly, and securely to any Kubernetes distribution. The enterprise-grade offerings of Red Hat AMQ Streams and IBM Event Streams are built on top of the underlying Strimzi Operator Custom Resources and take advantage of much of the same configuration specifications.

Clients external to the Kubernetes or OpenShift cluster which Strimzi et al are deployed to often use the following configurations:

Clients internal to the Kubernetes or OpenShift cluster which Strimzi et al are deployed to often use the following configurations: (note the difference as these clients do not require SSL-based encryption by default as they are local to the cluster)

NOTE: Recent Strimzi versions have added support for OAuth-based authentication, beyond just SASL or mTLS-based authentication. We have not widely encountered this in the field yet, but will address it once we do.

For more information on the specifics of configuring secure listeners on Strimzi-based Kafka clusters, you will want to visit the official Strimzi Docs. The examples in the documentation, as well as our demo scenario below, utilize three listeners with different security profiles (PLAINTEXT vs SSL transport layers / no authentication vs SASL-based authentication vs mutual TLS authentication / internal vs external clients) based upon the expected clients connecting to them.

For information on configuring users and access credentials, visit the Strimzi Docs page for User Authentication as the least-common-denominator between all three noted Strimzi variants. This method of user administration leverages the User Operator provided by Strimzi to manage users through Kubernetes Custom Resources.

Confluent Platform

Confluent provides many configuration options and deployment targets for their Confluent Platform on-premises Kafka offering. Our team’s main focus is Red Hat OpenShift-based deployments and as such, we focus on the capabilities made available to Confluent Platform through the Confluent Operator and SASL and mTLS-based authentication mechanisms with optional TLS-encrypted traffic.

Clients external to the Kubernetes or OpenShift cluster which Confluent Platform is deployed to most often use the following configurations:

Clients internal to the Kubernetes or OpenShift cluster which Confluent Platform is deployed to most often use the following configurations:

For information on configuring custom SASL users with the Confluent Operator, visit the Confluent Docs page for Authentication and Encryption.

IBM Event Streams on Cloud

IBM provides a hosted, SaaS-based Kafka offering as part of IBM Cloud and it is available via the IBM Cloud Catalog. It comes pre-configured with everything you need to immediately get started working with Kafka applications, along with authentication that is directly integrated into the IBM Cloud IAM.

As the IBM Event Streams on Cloud does not leverage a bootstrap address for client connectivity, the bootstrap.servers property is a list of broker endpoint addresses instead of a single broker bootstrap proxy endpoint.

For information on configuring users and access credentials, visit the IBM Cloud Docs page for Connecting to Event Streams and Kafka Clients.

Practical Use Case with Kafka binaries

Now we will walk through a straight-forward use case of deploying a vanilla open-source Strimzi cluster (which remember uses the same underlying configuration as IBM Event Streams V10 and Red Hat AMQ Streams), creating a SCRAM-based user, creating a topic, and interacting with that topic through the CLI tools provided by the Kafka binaries.

The only parameters we will be configuring from the downstream client perspective are the ones we covered above. This quick tutorial assumes access to a Red Hat OpenShift 4.5+ cluster. If you do not have one, you can access one easily via the IBM Open Labs.

NOTE: You will want to avoid the OpenShift Playground clusters available on Katacoda for this exercise, as they use a more complex SSL certificate setup than the scope of this article is meant to address.

Scenario setup

  • Access an OpenShift 4.5+ cluster via the OpenShift Console.
  • Go to Operators on the left hand side menu and click OperatorHub. Then, search for strimzi.
  • Click the Strimzi tile and click on Install.
  • Make sure you are installing from the stable channel and you are installing the operator in All namespaces on the cluster. Click Install.
  • After a couple of minutes, you should see the Strimzi operator successfully deployed and running.
  • Go to Projects on the left-hand side menu and click Create Project.
  • Once the project is created, you should be presented with a dashboard for it.

Create a Kafka cluster instance

  • Click Installed Operators under the Operators section in the left hand side menu. Click the Strimzi operator.
  • Click the Kafka tab and click Create Kafka.
  • Switch to the YAML view.
  • Update the listeners section within the .spec section with the following configuration:

The above listeners configuration will create an internal non-secured listener on port 9092, an internal TLS-secured listener on port 9093 and an external TLS-secured and SCRAM-SHA-512-authentication-required on port 9094. We will use this last port to access our Kafka cluster from outside of OpenShift. As a result, we will need a TLS certificate for the SSL connection plus a set of SCRAM-SHA-512 credentials to get authenticate with.

  • Make sure the entityOperator along with the topicOperator and the userOperator are also defined within the spec section. If they are not, add them.
  • Once you have updated the listeners section, click Create. After a minute or so, you should see the Kafka Cluster ready.

Create a Kafka topic via the Strimzi Operator.

  • Click the KafkaTopic tab and then click Create KafkaTopic:
  • Leave the defaults and click Create.

NOTE: Your Topic Name should stay my-topic for this simple example to align with the default ACLs for the KafkaUser you will create next.

  • You should see your topic created.

Create a KafkaUser via the Strimzi Operator

  • Click the KafkaUser tab and then click Create KafkaUser:
  • Click Authentication and switch the authentication type to scram-sha-512. Click Create.

This will create a Kafka user with a set of scram-sha-512 credentials so that it can access the Kafka Cluster from outside of OpenShift using the external listener on port 9094 you created earlier.

  • You should see the user created.

Generate configuration properties

On your terminal, execute the following commands:

  • Download the Kafka binaries to your sandbox environment.
  • Change project to kafka-security.
  • Export our configuration data to environment variables for easy reuse.
  • Setup our three basic SASL/SSL properties.

Since we are going to work with the Kafka cluster we deployed earlier from outside of the OpenShift cluster, we need to do that through the external SSL-secured and SCRAM-SHA-512-authentication-protected 9094 Kafka listener we defined earlier (the Strimzi operator will take care of creating the external access to that listener within the OpenShift cluster in the form of an OpenShift Route).

As a result, the security.protocol we need to set in the configuration file that the Kafka binaries scripts will later use is SASL_SSL. Then, we need to provide two things for this type of connection to a Kafka cluster. One is the SSL certificate for the secure communication and the other is the SCRAM-SHA-512 credentials. The first is provided through the ssl.truststore.* properties acquired in the next two steps. The other, the SCRAM-SHA-512 credentials, are provided through the sasl.* properties above.

  • Play the role of the Kafka Administrator and generate your own Java Truststore artifact. This is the certificate that will allow us to establish the SSL communication to our Kafka cluster.
  • Pass in the path to the Java Truststore artifact on the local filesystem (with absolute path).

Run the Kafka binaries

On your terminal, execute the following commands to validate our connection properties and interact with the Kafka brokers:

  • View topic data via bin/kafka-topic.sh tool:
  • Send data with the bin/kafka-console-producer.sh tool:

NOTE: Use `Ctrl+C` when you are done sending data to exit the console producer.

  • Read data with the bin/kafka-console-consumer.sh tool:

NOTE: Use `Ctrl+C` when you are done reading data to exit the console consumer.

With this understanding of Kafka client security configuration, you can now seamlessly switch between Strimzi to IBM Event Streams to Red Hat AMQ Streams to Confluent Platform to IBM Event Streams on Cloud to Lenses.io to kafdrop to kafkacat to Kafka-provided binaries and beyond by following the same process!

Conclusion

These same six settings will occur over and over again, from Producer Configs to Consumer Configs to Kafka Streams Configs to Kafka Connect Configs to Admin Client Configs to command-line configs. They are everywhere! Depending on your implementation language, platform, and tooling, there may be some assistance available in the SDKs, APIs, or CLIs to prevent the need of copying and pasting across the different Kafka component configurations, but once you get the configuration correct at the highest level, everything else in your event-driven applications seems much simpler to connect, debug, and validate at the lower levels!

I hope this whirlwind tour of practical Kafka security configuration options has been helpful and will become a handy reference for you to use going forward! Please let me know if you find this useful and would like to see more practical Kafka experiences documented here. In the meantime, head over to our Event-Driven Reference Architecture in the IBM Cloud Architecture Center that has the latest and greatest collection of all things event-driven and Kafka-based!

A special shoutout of thanks to Jesus Almaraz & Jerome Boyer for their reviews of, edits on, and polishing contributions to this article.

--

--

Rick Osowski
Rick Osowski

Written by Rick Osowski

Dad to B, Husband to A, Mental Health & Suicide Prevention Advocate, Magic: The Gathering Nerd, Principal Engineer — Kubernetes at Seagate Technology.