Skip to content

Kafka GUI for Apache Kafka to manage topics, topics data, consumers group, schema registry, connect and more...

License

Notifications You must be signed in to change notification settings

KarstenWintermann/kafkahq

 
 

Repository files navigation

KafkaHQ

Last Version License Docker Pull Github Downloads Github Start Build Status

Kafka GUI for Apache Kafka to manage topics, topics data, consumers group, schema registry, connect and more...

preview

Features

  • General
    • Works with modern Kafka cluster (1.0+)
    • Connection on standard or ssl, sasl cluster
    • Multi cluster
  • Topics
    • List
    • Configurations view
    • Partitions view
    • ACLS view
    • Consumers groups assignments view
    • Node leader & assignments view
    • Create a topic
    • Configure a topic
    • Delete a topic
  • Browse Topic datas
    • View data, offset, key, timestamp & headers
    • Automatic deserializarion of avro message encoded with schema registry
    • Configurations view
    • Logs view
    • Delete a record
    • Sort view
    • Filter per partitions
    • Filter with a starting time
    • Filter data with a search string
  • Consumer Groups (only with kafka internal storage, not with old Zookepper)
    • List with lag, topics assignments
    • Partitions view & lag
    • ACLS view
    • Node leader & assignments view
    • Display active and pending consumers groups
    • Delete a consumer group
    • Update consumer group offsets to start / end / timestamp
  • Schema Registry
    • List schema
    • Create / Update / Delete a schema
    • View and delete individual schema version
  • Connect
    • List connect definition
    • Create / Update / Delete a definition
    • Pause / Resume / Restart a definition or a task
  • Nodes
    • List
    • Configurations view
    • Logs view
    • Configure a node
  • ACLS
    • List principals
    • List principals topic & group acls
  • Authentification and Roles
    • Read only mode
    • BasicHttp with roles per user
    • User groups configuration
    • Filter topics with regexp for current groups
    • Ldap configuration to match KafkaHQ groups/roles

Quick preview

It will start a Kafka node, a Zookeeper node, a Schema Registry, a Connect, fill with some sample data, start a consumer group and a kafka stream & start KafkaHQ.

Installation

First you need a configuration files in order to configure KafkaHQ connections to Kafka Brokers.

Docker

docker run -d \
    -p 8080:8080 \
    -v /tmp/application.yml:/app/application.yml \
    tchiotludo/kafkahq
  • With -v /tmp/application.yml must be an absolute path to configuration file
  • Go to http://localhost:8080

Stand Alone

Running in Kubernetes (using a Helm Chart)

  • Clone the repository:
git clone https://github.com/tchiotludo/kafkahq && cd kafkahq/deploy/helm
  • Update helm values located in deploy/helm/values.yaml
    • configuration values will contains all related configuration that you can find in application.example.yml and will be store in a ConfigMap
    • secrets values will contains all sensitive configurations (with credentials) that you can find in application.example.yml and will be store in Secret
    • Both values will be merged at startup
  • Apply the chart:
helm install --name=kafkahq-release-name  .

Configuration

Configuration file can by default be provided in either Java properties, YAML, JSON or Groovy files. YML Configuration file example can be found here :application.example.yml

Kafka cluster configuration

  • kafkahq.connections is a key value configuration with :
    • key: must be an url friendly (letter, number, _, -, ... dot are not allowed here) string the identify your cluster (my-cluster-1 and my-cluster-2 is the example above)
    • properties: all the configurations found on Kafka consumer documentation. Most important is bootstrap.servers that is a list of host:port of your Kafka brokers.
    • schema-registry: (optional)
      • url: the schema registry url
      • basic-auth-username: schema registry basic auth username
      • basic-auth-password: schema registry basic auth password
      • properties: all the configurations for registry client, especially ssl configuration
    • connect: (optional list, define each connector as a element of a list)
      • name: connect name
      • url: connect url
      • basic-auth-username: connect basic auth username
      • basic-auth-password: connect basic auth password
      • ssl-trust-store: /app/truststore.jks
      • ssl-trust-store-password: trust-store-password
      • ssl-key-store: /app/truststore.jks
      • ssl-key-store-password: key-store-password

SSL Kafka Cluster with basic auth

Configuration example for kafka cluster secured by ssl for saas provider like aiven (full https & basic auth):

You need to generate a jks & p12 file from pem, cert files give by saas provider.

openssl pkcs12 -export -inkey service.key -in service.cert -out client.keystore.p12 -name service_key
keytool -import -file ca.pem -alias CA -keystore client.truststore.jks

Configurations will look like this example:

kafkahq:
  connections:
    ssl-dev:
      properties:
        bootstrap.servers: "{{host}}.aivencloud.com:12835"
        security.protocol: SSL
        ssl.truststore.location: {{path}}/avnadmin.truststore.jks
        ssl.truststore.password: {{password}}
        ssl.keystore.type: "PKCS12"
        ssl.keystore.location: {{path}}/avnadmin.keystore.p12
        ssl.keystore.password: {{password}}
        ssl.key.password: {{password}}
      schema-registry:
        url: "https://{{host}}.aivencloud.com:12838"
        basic-auth-username: avnadmin
        basic-auth-password: {{password}}
        properties: {}
      connect:
        - name: connect-1
          url: "https://{{host}}.aivencloud.com:{{port}}"
          basic-auth-username: avnadmin
          basic-auth-password: {{password}}

KafkaHQ configuration

Pagination

  • kafkahq.pagination.page-size number of topics per page (default : 25)

Topic List

  • kafkahq.topic.default-view is default list view (ALL, HIDE_INTERNAL, HIDE_INTERNAL_STREAM, HIDE_STREAM)
  • kafkahq.topic.internal-regexps is list of regexp to be considered as internal (internal topic can't be deleted or updated)
  • kafkahq.topic.stream-regexps is list of regexp to be considered as internal stream topic
  • kafkahq.topic.skip-consumer-groups disable loading of consumer group information when showing topics (true), default is to load the information

Topic creation default values

These parameters are the default values used in the topic creation page.

  • kafkahq.topic.retention Default retention in ms
  • kafkahq.topic.replication Default number of replica to use
  • kafkahq.topic.partition Default number of partition

Topic Data

  • kafkahq.topic-data.sort: default sort order (OLDEST, NEWEST) (default: OLDEST)
  • kafkahq.topic-data.size: max record per page (default: 50)
  • kafkahq.topic-data.poll-timeout: The time, in milliseconds, spent waiting in poll if data is not available in the buffer (default: 1000).

Security

  • kafkahq.security.default-roles: Roles available for all the user even unlogged user, roles available are :
    • topic/read
    • topic/insert
    • topic/delete
    • topic/config/update
    • node/read
    • node/config/update
    • topic/data/read
    • topic/data/insert
    • topic/data/delete
    • group/read
    • group/delete
    • group/offsets/update
    • acls/read
    • registry/read
    • registry/insert
    • registry/update
    • registry/delete
    • registry/version/delete

By default, security & roles is enabled by default but anonymous user have full access. You can completely disabled security with micronaut.security.enabled: false.

If you need a read-only application, simply add this to your configuration files :

kafkahq:
  security:
    default-roles:
      - topic/read
      - node/read
      - topic/data/read
      - group/read
      - registry/read
      - connect/read

Auth

Groups

Groups allow you to limit user

Define groups with specific roles for your users

  • kafkahq.security.groups: Groups list definition
    • group-name: Group identifier
      • roles: Roles list for the group
      • attributes.topics-filter-regexp: Regexp to filter topics available for current group

2 defaults group are available :

  • admin with all right
  • reader with only read acces on all KafkaHQ
Basic Auth
  • kafkahq.security.basic-auth: List user & password with affected roles
    • actual-username: Login of the current user as a yaml key (may be anything email, login, ...)
      • password: Password in sha256, can be converted with command echo -n "password" | sha256sum
      • groups: Groups for current user

Take care that basic auth will use session store in server memory. If your instance is behind a reverse proxy or a loadbalancer, you will need to forward the session cookie named SESSION and / or use sesssion stickiness

LDAP

Configure how the ldap groups will be matched in KafkaHQ groups

  • kafkahq.security.ldap.group: Ldap groups list
    • ldap-group-name: Ldap group name (same name as in ldap)
      • groups: KafkaHQ group list to be used for current ldap group

Example using online ldap test server

Configure ldap connection in micronaut

micronaut:
  security:
    ldap:
      default:
        enabled: true
        context:
          server: 'ldap://ldap.forumsys.com:389'
          managerDn: 'cn=read-only-admin,dc=example,dc=com'
          managerPassword: 'password'
        search:
          base: "dc=example,dc=com"
        groups:
          enabled: true
          base: "dc=example,dc=com"

Configure KafkaHQ groups and Ldap groups and users

kafkahq:
  security:
    groups:
      topic-reader: # Group name
        roles:  # roles for the group
          - topic/read
        attributes:
          # Regexp to filter topic available for group
          topics-filter-regexp: "test\\.reader.*"
      topic-writer: 
        roles:
          - topic/read
          - topic/insert
          - topic/delete
          - topic/config/update
        attributes:
          topics-filter-regexp: "test.*"
    ldap:
      group:
        mathematicians:
          groups:
            - topic-reader
        scientists:
          groups:
            - topic-reader
            - topic-writer
      user:
        franz:
          groups:
            - topic-reader
            - topic-writer

Server

  • kafkahq.server.base-path: if behind a reverse proxy, path to kafkahq with trailing slash (optional). Example: kafkahq is behind a reverse proxy with url http://my-server/kafkahq, set base-path: "/kafkahq/". Not needed if you're behind a reverse proxy with subdomain http://kafkahq.my-server/

Kafka admin / producer / consumer default properties

  • kafkahq.clients-defaults.{{admin|producer|consumer}}.properties: default configuration for admin producer or consumer. All properties from Kafka documentation is available.

Micronaut configuration

Since KafkaHQ is based on Micronaut, you can customize configurations (server port, ssl, ...) with Micronaut configuration. More information can be found on Micronaut documentation

Docker

KafkaHQ docker image support 3 environment variables to handle configuraiton :

  • KAFKAHQ_CONFIGURATION: a string that contains the full configuration in yml that will be written on /app/configuration.yml on container.
  • MICRONAUT_APPLICATION_JSON: a string that contains the full configuration in JSON format
  • MICRONAUT_CONFIG_FILES: a path to to a configuration file on container. Default path is /app/application.yml

How to mount configuration file

Take care when you mount configuration files to not remove kafkahq files located on /app. You need to explicitely mount the /app/application.yml and not mount the /app directory. This will remove the KafkaHQ binnaries and give you this error: /usr/local/bin/docker-entrypoint.sh: 9: exec: ./kafkahq: not found

volumeMounts:
- mountPath: /app/application.yml
  subPath: application.yml
  name: config
  readOnly: true

Monitoring endpoint

Several monitoring endpoint is enabled by default. You can disabled it or restrict access only for authenticated users following micronaut configuration below.

Development Environment

Early dev image

You can have access to last feature / bug fix with docker dev image automatically build on tag dev

docker pull tchiotludo/kafkahq:dev

The dev jar is not publish on GitHub, you have 2 solutions to have the dev jar :

Get it from docker image

docker pull tchiotludo/kafkahq:dev
docker run --rm --name=kafkahq -it tchiotludo/kafkahq:dev
docker cp kafkahq:/app/kafkahq.jar . 

Or build it with a ./gradlew shadowJar, the jar will be located here build/libs/kafkahq-*.jar

Development Server

A docker-compose is provide to start a development environnement. Just install docker & docker-compose, clone the repository and issue a simple docker-compose -f docker-compose-dev.yml up to start a dev server. Dev server is a java server & webpack-dev-server with live reload.

Who's using KafkaHQ

Credits

Many thanks to:

Jetbrains

License

Apache 2.0 © tchiotludo

About

Kafka GUI for Apache Kafka to manage topics, topics data, consumers group, schema registry, connect and more...

Resources

License

Stars

Watchers

Forks

Packages

No packages published

Languages

  • Java 65.5%
  • FreeMarker 22.7%
  • JavaScript 7.1%
  • CSS 4.4%
  • Smarty 0.2%
  • Shell 0.1%