A Python library for automatically optimizing Kafka producer configurations based on topic-specific recommendations.
Superstream Clients works as a Python import hook that intercepts Kafka producer creation and applies optimized configurations without requiring any code changes in your application. It dynamically retrieves optimization recommendations from Superstream and applies them based on impact analysis.
Works with any Python library that implements Kafka producers, including:
- kafka-python
- aiokafka
- confluent-kafka
- Faust
- FastAPI event publishers
- Celery Kafka backends
- Any custom wrapper around these Kafka clients
- Zero-code integration: No code changes required in your application
- Dynamic configuration: Applies optimized settings based on topic-specific recommendations
- Intelligent optimization: Identifies the most impactful topics to optimize
- Graceful fallback: Falls back to default settings if optimization fails
- Minimal overhead: Uses a single lightweight background thread (or async coroutine for aiokafka)
When initializing your Kafka producers, please ensure you pass the configuration as a mutable object. The Superstream library needs to modify the producer configuration to apply optimizations. The following initialization patterns are supported:
✅ Supported (Recommended):
# Using kafka-python
from kafka import KafkaProducer
producer = KafkaProducer(
bootstrap_servers=['localhost:9092'],
compression_type='snappy',
batch_size=16384
)
# Using aiokafka
from aiokafka import AIOKafkaProducer
producer = AIOKafkaProducer(
bootstrap_servers='localhost:9092',
compression_type='snappy',
batch_size=16384
)
# Using confluent-kafka
from confluent_kafka import Producer
producer = Producer({
'bootstrap.servers': 'localhost:9092',
'compression.type': 'snappy',
'batch.size': 16384
})
❌ Not Supported:
# Using frozen dictionaries or immutable configurations
from types import MappingProxyType
config = MappingProxyType({
'bootstrap.servers': 'localhost:9092'
})
producer = KafkaProducer(**config)
The Superstream library needs to modify your producer's configuration to apply optimizations based on your cluster's characteristics. This includes adjusting settings like compression, batch size, and other performance parameters. When the configuration is immutable, these optimizations cannot be applied.
pip install superclient && python -m superclient install_pth
That's it! Superclient will now automatically load and optimize all Kafka producers in your Python environment.
After installation, superclient works automatically. Just use your Kafka clients as usual:
# kafka-python
from kafka import KafkaProducer
producer = KafkaProducer(bootstrap_servers='localhost:9092')
# Automatically optimized!
# confluent-kafka
from confluent_kafka import Producer
producer = Producer({'bootstrap.servers': 'localhost:9092'})
# Automatically optimized!
# aiokafka
from aiokafka import AIOKafkaProducer
producer = AIOKafkaProducer(bootstrap_servers='localhost:9092')
# Automatically optimized!
When using Superstream Clients with containerized applications, include the package in your Dockerfile:
FROM python:3.8-slim
# Install superclient
RUN pip install superclient
RUN python -m superclient install_pth
# Your application code
COPY . /app
WORKDIR /app
# Run your application
CMD ["python", "your_app.py"]
SUPERSTREAM_TOPICS_LIST
: Comma-separated list of topics your application produces to
SUPERSTREAM_LATENCY_SENSITIVE
: Set to "true" to prevent any modification to linger.ms valuesSUPERSTREAM_DISABLED
: Set to "true" to disable optimizationSUPERSTREAM_DEBUG
: Set to "true" to enable debug logs
Example:
export SUPERSTREAM_TOPICS_LIST=orders,payments,user-events
export SUPERSTREAM_LATENCY_SENSITIVE=true
- Python 3.8 or higher
- Kafka cluster that is connected to the Superstream's console
- Read and write permissions to the
superstream.*
topics
Apache License 2.0