apache-kafka-streams

UnsatisfiedLinkError on Lib rocks DB dll when developing with Kafka Streams

最后都变了- 提交于 2019-11-29 07:20:41
I'm writing a Kafka Streams application on my development Windows machine. If I try to use the leftJoin and branch features of Kafka Streams I get the error below when executing the jar application: Exception in thread "StreamThread-1" java.lang.UnsatisfiedLinkError: C:\Users\user\AppData\Local\Temp\librocksdbjni325337723194862275.dll: Can't find dependent libraries at java.lang.ClassLoader$NativeLibrary.load(Native Method) at java.lang.ClassLoader.loadLibrary0(ClassLoader.java:1941) at java.lang.ClassLoader.loadLibrary(ClassLoader.java:1824) at java.lang.Runtime.load0(Runtime.java:809) at

Kafka KStreams - processing timeouts

走远了吗. 提交于 2019-11-29 06:58:57
I am attempting to use <KStream>.process() with a TimeWindows.of("name", 30000) to batch up some KTable values and send them on. It seems that 30 seconds exceeds the consumer timeout interval after which Kafka considers said consumer to be defunct and releases the partition. I've tried upping the frequency of poll and commit interval to avoid this: config.put(StreamsConfig.COMMIT_INTERVAL_MS_CONFIG, "5000"); config.put(StreamsConfig.POLL_MS_CONFIG, "5000"); Unfortunately these errors are still occurring: (lots of these) ERROR o.a.k.s.p.internals.RecordCollector - Error sending record to topic

UnsatisfiedLinkError: /tmp/snappy-1.1.4-libsnappyjava.so Error loading shared library ld-linux-x86-64.so.2: No such file or directory

旧时模样 提交于 2019-11-29 06:32:42
I am trying to run a Kafka Streams application in kubernetes. When I launch the pod I get the following exception: Exception in thread "streams-pipe-e19c2d9a-d403-4944-8d26-0ef27ed5c057-StreamThread-1" java.lang.UnsatisfiedLinkError: /tmp/snappy-1.1.4-5cec5405-2ce7-4046-a8bd-922ce96534a0-libsnappyjava.so: Error loading shared library ld-linux-x86-64.so.2: No such file or directory (needed by /tmp/snappy-1.1.4-5cec5405-2ce7-4046-a8bd-922ce96534a0-libsnappyjava.so) at java.lang.ClassLoader$NativeLibrary.load(Native Method) at java.lang.ClassLoader.loadLibrary0(ClassLoader.java:1941) at java.lang

Kafka multiple partition ordering

こ雲淡風輕ζ 提交于 2019-11-29 03:40:30
问题 I am aware that it is not possible to order multiple partitions in Kafka and that partition ordering is only guaranteed for a single consumer within a group (for a single partition). However with Kafka Streams 0.10 is it now possible to achieve this? If we use the timestamp feature so that each message in each partition maintains the order, at the consumer side, lets say with Kafka Streams 0.10 is this now possible? Assuming we receive all messages could we not sort all the partitions based

Why Apache Kafka Streams uses RocksDB and if how is it possible to change it?

大兔子大兔子 提交于 2019-11-29 02:20:57
问题 During investigation within new features in Apache Kafka 0.9 and 0.10, we had used KStreams and KTables. There is interesting fact, that Kafka uses RocksDB internally. See Introducing Kafka Streams: Stream Processing Made Simple. RocksDB is not written in JVN compatible language, so it needs careful handling of the deployment, as it needs extra shared library (OS dependent). And here there are simple questions: Why Apache Kafka Streams uses RocksDB? How is it possible to change it? I had

Kafka Streaming Concurrency?

↘锁芯ラ 提交于 2019-11-28 21:19:18
I have some basic Kafka Streaming code that reads records from one topic, does some processing, and outputs records to another topic. How does Kafka streaming handle concurrency? Is everything run in a single thread? I don't see this mentioned in the documentation. If it's single threaded, I would like options for multi-threaded processing to handle high volumes of data. If it's multi-threaded, I need to understand how this works and how to handle resources, like SQL database connections should be shared in different processing threads. Is Kafka's built-in streaming API not recommended for

Kafka: Consumer API vs Streams API

非 Y 不嫁゛ 提交于 2019-11-28 15:42:43
I recently started learning Kafka and end up with these questions. What is the difference between Consumer and Stream? For me, if any tool/application consume messages from Kafka is a consumer in the Kafka world. How Stream is different as this also consumes from or produce messages to Kafka? and why is it needed as we can write our own consumer application using Consumer API and process them as needed or send them to Spark from the consumer application? I did Google on this, but did not get any good answers for this. Sorry if this question is too trivial. Update April 09, 2018 : Nowadays you

Test Kafka Streams topology

南笙酒味 提交于 2019-11-28 11:23:40
I'm searching a way to test a Kafka Streams application. So that I can define the input events and the test suite shows me the output. Is this possible without a real Kafka setup? Update Kafka 1.1.0 (released 23-Mar-2018): KIP-247 added official test utils. Per the Upgrade Guide : There is a new artifact kafka-streams-test-utils providing a TopologyTestDriver , ConsumerRecordFactory , and OutputVerifier class. You can include the new artifact as a regular dependency to your unit tests and use the test driver to test your business logic of your Kafka Streams application. For more details, see

External system queries during Kafka Stream processing

梦想与她 提交于 2019-11-28 08:46:20
I'm trying to design a streaming architecture for streaming analytics. Requirements: RT and NRT streaming data input Stream processors implementing some financial analysis RT and NRT analysis output stream Reference data requests during stream processing I'm exploring Kafka and Kafka Streams for stream processing and RT/NRT realtime messaging. My question is: I need to perform some query to external systems (info providers, MongoDB etc etc) during stream pocessing. These queries could be both sync and async req-response, based on the external system characteristics. I've read this post

Print Kafka Stream Input out to console?

北战南征 提交于 2019-11-28 07:39:57
I've been looking through a lot of the Kafka documentation for a java application that I am working on. I've tried getting into the lambda syntax introduced in Java 8, but I am a little sketchy on that ground and don't feel too confident that it should be what I use as of yet. I've a Kafka/Zookeeper Service running without any troubles, and what I want to do is write a small example program that based on the input will write it out, but not do a wordcount as there are so many examples of already. As for sample data I will be getting a string of following structure: Example data This a sample