apache-storm

How to see the input and output in storm [closed]

血红的双手。 提交于 2019-12-11 11:24:28
问题 Closed . This question needs details or clarity. It is not currently accepting answers. Want to improve this question? Add details and clarify the problem by editing this post. Closed 5 years ago . I have installed Apache Storm and also executed the example word count topology. But there was no input or output shown. How can I see it? 回答1: Basic Understanding The manual is a must-read, especially the Concepts section will help you to get a basic understanding. You should not only install the

Storm UI topology not working

别说谁变了你拦得住时间么 提交于 2019-12-11 11:11:57
问题 We are executing a Storm topology in LocalCluster . Storm topology is executing fine and able to connect Storm UI (8090). But Storm UI is not displaying the running topology information. LocalCluster cluster = new LocalCluster(); and submitting like: bin/storm jar bin/StormTest-0.0.1-SNAPSHOT.jar com.abzooba.storm.twitter.TwitterTopologyCreator Twitter 回答1: LocalCluster does not have UI support... Thus the UI you are seeing belongs to a different Storm cluster. To be more precise:

storm spout didn't emit & result file empty

故事扮演 提交于 2019-12-11 09:39:27
问题 i'm new to storm i submitted different topologies form github like storm-starter and others successfully after some problem i faced it but some of the some spout didn't emit ! like that is that right ? or is there a problem ? where can i find the result after submitting topology ? i guessed in the result file in storm folder but it's empty ! this what i have in the bolt "one of them " 回答1: There is no explicit "result" file in storm. Your topology is responsible for handling/storing results

How to test Storm using Spock when emitting large output Tuples

无人久伴 提交于 2019-12-11 08:46:36
问题 So I have the following test: def "test execute(inputTuple) method emits an outputTuple containing a member ObjectType retrieved using the entity_id in inputTuple"() { given: Tuple inputTuple = Mock(Tuple); List<String> objectTypeIDsEmittedByPreviousBolt = new ArrayList<String>(); objectTypeIDsEmittedByPreviousBolt.add("member"); objectTypeIDsEmittedByPreviousBolt.add("1196"); 1 * inputTuple.getValues() >> objectTypeIDsEmittedByPreviousBolt; when: this.bolt.execute(inputTuple); then: 1 * this

Kafka in supervisor mode

会有一股神秘感。 提交于 2019-12-11 07:40:56
问题 I'm trying to run kafka in supervision mode so that it can start automatically in case of a shutdown. But all the examples of running kafka use shell scripts and the supervisord is not able to note which PID to monitor. Can anyone suggesthow to accomplish auto restart of kafka? 回答1: If you are on a Unix or Linux machine, then this is when /etc/inittab comes in handy. Or you might want to use daemontools. I don't know about Windows though. 回答2: We are running Kafka under Supervisord (http:/

Change Log4J Configuration for Apache Storm Topology

孤者浪人 提交于 2019-12-11 07:08:40
问题 I'm currently submitting Storm topologies programatically via my Java application by using the following command: Nimbus.Client client = NimbusClient.getConfiguredClient(stormConfigProvider.getStormConfig()).getClient(); client.submitTopology( this.topologyID.toString(), stormJarManager.getRemoteJarLocation(), JSONValue.toJSONString(stormConfigProvider.getStormConfig()), topology ); In my scenario, I have two kinds of topologies. Testing topologies and production topologies. For both kind of

Elasticsearch / Storm integration methods

自闭症网瘾萝莉.ら 提交于 2019-12-11 07:05:23
问题 Looking for a simple integration path between Elasticsearch and Apache Storm. Support for this is included in the elasticsearch-hadoop library, but this brings tons of dependencies on the Hadoop stack: from Hive to Cascading, that I simply don't need. Has anyone out there succeeded in this integration without bringing in elasticsearch-hadoop? Thanks. 回答1: In my project we're using rabbitmq river for indexing the storm output. It's very efficient and convenient way to write to elasticsearch.

Apache Storm Kafka Spout Lag Issue

非 Y 不嫁゛ 提交于 2019-12-11 06:37:17
问题 I am building a Java Spring application using Storm 1.1.2 and Kafka 0.11 to be launched in a Docker container. Everything in my topology works as planned but under a high load from Kafka, the Kafka lag increases more and more over time. My KafkaSpoutConfig: KafkaSpoutConfig<String,String> spoutConf = KafkaSpoutConfig.builder("kafkaContainerName:9092", "myTopic") .setProp(ConsumerConfig.GROUP_ID_CONFIG, "myGroup") .setProp(ConsumerConfig.VALUE_DESERIALIZER_CLASS_CONFIG, MyObjectDeserializer

kryo serialization over storm

折月煮酒 提交于 2019-12-11 04:49:27
问题 I need to serialize complex object (opencv:Mat) over apache storm (deployed in remote cluster). Can anyone suggest me a good tutorial custom kryo serialization or propose a solution on how to do this? Thanks in advance! 回答1: I have created a bean public class DataBean{ Mat imageMatrix; int id; public DataBean(){ } public DataBean(int id, Mat matrix) { setId(id); setImageMatrix(matrix); } public int getId() { return id; } public void setId(int id) { this.id = id; } public Mat getImageMatrix()

apache storm reliablity timeout configuration

旧城冷巷雨未停 提交于 2019-12-11 03:46:20
问题 I have a nodejs->kafka>storm->Mongo deployed in Linux Ubuntu. Everything is normal originally. Then I changed the method in storm worker which makes storm worker process message very slow, around 1 minute per message, I notice the message is sent again and again from storm. I revert back to original method, everything is fine. (original method process time is 90ms per message). I guess this is Storm reliability come into player. When message is not acknowledged, or time out, it sends message