Distributed Job scheduling, management, and reporting

前端 未结 6 1130
南旧
南旧 2021-01-02 05:20

I recently had a play around with Hadoop and was impressed with it\'s scheduling, management, and reporting of MapReduce jobs. It appears to make the distribution and execut

相关标签:
6条回答
  • 2021-01-02 06:04

    I guess you are looking for a workflow engine for CPU intensive tasks (also know "scientific workflow", e.g. http://www.extreme.indiana.edu/swf-survey). But I'm not sure how distributed do you want it to be. Usually all workflow engines have a "single point of failure".

    0 讨论(0)
  • 2021-01-02 06:05

    I have since found Spring Batch and Spring Batch Integration which appear to address many of my requirements. I will let you know how I get on.

    0 讨论(0)
  • 2021-01-02 06:08

    ProActive Scheduler seems to fit your requirements, especially the complex workflows you mentionned with tasks coordination. It is open source and Java based. You can use it to run anything, Hadoop jobs, scripts, Java code,...

    Disclaimer: I work for the company behind it

    0 讨论(0)
  • 2021-01-02 06:12

    I believe quite a few problems can be expressed as map-reduce problems.

    For problems that you can't modify to fit the structure your can look at setting up your own using Java's ExecutorService. But it will be limited to one JVM and it will be quite low level. It will allow for easy coordination and synchronization however.

    0 讨论(0)
  • 2021-01-02 06:14

    Take a look at Quartz. I think it supports stuff like managing jobs remotely and clustering several machines to run jobs.

    0 讨论(0)
  • 2021-01-02 06:17

    Try Redisson framework. It provides easy api to execute and schedule java.util.concurrent.Callable and java.lang.Runnable tasks. Here is documentation about distributed Executor service and Scheduler service

    0 讨论(0)
提交回复
热议问题