cluster-computing

Socket.io, cluster, express and sync events

血红的双手。 提交于 2019-12-04 02:47:29
I have a big problem sice 1 week. I try to convert my node.JS project actually run on single core to multi core with cluster. With websockets, at this moment, i have no problems for events but, for xhr-polling or jsonp-polling, i have big problems with socket.io on cluster mode. this is my server configuration : 00-generic.js 'use strict'; var http = require('http'), os = require('os'), cluster = require('cluster'); module.exports = function(done) { var app = this.express, port = process.env.PORT || 3000, address = '0.0.0.0'; if(this.env == 'test'){ port = 3030; } var self = this; var size =

Difference between “SOCK”, “PVM”, “MPI”, and “NWS” for the R SNOW package

隐身守侯 提交于 2019-12-04 01:24:54
The makeCluster function for the SNOW package has the different cluster types of " SOCK ", " PVM ", " MPI ", and " NWS " but I'm not very clear on the differences among them, and more specifically which would be best for my program. Currently I have a queue of tasks of different length going into a load balancing cluster with clusterApplyLB and am using a 64bit 32-core Windows machine. I am looking for a brief description of the differences among the four cluster types, which would be best for my use and why. Welcome to parallel programming. You may want to peruse the vignette of the excellent

how to specify error log file and output file in qsub

喜夏-厌秋 提交于 2019-12-03 22:14:32
I have a qsub script as #####----submit_job.sh---##### #!/bin/sh #$ -N job1 #$ -t 1-100 #$ -cwd SEEDFILE=/home/user1/data1 SEED=$(sed -n -e "$SGE_TASK_ID p" $SEEDFILE) /home/user1/run.sh $SEED The problem is-- it puts all error and output files (job1.eJOBID & job1.oJOBID) to the same directory from where I am running qsub submit_job.sh while I want to save these file (output and error log file in same different place (specified as $SEED_output). I tried to change the line as /home/user1/run.sh $SEED -o $SEED_output But it didn't work. Any suggestion ?? How can I specify path and name of

Running R in Batch Mode on Linux: Output Issues

让人想犯罪 __ 提交于 2019-12-03 16:40:41
I'm running an R program on a linux cluster because it is very demanding on my processor. My program is designed to output multiple (around 15) plots as PDF's into the folder from which the program gathers its input. I want my program to run in the background, and to continue running when I log out of the cluster. First, I tried this: cd /Users/The/Folder/With/My/RScript #changed working directory nohup ./BatchProgram.R & However, this didn't work because it appended the output to a file called nohup.out , and did not output any of the PDF's I need. Next I tried this: cd /Users/The/Folder/With

How are tasks distributed within a Spark cluster?

别说谁变了你拦得住时间么 提交于 2019-12-03 16:31:25
So I have an input that consists in a dataset and several ML algorithms (with parameter tuning) using scikit-learn. I have tried quite a few attempts on how to execute this as efficiently as possible but at this very moment I still don't have the proper infrastructure to assess my results. However, I lack some background on this area and I need help to get things cleared up. Basically I want to know how the tasks are distributed in a way that exploits as much as possible all the available resources, and what is actually done implicitly (for instance by Spark) and what isn't. This is my

In a hadoop cluster, should hive be installed on all nodes?

天涯浪子 提交于 2019-12-03 16:14:18
问题 I am a newbie to Hadoop / Hive and I have just started reading the docs. There are lots of blogs on installing Hadoop in cluster mode. Also, I know that Hive runs on top of Hadoop. My question is: Hadoop is installed on all the cluster nodes. Should I also install Hive on all the cluster nodes or only on the master node? 回答1: No, it is not something you install on worker nodes. Hive is a Hadoop client. Just run Hive according to the instructions you see at the Hive site. 回答2: Hive is

Deploying multiple Grails instances with shared cache and sessions?

牧云@^-^@ 提交于 2019-12-03 14:58:19
I am looking for a solution that allows me to deploy multiple load balanced Grails instances that have shared cache (EhCache Server ?) and sessions, is this possible ? I can't find any documentation on this (connecting to a common EhCache server or using Distributed EhCache, and sharing sessions (using EhCache too ?))... I'm looking for something that will work like multiple Rails instances with a common memcached and sessions/caches stored in the memcached... I was recently listening to a talk by Dave Klein , author of the book "Grails: A Quick Start Guide", and Mike Allen, Product Managemer

kill all child_process when node process is killed

柔情痞子 提交于 2019-12-03 14:56:46
问题 How do i make sure all child_process are killed when the parent process is killed. I have something like the below one. Even when the node process is kill i see that FFMPEG continues to run and the out.avi is generated. How can i stop FFMPEG from running after the node process exits. var args = "ffmpeg -i in.avi out.avi" child_process.exec(args , function(err, stdout,stderr){}); child_process.exec(args , function(err, stdout,stderr){}); 回答1: You need to listen for the process exit event and

Apache Flink streaming in cluster does not split jobs with workers

喜你入骨 提交于 2019-12-03 14:37:41
My objective is to setup a high throughput cluster using Kafka as source & Flink as the stream processing engine. Here's what I have done. I have setup a 2-node cluster the following configuration on the master and the slave. Master flink-conf.yaml jobmanager.rpc.address: <MASTER_IP_ADDR> #localhost jobmanager.rpc.port: 6123 jobmanager.heap.mb: 256 taskmanager.heap.mb: 512 taskmanager.numberOfTaskSlots: 50 parallelism.default: 100 Slave flink-conf.yaml jobmanager.rpc.address: <MASTER_IP_ADDR> #localhost jobmanager.rpc.port: 6123 jobmanager.heap.mb: 512 #256 taskmanager.heap.mb: 1024 #512

Hibernate Search in a Clustered Configuration?

喜你入骨 提交于 2019-12-03 14:30:03
I have a Java web app that I'm developing, using JBoss Seam as the application framework. I'd like to take advantage of Hibernate Search to provide entity searching capabilities. The integration has gone fine, and I'm getting closer to deployment. The setup for the app in production will be: 2 (or more) Tomcat app servers, load balanced PostgreSQL database on the backend My question is, how can I configure Hibernate Search so that the local Lucene indices are updated on App Server #2 when an entity is persisted on App Server #1, and vice versa? Does Hibernate Search have any sort of clustered