cluster-computing

Using a loop variable in a Bash script to pass different command-line arguments

五迷三道 提交于 2019-11-29 16:31:40
I have a C++ program to which I pass two doubles as inputs from the command line using int main(int argc, char *argv[]){ double a,b; a = atof(argv[1]); b = atof(argv[2]); further code..... I run the code on a cluster using the qsub utility and I have a Bash script named 'jobsub.sh` to submit the jobs which looks like this: #!/bin/csh -f hostname cd /home/roy/codes/3D # Change directory first -- replace Mysubdir set startdir = `pwd` # Remember the directory we're in if( ! -d /scratch/$USER ) then mkdir /scratch/$USER # Create scratch directory endif # If it does not exist #cp infile12 /scratch/

Cluster Failover

无人久伴 提交于 2019-11-29 15:44:21
I know I'm asking something very obvious about cluster failover. I read on redis.io that, if any master cluster node fails it will affect to other master nodes until slave come to take in charge. In my structure, I'm not defining any slave and just working with 3 masters. I'm thinking to modify the redis-trib.rb file, which will remove the defected server and will start the cluster with other 2 nodes. I'm confused about a couple of things, 1) Resharding Could not possible until failed server goes live 2) Minimum 3 node limitation for create cluster As per bit understanding, redis-trib.rb not

How to know deploy mode of PySpark application?

好久不见. 提交于 2019-11-29 10:36:29
I am trying to fix an issue with running out of memory, and I want to know whether I need to change these settings in the default configurations file ( spark-defaults.conf ) in the spark home folder. Or, if I can set them in the code. I saw this question PySpark: java.lang.OutofMemoryError: Java heap space and it says that it depends on if I'm running in client mode. I'm running spark on a cluster and monitoring it using standalone. But, how do I figure out if I'm running spark in client mode? If you are running an interactive shell, e.g. pyspark (CLI or via an IPython notebook), by default

How to run a job array in R using the rscript command from the command line? [closed]

雨燕双飞 提交于 2019-11-29 08:56:54
I am wondering how I might be able to run 500 parallel jobs in R using the Rscript function. I currently have an R file that has the header on top: args <- commandArgs(TRUE) B <- as.numeric(args[1]) Num.Cores <- as.numeric(args[2]) Outside of the R file, I wish to pass which of the 500 jobs are to be run, which is specified by B . Also, I would like to control the number of cores/CPUs available to each job, Num.Cores . I am wondering if there is software or guides that can allow this. I currently have a CentOS 7/Linux server and I know one way is to install Slurm. However, it is quite a hassle

LSF (bsub): how to specify a single “wrap-up” job to be run after all others finish?

给你一囗甜甜゛ 提交于 2019-11-29 07:19:24
BASIC PROBLEM: I want to submit N + 1 jobs to an LSF-managed Linux cluster in such a way that the ( N + 1)-st "wrap-up" job is not run until all the preceding N jobs have finished. EXTRA: If possible , it would be ideal if I could arrange matters so that the ( N + 1)-st ("wrap-up") job receives, as its first argument, a value of 0 (say) if all the previous N jobs terminated successfully, and a value different from 0 otherwise. This problem (or at least the part labeled "BASIC PROBLEM") is vastly simpler than what LSF's bsub appears to be designed to handle, so I have a hard time wading through

Injecting non-serializable application scoped bean as managed property of serializable session scoped bean in a cluster

有些话、适合烂在心里 提交于 2019-11-29 07:10:07
问题 I have the following managed beans : @ApplicationScoped public class ApplicationBean { // ... } @SessionScoped public class SessionBean implements Serializable { @ManagedProperty("#{applicationBean}") private ApplicationBean applicationBean; // ... } This is deployed to a server cluster with several nodes. What will happen when the HTTP session will be serialized on another node? ApplicationBean is not serialized because it doesn't implement Serializable . Will it be re-injected by

RabbitMQ cluster is not reconnecting after network failure

不羁岁月 提交于 2019-11-29 06:48:39
问题 I have RabbitMQ cluster with two nodes in production and the cluster is breaking with these error messages: =ERROR REPORT==== 23-Dec-2011::04:21:34 === ** Node rabbit@rabbitmq02 not responding ** ** Removing (timedout) connection ** =INFO REPORT==== 23-Dec-2011::04:21:35 === node rabbit@rabbitmq02 lost 'rabbit' =ERROR REPORT==== 23-Dec-2011::04:21:49 === Mnesia(rabbit@rabbitmq01): ** ERROR ** mnesia_event got {inconsistent_database, running_partitioned_network, rabbit@rabbitmq02} I tried to

ORA-01654: unable to extend index

前提是你 提交于 2019-11-29 06:07:04
问题 Calling all Oracle Gurus! I am in the process of clustering a well tested application on WebSphere. The application in question made it about half way through processing 1k of JMS messages from a queue before this happened. ---- Begin backtrace for Nested Throwables java.sql.SQLException: ORA-01654: unable to extend index DABUAT.INDEX1 by 128 in tablespace DABUAT_TBLSP at oracle.jdbc.driver.DatabaseError.throwSqlException(DatabaseError.java:112) at oracle.jdbc.driver.T4CTTIoer.processError

connection string to a SQL Server cluster

橙三吉。 提交于 2019-11-29 04:54:08
Could anyone point me or tell me how to write connection string to a SQL Server cluster instance? I want to establish trusted connection to a database (initial catalog) of the specific instance of a SQL Server cluster. I find for quite some time but can not find official answer from Google. I am asking for the connection string for C# ADO.Net code. I know how to write connection string to a normal one-machine non-cluster SQL Server instance. The connection string is written in the usual way; but in clustering you have 3 (or more) addresses: one for each node which constitues the cluster and

Create a cluster of co-workers' Windows 7 PCs for parallel processing in R?

放肆的年华 提交于 2019-11-29 03:33:08
问题 I am running the termstrc yield curve analysis package in R across 10 years of daily bond price data for 5 different countries. This is highly compute intensive, it takes 3200 seconds per country on a standard lapply, and if I use foreach and %dopar% (with doSNOW) on my 2009 i7 mac, using all 4 cores (8 with hyperthreading) I get this down to 850 seconds. I need to re-run this analysis every time I add a country (to compute inter-country spreads), and I have 19 countries to go, with many more