cluster-computing

Load Balancing (HAProxy or other) - Sticky Sessions

断了今生、忘了曾经 提交于 2019-11-27 19:08:16
I'm working on scaling out my app to multiple servers, and one requirement is that a client is always communicating with the same server (too much live data is used to allow bouncing between servers efficiently). My current setup is a small server cluster (using Linode). I have a frontend node running HAProxy using "balance source" so that an IP is always pointed towards the same node. I'm noticing that "balance source" is not a very even distribution. With my current test setup (2 backend servers), one server often has 3-4x as many connections when using a sample size of 80-100 source IPs. Is

connection string to a SQL Server cluster

廉价感情. 提交于 2019-11-27 18:47:12
问题 Could anyone point me or tell me how to write connection string to a SQL Server cluster instance? I want to establish trusted connection to a database (initial catalog) of the specific instance of a SQL Server cluster. I find for quite some time but can not find official answer from Google. I am asking for the connection string for C# ADO.Net code. I know how to write connection string to a normal one-machine non-cluster SQL Server instance. 回答1: The connection string is written in the usual

Singleton in Cluster environment

你离开我真会死。 提交于 2019-11-27 17:26:54
What is the best strategy to refactor a Singleton object to a cluster environment? We use Singleton to cache some custom information from Database. Its mostly read-only but gets refreshed when some particular event occurs. Now our application needs to be deployed in a Clustered environment. By definition, each JVM will have its own Singleton instance. So the cache may be out-of-sync between the JVM's when a refresh event occurs on a single node and its cache is refreshed. What is the best way to keep the cache's in sync? Thanks. Edit: The cache is mainly used to provide an autocomplete list

NodeJS|Cluster: How to send data from master to all or single child/workers?

|▌冷眼眸甩不掉的悲伤 提交于 2019-11-27 17:24:36
I have working (stock) script from node var cluster = require('cluster'); var http = require('http'); var numReqs = 0; if (cluster.isMaster) { // Fork workers. for (var i = 0; i < 2; i++) { var worker = cluster.fork(); worker.on('message', function(msg) { if (msg.cmd && msg.cmd == 'notifyRequest') { numReqs++; } }); } setInterval(function() { console.log("numReqs =", numReqs); }, 1000); } else { // Worker processes have a http server. http.Server(function(req, res) { res.writeHead(200); res.end("hello world\n"); // Send message to master process process.send({ cmd: 'notifyRequest' }); })

Easy way to use parallel options of scikit-learn functions on HPC

倖福魔咒の 提交于 2019-11-27 17:18:32
In many functions from scikit-learn implemented user-friendly parallelization. For example in sklearn.cross_validation.cross_val_score you just pass desired number of computational jobs in n_jobs argument. And for PC with multi-core processor it will work very nice. But if I want use such option in high performance cluster (with installed OpenMPI package and using SLURM for resource management) ? As I know sklearn uses joblib for parallelization, which uses multiprocessing . And, as I know (from this, for example, Python multiprocessing within mpi ) Python programs parallelized with

Run a hadoop cluster on docker containers

痴心易碎 提交于 2019-11-27 16:33:51
问题 I want to run a multi-node hadoop cluster, with each node inside a docker container on a different host. This image - https://github.com/sequenceiq/hadoop-docker works well to start hadoop in a pseudo distributed mode, what is the easiest way to modify this to have each node in a different container on a separate ec2 host? 回答1: I did this with two containers running master and slave nodes on two different ubuntu hosts. I did the networking between containers using weave. I have added the

How can I create a local multi-node Cassandra cluster on Windows 7 64 bit?

旧时模样 提交于 2019-11-27 15:25:38
I am looking for a set of instructions to create a local multi-node Cassandra 2.x cluster on a Window 7 64 bit PC. It should preferably use CCM “Cassandra Cluster Manager” and allow management using DataStax OpsCenter I followed the instructions in “ Getting Started with Apache Cassandra on Windows the Easy Way ” but they are for a single node cluster. EDIT: I got stuck on deploying OpsCenter agents on each node using CCM, any ideas? Articles used for this tutorial: CCM 2.0 and Windows Cassandra Wiki - Windows Development Setting up a multi-node Cassandra cluster on a single Windows machine

How to add a new node to my Elasticsearch cluster

冷暖自知 提交于 2019-11-27 14:29:06
问题 My cluster has a yellow health as it has only one single node, so the replicas remain unasigned simply because no other node is available to contain them. So I want to create/add another node so Elasticsearch can begin allocating replica’s to it. I've only one machine and I'm running ES as a service . I've found tons of site with some info but none of them is giving me clearly how can I add another node to ES. Can someone explain me which files do I've to edit and what commands do I've to

Running TensorFlow on a Slurm Cluster?

僤鯓⒐⒋嵵緔 提交于 2019-11-27 12:23:28
问题 I could get access to a computing cluster, specifically one node with two 12-Core CPUs, which is running with Slurm Workload Manager. I would like to run TensorFlow on that system but unfortunately I were not able to find any information about how to do this or if this is even possible. I am new to this but as far as I understand it, I would have to run TensorFlow by creating a Slurm job and can not directly execute python/tensorflow via ssh. Has anyone an idea, tutorial or any kind of source

How to change memory per node for apache spark worker

允我心安 提交于 2019-11-27 12:03:33
I am configuring an Apache Spark cluster. When I run the cluster with 1 master and 3 slaves, I see this on the master monitor page: Memory 2.0 GB (512.0 MB Used) 2.0 GB (512.0 MB Used) 6.0 GB (512.0 MB Used) I want to increase the used memory for the workers but I could not find the right config for this. I have changed spark-env.sh as below: export SPARK_WORKER_MEMORY=6g export SPARK_MEM=6g export SPARK_DAEMON_MEMORY=6g export SPARK_JAVA_OPTS="-Dspark.executor.memory=6g" export JAVA_OPTS="-Xms6G -Xmx6G" But the used memory is still the same. What should I do to change used memory? When using