cluster-computing

Spark to MongoDB via Mesos

北城以北 提交于 2019-12-12 04:59:13
问题 I am trying to connect Apache Spark to MongoDB using Mesos. Here is my architecture: - MongoDB: MongoDB Cluster of 2 shards, 1 config server and 1 query server. Mesos: 1 Mesos Master, 4 Mesos slaves Now I have installed Spark on just 1 node. There is not much information available on this out there. I just wanted to pose a few questions: - As per what I understand, I can connect Spark to MongoDB via mesos. In other words, I end up using MongoDB as a storage layer. Do I really Need Hadoop? Is

Got only one node after install Kubernetes cluster

强颜欢笑 提交于 2019-12-12 04:45:22
问题 Followed official guide to install Kubernetes cluster with kubeadm on Vagrant. https://kubernetes.io/docs/getting-started-guides/kubeadm/ master node1 node2 Master # kubeadm init --apiserver-advertise-address=192.168.33.200 # sudo cp /etc/kubernetes/admin.conf $HOME/ # sudo chown $(id -u):$(id -g) $HOME/admin.conf # export KUBECONFIG=$HOME/admin.conf # wget https://raw.githubusercontent.com/coreos/flannel/master/Documentation/kube-flannel.yml # kubectl apply -f kube-flannel.yaml Node1 and

IPython parallel on SGE cluster : Scikit-learn SVC parallel import causes engines to crash

你说的曾经没有我的故事 提交于 2019-12-12 03:44:50
问题 I use a SGE cluster with IPcontroller running on the head node, and ~50 engines running on the other nodes (submitted using QSUB). The engines are able to connect and register with the controller without any issues. I can also connect to the head node using SSH and view the engines IDs and running simple code. For e.g., this works perfectly well : %px %pylab inline parallel_result = lbView.map_sync(lambda x: x*rand(), range(32)) However, when I try to run the following line, then the engines

ipython with MPI clustering using machinefile

邮差的信 提交于 2019-12-12 03:33:00
问题 I have successfully configured mpi with mpi4py support across three nodes, as per testing of the hellowworld.py script in the mpi4py demo directory: gms@host:~/development/mpi$ mpiexec -f machinefile -n 10 python ~/development/mpi4py/demo/helloworld.py Hello, World! I am process 3 of 10 on host. Hello, World! I am process 1 of 10 on worker1. Hello, World! I am process 6 of 10 on host. Hello, World! I am process 2 of 10 on worker2. Hello, World! I am process 4 of 10 on worker1. Hello, World! I

[Storm][DRPC] Request failed

让人想犯罪 __ 提交于 2019-12-12 03:32:24
问题 We work with storm and use the DRPC to execute some algorithms on the cluster. When the duration time on each node is less than 60 seconds, there is no trouble: client receives correct result. However when we have to solve bigger problem with the same algorithm (then duration time is more than 60 seconds) we have the following message: Exception in thread "main" DRPCExecutionException(msg:Request failed) at backtype.storm.generated.DistributedRPC$execute_result$execute_resultStandardScheme

SGE Cluster - script fails after submission - works in terminal

谁说我不能喝 提交于 2019-12-12 03:16:35
问题 I have a script that I am trying to submit to a SGE cluster (on Redhat Linux). The very first part of the script defines the current folder from the full CWD path, as a variable to use downstream: #!/usr/bin/bash # #$ -cwd #$ -A username #$ -M user@server #$ -j y #$ -m aes #$ -N test #$ -o test.log.txt echo 'This is a test.' result="${PWD##*/}" echo $result In bash, this works as expected: CWD: -bash-4.1$ pwd /home/user/test Run script: -bash-4.1$ bash test.sh This is a test. test When I

How to utilize cluster processing on loops?

我们两清 提交于 2019-12-12 02:52:28
问题 Okay, I made this example on JSfiddle to mimic (really bad job), of my dilemma: https://jsfiddle.net/Dillybob/7zkL097q/2/ In the cluster module for nodejs, I want to utilize my child workers (processing power) as much as I can to help accomplish and complete my setInterval MonsterLoop faster . In my example above if I were to run that code inside my cluster's isWorker block, it would run 6 times every second. That seems counter-intuitive and would seem like it would use more processing power?

My gather Buffer is Empty?

风格不统一 提交于 2019-12-12 02:43:31
问题 I have problem gathering data from all process to process master "root" I can send the data MPI_Bcast but the on MPI_Gather I have all the problem in my countBuff I debug my output and this is what I have output brodcast data of 0 brodcast data of 1 MPI_Gather data rank 0 1 from 0 to 1.00 KM:-842150451,from 1.00 to 2.00 KM:-842150451,from 2.00 to 5.00 KM:-842150451,grater than 5.00 KM:-842150451 MPI_Type_free1 delete countBuff MPI_Finalize brodcast data of 2 MPI_Gather data rank 0 0 MPI

is it possible to a vCPU to use different CPUs from two different hardware computers

大兔子大兔子 提交于 2019-12-12 02:13:21
问题 I'v searched about this but i don't seem to get fair answer. lets say i wan't to create a vm that has a vCPU, and that vCPU must have 10 cores but i only have 2 computers with 5 cores of physical CPU for each. is it possible to create one vCPU by relaying on these two physical CPUs to perform like regular one physical CPU? Update 1: lets say i'm using virtualBox, and the term vCPU is referring to virtual cpu, and it's a well known term. Update 2: i'm asking this because i'm doing a little

Wildfly clustering with VirtualBox

北城以北 提交于 2019-12-12 01:55:38
问题 I am using VirtualBox on a WINDOWS7 as host of two DEBIAN7.7 guests, deb1 and deb2 . Each guest can comunicate with the other one. Using one guest browser I can see the Wildfly istance welcome page that's running on the other guest. I run each istance in standalone-ha mode, network interfaces have mutlicast enabled, I can see on Wildfly node named srv1 that the two istances build a cluster: ... ...ISPN000094: Received new cluster view: [srv2/web|3] (2) [srv2/web, srv1/web] where srv1 and srv2