load-balancing

How does Windows Azure perform load balancing?

試著忘記壹切 提交于 2019-12-20 17:31:22
问题 I was trying to read up on load balancing in Windows Azure, and all the information about it is extremely vague and non-specific. All I really want is a simple answer: how does Azure perform load balancing? Many applications use sessions or in-memory data to increase performance. With Azure, I'm not sure if this is possible, because (from what I can tell), Azure doesn't have any sort of "sticky" sessions. Even if it wasn't session-based, but simply IP-based balancing, that would be enough for

Do Azure Websites use sticky-ip load balancing?

懵懂的女人 提交于 2019-12-20 15:29:19
问题 I just stood up a website using the Windows Azure Websites preview. After doing so, I ran YSlow to make sure the score is what I expect it to be. When doing so, I get a message that reads "Use cookiless domains". Well, this is just an informational website, we don't even use session. So I check the http request and there's a cookie in there named "ARRAffinity". Some quick googling turns up this link: http://social.msdn.microsoft.com/Forums/en-US/windowsazurewebsitespreview/thread/0ba2c2f6

Writing a weighted load balancing algorithm

依然范特西╮ 提交于 2019-12-20 15:09:24
问题 I've to write a weighted load balancing algorithm and I'm looking for some references. Is there any book ? that you can suggest to understand such algorithms. Thanks! 回答1: A simple algorithm here isn't that complicated. Let's say you have a list of servers with the following weights: A 10 B 20 C 30 Where the higher weight represents it can handle more traffic. Just divide the amount of traffic sent to each server by the weight and sort smallest to largest. The server that comes out on top

Storm-Kafka multiple spouts, how to share the load?

巧了我就是萌 提交于 2019-12-20 12:38:09
问题 I am trying to share the task among the multiple spouts. I have a situation, where I'm getting one tuple/message at a time from external source and I want to have multiple instances of a spout, main intention behind is to share the load and increase performance efficiency. I can do the same with one Spout itself, but I want to share the load across multiple spouts. I am not able to get the logic to spread the load. Since the offset of messages will not be known until the particular spout

“Work stealing” vs. “Work shrugging”?

谁说我不能喝 提交于 2019-12-20 12:02:57
问题 Why is it that I can find lots of information on "work stealing" and nothing on "work shrugging" as a dynamic load-balancing strategy? By "work-shrugging" I mean pushing surplus work away from busy processors onto less loaded neighbours, rather than have idle processors pulling work from busy neighbours ("work-stealing"). I think the general scalability should be the same for both strategies. However I believe that it is much more efficient, in terms of latency & power consumption, to wake an

“Work stealing” vs. “Work shrugging”?

☆樱花仙子☆ 提交于 2019-12-20 12:00:31
问题 Why is it that I can find lots of information on "work stealing" and nothing on "work shrugging" as a dynamic load-balancing strategy? By "work-shrugging" I mean pushing surplus work away from busy processors onto less loaded neighbours, rather than have idle processors pulling work from busy neighbours ("work-stealing"). I think the general scalability should be the same for both strategies. However I believe that it is much more efficient, in terms of latency & power consumption, to wake an

Enable HTTPS on GCE/GKE

亡梦爱人 提交于 2019-12-20 11:54:07
问题 I am running web site with Kubernetes on Google Cloud. At the moment, everything is working well - through http. But I need https. I have several services and one of them is exposed to the outside world, let's call it web. As far as I know, this is the only service that needs to be modified. I tried to creating a static IP and TCP/SSL loadbalancer ssl-LB in the Networking section of GCP and using that LB in web.yaml, which I create. Creating the service gets stuck with: Error creating load

Enable HTTPS on GCE/GKE

无人久伴 提交于 2019-12-20 11:54:03
问题 I am running web site with Kubernetes on Google Cloud. At the moment, everything is working well - through http. But I need https. I have several services and one of them is exposed to the outside world, let's call it web. As far as I know, this is the only service that needs to be modified. I tried to creating a static IP and TCP/SSL loadbalancer ssl-LB in the Networking section of GCP and using that LB in web.yaml, which I create. Creating the service gets stuck with: Error creating load

Zero downtime deployment for Java apps

拟墨画扇 提交于 2019-12-20 10:39:09
问题 I am trying to build the very lightweight solution for zero downtime deployment for Java apps. For the sake of simplicity lets think that we have two servers. My solution is to use: On the "front" -- some load balancer (software) - I am thinking about HAProxy here. On the "back" - two servers, both running Tomcat with deployed application. When we are about to deploy new release We disable one of the servers with HAProxy, so only one server (let's call it server A, which is running old

Why does Elastic Load Balancing report 'Out of Service'?

别来无恙 提交于 2019-12-20 10:21:40
问题 I am trying to set up Elastic Load Balancing (ELB) in AWS to split the requests between multiple instances. I have created several images of my webserver based on the same AMI, and I am able to ssh into each individually and access the site via each distinct public DNS. I have added each of my instances to the load balancer, but they all come back with the Status: Out of Service because they failed the health check. I'm mostly confused because I can access each instance from its public DNS,