kubernetes-health-check

Are hitless rolling updates possible on GKE with externalTrafficPolicy: Local?

倾然丶 夕夏残阳落幕 提交于 2020-05-14 02:24:06
问题 I have a GKE cluster (1.12.10-gke.17). I'm running the nginx-ingress-controller with type: LoadBalancer . I've set externalTrafficPolicy: Local to preserve the source ip. Everything works great, except during rolling updates. I have maxSurge: 1 and maxUnavailable: 0 . My problem is that during a rolling update, I start getting request timeouts. I suspect the Google load balancer is still sending requests to the node where the pod is Terminating even though the health checks are failing. This

Multiple liveness probes in kuberenetes

此生再无相见时 提交于 2020-04-30 11:58:29
问题 I have a program which has multiple independent 1 components. It is trivial to add a liveness probe in all of the components, however it's not easy to have a single liveness probe which would determine the health of all of the program's components. How can I make kubernetes look at multiple liveness probes and restart the container when any of those are defunct? I know it can be achieved by adding more software, for example an additional bash script which does the liveness checks, but I am

Multiple liveness probes in kuberenetes

无人久伴 提交于 2020-04-30 11:56:08
问题 I have a program which has multiple independent 1 components. It is trivial to add a liveness probe in all of the components, however it's not easy to have a single liveness probe which would determine the health of all of the program's components. How can I make kubernetes look at multiple liveness probes and restart the container when any of those are defunct? I know it can be achieved by adding more software, for example an additional bash script which does the liveness checks, but I am

Kubernetes TCP Health Check

南笙酒味 提交于 2020-01-25 08:12:06
问题 I am building .NET core Console microservice and it has been suggested by Architect to use TCP Health Check instead of HTTP Health Check. Hence in order to implement TCP health Check, please find the below configuration that I have added in the OCP file(deploymentconfig section). Jenkins build was successful and also the deployment config roll out was successful. Query: How to ensure that probes are working properly.Is there a way to verify readiness and liveliness probes is being done in

Backup/Recover Kubernetes' current state (Namespaces/Pods/Controllers/etc)

白昼怎懂夜的黑 提交于 2020-01-01 19:31:08
问题 I would like to be able to get a description of my current state of the cluster so that in the future I would be able to recover from a failure. Aside from recreating all of the services from source/cli individually, what solutions are available? 回答1: I'd recommend Heptio Ark - https://github.com/heptio/ark. It's a general purpose disaster recovery solution for Kubernetes. It will back up all of your resources inside your cluster (pods, deployments, etc), and it can also take snapshots of

Setting up a Kuberentes cluster with HTTP Load balancing ingress for RStudio and Shiny results in error pages

一笑奈何 提交于 2020-01-01 05:41:09
问题 I'm attempting to create a cluster on Google Kubernetes Engine that runs nginx, RStudio server and two Shiny apps, following and adapting this guide. I have 4 workloads that are all green in the UI, deployed via: kubectl run nginx --image=nginx --port=80 kubectl run rstudio --image gcr.io/gcer-public/persistent-rstudio:latest --port 8787 kubectl run shiny1 --image gcr.io/gcer-public/shiny-googleauthrdemo:latest --port 3838 kubectl run shiny5 --image=flaviobarros/shiny-wordcloud --port=80 They

Kubernetes liveness probes with query string parameters

萝らか妹 提交于 2019-12-19 07:49:48
问题 I've looked over the documentation and browsed the source, but I can't seem to figure out how to do this. Is there any way to send query string parameters along with the path when implementing a Kubernetes liveness probe? The string I am sending, which looks something like this: /api/v1?q=... becomes URL-encoded and hits the server as: /api/v1%3fq=... As I have no such route on this particular API, I get a 404, and Kube reaps the pods after the allotted timeout. Is there any way to define

Monitoring a kubernetes job

一曲冷凌霜 提交于 2019-12-12 14:00:18
问题 I have kubernetes jobs that takes variable amount of time to complete. Between 4 to 8 minutes. Is there any way i can know when a job have completed, rather than waiting for 8 minutes assuming worst case. I have a test case that does the following: 1) Submits the kubernetes job. 2) Waits for its completion. 3) Checks whether the job has had the expected affect. Problem is that in my java test that submits the deployment job in the kubernetes, I am waiting for 8 minutes even if the job has

Why Kubernetes Pod gets into Terminated state giving Completed reason and exit code 0?

和自甴很熟 提交于 2019-12-10 18:28:21
问题 I am struggling to find any answer to this in the Kubernetes documentation. The scenario is the following: Kubernetes version 1.4 over AWS 8 pods running a NodeJS API (Express) deployed as a Kubernetes Deployment One of the pods gets restarted for no apparent reason late at night (no traffic, no CPU spikes, no memory pressure, no alerts...). Number of restarts is increased as a result of this. Logs don't show anything abnormal (ran kubectl -p to see previous logs, no errors at all in there)

GCE Ingress not picking up health check from readiness probe

我的未来我决定 提交于 2019-12-10 13:17:19
问题 When I create a GCE ingress, Google Load Balancer does not set the health check from the readiness probe. According to the docs (Ingress GCE health checks) it should pick it up. Expose an arbitrary URL as a readiness probe on the pods backing the Service. Any ideas why? Deployment: apiVersion: extensions/v1beta1 kind: Deployment metadata: name: frontend-prod labels: app: frontend-prod spec: selector: matchLabels: app: frontend-prod replicas: 3 strategy: rollingUpdate: maxSurge: 1