环境准备
部署 httpbin 服务:
 kubectl apply -f samples/httpbin/httpbin.yaml
       复制代码
 
部署 sleep 服务:
 kubectl apply -f samples/sleep/sleep.yaml 
       复制代码
 
httpbin 服务作为接收请求的服务端, sleep 服务作为发送请求的客户端。
设置超时时间
在 sleep 服务中向 httpbin 服务的发出请求:
 export SLEEP_POD=$(kubectl get pods -l app=sleep -o 'jsonpath={.items[0].metadata.name}')kubectl exec "$SLEEP_POD" -c sleep -- time curl -o /dev/null -sS -w "%{http_code}\n" http://httpbin.org/delay/5
       复制代码
 
返回结果如下:
 200real    0m 5.69suser    0m 0.00ssys     0m 0.00s
       复制代码
 
可以看到,请求大约在 5 秒返回 200 (OK)。
创建虚拟服务,访问httpbin 服务时,请求超时设置为 3 秒:
 kubectl apply -f - <<EOFapiVersion: networking.istio.io/v1alpha3kind: VirtualServicemetadata:  name: httpbinspec:  hosts:  - httpbin  http:  - route:    - destination:        host: httpbin    timeout: 3sEOF
       复制代码
 
再此次访问,返回结果如下:
 504real    0m 3.01suser    0m 0.00ssys     0m 0.00s
       复制代码
 
可以看到,在 3 秒后出现了 504 (Gateway Timeout)。 Istio 在 3 秒后切断了响应时间为 5 秒的httpbin 服务的请求。接下来,我们深入地看一下,Istio 是怎么切断请求的?
查看 Envoy 日志
执行一下命令,查看sleep 服务的 Envoy 日志:
 kubectl logs -l app=sleep -c istio-proxy
       复制代码
 
可以看到sleep服务对httpbin服务的调用的日志:
 {     "authority": "httpbin:8000",     "bytes_received": 0,     "bytes_sent": 24,     "connection_termination_details": null,     "downstream_local_address": "172.24.146.239:8000",     "downstream_remote_address": "172.24.158.25:40384",     "duration": 3001,     "method": "GET",     "path": "/delay/5",     "protocol": "HTTP/1.1",     "request_id": "5ef38816-7f49-48c8-9627-2416e1716293",     "requested_server_name": null,     "response_code": 504,     "response_code_details": "upstream_response_timeout",     "response_flags": "UT",     "route_name": null,     "start_time": "2022-07-01T09:40:13.882Z",     "upstream_cluster": "outbound|8000||httpbin.onemore.svc.cluster.local",     "upstream_host": "172.24.158.96:80",     "upstream_local_address": "172.24.158.25:32846",     "upstream_service_time": null,     "upstream_transport_failure_reason": null,     "user_agent": "curl/7.81.0-DEV",     "x_forwarded_for": null}
       复制代码
 
其中,response_flags为UT,表示上游(upstream)请求超时,也就是sleep服务检测到了httpbin服务的请求超时。
执行一下命令,查看httpbin 服务的 Envoy 日志:
 kubectl logs -l app=httpbin -c istio-proxy
       复制代码
 
可以看到httpbin服务被sleep服务调用的 Envoy 日志:
 {     "authority": "httpbin:8000",     "bytes_received": 0,     "bytes_sent": 0,     "connection_termination_details": null,     "downstream_local_address": "172.24.158.96:80",     "downstream_remote_address": "172.24.158.25:32846",     "duration": 2997,     "method": "GET",     "path": "/delay/5",     "protocol": "HTTP/1.1",     "request_id": "5ef38816-7f49-48c8-9627-2416e1716293",     "requested_server_name": "outbound_.8000_._.httpbin.onemore.svc.cluster.local",     "response_code": 0,     "response_code_details": "downstream_remote_disconnect",     "response_flags": "DC",     "route_name": "default",     "start_time": "2022-07-01T09:40:13.885Z",     "upstream_cluster": "inbound|80||",     "upstream_host": "172.24.158.96:80",     "upstream_local_address": "127.0.0.6:35701",     "upstream_service_time": null,     "upstream_transport_failure_reason": null,     "user_agent": "curl/7.81.0-DEV",     "x_forwarded_for": null}
       复制代码
 
其中,response_flags为DC,表示下游(downstream)连接中断,也就是sleep服务的调用请求被中断了。
深入分析
通过 Envoy 日志,我们可以做出一些分析和判断:
当httpbin服务的请求正常的时候,调用过程如下图:
当httpbin服务的请求超时的时候,调用过程如下图:
虽然,我们在httpbin服务上设置的请求超时时间,但实际上主动断开请求的却是sleep服务的 Envoy。
清理
 kubectl delete virtualservice httpbinkubectl delete -f samples/httpbin/httpbin.yamlkubectl delete -f samples/sleep/sleep.yaml 
       复制代码
 
最后,感谢你这么帅,还给我点赞。
评论