• envoy部分五:envoy的动态配置


    一、envoy动态配置介绍

    动态资源,是指由envoy通过xDS协议发现所需要的各项配置的机制,相关的配置信息保存 于称之为管理服务器(Management Server )的主机上,经由xDS API向外暴露;下面是一个 纯动态资源的基础配置框架。

    {
    "lds_config": "{...}",
    "cds_config": "{...}",
    "ads_config": "{...}"
    }

    xDS API为Envoy提供了资源的动态配置机制,它也被称为Data Plane API。

    Envoy支持三种类型的配置信息的动态发现机制,相关的发现服务及其相应的API联合起来 称为xDS API。

    1)基于文件系统发现:指定要监视的文件系统路径
    2)通过查询一到多个管理服务器(Management Server)发现:通过DiscoveryRequest协议报文发送请求,并要求服务方以DiscoveryResponse协议报文进行响应
    (1)gRPC服务:启动gRPC流
    (2)REST服务:轮询REST-JSON URL

    v3 xDS支持如下几种资源类型

    envoy.config.listener.v3.Listener
    envoy.config.route.v3.RouteConfiguration
    envoy.config.route.v3.ScopedRouteConfiguration
    envoy.config.route.v3.VirtualHost
    envoy.config.cluster.v3.Cluster
    envoy.config.endpoint.v3.ClusterLoadAssignment
    envoy.extensions.transport_sockets.tls.v3.Secret
    envoy.service.runtime.v3. Runtime

    二、DS API介绍

    Envoy对xDS API的管理由后端服务器实现,包括LDS、CDS、RDS、SRDS(Scoped Route)、VHDS (Virtual Host)、EDS、SDS、RTDS(Runtime )等。

    1)所有这些API都提供了最终的一致性,并且彼此间不存在相互影响;
    2)部分更高级别的操作(例如执行服务的A/B部署)需要进行排序以防止流量被丢弃,因此,基于一个管理服务器提供多类API时还需要使用聚合发现服务(ADS )API。
       ADS API允许所有其他API通过来自单个管理服务器的单个gRPC双向流进行编组,从而允许对操作进行确定性排序

    xDS的各API还支持增量传输机制,包括ADS

     

    三、Bootstrap node

    一个Management Server 实例可能需要同时响应多个不同的Envoy实例的资源发现请求。

    1)Management Server上的配置需要为适配到不同的Envoy实例
    2)Envoy 实例请求发现配置时,需要在请求报文中上报自身的信息
    (1)例如id、cluster、metadata和locality等
    (2)这些配置信息定义在Bootstrap配置文件中

    专用的顶级配置段“node{…}”中配置

    node:
      id: … # An opaque node identifier for the Envoy node. 
      cluster: … # Defines the local service cluster name where Envoy is running.
      metadata: {…} # Opaque metadata extending the node identifier. Envoy will pass this directly to the management server .
      locality: # Locality specifying where the Envoy instance is running.
        region: …
        zone: …
        sub_zone: …
      user_agent_name: … # Free-form string that identifies the entity requesting config. E.g . “envoy ” or “grpc”
      user_agent_version: … # Free-form string that identifies the version of the entity requesting config. E.g . “1.12.2” or “abcd1234” , or “SpecialEnvoyBuild ” 
      user_agent_build_version: # Structured version of the entity requesting config.
        version: …
        metadata: {…}
      extensions: [ ] # List of extensions and their versions supported by the node.
      client_features: [ ]
      listening_addresses: [ ] # Known listening ports on the node as a generic hint to the management server for filtering listeners to be returned

    四、API的流程

    1、对于典型的HTTP路由方案,xDS API的Management Server 需要为其客户端(Envoy实例)配 置的核心资源类型为Listener、RouteConfiguration、Cluster和ClusterLoadAssignment四个。每个Listener资源可以指向一个RouteConfiguration资源,该资源可以指向一个或多个Cluster资源,并且每个Cluster资源可以指向一个ClusterLoadAssignment资源。

    2、Envoy实例在启动时请求加载所有Listener和Cluster资源,而后,再获取由这些Listener和 Cluster所依赖的RouteConfiguration和ClusterLoadAssignment配置;此种场景中,Listener资源和Cluster资源分别代表着客户端配置树上的“根(root)”配置,因而 可并行加载。

    3、类型gRPC一类的非代理式客户端可以仅在启动时请求加载其感兴趣的Listener资源, 而后再加载这些特定Listener相关的RouteConfiguration资源;再然后,是这些 RouteConfiguration资源指向的Cluster资源,以及由这些Cluster资源依赖的 ClusterLoadAssignment资源;该种场景中,Listener资源是客户端整个配置树的“根”。

    五、Envoy的配置方式

    Envoy的架构支持非常灵活的配置方式:简单部署场景可以使用纯静态配置,而更复 杂的部署场景则可以逐步添加需要的动态配置机制

    1、纯静态配置:用户自行提供侦听器、过滤器链、集 群及HTTP路由(http代理场景),上 游端点的 发 现仅可通过DNS服务进行,且配置的重新加载必须通过内置的热 重启( hot restart)完成;
    
    2、仅使用EDS:EDS提供的端点发现功能可有效规避DNS的限制( 响应中的 最大记录 数等);
    
    3、使用EDS和CDS:CDS能够让Envoy以优雅的方式添加、更新和删除 上游集群 ,于是, 初始配置 时, Envoy无须事先了解所有上游集群;
    
    4、EDS、CDS和RDS:动态发现路由配置;RDS与EDS、CDS一起使用时 ,为用户 提供了构 建复杂路 由 拓扑的能力(流量转移、蓝/绿部署等);
    
    5、EDS、CDS、RDS和LDS:动态发现侦听器配置,包括内嵌的过滤 器链;启 用此四种 发现服务 后,除 了较罕见的配置变动、证书轮替或更新Envoy程序之外,几 乎无须再 热重启Envoy;
    
    6、EDS、CDS、RDS、LDS和SDS:动态发现侦听器密钥相关的证书、 私钥及TLS会 话票据, 以及对证 书 验证逻辑的配置(受信任的根证书和撤销机制等 );

    六、Envoy资源的配置源(ConfigSource)

    1、配置源(ConfigSource)用于指定资源配置数据的来源,用于为Listener、Cluster、Route、 Endpoint、Secret和VirtualHost等资源提供配置数据。

    2、目前,Envoy支持的资源配置源只能是path、api_config_source或ads其中之一。

    3、api_config_source或ads的数据来自于xDS API Server,即Management Server。

     

    七、基于文件系统的订阅

    为Envoy提供动态配置的最简单方法是将其放置在ConfigSource中显式指定的文件路径中

    1)Envoy将使用inotify(Mac OS X上的kqueue)来监视文件的更改,并在更新时解析文件中的DiscoveryResponse 报文
    2)二进制protobufs,JSON,YAML和proto文本都是DiscoveryResponse 所支持的数据格式
    
    提示
    1)除了统计计数器和日志以外,没有任何机制可用于文件系统订阅ACK/NACK更新
    2)若配置更新被拒绝,xDS API的最后一个有效配置将继续适用

    1、基于eds实现Envoy基本全动态的配置方式

    以EDS为例,Cluster为静态定义,其各Endpoint通过EDS动态发现

    集群定义格式

    # Cluster中的endpoint配置格式
    
    clusters:
    - name:
      ...
      eds_cluster_config:
        service_name:
        eds_config:
          path: ... # ConfigSource,支持使用path, api_config_source或ads三者之一;

    cluster的配置

    # 类似如下纯静态格式的Cluster定义
    clusters:
    - name: webcluster
      connect_timeout: 0.25s
      type: STATIC #类型为静态
      lb_policy: ROUND_ROBIN
      load_assignment:
        cluster_name: webcluster
        endpoints:
        - lb_endpoints:
          - endpoint:
            address:
              socket_address:
                address: 172.31.11.11
                port_value: 8080
    
    #使用了EDS的配置
    clusters:
    - name: targetCluster
      connect_timeout: 0.25s
      lb_policy: ROUND_ROBIN
      type: EDS #类型为EDS
      eds_cluster_config:
        service_name: webcluster
        eds_config:
          path: '/etc/envoy/eds.yaml' # 指定订阅的文件路径
    #提示:文件后缀名为conf,则资源要以json格式定义;文件后缀名为yaml,则资源需要以yaml格式定义;另外,动态配置中,各Envoy实例需要有惟的id标识;

    EDS的配置

    1)文件/etc/envoy/eds.yaml中以Discovery Response报文的格式给出响应实例,例如,下面的 配置示例用于存在地址172.31.11.11某上游服务器可提供服务时

    2) 响应报文需要以JSON格式给出

    resources:
      - "@type": type.googleapis.com/envoy.config.endpoint.v3.ClusterLoadAssignment
        cluster_name: webcluster
        endpoints:
        - lb_endpoints:
          - endpoint:
            address:
              socket_address:
                address: 172.31.11.11
                port_value: 8080

    随后,修改该文件,将172.31.11.12也添加进后端端点列表中,模拟配置变动

    - endpoint:
            address:
              socket_address:
                address: 172.31.11.12
                port_value: 8080

    2、基于lds和cds实现Envoy基本全动态的配置方式

    1)各Listener的定义以Discovery Response 的标准格式保存于一个文件中。

    2) 各Cluster的定义同样以Discovery Response的标准格式保存于另一文件中。

    如下面Envoy Bootstrap配置文件示例中的文件引用

    node:
      id: envoy_front_proxy 
      cluster: MageEdu_Cluster
    
    admin:
      profile_path: /tmp/envoy.prof 
      access_log_path: /tmp/admin_access.log
      address: 
        socket_address: 
          address: 0.0.0.0
          port_value: 9901
    
    dynamic_resources:
      lds_config:
        path: /etc/envoy/conf.d/lds.yaml 
      cds_config: 
        path: /etc/envoy/conf.d/cds.yaml 

    lds的Discovery Response格式的配置示例如下

    resources:
    - "@type": type.googleapis.com/envoy.config.listener.v3.Listener
      name: listener_http
      address: 
        socket_address: { address: 0.0.0.0, port_value: 80 }
      filter_chains:
      - filters:
        name: envoy.http_connection_manager
        typed_config: 
          "@type":type.googleapis.com/envoy.extensions.filters.network.http_connection_manag
    er.v3.HttpConnectionManager
          stat_prefix: ingress_http
          route_config:
            name: local_route 
            virtual_hosts:
            - name: local_service
              domains: ["*"]
              routes:
              - match:
                  prefix: "/"
                route:
                  cluster: webcluster
          http_filters:
          - name: envoy.filters.http.router

    cds的Discovery Response格式的配置示例如下

    resources:
    - "@type": type.googleapis.com/envoy.config.cluster.v3.Cluster
      name: webcluster
      connect_timeout: 1s
      type: STRICT_DNS
      load_assignment:
      cluster_name: webcluster
        endpoints:
        - lb_endpoints:
          - endpoint:
              address: 
                socket_address: 
                  address: webserver01
                  port_value: 8080
          - endpoint:
              address: 
                socket_address: 
                  address: webserver02
                  port_value: 8080

    八、基于gRPC的动态配置

    1、gRPC的介绍

    1)、Enovy支持为每个xDS API独立指定gRPC ApiConfigSource,它指向与管理服务器对应的某 上游集群。

    (1)这将为每个xDS资源类型启动一个独立的双向gRPC流,可能会发送给不同的管理服务器
    (2)每个流都有自己独立维护的 资源版本,且不存在跨资源类型的共享版本机制;
    (3)在不使用ADS的情况下,每个资源类型可能具有不同的版本,因为Envoy API允许指向不同的 EDS/RDS资源配置并对应不同的ConfigSources

     

    2)、API的交付方式采用最终一致性机制。

    2、基于gRPC的动态配置格式

    以LDS为例,它配置Listener以动态方式发现和加载,而内部的路由可由发现的Listener直接 提供,也可配置再经由RDS发现。

    下面为LDS配置格式,CDS等的配置格式类同:

    dynamic_resouces:
      lds_config:
      api_config_source:
        api_type: ... # API 可经由REST或gRPC获取,支持的类型包括REST、gRPC和delta_gRPC
        resource_api_version: ... # xDS 资源的API版本,对于1.19及之后的Envoy版本,要使用v3;
        rate_limit_settings: {...} # 速率限制
        grpc_services: # 提供grpc服务的一到多个服务源
          transport_api_version : ... # xDS 传输协议使用的API版本,对于1.19及之后的Envoy版本,要使用v3;
          envoy_grpc: # Envoy内建的grpc客户端,envoy_grpc和google_grpc二者仅能用其一;
            cluster_name: ... # grpc集群的名称;
          google_grpc: # Google的C++ grpc客户端
          timeout: ... # grpc超时时长;
         
    注意:
    提供gRPC API服务的Management Server (控制平面)也需要定义为Envoy上的集群, 并由envoy实例通过xDS API进行请求;
    (1)通常,这些管理服务器需要以静态资源的格式提供;
    (2)类似于,DHCP协议的Server 端的地址必须静态配置,而不能经由DHCP协议获取;

    3、基于GRPC管理服务器订阅

    基于gRPC的订阅功能需要向专用的Management Server请求配置信息。

    下面的示例配置使用了lds和cds 分别动态获取Listener和Cluster相关的配置。

    node:
      id: envoy_front_proxy 
      cluster: webcluster
      
    admin:
      profile_path: /tmp/envoy.prof
      access_log_path: /tmp/admin_access.log
      address:
        socket_address:
          address: 0.0.0.0
          port_value: 9901
    
    dynamic_resources: 
      lds_config:  #定义动态获取listener
        resource_api_version: V3
        api_config_source:
          api_type: GRPC  #使用GRPC模式
          transport_api_version: V3
          grpc_services:
          - envoy_grpc:
            cluster_name: xds_cluster #调用grpc服务器
            
      cds_config: #定义动态获取cluster
        resource_api_version: V3
        api_config_source:
          api_type: GRPC  #使用GRPC模式
          transport_api_version: V3
          grpc_services:
          - envoy_grpc:
            cluster_name: xds_cluster #调用grpc服务器
    
    #xds_cluster需要静态配置
    static_resources:
      clusters:
        - name: xds_cluster
          connect_timeout: 0.25s
          type: STRICT_DNS
          # Used to provide extension-specific protocol options for upstream connections. 
          typed_extension_protocol_options: 
            envoy.extensions.upstreams.http.v3.HttpProtocolOptions: 
              "@type": type.googleapis.com/envoy.extensions.upstreams.http.v3.HttpProtocolOptions 
               explicit_http_config:
               http2_protocol_options: {}
          lb_policy: ROUND_ROBIN
          load_assignment:
            cluster_name: xds_cluster
            endpoints:
            - lb_endpoints:
              - endpoint:
                  address:
                    socket_address:
                      address: xdsserver-IP
                      port_value: 18000

    4、ADS

    通过上面的交互顺序保证MS资源分发时的流量丢弃是一项很有挑战的工作,而ADS允许单 一MS通过单个gRPC流提供所有的API更新。

    1)配合仔细规划的更新顺序,ADS可规避更新过程中流量丢失
    2)使用 ADS,在单个流上可通过类型 URL 来进行复用多个独立的 DiscoveryRequest/DiscoveryResponse 序列

     

    使用了ads分别动态获取Listener和Cluster相关的配置

    node:
      id: envoy_front_proxy 
      cluster: webcluster
      
    admin:
      profile_path: /tmp/envoy.prof
      access_log_path: /tmp/admin_access.log
      address:
        socket_address:
          address: 0.0.0.0
          port_value: 9901
    
    dynamic_resources:
      ads_config:
        api_type: GRPC
        transport_api_version: V3
        grpc_services:
        - envoy_grpc:
          cluster_name: xds_cluster
          set_node_on_first_message_only: true
    
      cds_config:
        resource_api_version: V3
        ads: {}
        lds_config:
          resource_api_version: V3
          ads: {}
     
    
    
     #xds_cluster需要静态配置
    static_resources:
      clusters:
        - name: xds_cluster
          connect_timeout: 0.25s
          type: STRICT_DNS
          # Used to provide extension-specific protocol options for upstream connections. 
          typed_extension_protocol_options: 
            envoy.extensions.upstreams.http.v3.HttpProtocolOptions: 
              "@type": type.googleapis.com/envoy.extensions.upstreams.http.v3.HttpProtocolOptions 
               explicit_http_config:
               http2_protocol_options: {}
          lb_policy: ROUND_ROBIN
          load_assignment:
            cluster_name: xds_cluster
            endpoints:
            - lb_endpoints:
              - endpoint:
                  address:
                    socket_address:
                      address: xdsserver-IP
                      port_value: 18000

    九、REST-JSON轮询订阅

    1、REST-JSON介绍

    1)、通过REST端点进行的同步(长)轮询也可用于xDS单例API

    2)、上面的消息顺序是类似的,除了没有维护到管理服务器的持久流

    3)、预计在任何时间点只有一个未完成的请求,因此响应nonce在REST-JSON中是可选的 p proto3的JSON规范转换用于编码DiscoveryRequest和DiscoveryResponse消息。

    4)、ADS不适用于REST-JSON轮询

    5)、当轮询周期设置为较小的值时,为了进行长轮询,则还需要避免发送DiscoveryResponse ,除 非发生了对底层资源的更改

    2、以LDS为例,它用于发现并配置Listener

    基于REST订阅的LDS配置格式,CDS等其它配置类似

    dynamic_resources:
      lds_config:
      resource_api_version: … # xDS 资源配置遵循的API版本,v1.19版本及以后仅支持V3 ;
      api_config_source:
        transport_api_version: ... # xDS 传输协议中使用API版本,v1.19版本及以后仅支持V3;
        api_type: ... # API 可经由REST或gRPC获取,支持的类型包括REST、GRPC和DELTA_GRPC
        cluster_names: ... # 提供服务的集群名称列表,仅能与REST类型的API一起使用;多个集群用于冗余之目的,故障时将循环访问;
        refresh_delay: ... # REST API 轮询时间间隔;
        request_timeout: ... # REST API请求超时时长,默认为1s;

    注意:提供REST API服务的管理服务器也需要定义为Envoy上的集群,并由LDS等相关 的动态发现服务进行调用;但这些管理服务器需要以静态配置的格式提供;

    3、基于REST管理服务器订阅

    使用了ads分别动态获取Listener和Cluster相关的配置

    node:
      id: envoy_front_proxy
      cluster: webcluster
    
    admin:
      profile_path: /tmp/envoy.prof
      access_log_path: /tmp/admin_access.log
      address:
        socket_address:
          address: 0.0.0.0
          port_value: 9901
    
    dynamic_resources:
      cds_config:
        resource_api_version: V3
        api_config_source:
          api_type: REST
          transport_api_version: V3 refresh_delay: {nanos: 500000000} # 1/2s cluster_names:
          - xds_cluster
    
      lds_config:
        resource_api_version: V3
        api_config_source:
        api_type: REST
          transport_api_version: V3 refresh_delay: {nanos: 500000000} # 1/2s cluster_names:
          - xds_cluster

    十、实验案例

    1、cluster-static-dns-discovery

    实验环境

    三个Service:
    
    envoy:Front Proxy,地址由docker-compose动态分配
    webserver01:第一个后端服务,地址由docker-compose动态分配,且将webserver01解析到该地址
    webserver02:第二个后端服务,地址由docker-compose动态分配,且将webserver02解析到该地址

    envoy.yaml

    admin:
      access_log_path: "/dev/null"
      address:
        socket_address:
          address: 0.0.0.0
          port_value: 9901
    
    static_resources:
      listeners:
      - name: listener_0
        address:
          socket_address: { address: 0.0.0.0, port_value: 80 }
        filter_chains:
        - filters:
          - name: envoy.filters.network.http_connection_manager
            typed_config:
              "@type": type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager
              stat_prefix: ingress_http
              codec_type: AUTO
              route_config:
                name: local_route
                virtual_hosts:
                - name: web_service_1
                  domains: ["*"]
                  routes:
                  - match: { prefix: "/" }
                    route: { cluster: local_cluster }
              http_filters:
              - name: envoy.filters.http.router
              
    clusters:
      - name: local_cluster
        connect_timeout: 0.25s
        type: STRICT_DNS
        dns_lookup_family: V4_ONLY
        lb_policy: ROUND_ROBIN
        load_assignment:
          cluster_name: local_cluster
          endpoints:
          - lb_endpoints:
            - endpoint:
                address:
                  socket_address: { address: webserver01, port_value: 8080 }
            - endpoint:
                address:
                  socket_address: { address: webserver02, port_value: 8080 }

    docker-compose.yaml

    version: '3.3'
      
    services:
      envoy:
        image: envoyproxy/envoy-alpine:v1.20.0
        environment:
          - ENVOY_UID=0
        volumes:
        - ./envoy.yaml:/etc/envoy/envoy.yaml
        networks:
          envoymesh:
            aliases:
            - front-proxy
        depends_on:
        - webserver01
        - webserver02
    
      webserver01:
        image: ikubernetes/demoapp:v1.0
        environment:
          - PORT=8080
        hostname: webserver01
        networks:
          envoymesh:
            aliases:
            - webserver01
    
      webserver02:
        image: ikubernetes/demoapp:v1.0
        environment:
          - PORT=8080
        hostname: webserver02
        networks:
          envoymesh:
            aliases:
            - webserver02
    
    networks:
      envoymesh:
        driver: bridge
        ipam:
          config:
            - subnet: 172.31.10.0/24

    实验验证

    docker-compose up

    克隆窗口来验证

    root@test:~# front_proxy_ip=$(docker container inspect --format '{{ $network := index .NetworkSettings.Networks "cluster-static-dns-discovery_envoymesh" }}{{ $network.IPAddress}}' cluster-static-dns-discovery_envoy_1)
    root@test:~# echo $front_proxy_ip 
    172.31.10.4
    root@test:~# curl $front_proxy_ip
    iKubernetes demoapp v1.0 !! ClientIP: 172.31.10.4, ServerName: webserver01, ServerIP: 172.31.10.2!
    root@test:~# curl $front_proxy_ip
    iKubernetes demoapp v1.0 !! ClientIP: 172.31.10.4, ServerName: webserver02, ServerIP: 172.31.10.3!
    
    #可以通过admin interface了解集群的相关状态,尤其是获取的各endpoint的相关信息
    root@test:~# curl http://${front_proxy_ip}:9901/clusters
    local_cluster::observability_name::local_cluster
    local_cluster::default_priority::max_connections::1024
    local_cluster::default_priority::max_pending_requests::1024
    local_cluster::default_priority::max_requests::1024
    local_cluster::default_priority::max_retries::3
    local_cluster::high_priority::max_connections::1024
    local_cluster::high_priority::max_pending_requests::1024
    local_cluster::high_priority::max_requests::1024
    local_cluster::high_priority::max_retries::3
    local_cluster::added_via_api::false
    local_cluster::172.31.10.2:8080::cx_active::0
    local_cluster::172.31.10.2:8080::cx_connect_fail::0
    local_cluster::172.31.10.2:8080::cx_total::2
    local_cluster::172.31.10.2:8080::rq_active::0
    local_cluster::172.31.10.2:8080::rq_error::0
    local_cluster::172.31.10.2:8080::rq_success::2
    local_cluster::172.31.10.2:8080::rq_timeout::0
    local_cluster::172.31.10.2:8080::rq_total::2
    local_cluster::172.31.10.2:8080::hostname::webserver01
    local_cluster::172.31.10.2:8080::health_flags::healthy
    local_cluster::172.31.10.2:8080::weight::1
    local_cluster::172.31.10.2:8080::region::
    local_cluster::172.31.10.2:8080::zone::
    local_cluster::172.31.10.2:8080::sub_zone::
    local_cluster::172.31.10.2:8080::canary::false
    local_cluster::172.31.10.2:8080::priority::0
    local_cluster::172.31.10.2:8080::success_rate::-1.0
    local_cluster::172.31.10.2:8080::local_origin_success_rate::-1.0
    local_cluster::172.31.10.3:8080::cx_active::0
    local_cluster::172.31.10.3:8080::cx_connect_fail::0
    local_cluster::172.31.10.3:8080::cx_total::3
    local_cluster::172.31.10.3:8080::rq_active::0
    local_cluster::172.31.10.3:8080::rq_error::0
    local_cluster::172.31.10.3:8080::rq_success::3
    local_cluster::172.31.10.3:8080::rq_timeout::0
    local_cluster::172.31.10.3:8080::rq_total::3
    local_cluster::172.31.10.3:8080::hostname::webserver02
    local_cluster::172.31.10.3:8080::health_flags::healthy
    local_cluster::172.31.10.3:8080::weight::1
    local_cluster::172.31.10.3:8080::region::
    local_cluster::172.31.10.3:8080::zone::
    local_cluster::172.31.10.3:8080::sub_zone::
    local_cluster::172.31.10.3:8080::canary::false
    local_cluster::172.31.10.3:8080::priority::0
    local_cluster::172.31.10.3:8080::success_rate::-1.0
    local_cluster::172.31.10.3:8080::local_origin_success_rate::-1.0

    2、eds-filesystem

    实验环境

    五个Service:
    
    envoy:Front Proxy,地址为172.31.11.2
    webserver01:第一个后端服务
    webserver01-sidecar:第一个后端服务的Sidecar Proxy,地址为172.31.11.11
    webserver02:第二个后端服务
    webserver02-sidecar:第二个后端服务的Sidecar Proxy,地址为172.31.11.12

    front-envoy.yaml

    node:
      id: envoy_front_proxy
      cluster: MageEdu_Cluster
    
    admin:
      profile_path: /tmp/envoy.prof
      access_log_path: /tmp/admin_access.log
      address:
        socket_address:
           address: 0.0.0.0
           port_value: 9901
    
    static_resources:
      listeners:
      - name: listener_0
        address:
          socket_address: { address: 0.0.0.0, port_value: 80 }
        filter_chains:
        - filters:
          - name: envoy.filters.network.http_connection_manager
            typed_config:
              "@type": type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager
              stat_prefix: ingress_http
              codec_type: AUTO
              route_config:
                name: local_route
                virtual_hosts:
                - name: web_service_01
                  domains: ["*"]
                  routes:
                  - match: { prefix: "/" }
                    route: { cluster: webcluster }
              http_filters:
              - name: envoy.filters.http.router
    
      clusters:
      - name: webcluster
        connect_timeout: 0.25s
        type: EDS
        lb_policy: ROUND_ROBIN
        eds_cluster_config:
          service_name: webcluster
          eds_config:
            path: '/etc/envoy/eds.conf.d/eds.yaml' 

    envoy-sidecar-proxy.yaml

    admin:
      profile_path: /tmp/envoy.prof
      access_log_path: /tmp/admin_access.log
      address:
        socket_address:
           address: 0.0.0.0
           port_value: 9901
    
    static_resources:
      listeners:
      - name: listener_0
        address:
          socket_address: { address: 0.0.0.0, port_value: 80 }
        filter_chains:
        - filters:
          - name: envoy.filters.network.http_connection_manager
            typed_config:
              "@type": type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager
              stat_prefix: ingress_http
              codec_type: AUTO
              route_config:
                name: local_route
                virtual_hosts:
                - name: local_service
                  domains: ["*"]
                  routes:
                  - match: { prefix: "/" }
                    route: { cluster: local_cluster }
              http_filters:
              - name: envoy.filters.http.router
    
      clusters:
      - name: local_cluster
        connect_timeout: 0.25s
        type: STATIC
        lb_policy: ROUND_ROBIN
        load_assignment:
          cluster_name: local_cluster
          endpoints:
          - lb_endpoints:
            - endpoint:
                address:
                  socket_address: { address: 127.0.0.1, port_value: 8080 }

    eds.conf.d目录中的文件

    eds.yaml

    resources:
    - "@type": type.googleapis.com/envoy.config.endpoint.v3.ClusterLoadAssignment
      cluster_name: webcluster
      endpoints:
      - lb_endpoints:
        - endpoint:
            address:
              socket_address:
                address: 172.31.11.11
                port_value: 8080

    eds.yaml.v1

    resources:
    - "@type": type.googleapis.com/envoy.config.endpoint.v3.ClusterLoadAssignment
      cluster_name: webcluster
      endpoints:
      - lb_endpoints:
        - endpoint:
            address:
              socket_address:
                address: 172.31.11.11
                port_value: 8080

    eds.yaml.v2

    version_info: '2'
    resources:
    - "@type": type.googleapis.com/envoy.config.endpoint.v3.ClusterLoadAssignment
      cluster_name: webcluster
      endpoints:
      - lb_endpoints:
        - endpoint:
            address:
              socket_address:
                address: 172.31.11.11
                port_value: 8080
        - endpoint:
            address:
              socket_address:
                address: 172.31.11.12
                port_value: 8080

    docker-compose.yaml

    version: '3.3'
      
    services:
      envoy:
        image: envoyproxy/envoy-alpine:v1.20.0
        environment:
          - ENVOY_UID=0
        volumes:
        - ./front-envoy.yaml:/etc/envoy/envoy.yaml
        - ./eds.conf.d/:/etc/envoy/eds.conf.d/
        networks:
          envoymesh:
            ipv4_address: 172.31.11.2
            aliases:
            - front-proxy
        depends_on:
        - webserver01-sidecar
        - webserver02-sidecar
    
      webserver01-sidecar:
        image: envoyproxy/envoy-alpine:v1.20.0
        environment:
          - ENVOY_UID=0
        volumes:
        - ./envoy-sidecar-proxy.yaml:/etc/envoy/envoy.yaml
        hostname: webserver01
        networks:
          envoymesh:
            ipv4_address: 172.31.11.11
            aliases:
            - webserver01-sidecar
    
      webserver01:
        image: ikubernetes/demoapp:v1.0
        environment:
          - PORT=8080
          - HOST=127.0.0.1
        network_mode: "service:webserver01-sidecar"
        depends_on:
        - webserver01-sidecar
    
      webserver02-sidecar:
        image: envoyproxy/envoy-alpine:v1.20.0
        environment:
          - ENVOY_UID=0
        volumes:
        - ./envoy-sidecar-proxy.yaml:/etc/envoy/envoy.yaml
        hostname: webserver02
        networks:
          envoymesh:
            ipv4_address: 172.31.11.12
            aliases:
            - webserver02-sidecar
            
       webserver02:
        image: ikubernetes/demoapp:v1.0
        environment:
          - PORT=8080
          - HOST=127.0.0.1
        network_mode: "service:webserver02-sidecar"
        depends_on:
        - webserver02-sidecar
    
    networks:
      envoymesh:
        driver: bridge
        ipam:
          config:
            - subnet: 172.31.11.0/24

    实验验证

    docker-compose up

    克隆窗口

    # 查看Cluster中的Endpoint信息 
     root@test:~# curl 172.31.11.2:9901/clusters
    webcluster::observability_name::webcluster
    webcluster::default_priority::max_connections::1024
    webcluster::default_priority::max_pending_requests::1024
    webcluster::default_priority::max_requests::1024
    webcluster::default_priority::max_retries::3
    webcluster::high_priority::max_connections::1024
    webcluster::high_priority::max_pending_requests::1024
    webcluster::high_priority::max_requests::1024
    webcluster::high_priority::max_retries::3
    webcluster::added_via_api::false
    webcluster::172.31.11.11:8080::cx_active::0
    webcluster::172.31.11.11:8080::cx_connect_fail::0
    webcluster::172.31.11.11:8080::cx_total::0
    webcluster::172.31.11.11:8080::rq_active::0
    webcluster::172.31.11.11:8080::rq_error::0
    webcluster::172.31.11.11:8080::rq_success::0
    webcluster::172.31.11.11:8080::rq_timeout::0
    webcluster::172.31.11.11:8080::rq_total::0
    webcluster::172.31.11.11:8080::hostname::
    webcluster::172.31.11.11:8080::health_flags::healthy
    webcluster::172.31.11.11:8080::weight::1
    webcluster::172.31.11.11:8080::region::
    webcluster::172.31.11.11:8080::zone::
    webcluster::172.31.11.11:8080::sub_zone::
    webcluster::172.31.11.11:8080::canary::false
    webcluster::172.31.11.11:8080::priority::0
    webcluster::172.31.11.11:8080::success_rate::-1.0
    webcluster::172.31.11.11:8080::local_origin_success_rate::-1.0
    
    
    
    # 接入front proxy envoy容器的交互式接口,修改eds.conf文件中的内容,将另一个endpoint添加进文件中;
    root@test:~# docker exec -it eds-filesystem_envoy_1 /bin/sh
    / # cd /etc/envoy/eds.conf.d/
    /etc/envoy/eds.conf.d # cat eds.yaml.v2 > eds.yaml
    # 运行下面的命令强制激活文件更改,以便基于inode监视的工作机制可被触发
    /etc/envoy/eds.conf.d # mv eds.yaml temp && mv temp eds.yaml
    
    # 再次查看Cluster中的Endpoint信息 
    root@test:~# curl 172.31.11.2:9901/clusters
    webcluster::observability_name::webcluster
    webcluster::default_priority::max_connections::1024
    webcluster::default_priority::max_pending_requests::1024
    webcluster::default_priority::max_requests::1024
    webcluster::default_priority::max_retries::3
    webcluster::high_priority::max_connections::1024
    webcluster::high_priority::max_pending_requests::1024
    webcluster::high_priority::max_requests::1024
    webcluster::high_priority::max_retries::3
    webcluster::added_via_api::false
    webcluster::172.31.11.11:8080::cx_active::0
    webcluster::172.31.11.11:8080::cx_connect_fail::0
    webcluster::172.31.11.11:8080::cx_total::0
    webcluster::172.31.11.11:8080::rq_active::0
    webcluster::172.31.11.11:8080::rq_error::0
    webcluster::172.31.11.11:8080::rq_success::0
    webcluster::172.31.11.11:8080::rq_timeout::0
    webcluster::172.31.11.11:8080::rq_total::0
    webcluster::172.31.11.11:8080::hostname::
    webcluster::172.31.11.11:8080::health_flags::healthy
    webcluster::172.31.11.11:8080::weight::1
    webcluster::172.31.11.11:8080::region::
    webcluster::172.31.11.11:8080::zone::
    webcluster::172.31.11.11:8080::sub_zone::
    webcluster::172.31.11.11:8080::canary::false
    webcluster::172.31.11.11:8080::priority::0
    webcluster::172.31.11.11:8080::success_rate::-1.0
    webcluster::172.31.11.11:8080::local_origin_success_rate::-1.0
    webcluster::172.31.11.12:8080::cx_active::0
    webcluster::172.31.11.12:8080::cx_connect_fail::0
    webcluster::172.31.11.12:8080::cx_total::0
    webcluster::172.31.11.12:8080::rq_active::0
    webcluster::172.31.11.12:8080::rq_error::0
    webcluster::172.31.11.12:8080::rq_success::0
    webcluster::172.31.11.12:8080::rq_timeout::0
    webcluster::172.31.11.12:8080::rq_total::0
    webcluster::172.31.11.12:8080::hostname::
    webcluster::172.31.11.12:8080::health_flags::healthy
    webcluster::172.31.11.12:8080::weight::1
    webcluster::172.31.11.12:8080::region::
    webcluster::172.31.11.12:8080::zone::
    webcluster::172.31.11.12:8080::sub_zone::
    webcluster::172.31.11.12:8080::canary::false
    webcluster::172.31.11.12:8080::priority::0
    webcluster::172.31.11.12:8080::success_rate::-1.0
    webcluster::172.31.11.12:8080::local_origin_success_rate::-1.0
    #新节点172.31.11.12已经添加

    3、lds-cds-filesystem

    实验环境

    五个Service:
    
    envoy:Front Proxy,地址为172.31.12.2
    webserver01:第一个后端服务
    webserver01-sidecar:第一个后端服务的Sidecar Proxy,地址为172.31.12.11
    webserver02:第二个后端服务
    webserver02-sidecar:第二个后端服务的Sidecar Proxy,地址为172.31.12.12

    front-envoy.yaml

    node:
      id: envoy_front_proxy
      cluster: MageEdu_Cluster
    
    admin:
      profile_path: /tmp/envoy.prof
      access_log_path: /tmp/admin_access.log
      address:
        socket_address:
           address: 0.0.0.0
           port_value: 9901
    
    dynamic_resources:
      lds_config:
        path: /etc/envoy/conf.d/lds.yaml
      cds_config:
        path: /etc/envoy/conf.d/cds.yaml

    envoy-sidecar-proxy.yaml

    admin:
      profile_path: /tmp/envoy.prof
      access_log_path: /tmp/admin_access.log
      address:
        socket_address:
           address: 0.0.0.0
           port_value: 9901
    
    static_resources:
      listeners:
      - name: listener_0
        address:
          socket_address: { address: 0.0.0.0, port_value: 80 }
        filter_chains:
        - filters:
          - name: envoy.filters.network.http_connection_manager
            typed_config:
              "@type": type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager
              stat_prefix: ingress_http
              codec_type: AUTO
              route_config:
                name: local_route
                virtual_hosts:
                - name: local_service
                  domains: ["*"]
                  routes:
                  - match: { prefix: "/" }
                    route: { cluster: local_cluster }
              http_filters:
              - name: envoy.filters.http.router
    
    clusters:
      - name: local_cluster
        connect_timeout: 0.25s
        type: STATIC
        lb_policy: ROUND_ROBIN
        load_assignment:
          cluster_name: local_cluster
          endpoints:
          - lb_endpoints:
            - endpoint:
                address:
                  socket_address: { address: 127.0.0.1, port_value: 8080 }

    docker-compose.yaml

    version: '3.3'
      
    services:
      envoy:
        image: envoyproxy/envoy-alpine:v1.20.0
        environment:
          - ENVOY_UID=0
        volumes:
        - ./front-envoy.yaml:/etc/envoy/envoy.yaml
        - ./conf.d/:/etc/envoy/conf.d/
        networks:
          envoymesh:
            ipv4_address: 172.31.12.2
            aliases:
            - front-proxy
        depends_on:
        - webserver01
        - webserver01-app
        - webserver02
        - webserver02-app
    
      webserver01:
        #image: envoyproxy/envoy-alpine:v1.18-latest
        image: envoyproxy/envoy-alpine:v1.20.0
        environment:
          - ENVOY_UID=0
        volumes:
        - ./envoy-sidecar-proxy.yaml:/etc/envoy/envoy.yaml
        hostname: webserver01
        networks:
          envoymesh:
            ipv4_address: 172.31.12.11
            aliases:
            - webserver01-sidecar
    
      webserver01-app:
        image: ikubernetes/demoapp:v1.0
        environment:
          - PORT=8080
          - HOST=127.0.0.1
        network_mode: "service:webserver01"
        depends_on:
        - webserver01
    
      webserver02:
        #image: envoyproxy/envoy-alpine:v1.18-latest
        image: envoyproxy/envoy-alpine:v1.20.0
        environment:
          - ENVOY_UID=0
        volumes:
        - ./envoy-sidecar-proxy.yaml:/etc/envoy/envoy.yaml
        hostname: webserver02
        networks:
          envoymesh:
            ipv4_address: 172.31.12.12
            aliases:
            - webserver02-sidecar
    
      webserver02-app:
        image: ikubernetes/demoapp:v1.0
        environment:
          - PORT=8080
          - HOST=127.0.0.1
        network_mode: "service:webserver02"
        depends_on:
        - webserver02
    
    networks:
      envoymesh:
        driver: bridge
        ipam:
          config:
            - subnet: 172.31.12.0/24

    conf.d目录中的文件

    cds.yaml

    resources:
    - "@type": type.googleapis.com/envoy.config.cluster.v3.Cluster
      name: webcluster
      connect_timeout: 1s
      type: STRICT_DNS
      load_assignment:
        cluster_name: webcluster
        endpoints:
        - lb_endpoints:
          - endpoint:
              address:
                socket_address:
                  address: webserver01
                  port_value: 8080
          #- endpoint:
          #    address:
          #      socket_address:
          #        address: webserver02
          #        port_value: 8080

    lds.yaml

    resources:
    - "@type": type.googleapis.com/envoy.config.listener.v3.Listener
      name: listener_http
      address:
        socket_address: { address: 0.0.0.0, port_value: 80 }
      filter_chains:
      - filters:
          name: envoy.http_connection_manager
          typed_config:
            "@type": type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager
            stat_prefix: ingress_http
            route_config:
              name: local_route
              virtual_hosts:
              - name: local_service
                domains: ["*"]
                routes:
                - match:
                    prefix: "/"
                  route:
                    cluster: webcluster
            http_filters:
            - name: envoy.filters.http.router

    实验验证

    docker-compose up

    克隆窗口

    # 查看Cluster的信息 
    root@test:~# curl 172.31.12.2:9901/clusters
    webcluster::observability_name::webcluster
    webcluster::default_priority::max_connections::1024
    webcluster::default_priority::max_pending_requests::1024
    webcluster::default_priority::max_requests::1024
    webcluster::default_priority::max_retries::3
    webcluster::high_priority::max_connections::1024
    webcluster::high_priority::max_pending_requests::1024
    webcluster::high_priority::max_requests::1024
    webcluster::high_priority::max_retries::3
    webcluster::added_via_api::true
    webcluster::172.31.12.11:8080::cx_active::0
    webcluster::172.31.12.11:8080::cx_connect_fail::0
    webcluster::172.31.12.11:8080::cx_total::0
    webcluster::172.31.12.11:8080::rq_active::0
    webcluster::172.31.12.11:8080::rq_error::0
    webcluster::172.31.12.11:8080::rq_success::0
    webcluster::172.31.12.11:8080::rq_timeout::0
    webcluster::172.31.12.11:8080::rq_total::0
    webcluster::172.31.12.11:8080::hostname::webserver01
    webcluster::172.31.12.11:8080::health_flags::healthy
    webcluster::172.31.12.11:8080::weight::1
    webcluster::172.31.12.11:8080::region::
    webcluster::172.31.12.11:8080::zone::
    webcluster::172.31.12.11:8080::sub_zone::
    webcluster::172.31.12.11:8080::canary::false
    webcluster::172.31.12.11:8080::priority::0
    webcluster::172.31.12.11:8080::success_rate::-1.0
    webcluster::172.31.12.11:8080::local_origin_success_rate::-1.0
    
    # 查看Listener的信息 
    root@test:~# curl 172.31.12.2:9901/listeners
    listener_http::0.0.0.0:80
    
    # 接入front proxy envoy容器的交互式接口
    root@test:~# docker exec -it lds-cds-filesystem_envoy_1 /bin/sh
    / # cd /etc/envoy/conf.d/
    
    # 修改cds.yaml的内容,cds.yaml中添加一个节点
    /etc/envoy/conf.d # cat cds.yaml 
    resources:
    - "@type": type.googleapis.com/envoy.config.cluster.v3.Cluster
      name: webcluster
      connect_timeout: 1s
      type: STRICT_DNS
      load_assignment:
        cluster_name: webcluster
        endpoints:
        - lb_endpoints:
          - endpoint:
              address:
                socket_address:
                  address: webserver01
                  port_value: 8080
          - endpoint:
              address:
                socket_address:
                  address: webserver02
                  port_value: 8080
    
    #运行类似下面的命令强制激活文件更改,以便基于inode监视的工作机制可被触发
    /etc/envoy/conf.d # mv cds.yaml temp && mv temp cds.yaml
    
    ## 再次验证相关的配置信息
    root@test:~# curl 172.31.12.2:9901/clusters
    webcluster::observability_name::webcluster
    webcluster::default_priority::max_connections::1024
    webcluster::default_priority::max_pending_requests::1024
    webcluster::default_priority::max_requests::1024
    webcluster::default_priority::max_retries::3
    webcluster::high_priority::max_connections::1024
    webcluster::high_priority::max_pending_requests::1024
    webcluster::high_priority::max_requests::1024
    webcluster::high_priority::max_retries::3
    webcluster::added_via_api::true
    webcluster::172.31.12.11:8080::cx_active::0
    webcluster::172.31.12.11:8080::cx_connect_fail::0
    webcluster::172.31.12.11:8080::cx_total::0
    webcluster::172.31.12.11:8080::rq_active::0
    webcluster::172.31.12.11:8080::rq_error::0
    webcluster::172.31.12.11:8080::rq_success::0
    webcluster::172.31.12.11:8080::rq_timeout::0
    webcluster::172.31.12.11:8080::rq_total::0
    webcluster::172.31.12.11:8080::hostname::webserver01
    webcluster::172.31.12.11:8080::health_flags::healthy
    webcluster::172.31.12.11:8080::weight::1
    webcluster::172.31.12.11:8080::region::
    webcluster::172.31.12.11:8080::zone::
    webcluster::172.31.12.11:8080::sub_zone::
    webcluster::172.31.12.11:8080::canary::false
    webcluster::172.31.12.11:8080::priority::0
    webcluster::172.31.12.11:8080::success_rate::-1.0
    webcluster::172.31.12.11:8080::local_origin_success_rate::-1.0
    webcluster::172.31.12.12:8080::cx_active::0
    webcluster::172.31.12.12:8080::cx_connect_fail::0
    webcluster::172.31.12.12:8080::cx_total::0
    webcluster::172.31.12.12:8080::rq_active::0
    webcluster::172.31.12.12:8080::rq_error::0
    webcluster::172.31.12.12:8080::rq_success::0
    webcluster::172.31.12.12:8080::rq_timeout::0
    webcluster::172.31.12.12:8080::rq_total::0
    webcluster::172.31.12.12:8080::hostname::webserver02
    webcluster::172.31.12.12:8080::health_flags::healthy
    webcluster::172.31.12.12:8080::weight::1
    webcluster::172.31.12.12:8080::region::
    webcluster::172.31.12.12:8080::zone::
    webcluster::172.31.12.12:8080::sub_zone::
    webcluster::172.31.12.12:8080::canary::false
    webcluster::172.31.12.12:8080::priority::0
    webcluster::172.31.12.12:8080::success_rate::-1.0
    webcluster::172.31.12.12:8080::local_origin_success_rate::-1.0
    #新节点已经添加

    4、ads-grpc

    实验环境

    六个Service:
    
    envoy:Front Proxy,地址为172.31.16.2
    webserver01:第一个后端服务
    webserver01-sidecar:第一个后端服务的Sidecar Proxy,地址为172.31.16.11
    webserver02:第二个后端服务
    webserver02-sidecar:第二个后端服务的Sidecar Proxy,地址为172.31.16.12
    xdsserver: xDS management server,地址为172.31.16.5

    front-envoy.yaml

    node:
      id: envoy_front_proxy
      cluster: webcluster
    
    admin:
      profile_path: /tmp/envoy.prof
      access_log_path: /tmp/admin_access.log
      address:
        socket_address:
           address: 0.0.0.0
           port_value: 9901
    
    dynamic_resources:
      ads_config:
        api_type: GRPC
        transport_api_version: V3
        grpc_services:
        - envoy_grpc:
            cluster_name: xds_cluster
        set_node_on_first_message_only: true
      cds_config:
        resource_api_version: V3
        ads: {}
      lds_config:
        resource_api_version: V3
        ads: {}
    static_resources:
      clusters:
      - name: xds_cluster
        connect_timeout: 0.25s
        type: STRICT_DNS
        # The extension_protocol_options field is used to provide extension-specific protocol options for upstream connections. 
        typed_extension_protocol_options:
          envoy.extensions.upstreams.http.v3.HttpProtocolOptions:
            "@type": type.googleapis.com/envoy.extensions.upstreams.http.v3.HttpProtocolOptions
            explicit_http_config:
              http2_protocol_options: {}
        lb_policy: ROUND_ROBIN
        load_assignment:
          cluster_name: xds_cluster
          endpoints:
          - lb_endpoints:
            - endpoint:
                address:
                  socket_address:
                    address: xdsserver
                    port_value: 18000

    envoy-sidecar-proxy.yaml

    admin:
      profile_path: /tmp/envoy.prof
      access_log_path: /tmp/admin_access.log
      address:
        socket_address:
           address: 0.0.0.0
           port_value: 9901
    
    static_resources:
      listeners:
      - name: listener_0
        address:
          socket_address: { address: 0.0.0.0, port_value: 80 }
        filter_chains:
        - filters:
          - name: envoy.filters.network.http_connection_manager
            typed_config:
              "@type": type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager
              stat_prefix: ingress_http
              codec_type: AUTO
              route_config:
                name: local_route
                virtual_hosts:
                - name: local_service
                  domains: ["*"]
                  routes:
                  - match: { prefix: "/" }
                    route: { cluster: local_cluster }
              http_filters:
              - name: envoy.filters.http.router
      clusters:
        - name: local_cluster
          connect_timeout: 0.25s
          type: STATIC
          lb_policy: ROUND_ROBIN
          load_assignment:
            cluster_name: local_cluster
            endpoints:
            - lb_endpoints:
              - endpoint:
                  address:
                    socket_address: { address: 127.0.0.1, port_value: 8080 }

    docker-compose.yaml

    version: '3.3'
      
    services:
      envoy:
        image: envoyproxy/envoy-alpine:v1.18-latest
        volumes:
        - ./front-envoy.yaml:/etc/envoy/envoy.yaml
        networks:
          envoymesh:
            ipv4_address: 172.31.16.2
            aliases:
            - front-proxy
        depends_on:
        - webserver01
        - webserver02
        - xdsserver
    
      webserver01:
        image: ikubernetes/demoapp:v1.0
        environment:
          - PORT=8080
          - HOST=127.0.0.1
        hostname: webserver01
        networks:
          envoymesh:
            ipv4_address: 172.31.16.11
    
      webserver01-sidecar:
        image: envoyproxy/envoy-alpine:v1.18-latest
        volumes:
        - ./envoy-sidecar-proxy.yaml:/etc/envoy/envoy.yaml
        network_mode: "service:webserver01"
        depends_on:
        - webserver01
    
      webserver02:
        image: ikubernetes/demoapp:v1.0
        environment:
          - PORT=8080
          - HOST=127.0.0.1
        hostname: webserver02
        networks:
          envoymesh:
            ipv4_address: 172.31.16.12
    
      webserver02-sidecar:
        image: envoyproxy/envoy-alpine:v1.18-latest
        volumes:
        - ./envoy-sidecar-proxy.yaml:/etc/envoy/envoy.yaml
        network_mode: "service:webserver02"
        depends_on:
        - webserver02
    
      xdsserver:
        image: ikubernetes/envoy-xds-server:v0.1
        environment:
          - SERVER_PORT=18000
          - NODE_ID=envoy_front_proxy
          - RESOURCES_FILE=/etc/envoy-xds-server/config/config.yaml
        volumes:
        - ./resources:/etc/envoy-xds-server/config/
        networks:
          envoymesh:
            ipv4_address: 172.31.16.5
            aliases:
            - xdsserver
            - xds-service
        expose:
        - "18000"
        
    networks:
      envoymesh:
        driver: bridge
        ipam:
          config:
            - subnet: 172.31.16.0/24

    resources目录中的文件

    config.yaml

    name: myconfig
    spec:
      listeners:
      - name: listener_http
        address: 0.0.0.0
        port: 80
        routes:
        - name: local_route
          prefix: /
          clusters:
          - webcluster
      clusters:
      - name: webcluster
        endpoints:
        - address: 172.31.16.11
          port: 8080

    config.yaml-v2

    name: myconfig
    spec:
      listeners:
      - name: listener_http
        address: 0.0.0.0
        port: 80
        routes:
        - name: local_route
          prefix: /
          clusters:
          - webcluster
      clusters:
      - name: webcluster
        endpoints:
        - address: 172.31.16.11
          port: 8080
        - address: 172.31.16.12
          port: 8080

    实验验证

    docker-compose up

    克隆窗口

    # 查看Cluster及Endpoints信息;
    root@test:~# curl 172.31.16.2:9901/clusters
    webcluster::observability_name::webcluster
    webcluster::default_priority::max_connections::1024
    webcluster::default_priority::max_pending_requests::1024
    webcluster::default_priority::max_requests::1024
    webcluster::default_priority::max_retries::3
    webcluster::high_priority::max_connections::1024
    webcluster::high_priority::max_pending_requests::1024
    webcluster::high_priority::max_requests::1024
    webcluster::high_priority::max_retries::3
    webcluster::added_via_api::true
    webcluster::172.31.16.11:8080::cx_active::0
    webcluster::172.31.16.11:8080::cx_connect_fail::0
    webcluster::172.31.16.11:8080::cx_total::0
    webcluster::172.31.16.11:8080::rq_active::0
    webcluster::172.31.16.11:8080::rq_error::0
    webcluster::172.31.16.11:8080::rq_success::0
    webcluster::172.31.16.11:8080::rq_timeout::0
    webcluster::172.31.16.11:8080::rq_total::0
    webcluster::172.31.16.11:8080::hostname::
    webcluster::172.31.16.11:8080::health_flags::healthy
    webcluster::172.31.16.11:8080::weight::1
    webcluster::172.31.16.11:8080::region::
    webcluster::172.31.16.11:8080::zone::
    webcluster::172.31.16.11:8080::sub_zone::
    webcluster::172.31.16.11:8080::canary::false
    webcluster::172.31.16.11:8080::priority::0
    webcluster::172.31.16.11:8080::success_rate::-1.0
    webcluster::172.31.16.11:8080::local_origin_success_rate::-1.0
    xds_cluster::observability_name::xds_cluster
    xds_cluster::default_priority::max_connections::1024
    xds_cluster::default_priority::max_pending_requests::1024
    xds_cluster::default_priority::max_requests::1024
    xds_cluster::default_priority::max_retries::3
    xds_cluster::high_priority::max_connections::1024
    xds_cluster::high_priority::max_pending_requests::1024
    xds_cluster::high_priority::max_requests::1024
    xds_cluster::high_priority::max_retries::3
    xds_cluster::added_via_api::false
    xds_cluster::172.31.16.5:18000::cx_active::1
    xds_cluster::172.31.16.5:18000::cx_connect_fail::0
    xds_cluster::172.31.16.5:18000::cx_total::1
    xds_cluster::172.31.16.5:18000::rq_active::3
    xds_cluster::172.31.16.5:18000::rq_error::0
    xds_cluster::172.31.16.5:18000::rq_success::0
    xds_cluster::172.31.16.5:18000::rq_timeout::0
    xds_cluster::172.31.16.5:18000::rq_total::3
    xds_cluster::172.31.16.5:18000::hostname::xdsserver
    xds_cluster::172.31.16.5:18000::health_flags::healthy
    xds_cluster::172.31.16.5:18000::weight::1
    xds_cluster::172.31.16.5:18000::region::
    xds_cluster::172.31.16.5:18000::zone::
    xds_cluster::172.31.16.5:18000::sub_zone::
    xds_cluster::172.31.16.5:18000::canary::false
    xds_cluster::172.31.16.5:18000::priority::0
    xds_cluster::172.31.16.5:18000::success_rate::-1.0
    xds_cluster::172.31.16.5:18000::local_origin_success_rate::-1.0
    
    #查看动态Clusters的相关信息
    root@test:~# curl -s 172.31.16.2:9901/config_dump | jq '.configs[1].dynamic_active_clusters'
    [
      {
        "version_info": "411",
        "cluster": {
          "@type": "type.googleapis.com/envoy.config.cluster.v3.Cluster",
          "name": "webcluster",
          "type": "EDS",
          "eds_cluster_config": {
            "eds_config": {
              "api_config_source": {
                "api_type": "GRPC",
                "grpc_services": [
                  {
                    "envoy_grpc": {
                      "cluster_name": "xds_cluster"
                    }
                  }
                ],
                "set_node_on_first_message_only": true,
                "transport_api_version": "V3"
              },
              "resource_api_version": "V3"
            }
          },
          "connect_timeout": "5s",
          "dns_lookup_family": "V4_ONLY"
        },
        "last_updated": "2021-12-02T07:47:28.765Z"
      }
    ]
    
    # 查看Listener列表
    root@test:~# curl 172.31.16.2:9901/listeners
    listener_http::0.0.0.0:80
    
    #查看动态的Listener信息
    root@test:~# curl -s 172.31.16.2:9901/config_dump?resource=dynamic_listeners | jq '.configs[0].active_state.listener.address'
    {
      "socket_address": {
        "address": "0.0.0.0",
        "port_value": 80
      }
    }
    
    # 接入xdsserver容器的交互式接口,修改config.yaml文件中的内容,将另一个endpoint添加进文件中,或进行其它修改;
    root@test:/apps/servicemesh_in_practise-develop/Dynamic-Configuration/ads-grpc# docker-compose exec xdsserver sh
    / # cd /etc/envoy-xds-server/config/
    /etc/envoy-xds-server/config #/etc/envoy-xds-server/config # cat config.yaml-v2 > config.yaml
    #提示:以上修改操作也可以直接在宿主机上的存储卷目录中进行。
    
    # 再次查看Cluster中的Endpoint信息
    root@test:/apps/servicemesh_in_practise-develop/Dynamic-Configuration/ads-grpc# curl 172.31.16.2:9901/clusters
    webcluster::observability_name::webcluster
    webcluster::default_priority::max_connections::1024
    webcluster::default_priority::max_pending_requests::1024
    webcluster::default_priority::max_requests::1024
    webcluster::default_priority::max_retries::3
    webcluster::high_priority::max_connections::1024
    webcluster::high_priority::max_pending_requests::1024
    webcluster::high_priority::max_requests::1024
    webcluster::high_priority::max_retries::3
    webcluster::added_via_api::true
    webcluster::172.31.16.11:8080::cx_active::0
    webcluster::172.31.16.11:8080::cx_connect_fail::0
    webcluster::172.31.16.11:8080::cx_total::0
    webcluster::172.31.16.11:8080::rq_active::0
    webcluster::172.31.16.11:8080::rq_error::0
    webcluster::172.31.16.11:8080::rq_success::0
    webcluster::172.31.16.11:8080::rq_timeout::0
    webcluster::172.31.16.11:8080::rq_total::0
    webcluster::172.31.16.11:8080::hostname::
    webcluster::172.31.16.11:8080::health_flags::healthy
    webcluster::172.31.16.11:8080::weight::1
    webcluster::172.31.16.11:8080::region::
    webcluster::172.31.16.11:8080::zone::
    webcluster::172.31.16.11:8080::sub_zone::
    webcluster::172.31.16.11:8080::canary::false
    webcluster::172.31.16.11:8080::priority::0
    webcluster::172.31.16.11:8080::success_rate::-1.0
    webcluster::172.31.16.11:8080::local_origin_success_rate::-1.0
    webcluster::172.31.16.12:8080::cx_active::0
    webcluster::172.31.16.12:8080::cx_connect_fail::0
    webcluster::172.31.16.12:8080::cx_total::0
    webcluster::172.31.16.12:8080::rq_active::0
    webcluster::172.31.16.12:8080::rq_error::0
    webcluster::172.31.16.12:8080::rq_success::0
    webcluster::172.31.16.12:8080::rq_timeout::0
    webcluster::172.31.16.12:8080::rq_total::0
    webcluster::172.31.16.12:8080::hostname::
    webcluster::172.31.16.12:8080::health_flags::healthy
    webcluster::172.31.16.12:8080::weight::1
    webcluster::172.31.16.12:8080::region::
    webcluster::172.31.16.12:8080::zone::
    webcluster::172.31.16.12:8080::sub_zone::
    webcluster::172.31.16.12:8080::canary::false
    webcluster::172.31.16.12:8080::priority::0
    webcluster::172.31.16.12:8080::success_rate::-1.0
    webcluster::172.31.16.12:8080::local_origin_success_rate::-1.0
    xds_cluster::observability_name::xds_cluster
    xds_cluster::default_priority::max_connections::1024
    xds_cluster::default_priority::max_pending_requests::1024
    xds_cluster::default_priority::max_requests::1024
    xds_cluster::default_priority::max_retries::3
    xds_cluster::high_priority::max_connections::1024
    xds_cluster::high_priority::max_pending_requests::1024
    xds_cluster::high_priority::max_requests::1024
    xds_cluster::high_priority::max_retries::3
    xds_cluster::added_via_api::false
    xds_cluster::172.31.16.5:18000::cx_active::1
    xds_cluster::172.31.16.5:18000::cx_connect_fail::0
    xds_cluster::172.31.16.5:18000::cx_total::1
    xds_cluster::172.31.16.5:18000::rq_active::3
    xds_cluster::172.31.16.5:18000::rq_error::0
    xds_cluster::172.31.16.5:18000::rq_success::0
    xds_cluster::172.31.16.5:18000::rq_timeout::0
    xds_cluster::172.31.16.5:18000::rq_total::3
    xds_cluster::172.31.16.5:18000::hostname::xdsserver
    xds_cluster::172.31.16.5:18000::health_flags::healthy
    xds_cluster::172.31.16.5:18000::weight::1
    xds_cluster::172.31.16.5:18000::region::
    xds_cluster::172.31.16.5:18000::zone::
    xds_cluster::172.31.16.5:18000::sub_zone::
    xds_cluster::172.31.16.5:18000::canary::false
    xds_cluster::172.31.16.5:18000::priority::0
    xds_cluster::172.31.16.5:18000::success_rate::-1.0
    xds_cluster::172.31.16.5:18000::local_origin_success_rate::-1.0
    #新节点已经添加

    5、lds-cds-grpc

    实验环境

    六个Service:
    
    envoy:Front Proxy,地址为172.31.15.2
    webserver01:第一个后端服务
    webserver01-sidecar:第一个后端服务的Sidecar Proxy,地址为172.31.15.11
    webserver02:第二个后端服务
    webserver02-sidecar:第二个后端服务的Sidecar Proxy,地址为172.31.15.12
    xdsserver: xDS management server,地址为172.31.15.5

    envoy_front_proxy.yaml

    node:
      id: envoy_front_proxy
      cluster: webcluster
    
    admin:
      profile_path: /tmp/envoy.prof
      access_log_path: /tmp/admin_access.log
      address:
        socket_address:
           address: 0.0.0.0
           port_value: 9901
    
    dynamic_resources:
      lds_config:
        resource_api_version: V3
        api_config_source:
          api_type: GRPC
          transport_api_version: V3
          grpc_services:
          - envoy_grpc:
              cluster_name: xds_cluster
    
      cds_config:
        resource_api_version: V3
        api_config_source:
          api_type: GRPC
          transport_api_version: V3
          grpc_services:
          - envoy_grpc:
              cluster_name: xds_cluster
    
    static_resources:
      clusters:
      - name: xds_cluster
        connect_timeout: 0.25s
        type: STRICT_DNS
        # The extension_protocol_options field is used to provide extension-specific protocol options for upstream connections. 
        typed_extension_protocol_options:
          envoy.extensions.upstreams.http.v3.HttpProtocolOptions:
            "@type": type.googleapis.com/envoy.extensions.upstreams.http.v3.HttpProtocolOptions
            explicit_http_config:
              http2_protocol_options: {}
        lb_policy: ROUND_ROBIN
        load_assignment:
          cluster_name: xds_cluster
          endpoints:
          - lb_endpoints:
            - endpoint:
                address:
                  socket_address:
                    address: xdsserver
                    port_value: 18000

    envoy-sidecar-proxy.yaml

    admin:
      profile_path: /tmp/envoy.prof
      access_log_path: /tmp/admin_access.log
      address:
        socket_address:
           address: 0.0.0.0
           port_value: 9901
    
    static_resources:
      listeners:
      - name: listener_0
        address:
          socket_address: { address: 0.0.0.0, port_value: 80 }
        filter_chains:
        - filters:
          - name: envoy.filters.network.http_connection_manager
            typed_config:
              "@type": type.googleapis.com/envoy.extensions.filters.network.http_connection_manager.v3.HttpConnectionManager
              stat_prefix: ingress_http
              codec_type: AUTO
              route_config:
                name: local_route
                virtual_hosts:
                - name: local_service 
                  domains: ["*"]
                  routes:
                  - match: { prefix: "/" }
                    route: { cluster: local_cluster }
              http_filters:
              - name: envoy.filters.http.router
    
      clusters:
      - name: local_cluster
        connect_timeout: 0.25s
        type: STATIC
        lb_policy: ROUND_ROBIN
        load_assignment:
          cluster_name: local_cluster
          endpoints:
          - lb_endpoints:
            - endpoint:
                address:
                  socket_address: { address: 127.0.0.1, port_value: 8080 }

    docker-compose.yaml

    version: '3.3'
    
    services:
      envoy:
        image: envoyproxy/envoy-alpine:v1.20.0
        environment:
          - ENVOY_UID=0
        volumes:
        - ./front-envoy.yaml:/etc/envoy/envoy.yaml
        networks:
          envoymesh:
            ipv4_address: 172.31.15.2
            aliases:
            - front-proxy
        depends_on:
        - webserver01
        - webserver02
        - xdsserver
    
      webserver01:
        image: ikubernetes/demoapp:v1.0
        environment:
          - PORT=8080
          - HOST=127.0.0.1
        hostname: webserver01
        networks:
          envoymesh:
            ipv4_address: 172.31.15.11
    
      webserver01-sidecar:
        image: envoyproxy/envoy-alpine:v1.20.0
        environment:
          - ENVOY_UID=0
        volumes:
        - ./envoy-sidecar-proxy.yaml:/etc/envoy/envoy.yaml
        network_mode: "service:webserver01"
        depends_on:
        - webserver01
    
      webserver02:
        image: ikubernetes/demoapp:v1.0
        environment:
          - PORT=8080
          - HOST=127.0.0.1
        hostname: webserver02
        networks:
          envoymesh:
            ipv4_address: 172.31.15.12
    
      webserver02-sidecar:
        image: envoyproxy/envoy-alpine:v1.20.0
        environment:
          - ENVOY_UID=0
        volumes:
        - ./envoy-sidecar-proxy.yaml:/etc/envoy/envoy.yaml
        network_mode: "service:webserver02"
        depends_on:
        - webserver02
    
      xdsserver:
        image: ikubernetes/envoy-xds-server:v0.1
        environment:
          - SERVER_PORT=18000
          - NODE_ID=envoy_front_proxy
          - RESOURCES_FILE=/etc/envoy-xds-server/config/config.yaml
        volumes:
        - ./resources:/etc/envoy-xds-server/config/
        networks:
          envoymesh:
            ipv4_address: 172.31.15.5
            aliases:
            - xdsserver
            - xds-service
        expose:
        - "18000"
    
    networks:
      envoymesh:
        driver: bridge
        ipam:
          config:
            - subnet: 172.31.15.0/24

    resources目录文件

    config.yaml

    name: myconfig
    spec:
      listeners:
      - name: listener_http
        address: 0.0.0.0
        port: 80
        routes:
        - name: local_route
          prefix: /
          clusters:
          - webcluster
      clusters:
      - name: webcluster
        endpoints:
        - address: 172.31.15.11
          port: 8080

    config.yaml-v2

    name: myconfig
    spec:
      listeners:
      - name: listener_http
        address: 0.0.0.0
        port: 80
        routes:
        - name: local_route
          prefix: /
          clusters:
          - webcluster
      clusters:
      - name: webcluster
        endpoints:
        - address: 172.31.15.11
          port: 8080
        - address: 172.31.15.12
          port: 8080

    实验验证

    docker-compose up

    克隆窗口

    # 查看Cluster及Endpoints信息
    root@test:/apps/servicemesh_in_practise-develop/Dynamic-Configuration/ads-grpc# curl 172.31.15.2:9901/clusters
    xds_cluster::observability_name::xds_cluster
    xds_cluster::default_priority::max_connections::1024
    xds_cluster::default_priority::max_pending_requests::1024
    xds_cluster::default_priority::max_requests::1024
    xds_cluster::default_priority::max_retries::3
    xds_cluster::high_priority::max_connections::1024
    xds_cluster::high_priority::max_pending_requests::1024
    xds_cluster::high_priority::max_requests::1024
    xds_cluster::high_priority::max_retries::3
    xds_cluster::added_via_api::false
    xds_cluster::172.31.15.5:18000::cx_active::1
    xds_cluster::172.31.15.5:18000::cx_connect_fail::0
    xds_cluster::172.31.15.5:18000::cx_total::1
    xds_cluster::172.31.15.5:18000::rq_active::4
    xds_cluster::172.31.15.5:18000::rq_error::0
    xds_cluster::172.31.15.5:18000::rq_success::0
    xds_cluster::172.31.15.5:18000::rq_timeout::0
    xds_cluster::172.31.15.5:18000::rq_total::4
    xds_cluster::172.31.15.5:18000::hostname::xdsserver
    xds_cluster::172.31.15.5:18000::health_flags::healthy
    xds_cluster::172.31.15.5:18000::weight::1
    xds_cluster::172.31.15.5:18000::region::
    xds_cluster::172.31.15.5:18000::zone::
    xds_cluster::172.31.15.5:18000::sub_zone::
    xds_cluster::172.31.15.5:18000::canary::false
    xds_cluster::172.31.15.5:18000::priority::0
    xds_cluster::172.31.15.5:18000::success_rate::-1.0
    xds_cluster::172.31.15.5:18000::local_origin_success_rate::-1.0
    webcluster::observability_name::webcluster
    webcluster::default_priority::max_connections::1024
    webcluster::default_priority::max_pending_requests::1024
    webcluster::default_priority::max_requests::1024
    webcluster::default_priority::max_retries::3
    webcluster::high_priority::max_connections::1024
    webcluster::high_priority::max_pending_requests::1024
    webcluster::high_priority::max_requests::1024
    webcluster::high_priority::max_retries::3
    webcluster::added_via_api::true
    webcluster::172.31.15.11:8080::cx_active::0
    webcluster::172.31.15.11:8080::cx_connect_fail::0
    webcluster::172.31.15.11:8080::cx_total::0
    webcluster::172.31.15.11:8080::rq_active::0
    webcluster::172.31.15.11:8080::rq_error::0
    webcluster::172.31.15.11:8080::rq_success::0
    webcluster::172.31.15.11:8080::rq_timeout::0
    webcluster::172.31.15.11:8080::rq_total::0
    webcluster::172.31.15.11:8080::hostname::
    webcluster::172.31.15.11:8080::health_flags::healthy
    webcluster::172.31.15.11:8080::weight::1
    webcluster::172.31.15.11:8080::region::
    webcluster::172.31.15.11:8080::zone::
    webcluster::172.31.15.11:8080::sub_zone::
    webcluster::172.31.15.11:8080::canary::false
    webcluster::172.31.15.11:8080::priority::0
    webcluster::172.31.15.11:8080::success_rate::-1.0
    webcluster::172.31.15.11:8080::local_origin_success_rate::-1.0
    
    #或者查看动态Clusters的相关信息
    root@test:/apps/servicemesh_in_practise-develop/Dynamic-Configuration/ads-grpc# curl -s 172.31.15.2:9901/config_dump | jq '.configs[1].dynamic_active_clusters'
    [
      {
        "version_info": "411",
        "cluster": {
          "@type": "type.googleapis.com/envoy.config.cluster.v3.Cluster",
          "name": "webcluster",
          "type": "EDS",
          "eds_cluster_config": {
            "eds_config": {
              "api_config_source": {
                "api_type": "GRPC",
                "grpc_services": [
                  {
                    "envoy_grpc": {
                      "cluster_name": "xds_cluster"
                    }
                  }
                ],
                "set_node_on_first_message_only": true,
                "transport_api_version": "V3"
              },
              "resource_api_version": "V3"
            }
          },
          "connect_timeout": "5s",
          "dns_lookup_family": "V4_ONLY"
        },
        "last_updated": "2021-12-02T08:05:20.650Z"
      }
    ]
    
    # 查看Listener列表
    root@test:/apps/servicemesh_in_practise-develop/Dynamic-Configuration/ads-grpc# curl 172.31.15.2:9901/listeners
    listener_http::0.0.0.0:80
    
    #或者查看动态的Listener信息
    root@test:/apps/servicemesh_in_practise-develop/Dynamic-Configuration/ads-grpc# curl -s 172.31.15.2:9901/config_dump?resource=dynamic_listeners | jq '.configs[0].active_state.listener.address'
    {
      "socket_address": {
        "address": "0.0.0.0",
        "port_value": 80
      }
    }
    
    # 接入xdsserver容器的交互式接口,修改config.yaml文件中的内容,将另一个endpoint添加进文件中,或进行其它修改;
    root@test:/apps/servicemesh_in_practise-develop/Dynamic-Configuration/ads-grpc# docker exec -it lds-cds-grpc_xdsserver_1 sh
    
    / # cd /etc/envoy-xds-server/config
    /etc/envoy-xds-server/config # cat config.yaml-v2 > config.yaml
    #提示:以上修改操作也可以直接在宿主机上的存储卷目录中进行。
    
    # 再次查看Cluster中的Endpoint信息 
    root@test:/apps/servicemesh_in_practise-develop/Dynamic-Configuration/ads-grpc# curl 172.31.15.2:9901/clusters
    xds_cluster::observability_name::xds_cluster
    xds_cluster::default_priority::max_connections::1024
    xds_cluster::default_priority::max_pending_requests::1024
    xds_cluster::default_priority::max_requests::1024
    xds_cluster::default_priority::max_retries::3
    xds_cluster::high_priority::max_connections::1024
    xds_cluster::high_priority::max_pending_requests::1024
    xds_cluster::high_priority::max_requests::1024
    xds_cluster::high_priority::max_retries::3
    xds_cluster::added_via_api::false
    xds_cluster::172.31.15.5:18000::cx_active::1
    xds_cluster::172.31.15.5:18000::cx_connect_fail::0
    xds_cluster::172.31.15.5:18000::cx_total::1
    xds_cluster::172.31.15.5:18000::rq_active::4
    xds_cluster::172.31.15.5:18000::rq_error::0
    xds_cluster::172.31.15.5:18000::rq_success::0
    xds_cluster::172.31.15.5:18000::rq_timeout::0
    xds_cluster::172.31.15.5:18000::rq_total::4
    xds_cluster::172.31.15.5:18000::hostname::xdsserver
    xds_cluster::172.31.15.5:18000::health_flags::healthy
    xds_cluster::172.31.15.5:18000::weight::1
    xds_cluster::172.31.15.5:18000::region::
    xds_cluster::172.31.15.5:18000::zone::
    xds_cluster::172.31.15.5:18000::sub_zone::
    xds_cluster::172.31.15.5:18000::canary::false
    xds_cluster::172.31.15.5:18000::priority::0
    xds_cluster::172.31.15.5:18000::success_rate::-1.0
    xds_cluster::172.31.15.5:18000::local_origin_success_rate::-1.0
    webcluster::observability_name::webcluster
    webcluster::default_priority::max_connections::1024
    webcluster::default_priority::max_pending_requests::1024
    webcluster::default_priority::max_requests::1024
    webcluster::default_priority::max_retries::3
    webcluster::high_priority::max_connections::1024
    webcluster::high_priority::max_pending_requests::1024
    webcluster::high_priority::max_requests::1024
    webcluster::high_priority::max_retries::3
    webcluster::added_via_api::true
    webcluster::172.31.15.11:8080::cx_active::0
    webcluster::172.31.15.11:8080::cx_connect_fail::0
    webcluster::172.31.15.11:8080::cx_total::0
    webcluster::172.31.15.11:8080::rq_active::0
    webcluster::172.31.15.11:8080::rq_error::0
    webcluster::172.31.15.11:8080::rq_success::0
    webcluster::172.31.15.11:8080::rq_timeout::0
    webcluster::172.31.15.11:8080::rq_total::0
    webcluster::172.31.15.11:8080::hostname::
    webcluster::172.31.15.11:8080::health_flags::healthy
    webcluster::172.31.15.11:8080::weight::1
    webcluster::172.31.15.11:8080::region::
    webcluster::172.31.15.11:8080::zone::
    webcluster::172.31.15.11:8080::sub_zone::
    webcluster::172.31.15.11:8080::canary::false
    webcluster::172.31.15.11:8080::priority::0
    webcluster::172.31.15.11:8080::success_rate::-1.0
    webcluster::172.31.15.11:8080::local_origin_success_rate::-1.0
    webcluster::172.31.15.12:8080::cx_active::0
    webcluster::172.31.15.12:8080::cx_connect_fail::0
    webcluster::172.31.15.12:8080::cx_total::0
    webcluster::172.31.15.12:8080::rq_active::0
    webcluster::172.31.15.12:8080::rq_error::0
    webcluster::172.31.15.12:8080::rq_success::0
    webcluster::172.31.15.12:8080::rq_timeout::0
    webcluster::172.31.15.12:8080::rq_total::0
    webcluster::172.31.15.12:8080::hostname::
    webcluster::172.31.15.12:8080::health_flags::healthy
    webcluster::172.31.15.12:8080::weight::1
    webcluster::172.31.15.12:8080::region::
    webcluster::172.31.15.12:8080::zone::
    webcluster::172.31.15.12:8080::sub_zone::
    webcluster::172.31.15.12:8080::canary::false
    webcluster::172.31.15.12:8080::priority::0
    webcluster::172.31.15.12:8080::success_rate::-1.0
    webcluster::172.31.15.12:8080::local_origin_success_rate::-1.0
    #新节点已加入
    I have a dream so I study hard!!!
  • 相关阅读:
    idea spring boot启动项目上面有红色叉
    hibernate Criteria中多个or和and的用法 and ( or or)
    CAS Server 4.2.7(自定义密码验证) 部署
    Web应用系统集成CAS-rest指南
    用DBMS_REDEFINITION将普通表转换为分区表
    windows编译libevent时报告“缺少print_winsock_errors.obj”的解决
    Oracle表增删分区的脚本
    libcassandra开发示例
    关于MongoDB API的两个小经验
    C/C++开发Cassandra的一些经验
  • 原文地址:https://www.cnblogs.com/yaokaka/p/15633910.html
Copyright © 2020-2023  润新知