-
Notifications
You must be signed in to change notification settings - Fork 200
encap_decap_gre
This test verifies "EthoCWoMPLSoGRE" forwarding of IP traffic using policy-forwarding configuration.
"EthoCWoMPLSoGRE" describes the encapsulation for IPv4/IPv6 packets, including the Ethernet header, all contained within GRE and MPLS headers. In addition 4-byte zero control word (CW) is inserted between the MPLS header and the inner Ethernet header.
+------------------------------------+
| Outer IP Header |
+------------------------------------+
| GRE Header |
+------------------------------------+
| MPLS Label Stack |
+------------------------------------+
| Control Word (CW) |
| (4 bytes - all '0') |
+------------------------------------+
| Inner Ethernet Header |
+------------------------------------+
| Inner IP Header |
+------------------------------------+
| Layer-4 Header |
+------------------------------------+
| Payload |
+------------------------------------+
DUT has an ingress and 2 egress aggregate interfaces.
| | --eBGP-- | ATE Ports 2,3 |
[ ATE Ports 1 ]---- | DUT | | |
| | --eBGP-- | ATE Port 4,5 |
Test uses aggregate 802.3ad bundled interfaces (Aggregate Interfaces).
-
Ingress Port: Traffic is generated from Aggregate1 (ATE Ports 1).
-
Egress Ports: Aggregate2 (ATE Ports 2,3) and Aggregate3 (ATE Ports 4,5) are used as the destination ports for encapsulated traffic.
Aggregate 1 "customer interface" is the ingress port that could either have port mode configuration or attachment mode configuration as described below.
EACH test should be run twice - once with port mode configuration and once with attachment mode configuration.
- Configure DUT port 1 to be a member of aggregate interface named "customer interface"
- "customer interface" is a static Layer 2 bundled interface part of pseudowire that accepts packets from all VLANs.
- MTU default 9216
- Configure DUT port 1 to be a member of aggregate interface named "customer interface"
- Create a sub interface of the aggregated interface and assign a VLAN ID to it.
- This sub interface will be a static Layer 2 bundled interface part of pseudowire that accepts packets from vlan ID associated with it.
- MTU default 9216
-
Policy-forwarding enabling EthoMPLSoGRE encapsulation of all incoming traffic:
-
The forwarding policy must allow forwarding of incoming traffic across 16 tunnels. 16 Tunnels has 16 source address and a single tunnel destination.
-
Source address must be configurable as:
- Loopback address OR
- 16 source addresses corresponding to a single tunnel destinations to achieve maximum entropy.
-
DSCP of the innermost IP packet header must be preserved during encapsulation
-
DSCP of the GRE/outermost IP header must be configurable (Default TOS 96) during encapsulation
-
TTL of the outer GRE must be configurable (Default TTL 64)
-
QoS Hardware queues for all traffic must be configurable (default QoS hardaware class selected is 3)
-
- "Customer interface" is endpoint 1 and endpoint 2 is an IP address pointing towards Aggregate2, Aggregare3
- Two unique static MPLS label for local label and remote label.
- Enable control word
-
IPV4 and IPV6 addresses
-
MTU (default 9216)
-
LACP Member link configuration
-
Lag id
-
LACP (default: period short)
-
Carrier-delay (default up:3000 down:150)
-
Statistics load interval (default:30 seconds)
- Create static route for tunnel destination pointing towards Aggregate 2.
- Static mapping of MPLS label for encapsulation must be configurable
- Entire Label block must be reallocated for static MPLS
- Labels from start/end/mid ranges must be usable and configured corresponding to EthoMPLSoGRE encapsulation
PF-1.23.2: Verify PF EthoMPLSoGRE encapsulate action for unencrytped IPv4, IPv6 traffic with entropy on ethernet headers
- Generate 1000 different traffic flows on ATE Port 1 at line rate with a mix of both IPV4 and IPv6 traffic. Use 64, 128, 256, 512, 1024 MTU bytes frame size.
- Flows should have entropy on Source MAC address, Destination MAC address. Other headers are fixed.
Verify:
- All traffic received on Aggregate2, Aggregate3 are EthoCWoMPLSoGRE-encapsulated.
- No packet loss when forwarding.
- Traffic equally load-balanced across 16 GRE destinations and distributed equally across 2 egress ports.
Run the test separately for both port mode and attachment mode "customer interface" configuration.
PF-1.23.3: Verify no hashing of EthoCWoMPLSoGRE encapsulation for unencrytped IPv4 traffic without entropy
- Generate single traffic flow on ATE Port 1 at line rate with IPV4 traffic. Use 64, 128, 256, 512, 1024 MTU bytes frame size.
- Flows should have NOT have entropy on any headers.
Verify:
- All traffic received on either Aggregate2 only or Aggregate3 and is EthoCWoMPLSoGRE-encapsulated.
- No hashing of traffic between Aggregate2 and Aggregate3.
- No packet loss when forwarding.
- Traffic only takes one of the 16 GRE destinations.
Run the test separately for both port mode and attachment mode "customer interface" configuration.
- Generate 1000 different traffic flows on ATE Port 1 at line rate with a mix of both IPV4 and IPv6 traffic. Use 64, 128, 256, 512, 1024 MTU bytes frame size.
- Flows are MACSec encrypted when sent from ATE Port 1.
- MACSec should encrypt all headers and paylaod of traffic flow, including source mac, destination mac and VLAN tag.
- Flows should have entropy on Source MAC address, Destination MAC address. Other headers are fixed.
Verify:
- All traffic received on either Aggregate2 only or Aggregate3 only and is EthoCWoMPLSoGRE-encapsulated.
- No hashing of traffic between Aggregate2 and Aggregate3
- No packet loss when forwarding.
- Traffic only takes one of the 16 GRE destinations.
Run the test separately for both port mode and attachment mode "customer interface" configuration.
- Use the same traffic profile as PF-1.23.2. However, set the packet size to 9000 bytes
Verify:
- All traffic received on Aggregate2, Aggregate3 are EthoCWoMPLSoGRE-encapsulated.
- No packet loss when forwarding.
- Traffic equally load-balanced across 16 GRE destinations and distributed equally across 2 egress ports.
Run the test separately for both port mode and attachment mode "customer interface" configuration.
- Use the same traffic profile as PF-1.23.2.
Verify:
- Verify a “0” (32-bit field ) control word is inserted between the MPLS label stack and the Layer 2 payload (the Ethernet frame).
- All traffic received on Aggregate2, Aggregate3 are EthoCWoMPLSoGRE-encapsulated.
- No packet loss when forwarding.
- Traffic equally load-balanced across 16 GRE destinations and distributed equally across 2 egress ports.
Run the test separately for both port mode and attachment mode "customer interface" configuration.
- Use the same traffic profile as PF-1.23.4
Verify:
- Verify a “0” (32-bit field ) control word is inserted between the MPLS label stack and the Layer 2 payload (the Ethernet frame).
- All traffic received on either Aggregate2 only or Aggregate3 only and is EthoCWoMPLSoGRE-encapsulated.
- No hashing of traffic between Aggregate2 and Aggregate3
- No packet loss when forwarding.
- Traffic only takes one of the 16 GRE destinations.
Run the test separately for both port mode and attachment mode "customer interface" configuration.
- Use the same traffic profile as PF-1.23.2.
Verify:
- TOS of encapsulated packets is set to 96.
- All traffic received on Aggregate2, Aggregate3 are EthoCWoMPLSoGRE-encapsulated.
- No packet loss when forwarding.
- Traffic equally load-balanced across 16 GRE destinations and distributed equally across 2 egress ports.
Run the test separately for both port mode and attachment mode "customer interface" configuration.
- Use the same traffic profile as PF-1.23.2.
- Start the traffic profile on ATE.
- Shutdown or remove a single GRE tunnel destination on the DUT.
Verify:
- All traffic received on Aggregate2, Aggregate3 are EthoCWoMPLSoGRE-encapsulated.
- No packet loss when forwarding.
- Traffic load-balanced across remaining 15 GRE destinations and distributed equally across 2 egress ports.
Generate traffic on ATE Aggregate2 and Aggregate3 having:
-
Outer source address: random combination of 1000+ IPV4 source addresses from 100.64.0.0/22
-
Outer destination address: Traffic must fall within the configured GRE tunnel sources in PF-1.23.1 so it cuold be decapsulated.
-
MPLS Label: Should be same as the local label configured in PF-1.23.1 Inner payload:
-
Both IPV4 and IPV6 unicast payloads, with random source address, destination address, TCP/UDP source port and destination ports
-
Use 64, 128, 256, 512, 1024.. MTU bytes frame size
-
Verify:
-
All traffic received on Aggregate2 and Aggregate3 gets decapsulated and forwarded as IPV4/IPV6 unicast on the respective egress interfaces under Aggregate1
-
No packet loss when forwarding with counters incrementing corresponding to traffic
Run the test separately for both port mode and attachment mode "customer interface" configuration.
-
Use the same traffic profile as PF-1.23.12.
-
Ensure inner payload Ethernet header has a VLAN tag attached to it.
-
Verify:
-
In port mode configuration: Traffic flow is decapsulated and mapped to the resperctive pseudowire and egress interface based on the MPLS label. Inner payload VLAN tag is retained after decapsulation.
-
In attachment mode configuration: Traffic flow is decapsulated and mapped to the resperctive pseudowire and egress interface based on the MPLS label. VLAN tag of decapsulated packet is same as the VLAN-ID associated with the egress sub-interface.
-
No packet loss when forwarding with counters incrementing corresponding to traffic
Run the test separately for both port mode and attachment mode "customer interface" configuration.
Port mode Interface configs
{
"interfaces": {
"interface": [
{
"config": {
"description": "CLOUD-CSI",
"enabled": true,
"mtu": 9080,
"name": "Bundle-Ether8",
"type": "ieee8023adLag"
},
"name": "Bundle-Ether8",
"rates": {
"config": {
"load-interval": 30
}
},
"subinterfaces": {
"subinterface": [
{
"config": {
"description": "CLOUD-CSI",
"enabled": true,
"index": 0
},
"index": 0,
"ipv4": {
"config": {
"mtu": 9066
}
},
"ipv6": {
"config": {
"mtu": 9066
}
}
}
]
},
"aggregation": {
"config": {
"lag-type": "STATIC"
}
}
}
]
}
}VLAN mode Interface configs
{
"interfaces": {
"interface": [
{
"config": {
"description": "CLOUD-CSI",
"enabled": true,
"mtu": 9080,
"name": "Bundle-Ether9",
"type": "ieee8023adLag"
},
"name": "Bundle-Ether9",
"rates": {
"config": {
"load-interval": 30
}
},
"subinterfaces": {
"subinterface": [
{
"config": {
"description": "CLOUD-GEO-PRIVATE [T=qp1309122726287]",
"enabled": true,
"index": 1709
},
"index": 1709,
"vlan": {
"match": {
"single-tagged": {
"config": {
"vlan-id": 1709
}
}
}
}
}
]
}
}
]
}
}Pseudowire configs Port mode
{
"network-instances": {
"network-instance": [
{
"name": "GEO_4",
"config": {
"name": "GEO_4",
"type": "L2P2P"
},
"connection-points": {
"connection-point": [
{
"connection-point-id": "GEO_4",
"config": {
"connection-point-id": "GEO_4"
},
"endpoints": {
"endpoint": [
{
"endpoint-id": "LOCAL",
"config": {
"endpoint-id": "LOCAL",
"type": "LOCAL"
},
"local": {
"config": {
"interface": "Bundle-Ether9",
"subinterface": 0
}
}
},
{
"endpoint-id": "REMOTE",
"config": {
"endpoint-id": "REMOTE",
"type": "REMOTE"
},
"remote": {
"config": {
"virtual-circuit-identifier": 4
}
}
}
]
}
}
]
}
}
]
},
"interfaces": {
"interface": [
{
"name": "Bundle-Ether9",
"config": {
"name": "Bundle-Ether9",
"type": "IANA_INTERFACE_TYPE:ieee8023adLag",
"enabled": true
},
"ethernet": {
"config": {
"aggregate-id": "Bundle-Ether9"
}
},
"subinterfaces": {
"subinterface": [
{
"index": 0,
"config": {
"index": 0,
"description": "L2P2P Service GEO_4 - Local Endpoint"
}
}
]
}
}
]
}
}Pseudowire configs VLAN mode
{
"network-instances": {
"network-instance": [
{
"name": "GEO_4",
"config": {
"name": "GEO_4",
"type": "L2P2P"
},
"connection-points": {
"connection-point": [
{
"connection-point-id": "GEO_4",
"config": {
"connection-point-id": "GEO_4"
},
"endpoints": {
"endpoint": [
{
"endpoint-id": "LOCAL",
"config": {
"endpoint-id": "LOCAL",
"type": "LOCAL"
},
"local": {
"config": {
"interface": "Bundle-Ether9",
"subinterface": 1709
}
}
},
{
"endpoint-id": "REMOTE",
"config": {
"endpoint-id": "REMOTE",
"type": "REMOTE"
},
"remote": {
"config": {
"virtual-circuit-identifier": 4
}
}
}
]
}
}
]
}
}
]
},
"interfaces": {
"interface": [
{
"name": "Bundle-Ether9",
"config": {
"name": "Bundle-Ether9",
"type": "IANA_INTERFACE_TYPE:ieee8023adLag",
"enabled": true
},
"ethernet": {
"config": {
"aggregate-id": "Bundle-Ether9"
}
},
"subinterfaces": {
"subinterface": [
{
"index": 1709,
"config": {
"index": 1709,
"description": "L2P2P Service GEO_4 - Local Endpoint (VLAN implied by index)"
}
}
]
}
}
]
}
}Tunnels/Next-hop group configs
{
"network-instances": {
"network-instance": [
{
"name": "DEFAULT",
"config": {
"name": "DEFAULT"
},
"static": {
"next-hop-groups": {
"next-hop-group": [
{
"config": {
"name": "MPLS_in_GRE_Encap"
},
"name": "MPLS_in_GRE_Encap",
"next-hops": {
"next-hop": [
{
"index": "1",
"config": {
"index": "1"
}
},
{
"index": "2",
"config": {
"index": "2"
}
}
]
}
}
]
},
"next-hops": {
"next-hop": [
{
"index": "1",
"config": {
"index": "1",
"next-hop": "nh_ip_addr_1",
"encap-headers": {
"encap-header": [
{
"index": "1",
"type": "GRE",
"config": {
"dst-ip": "outer_ipv4_dst_def",
"src-ip": "outer_ipv4_src1",
"dscp": "outer_dscp",
"ip-ttl": "outer_ip-ttl"
}
}
]
}
}
},
{
"index": "2",
"config": {
"index": "2",
"next-hop": "nh_ip_addr_2",
"encap-headers": {
"encap-header": [
{
"index": "2",
"type": "GRE",
"config": {
"dst-ip": "outer_ipv4_dst_def",
"src-ip": "outer_ipv4_src2",
"dscp": "outer_dscp",
"ip-ttl": "outer_ip-ttl"
}
}
]
}
}
}
]
}
}
}
]
}
}paths:
# interface configs
/interfaces/interface/config/description:
/interfaces/interface/config/enabled:
/interfaces/interface/config/mtu:
/interfaces/interface/config/name:
/interfaces/interface/config/type:
/interfaces/interface/rates/config/load-interval:
/interfaces/interface/subinterfaces/subinterface/config/description:
/interfaces/interface/subinterfaces/subinterface/config/enabled:
/interfaces/interface/subinterfaces/subinterface/config/index:
/interfaces/interface/subinterfaces/subinterface/ipv4/config/mtu:
/interfaces/interface/subinterfaces/subinterface/ipv6/config/mtu:
/interfaces/interface/aggregation/config/lag-type:
# psuedowire configs
/network-instances/network-instance/config/name:
/network-instances/network-instance/config/type:
/network-instances/network-instance/connection-points/connection-point/config/connection-point-id:
/network-instances/network-instance/connection-points/connection-point/endpoints/endpoint/config/endpoint-id:
/network-instances/network-instance/connection-points/connection-point/endpoints/endpoint/local/config/interface:
/network-instances/network-instance/connection-points/connection-point/endpoints/endpoint/local/config/subinterface:
/network-instances/network-instance/connection-points/connection-point/endpoints/endpoint/remote/config/virtual-circuit-identifier:
#TODO: Add new OCs for labels and next-hop-group under connection-point
#/network-instances/network-instance/connection-points/connection-point/endpoints/endpoint/local/config/local-label
#/network-instances/network-instance/connection-points/connection-point/endpoints/endpoint/remote/config/remote-label
#/network-instances/network-instance/connection-points/connection-point/endpoints/endpoint/remote/config/next-hop-group
#Tunnels/Next-hop group configs
#TODO: Add new OC for GRE encap headers
#/network-instances/network-instance/static/next-hop-groups/next-hop-group/nexthops/nexthop/config/index:
#/network-instances/network-instance/static/next-hop-groups/next-hop-group/nexthops/nexthop/config/next-hop:
#/network-instances/network-instance/static/next-hop-groups/next-hop-group/nexthops/nexthop/encap-headers/encap-header/config/index:
#/network-instances/network-instance/static/next-hop-groups/next-hop-group/nexthops/nexthop/encap-headers/encap-header/gre/config/type:
#/network-instances/network-instance/static/next-hop-groups/next-hop-group/nexthops/nexthop/encap-headers/encap-header/gre/config/dst-ip:
#/network-instances/network-instance/static/next-hop-groups/next-hop-group/nexthops/nexthop/encap-headers/encap-header/gre/config/src-ip:
#/network-instances/network-instance/static/next-hop-groups/next-hop-group/nexthops/nexthop/encap-headers/encap-header/gre/config/dscp:
#/network-instances/network-instance/static/next-hop-groups/next-hop-group/nexthops/nexthop/encap-headers/encap-header/gre/config/ip-ttl:
#/network-instances/network-instance/static/next-hop-groups/next-hop-group/nexthops/nexthop/encap-headers/encap-header/gre/config/index:
# Telemetry paths
/interfaces/interface/state/counters/in-discards:
/interfaces/interface/state/counters/in-errors:
/interfaces/interface/state/counters/in-multicast-pkts:
/interfaces/interface/state/counters/in-pkts:
/interfaces/interface/state/counters/in-unicast-pkts:
/interfaces/interface/state/counters/out-discards:
/interfaces/interface/state/counters/out-errors:
/interfaces/interface/state/counters/out-multicast-pkts:
/interfaces/interface/state/counters/out-pkts:
/interfaces/interface/state/counters/out-unicast-pkts:
/interfaces/interface/subinterfaces/subinterface/state/counters/in-discards:
/interfaces/interface/subinterfaces/subinterface/state/counters/in-errors:
/interfaces/interface/subinterfaces/subinterface/state/counters/in-multicast-pkts:
/interfaces/interface/subinterfaces/subinterface/state/counters/in-pkts:
/interfaces/interface/subinterfaces/subinterface/state/counters/in-unicast-pkts:
/interfaces/interface/subinterfaces/subinterface/state/counters/out-discards:
/interfaces/interface/subinterfaces/subinterface/state/counters/out-errors:
/interfaces/interface/subinterfaces/subinterface/state/counters/out-multicast-pkts:
/interfaces/interface/subinterfaces/subinterface/state/counters/out-pkts:
/interfaces/interface/subinterfaces/subinterface/state/counters/out-unicast-pkts:
# Config paths for GRE decap
/network-instances/network-instance/policy-forwarding/policies/policy/rules/rule/action/config/decapsulate-gre:
rpcs:
gnmi:
gNMI.Set:
union_replace: true
replace: true
gNMI.Subscribe:
on_change: true- MFF
- FFF
-
Home
- Test Plans
- ACCTZ-1.1: Record Subscribe Full
- ACCTZ-2.1: Record Subscribe Partial
- ACCTZ-3.1: Record Subscribe Non-gRPC
- ACCTZ-4.1: Record History Truncation
- ACCTZ-4.2: Record Payload Truncation
- ACCTZ-5.1: gNSI.acctz.v1 (Accounting) Test RecordSubscribe Idle Timeout - client becomes silent
- ACCTZ-6.1: gNSI.acctz.v1 (Accounting) Test RecordSubscribe Idle Timeout - DoA client
- ACCTZ-7.1: gNSI.acctz.v1 (Accounting) Test Accounting Authentication Failure - Multi-transaction
- ACCTZ-8.1: gNSI.acctz.v1 (Accounting) Test Accounting Authentication Failure - Uni-transaction
- ACCTZ-9.1: gNSI.acctz.v1 (Accounting) Test Accounting Privilege Escalation
- ACCTZ-10.1: gNSI.acctz.v1 (Accounting) Test Accounting Authentication Error - Multi-transaction
- ACL-1.1: ACL match based on L3/L4 fields and DSCP value
- ACL-1.2: ACL Update (Make-before-break)
- ACL-1.3: Large Scale ACL with TCAM profile
- AFT-1.1: AFTs Base
- AFT-1.2: AFTs slow collector
- AFT-1.3: AFTs collector Flap
- AFT-2.1: AFTs Prefix Counters
- AFT-3.1: AFTs Atomic Flag Check
- AFT-5.1: AFTs DUT Reboot
- attestz-1: General enrollz and attestz tests
- Authz: General Authz (1-4) tests
- BMP-1.1: BMP Session Establishment and Telemetry Test
- BMP-2.7: BMP Pre Policy Test
- BMP-2.8: BMP Post Policy Test
- bootz: General bootz bootstrap tests
- CERTZ-1: gNSI Client Certificate Tests
- Certz-2: Server Certificate
- Certz-3: Server Certificate Rotation
- Certz-4: Trust Bundle
- Certz-5: Trust Bundle Rotation
- CFM-1.1: CFM over ETHoCWoMPLSoGRE
- CNTR-1: Basic container lifecycle via
gnoi.Containerz. - CNTR-2: Container network connectivity tests
- CPT-1.1: Interface based ARP policer
- Credentialz-1: Password console login
- Credentialz-2: SSH Password Login Disallowed
- Credentialz-3: Host Certificates
- Credentialz-4: SSH Public Key Authentication
- Credentialz-5: Hiba Authentication
- DP-1.2: QoS policy feature config
- DP-1.3: QoS ECN feature config
- DP-1.4: QoS Interface Output Queue Counters
- DP-1.5: Egress Strict Priority scheduler with bursty traffic
- DP-1.7: One strict priority queue traffic test
- DP-1.8: Two strict priority queue traffic test
- DP-1.9: WRR traffic test
- DP-1.10: Mixed strict priority and WRR traffic test
- DP-1.11: Bursty traffic test
- DP-1.12: ECN enabled traffic test
- DP-1.13: DSCP and ECN bits are copied over during IPinIP encap and decap
- DP-1.14: QoS basic test
- DP-1.15: Egress Strict Priority scheduler
- DP-1.16: Ingress traffic classification and rewrite
- DP-1.17: DSCP Transparency with ECN
- DP-1.19: Egress traffic DSCP rewrite
- DP-2.2: QoS scheduler with 1 rate 2 color policer, classifying on next-hop group
- DP-2.4: Police traffic on input matching all packets using 1 rate, 2 color marker
- DP-2.5: Police traffic on input matching all packets using 2 rate, 3 color marker
- DP-2.6: Police traffic on input matching all packets using 2 rate, 3 color marker with classifier
- enrollz-1: enrollz test for TPM 2.0 HMAC-based Enrollment flow
- enrollz-2: enrollz test for TPM 1.2 Enrollment flow
- example-0.1: Topology Test
- FP-1.1: Power admin DOWN/UP Test
- gNMI-1.1: cli Origin
- gNMI-1.2: Benchmarking: Full Configuration Replace
- gNMI-1.3: Benchmarking: Drained Configuration Convergence Time
- gNMI-1.4: Telemetry: Inventory
- gNMI-1.5: Telemetry: Port Speed Test
- gNMI-1.6: System gRPC Servers running in more than one network-instance
- gNMI-1.8: Configuration Metadata-only Retrieve and Replace
- gNMI-1.9: Get requests
- gNMI-1.10: Telemetry: Basic Check
- gNMI-1.11: Telemetry: Interface Packet Counters
- gNMI-1.12: Mixed OpenConfig/CLI Origin
- gNMI-1.13: Optics Telemetry, Instant, threshold, and miscellaneous static info
- gNMI-1.14: OpenConfig metadata consistency during large config push
- gNMI-1.15: Set Requests
- gNMI-1.16: Fabric redundnacy test
- gNMI-1.17: Controller card redundancy test
- gNMI-1.18: gNMI subscribe with sample mode for backplane capacity counters
- gNMI-1.19: ConfigPush and ConfigPull after Control Card switchover
- gNMI-1.20: Telemetry: Optics Thresholds
- gNMI-1.21: Integrated Circuit Hardware Resource Utilization Test
- gNMI-1.22: Controller card port attributes
- gNMI-1.23: Telemetry: Aggregate Interface Counters
- gNMI-1.24: gNMI Leaf-List Update Test
- gNMI-1.25: Telemetry: Interface Last Change Timestamp
- gNMI-1.26: Carrier Transitions Test
- gNMI-1.27: gNMI Sample Mode Test
- GNMI-2: gnmi_subscriptionlist_test
- gNOI-2.1: Packet-based Link Qualification on 100G and 400G links
- gNOI-3.1: Complete Chassis Reboot
- gNOI-3.2: Per-Component Reboot
- gNOI-3.3: Supervisor Switchover
- gNOI-3.4: Chassis Reboot Status and Reboot Cancellation
- gNOI-4.1: Software Upgrade
- gNOI-5.1: Ping Test
- gNOI-5.2: Traceroute Test
- gNOI-5.3: Copying Debug Files
- gNOI-6.1: Factory Reset
- gNOI-7.1: BootConfig
- gNPSI-1: Sampling and Subscription Check
- HA-1.0: Telemetry: Firewall High Availability.
- Health-1.1: Generic Health Check
- Health-1.2: Healthz component status paths
- INT-1.1: Interface Performance
- IPSEC-1.1: IPSec with MACSec over aggregated links.
- IPSEC-1.2: IPSec Scaling with MACSec over aggregated links.
- IPSEC-1.3: IPSec Packet-Order with MACSec over aggregated links.
- MGT-1: Management HA solution test
- MPLS-1.1: MPLS label blocks using ISIS
- MPLS-1.2: MPLS Traffic Class Marking
- MPLS-2.2: MPLS forwarding via static LSP to BGP next-hop.
- MTU-1.3: Large IP Packet Transmission
- MTU-1.4: Large IP Packet through GRE/GUE tunnel Transmission
- MTU-1.5: Path MTU handing
- OC-1.2: Default Address Families
- OC-26.1: Network Time Protocol (NTP)
- P4RT-1.1: Base P4RT Functionality
- P4RT-1.2: P4RT Daemon Failure
- P4RT-1.3: P4RT behavior when a device/node is dowm
- P4RT-2.1: P4RT Election
- P4RT-2.2: P4RT Metadata Validation
- P4RT-3.1: Google Discovery Protocol: PacketIn
- P4RT-3.2: Google Discovery Protocol: PacketOut
- P4RT-3.21: Google Discovery Protocol: PacketOut with LAG
- P4RT-5.1: Traceroute: PacketIn
- P4RT-5.2: Traceroute Packetout
- P4RT-5.3: Traceroute: PacketIn With VRF Selection
- P4RT-6.1: Required Packet I/O rate: Performance
- P4RT-7.1: LLDP: PacketIn
- P4RT-7.2: LLDP: PacketOut
- PF-1.1: IPv4/IPv6 policy-forwarding to indirect NH matching DSCP/TC.
- PF-1.2: Policy-based traffic GRE Encapsulation to IPv4 GRE tunnel
- PF-1.3: Policy-based IPv4 GRE Decapsulation
- PF-1.4: GUEv1 Decapsulation rule using destination-address-prefix-set and TTL and DSCP behavior test
- PF-1.6: Policy based VRF selection for IPV4/IPV6
- PF-1.7: Decapsulate MPLS in GRE and UDP
- PF-1.8: Ingress handling of TTL
- PF-1.9: Egress handling of TTL
- PF-1.11: Rewrite the ingress innner packet TTL
- PF-1.12: MPLSoGRE IPV4 decapsulation of IPV4/IPV6 payload
- PF-1.13: MPLSoGRE IPV4 decapsulation of IPV4/IPV6 payload scale test
- PF-1.14: MPLSoGRE IPV4 encapsulation of IPV4/IPV6 payload
- PF-1.15: MPLSoGRE IPV4 encapsulation of IPV4/IPV6 payload scale test
- PF-1.16: MPLSoGRE IPV4 encapsulation IPV4/IPV6 local proxy test
- PF-1.17: MPLSoGRE and MPLSoGUE MACsec
- PF-1.18: MPLSoGRE and MPLSoGUE QoS
- PF-1.19: MPLSoGUE IPV4 decapsulation of IPV4/IPV6 payload
- PF-1.20: MPLSoGUE IPV4 decapsulation of IPV4/IPV6 payload scale test
- PF-1.21: Configurable IPv6 flow labels corresponding to IPV6 tunnels
- PF-1.22: GUEv1 Decapsulation and ECMP test for IPv4 and IPv6 payload
- PF-1.23: EthoCWoMPLSoGRE IPV4 forwarding of IPV4/IPV6 payload
- PF-1.24: Add and remove interface bound to PBF
- PF-2.3: Multiple VRFs and GUE DECAP in Default VRF
- PLT-1.1: Interface breakout Test
- PLT-1.2: Parent component validation test
- PLT-1.3: OnChange Subscription Test for Breakout Interfaces
- Replay-1.0: Record/replay presession test
- Replay-1.1: Record/replay diff command trees test
- Replay-1.2: P4RT Replay Test
- RT-1.1: Base BGP Session Parameters
- RT-1.2: BGP Policy & Route Installation
- RT-1.3: BGP Route Propagation
- RT-1.4: BGP Graceful Restart
- RT-1.5: BGP Prefix Limit
- RT-1.7: Local BGP Test
- RT-1.8: BGP Route Reflector Test at scale
- RT-1.10: BGP Keepalive and HoldTimer Configuration Test
- RT-1.11: BGP remove private AS
- RT-1.12: BGP always compare MED
- RT-1.14: BGP Long-Lived Graceful Restart
- RT-1.15: BGP Addpath on scale with and without routing policy
- RT-1.19: BGP 2-Byte and 4-Byte ASN support
- RT-1.21: BGP TCP MSS and PMTUD
- RT-1.23: BGP AFI SAFI OC DEFAULTS
- RT-1.24: BGP 2-Byte and 4-Byte ASN support with policy
- RT-1.25: Management network-instance default static route
- RT-1.26: Basic static route support
- RT-1.27: Static route to BGP redistribution
- RT-1.28: BGP to IS-IS redistribution
- RT-1.29: BGP chained import/export policy attachment
- RT-1.30: BGP nested import/export policy attachment
- RT-1.31: BGP 3 levels of nested import/export policy with match-set-options
- RT-1.32: BGP policy actions - MED, LocPref, prepend, flow-control
- RT-1.33: BGP Policy with prefix-set matching
- RT-1.34: BGP route-distance configuration
- RT-1.35: BGP Graceful Restart Extended route retention (ExRR)
- RT-1.51: BGP multipath ECMP
- RT-1.52: BGP multipath UCMP support with Link Bandwidth Community
- RT-1.53: prefix-list test
- RT-1.54: BGP Override AS-path split-horizon
- RT-1.55: BGP session mode (active/passive)
- RT-1.63: BGP Multihop
- RT-1.64: BGP Import/Export Policy (Control plane only) Functional Test Case
- RT-1.65: BGP scale test
- RT-1.66: IPv4 Static Route with IPv6 Next-Hop
- RT-2.1: Base IS-IS Process and Adjacencies
- RT-2.2: IS-IS LSP Updates
- RT-2.6: IS-IS Hello-Padding enabled at interface level
- RT-2.7: IS-IS Passive is enabled at interface level
- RT-2.8: IS-IS metric style wide not enabled
- RT-2.9: IS-IS metric style wide enabled
- RT-2.10: IS-IS change LSP lifetime
- RT-2.11: IS-IS Passive is enabled at the area level
- RT-2.12: Static route to IS-IS redistribution
- RT-2.13: Weighted-ECMP for IS-IS
- RT-2.14: IS-IS Drain Test
- RT-2.15: IS-IS Extensions for Segment Routing
- RT-2.16: IS-IS Graceful Restart Helper
- RT-3.1: Policy based VRF selection
- RT-3.2: Multiple <Protocol, DSCP> Rules for VRF Selection
- RT-3.52: Multidimensional test for Static GUE Encap/Decap based on BGP path selection and selective DSCP marking
- RT-3.53: Static route based GUE Encapsulation to IPv6 tunnel
- RT-4.10: AFTs Route Summary
- RT-4.11: AFTs Route Summary
- RT-5.1: Singleton Interface
- RT-5.2: Aggregate Interfaces
- RT-5.3: Aggregate Balancing
- RT-5.4: Aggregate Forwarding Viable
- RT-5.5: Interface hold-time
- RT-5.6: Interface Loopback mode
- RT-5.7: Aggregate Not Viable All
- RT-5.8: IPv6 Link Local
- RT-5.9: Disable IPv6 ND Router Arvetisment
- RT-5.10: IPv6 Link Local generated by SLAAC
- RT-5.11: LACP Intervals
- RT-5.12: Suppress IPv6 ND Router Advertisement [Depreciated]
- RT-5.13: Flow control test
- RT-6.1: Core LLDP TLV Population
- RT-7.1: BGP default policies
- RT-7.2: BGP Policy Community Set
- RT-7.3: BGP Policy AS Path Set
- RT-7.4: BGP Policy AS Path Set and Community Set
- RT-7.5: BGP Policy - Match and Set Link Bandwidth Community
- RT-7.6: BGP Link Bandwidth Community - Cumulative
- RT-7.8: BGP Policy Match Standard Community and Add Community Import/Export Policy
- RT-7.9: BGP ECMP for iBGP with IS-IS protocol nexthop
- RT-7.10: Routing policy statement insertion and removal
- RT-7.11: BGP Policy - Import/Export Policy Action Using Multiple Criteria
- RT-7.51: BGP Auto-Generated Link-Bandwidth Community
- RT-8: Singleton with breakouts
- RT-10.1: Default Route Generation based on 192.0.0.0/8 Presence
- RT-10.2: Non-default Route Generation based on 192.168.2.2/32 Presence in ISIS
- RT-14.2: GRIBI Route Test
- SEC-3.1: Authentication
- SFLOW-1: sFlow Configuration and Sampling
- SR-1.1: Transit forwarding to Node-SID via ISIS
- SR-1.2: Egress Node Forwarding for MPLS traffic with Explicit Null label
- Storage-1.1: Storage File System Check
- SYS-1.1: Test default COPP policy thresholds for Arista
- SYS-2.1: Ingress control-plane ACL.
- SYS-3.1: AAA and TACACS+ Configuration Verification Test Suite
- SYS-4.1: System Mount Points State Verification
- System-1.1: System banner test
- System-1.2: System g protocol test
- System-1.3: System hostname test
- System-1.4: System time test
- System-1.5: System software-version test
- TE-1.1: Static ARP
- TE-1.2: My Station MAC
- TE-2.1: gRIBI IPv4 Entry
- TE-2.2: gRIBI IPv4 Entry With Aggregate Ports
- TE-3.1: Base Hierarchical Route Installation
- TE-3.2: Traffic Balancing According to Weights
- TE-3.3: Hierarchical weight resolution
- TE-3.5: Ordering: ACK Received
- TE-3.6: ACK in the Presence of Other Routes
- TE-3.7: Base Hierarchical NHG Update
- TE-3.31: Hierarchical weight resolution with PBF
- TE-4.1: Base Leader Election
- TE-4.2: Persistence Mode
- TE-5.1: gRIBI Get RPC
- TE-6.1: Route Removal via Flush
- TE-6.2: Route Removal In Non Default VRF
- TE-6.3: Route Leakage between Non Default VRF
- TE-8.1: DUT Daemon Failure
- TE-8.2: Supervisor Failure
- TE-9.2: MPLS based forwarding Static LSP
- TE-9.3: FIB FAILURE DUE TO HARDWARE RESOURCE EXHAUST
- TE-9: gRIBI MPLS Compliance
- TE-10: gRIBI MPLS Forwarding
- TE-11.1: Backup NHG: Single NH
- TE-11.2: Backup NHG: Multiple NH
- TE-11.3: Backup NHG: Actions
- TE-11.21: Backup NHG: Multiple NH with PBF
- TE-11.31: Backup NHG: Actions with PBF
- TE-13.1: gRIBI route ADD during Failover
- TE-13.2: gRIBI route DELETE during Failover
- TE-14.1: gRIBI Scaling
- TE-14.2: encap and decap scale
- TE-15.1: gRIBI Compliance
- TE-16.1: basic encapsulation tests
- TE-16.2: encapsulation FRR scenarios
- TE-16.3: encapsulation FRR scenarios
- TE-17.1: VRF selection policy driven TE
- TE-18.1: gRIBI MPLS-in-UDP Encapsulation
- TE-18.3: MPLS in UDP Encapsulation Scale Test
- TE-18.4: ECMP hashing on outer and inner packets with MPLSoUDP encapsulation
- TestID-16.4: gRIBI to BGP Route Redistribution for IPv4
- TR-6.1: Remote Syslog feature config
- TR-6.2: Local logging destinations
- TRANSCEIVER-1.1: Telemetry: 400ZR Chromatic Dispersion(CD) telemetry values streaming
- TRANSCEIVER-1.2: Telemetry: 400ZR_PLUS Chromatic Dispersion(CD) telemetry values streaming
- TRANSCEIVER-3.1: Telemetry: 400ZR Optics firmware version streaming
- TRANSCEIVER-3.2: Telemetry: 400ZR_PLUS Optics firmware version streaming
- TRANSCEIVER-4.1: Telemetry: 400ZR RX input and TX output power telemetry values streaming.
- TRANSCEIVER-4.2: Telemetry: 400ZR_PLUS RX input and TX output power telemetry values streaming.
- TRANSCEIVER-5.1: Configuration: 400ZR channel frequency, output TX launch power and operational mode setting.
- TRANSCEIVER-5.2: Configuration: 400ZR_PLUS channel frequency, output TX launch power and operational mode setting.
- TRANSCEIVER-6.1: Telemetry: 400ZR Optics performance metrics (pm) streaming.
- TRANSCEIVER-6.2: Telemetry: 400ZR_PLUS Optics performance metrics (pm) streaming.
- TRANSCEIVER-7.1: Telemetry: 400ZR Optics inventory info streaming
- TRANSCEIVER-7.2: Telemetry: 400ZR_PLUS Optics inventory info streaming
- TRANSCEIVER-8.1: Telemetry: 400ZR Optics module temperature streaming.
- TRANSCEIVER-8.2: Telemetry: 400ZR_PLUS Optics module temperature streaming.
- TRANSCEIVER-9.1: Telemetry: 400ZR TX laser bias current telemetry values streaming.
- TRANSCEIVER-9.2: Telemetry: 400ZR_PLUS TX laser bias current telemetry values streaming.
- TRANSCEIVER-10.1: Telemetry: 400ZR Optics FEC(Forward Error Correction) Uncorrectable Frames Streaming.
- TRANSCEIVER-10.2: Telemetry: 400ZR_PLUS Optics FEC(Forward Error Correction) Uncorrectable Frames Streaming.
- TRANSCEIVER-11.1: Telemetry: 400ZR Optics logical channels provisioning and related telemetry.
- TRANSCEIVER-11.2: Telemetry: 400ZR_PLUS Optics logical channels provisioning and related telemetry.
- TRANSCEIVER-12.1: Telemetry: 400ZR Transceiver Supply Voltage streaming.
- TRANSCEIVER-12.2: Telemetry: 400ZR_PLUS Transceiver Supply Voltage streaming.
- TRANSCEIVER-13.1: Configuration: 400ZR Transceiver Low Power Mode Setting.
- TRANSCEIVER-13.2: Configuration: 400ZR_PLUS Transceiver Low Power Mode Setting.
- TRANSCEIVER-101: Telemetry: ZR platform OC paths streaming.
- TRANSCEIVER-102: Telemetry: ZR terminal-device OC paths streaming.
- TRANSCEIVER-103: Telemetry: ZR Plus platform OC paths streaming.
- TRANSCEIVER-104: Telemetry: ZR Plus terminal-device OC paths streaming.
- TRANSCEIVER-105: Telemetry: ZR platform OC paths streaming.
- TRANSCEIVER-106: Telemetry: ZR terminal-device OC paths streaming.
- TRANSCEIVER-107: Telemetry: ZR Plus platform OC paths streaming.
- TRANSCEIVER-108: Telemetry: ZR Plus terminal-device OC paths streaming.
- TUN-1.3: Interface based IPv4 GRE Encapsulation
- TUN-1.4: Interface based IPv6 GRE Encapsulation
- TUN-1.6: Tunnel End Point Resize for Ecapsulation - Interface Based GRE Tunnel
- TUN-1.9: GRE inner packet DSCP
- URPF-1.1: uRPF validation from non-default network-instance
- Test Plans