Defects Closed with Code Changes

The following defects, which were previously disclosed as open, were resolved in Extreme Fabric Automation 2.5.0 and 2.5.1.

Parent Defect ID: EFA-5732 Issue ID: EFA-5732
Severity: S2 - High
Product: Extreme Fabric Automation Reported in Release: EFA 2.2.0
Symptom: When firmware download is in progress, fabric delete command is accepted without an error.
Condition: If fabric delete command is submitted when firmware download is in progress, it fails.
Workaround:

Allow firmware download process to complete. Status of the same can be checked using command

efa inventory device firmware-download show --fabric {fabric name}

Recovery: Fabric can be deleted once the firmware download is completed
Parent Defect ID: EFA-6501 Issue ID: EFA-6501
Severity: S3 - Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.3.0
Symptom: Configuration Drift for VRF still shown in "cfg-in-sync" though its child configuration are drifted on SLX switch.
Condition:

With below steps issue can be observed.

1) Create VRF/EPG having route target, static route and bgp configuration.

2) Introduce drift in VRF route target or static route or bgp configuration on SLX switch.

3) Update device from efa command "efa inventory device update --ip <device ip>"

4) Check device drift using efa command as "efa inventory drift-reconcile execute --ip <device ip>"

5) VRF shows as "cfg-in-sync" though its child configuration was drifted.

Workaround: None
Recovery: After drift and reconcile all EFA and device configuration will be in sync.
Parent Defect ID: EFA-7324 Issue ID: EFA-7324
Severity: S3 - Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.3.0
Symptom: Continuous create/delete of BGP peer-group and peer can finally cause CLI errors
Condition: When create and delete BGP peer/peer-group is repeatedly done in a loop. This will cause inventory does not have chance to update its DB with the current design so DB can be out of sync between inventory and tenant. When other events happen such as timer collection from inventory to sweep config to tenant, it can cause issues on tenant DB where CLI can fail.
Workaround: Avoid such cycles of operations
Recovery: Delete the BGP peer/peer-group in problem and recreate them again.
Parent Defect ID: EFA-8090 Issue ID: EFA-8090
Severity: S2 - High
Product: Extreme Fabric Automation Reported in Release: EFA 2.4.0
Symptom: When a fabric containing more than 15 newly registered devices is deployed using the CLI 'efa fabric configure', an attempt to add ports of any of these devices to a tenant within 5 minutes may fail. The error will indicate that the ports have not yet been registered in the fabric
Condition: Attempt to add device ports of a recently configured fabric to a tenant may fail with an error indication that the ports have not yet been registered in the fabric
Workaround: Wait for up to 5 minutes after deploying the fabric before adding ports to a tenant
Recovery: This is a transient error. Rerunning the port-add operation after a maximum wait time of 5 minutes will succeed
Parent Defect ID: EFA-8152 Issue ID: EFA-8152
Severity: S2 - High
Product: Extreme Fabric Automation Reported in Release: EFA 2.4.0
Symptom: While graceful-restart(GR) updating with value TRUE and inflight transition triggered as a part of EFA rollover then update will continue as a part of inflight transition.
Condition: Update GR with value TRUE and perform EFA rollover on HA setup.
Parent Defect ID: EFA-8155 Issue ID: EFA-8155
Severity: S2 - High
Product: Extreme Fabric Automation Reported in Release: EFA 2.4.0
Symptom: "cluster-client auto" is not configured under port channel for first reloaded device.
Condition:

Execute below steps to hit this condition

1) Create fabric on MCT paired device

2) Create Tenant/PO/VRF/EPG

3) Enable MM mode on both device

4) Perform EFA backup

5) Delete EPG/VRF/PO/Tenant

6) Delete fabric

7) Restore EFA backup

8) Reload device one by one

After these steps check PO on both device, "cluster-client auto" will not configured on first reloaded device.

Workaround: Instead of reload device in Step (8), perform manual DRC using inventory CLI as "efa inventory drift-reconcile execute --ip <device ip> --reconcile" for each device.
Parent Defect ID: EFA-8257 Issue ID: EFA-8257
Severity: S2 - High
Product: Extreme Fabric Automation Reported in Release: EFA 2.4.0
Symptom: EFA is not able to detect drift for configuration like VRF/VE/VLAN/EVPN
Condition:

Please follow below steps,

1) Create tenant/VRF/PO/EPG

2) As soon as EPG creation pushed configuration on device, remove them from device.

3) Check drift using inventory CLI as "efa inventory drift-reconcile execute --ip --device-ip <device ip>"

Workaround: As this is timing issue so we need to wait for 1 min before remove configurations from device.
Recovery: We need to delete EPG and recreate it again.
Parent Defect ID: EFA-8269 Issue ID: EFA-8269
Severity: S3 - Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.4.0
Symptom: EPG app-state moved to cfg-refresh-err after epg delete and admin up
Condition:

1) Configure clos fabric (Medium scale fabric)

2) Create tenant

3) Admin down the devices

4) Create port-channels, vrfs and epgs

5) Admin up the following devices

Wait for the DRC to be success

6) Repeat step 3

Wait for the devices to put into maintenance mode

7) Create bgp peer-group and dynamic peers

8) Delete all epg's

9) Repeat step 5

10) Vrfs are getting deleted from admin up devices

11) EPG app-state move to cfg-refresh-err

Recovery: Delete the EPGs in cfg-refresh-err state and recreate them.
Parent Defect ID: EFA-8273 Issue ID: EFA-8273
Severity: S3 - Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.4.0
Symptom: EPG Update "vrf-add" operation gives success when EPG is in "vrf-delete-pending" state
Condition: Perform EPG Update "vrf-add" operation on an EPG in "vrf-delete-pending" state
Workaround: No workaround
Recovery: User needs to remove the VRF from EPG using EPG update "vrf-delete" operation before attempting the "vrf-add" operation.
Parent Defect ID: EFA-8315 Issue ID: EFA-8315
Severity: S2 - High
Product: Extreme Fabric Automation Reported in Release: EFA 2.4.0
Symptom: User adds ports in empty EPG and immediately deletes them. The following adding ports into EPG can have error as duplicate entry
Condition:

1) Add ports in empty EPG

2) Delete ports from epg right away

3) Add ports into EPG. Which can have error.

Workaround: After adding ports into EPG, wait certain time before trying to delete ports from EPG.
Recovery: Delete the EPG and recreate again.
Parent Defect ID: EFA-8322 Issue ID: EFA-8322
Severity: S3 - Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.4.0
Symptom: EPG Update "anycast-ip-delete" operation gives different output/result when one of the EPG device is admin down
Condition:

1) Create L3 EPG with anycast-ip/anycast-ipv6

2) Take one of EPG device administratively down

3) Bring device admin up which was taken down in previous step

4) While device is coming up administratively, try EPG Update "anycast-ip-delete" operation

Workaround: No workaround
Recovery: No recovery as such. Wait for device to be completely up before trying EPG Update "anycast-ip-delete" operation
Parent Defect ID: EFA-8334 Issue ID: EFA-8334
Severity: S3 - Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: system backup and restore causes epg state to be in cfg-refresh-err
Condition: Tenant DB and inventory DB needs time to be in sync. In a busy and scaled system this in sync can take much longer time to finish. Backup DB during DB un-sync window can cause system saves the DBs for tenant and inventory which are not synced yet and following restore will have issues.
Workaround: If there's a need to make system backup, please execute the backup after system have not made any new config for few minutes. It's needed for the inventory and tenant databases to be in sync before executing system backup. In a busy system the DB sync can take longer to finish.
Recovery: Delete the EPGs which report errors and recreate them.
Parent Defect ID: EFA-8335 Issue ID: EFA-8335
Severity: S2 - High
Product: Extreme Fabric Automation Reported in Release: EFA 2.4.0
Symptom: system backup and restore causes following manual DRC has errors
Condition: Tenant DB and inventory DB needs time to be in sync. In a busy and scaled system this in sync can take much longer time to finish. Backup DB during this window will cause system saves the DB for tenant and inventory which is not synced yet and following restore will have issues.
Workaround: If there's a need to make system backup, please execute the backup after system have not made any new config for few minutes. It's needed for the inventory and tenant databases to be in sync before executing system backup. In a busy system the DB sync can take longer to finish.
Recovery: Delete the epg or tenant with problem and recreate them.
Parent Defect ID: EFA-8443 Issue ID: EFA-8443
Severity: S3 - Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.4.0
Symptom: For Tenant created with L3 port having multiple ip-address associated with it, "efa tenant show" will have repeated entries of that L3 port.
Condition:

Steps to reproduce issue:

1) Assign multiple IPs to the physical port on SLX.

2) Create Tenant using same L3 port.

3) Check Tenant show output.

L3 ports having multiple IPs will have repeated entry in the "efa tenant show" output.

Workaround: No workaround.
Recovery: Recovery can be done by removing all but one IP from the L3 port on SLX followed by an inventory device update.
Parent Defect ID: EFA-8465 Issue ID: EFA-8465
Severity: S2 - High
Product: Extreme Fabric Automation Reported in Release: EFA 2.4.0
Symptom: The "efa inventory device firmware-download prepare add" command fails with "Please specify 'fullinstall' option in firmware download cmdline as GLIBC versions change".
Condition: Upgrading the SLX firmware from 20.1.2x to 20.2.x requires a 'fullinstall' firmware download in order to proceed.
Workaround: There is no workaround from EFA. The firmware download fullinstall must be carried out individually on each SLX device.
Parent Defect ID: EFA-8507 Issue ID: EFA-8507
Severity: S2 - High
Product: Extreme Fabric Automation Reported in Release: EFA 2.4.0
Symptom: Certain vlans are missed in configuration when stacks are created in quick succession within a script with no delay.
Condition: 10 stack creations without much delay leads to missing configuration. Trunk Subport update is not generated from neutron. Issue is seen with only one controller and not seen when more delay is introduced between stack creations. Trunk also remains in DOWN state.
Workaround: Workaround is to have delay between stack creation.
Recovery:

Remove the Trunk Parent port added to the VM and add it back again.

e.g.

Max-L2-ss3VirtIoVM2_Test1==> VM Name

Max-L2-ss3VirtIoTrunkPort2_Test1 ==> Parent Port of the Sub Port that is down

openstack server remove port Max-L2-ss3VirtIoVM2_Test1 Max-L2-ss3VirtIoTrunkPort2_Test1

openstack server add port Max-L2-ss3VirtIoVM2_Test1 Max-L2-ss3VirtIoTrunkPort2_Test1

Parent Defect ID: EFA-8512 Issue ID: EFA-8512
Severity: S3 - Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.4.0
Symptom: On SLX there can have partial config of neighbor under router bgp. The "show running command router bgp" from SLX shows invalid command "neighbor pg1" (Assume the bgp-group name is pg1). There's no corresponding command to delete this.
Condition: It's found if issue netconf RPC to SLX device with BGP peer group delete operation which the peer-group does not exist, SLX will create the invalid "neighbor pg1".
Workaround: Under some admin-down device scenario, avoid delete the same bgp-peer more than once.
Recovery:

On SLX use the following commands to get rid of the partial bgp-peer.

SLX(config)# router bgp

SLX(config-bgp-router)# neighbor pg1 peer-group

SLX(config-bgp-router)# no neighbor pg1 peer-group

Parent Defect ID: EFA-8517 Issue ID: EFA-8517
Severity: S3 – Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: Config deployment on EFA takes longer time after stack create complete
Condition:
Parent Defect ID: EFA-8526 Issue ID: EFA-8526
Severity: S3 - Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.4.0
Symptom: VRF Update "centralized-router-add" fail with error "[x, y] are MCT pair. Update the VRF with both devices together as centralized routers"
Condition:

1) In a CLOS fabric setup with MCT pair of border-leafs, create VRF with routing-type as centralized and select MCT pair of border-leafs as centralized routers.

2) Remove one of the MCT pair border-leaf from the fabric

3) Add same/different border-leaf to the fabric and run fabric configure command

4) Wait for sometime and run VRF Update "centralized-router-add" operation to add newly added border-leaf as centralized router

Workaround: Run VRF Update "centralized-router-add" operation and specify both nodes of MCT pair border-leafs as centralized routers.
Recovery: No recovery is required.
Parent Defect ID: EFA-8573 Issue ID: EFA-8573
Severity: S3 - Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.4.0
Symptom: In few cases, networks in EPG will remain in cfg-in-sync state even if they are created with partial success topology (MCT pair with one admin-up device and one admin-down device).
Condition:

The issue is seen with the below steps

1) Configure a fabric

2) Create Tenant

3) Create multi-homed portchannel

4) Bring one of the devices of the MCT pair(having the PO created in step 3) admin-down to create a partial success topology

5) Create EPGs on the partial success topology

Recovery: Bring all the devices in admin-up state. It should push all the configs on devices and everything will be in cfg-in-sync.
Parent Defect ID: EFA-8628 Issue ID: EFA-8628
Severity: S3 - Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.4.0
Symptom: Tenant does not contain any Ports, state of EPG is "vrf-delete-pending" and EPG Update "vrf-delete" fail with error- "EPG cannot be updated with tenant having no ports"
Condition:

Step1) Create Fabric and Tenant

Step2) Create VRF using Routing type as Centralized and Border-leaf devices as Centralized Routers

Step3) Create EPG using VRF created in Step2 with Physical Ports/Portchannel from leaf devices

Step4) Delete all the devices from Fabric and/or Inventory

Step5) Perform EPG Update "vrf-delete" operation

Workaround: No workaround is required.
Recovery: Delete and re-create EPG(s).
Parent Defect ID: EFA-8665 Issue ID: EFA-8665
Severity: S3 - Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.4.1
Symptom: VRF configuration present on Border Leaf devices
Condition:

1) Create Fabric and Tenant

2) Create VRF using Routing-type as "Centralized" and Border-leaf devices as Centralized-Routers

3) Create EPG(s) using VRF created in Step2 and Physical Ports/Portchannel from Leaf devices

4) Remove Leaf devices (having the Physical Ports/Portchannel used in the EPG) from Fabric/Inventory

Workaround: No workaround.
Recovery: Manually clean up VRF configuration from device
Parent Defect ID: EFA-8669 Issue ID: EFA-8669
Severity: S2 - High
Product: Extreme Fabric Automation Reported in Release: EFA 2.4.0
Symptom: EFA is not reachable after secondary node down/up of 8720. (kube-system pods are not responding)
Condition: This issue occurs only when the management cables are unplugged for under 10 seconds and plugged back in. What happens is that the EFA failover keepalived scripts, which does labeling and then switches the pod over to the other node, doesn't finish to completion causing the node labeling to not reflect right status where the VIP is assigned.
Workaround: Either keep the management ports unplugged for more than 20 seconds or reboot the tpvm.
Recovery: Reboot the tpvm.
Parent Defect ID: EFA-8701 Issue ID: EFA-8701
Severity: S2 - High
Product: Extreme Fabric Automation Reported in Release: EFA 2.4.1
Symptom: EFA becomes unresponsive and the OS reports that there is no disk space.
Condition: Galera is technology used by mysql to keep data in sync between nodes. Occasionally, such a transfer between nodes can fail, resulting in a log file of the failed transaction, and some binary data. These accumulate in /apps/efadata/mysql. EFA will tar.gz these logs hourly and delete; but if there is a rapid accumulation, this process can be slow, and under extreme circumstances, not complete before the next hour starts. Eventually, the system is unable to proceed.
Workaround: Delete files matching GRA_*.log in /apps/efadata/mysql. These are not needed by EFA.
Recovery: Delete files matching GRA_*.log in /apps/efadata/mysql. These are not needed by EFA. Kill all tar czvf processes working on GRA*.log files.
Parent Defect ID: EFA-8773 Issue ID: EFA-8773
Severity: S2 - High
Product: Extreme Fabric Automation Reported in Release: EFA 2.4.1
Symptom: When the address family config is manually removed from the SLX followed by DRC, the DRC fails with the error - "Error: VRF Address Family not configured".
Condition:

Steps to reproduce:

1. Configure a L3 EPG via EFA which results in VRF configuration on the SLX.

2. Remove the address-family configuration from SLX ( this would remove bgp address-family configuration also).

3. Initiate the DRC to push back the configuration from EFA.

4. DRC fails with error "Error: VRF Address Family not configured".

Workaround: No workaround.
Recovery: Configure the address family on the switch manually and then trigger DRC. The configuration will be reconciled.
Parent Defect ID: EFA-8802 Issue ID: EFA-8802
Severity: S3 - Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.4.1
Symptom: Same epg will be allowed to update with different port properties value, however the first configured value will take effect
Condition:

1. Create L3 epg e1 with port-group P1 port-group property switch port mode trunk

2. Update epg e1 by port-group-add P2 and port-group property switch port mode as trunk

3. Update same epg e1 by port-group-add P2 and different port-group property switch port mode trunk-no-default-native

Above steps are applicable for all port properties like switch port mode , switchport-native-vlan-tagging, switchport-native-vlan, single-homed-bfd-session-type. No error is seen while updating with different port property

Workaround: Updating the epg with similar port property values since idempotency doesn't work
Parent Defect ID: EFA-8827 Issue ID: EFA-8827
Severity: S3 - Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.4.1
Symptom: CEP port will remain in "cfg-refreshed" state even after the DRC is successful.
Condition:

1. Create EPG with CEP ports.

2. Configure "MM enable on reboot" on the SLX and modify any of the CEP port properties.

3. Reboot the SLX.

4. MM triggered DRC will reconcile all the port properties but the CEP will remain in cfg-refreshed state.

Recovery: CEP in cfg-refreshed state can be removed from the EPG and readded to the EPG.
Parent Defect ID: EFA-8848 Issue ID: EFA-8848
Severity: S3 - Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.4.1
Symptom: When a configuration drift is introduced on a physical port on the SLX device, followed by a DRC on the device, the reconcile status displays "Portchannel" and "Vlan" reconciliation status as success even though the same were not drifted.
Condition:

Below are the steps to reproduce the issue:

1) Create Fabric/Tenant using MCT paired SLX devices.

2) Create an EPG using the physical port (P1) from an MCT SLX device (D1).

3) Introduce drift on port (P1) on the SLX device(D1).

4) Perform DRC of the SLX device (D1).

5) DRC executed successfully and all configuration pushed on the SLX device.

6) DRC output displays "Portchannel" and "Vlan" reconcilation status as success even though the same were not drifted.

Parent Defect ID: EFA-8966 Issue ID: EFA-8966
Severity: S4 - Low
Product: Extreme Fabric Automation Reported in Release: EFA 2.4.0
Symptom: BGP peer group deletion fails when the deletion is attempted using a tenant not owning the BGP peer group.
Condition: Deletion of the BGP peer group which is not owned by an existing tenant.
Workaround: No workaround
Parent Defect ID: EFA-9009 Issue ID: EFA-9009
Severity: S2 - High
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: Creation or update of an EPG involving a port group comprising more than 50 ports and 100 or more VLANs may take more than 6 minutes and may fail
Condition: When a user sends a Tenant EPG command to create or update an EPG involving a port group comprising more than 50 ports and 100 or more VLANs, the command may fail with the error "Error : EPG: <epg-name> Save for devices failed"
Workaround: Split such EPG create or update request to add not more than 50 VLANs at a time.
Recovery: None.
Parent Defect ID: EFA-9065 Issue ID: EFA-9065
Severity: S2 – High
Product: Extreme Fabric Automation Reported in Release: EFA 2.4.3
Symptom: EFA Port Channel remains in cfg-refreshed state when the port-channel is created immediately followed by the EPG create using that port-channel
Condition:

Below are the steps to reproduce the issue:

1. Create port-channel po1 under the ownership of tenant1

2. Create endpoint group with po1 under the ownership of tenant1

3. After step 2 begins and before step 2 completes, the raslog event w.r.t. step 1 i.e. port-channel creation is received. This Ralsog event is processed after step 2 is completed

Parent Defect ID: EFA-9346 Issue ID: EFA-9346
Severity: S3 – Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: When the MCT Port channel membership is changed via config changes on SLX (out of band), Fabric service will not mark the device to be in cfg-refreshed state
Condition: This gives the user an incorrect impression that the status of the device is cfg-in-sync.
Parent Defect ID: EFA-9400 Issue ID: EFA-9400
Severity: S3 - Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.4.4
Symptom: When the EPG is in "port-group-delete-pending" state, the subsequent "port-group-delete" operations will not clean up any configurations from the admin up devices.
Condition:

The issue can be reproduced with the below steps:

1. Create an EPG with CEEP port channels pg1, pg2.

2. Admin down one of the MCT nodes.

3. Deletion of pg1 from the EPG (created in step 1) will clean up all configs from the admin up device and the EPG moves to the "port-group-delete-pending" state.

4. Deletion of pg2 from the EPG will not clean any of the configurations from admin up device since the EPG is in the "port-group-delete-pending" state.

Workaround:

Perform a single port-group-delete operation with all the port-groups belonging to the EPG, instead of multiple port-group-delete

operations with each iteration containing some of the port-groups belonging to the EPG

Recovery: Bring all the devices in admin-up state and then perform the port-group-delete operation on the EPG which will clean all configs on both the nodes
Parent Defect ID: EFA-9443 Issue ID: EFA-9443
Severity: S2 - High
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: resilient-hash max-path is seen on the vrf even though it was deleted from the VRF (when the device was down) using VRF update rh-max-path-delete operation followed by the admin up of the device.
Condition:

Below are the steps to reproduce the issue:

1. Configure Tenant and VRF with rh-max-path.

2. Create EPG using VRF created in step1.

3. Admin-down one of the MCT devices.

4. Execute VRF Update rh-max-path-delete operation.

5. Bring the admin down device administratively up.

Workaround:

1. Execute VRF Update rh-max-path-add to add resilient-hash max-path again to the VRF.

2. Execute VRF Update rh-max-path-delete to remove resilient-hash max-path from the VRF.

Recovery: No recovery is required
Parent Defect ID: EFA-9451 Issue ID: EFA-9451
Severity: S2 - High
Product: Extreme Fabric Automation Reported in Release: EFA 2.4.3
Symptom: In a brownfield deployment with the existing/stale MCT cluster-client configuration imported into EFA, the EPG creation fails when the MCT cluster-client ID conflicts with the MCT cluster-client ID already consumed by an out-of-band created MCT cluster-client.
Condition:

Below are the probable steps to reproduce the issue:

1. Dual-homed (MCT) SLX has a brownfield MCT cluster-client configuration which is imported into EFA DB during "efa fabric configure" on the SLX.

2. Create tenant on the fabric configured in step1.

3. Create multi-homed port channels (which are already configured as MCT cluster-clients on the SLX) under the ownership of the tenant created in step 2.

4. Create EPG using the multi-homed port-channels and the same fails with the error

"Port :<slx-mgmt-ip>:port-channel:<po-id> ClientID : <1000+po-id> already configured, conflicting with tenant-service generated ID <po-id>".

Workaround: No workaround
Recovery:

Below are the recovery steps:

1. Delete the "cluster" configuration from one of the MCT SLX nodes and perform "inventory device update" for the SLX.

2. Delete the "cluster" configuration from the other MCT SLX node and perform "inventory device update" for the SLX.

3. Execute DRC for both the MCT SLX nodes.

Parent Defect ID: EFA-9467 Issue ID: EFA-9467
Severity: S2 - High
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: When a BGP peer-group create with an invalid MD5 password is attempted on an SLX device via EFA, the creation will fail with a valid error from SLX, resulting in the rollback of the failed operation which further results in a stale/partial configuration "neighbor <peer-group-name>" on the SLX.
Condition:

Below are the steps to reproduce the issue:

1. Configure a CLOS/Non-CLOS fabric.

2. Create a tenant on the fabric configured in step 1.

3. Create a BGP peer-group (under the ownership of the tenant created in step 2) with an invalid md5 password, which fails with a valid error from SLX device.

4. Failure of (3) results in the "rollback" operation resulting in the removal of peer-group configuration from the SLX device, which further results in a stale/partial SLX configuration "neighbor <peer-group-name>".

Workaround: No workaround.
Recovery:

On the SLX, use the following commands to remove the partial bgp peer-group configuration.

SLX(config)# router bgp

SLX(config-bgp-router)# neighbor pg1 peer-group

SLX(config-bgp-router)# no neighbor pg1 peer-group

Parent Defect ID: EFA-9487 Issue ID: EFA-9487
Severity: S2 - High
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: When bgp static peers are created (on a partial success topology) with the target device being admin down device followed by admin up of the admin down device, the dev-state/app-state continue to be not-provisioned/cfg-refreshed instead of provisioned/cfg-in-sync.
Condition:

This issue can be reproduced with the below steps.

1. Admin down one of the SLX devices of an MCT pair.

2. Create BGP static peers with md5-password with the admin down SLX device as the target device.

3. Admin up the SLX device (which was admin down as mentioned in step 1).

4. After admin up, the BGP static peers' dev-state/app-state moves to not-provisioned/cfg-refreshed instead of it being provisioned/cfg-in-sync.

Workaround: Create the bgp static peers with md5-password when all the devices are in admin-up state.
Recovery:

1. Bring all the devices to the admin up state.

2. Delete the BGP static peers which were created with md5-password when one of the devices was admin down.

3. Recreate the BGP static peers which were created with md5-password when one of the devices was admin down.

Parent Defect ID: EFA-9932 Issue ID: EFA-9932
Severity: S2 – High
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: System API throws a 502 error after a restore is run.
Condition: After executing 'efa system restore', triggering any system API sometimes yield a 502 response.
Parent Defect ID: EFA-9941 Issue ID: EFA-9941
Severity: S3 – Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.4.3
Symptom: EPG create with a CEP port (already used in another EPG) fails with the NETCONF RPC Error "NOTAKNOWNRESOURCEID"
Condition:

1. Create fabric and tenant.

2. Create EPG1 using CEP port.

3. Create EPG2 using the same CEP port.

Parent Defect ID: EFA-9944 Issue ID: EFA-9944
Severity: S3 – Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: If port channel speed is modified on the device. In the mean time member ports are deleted from PO. Issue DRC from EFA, the DRC will fail with error and PO speed is not get reconciled: "10.20.246.3:ethernet:0/19 to Portchannel po2 failed due to netconf rpc [error] %Error: Port-channel should be admin down for speed to be configured"
Condition:

Steps to reproduce:

1. Create tenant and PO

2. Modify po speed on SLX

config t

no int po 112

interface Port-channel 112

speed 100

no shutdown

3. Update inventory

4. Manual DRC

Parent Defect ID: EFA-9945 Issue ID: EFA-9945
Severity: S3 – Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: Make config changes of channel group mode on SLX for the interface belongs to some PO. Issue DRC from EFA. Drift and reconcile will not detect this change and correct it.
Condition:

Steps to reproduce:

Create tenant and PO

Modify channel group mode on SLX

config t

interface Ethernet 0/17-18

no channel-group

channel-group 111 mode passive type standard

Update inventory

Manual DRC

Parent Defect ID: EFA-9968 Issue ID: EFA-9968
Severity: S2 – High
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: DRC fail with error "Delete and Update operations are not supported in a singled transaction. Please try them individually."
Condition:

Below are the steps to reproduce:

1. Create Tenant

2. Create VRF with max-path and graceful-restart-enable

3. Create EPG using VRF created in step 1

4. Take one of the SLX devices to administratively down state

5. Perform VRF Update max-path-add operation to add a different max-path value

6. Perform VRF Update graceful-restart-update to disable graceful-restart

7. Admin up the SLX device which was made administratively down in step 4 and wait for DRC to complete

Parent Defect ID: EFA-9974 Issue ID: EFA-9974
Severity: S3 – Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: [SQA] DRC after firmware download failed with reason 'drift and reconcile failed due to efa failover'
Condition:
Parent Defect ID: EFA-9988 Issue ID: EFA-9988
Severity: S2 – High
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: If the SLX device is in firmware-download-in-progress state, then the DRC (drift and reconcile) fails, But the failure reason is not shown in the DRC output
Condition: Trigger drift and reconcile from EFA for the SLX device when the SLX device is in the firmware-download-in-progress-state
Parent Defect ID: EFA-10016 Issue ID: EFA-10016
Severity: S3 – Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom:

BGP Peer creation fails with the error "failed to fetch BgpAddressFamily data" because of the intermittent connectivity loss of EFA with SLX.

Rollback also failed leaving the stale config on SLX.

Condition:

1. Create tenant, po, vrf and epg

2. Create bgp peer group

3. Create bgp peers

Create fails because of an intermittent connection issue

Parent Defect ID: EFA-10018 Issue ID: EFA-10018
Severity: S3 – Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: Deployment may fail in the step "Checking default gateway reachability on all nodes" due to network reachability issue with gateway
Condition: Network reachability issue with gateway will cause deployment failure
Parent Defect ID: EFA-10022 Issue ID: EFA-10022
Severity: S3 – Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: questions on EFA for static vs dynamic LAG delay
Condition:
Parent Defect ID: EFA-10041 Issue ID: EFA-10041
Severity: S2 – High
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: The "efa inventory device tpvm-upgrade execute" command to upgrade to a new tpvm version will result in a failed tpvm upgrade where the previous tpvm image will be rolled back and restored.
Condition: The "trusted peer" TPVM running-configuration has not been applied to either SLX device hosting the TPVMs installed with an EFA multi-node deployment.
Parent Defect ID: EFA-10052 Issue ID: EFA-10052
Severity: S3 – Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: After upgrade EFA from pre-2.5.0 to 2.5.0, the "Fabric Status" in "efa fabric show" is shown as "configure-success".
Condition: There is a new field "Fabric Status" introduced in 2.5.0. After upgrade from pre-2.5.0 to 2.5.0, the "Fabric Status" in "efa fabric show" is showing as "configure-success".
Parent Defect ID: EFA-10064 Issue ID: EFA-10064
Severity: S2 – High
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: During the required fullinstall firmware download of SLXOS from 20.2.3f to 20.3.2a the TPVM configuration through exec-mode commands are not converted to the running-config.
Condition: When a firmware download is run using the fullinstall option.
Parent Defect ID: EFA-10067 Issue ID: EFA-10067
Severity: S2 – High
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: In a node replacement scenario, the standby node will not have a TPVM configured nor running. When the "efa inventory device tpvm-upgrade" command is run against this replacement node, the TPVM deployment and upgrade to the new TPVM version fails.
Condition: The TPVM is neither configured nor running on the switch.
Parent Defect ID: EFA-10071 Issue ID: EFA-10071
Severity: S3 – Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: [EFA] - High CPU on logrotate - making EFA slow
Condition:
Parent Defect ID: EFA-10073 Issue ID: EFA-10073
Severity: S3 – Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: Getting Error : Port :172.16.18.3:port-channel:1 ClientID : 1001 already configured, conflicting with tenant-service generated ID 1
Condition:
Parent Defect ID: EFA-10094 Issue ID: EFA-10094
Severity: S3 - Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: In a fabric-wide firmware download, other firmware download operations such as firmware commit and restore error out stating that a firmware download execution is in progress. However, after the active EFA node switch is upgraded and rebooted, this check is no longer enforced and other firmware download operations are allowed even though the fabric-wide firmware download operation is still in progress.
Condition: A fabric-wide firmware download execution where the active EFA node switch is included.
Parent Defect ID: EFA-10099 Issue ID: EFA-10099
Severity: S2 – High
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom:

When the md5-password is updated on an already provisioned fabric, the existing tenant vrf backup routing bgp neighbours will be updated with the new md5-password followed by the clearing of the bgp neighbours.

Updation of md5-password and clearing of the corresponding bgp neighbours happens one SLX device at a time, hence resulting in the session being down time till the process is complete for both the devices of the MCT pair.

Condition:

1. Configure fabric with the fabric setting backup routing enabled and with the md5-password fabric setting

2. Configure tenant under the fabric

3. Configure VRF and L3 EPG (using the VRF and under the ownership of the tenant), which results in the creation of the backup routing bgp neighbours (for the tenant vrf) using the md5-password provided at the fabric setting

4. Update md5-password on the already provisioned fabric followed by "fabric configure"

Parent Defect ID: EFA-10115 Issue ID: EFA-10115
Severity: S3 - Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: Redeployment of EFA 2.5.0 fails on a multi-node when one of the nodes in the cluster is changed.
Condition: After un-deployment of multi-node EFA, if a fresh deployment is triggered after changing one of the nodes or its IP in the cluster, then installation fails.
Parent Defect ID: EFA-10121 Issue ID: EFA-10121
Severity: S2 – High
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: TPVM and Apps/EFA running in TPVM, will be removed after copy default-config startup-config" and "reload system" commands on SLX.
Condition: If copy default-config startup-config is run on the SLX device that is hosting a TPVM, the TPVM and applications/EFA running within, will be removed after "reload system"
Parent Defect ID: EFA-10126 Issue ID: EFA-10126
Severity: S3 - Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: Users might see a failure message when doing a node replacement when running the installer in GUI mode, when the progress is around 18% although the node replacement proceeds
Condition: This error message is not expected and is not consistently seen, however this issue is not a functional issue but more of a display issue
Parent Defect ID: EFA-10135 Issue ID: EFA-10135
Severity: S3 – Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: Efa system - manual setting of backup schedule not taken into use
Condition:
Parent Defect ID: EFA-10137 Issue ID: EFA-10137
Severity: S3 – Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: Efa fabric device add-bulk failed with Validate Fabric [Failed] Missing Links
Condition:
Parent Defect ID: EFA-10139 Issue ID: EFA-10139
Severity: S3 – Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: efa fabric device add-bulk failed with Server unreachable during leaf-pair expansion
Condition:
Parent Defect ID: EFA-10141 Issue ID: EFA-10141
Severity: S3 – Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: [EFA] - HA Deployment 2.5.0 - Configuring database server Failed
Condition:
Parent Defect ID: EFA-10142 Issue ID: EFA-10142
Severity: S3 – Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: Journals are kept under pending status even though dynamic marking of failure when the parent journal fails
Condition:
Parent Defect ID: EFA-10143 Issue ID: EFA-10143
Severity: S3 – Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: Modifications needed in the current efa-journal Table display entries
Condition:
Parent Defect ID: EFA-10206 Issue ID: EFA-10206
Severity: S3 – Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: [EFA 2.4 -> 2.5] KeepAlived log files changed to check reachability of Peer or GW is not updated any more
Condition:
Parent Defect ID: EFA-10285 Issue ID: EFA-10285
Severity: S3 – Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: [EFA] Slow upgrade of SLX OS on EFA - 6 hours for 18 switches
Condition:
Parent Defect ID: EFA-10289 Issue ID: EFA-10289
Severity: S3 – Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: EFA performs "copy running-config startup-config" after firmware download is started ( just before reload) which causes config loss during SLX full install scenario
Condition:
Parent Defect ID: EFA-10304 Issue ID: EFA-10304
Severity: S3 – Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: NTP configuration present on SLX is not considered by EFA for DRC
Condition:
Parent Defect ID: EFA-10358 Issue ID: EFA-10358
Severity: S3 - Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: SLX upgrade to 20.3.2A - traffic outage + Maintenance Mode Disable Failed
Condition: