Open Defects

The following defects are open in Extreme Fabric Automation 2.5.0.

Parent Defect ID: EFA-10126 Issue ID: EFA-10126
Severity: S3 - Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: Users might see a failure message when doing a node replacement when running the installer in GUI mode, when the progress is around 18% although the node replacement proceeds
Condition: This error message is not expected and is not consistently seen, however this issue is not a functional issue but more of a display issue
Workaround: Since the installation and node replacement proceeds fine and succeeds, users can just ignore the briefly seen failed message on the installer. This will be fixed in an upcoming release.
Parent Defect ID: EFA-10121 Issue ID: EFA-10121
Severity: S2 - High
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: TPVM and Apps/EFA running in TPVM, will be removed after copy default-config startup-config" and "reload system" commands on SLX.
Condition: If copy default-config startup-config is run on the SLX device that is hosting a TPVM, the TPVM and applications/EFA running within, will be removed after "reload system"
Workaround: Avoid using "copy default-config startup-config"
Recovery:

This recovery procedure assumes multi-node EFA deployment on SLX1/TPVM1 and SLX2/TPVM2..

TPVM1 is the TPVM that was removed on SLX1, and SLX2/TPVM2 is the second device and now running active EFA.

On SLX1 perform the following steps.

1. Reinstall TPVM1 using config mode ​with the same configuration as before.

2. On SLX2/TPVM2, setup trusted peer with TPVM1. "SLX2(config-tpvm-TPVM)# trusted-peer ip TPVM1 password <password>"

3. On TPVM2, perform efa upgrade with node replacement using the following steps.

SLX2# efa deploy --graphics no

Starting "efa deploy"...

Step 1: Checking if TPVM is deployed...

Step 2: Get IP Address assigned to TPVM to deploy EFA

IP Address of the TPVM TPVM2

Step 3: Checking for EFA packages in /efaboot directory

Step 4: Deploying EFA package efa-2.5.0.tar.gz on TPVM2

Step 5: Checking for EFA Stack...

Previous Stack found

Are you sure you want to re-deploy EFA? (yes/no)

yes

Step 6: Re-deploying EFA

Would you like to configure additional IP address(es) for the HA health ping check? (yes/no)

no

Would you like to configure additional management IP networks? (yes/no)

no

Would you like to configure additional management IP network routes? (yes/no)

no

Please choose: 1 Multi Node Build Upgrade 2 Multi Node Build Upgrade With Node Replacement

2

Enter the replacement peer node (IP/Hostname):

TPVM1

TPVM1-IP server is reachable...

You have chosen:

- to redeploy EFA at version 2.5.0 build GA

- with peer TPVM1 and VIP EFA-VIP

- with node replacement

Do you wish to proceed? (yes/no)

yes

...

...

Waiting for EFA services

Waiting for EFA containers to start

Extreme Fabric Automation Stack has been upgraded successfully.

SLX2#

Once efa upgrade with node replacement is complete, verify on both nodes, efa is running and services have started using efactl status.

Parent Defect ID: EFA-10115 Issue ID: EFA-10115
Severity: S3 - Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: Redeployment of EFA 2.5.0 fails on a multi-node when one of the nodes in the cluster is changed.
Condition: After un-deployment of multi-node EFA, if a fresh deployment is triggered after changing one of the nodes or its IP in the cluster, then installation fails.
Workaround: Edit /etc/fstab file and remove any mount point entries of old IPs from both the nodes in the multi-node setup.
Parent Defect ID: EFA-10110 Issue ID: EFA-10110
Severity: S2 - High
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: EFA fabric and tenant operations are not blocked when (manual) DRC operation is triggered and in-progress.
Condition: DRC operations may fail/timeout with fabric/tenant operations taking longer time to complete.
Workaround: Do not run fabric configure or tenant operations when manual DRC for a device is in progress.
Parent Defect ID: EFA-10109 Issue ID: EFA-10109
Severity: S3 - Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: Global BFD timer values will be applied to multi-hop BFD sessions in multi-rack setups.
Condition: This behavior is seen only for multi-rack, multi-hop BFD sessions.
Workaround: Set required bfd timer values under appropriate peer-group using SLX CLI
Recovery: Set required bfd timer values under appropriate peer-group using SLX CLI
Parent Defect ID: EFA-10099 Issue ID: EFA-10099
Severity: S2 - High
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom:

When the md5-password is updated on an already provisioned fabric, the existing tenant vrf backup routing bgp neighbours will be updated with the new md5-password followed by the clearing of the bgp neighbours.

Updation of md5-password and clearing of the corresponding bgp neighbors happens one SLX device at a time, hence resulting in the session being down time till the process is complete for both the devices of the MCT pair.

Condition:

1. Configure fabric with the fabric setting backup routing enabled and with the md5-password fabric setting

2. Configure tenant under the fabric

3. Configure VRF and L3 EPG (using the VRF and under the ownership of the tenant), which results in the creation of the backup routing bgp neighbors (for the tenant vrf) using the md5-password provided at the fabric setting

4. Update md5-password on the already provisioned fabric followed by "fabric configure"

Workaround: No workaround
Parent Defect ID: EFA-10093 Issue ID: EFA-10093
Severity: S3 - Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: Deletion of the VLAN/BD based L3 EPGs in epg-delete-pending state will result in creation and then deletion of the VLAN/BD on the admin up device where the VLAN/BD was already removed
Condition:

Issue occurs with the below steps:

1. Create L3 EPG with VLAN/BD X on an MCT pair

2. Admin down one of the devices of the MCT pair

3. Delete the L3 EPG. This results in the L3 configuration removal (corresponding to the L3 EPG getting deleted) from the admin up device and no config changes happen on the admin down device and the EPG transits to epg-delete-pending state

4. Admin up the device which was made admin down in step 2

5. Delete the L3 EPG which transited to epg-delete-pending state in step 3

Recovery: Not needed
Parent Defect ID: EFA-10067 Issue ID: EFA-10067
Severity: S2 - High
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: In a node replacement scenario, the standby node will not have a TPVM configured nor running. When the "efa inventory device tpvm-upgrade" command is run against this replacement node, the TPVM deployment and upgrade to the new TPVM version fails.
Condition: The TPVM is neither configured nor running on the switch.
Parent Defect ID: EFA-10064 Issue ID: EFA-10064
Severity: S2 - High
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: During the required fullinstall firmware download of SLXOS from 20.2.3f to 20.3.2a the TPVM configuration through exec-mode commands are not converted to the running-config.
Condition: When a firmware download is run using the fullinstall option.
Parent Defect ID: EFA-10063 Issue ID: EFA-10063
Severity: S3 - Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: Deleting device from EFA Inventory would not bring up interface to admin state 'up' after unconfiguring breakout configuration
Condition: This condition occurs when there is breakout configuration present on device that is being deleted from EFA Inventory
Workaround: Manually bring the admin-state up on the interface, if required
Recovery: Manually bring the admin-state up on the interface, if required
Parent Defect ID: EFA-10062 Issue ID: EFA-10062
Severity: S3 - Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: Removing device from Inventory would not clean up breakout configuration on interfaces that are part of port channels.
Condition: This condition occurs when there is breakout configuration present on device that is being deleted from EFA Inventory, such that those breakout configurations are on interfaces that are part of port-channels
Workaround: Manually remove the breakout configuration, if required.
Recovery: Manually remove the breakout configuration, if required.
Parent Defect ID: EFA-10048 Issue ID: EFA-10048
Severity: S2 - High
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom:

EPG: epgev10 Save for devices failed

When concurrent EFA tenant EPG create or update operation is requested where the commands involve large number of vlans and/or ports, one of them could fail with the error "EPG: <epg-name> Save for devices Failed".

Condition: The failure is reported when concurrent DB write operation are done by EFA Tenant service as part of the command execution.
Workaround: This is a transient error and there is no workaround. The failing command can be executed once again and it will succeed.
Recovery: The failing command can be rerun separately and it will succeed.
Parent Defect ID: EFA-10041 Issue ID: EFA-10041
Severity: S2 - High
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: The "efa inventory device tpvm-upgrade execute" command to upgrade to a new tpvm version will result in a failed tpvm upgrade where the previous tpvm image will be rolled back and restored.
Condition: The "trusted peer" TPVM running-configuration has not been applied to either SLX device hosting the TPVMs installed with an EFA multi-node deployment.
Workaround: Ensure the "trusted peer" TPVM running-configuration is configured on one of the SLX devices.
Recovery: Configure the "trusted peer" TPVM running-configuration and rerun the TPVM upgrade command.
Parent Defect ID: EFA-10026 Issue ID: EFA-10026
Severity: S3 - Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: 'efa inventory device interface unset-fec' command will set the fec mode to 'auto-negotiation' instead of removing fec configuration.
Condition: Once fec mode is set on the interface, the configuration cannot be removed. 'efa inventory device interface unset-fec' command will set the fec mode to 'auto-negotiation' instead of removing fec configuration. This is because 'no fec mode' command is no longer supported on SLX.
Workaround: Default value for fec-mode is 'auto-negotiation' and will show up as-is in the output of 'show running-config'. Users can set a different value using 'efa inventory device interface set-fec', if required.
Recovery: Default value for fec-mode is 'auto-negotiation' and will show up as-is in the output of 'show running-config'. Users can set a different value using 'efa inventory device interface set-fec', if required.
Parent Defect ID: EFA-10018 Issue ID: EFA-10018
Severity: S3 - Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: Deployment may fail in the step "Checking default gateway reachability on all nodes" due to network reachability issue with gateway
Condition: Network reachability issue with gateway will cause deployment failure
Workaround: Prior to performing the deployment, login to the new tpvm and ping the gateway until it is reachable. Once it is, the deployment will then run fine.
Recovery: Cancel the deployment & Login to the redeployed TPVM. Ping the gateway IP until it is connected. Once it is, the deployment will run fine.
Parent Defect ID: EFA-10016 Issue ID: EFA-10016
Severity: S3 - Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom:

BGP Peer creation fails with the error "failed to fetch BgpAddressFamily data" because of the intermittent connectivity loss of EFA with SLX.

Rollback also failed leaving the stale config on SLX.

Condition:

1. Create tenant, po, vrf and epg

2. Create bgp peer group

3. Create bgp peers

Create fails because of an intermittent connection issue

Workaround:
Recovery:

1. Delete the peer using force option so all stale config is removed.

2. Recreate the bgp peer

Parent Defect ID: EFA-9990 Issue ID: EFA-9990
Severity: S2 - High
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: EPG update ctag-range-add operation with the existing ctag-range (i.e. ctag1, ctag2) and modified native vlan (ctag2) succeeds without any effect
Condition:

Below are the steps to reproduce the issue:

1. Create Endpoint group with ctag1, ctag2 and native vlan as ctag1

2. Update the Endpoint group (created in step 1) using ctag-range-add operation with the same set of ctags (i.e. ctag1, ctag2) and different native VLAN ctag2

Workaround: If user intends to modify the native vlan from ctag1 to ctag2 in an EPG, then the user will need to remove ctag1 (using ctag-range-delete) from the EPG and add ctag2 (using ctag-range-add) as native vlan to the EPG
Parent Defect ID: EFA-9988 Issue ID: EFA-9988
Severity: S2 - High
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: If the SLX device is in firmware-download-in-progress state, then the DRC (drift and reconcile) fails, But the failure reason is not shown in the DRC output
Condition: Trigger drift and reconcile from EFA for the SLX device when the SLX device is in the firmware-download-in-progress-state
Workaround: Wait for the firware-download to complete and then execute DRC
Parent Defect ID: EFA-9968 Issue ID: EFA-9968
Severity: S2 - High
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: DRC fail with error "Delete and Update operations are not supported in a singled transaction. Please try them individually."
Condition:

Below are the steps to reproduce:

1. Create Tenant

2. Create VRF with max-path and graceful-restart-enable

3. Create EPG using VRF created in step 1

4. Take one of the SLX devices to administratively down state

5. Perform VRF Update max-path-add operation to add a different max-path value

6. Perform VRF Update graceful-restart-update to disable graceful-restart

7. Admin up the SLX device which was made administratively down in step 4 and wait for DRC to complete

Workaround: No workaround
Recovery:

Below are the steps to recover:

1. Perform VRF Update graceful-restart-update to enable graceful-restart

2. Admin up the SLX device which was made administratively down and wait for DRC to complete

3. Perform VRF Update graceful-restart-update to disable graceful-restart

Parent Defect ID: EFA-9966 Issue ID: EFA-9966
Severity: S2 - High
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom:

In 2.4.x, adding the same ports/pos with different native vlan was allowed without any validation, this leads to two possible issues

1. Adding the same ports/PO to multiple epgs with different native-vlan VLAN results in the initial input native vlan being over written with the latest native vlan value onto the SLX

2. Adding the same port/po to multiple epgs, the first epg being created with native vlan and the second epg being created without native vlan, will result in removal of native vlan configured as part of the first epg

The same inconsistency gets carried over from 2.4.x to 2.5.0

Condition:

Case1:

1. Create fabric, tenant

2. Create PO1 with Device1Port1

3. Create PO2 with Device1Port2

4. Create EPG1 with PO1, PO2, CTAG1, CTAG2 with "switchport mode trunk", and native vlan as CTAG1

5. Create EPG2 with PO2, CTAG3, CTAG4 with native vlan CTAG3

6. On the SLX, PO2 initial native vlan CTAG1 gets replaced with CTAG3, while PO1 will have the initial native vlan CTAG1

Case2:

1. Create fabric, tenant.

2. Create PO1 with Device1Port1

3. Create PO2 with Device1Port2

4. Create EPG1 with PO1, PO2, CTAG1, CTAG2 with "switchport mode trunk" and native vlan as CTAG1

5. Create EPG2 with PO2 CTAG3, CTAG4 without native vlan

6. On the SLX, PO2 initial native vlan CTAG1 gets replaced with the default value, while PO1 will have the initial native vlan CTAG1

Workaround: The workaround is not available.
Recovery:

1. Identify EPGs with common PO/port with conflicting native vlan

2. Delete conflicting ctag (that needs to be configured as native vlan but isnt configured as one) using EPG update "ctag-range-delete" operation

3. Add ctag back with correct native vlan (to be configured on SLX) using EPG update ctag-range-add" operation

Parent Defect ID: EFA-9952 Issue ID: EFA-9952
Severity: S2 - High
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom:

network-property delete failed

When concurrent EFA tenant EPG delete operations are requested where the commands involve large number of vlans and/or ports, one of them could fail with the error "EPG network-property delete failed"

Condition: The failure is reported when concurrent DB write operation are done by EFA Tenant service as part of the command execution.
Workaround: This is a transient error and there is no workaround. The failing command can be executed once again and it will succeed.
Recovery: The failing command can be rerun separately and it will succeed
Parent Defect ID: EFA-9945 Issue ID: EFA-9945
Severity: S3 - Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: Make config changes of channel group mode on SLX for the interface belongs to some PO. Issue DRC from EFA. Drift and reconcile will not detect this change and correct it.
Condition:

Steps to reproduce:

Create tenant and PO

Modify channel group mode on SLX

config t

interface Ethernet 0/17-18

no channel-group

channel-group 111 mode passive type standard

Update inventory

Manual DRC

Workaround: Avoid making manual changes on device for channel group mode. This will be supported in later release.
Parent Defect ID: EFA-9944 Issue ID: EFA-9944
Severity: S3 - Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: If port channel speed is modified on the device. In the mean time member ports are deleted from PO. Issue DRC from EFA, the DRC will fail with error and PO speed is not get reconciled: "10.20.246.3:ethernet:0/19 to Portchannel po2 failed due to netconf rpc [error] %Error: Port-channel should be admin down for speed to be configured"
Condition:

Steps to reproduce:

1. Create tenant and PO

2. Modify po speed on SLX

config t

no int po 112

interface Port-channel 112

speed 100

no shutdown

3. Update inventory

4. Manual DRC

Workaround: On device avoid making PO speed change and remove member ports from PO at the same time. EFA will have support in later release.
Recovery: Delete the PO from device and issue DRC.
Parent Defect ID: EFA-9941 Issue ID: EFA-9941
Severity: S3 - Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.4.3
Symptom: EPG create with a CEP port (already used in another EPG) fails with the NETCONF RPC Error "NOTAKNOWNRESOURCEID"
Condition:

1. Create fabric and tenant.

2. Create EPG1 using CEP port.

3. Create EPG2 using the same CEP port.

Recovery: Delete the EPG using the common CEP port and recreate the EPG
Parent Defect ID: EFA-9932 Issue ID: EFA-9932
Severity: S2 - High
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: System API throws a 502 error after a restore is run.
Condition: After executing 'efa system restore', triggering any system API sometimes yield a 502 response.
Workaround: Run 'efa status' to confirm EFA is running and re-execute the commands.
Parent Defect ID: EFA-9930 Issue ID: EFA-9930
Severity: S2 - High
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: Periodic backup happens according to the system timezone.
Condition: If the nodes in HA are not configured in the same timezone, then periodic backup is scheduled according to the timezone of the active node. When a failover happens, the schedule is changed to the timezone of the new active node.
Workaround: Configure the same timezone on both the nodes in a multi-node installation
Parent Defect ID: EFA-9906 Issue ID: EFA-9906
Severity: S2 - High
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: When concurrent EFA tenant EPG create or update operation is requested where the commands involve large number of vlans and/or ports, one of them could fail with the error "EPG: <epg-name> Save for Vlan Records save Failed".
Condition: The failure is reported when concurrent DB write operation are done by EFA Tenant service as part of the command execution.
Workaround: This is a transient error and there is no workaround. The failing command can be executed once again and it will succeed.
Recovery: The failing command can be rerun separately and it will succeed.
Parent Defect ID: EFA-9874 Issue ID: EFA-9874
Severity: S3 - Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: When EPG is in the "anycast-ip-delete-pending" state and the user performs "epg configure", it will succeed without actually removing anycast-ip from SLX.
Condition:

Below are the steps to reproduce the issue:

1) Configure EPG with VRF, VLAN and anycast-ip (ipv4/ipv6) on a single rack Non-CLOS fabric.

2) Bring one of the devices to admin-down.

3) EPG Update anycast-ip-delete for anycast-ip ipv4 or ipv6. This will put EPG in "anycast-ip-delete-pending" state.

4) Bring the admin-down device to admin-up.

5) In this state, the only allowed operations on EPG are "epg configure" and EPG update "anycast-ip-delete".

6) Perform "epg configure --name <epg-name> --tenant <tenant-name>".

Workaround: No workaround.
Recovery: Perform the same anycast-ip-delete operation when both devices are admin-up.
Parent Defect ID: EFA-9813 Issue ID: EFA-9813
Severity: S2 - High
Product: Extreme Fabric Automation Reported in Release: EFA 2.4.3
Symptom: When doing RMA of device the port connections for the new device must be identical.
Condition: New device's port connections were not identical to the device being RMAed.
Workaround: When doing RMA of device the port connections for the new device must be identical.
Parent Defect ID: EFA-9799 Issue ID: EFA-9799
Severity: S3 - Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: 'efa status' response shows standby node status as 'UP' when node is still booting up
Condition: If SLX device is reloaded where EFA standby node resides, then 'efa status' command will still show the status of standby as UP.
Workaround: Retry the same command after sometime.
Parent Defect ID: EFA-9758 Issue ID: EFA-9758
Severity: S2 - High
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: When user modifies the remote-asn of BGP peer out of band, drift and reconcile is not reconciling the intended remote-asn of BGP peer configuration.
Condition: Issue will seen if the user modifies the remote ASN of BGP peer through out of band means, DRC is not reconciling the remote ASN.
Workaround: Login to the device where the remote ASN is modified and revert it back to what EFA has configured.
Recovery: Revert the remote ASN of BGP peer on the device through SLX CLI to what EFA has configured previously.
Parent Defect ID: EFA-9674 Issue ID: EFA-9674
Severity: S3 - Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.4.2
Symptom: Creation and deletion of stacks can result in failure. Network create fails as the previous network with same VLAN is not deleted.
Condition: Network is deleted and created in quick succession. Since the EFA processing takes time to delete the network at EFA, another call made for network create with same vlan id is also processed. This network create call will end in failure.
Workaround: Add delay between delete and create of stacks to allow more time for efa processing.
Recovery: Cleanup and recreate the failed network/stack at openstack
Parent Defect ID: EFA-9669 Issue ID: EFA-9669
Severity: S2 - High
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom:

Network and router creation during EFA HA fail-over, stale entries

Steps:

Started script/stack to create networks and routers

Did HA fail over when network/router creation was in progress

Few network and router creations failed.

Condition:

EFA HA Failover during stack creation can result in failed network and router entries at OpenStack.

EFA services will be down for few minutes during HA failover, resulting in the failures.

Workaround: No workaround. This is expected behavior during EFA HA failover.
Recovery: Delete and recreate the network/router entries at OpenStack after EFA HA failover is complete. Use 'efa-health show' to check EFA HA status.
Parent Defect ID: EFA-9645 Issue ID: EFA-9645
Severity: S2 - High
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: When the fabric setting is updated with this particular password "password$\n", md5 password doesn't get configured on the backup routing neighbors that was already created.
Condition:

1. Configure fabric

2. Create tenant, po, vrf and epg

3. Update fabric setting with "password$\n" and configure fabric

4. MD5 password is not configured on backup routing neighbors under BGP address family ipv4/ipv6 vrf

Workaround: Update the fabric setting with any other password combination that does not include "$\n" combination.
Recovery: Update the fabric setting with any other password combination that does not include "$\n" combination.
Parent Defect ID: EFA-9591 Issue ID: EFA-9591
Severity: S2 - High
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: When certain BGP sessions are not in ESTABLISHED state after clearing the BGP sessions as part of fabric configure, we see this issue.
Condition: This condition was seen when "efa fabric configure --name <fabric name>" was issued after modifying the MD5 password.
Workaround: Wait for BGP sessions to be ready. Check the status of BGP sessions using "efa fabric topology show underlay --name <fabric name>"
Recovery: Wait for BGP sessions to be ready. Check the status of BGP sessions using "efa fabric topology show underlay --name <fabric name>"
Parent Defect ID: EFA-9576 Issue ID: EFA-9576
Severity: S2 - High
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: Deletion of the tenant by force followed by the recreation of the tenant and POs can result in the error "Po number <id> not available on the devices".
Condition:

Below are the steps to reproduce the issue:

1. Create tenant and PO.

2. Delete the tenant using the "force" option.

3. Recreate the tenant and recreate the PO in the short time window.

Workaround: Avoid performing tenant/PO create followed by tenant delete followed by the tenant and PO recreate in the short time window.
Recovery: Execute inventory device prior to the PO creation.
Parent Defect ID: EFA-9570 Issue ID: EFA-9570
Severity: S2 - High
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: Add Device Failed because ASN used in border leaf showing conflict
Condition: If there are more than one pair of Leaf/border leaf devices then devices which are getting added first will get the first available ASN in ascending order and in subsequent addition of devices if one of device is trying to allocate the same ASN because of brownfield scenario then EFA will throw an error of conflicting ASN
Workaround:

Add the devices to fabric in following sequence

1)First add brownfield devices which have preconfigured configs

2)Add remaining devices which don't have any configs stored

Recovery:

Removing the devices and adding the devices again to fabric in following sequence

1)First add brownfield devices which have preconfigured configs

2)Add remaining devices which don't have any configs stored

Parent Defect ID: EFA-9456 Issue ID: EFA-9456
Severity: S3 - Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.4.3
Symptom: Issue is seen when the devices which are being added to fabric have IP addresses already configured on interfaces and are conflicting with what EFA assigns.
Condition: Issue will be observed if devices being added to fabric have IP addresses assigned on interfaces and these IP addresses are already reserved by EFA for other devices.
Workaround:

Delete the IP addresses on interfaces of devices having conflicting configuration so that new IP addresses can be reserved for these devices. One way to clear the device configuration is using below commands:

1. Register the device with inventory

efa inventory device register --ip <ip1, ip2> --username admin --password password

2. Issue debug clear "efa fabric debug clear-config --device <ip1, ip2>"

Recovery:

Delete the IP addresses on interfaces of devices having conflicting configuration so that new IP addresses can be reserved for these devices. One way to clear the device configuration is using below commands:

1. Register the device with inventory

efa inventory device register --ip <ip1, ip2> --username admin --password password

2. Issue debug clear "efa fabric debug clear-config --device <ip1, ip2>"

3. Add the devices to fabric

Parent Defect ID: EFA-9439 Issue ID: EFA-9439
Severity: S2 - High
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: Dev-State and App-State of EPG Networks are not-provisioned and cfg-ready
Condition:

Below are the steps to reproduce the issue:

1) Create VRF with local-asn

2) Create EPG using the VRF created in step 1

3) Take one of the SLX devices to administratively down state

4) Perform VRF Update "local-asn-add" to different local-asn than the one configured during step 1

5) Perform VRF Update "local-asn-add" to the same local-asn that is configured during step 1

6) Admin up the SLX device which was made administratively down in step 3 and wait for DRC to complete

Workaround: No workaround as such.
Recovery:

Following are the steps to recover:

1) Log in to SLX device which was made admin down and then up

2) Introduce local-asn configuration drift under "router bgp address-family ipv4 unicast" for the VRF

3) Execute DRC for the device

Parent Defect ID: EFA-9398 Issue ID: EFA-9398
Severity: S3 - Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.4.3
Symptom: Upon failed EFA upgrade, unable to revert to previous version of EFA.
Condition: Upon failed EFA upgrade, unable to revert to previous version of EFA.
Workaround: Users can select "No" and re-try the operation using `efa deploy` on SLX or `source deployment.sh` from a server based deployment.
Parent Defect ID: EFA-9346 Issue ID: EFA-9346
Severity: S3 - Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: When the MCT Port channel membership is changed via config changes on SLX (out of band), Fabric service will not mark the device to be in cfg-refreshed state
Condition: This gives the user an incorrect impression that the status of the device is cfg-in-sync.
Workaround: Users can run manual Drift/reconcile command which will identify the drift and fix it. This condition will happen only when PO membership has changed out of band.
Parent Defect ID: EFA-9288 Issue ID: EFA-9288
Severity: S2 - High
Product: Extreme Fabric Automation Reported in Release: EFA 2.4.4
Symptom: When device is removed from a non-clos fabric with multiple racks, BGP configuration is not removed.
Condition: Issue is seen ONLY in multi-rack non-CLOS fabric. When device is removed from a non-clos fabric with multiple racks, BGP configuration is not removed.
Workaround: Remove the configuration explicitly from device that is being removed from fabric by issuing "no router bgp" on SLX device.
Recovery: There is no recovery required as the device is decommissioned/removed from fabric. To remove stale BGP configuration from device user can issue "no router bgp" on the device directly.
Parent Defect ID: EFA-9065 Issue ID: EFA-9065
Severity: S2 - High
Product: Extreme Fabric Automation Reported in Release: EFA 2.4.3
Symptom: EFA Port Channel remains in cfg-refreshed state when the port-channel is created immediately followed by the EPG create using that port-channel
Condition:

Below are the steps to reproduce the issue:

1. Create port-channel po1 under the ownership of tenant1

2. Create endpoint group with po1 under the ownership of tenant1

3. After step 2 begins and before step 2 completes, the raslog event w.r.t. step 1 i.e. port-channel creation is received. This Ralsog event is processed after step 2 is completed

Recovery:

1. Introduce switchport or switchport-mode drift on the SLX for the port-channel which is in cfg-refreshed state

2. Perform manual DRC to bring back the cfg-refreshed port-channel back to cfg-in-sync

Parent Defect ID: EFA-9010 Issue ID: EFA-9010
Severity: S2 - High
Product: Extreme Fabric Automation Reported in Release: EFA 2.4.2
Symptom:

Creation of 100 Openstack VM/stacks fails at the rate of 10 stacks/min

One stack has 1 VM , 2 networks and 3 Ports (2 dhcp and one nova port)

Condition:

100 openstack stacks created at the rate of 10 stacks/min are sent to the EFA.

The EFA processing requests at such high case resuts in overwhelming the CPU,

Since the EFA cannot handle requests at such high rates, backlog of requests are created. This eventually results in VM reschedules and failure to complete some stacks with errors.

Workaround: 100 openstack stacks can be created with lower rate of creation consistently eg 3 stacks/min
Recovery: Delete the failed or all openstack stacks and create them at lower rate of creation e.g 3 stacks/min
Parent Defect ID: EFA-8904 Issue ID: EFA-8904
Severity: S2 - High
Product: Extreme Fabric Automation Reported in Release: EFA 2.4.2
Symptom: Single node deployment fails with 'DNS resolution failed.'
Condition: After a multi-node deployment and then un-deployment is done on a server, if single-node deployment is tried on the same server, the installer exits with the error, 'DNS resolution failed.'
Workaround: After un-deployment of the multi-node installation, perform a reboot of the server/TPVM.
Parent Defect ID: EFA-8535 Issue ID: EFA-8535
Severity: S3 - Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.4.0
Symptom: On a single-node installation of TPVM, after ip-change, EFA is not operational.
Condition: After IP change of the host system, if 'efa-change-ip' script is run by a different user other than the installation user, in that case, EFA is not operational.
Workaround: Restart k3s service using the command 'sudo systemctl restart k3s'
Parent Defect ID: EFA-8448 Issue ID: EFA-8448
Severity: S3 - Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.4.0
Symptom:

When the ports provided by the user in “tenant update port-delete operation” contains all the ports owned by the port-channel, the PO goes into delete pending state. However, the ports are not deleted from the PO.

They get deleted from the tenant though.

Condition: This issue is seen when the ports provided by the user in “tenant update port-delete operation” contains all the ports owned by the port-channel resulting in an empty PO.
Workaround: User needs to provide ports for “tenant update port-delete operation” which do not result in an empty PO i.e. PO needs to have at least 1 member port.
Recovery: Add the ports back using "tenant port-add operation" so that the port-channel has at least 1 member port. The use "efa configure tenant port-channel" to bring the po back to stable state.
Parent Defect ID: EFA-8297 Issue ID: EFA-8297
Severity: S3 - Medium
Product: Extreme Fabric Automation Reported in Release: EFA 2.4.0
Symptom:

EPG update anycast-ip-delete operation succeeded for deletion of provisioned anycast-ip for admin-down device.

This issue is observed only if an update anycast-ip-add operation is performed after device is put in admin down state and the new config is in non-provisioned state followed by anycast-ip-delete operation for already configured anycast-ip.

Condition:

Steps to reproduce issue:

1) Configure EPG with anycast-ip (ipv4/ipv6)

2) Make one device admin-down

3) Anycast-ip update-add new anycast-ip (ipv6/ipv4)

4) Update-delete provisioned anycast-ip configured in step-1 (ipv4/ipv6)

Step (4) should fail as IP is already configured on the device and trying to delete it should fail as part of APS.

Workaround: No workaround for this.
Recovery: Recovery can be done by configuring EPG again with the required configuration using efa or cleaning device config for anycast-ip on the switch.

Parent Defect ID: EFA-5928 Issue ID: EFA-5928
Severity: S2 - High
Product: Extreme Fabric Automation Reported in Release: EFA 2.2.0
Symptom: Configuring devices to default startup-config and adding them to a non-clos fabric does not enable all MCT ports resulting into fabric validation failure for missing link
Condition: Added devices immediately after setting to default startup config
Workaround:

Remove the devices from fabric and re-add

efa fabric device remove --name <fabric-name> --ip <device-ips>

efa inventory device delete --ip <device-ips>

efa fabric device add-bulk --name <fabric-name> --rack <rack-name> --username <username> --password <password> --ip <device-ips>

Parent Defect ID: EFA-5592 Issue ID: EFA-5592
Severity: S2 - High
Product: Extreme Fabric Automation Reported in Release: EFA 2.2.0
Symptom: Certificates need to be manually imported on replaced equipment in-order to perform RMA.
Condition: RMA/replaced equipment will not have ssh key and auth certificate, in-order to replay the configuration on new switch user needs to import the certificates manually.
Workaround:

import certificate manually

efa certificates device install --ips x,y --certType