Open Defects

The following defects are open in Extreme Fabric Automation 2.7.0.

Parent Defect ID: EFA-9439 Issue ID: EFA-9439
Severity: S2 - Major
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: Dev-State and App-State of EPG Networks are not-provisioned and cfg-ready
Condition:

Below are the steps to reproduce the issue:

1) Create VRF with local-asn

2) Create EPG using the VRF created in step 1

3) Take one of the SLX devices to administratively down state

4) Perform VRF Update "local-asn-add" to different local-asn than the one configured during step 1

5) Perform VRF Update "local-asn-add" to the same local-asn that is configured during step 1

6) Admin up the SLX device which was made administratively down in step 3 and wait for DRC to complete

Workaround: No workaround as such.
Recovery:

Following are the steps to recover:

1) Log in to SLX device which was made admin down and then up

2) Introduce local-asn configuration drift under "router bgp address-family ipv4 unicast" for the VRF

3) Execute DRC for the device

Parent Defect ID: EFA-9570 Issue ID: EFA-9570
Severity: S2 - Major
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: Add Device Failed because ASN used in border leaf showing conflict
Condition: If there are more than one pair of Leaf/border leaf devices then devices which are getting added first will get the first available ASN in ascending order and in subsequent addition of devices if one of device is trying to allocate the same ASN because of brownfield scenario then EFA will throw an error of conflicting ASN
Workaround:

Add the devices to fabric in the following sequence

1)First add devices that have preconfigured configs

2)Add remaining devices that don't have any configs stored

Recovery:

Removing the devices and adding the devices again to fabric in following sequence

1)First add devices that have preconfigured configs

2)Add remaining unconfigured devices.

Parent Defect ID: EFA-9591 Issue ID: EFA-9591
Severity: S2 - Major
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: "efa fabric configure" fails with error after previously changing the fabric password in the configured fabric
Condition: This condition was seen when "efa fabric configure --name <fabric name>" was issued after modifying the MD5 password. Issue is observed when certain BGP sessions are not in an ESTABLISHED state after clearing the BGP sessions as part of fabric configure.
Workaround: Wait for BGP sessions to be ready by checking the status of BGP sessions using "efa fabric topology show underlay --name <fabric name>"
Recovery: Wait for BGP sessions to be ready. Check the status of BGP sessions using "efa fabric topology show underlay --name <fabric name>"
Parent Defect ID: EFA-10062 Issue ID: EFA-10062
Severity: S3 - Moderate
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: Removing a device from Inventory does not clean up breakout configuration on interfaces that are part of port channels.
Condition: This condition occurs when there is breakout configuration present on device that is being deleted from Inventory, such that those breakout configurations are on interfaces that are part of port-channels
Workaround: Manually remove the breakout configuration, if required.
Recovery: Manually remove the breakout configuration, if required.
Parent Defect ID: EFA-10063 Issue ID: EFA-10063
Severity: S3 - Moderate
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.0
Symptom: Deleting device from EFA Inventory does not bring up the interface to admin state 'up' after unconfiguring breakout configuration
Condition: This condition occurs when there is a breakout configuration present on the device that is being deleted from EFA Inventory
Workaround: Manually bring the admin-state up on the interface, if required
Recovery: Manually bring the admin-state up on the interface, if required
Parent Defect ID: EFA-12133 Issue ID: EFA-12133
Severity: S2 - Major
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.5
Symptom: On the last port-group delete, VRF VRF1 is not cleaned up from the device, when the VRF is shared across the EPGs.
Condition:

Below are the steps to reproduce the issue:

1. Create L3 EPG EPG1 with Device1Port1 and VRF1.

2. Create L3 EPG EPG2 with Device1Port2, Device2Port1, and VRF1

3. Update EPG EPG2 with "port-group-delete" of Device1Port2

4. Update EPG EPG1 with "port-group-delete" of Device1Port1. This is the last port getting deleted from the device which should have resulted in the deletion of the VRF VRF1 from the Device1.

Recovery:

Recovery way 1:

1. Delete EPG1.

2. EPG2 update with port-group add D1P2 and then remove D1P2 from EPG.

After the port removal D1P2 (last-port) vrf will be removed from the device.

Recovery way 2:

Manually remove the vrf from the device.. Inventory update.

Parent Defect ID: EFA-12237 Issue ID: EFA-12237
Severity: S2 - Major
Product: Extreme Fabric Automation Reported in Release: EFA 2.5.4
Symptom: EPG update port-group-delete operation results in the runtime error "Execution error: service is not available or internal server error has occurred, please try again later"
Condition:

Below are the steps to reproduce the issue:

1. Create a BD based tenant under a CLOS or Non-CLOS fabric.

2. Create a BD based EPG (under the ownership of the tenant created in step 1) with some ctags and some member port-channels.

3. For the reasons unknown, the BD (Bridge Domain) configuration pertaining to one of the member port-channel got deleted from the EFA DB, causing the DB to be in an inconsistent state.

4. Execute EPG update "port-group-delete" operation to remove the member port-channel whose BD configuration is inconsistent.

Recovery:

No recovery through EFA CLI.

The inconsistent DB needs to be corrected by creating dummy BD (Bridge Domain) entries in the database followed by EPG update "port-group-delete".

Parent Defect ID: EFA-12600 Issue ID: EFA-12600
Severity: S3 - Moderate
Product: Extreme Fabric Automation Reported in Release: EFA 2.3.2
Symptom: EFA certificates have expired or about to expire.
Recovery:

Perform the following steps to renew EFA certificates on 2.3.x and 2.4.x.

1. cd /apps/efa/efacerts

2. Generate a new certificate using the tls.key from <IP>-certs folder

# openssl req -new -sha256 -key <IP>-certs/tls.key -subj "/CN=efa.extremenetworks.com" | openssl x509 -req -sha256 \

-CA /apps/efa/efacerts/extreme-ca.cert.pem \

-CAkey /apps/efa/efacerts/extreme-ca.key.pem \

-CAcreateserial \

-out newtls.crt -days 365 \

-extensions v3_req -extfile /apps/efa/efacerts/extreme-openssl.cnf-san

3. Install the new certificate into traefik using the following shell script

# if ./install_efa_certs_st.sh --cert /apps/efa/efacerts/newtls.crt --key /apps/efa/efacerts/<IP>-certs/tls.key ; then

cp newtls.crt <IP>-certs/tls.crt

cp newtls.crt /apps/efadata/certs/own/tls.crt

echo SUCCESS

fi

4. Wait for the traefik pod to restart (1-2min)

5. Restart the running goraslog pod

Parent Defect ID: EFA-12710 Issue ID: EFA-12710
Severity: S2 - Major
Product: Extreme Fabric Automation Reported in Release: EFA 2.7.0
Symptom:

With rollback , Multi-node upgrade with node replacement is not supported.

This happens when user gives command efa deploy with-rollback for multi-node replacement, user is given option 1) Multi Node Build Upgrade and 2) Multi Node Build Upgrade With Node Replacement . If user chooses the option 2 Multi Node Build Upgrade With Node Replacement . Installer prompts

"With rollback , replacement upgrade not supported

Do you wish to restart the install? (yes/no)"

Condition: This happens when user gives command efa deploy with-rollback for Multi Node Build Upgrade With Node Replacement
Workaround:

When the installer prompts,

1) Multi Node Build Upgrade and 2) Multi Node Build Upgrade With Node Replacement .

if User presses option 2 ,

Installer promtps

"With rollback , replacement upgrade not supported

Do you wish to restart the install? (yes/no)"

User can input no to halt the installation and press yes if he wants to go for option 1) Multi Node Build Upgrade

Recovery:

When the installer prompts,

1) Multi Node Build Upgrade and 2) Multi Node Build Upgrade With Node Replacement .

if User presses option 2 ,

Installer promtps

"With rollback , replacement upgrade not supported

Do you wish to restart the install? (yes/no)"

User can input no to halt the installation

Parent Defect ID: EFA-12777 Issue ID: EFA-12777
Severity: S2 - Major
Product: Extreme Fabric Automation Reported in Release: EFA 2.7.0
Symptom: "efa inventory drift-reconcile execute --ip <device-ip>" fails with the status "tenant-dr-timeout" as the time taken to reconcile the scale drifted tenant configuration is beyond the timeout
Condition:

Below are the steps to reproduce the issue:

1. Introduce drift w.r.t the scaled tenant configuration

e.g. 100 POs, 100 VRFs, 200 EPGs (with around 100 ctags), 300 BGP Peers, 100 BGP Peer-Groups

2. Execute "efa inventory drift-reconcile execute --ip <device-ip>"

Workaround: No workaround
Recovery:

Below steps needs to be re-executed on the device for which the DRC had failed with the status "tenant-dr-timeout"

1. "efa inventory device update --ip <device-ip>"

2. "efa inventory drift-reconcile execute --ip <device-ip>"

Parent Defect ID: EFA-12823 Issue ID: EFA-12823
Severity: S2 - Major
Product: Extreme Fabric Automation Reported in Release: EFA 2.7.0
Symptom: Prefix Independent Convergence is not detected as drifted and it will not get reconciled.
Condition: Occasionally when preforming DRC in maintenance mode on reboot state, the Prefix Independent Convergence fails to detect refreshed state.
Workaround: Working around the drift detection is done by validating Prefix Independent Convergence has not drifted before reloading into maintenance mode.
Recovery: When Prefix Independent Convergence is not being detected as refreshed, re-configuring the value will allow drift detection to function.
Parent Defect ID: EFA-13036 Issue ID: EFA-13036
Severity: S2 - Major
Product: Extreme Fabric Automation Reported in Release: EFA 2.7.0
Symptom: EFA created Mirror Session with Global VLANs as the source is not deleted
Condition:

Below are the steps to reproduce the issue:

1) Create Tenant and EndpointGroup with ctag-range

2) Create Mirror Session with Global VLANs (VLANs to be chosen from the ctag-range mentioned in step 1) as a mirror source

3) Delete EndpointGroup (created in Step1) with the force option

Workaround: No workaround
Recovery: Delete Mirror Session using "efa tenant service mirror session delete -- tenant <tenant-name> --name <mirror-session-name>"
Parent Defect ID: EFA-13080 Issue ID: EFA-13080
Severity: S3 - Moderate
Product: Extreme Fabric Automation Reported in Release: EFA 2.7.0
Symptom: Some configurations when removed from the device will not cause the EFA to move the device to cfg-refresh state.
Condition:

Case1: EFA doesnt move device to config-refresh on device update. Config does NOT Reconcile on drift-reconcile:

Fabric devices app state should be cfg-refreshed, after peer-keepalive configuration removed manually in slx

Fabric devices app state should be cfg-refreshed, after "ip address" configuration under a fabric interface removed manually in slx

Fabric devices app state should be cfg-refreshed, after loopback interface configuration under a fabric is removed manually in slx

Fabric devices app state should be cfg-refreshed, after "address-family l2vpn evpn" is removed

Fabric devices app state should be cfg-refreshed, after "no neighbor <ip> next-hop-self" is removed

Case2 : EFA doesnt move device to config-refresh on device update. Config does Reconcile on drift-reconcile:

Fabric devices app state should be cfg-refreshed, after "maximum-paths" configuration under router bgp removed manually in slx

Fabric devices app state should be cfg-refreshed, after "graceful-restart" configuration under router bgp removed manually in slx

Workaround:

Manually configure the device back in case 1

Run drift-reconcile OR manually configure device back will recover the config in case 2

Parent Defect ID: EFA-13083 Issue ID: EFA-13083
Severity: S3 - Moderate
Product: Extreme Fabric Automation Reported in Release: EFA 2.7.0
Symptom: "efa tenant po show", "efa tenant vrf show", "efa tenant epg show", "efa tenant service mirror session show" doesn't show the configuration in cfg-refreshed state even though the configurations (which should have been present but) are not present on the SLX
Condition:

Below are the steps to reproduce the issue

1. Configure fabric, tenant, po, vrf, epg, mirror session

2. Execute "efa system backup"

3. Delete the devices from inventory

4. Execute "efa system restore" using the backup taken in step 2

5. Execute "efa tenant po show", "efa tenant vrf show", "efa tenant epg show", "efa tenant service mirror session show"

Recovery: Execute "efa inventory device update --ip <device-ip>" and then check the output of "efa tenant po show", "efa tenant vrf show", "efa tenant epg show", "efa tenant service mirror session show" to see the configurations in cfg-refreshed state
Parent Defect ID: EFA-13124 Issue ID: EFA-13124
Severity: S2 - Major
Product: Extreme Fabric Automation Reported in Release: EFA 2.7.0
Symptom: When endpoint group create or update operation REST requests of multiple endpoint groups each with 50+ ctags are issued concurrently, one or two of the requests may fail with "Error 1452: Cannot add or update a child row: a foreign key constraint fails" or with an error indicating database timeout or an error indicating failure of network property delete.
Condition: When multiple endpoint group requests are processed concurrently, some of the database requests initiated by EFA may cause database to abort one of the request with the above mentioned error
Workaround: Execute the commands sequentially
Recovery: EFA database and SLX device configurations are always not affected by this error and hence no recovery is required. The failed commands shall be rerun sequentially to successful completion of the expected operations
Parent Defect ID: EFA-13158 Issue ID: EFA-13158
Severity: S2 - Major
Product: Extreme Fabric Automation Reported in Release: EFA 2.7.0
Symptom: "efa inventory drift-reconcile execute --ip <device-ip>" fails with the error "Error: Monitor session already configured"
Condition:

Below are the steps to reproduce the issue

1. Create a mirror session on an SLX device using EFA

2. Modify the mirror session configuration on SLX by changing any of the attributes e.g. source, destination, direction, etc.

3. Perform "efa inventory drift-reconcile execute --ip <device-ip> --reconcile"

Recovery: Delete existing monitor session from the device and then execute "efa inventory drift-reconcile execute --ip <device-ip>"
Parent Defect ID: EFA-13178 Issue ID: EFA-13178
Severity: S2 - Major
Product: Extreme Fabric Automation Reported in Release: EFA 2.7.0
Symptom: Fabric configuration failed to reconcile when DRC was on-going and user initiated a EFA backup
Condition: EFA's backup needs to stop services to ensure that the database is in quiet state, so that the backup is consistent.
Workaround: Users should run a backup once the devices are completed going through DRC
Recovery: Recovery would be to run DRC operation on that device again once the backup is completed.
Parent Defect ID: EFA-13187 Issue ID: EFA-13187
Severity: S2 - Major
Product: Extreme Fabric Automation Reported in Release: EFA 2.7.0
Symptom: EFA REST CLI to fetch list of tenants may fail with error "Error : Cannot find Tenant <tenant-name>" when there are large number of tenants and epgs configured on them.
Condition: When there are large number of tenants with ports spanning across 30+ devices with multiple EPGs configured, database access can be under heavy load causing the above error
Workaround: This is a transient error. Re-execute the same REST command. If this error is observed, information about individual tenant objects can be fetched by 'efa tenant show --name <tenant-name>'
Recovery: There is no recovery required as the EFA and SLX configurations are not altered as part of this issue
Parent Defect ID: EFA-13254 Issue ID: EFA-13254
Severity: S3 - Moderate
Product: Extreme Fabric Automation Reported in Release: EFA 2.6.1
Symptom: 3 of EFA Pods fails Liveliness / Rediness checks causing Init containers to stop and causing crashloopback
Parent Defect ID: EFA-13281 Issue ID: EFA-13281
Severity: S2 - Major
Product: Extreme Fabric Automation Reported in Release: EFA 2.7.0
Symptom: When 'efa system backup' command is executed with copy to the remote system enabled and the error 'local error: tls: bad record MAC' is seen.
Workaround: The backup will be copied to the remote location and the error is harmless.
Parent Defect ID: EFA-13291 Issue ID: EFA-13291
Severity: S2 - Major
Product: Extreme Fabric Automation Reported in Release: EFA 2.7.0
Symptom: Trusted-Peer IP is configured as its own TPVM IP address during tpvm-upgrade
Workaround:
Parent Defect ID: EFA-13322 Issue ID: EFA-13322
Severity: S3 - Moderate
Product: Extreme Fabric Automation Reported in Release: EFA 2.7.0
Symptom:

At the end of upgrade, the installer displays a

message associated with a fresh install and not an upgraded install. The installer will show

"Extreme Fabric Automation Stack is now deployed and ready"

instead of

"Extreme Fabric Automation Stack has been upgraded successfully"

This is harmless and the upgrade procedure is unaffected.

Condition: When a single-node installation of EFA is upgraded using the SLX CLI with-rollback option
Parent Defect ID: EFA-13339 Issue ID: EFA-13339
Severity: S3 - Moderate
Product: Extreme Fabric Automation Reported in Release: EFA 2.7.0
Symptom: The EFA notification service does not send a syslog alert message when an EFA inventory device firmware-download operation fails.
Condition: The user attempts to prepare a device for a firmware download using "efa inventory device firmware-download prepare add --ip <device IP>" when the device's management connectivity is unreachable.
Workaround: Although the syslog alert message is not available, both the CLI and REST response contain an appropriate error message about the reason for the firmware-download prepare error and the device's connectivity issue.
Recovery: None
Parent Defect ID: EFA-13362 Issue ID: EFA-13362
Severity: S3 - Moderate
Product: Extreme Fabric Automation Reported in Release: EFA 2.7.0
Symptom: If the use manually removes "member-vlan-all" or " member-bd-all", from the SLX followed by DRC, then the entire cluster configuration will be deleted and recreated.
Condition:

1. Create a single rack fabric

2. If the use manually removes "member-vlan-all" or " member-bd-all", from the SLX followed by DRC.

Workaround: Manually restore the member-vlan or member-bridge-all configuration in cluster.
Parent Defect ID: EFA-13367 Issue ID: EFA-13367
Severity: S2 - Major
Product: Extreme Fabric Automation Reported in Release: EFA 2.7.0
Symptom: DRC does not restore config after MCT PO member port shutdown manually in SLX
Condition:

Execute in device command 'shutdown' of member port of MCT PO.

EFA device update brings the device to 'cfg-refresh'

Execute Drift-reconcile for device does not bring the member port back up ie 'no shutdown' is not issued on member port interface.

Workaround: Execute on device 'no shutdown' on the interface manually on the device.
Recovery: Execute 'efa fabric configure' for fabric to bring devices in efa to in-sync state.
Parent Defect ID: EFA-13370 Issue ID: EFA-13370
Severity: S3 - Moderate
Product: Extreme Fabric Automation Reported in Release: EFA 2.7.0
Symptom: No alert messages are sent to configured syslog relp subscribers when there is a failure while configuring inventory device interface settings such as admin state, speed, mtu etc.
Condition: The lack of logging occurs when there is a failure while configuring inventory device interface settings, such as a failure to establish connection to a device.
Workaround: The failure is logged in the inventory log file and can be reviewed there.
Recovery: Once the failure condition is corrected there is no need for an alert message to be sent to configured syslog relp subscribers.
Parent Defect ID: EFA-13379 Issue ID: EFA-13379
Severity: S3 - Moderate
Product: Extreme Fabric Automation Reported in Release: EFA 2.7.0
Symptom: Some of the backup routing neighbors get stuck in cfg-refreshed state and the same can be seen in the DRC output
Condition:

Below are the steps to reproduce the issue

1. Create and configure a fabric with backup routing enabled

2. Create a Tenant, a PO, 5 VRFs and 5 EPGs

3. Update the fabric setting (for the fabric created in step 1) with md5 password

4. Perform fabric configure

Recovery:

1. Remove backup routing neighbor md5-password from the VRFs which are in cfg-refreshed state (as seen in the DRC output)

2. Execute DRC to reconcile the configs and to move the cfg-refreshed configurations to cfg-in-sync