Verify the Running System and Services

After any of the following scenarios, wait 10 minutes for EFA micro-services to be operational before you run EFA commands.
  • Powering on the OVA
  • Rebooting the OVA
  • Rebooting the TPVM
  • Rebooting the SLX (which also reboots the TPVM)
  • Rebooting the server on which the EFA is installed

You can use various commands and scripts to verify the status of the EFA system, to help troubleshoot, and to view details of EFA nodes, PODs, and services.

  1. Verify the K3s installation in a TPVM.
    1. Run the show efa status command from the SLX command prompt.
      device# show efa status
      ===================================================
                     EFA version details                  
      ===================================================
      Version : 2.7.0
      Build: 30
      Time Stamp: 22-04-23:12:37:36
      Mode: Secure
      Deployment Type: multi-node
      Deployment Platform: TPVM
      Virtual IP: 10.20.59.183
      Node IPs: 10.20.59.181,10.20.59.182
      --- Time Elapsed: 12.253203ms ---
      
      ===================================================
                     EFA Status                          
      ===================================================
      +-----------+---------+--------+--------------+
      | Node Name | Role    | Status | IP           |
      +-----------+---------+--------+--------------+
      | node-1    | active  | up     | 10.20.59.182 |
      +-----------+---------+--------+--------------+
      | node-2    | standby | up     | 10.20.59.181 |
      +-----------+---------+--------+--------------+
      --- Time Elapsed: 6.888985085s ---
    Output varies by type of deployment, such as single-node or multi-node, and the services that are installed.
  2. View details of EFA nodes, PODs, and services.
    1. Run the efa status command.

      On a multi-node installation:

      +-----------+---------+--------+---------------+
      | Node Name | Role    | Status | IP            |
      +-----------+---------+--------+---------------+
      | tpvm2     | active  | up     | 10.20.216.242 |
      +-----------+---------+--------+---------------+
      | tpvm1     | standby | up     | 10.20.216.241 |
      +-----------+---------+--------+---------------+
      --- Time Elapsed: 4.277420974s ---

      On a single-node installation:

      +-----------+--------+--------+-------------+
      | Node Name | Role   | Status | IP          |
      +-----------+--------+--------+-------------+
      | efa       | active | up     | 10.21.90.43 |
      +-----------+--------+--------+-------------+
      --- Time Elapsed: 1.461512261s ---
      These examples show only a few of all possible rows of detail.
  3. Verify that all PODs are in a running state.
    1. Run the k3s kubectl get pods -n efa command.
      extreme@node-1:~$ k3s kubectl -n efa get pods 
      
      NAME                           READY   STATUS    RESTARTS   AGE
      efa-api-docs-m6dgg             1/1     Running   0          19h
      gosnmp-service-vmmf7           1/1     Running   0          19h
      goopenstack-service-lf45g      1/1     Running   0          19h
      rabbitmq-ckvrq                 1/1     Running   0          19h
      gorbac-service-2v8qh           1/1     Running   0          19h
      gonotification-service-9tjbg   1/1     Running   0          19h
      goauth-service-8zltv           1/1     Running   0          19h
      goinventory-service-gsxzb      1/1     Running   0          19h
      gohyperv-service-ng95v         1/1     Running   0          19h
      govcenter-service-rt9cc        1/1     Running   0          19h
      gofabric-service-nnv4m         1/1     Running   0          19h
      gosystem-service-wxnww         1/1     Running   0          19h
      gopolicy-service-hrl7d         1/1     Running   0          19h
      goraslog-service-fqdjm         1/1     Running   0          19h
      gotenant-service-v5nq2         1/1     Running   0          19h
      In a multi-node installation, only the pods on the active node are in "Running" status.
  4. Verify the status of the Authentication service.
    1. Run the systemctl status hostauth.service script.
      $ systemctl status hostauth.service
      hostauth.service - OS Auth Service
      Loaded: loaded (/lib/systemd/system/hostauth.service; enabled; vendor preset: enabled)
      Active: active (running) since Thu 2020-04-23 07:56:20 UTC; 23 h ago
      Main PID: 23839 (hostauth)
      Tasks: 5
      CGroup: /system.slice/hostauth.service
              23839 /apps/bin/hostauth
      
      Apr 23 07:56:20 tpvm2 systemd[1]: Started OS Auth Service
  5. Restart a service using the efactl restart-service <service-name> command.
  6. Identify the active node that serves as the database for Kubernetes clusters.
    1. Run the ip addr show command from all nodes.
    2. Verify that on one of the Ethernet interfaces, the virtual IP address shows up as the secondary IP address.