Verify the Running System and Services

Before you begin

After any of the following scenarios, wait 10 minutes for EFA micro-services to be operational before you run EFA commands.
  • Powering on the OVA
  • Rebooting the OVA
  • Rebooting the TPVM
  • Rebooting the SLX (which also reboots the TPVM)
  • Rebooting the server on which the EFA is installed

About this task

You can use various commands and scripts to verify the status of the EFA system, to help troubleshoot, and to view details of EFA nodes, PODs, and services.

Procedure

  1. Verify the K3s installation in a TPVM.
    1. Run the show efa status command from the SLX command prompt.
      device# show efa status
      ===================================================
                     EFA version details                  
      ===================================================
      Version : 3.1.0
      Build: GA
      Time Stamp: 22-11-10:17:11:29
      Mode: Secure
      Deployment Type: multi-node
      Deployment Platform: TPVM
      Deployment Suite: Fabric Automation
      Virtual IP: 10.20.54.87
      Node IPs: 10.20.54.88,10.20.54.89
      Node IPv6s: ,
      --- Time Elapsed: 9.30156ms ---
      
      ===================================================
                     EFA Status
      ===================================================
      +-----------+---------+--------+-------------+
      | Node Name | Role    | Status | IP          |
      +-----------+---------+--------+-------------+
      | node-1    | active  | up     | 10.20.54.88 |
      +-----------+---------+--------+-------------+
      | node-2    | standby | up     | 10.20.54.89 |
      +-----------+---------+--------+-------------+
      --- Time Elapsed: 19.438967114s ---
    Output varies by type of deployment, such as single-node or multi-node, and the services that are installed.
  2. View details of EFA nodes, PODs, and services.
    1. Run the efa status command.

      On a multi-node installation:

      +-----------+---------+--------+---------------+
      | Node Name | Role    | Status | IP            |
      +-----------+---------+--------+---------------+
      | tpvm2     | active  | up     | 10.20.216.242 |
      +-----------+---------+--------+---------------+
      | tpvm1     | standby | up     | 10.20.216.241 |
      +-----------+---------+--------+---------------+
      --- Time Elapsed: 4.277420974s ---

      On a single-node installation:

      +-----------+--------+--------+-------------+
      | Node Name | Role   | Status | IP          |
      +-----------+--------+--------+-------------+
      | efa       | active | up     | 10.21.90.43 |
      +-----------+--------+--------+-------------+
      --- Time Elapsed: 1.461512261s ---
      These examples show only a few of all possible rows of detail.
  3. Verify that all PODs are in a running state.
    1. Run the k3s kubectl get pods -n efa command.
      (efa:extreme)extreme@node-1:~$ k3s kubectl get pods -n efa -o wide
      
      NAME                         READY STATUS   RESTARTS AGE   IP            NODE NOMINATED READINESS
                                                                                    NODE      GATES
      efa-api-docs-z84wn           1/1   Running  0        5h3m  10.42.194.72  efa  <none>    <none>
      gosystem-service-t4h2b       1/1   Running  0        5h3m  10.42.194.74  efa  <none>    <none>
      rabbitmq-vn27v               1/1   Running  0        5h4m  10.42.194.69  efa  <none>    <none>
      goinventory-service-vpdj7    1/1   Running  0        5h3m  10.42.194.75  efa  <none>    <none>
      goauth-service-g76c4         1/1   Running  0        5h3m  10.42.194.71  efa  <none>    <none>
      gorbac-service-jzcnf         1/1   Running  0        5h3m  10.42.194.70  efa  <none>    <none>
      gofaultmanager-service-wzwgp 1/1   Running  0        5h3m  10.42.194.73  efa  <none>    <none>
      gotenant-service-qmvzb       1/1   Running  0        5h3m  10.42.194.78  efa  <none>    <none>
      gonotification-service-h9ms2 1/1   Running  0        5h2m  10.20.54.87   efa  <none>    <none>
      goraslog-service-rvjsj       1/1   Running  0        5h3m  10.20.54.87   efa  <none>    <none>
      gofabric-service-6c4qs       1/1   Running  0        5h3m  10.42.194.76  efa  <none>    <none>
      gopolicy-service-g78bh       1/1   Running  0        5h3m  10.42.194.77  efa  <none>    <none>
      gosnmp-service-x86sn         1/1   Running  0        5h1m  10.20.54.87   efa  <none>    <none>
      (efa:extreme)extreme@node-1:~$
  4. Verify the status of the Authentication service.
    1. Run the systemctl status hostauth.service script.
      $ systemctl status hostauth.service
      hostauth.service - OS Auth Service
      Loaded: loaded (/lib/systemd/system/hostauth.service; enabled; vendor preset: enabled)
      Active: active (running) since Thu 2020-04-23 07:56:20 UTC; 23 h ago
      Main PID: 23839 (hostauth)
      Tasks: 5
      CGroup: /system.slice/hostauth.service
              23839 /apps/bin/hostauth
      
      Apr 23 07:56:20 tpvm2 systemd[1]: Started OS Auth Service
  5. Restart a service using the efactl restart-service <service-name> command.
  6. Identify the active node that serves as the database for Kubernetes clusters.
    1. Run the ip addr show command from all nodes.
    2. Verify that on one of the Ethernet interfaces, the virtual IP address shows up as the secondary IP address.