No relevant resource is found in the selected language.

This site uses cookies. By continuing to browse the site you are agreeing to our use of cookies. Read our privacy policy>Search

Reminder

To have a better experience, please upgrade your IE browser.

upgrade

HUAWEI CLOUD Stack 6.5.0 Alarm and Event Reference 04

Rate and give feedback:
Huawei uses machine translation combined with human proofreading to translate this document to different languages in order to help you better understand the content of this document. Note: Even the most advanced machine translation cannot match the quality of professional translators. Huawei shall not bear any responsibility for translation accuracy and it is recommended that you refer to the English document (a link for which has been provided).
ALM-15415 Abnormal Instance Status of Physical Multi-Tenant Kafka

ALM-15415 Abnormal Instance Status of Physical Multi-Tenant Kafka

Alarm Description

If the Kafka instance status is abnormal, the statuses of multiple broker processes in Kafka are abnormal.

Attribute

Alarm ID

Alarm Severity

Alarm Type

15415

Major

Integrity violation

Alarm Parameters

Parameter

Description

hostIP

Specifies the node for which the alarm is reported.

instanceName

Specifies the Kafka instance name

Impact on the System

If the Kafka status is abnormal, the following functions may become unavailable: topic creation and deletion and message consumption and confirmation.

Possible Causes

  • The connection between the Server node and the node of physical multi-tenant Kafka is abnormal.
  • The Kafka process is running abnormally.

Procedure

  1. Use PuTTY to log in to the manage_lb1_ip node.

    The default username is paas, and the default password is QAZ2wsx@123!.

  2. Run the following command and enter the password of the root user to switch to the root user:

    su - root

    Default password: QAZ2wsx@123!

  3. Run the following command to obtain the project ID:

    kubectl get node | grep dms

    1792b1d9475f4999bd32551d7c43a35e   dms-np-phy-kafka-9eeb4814-ht6nq                  Ready      <none>    6d        v2.3.27-FusionStage6.5.RP3-B100-dirty

    In the preceding information, the content in bold is the obtained project ID.

  4. Query the IP address of the node where the physical multi-tenant Kafka is installed.

    • Region Type III scenario
      1. Run the following command to query the IP address of the node where the physical multi-tenant Kafka is installed:

        kubectl get po -n 1792b1d9475f4999bd32551d7c43a35e -owide | grep dms-kafka-phy

        In the preceding command, 1792b1d9475f4999bd32551d7c43a35e is the project ID obtained in 3.

        dms-kafka-phy-9eeb4814-69fcb99f45-2n6m7           1/1       Running   0          6d        192.168.34.139   dms-np-phy-kafka-9eeb4814-ht6nq
        dms-kafka-phy-9eeb4814-69fcb99f45-d62tb           1/1       Running   0          6d        192.168.34.24    dms-np-phy-kafka-9eeb4814-hw9qb
        dms-kafka-phy-9eeb4814-69fcb99f45-l7l5q           1/1       Running   0          6d        192.168.34.179   dms-np-phy-kafka-9eeb4814-t2sc4

        In the preceding information, the content in bold indicates the pod of physical multi-tenant Kafka, and the IP address and the name of the node where the physical multi-tenant Kafka is installed, respectively.

    • Region Type I and Region Type II scenarios
      1. Perform the operations provided in 4.a.
      2. Run the following command to query the IP address of the node where the physical multi-tenant Kafka is installed:

        kubectl get node -owide | grep dms-np-phy-kafka-9eeb4814-ht6nq

  5. Run the following command to log in to the node where the physical multi-tenant Kafka is installed:

    ssh 192.168.34.139

    NOTE:

    192.168.34.139 is the IP address obtained in 4.

  6. Check the running status of the Kafka process. If the process is in the D, T, or Z state, the process is abnormal. In this case, view the corresponding logs in the /var/log/dms/kafka-logs/server.log directory to analyze the cause of the exception.
  7. If the cause cannot be located by analyzing logs, contact technical support for assistance.

Alarm Clearing

This alarm will be automatically cleared after the fault is rectified.

Related Information

None

Translation
Download
Updated: 2019-08-30

Document ID: EDOC1100062365

Views: 34349

Downloads: 31

Average rating:
This Document Applies to these Products
Related Documents
Related Version
Share
Previous Next