1 Introduction
This section contains the purpose and scope of the document, and also the revision information.
1.1 Document Purpose and Scope
The purpose of this document is to provide the instructions and tools needed to automate CUDB log collection and log analysis. This document requires knowledge of the product. It is addressed to both Ericsson personnel and system administrators. If the contents of this document and the Preventive Maintenance, Logchecker Found Error(s) document are not enough to fix a fault, then contact next level of maintenance support.
2 Overview
CUDB Logchecker is a software monitoring component on top of the current monitoring processes that aims to work as a preventive maintenance tool.
3 Usage Information
This section provides usage information about CUDB Logchecker.
3.1 General Information
CUDB Logchecker consists of the following two scripts:
| Note: |
These scripts can be executed only on the active System Controller
(SC). |
3.2 Automatic Log Collection and Log Analysis
CUDB log collection starts automatically every day at 00:25 and 12:25.
At 00:50 and 12:50, there is a scheduled CUDB log analysis, which saves the detailed result under the following location:
/home/cudb/monitoring/preventiveMaintenance/cron_analysis.<SC_NAME>.log
In the above path, <SC_NAME> can be SC_2_1 or SC_2_2.
The automatic log analysis can also send alarms: see General Information for more information.
| Note: |
Automatic log collection and analysis is performed only in
the active SC. |
The alarm severity is calculated by the number and the severity of faults found during the last analysis (that is, multiple minor faults can result in a major alarm). All these faults are weighted by CUDB Logchecker, then the severity of alarm is calculated and set, using the weight as input.
The severity levels of the alarm can be as follows:
3.2.1 Defining Custom Monitoring Intervals
It is possible to define custom monitoring intervals for CUDB Logchecker. Specify monitoring for both SCs separately.
Configuration can be checked using the crontab -l command, as shown in Example 1.
Define the same configuration files on both of the SCs. The paths of the files are as follows:
The following options are accepted, as shown in Example 2:
| Note: |
Follow the standard cron expression format when defining the value of
options. Configuration files must take effect after the SC is rebooted. Use the
crontab -e command to activate cron changes
immediately, without rebooting the SC and defining configuration files. |
If the configuration files do not exist, the following default options take effect:
The absence of a configuration file means that the default values are used.
| Note: |
Always leave at least 25 minutes between log collection and log analysis, in order to let
the log collection finish before the log analysis starts. |
Example 1 Configuration Check
CUDB81 SC_2_1# crontab -l # DO NOT EDIT THIS FILE - edit the master and reinstall. # (/var/spool/cron/tabs/root installed on Wed Oct 17 13:59:03 2012) # (Cron version V5.0 -- $Id: crontab.c,v 1.12 2004/01/23 18:56:42 vixie Exp $) 0,15,30,45 * * * * /home/cudb/oam/performanceMgmt/appCounters/scripts/appCounters.cron >> /dev/null 25 0,12 * * * /bin/bash /opt/ericsson/cudb/OAM/bin/cudbGetLogs 50 0,12 * * * /bin/bash /opt/ericsson/cudb/OAM/bin/cudbAnalyser --auto-check --send-alarm --save-counter > \ /home/cudb/monitoring/preventiveMaintenance/cron_analysis.SC_2_1.log 37 0 * * * /bin/bash /opt/ericsson/cudb/Monitors/bin/cudbCheckConsistency --locked --alarms >/dev/null 2>&1 || true 7 0 * * * /bin/bash /opt/ericsson/cudb/Monitors/bin/cudbCheckReplication --locked --alarms >/dev/null 2>&1 || true */1 * * * * /opt/ericsson/cudb/Monitors/keepAlive/bin/keepAlive_monitor.sh 2>&1
Example 2 Configuration File
CUDB45 SC_2_1# cat logchecker.SC_2_1.conf getlogs_schedule=25 0,4,8,12,16,20 * * * analyser_schedule=50 0,4,8,12,16,20 * * *
3.3 Manual Log Collection and Analysis
It is possible to use the CUDB Logchecker to help troubleshooting. For more details, refer to CUDB Troubleshooting Guide.
4 Troubleshooting cudbAnalyser Results
The following sections describe the possible actions to take for each potential issue found during log analysis.
4.1 cudbSystemStatus
Table 1 shows the cudbSystemStatus faults in different scenarios. For more information on cudbSystemStatus, refer to the cudbSystemStatus section of CUDB Node Commands and Parameters.
|
Fault |
Example Printout |
Action |
|---|---|---|
|
Process not running as expected.(1) |
[ERROR] cudbSystemStatus: CUDB Process is not running as expected (Severity: Warning) [-W-] CudbNotifications process.....................Not running in: PL0 Shows which CUDB process is not running as expected. |
Run the cudbSystemStatus -p command to check if the problem is resolved. If the printout still shows that a CUDB process is not running, contact the next level of support. |
|
Replication delay of 1000-9999.(2) |
[ERROR] cudbSystemStatus: replication delay: 1000-9999 (Severity: Warning) Replication in DSG3(Node=246--Chan=1).. OK -- Delay = 1021 Shows which replication channel has a delay. |
Run the cudbSystemStatus -R command to verify that the replication delay is decreasing. If the replication delay is continuously increasing due to high load, consider reallocation. If the replication delay is not decreasing, contact the next level of support. |
|
Replication delay > 10000. |
[ERROR] cudbSystemStatus: replication delay > 10000 (Severity: Minor) Replication in DSG3(Node=246--Chan=1).. OK -- Delay = 20001 Shows which replication channel has a delay. |
Run the cudbSystemStatus -R command to verify that the replication delay is decreasing. If the replication delay is continuously increasing due to high load, consider reallocation. If the replication delay is not decreasing, contact the next level of support. |
|
BC server is not in a good shape. |
[ERROR] cudbSystemStatus: BC Server is not in a good shape. (Severity: Major) BC server in PL_2_5 ......... not running |
Run the cudbSystemStatus -b command to check if the problem persists. If yes, contact the next level of support. |
|
Unexpected MySQL connection state. |
[ERROR] cudbSystemStatus: Unexpected MySQL connection state(Severity: Critical) [-W-] MySQL Access Server connection Fault in....: DS6_0 |
Some maintenance actions may lead to this alert. If there are no maintenance actions ongoing at the time cudbSystemStatus is checked and the fault persists, contact the next level of support. |
|
Unexpected cluster status. |
[ERROR] cudbSystemStatus: Unexpected cluster status in CUDB system (Severity: Critical) 131080-1Unreachable 141080-1Unreachable [-W-]There are Clusters in wrong state |
Some maintenance actions may lead to this alert. If there are no maintenance actions ongoing at the time cudbSystemStatus is checked and the fault persists, contact the next level of support. |
|
Wrong cluster state or replication state. |
[ERROR] cudbSystemStatus: Wrong cluster state or replication state found in CUDB system (Severity: Critical) DSG13 is in wrong state DSG14 is in wrong state |
Run the cudbSystemStatus -r command to check if the problem persists. If yes, contact the next level of support. |
4.2 NDB
Table 2 shows the NDB faults due to the execution of cudbAnalyser in different scenarios.
|
Fault |
Example Printout |
Action |
|---|---|---|
|
NDB cluster logs. |
[ERROR] NDB: Errors found in NDB cluster logs (Severity: Warning) > 2013-07-17 09:55:51 [MgmtSrvr] ALERT \ -- Node 3: Arbitration check won - node group majority |
If this printout comes without any other alarm or explanation for the last 12-hour monitoring period, contact the next level of support to investigate the cause. |
|
ndb_out logs. |
[ERROR] NDB: Errors found in ndb_out logs (Severity: Warning) [ndbd] ALERT -- Node 4: Forced node shutdown completed. Caused by error 2303: 'System error, node killed during node restart by other node(Internal error, programming error or missing error message, please report a bug). Temporary error, restart node'. |
If this printout comes without any other alarm or explanation for the last 12-hour monitoring period, contact the next level of support to investigate the cause. |
|
Database cluster hanging in phase 4. |
[ERROR] NDB: Mysql hanging in Phase 4 (Severity: Warning) > 2016-03-18 19:26:56 [ndbd] INFO -- refing dict lock to 4 |
It is recommended to contact the next level of support to investigate the cause. |
4.3 OS
Table 3 shows OS faults in different scenarios.
|
Fault |
Example Printout |
Action |
|---|---|---|
|
High CPU load.(1) |
[ERROR] OS: high CPU load (Severity: Major) CUDB119 oam2 00:26am up 6 days 16:10, 0 users, load average: 25.35, 17.76, 14.16 |
Log in to the affected blade or VM, and run the top command to check if the load is still high. If a software fault is suspected, contact the next level of support. |
|
procs_blocked > 9, indicating high load, or hanging IO. |
[ERROR] OS: Kernel printout: procs_blocked > 9, indicating high load, or hanging IO (Severity: Major) CUDB100 oam2 procs_blocked 10 This kernel printout shows which blade or VM is affected. |
Check if there is another error printout by CUDB Logchecker that pinpoints an infrastructure or load related error. Log in to the affected blade or VM and run the dmesg command to see if the problem is caused by IO issues. Log in to the affected blade or VM and run the top command to see if the problem is caused by high load. It is recommended to contact the next level of support to investigate the issue.(2) |
|
dmesg shows IO error or EXT3 filesystem error. |
[ERROR] OS: dmesg shows IO error or EXT3 filesystem error (Severity: Major) Feb 2 20:55:10 PL_2_5 kernel: end_request: I/O error, dev sda, sector 112313 Feb 2 20:55:10 PL_2_5 kernel: Buffer I/O error on device sda1, logical block 14039 Shows which blade or VM is affected. |
Contact the next level of support to investigate the issue.(2) |
|
dmesg shows filesystem is mounted as read-only. |
[ERROR] OS: dmesg shows that filesystem is remounted as read-only (Severity: Major) Feb 2 20:55:15 PL_2_5 kernel: Remounting filesystem read-only Shows which blade or VM is affected. |
Contact the next level of support to investigate the issue.(2) |
|
Network stat shows errors. |
[ERROR] OS: network stat shows errors (Severity: Minor) > CUDB124 oam1 eth3/statistics/rx_dropped 16467 K packets > CUDB124 oam1 bond0/statistics/rx_dropped 16467 K packets > CUDB124 oam2 eth3/statistics/rx_dropped 803 K packets > CUDB124 oam2 bond0/statistics/rx_dropped 803 K packets Shows which blade or VM is affected. |
Contact the next level of support to investigate the issue.(2) |
|
For CUDB systems deployed on native BSP 8100, SMART (Self-Monitoring, Analysis and Reporting Technology) data indicates a possible imminent drive failure.(3) |
[ERROR] OS: SMART Health Status is NOT OK! (Severity: Major) CUDB136 oam1 SMART Health Status: HARDWARE IMPENDING FAILURE DATA ERROR RATE TOO HIGH [asc=5d, ascq=12] CUDB136 PL2 SMART Health Status: HARDWARE IMPENDING FAILURE DATA ERROR RATE TOO HIGH [asc=5d, ascq=12] |
A drive in the affected blade shows a pre-failure condition. To avoid blade failure, the blade containing the drive must be replaced.(4) |
|
New core dumps were generated. |
[ERROR] OS: New core dumps were generated under /local2/dumps in the PLs and under /cluster/dumps in PL_2_5 (More details inside <cudb_pl_core_dumps> in log files) (Severity: Major) > -rw------- 1 root root 367984640 Feb 12 15:10 slapd.31928.PL_2_3.core |
Contact the next level of support to investigate the issue. |
|
New printouts regarding system Out of Memory Killer. |
[ERROR] OS: New printouts regarding system Out of Memory Killer (Severity: Warning) > May 10 17:37:05 PL_2_3 kernel: [191935.037336] Out of memory: Kill process 26499 (slapd) score 452 or sacrifice child |
Contact the next level of support to investigate the issue. |
|
dmesg has changed. |
[ERROR] OS: dmesg has changed (Severity: Warning) > CUDB31 PL0 [1052612.453321] Clock: inserting leap second 23:59:60 UTC |
Contact the next level of support. |
|
There are more than one NTP peers configured for synchronization. |
[ERROR] OS: syncing from different ntp peers? (Severity: Warning) |
|
|
Payload blade has been started with the parameter NUMA enabled (Non-Uniform Memory Access). |
[ERROR] OS: Kernel printout: payload blade not started with numa=off (Severity: Warning) > CUDB31 DS10_1 initrd=netboot_initrd i8042.noaux i8042.nokbd i8042.nomux panic=10 console=tty0 console=ttyS0,115200 cluster=(type=payload,disk_cache=0,clean_rootfs=0) numa=off transparent_hugepage=madvise intel_idle.max_cstate=1 BOOT_IMAGE=vmlinuz ip=192.168.0.26:192.168.0.101:0.0.0.0:255.255.255.0 BOOTIF=01-90-38-09-8e-62-33 |
Contact the next level of support. |
|
A hardware error has been detected. |
[ERROR] kernel: Hardware Error found (Severity: Major) Jul 19 18:35:16 SC_2_1 kernel: [ 8757.942729] [Hardware Error]: Machine check events logged |
Contact the next level of support. |
4.4 Config
Table 4 shows config fault in different scenarios.
|
Fault |
Example Printout |
Action |
|---|---|---|
|
Error in custom cudb config checks. |
[ERROR] Config: Error in custom cudb config checks (Severity: Major) ERROR: cluster alarm -l -a alarm list is not empty Shows which config checks have failed. |
Contact the next level of support to investigate the issue. |
[ERROR] CONFIG: ldapfe log rotation not defined. |
Contact the next level of support. |
4.5 SWM
Table 5 shows SWM fault in different scenarios.
|
Fault |
Example Printout |
Action |
|---|---|---|
|
Multiple revisions from package: [package name]. |
[ERROR] SWM: Multiple revisions from package (Severity: Minor) CUDB_NODE_CONFIG-CXP9015320 |
Check cmw-repository-list on the node, and verify that the package shown in the printout has still more than one versions. Verify that there is no ongoing SW update or upgrade on the CUDB node, as the printout is normal in these cases. Contact the next level of support to investigate and resolve the issue.(1) |
4.6 Database Cluster
Table 6 shows database cluster faults in different scenarios.
|
Fault |
Example Printout |
Action |
|---|---|---|
|
Binlog is not written. |
[ERROR] MYSQL: Binlog is not written (Severity: Major) CUDB41 DS1_0 0 Shows the blade or VM where the binlog is not written. |
Run manual log collection and manual log analysis to verify that the problem is still present.(1) (2) |
|
Incorrect key file for table. |
[ERROR] MYSQL: Incorrect key file for table (Severity: Major) Sep 16 09:50:54 PL_2_5 mysqld: 110916 9:50:54 [ERROR] mysqld: Incorrect key file for table './mysql/ndb_binlog_index.MYI'; try to repair it Sep 16 09:50:54 PL_2_5 mysqld: 110916 9:50:54 [ERROR] mysqld: Incorrect key file for table './mysql/ndb_binlog_index.MYI'; try to repair it |
Contact the next level of support. |
|
Unstable NTP service. |
[ERROR] NDB: Time moved forward with 200 milliseconds (Severity: Warning) > Time moved forward with 200 milliseconds |
|
|
Data node has been crushed. |
[ERROR] NDB: There are new entries in ndb error dumps (Severity: Warning) > CUDB31 PL0 Current byte-offset of file-pointer is: 3063 |
Contact the next level of support. |
|
MYSQL cluster logs. |
[ERROR] MYSQL: Errors found in mysql cluster logs (Severity: Warning) > May 11 13:09:02 PL_2_3 mysqld: 2017-05-11 13:09:02 0 [Warning] TIMESTAMP with implicit DEFAULT value is deprecated. Please use --explicit_defaults_for_timestamp server option (see documentation for more details). |
This is a generic statement mysql cluster logs errors and can be provoked by any unexpected mysql behavior. If this printout appears without any other alarm or likely cause for the previous 12-hour monitoring period, contact the next level of support. |
4.7 LDAP
Table 7 shows LDAP faults in different scenarios.
|
Fault |
Example Printout |
Action |
|---|---|---|
|
Traffic bursts or the client is not able to handle responses fast enough. |
[ERROR] LDAP: Deferring operations found in ldap logs (Severity: Critical) > Sep 8 00:33:23 PL_2_16 slapd[23297]: connection_input: conn=30556 deferring operation: binding > Sep 8 00:36:40 PL_2_18 slapd[25017]: connection_input: conn=30623 deferring operation: binding |
Contact the next level of support to check the LDAP FE log files of the affected blades. |
|
Various errors found. |
[ERROR] LDAP: various errors found in LDAP logs (Severity: Critical) > May 12 00:17:29 PL_2_3 slapd[17513]: conn=2611 op=1 MODIFY RESULT tag=103 err=53 text=No available master replica for DSG 3 req=serv=csps,mscId=39,ou=multiSCs,CUDBNode=47,dc=ericsson,dc=com [ERROR] LDAP: err=52 errors found in LDAP logs (Severity: Critical) > May 12 01:14:46 PL_2_8 slapd[10961]: conn=6458 op=0 BIND RESULT tag=97 err=52 text=CUDB node 47 is temporarily out of service req=cn=manager,ou=ft,o=cudb,c=es |
Investigate the error codes in the log output.(1) |
|
[ERROR] LDAP: err=51 errors found in LDAP logs (Severity: Warning) > CUDB31 DS10_1 Aug 18 04:29:51 PL_2_3 slapd[10356]: conn=2155516 op=22 SEARCH RESULT tag=101 err=51 nentries=0 text=LDAP server overloaded in node 11 |
If the alert is produced frequently, contact the next level of support for a deeper analysis. (1) |
|
Unavailable. |
[ERROR] LDAP: err=52 errors found in LDAP logs (Severity: Critical) > May 12 01:14:46 PL_2_8 slapd[10961]: conn=6458 op=0 BIND RESULT tag=97 err=52 text=CUDB node 47 is temporarily out of service req=cn=manager,ou=ft,o=cudb,c=es |
|
|
Unwilling to perform. |
[ERROR] LDAP: err=53 errors found in LDAP logs (Severity: Critical) > May 12 00:17:29 PL_2_3 slapd[17513]: conn=2611 op=1 MODIFY RESULT tag=103 err=53 text=No available master replica for DSG 3 req=serv=csps,mscId=39,ou=multiSCs,CUDBNode=47,dc=ericsson,dc=com |
No special actions are expected from the LDAP client. CUDB is handling the internal fault. |
|
[ERROR] LDAP: err=80 errors found in LDAP logs Sep 13 10:20:39 PL_2_9 slapd[8882]: conn=305292 op=4686 RESULT tag=107 err=80 text=REDO buffers overloaded (increase RedoBuffer) Sep 13 10:20:40 PL_2_9 slapd[8882]: conn=310253 op=1875 RESULT tag=107 err=80 text=REDO buffers overloaded (increase RedoBuffer) Action: |
The possible cause of this error is the temporary unavailability of an internal component, or the temporary internal overload of a specific partition or node. Reduce the traffic or provisioning load to stabilize the system or speed up application and lower the number terminal retries rejected by CUDB with error 80.(2) |
|
REDO buffers overloaded. |
[ERROR] LDAP: err=80 errors with REDO buffers overloaded found in LDAP logs Sep 13 10:20:39 PL_2_9 slapd[8882]: conn=305292 op=4686 RESULT tag=107 err=80 text=REDO buffers overloaded (increase RedoBuffer) Sep 13 10:20:40 PL_2_9 slapd[8882]: conn=310253 op=1875 RESULT tag=107 err=80 text=REDO buffers overloaded (increase RedoBuffer) |
|
4.8 System Monitor
Table 8 shows System Monitor faults in different scenarios.
|
Fault |
Example Printout |
Action |
|---|---|---|
|
System Monitor logs indicate there were major errors. |
[ERROR] System Monitor logs indicate there were major errors (Severity: Major) > Apr 19 18:21:33 SC_2_1 SM[23399]: INFO Auxiliar connection with site 2 has changed to LOST > Apr 19 18:21:34 SC_2_1 SM[23399]: INFO Connection status between SM leader in site 1 (S1-N202-I1) and BC service in site 2 has change to LOST |
Run the cudbSystemStatus -b command to check if the problem persists. If yes, contact the next level of support. |
|
System Monitor logs indicate there were some warnings. |
[ERROR] System Monitor logs indicate there were some warnings [Towards site 2; Event count 14] (Severity: Warning) > Apr 20 01:10:25 SC_2_1 SM[23399]: INFO Auxiliar connection with site 2 has changed to SUSPENDED > Apr 20 01:10:26 SC_2_1 SM[23399]: INFO Auxiliar connection with site 2 has changed to RECONNECTED |
Run the cudbSystemStatus -b command to check if the problem persists. If yes, contact the next level of support. |
4.9 BC Server
Table 9 shows BC Server faults in different scenarios.
|
Fault |
Example Printout |
Action |
|---|---|---|
|
BC Server logs indicate there were minor errors. |
[ERROR] BC Server logs indicate there were minor errors (Severity: Minor) > Apr 20 09:25:37 SC_2_2 BC[myid:11]: INFO Got user-level KeeperException when processing sessionid:0xa52fbde985d104a type:ping cxid:0xfffffffffffffffe zxid:0xfffffffffffffffe txntype:unknown reqpath:n/a Error Path:null Error:KeeperErrorCode = Session moved |
It is recommended to contact the next level of support to investigate the issue. |
|
BC Server logs indicate there were major errors. |
[ERROR] BC Server logs indicate there were major errors (Severity: Major) > Feb 18 13:17:46 SC_2_2 BC[myid:11]: ERROR Unexpected exception causing shutdown while sock still open > Feb 18 13:17:47 SC_2_2 BC[myid:11]: INFO Shutting down > Feb 18 13:17:47 SC_2_2 BC[myid:11]: INFO Shutdown called |
It is recommended to contact the next level of support to investigate the issue. |
4.10 Master Unavailability
Table 10 shows Master Unavailability fault in different scenarios.
|
Fault |
Example Printout |
Action |
|---|---|---|
|
Master is not available for DSG. |
[ERROR] Error: Master is not available for DSG (Severity: Major) > Feb 12 14:45:06 SC_2_1 CS[14068]: Monitoring [ClusterSupervisor]: INFO - N203D0-6-BCClient: Enter getMasterStatus: masterless > Feb 12 14:45:06 SC_2_1 CS[14068]: Monitoring [ClusterSupervisor]: INFO - N203D0-6-BCClient: Information read in node /cudb/masterList/D0 -> masterless |
Run the cudbSystemStatus -R command to check if the problem persists. If yes, contact the next level of support. |
4.11 Software Platform
Table 11 shows Software Platform faults in different scenarios.
|
Fault |
Example Printout |
Action |
|---|---|---|
|
Not all SU have Operational State Enabled. |
[ERROR] SAF: Not all SU have Operational State Enabled (Severity: Warning) saAmfNodeOperState."safAmfNode=SC-1,safAmfCluster=myAmfCluster": Disabled |
Run the cudbHaState command to check if the problem persists. If yes, contact the next level of support. |
|
There are unassigned HA states. |
[ERROR] SAF: there are unassigned HA states (Severity: Warning) saAmfSISUHAState."safSu=SC-1,safSg=NWA,safApp=ERIC-ldews.pmcounters.sc"."safSi=ldews.pmcounters.sc-NWA-1"."safSi=NoRed8": unassigned(3) |
Run the cudbHaState command to check if the problem persists. If yes, contact the next level of support. |
|
One or more nodes are locked in the cluster. |
[ERROR] SAF: One or more node is locked in the cluster (Severity: Warning) saAmfNodeAdminState."safAmfNode=SC-1,safAmfCluster=myAmfCluster": Locked |
Run the cudbHaState command to check if the problem persists. If yes, contact the next level of support. |
|
There are active cluster alarms. |
[ERROR] SAF: There are active cluster alarms (Severity: Major) Node Hostname Severity Type Problem Information 1 SC_2_1 Major 2 Ethernet Bonding Bonding degraded on bond0 (link down on eth4) |
Run the cluster alarm -a -l command to check which alarms are in the system. If no reason is found for the alarms, contact the next level of support. |
4.12 Alarm
Table 12 shows ALARM faults in different scenarios.
|
Fault |
Example Printout |
Action |
|---|---|---|
|
There are active alarms. |
[ERROR] ALARM: There are active alarms (Severity: Critical) Timestamp First : Mon Nov 28 05:15:05 GMT 2016 Timestamp Last : Mon Nov 28 05:15:05 GMT 2016 Active Description : Root Login Failed @172.17.17.10 |
Run the fmactivealarms command to see all active alarms.(1) |
4.13 Log
Table 13 shows LOG faults in different scenarios.
|
Fault |
Example Printout |
Action |
|---|---|---|
|
Potentially harmful situation. |
[ERROR] LOG: New printouts with EMERG level (Severity: Warning) > May 12 07:48:35 SC_2_1 clsupervisor[16467] : Monitoring : EMERG - main: Failed to read configuration |
This is generic error message that can refer to any component. Refer to CUDB Node Logging Events for troubleshooting information. If the cause cannot be determined, contact the next level of support. |
|
Error conditions. |
[ERROR] LOG: New printouts with ERROR level (Severity: Warning) > May 11 13:08:56 SC_2_1 clsupervisor[13154] : Monitoring : INFO - N8D0-cl_control: testSQLNodeConnections failed for node 49 a1 on host 10.22.8.3 port 15000, user root with error: ErrCode: 2003 Err message: Can't connect to MySQL server on '10.22.8.3' (111) SQLState: HY000 |
This is generic error message that can refer to any component. Refer to CUDB Node Logging Events for troubleshooting information. If the cause cannot be determined, contact the next level of support. |
|
Component or subcomponent is unusable. |
[ERROR] LOG: New printouts with WARNING level (Severity: Warning) > May 11 13:08:57 SC_2_1 clsupervisor[13154] : Monitoring : WARNING - N8D0-cl_control: Clean stop procedure for Mysql server with ndb node id 49 failed on 10.22.8.3 with error: ErrCode: 2003 Err message: Can't connect to MySQL server on '10.22.8.3' (111) SQLState: HY000. Killing the server. |
This is generic error message that can refer to any component. Refer to CUDB Node Logging Events for troubleshooting information. If the cause cannot be determined, contact the next level of support. |
Reference List
- Preventive Maintenance, Logchecker Found Error(s)
- CUDB Node Preventive Maintenance
- CUDB Troubleshooting Guide
- CUDB Node Commands and Parameters
- Server Platform, Blade Replacement
- CUDB LDAP Interwork Description
- CUDB LDAP Data Access
- Data Collection Guideline for CUDB
- CUDB Node Fault Management Configuration Guide
- CUDB Node Logging Events
- CUDB Glossary of Terms and Acronyms

Contents