Spec-Zone .ru
спецификации, руководства, описания, API
|
This section provides a simplified outline of the steps involved when MySQL Cluster data nodes are started. More
complete information can be found in NDB
Internals Guide.
These phases are the same as those reported in the output from the
command in the management client (see Section 17.5.2, "Commands in the MySQL
Cluster Management Client"). These start phases are also reported in the node_id
STATUSstart_phase
column of the ndbinfo.nodes
table.
Start types. There are several different startup types and modes, as shown in the following list:
Initial start. The cluster starts with a clean file system on all data
nodes. This occurs either when the cluster started for the very first time, or when all data nodes are
restarted using the --initial
option.
Disk Data files are not removed when restarting a node using --initial
.
System restart. The cluster starts and reads data stored in the data nodes. This occurs when the cluster has been shut down after having been in use, when it is desired for the cluster to resume operations from the point where it left off.
Node restart. This is the online restart of a cluster node while the cluster itself is running.
Initial node restart. This is the same as a node restart, except that the node is reinitialized and started with a clean file system.
Setup and initialization (phase -1). Prior to startup, each data node (ndbd process) must be initialized. Initialization consists of the following steps:
Obtain a node ID
Fetch configuration data
Allocate ports to be used for inter-node communications
Allocate memory according to settings obtained from the configuration file
When a data node or SQL node first connects to the management node, it reserves a cluster node ID. To make sure that no other node allocates the same node ID, this ID is retained until the node has managed to connect to the cluster and at least one ndbd reports that this node is connected. This retention of the node ID is guarded by the connection between the node in question and ndb_mgmd.
After each data node has been initialized, the cluster startup process can proceed. The stages which the cluster goes through during this process are listed here:
Phase 0. The NDBFS
and NDBCNTR
blocks start (see NDB
Kernel Blocks--initial
option.
Phase 1. In this stage, all remaining NDB
kernel blocks are started. MySQL Cluster connections are set up,
inter-block communications are established, and heartbeats are started. In the case of a node restart,
API node connections are also checked.
When one or more nodes hang in Phase 1 while the remaining node or nodes hang in Phase 2, this often indicates network problems. One possible cause of such issues is one or more cluster hosts having multiple network interfaces. Another common source of problems causing this condition is the blocking of TCP/IP ports needed for communications between cluster nodes. In the latter case, this is often due to a misconfigured firewall.
Phase 2. The NDBCNTR
kernel block checks the
states of all existing nodes. The master node is chosen, and the cluster schema file is initialized.
Phase 3. The DBLQH
and DBTC
kernel blocks set up communications between them. The startup type is determined; if this is a restart,
the DBDIH
block obtains permission to perform the restart.
Phase 4. For an initial start or initial node restart, the redo log files
are created. The number of these files is equal to NoOfFragmentLogFiles
.
For a system restart:
Read schema or schemas.
Read data from the local checkpoint.
Apply all redo information until the latest restorable global checkpoint has been reached.
For a node restart, find the tail of the redo log.
Phase 5. Most of the database-related portion of a data node start is performed during this phase. For an initial start or system restart, a local checkpoint is executed, followed by a global checkpoint. Periodic checks of memory usage begin during this phase, and any required node takeovers are performed.
Phase 6. In this phase, node groups are defined and set up.
Phase 7. The arbitrator node is selected and begins to function. The next
backup ID is set, as is the backup disk write speed. Nodes reaching this start phase are marked as Started
. It is now possible for API nodes (including SQL nodes) to
connect to the cluster.
Phase 8. If this is a system restart, all indexes are rebuilt (by DBDIH
).
Phase 9. The node internal startup variables are reset.
Phase 100 (OBSOLETE). Formerly, it was at this point during a node restart or initial node restart that API nodes could connect to the node and begin to receive events. Currently, this phase is empty.
Phase 101. At this point in a node restart or initial node restart, event
delivery is handed over to the node joining the cluster. The newly-joined node takes over responsibility
for delivering its primary data to subscribers. This phase is also referred to as SUMA
handover phase.
After this process is completed for an initial start or system restart, transaction handling is enabled. For a node restart or initial node restart, completion of the startup process means that the node may now act as a transaction coordinator.