Concept of Cluster: Prepared By: Vinutha C
Concept of Cluster: Prepared By: Vinutha C
Terminologies:
Nodes – Nodes are the individual physical servers, which hold all the files and data.
Shared Drive – It is an external disk drive physically plugged into all nodes in the cluster. IT
store data, applications, resources and services..
Public Network: It is a network through which client access to the Nodes of the Cluster
Internal network: It is a network via which the inter node communication occurs.
Quorum: It is a resource which stores the cluster configuration, database, checkpoints and
recovery log. Usually a separate disk is dedicated for a quorum.
Virtual Server – A Virtual server is a group that has an IP address, a network name and the
applications to be accessed by the clients of the Virtual Server
Resources implies –
1. Physical hardware devices – disk drives,
2. Logical entities – IP addresses, network names, applications and services.
Failover – It is a process of moving a group of resources from one node to another manually or
automatic.
Failback – It is a process of returning a group of resources from one node to the node on which it
was running before the failover occurred
The data on the quorum resource includes a set of cluster configuration information plus records
(sometimes called checkpoints) of the most recent changes made to that configuration. A node
coming online after an outage can use the quorum resource as the definitive source for recent
changes in the configuration.
The sections that follow describe the three different types of server clusters:
Single Quorum Device Cluster (standard quorum cluster)
Because single quorum device clusters are the most widely used cluster, this Technical
Reference focuses on this type of cluster.
1. Active/Active Cluster:
In an Active/Active cluster, when one node fails or is taken offline, the other node in the cluster
takes over for the failed node. There is no concept of a primary or backup node: both nodes
provide some service, should one of these nodes fail, the other must also assume the failed node's
services.However, because the failover causes the remaining node to take on additional
processing operations, the overall performance of your Exchange 2000 cluster or Exchange 2003
cluster may be reduced. For performance, availability, and scalability reasons, Active/Passive
cluster configurations are a better option than Active/Active configurations.
In a typical Active/Active cluster configuration, each node owns one Exchange virtual server.
However, if the number of Exchange virtual servers in the cluster is greater than or equal to the
number of nodes in the cluster, and even when all the Exchange virtual servers are owned by
only one node in the cluster, the cluster is still considered an Active/Active cluster.
2. Active/Passive Cluster:
In case of Active Passive, one node owns the services, the other one remains inoperative. When
the primary node fails, the secondary or backup node takes the resources and reactivates the
services, while the ex-primary remains in turn inoperative. This is a configuration where only
one node is operative at any point of time in case of two node cluster.
High availability (HA) cluster computing is often used in this way, to create a redundant
network which will be accessible to users at all times, with failsafes in case parts of the
cluster breakdown.
Load balancing clusters are designed to address a large load of incoming requests,
coordinating requests in a way which will maximize efficiency and usability.
Load-balancing clusters
Load-balancing when multiple computers are linked together to share computational
workload or function as a single virtual computer.
Logically, from the user side, they are multiple machines, but function as a single virtual
machine. Requests initiated from the user are managed by, and distributed among, all the
standalone computers to form a cluster. This results in balanced computational work
among different machines, improving the performance of the cluster system.
Compute clusters
Clusters are used primarily for computational purposes, rather than handling IO-oriented
operations such as web service or databases. For instance, a cluster might support
computational simulations of weather or vehicle crashes.
The primary distinction within compute clusters is how tightly-coupled the individual
nodes are. For instance, a single compute job may require frequent communication
among nodes - this implies that the cluster shares a dedicated network, is densely located,
and probably has homogenous nodes. This cluster design is usually referred to as
Beowulf Cluster.
The other extreme is where a compute job uses one or few nodes, and needs little or no
inter-node communication. This category is sometimes called "Grid" computing. Tightly-
coupled compute clusters are designed for work that might traditionally have been called
"supercomputing
Cluster computing can also be used as a relatively low-cost form of parallel processing
for scientific and other applications that lend themselves to parallel operations. An early
and well-known example was the Beowulf project in which a number of off-the-shelf
PCs were used to form a cluster for scientific applications.
References:
www.microsoft.com