View on GitHub

Notes

reference notes

Parallel Computing Architecture

Flynn’s Classical Taxonomy

Flynn’s Taxonomy is a widely used classification system for parallel computers, first introduced in 1966. It categorizes parallel computer architectures based on two independent dimensions: Instruction and Data, each having two possible states: Single or Multiple.

Flynn Taxonomy - Categories:

categories

Taxonomy of Parallel Processor Architectures:

2

Single Instruction, Single Data (SISD)

3

Description:

Single Instruction, Multiple Data (SIMD)

4

Description:

Multiple Instruction, Single Data (MISD)

5

Description:

Multiple Instruction, Multiple Data (MIMD)

6

Description:

Shared Memory

Symmetric Multiprocessor (SMP)

Advantages:

Organization: 7

Key design issues:

Nonuniform Memory Access (NUMA)

NUMA (Nonuniform Memory Access) is a parallel computing architecture often created by physically linking two or more SMPs (Symmetric Multiprocessors). In NUMA, all processors have access to all parts of main memory, but the access time depends on the location of the memory.

Characteristics:

8

Design Issues:

Solutions:

Pros:

Cons:

Distributed Memory

Distributed Memory is an approach where a group of interconnected computers work together as a unified computing resource. Each computer in the cluster operates independently and has its own local memory, which is not shared. CPUs in this architecture need explicit access to data in other processors, which must be defined by the programmer.

9

Advantages:

Disadvantages:

Cluster Configuration

Clusters can be configured in various ways:

  1. Standby server with no shared disk: Two-node interconnected setup with high-speed links, LAN, and WAN communication.

  2. Shared disk: Two-node interconnected setup with high-speed links, LAN, WAN communication, and a disk subsystem such as a RAID system.

Clustering Methods:

Clusters compared to SMP:

Characteristic Clusters SMP (Symmetric Multiprocessor)
Number of Processors Multiple processors Multiple processors
Configuration Need to deal with different computer range and specifications Easier to manage and configure
Physical Space More space to allocate to each node Located on computer motherboard
Power Consumption Power required by each node Power required by each processor
Stability Not stable – depends on connection and node’s availability Stable and established
Scalability Big and expandable Small-scale/single computing

Hybrid Distributed Shared Memory

This architecture applies both shared and distributed memory architectures. It combines a shared memory component (such as a shared memory machine or GPUs) with a distributed memory component (networking of multiple shared memory/GPU machines). Each node in the distributed memory component only knows about its own memory, and network communication handles data transfer between nodes.