Parallel Computing

Parallel computing is a type of computer structure in which multiple processes are executed at the same time. It is the opposite of serial computing, in which one task is broken down into a set of instructions that are processed individually in sequential order. Parallel computing is closely related to concurrent computing, but they are distinct concepts; with the former, all of the computational tasks are interrelated, while the latter deals with processes that are unrelated or significantly varied in nature.

Parallel computing was originally reserved for high-performance computing environments, but it has become the primary framework for all computer architectures. In fact, it’s rare to find a server, laptop computer, smartphone, or other modern device that does not contain a multi-core processor. This is because parallel computing makes all computer processes more efficient, thereby saving time and money. It also enables computers to solve exceptionally large and complex problems that would otherwise be impossible to process.

Types of parallel computing approaches

There are four main types of parallel computing:

  • Bit-level parallelism, which increases the amount of information the processor can handle, effectively decreasing the number of instructions that are required to complete a single task
  • Instruction-level parallelism, which executes multiple instruction sequences at the same time, either through hardware (dynamic parallelism) or software (static parallelism)
  • Task parallelism, which performs multiple simultaneous tasks on the same set of data
  • Superword level parallelism, which exploits parallelism of inline code through vectorization techniques

Types of parallel computing architecture

Parallel computers can be classified based on four types of architecture:

  • Multicore computing, in which multiple processing units (called cores) are housed on the same chip
  • Symmetric multiprocessing, in which a bus connects multiple separate but identical processing units
  • Distributed computing, in which processing units exist in separate computer networks but coordinate efforts via HTTP, RPC, or similar protocols
  • Massively parallel computing, in which a large volume of networked processing units (100+) are housed in a single computer, usually a supercomputer

 

Related Link

9 Parallel Processing Examples & Applications

Kaiti Norton
Kaiti Norton is a Nashville-based Content Writer for TechnologyAdvice, a full-service B2B media company. She is passionate about helping brands build genuine connections with their customers through relatable, research-based content. When she's not writing about technology, she's sharing her musings about fashion, cats, books, and skincare on her blog.

Top Articles

The Complete List of 1500+ Common Text Abbreviations & Acronyms

Text Abbreviations reviewed by Web Webster   From A3 to ZZZ we list 1,559 SMS, online chat, and text abbreviations to help you translate and understand...

Windows Operating System History & Versions

The Windows operating system (Windows OS) refers to a family of operating systems developed by Microsoft Corporation. We look at the history of Windows...

How to Create a Website Shortcut on Your Desktop

Website Shortcut on Your Desktop reviewed by Web Webster   This Webopedia guide will show you how to create a website shortcut on your desktop using...

Generations of Computers (1st to 5th)

Reviewed by Web Webster Learn about each of the 5 generations of computers and major technology developments that have led to the computing devices that...

Pipeline

A pipeline is a figurative measurement of client or customer progress through a...

Direct Marketing

Direct marketing is a strategy for selling products and services directly...

Compact Disc

A compact disc (CD) is a polycarbonate disc measuring 120 mm in diameter...