MCQs | High Performance Computing (HPC)/High Performance Computing (HPC) MCQ Set 2 Sample Test,Sample questions

Question:
A good mapping does not depends on which following factor

1. knowledge of task sizes

2.the size of data associated with tasks

3.characteristics of inter-task interactions

4.task overhead


Question:
All nodes collects _____ message corresponding to √p nodes to their respectively

1. √p

2.p

3. p+1

4. p-1


Question:
All-to-all personalized communication can be used in ____

1.fourier transform

2.matrix transpose

3.sample sort

4.All of the above


Question:
All-to-one communication (reduction) is the dual of ______ broadcast.

1.all-to-all

2. one-to-all

3.one-to-one

4.all-to-one


Question:
Cost Analysis on a mesh is

1. A. 2ts(sqrt(p) + 1) + twm(p - 1)

2.2tw(sqrt(p) + 1) + tsm(p - 1)

3.2tw(sqrt(p) - 1) + tsm(p - 1)

4.2ts(sqrt(p) - 1) + twm(p - 1)


Question:
Goal of good algorithm is to implement commonly used _____ pattern.

1. communication

2.interaction

3.parallel

4.regular


Question:
If we port algorithm to higher dimemsional network it would cause

1.error

2.contention

3.recursion

4.none


Question:
In ------------task are defined before starting the execution of the algorithmting?

1. dynamic task

2.static task

3.regular task

4.one way task


Question:
In a broadcast and reduction on a balanced binary tree reduction is done in ______

1.recursive order

2. straight order

3.vertical order

4.parallel order


Question:
In one -to- all broadcast there is

1. divide and conquer type algorithm

2.sorting type algorithm

3.searching type algorithm

4. simple algorithm


Question:
In Scatter Operation on Hypercube, on each step, the size of the messages communicated is ____

1. tripled

2.halved

3.doubled

4.no change


Question:
In the scatter operation ____ node send message to every other node

1.single

2.double

3.triple

4.none


Question:
In the second phase of 2D Mesh All to All, the message size is ___

1.m

2.p*sqrt(m)

3.p

4.m*sqrt(p)


Question:
In this decomposition problem decomposition goes hand in hand with its execution

1.data decomposition

2. recursive decomposition

3.explorative decomposition

4.speculative decomposition


Question:
It is not possible to port ____ for higher dimensional network

1.algorithm

2.hypercube

3.both

4.none


Question:
Reduction can be used to find the sum, product, maximum, minimum of _____ of numbers.

1.tuple

2.list

3.sets

4.all of above


Question:
subsets of processes in ______ interaction.

1.global

2.local

3.wide

4.variable


Question:
threads being block altogether and being executed in the sets of 32 threads called a

1. thread block

2.32 thread

3.32 block

4.unit block


Question:
When the topological sort of a graph is unique?

1.when there exists a hamiltonian path in the graph

2.in the presence of multiple nodes with indegree 0

3.in the presence of single node with indegree 0

4.in the presence of single node with outdegree 0


Question:
Which of the following is not a form of parallelism supported by CUDA

1.vector parallelism - floating point computations are executed in parallel on wide vector units

2. thread level task parallelism - different threads execute a different tasks

3.block and grid level parallelism - different blocks or grids execute different tasks

4.data parallelism - different threads and blocks process different parts of data in memory


Question:
which of the following is not a granularity type

1.course grain

2.large grain

3.medium grain

4.fine grain


Question:
which of the following is not an example of explorative decomposition

1.n queens problem

2.15 puzzal problem

3. tic tac toe

4.quick sort


Question:
__ can be performed in an identical fashion by inverting the process.

1.recursive doubling

2.reduction

3.broadcast

4.none of these


Question:
accumulate results and send with the same pattern is...

1.broadcast

2.naive approach

3.recursive doubling

4.reduction symmetric


Question:
All processes participate in a single ______ interaction operation.

1.global

2.local

3.wide

4.variable


Question:
all processes that have the data can send it again is

1.recursive doubling

2. naive approach

3.reduction

4.All of the above


Question:
All-to-all broadcast algorithm for the 2D mesh is based on the

1. linear array algorithm

2. ring algorithm

3.both

4.none


Question:
blocking optimization is used to improve temmporal locality for reduce

1. hit miss

2.misses

3.hit rate

4.cache misses


Question:
Broadcast and reduction operations on a mesh is performed

1. along the rows

2.along the columns

3.both a and b concurrently

4.none of these


Question:
Communication between two directly link nodes

1.cut-through routing

2.store-and-forward routing

3.nearest neighbour communication

4.none


Question:
Cost Analysis on a ring is

1. (ts + twm)(p - 1)

2. (ts - twm)(p + 1)

3. (tw + tsm)(p - 1)

4.(tw - tsm)(p + 1)


Question:
CUDA thought that 'unifying theme' of every form of parallelism is

1.cda thread

2.pta thread

3.cuda thread

4.cud thread


Question:
Each node first sends to one of its neighbours the data it need to....

1. broadcast

2.identify

3.verify

4.none


Question:
efficiency of data parallel algorithm depends on the

1.efficient implementation of the algorithm

2.efficient implementation of the operation

3.both

4.none


Question:
every node has to know when to communicate that is

1.call the procedure

2.call for broadcast

3. call for communication

4.call the congestion


Question:
every node on the linear array has the data and broadcast on the columns with the linear array algorithm in _____

1. parallel

2.vertical

3.horizontal

4.all


Question:
For sake of simplicity, the number of nodes is a power of

1.1

2.2

3.3

4.4


Question:
Generalization of broadcast in Which each processor is

1.source as well as destination

2.only source

3.only destination

4.none


Question:
Group communication operations are built using which primitives?

1.one to all

2.all to all

3.point to point

4.none of these


Question:
if "X" is the message to broadcast it initially resides at the source node

1.1

2.2

3.8

4.0


Question:
In a balanced binary tree processing nodes is equal to

1.. leaves

2. number of elemnts

3.branch

4.none


Question:
In a eight node ring, node ____ is source of broadcast

1.1

2.2

3.8

4.0


Question:
In All to All on Hypercube, The size of the message to be transmitted at the next step is ____ by concatenating the received message with their current data

1.doubled

2.tripled

3.halfed

4.no change


Question:
In collective communication operations, collective means

1.involve group of processors

2.involve group of algorithms

3. involve group of variables

4.none of these


Question:
In task dependency graph longest directed path between any pair of start and finish node is called as --------------

1.. total work

2. critical path

3.task path

4.task path


Question:
In the first phase of 2D Mesh All to All, the message size is ___

1. p

2.m*sqrt(p)

3.m

4.p*sqrt(m) discuss


Question:
kernel may contain only host code

1.a code known as grid which runs on GPU consisting of a set of A. 32 thread

2.unit block

3.32 block

4. thread block


Question:
logical operators used in algorithm are

1. xor

2.and

3.both

4.none


Question:
Nides with zero in i least significant bits participate in _______

1.algorithm

2.broadcast

3.communication

4.searching


Question:
one to all broadcast use

1.recursive doubling

2. simple algorithm

3.both

4.none


Question:
One-to-All Personalized Communication operation is commonly called ___

1.gather operation

2.concatenation

3. scatter operation

4.none


Question:
only connections between single pairs of nodes are used at a time is

1. good utilization

2.poor utilization

3. massive utilization

4.medium utilization


Question:
Renaming relative to the source is _____ the source.

1. xor

2.xnor

3.and

4.nand


Question:
Renaming relative to the source is _____ the source.

1.xor

2.xnor

3.and

4.nand


Question:
Similar communication pattern to all-to-all broadcast except in the_____

1. reverse order

2. parallel order

3.straight order

4.vertical order


Question:
source ____ is bottleneck.

1.process

2.algorithm

3.list

4.tuple


Question:
Task dependency graph is ------------------

1.directed

2.undirected

3.directed acyclic

4.undirected acyclic


Question:
The ____ do not snoop the messages going through them.

1. nodes

2.variables

3.tuple

4.list


Question:
The algorithm terminates in _____ steps

1.p

2.p+1

3.p+2

4. p-1


Question:
The all-to-all broadcast on Hypercube needs ____ steps

1.p

2.sqrt(p) - 1

3.log p

4.none


Question:
The dual of all-to-all broadcast is

1.all-to-all reduction

2.all-to-one reduction

3.both

4.none


Question:
The dual of the scatter operation is the

1.concatenation

2.gather operation

3.both

4.none


Question:
The gather Operation is exactly the inverse of _____

1.scatter operation

2. recursion operation

3.execution

4.none


Question:
the procedure is disturbed and require only point-to-point _______

1.synchronization

2.communication

3.both

4.none


Question:
the procedure is disturbed and require only point-to-point _______

1.synchronization

2.communication

3.both

4.none


Question:
The processors compute ______ product of the vector element and the loval matrix

1.local

2.global

3.both

4.none


Question:
The second communication phase is a columnwise ______ broadcast of consolidated

1.all-to-all

2.one -to-all

3.all-to-one

4.point-to-point


Question:
The style of parallelism supported on GPUs is best described as

1.misd - multiple instruction single data

2.simt - single instruction multiple thread

3.sisd - single instruction single data

4.mimd


Question:
using different links every time and forwarding in parallel again is

1. better for congestion

2.better for reduction

3.better for communication

4.better for algorithm


Question:
What is a high performance multi-core processor that can be used to accelerate a wide variety of applications using parallel computing.

1.cpu

2.dsp

3.gpu

4.clu


Question:
Which is also called "Total Exchange" ?

1.all-to-all broadcast

2.all-to-all personalized communication

3.all-to-one reduction

4.none


Question:
Which is known as Broadcast?

1.one-to-one

2.one-to-all

3.all-to-all

4.all-to-one


Question:
Which is known as Reduction?

1.. all-to-one

2.all-to-all

3. one-to-one

4.one-to-all


Question:
Which of the following correctly describes a GPU kernel

1. a kernel may contain a mix of host and gpu code

2.all thread blocks involved in the same computation use the same kernel

3. a kernel is part of the gpus internal micro-operating system, allowing it to act as in independent host

4.kernel may contain only host code


Question:
which of the following is a an example of data decomposition

1.matrix multiplication

2. merge sort

3.quick sort

4.15 puzzal


Question:
which of the following is not a granularity type

1. course grain

2. large grain

3. medium grain

4.fine grain


Question:
which of the following is not an parallel algorithm model

1.data parallel model

2.task graph model

3.task model

4.work pool model


Question:
which of the following is not the array distribution method of data partitioning

1.block

2.cyclic

3.block cyclic

4.chunk


Question:
which problems can be handled by recursive decomposition

1.backtracking

2.greedy method

3.divide and conquer problem

4.branch and bound


Question:
wimpleat way to send p-1 messages from source to the other p-1 processors

1.algorithm

2.communication

3.concurrency

4.receiver


More MCQS

  1. Green Computing (GC)
  2. High Performance Computing (HPC) MCQ Set 1
  3. High Performance Computing (HPC) MCQ Set 2
Search