1. What does MPI stand for in the context of Parallel and Distributed Computing?
(A) Multiple Processor Interface
(B) Message Passing Interface
(C) Multi-Threaded Processing Interface
(D) Modular Parallel Integration
2. In MPI, which function is commonly used to send a message from one process to another?
(A) MPI_Receive
(B) MPI_Send
(C) MPI_Comm_rank
(D) MPI_Bcast
3. What is the purpose of MPI_Comm_size function in MPI?
(A) To initialize MPI communication
(B) To get the rank of the process
(C) To determine the size of the communicator
(D) To finalize MPI communication
4. Which MPI function is used to initialize the MPI environment?
(A) MPI_Initialize
(B) MPI_Init
(C) MPI_Comm_init
(D) MPI_Start
5. In MPI, what does the term “rank” refer to?
(A) The size of the communicator
(B) The process identifier
(C) The message size
(D) The type of data being sent
6. What MPI function is used to obtain the rank of the calling process within a communicator?
(A) MPI_Rank
(B) MPI_Get_rank
(C) MPI_Comm_rank
(D) MPI_Process_rank
7. In MPI, which function is used for collective communication to broadcast data from one process to all others?
(A) MPI_Scatter
(B) MPI_Gather
(C) MPI_Bcast
(D) MPI_Reduce
8. What is the purpose of MPI_Finalize function?
(A) To close MPI communication channels
(B) To signal the end of the MPI program
(C) To synchronize MPI processes
(D) To free up memory allocated for MPI communication
9. Which MPI function is used for blocking point-to-point communication to receive a message?
(A) MPI_Recv
(B) MPI_Send
(C) MPI_Irecv
(D) MPI_Isend
10. In MPI, what is the purpose of MPI_Barrier function?
(A) To initialize MPI communication
(B) To synchronize processes within a communicator
(C) To broadcast data from one process to all others
(D) To reduce data across all processes
11. What MPI function is used for non-blocking point-to-point communication to send a message?
(A) MPI_Recv
(B) MPI_Send
(C) MPI_Irecv
(D) MPI_Isend
12. In MPI, what is the purpose of MPI_Reduce function?
(A) To initialize MPI communication
(B) To synchronize processes within a communicator
(C) To reduce data across all processes
(D) To broadcast data from one process to all others
13. Which MPI function is used for gathering data from all processes within a communicator?
(A) MPI_Scatter
(B) MPI_Gather
(C) MPI_Bcast
(D) MPI_Reduce
14. In MPI, what does the term “communicator” refer to?
(A) The message size
(B) The process identifier
(C) A group of processes that can communicate with each other
(D) The size of the communicator
15. What is the purpose of MPI_Sendrecv function in MPI?
(A) To send and receive messages concurrently
(B) To synchronize processes within a communicator
(C) To broadcast data from one process to all others
(D) To reduce data across all processes
16. In MPI, what is the primary purpose of MPI_Scatter function?
(A) To initialize MPI communication
(B) To synchronize processes within a communicator
(C) To scatter data from one process to all others
(D) To broadcast data from one process to all others
17. What MPI function is used for non-blocking point-to-point communication to receive a message?
(A) MPI_Recv
(B) MPI_Send
(C) MPI_Irecv
(D) MPI_Isend
18. In MPI, which function is used to create a new communicator?
(A) MPI_Comm_split
(B) MPI_Comm_rank
(C) MPI_Comm_create
(D) MPI_Comm_dup
19. What is the purpose of MPI_Gather function in MPI?
(A) To scatter data from one process to all others
(B) To gather data from all processes within a communicator
(C) To broadcast data from one process to all others
(D) To reduce data across all processes
20. In MPI, what is the purpose of MPI_Comm_rank function?
(A) To get the rank of the process
(B) To determine the size of the communicator
(C) To initialize MPI communication
(D) To finalize MPI communication
21. Which MPI function is used for collective communication to scatter data from one process to all others?
(A) MPI_Scatter
(B) MPI_Gather
(C) MPI_Bcast
(D) MPI_Reduce
22. In MPI, what does the term “blocking communication” mean?
(A) Processes are not allowed to communicate
(B) Processes can only communicate within a communicator
(C) Processes are synchronized until the communication completes
(D) Processes can communicate concurrently without synchronization
23. What is the purpose of MPI_Comm_dup function in MPI?
(A) To create a new communicator
(B) To duplicate an existing communicator
(C) To synchronize processes within a communicator
(D) To finalize MPI communication
24. In MPI, what does the term “non-blocking communication” mean?
(A) Processes are not allowed to communicate
(B) Processes can only communicate within a communicator
(C) Processes can communicate concurrently