In computing, a process that is blocked is waiting for some event, such as a resource becoming available or the completion of an I/O operation.[1] Once the event occurs for which the process is waiting ("is blocked on"), the process is advanced from blocked state to an imminent one, such as runnable.

In a multitasking computer system, individual tasks, or threads of execution, must share the resources of the system. Shared resources include: the CPU, network and network interfaces, memory and disk.

When one task is using a resource, it is generally not possible, or desirable, for another task to access it. The techniques of mutual exclusion are used to prevent this concurrent use. When the other task is blocked, it is unable to execute until the first task has finished using the shared resource.

Programming languages and scheduling algorithms are designed to minimize the over-all effect of blocking. A process that blocks may prevent local work-tasks from progressing. In this case "blocking" often is seen as not wanted.[2] However, such work-tasks may instead have been assigned to independent processes, where halting one has little to no effect on the others, since scheduling will continue. An example is "blocking on a channel" where passively waiting for the other part (i.e. no polling or spin loop) is part of the semantics of channels.[3] Correctly engineered, any of these may be used to implement reactive systems.[clarification needed]

Deadlock means that processes pathologically wait for each other in a circle. As such it is not directly associated with blocking.

See also

edit

References

edit
  1. ^ Stallings, William (2004). Operating Systems: Internals and Design Principles (5th ed.). Prentice Hall. ISBN 978-0131479548.
  2. ^ C++ and Beyond 2012: Herb Sutter - C++ Concurrency
  3. ^ Rob Pike (2012-07-02). Go Concurrency Patterns. Google I/O 2012. Google for Developers.
  NODES
INTERN 1
Note 1