Parallel Processing and Parallel Algorithms: Theory and ComputationSpringer Science & Business Media, 10 դեկ, 1999 թ. - 566 էջ Motivation It is now possible to build powerful single-processor and multiprocessor systems and use them efficiently for data processing, which has seen an explosive ex pansion in many areas of computer science and engineering. One approach to meeting the performance requirements of the applications has been to utilize the most powerful single-processor system that is available. When such a system does not provide the performance requirements, pipelined and parallel process ing structures can be employed. The concept of parallel processing is a depar ture from sequential processing. In sequential computation one processor is in volved and performs one operation at a time. On the other hand, in parallel computation several processors cooperate to solve a problem, which reduces computing time because several operations can be carried out simultaneously. Using several processors that work together on a given computation illustrates a new paradigm in computer problem solving which is completely different from sequential processing. From the practical point of view, this provides sufficient justification to investigate the concept of parallel processing and related issues, such as parallel algorithms. Parallel processing involves utilizing several factors, such as parallel architectures, parallel algorithms, parallel programming lan guages and performance analysis, which are strongly interrelated. In general, four steps are involved in performing a computational problem in parallel. The first step is to understand the nature of computations in the specific application domain. |
Բովանդակություն
II | 3 |
IV | 4 |
V | 6 |
VI | 8 |
VII | 15 |
VIII | 25 |
IX | 32 |
X | 37 |
LXXVIII | 264 |
LXXIX | 277 |
LXXXI | 286 |
LXXXII | 296 |
LXXXIII | 304 |
LXXXIV | 308 |
LXXXV | 314 |
LXXXVI | 317 |
XI | 40 |
XII | 41 |
XIII | 45 |
XIV | 52 |
XV | 56 |
XVI | 59 |
XVIII | 60 |
XIX | 68 |
XX | 69 |
XXI | 70 |
XXII | 72 |
XXIII | 73 |
XXIV | 74 |
XXVI | 76 |
XXVII | 77 |
XXVIII | 78 |
XXIX | 79 |
XXX | 81 |
XXXI | 84 |
XXXII | 85 |
XXXIII | 92 |
XXXIV | 102 |
XXXV | 104 |
XXXVI | 107 |
XXXVII | 111 |
XXXIX | 114 |
XL | 115 |
XLI | 119 |
XLII | 122 |
XLIII | 125 |
XLIV | 127 |
XLV | 128 |
XLVII | 130 |
XLVIII | 131 |
XLIX | 132 |
L | 133 |
LI | 135 |
LII | 139 |
LIV | 142 |
LV | 151 |
LVI | 167 |
LVII | 179 |
LVIII | 193 |
LIX | 206 |
LX | 216 |
LXI | 219 |
LXIII | 222 |
LXIV | 224 |
LXV | 228 |
LXVI | 229 |
LXVII | 230 |
LXVIII | 232 |
LXIX | 234 |
LXX | 235 |
LXXI | 236 |
LXXII | 239 |
LXXIII | 241 |
LXXIV | 246 |
LXXV | 255 |
LXXVI | 261 |
LXXXVII | 321 |
LXXXIX | 322 |
XC | 326 |
XCI | 331 |
XCII | 335 |
XCIII | 338 |
XCIV | 343 |
XCV | 351 |
XCVI | 353 |
XCVII | 357 |
XCIX | 359 |
C | 365 |
CI | 369 |
CII | 372 |
CIII | 376 |
CIV | 382 |
CV | 392 |
CVI | 394 |
CVII | 399 |
CVIII | 407 |
CIX | 409 |
CX | 413 |
CXII | 414 |
CXIII | 416 |
CXIV | 420 |
CXV | 425 |
CXVI | 427 |
CXVII | 431 |
CXVIII | 441 |
CXX | 443 |
CXXI | 455 |
CXXII | 466 |
CXXIII | 476 |
CXXIV | 479 |
CXXV | 482 |
CXXVI | 484 |
CXXVII | 485 |
CXXVIII | 488 |
CXXIX | 491 |
CXXX | 493 |
CXXXI | 496 |
CXXXII | 498 |
CXXXIII | 499 |
CXXXIV | 503 |
CXXXVII | 504 |
CXXXVIII | 506 |
CXXXIX | 507 |
CXL | 510 |
CXLI | 512 |
CXLII | 515 |
CXLIII | 520 |
CXLIV | 526 |
CXLV | 529 |
CXLVI | 533 |
CXLVII | 534 |
CXLVIII | 537 |
CXLIX | 549 |
553 | |
557 | |
Այլ խմբագրություններ - View all
Parallel Processing and Parallel Algorithms: Theory and Computation Seyed H Roosta Դիտել հնարավոր չէ - 2012 |
Common terms and phrases
adjacency matrix associated breadth-first search buffer C-Linda called chare color communication compiler concurrent consists Control Unit created data flow architectures data parallel data type defined depth-first search distributed edge endfor endif evaluation example Fortran functional programming graph G hypercube implementation indicates input instruction integer interconnection network iteration loop machine mapping MasPar matching matrix multiplication meaning memory location mesh message-passing method MIMD MIMD computers minimum spanning tree Modula-2 module multiprocessor node number of processors operating system output parallel algorithm parallel architectures parallel computation parallel processing parallel programming partitioning performed pipeline prefix sums primitives problem proc procedure processing elements programming language Prolog provides queue recursive represented result semaphore sequence sequential shared shared-memory shortest path SIMD SISAL solution specify speedup statement step synchronization systolic array task thread tion tuple space utilizing variable vector vertex vertices
Վկայակոչումներ այս գրքի մասին
Cit 2004: Proceedings Of The 7th International Conference On Information ... Gautam Das,V.P. Gulati Դիտել հնարավոր չէ - 2004 |