The Resource High-performance computing : paradigm and infrastructure, edited by Laurence T. Yang, Minyi Guo
High-performance computing : paradigm and infrastructure, edited by Laurence T. Yang, Minyi Guo
Resource Information
The item High-performance computing : paradigm and infrastructure, edited by Laurence T. Yang, Minyi Guo represents a specific, individual, material embodiment of a distinct intellectual or artistic creation found in University of Missouri-St. Louis Libraries.This item is available to borrow from 1 library branch.
Resource Information
The item High-performance computing : paradigm and infrastructure, edited by Laurence T. Yang, Minyi Guo represents a specific, individual, material embodiment of a distinct intellectual or artistic creation found in University of Missouri-St. Louis Libraries.
This item is available to borrow from 1 library branch.
- Summary
- With hyperthreading in Intel processors, hypertransport links in next generation AMD processors, multi-core silicon in today's high-end microprocessors from IBM and emerging grid computing, parallel and distributed computers have moved into the mainstream
- Language
- eng
- Extent
- 1 online resource (xxxviii, 778 pages)
- Contents
-
- Cover HIGH-PERFORMANCE COMPUTING Contents Preface Contributors PART 1 Programming Model 1 ClusterGOP: A High-Level Programming Environment for Clusters 1.1 Introduction 1.2 GOP Model and ClusterGOP Architecture 1.3 VisualGOP 1.4 The ClusterGOP Library 1.5 MPMD programming Support 1.6 Programming Using ClusterGOP 1.7 Summary 2 The Challenge of Providing A High-Level Programming Model for High-Performance Computing 2.1 Introduction 2.2 HPC Architectures 2.3 HPC Programming Models: The First Generation 2.4 The Second Generation of HPC Programming Models 2.5 OpenMP for DMPs 2.6 Experiments with OpenMP on DMPs 2.7 Conclusions 3 SAT: Toward Structured Parallelism Using Skeletons 3.1 Introduction 3.2 SAT: A Methodology Outline 3.3 Skeletons and Collective Operations 3.4 Case Study: Maximum Segment SUM (MSS) 3.5 Performance Aspect in SAT 3.6 Conclusions and Related Work 4 Bulk-Synchronous Parallelism: An Emerging Paradigm of High-Performance Computing 4.1 The BSP Model 4.2 BSP Programming 4.3 Conclusions 5 Cilk Versus MPI: Comparing Two Parallel Programming Styles on Heterogeneous Systems 5.1 Introduction 5.2 Experiments 5.3 Results 5.4 Conclusion 6 Nested Parallelism and Pipelining in OpenMP 6.1 Introduction 6.2 OpenMP Extensions for Nested Parallelism 6.3 OpenMP Extensions For Thread Synchronization 6.4 Summary 7 OpenMP for Chip Multiprocessors 7.1 Introduction 7.2 3SoC Architecture Overview 7.3 The OpenMp Compiler/Translator 7.4 Extensions to OpenMP for DSEs 7.5 Optimization for OpenMP 7.6 Implementation 7.7 Performance Evaluation 7.8 Conclusions PART 2 Architectural and System Support 8 Compiler and Run-Time Parallelization Techniques for Scientific Computations on Distributed-Memory Parallel Computers 8.1 Introduction 8.2 Background Material 8.3 Compiling Regular Programs on DMPCs 8.4 Compiler and Run-Time Support for Irregular Programs 8.5 Library Support for Irregular Applications 8.6 Related Works 8.7 Concluding Remarks 9 Enabling Partial-Cache Line Prefetching through Data Compression 9.1 Introduction 9.2 Motivation of Partial Cache-Line Prefetching 9.3 Cache Design Details 9.4 Experimental Results 9.5 Related Work 9.6 Conclusion 10 MPI Atomicity and Concurrent Overlapping I/O 10.1 Introduction 10.2 Concurrent Overlapping I/O 10.3 Implementation Strategies 10.4 Experiment Results 10.5 Summary 11 Code Tiling: One Size Fits All 11.1 Introduction 11.2 Cache Model 11.3 Code Tiling 11.4 Data Tiling 11.5 Finding Optimal Tile Sizes 11.6 Experimental Results 11.7 Related Work 11.8 Conclusion 12 Data Conversion for Heterogeneous Migration/Checkpointing 12.1 Introduction 12.2 Migration and Checkpointing 12.3 Data Conversion 12.4 Coarse-Grain Tagged RMR in MigThread 12.5 Microbenchmarks and Experiments 12.6 Related Work 12.7 Conclusions and Future Work 13 Receiving-Message Prediction and Its Speculative Execution 13.1 Background 13.2 Receiving-Message Prediction Method 13.3 Implementation of the Method in the MIPI Libraries 13.4 Experimental Results 13.5 Conclusing Remarks 14 An Investigation of the Applicability of Distributed FPGAs to High-Performance Computing 14.1 Introduction 14.2 High Performance Computing with Cluster Computing 14.3 Reconfigurable Computing with EPGAs 14.4 DRMC: A Distributed Reconfigurable Metacomputer 14.5 Algorithms Suited to the Implementation on FPGAs/DRMC 14.6 Algorithms Not Suited to the Implementation on FPGAs/DRMC 14.7 Summary PART 3 Scheduling and Resource Management 15 Bandwidth-Aware Resource Allocation for Heterogeneous Computing Systems to Maximize Throughput 15.1 Introduction 15.2 Related Work 15.3 System Model and Problem Statement 15.4 Resource Allocation to Maximize System Throughput 15.5 Experimental Results 15.6 Conclusion 16 Scheduling Algorithms with Bus Bandwidth Considerations for SMPs 16.1 Intr
- Isbn
- 9780471654711
- Label
- High-performance computing : paradigm and infrastructure
- Title
- High-performance computing
- Title remainder
- paradigm and infrastructure
- Statement of responsibility
- edited by Laurence T. Yang, Minyi Guo
- Subject
-
- COMPUTERS -- Systems Architecture | Distributed Systems & Computing
- Electronic books
- Electronic data processing -- Distributed processing
- Electronic data processing -- Distributed processing
- Grid Computing
- High performance computing
- High performance computing
- Parallel processing (Electronic computers)
- Parallel processing (Electronic computers)
- Processamento eletrônico de dados
- Sistemas distribuídos
- Supercomputer
- Arquitetura e organização de computadores
- Language
- eng
- Summary
- With hyperthreading in Intel processors, hypertransport links in next generation AMD processors, multi-core silicon in today's high-end microprocessors from IBM and emerging grid computing, parallel and distributed computers have moved into the mainstream
- Action
- digitized
- Cataloging source
- TEFOD
- http://library.link/vocab/creatorName
- Yang, Laurence Tianruo
- Dewey number
- 004/.35
- Illustrations
- illustrations
- Index
- index present
- Language note
- English
- LC call number
- QA76.88
- LC item number
- .H538 2005
- Literary form
- non fiction
- Nature of contents
-
- dictionaries
- bibliography
- http://library.link/vocab/relatedWorkOrContributorName
- Guo, Minyi
- Series statement
- Wiley series on parallel and distributed computing
- http://library.link/vocab/subjectName
-
- High performance computing
- Parallel processing (Electronic computers)
- Electronic data processing
- COMPUTERS
- Electronic data processing
- High performance computing
- Parallel processing (Electronic computers)
- Grid Computing
- Supercomputer
- Sistemas distribuídos
- Arquitetura e organização de computadores
- Processamento eletrônico de dados
- Label
- High-performance computing : paradigm and infrastructure, edited by Laurence T. Yang, Minyi Guo
- Bibliography note
- Includes bibliographical references and index
- Carrier category
- online resource
- Carrier category code
-
- cr
- Carrier MARC source
- rdacarrier
- Color
- multicolored
- Content category
- text
- Content type code
-
- txt
- Content type MARC source
- rdacontent
- Contents
- Cover HIGH-PERFORMANCE COMPUTING Contents Preface Contributors PART 1 Programming Model 1 ClusterGOP: A High-Level Programming Environment for Clusters 1.1 Introduction 1.2 GOP Model and ClusterGOP Architecture 1.3 VisualGOP 1.4 The ClusterGOP Library 1.5 MPMD programming Support 1.6 Programming Using ClusterGOP 1.7 Summary 2 The Challenge of Providing A High-Level Programming Model for High-Performance Computing 2.1 Introduction 2.2 HPC Architectures 2.3 HPC Programming Models: The First Generation 2.4 The Second Generation of HPC Programming Models 2.5 OpenMP for DMPs 2.6 Experiments with OpenMP on DMPs 2.7 Conclusions 3 SAT: Toward Structured Parallelism Using Skeletons 3.1 Introduction 3.2 SAT: A Methodology Outline 3.3 Skeletons and Collective Operations 3.4 Case Study: Maximum Segment SUM (MSS) 3.5 Performance Aspect in SAT 3.6 Conclusions and Related Work 4 Bulk-Synchronous Parallelism: An Emerging Paradigm of High-Performance Computing 4.1 The BSP Model 4.2 BSP Programming 4.3 Conclusions 5 Cilk Versus MPI: Comparing Two Parallel Programming Styles on Heterogeneous Systems 5.1 Introduction 5.2 Experiments 5.3 Results 5.4 Conclusion 6 Nested Parallelism and Pipelining in OpenMP 6.1 Introduction 6.2 OpenMP Extensions for Nested Parallelism 6.3 OpenMP Extensions For Thread Synchronization 6.4 Summary 7 OpenMP for Chip Multiprocessors 7.1 Introduction 7.2 3SoC Architecture Overview 7.3 The OpenMp Compiler/Translator 7.4 Extensions to OpenMP for DSEs 7.5 Optimization for OpenMP 7.6 Implementation 7.7 Performance Evaluation 7.8 Conclusions PART 2 Architectural and System Support 8 Compiler and Run-Time Parallelization Techniques for Scientific Computations on Distributed-Memory Parallel Computers 8.1 Introduction 8.2 Background Material 8.3 Compiling Regular Programs on DMPCs 8.4 Compiler and Run-Time Support for Irregular Programs 8.5 Library Support for Irregular Applications 8.6 Related Works 8.7 Concluding Remarks 9 Enabling Partial-Cache Line Prefetching through Data Compression 9.1 Introduction 9.2 Motivation of Partial Cache-Line Prefetching 9.3 Cache Design Details 9.4 Experimental Results 9.5 Related Work 9.6 Conclusion 10 MPI Atomicity and Concurrent Overlapping I/O 10.1 Introduction 10.2 Concurrent Overlapping I/O 10.3 Implementation Strategies 10.4 Experiment Results 10.5 Summary 11 Code Tiling: One Size Fits All 11.1 Introduction 11.2 Cache Model 11.3 Code Tiling 11.4 Data Tiling 11.5 Finding Optimal Tile Sizes 11.6 Experimental Results 11.7 Related Work 11.8 Conclusion 12 Data Conversion for Heterogeneous Migration/Checkpointing 12.1 Introduction 12.2 Migration and Checkpointing 12.3 Data Conversion 12.4 Coarse-Grain Tagged RMR in MigThread 12.5 Microbenchmarks and Experiments 12.6 Related Work 12.7 Conclusions and Future Work 13 Receiving-Message Prediction and Its Speculative Execution 13.1 Background 13.2 Receiving-Message Prediction Method 13.3 Implementation of the Method in the MIPI Libraries 13.4 Experimental Results 13.5 Conclusing Remarks 14 An Investigation of the Applicability of Distributed FPGAs to High-Performance Computing 14.1 Introduction 14.2 High Performance Computing with Cluster Computing 14.3 Reconfigurable Computing with EPGAs 14.4 DRMC: A Distributed Reconfigurable Metacomputer 14.5 Algorithms Suited to the Implementation on FPGAs/DRMC 14.6 Algorithms Not Suited to the Implementation on FPGAs/DRMC 14.7 Summary PART 3 Scheduling and Resource Management 15 Bandwidth-Aware Resource Allocation for Heterogeneous Computing Systems to Maximize Throughput 15.1 Introduction 15.2 Related Work 15.3 System Model and Problem Statement 15.4 Resource Allocation to Maximize System Throughput 15.5 Experimental Results 15.6 Conclusion 16 Scheduling Algorithms with Bus Bandwidth Considerations for SMPs 16.1 Intr
- Control code
- 71353018
- Dimensions
- unknown
- Extent
- 1 online resource (xxxviii, 778 pages)
- Form of item
- online
- Isbn
- 9780471654711
- Media category
- computer
- Media MARC source
- rdamedia
- Media type code
-
- c
- Other control number
- 10.1002/0471732710
- Other physical details
- illustrations
- http://library.link/vocab/ext/overdrive/overdriveId
- 04563e26-ed8a-4991-84ab-19f271d7fc58
- Reproduction note
- Electronic reproduction.
- Specific material designation
- remote
- System control number
- (OCoLC)71353018
- System details
- Master and use copy. Digital master created according to Benchmark for Faithful Digital Reproductions of Monographs and Serials, Version 1. Digital Library Federation, December 2002.
- Label
- High-performance computing : paradigm and infrastructure, edited by Laurence T. Yang, Minyi Guo
- Bibliography note
- Includes bibliographical references and index
- Carrier category
- online resource
- Carrier category code
-
- cr
- Carrier MARC source
- rdacarrier
- Color
- multicolored
- Content category
- text
- Content type code
-
- txt
- Content type MARC source
- rdacontent
- Contents
- Cover HIGH-PERFORMANCE COMPUTING Contents Preface Contributors PART 1 Programming Model 1 ClusterGOP: A High-Level Programming Environment for Clusters 1.1 Introduction 1.2 GOP Model and ClusterGOP Architecture 1.3 VisualGOP 1.4 The ClusterGOP Library 1.5 MPMD programming Support 1.6 Programming Using ClusterGOP 1.7 Summary 2 The Challenge of Providing A High-Level Programming Model for High-Performance Computing 2.1 Introduction 2.2 HPC Architectures 2.3 HPC Programming Models: The First Generation 2.4 The Second Generation of HPC Programming Models 2.5 OpenMP for DMPs 2.6 Experiments with OpenMP on DMPs 2.7 Conclusions 3 SAT: Toward Structured Parallelism Using Skeletons 3.1 Introduction 3.2 SAT: A Methodology Outline 3.3 Skeletons and Collective Operations 3.4 Case Study: Maximum Segment SUM (MSS) 3.5 Performance Aspect in SAT 3.6 Conclusions and Related Work 4 Bulk-Synchronous Parallelism: An Emerging Paradigm of High-Performance Computing 4.1 The BSP Model 4.2 BSP Programming 4.3 Conclusions 5 Cilk Versus MPI: Comparing Two Parallel Programming Styles on Heterogeneous Systems 5.1 Introduction 5.2 Experiments 5.3 Results 5.4 Conclusion 6 Nested Parallelism and Pipelining in OpenMP 6.1 Introduction 6.2 OpenMP Extensions for Nested Parallelism 6.3 OpenMP Extensions For Thread Synchronization 6.4 Summary 7 OpenMP for Chip Multiprocessors 7.1 Introduction 7.2 3SoC Architecture Overview 7.3 The OpenMp Compiler/Translator 7.4 Extensions to OpenMP for DSEs 7.5 Optimization for OpenMP 7.6 Implementation 7.7 Performance Evaluation 7.8 Conclusions PART 2 Architectural and System Support 8 Compiler and Run-Time Parallelization Techniques for Scientific Computations on Distributed-Memory Parallel Computers 8.1 Introduction 8.2 Background Material 8.3 Compiling Regular Programs on DMPCs 8.4 Compiler and Run-Time Support for Irregular Programs 8.5 Library Support for Irregular Applications 8.6 Related Works 8.7 Concluding Remarks 9 Enabling Partial-Cache Line Prefetching through Data Compression 9.1 Introduction 9.2 Motivation of Partial Cache-Line Prefetching 9.3 Cache Design Details 9.4 Experimental Results 9.5 Related Work 9.6 Conclusion 10 MPI Atomicity and Concurrent Overlapping I/O 10.1 Introduction 10.2 Concurrent Overlapping I/O 10.3 Implementation Strategies 10.4 Experiment Results 10.5 Summary 11 Code Tiling: One Size Fits All 11.1 Introduction 11.2 Cache Model 11.3 Code Tiling 11.4 Data Tiling 11.5 Finding Optimal Tile Sizes 11.6 Experimental Results 11.7 Related Work 11.8 Conclusion 12 Data Conversion for Heterogeneous Migration/Checkpointing 12.1 Introduction 12.2 Migration and Checkpointing 12.3 Data Conversion 12.4 Coarse-Grain Tagged RMR in MigThread 12.5 Microbenchmarks and Experiments 12.6 Related Work 12.7 Conclusions and Future Work 13 Receiving-Message Prediction and Its Speculative Execution 13.1 Background 13.2 Receiving-Message Prediction Method 13.3 Implementation of the Method in the MIPI Libraries 13.4 Experimental Results 13.5 Conclusing Remarks 14 An Investigation of the Applicability of Distributed FPGAs to High-Performance Computing 14.1 Introduction 14.2 High Performance Computing with Cluster Computing 14.3 Reconfigurable Computing with EPGAs 14.4 DRMC: A Distributed Reconfigurable Metacomputer 14.5 Algorithms Suited to the Implementation on FPGAs/DRMC 14.6 Algorithms Not Suited to the Implementation on FPGAs/DRMC 14.7 Summary PART 3 Scheduling and Resource Management 15 Bandwidth-Aware Resource Allocation for Heterogeneous Computing Systems to Maximize Throughput 15.1 Introduction 15.2 Related Work 15.3 System Model and Problem Statement 15.4 Resource Allocation to Maximize System Throughput 15.5 Experimental Results 15.6 Conclusion 16 Scheduling Algorithms with Bus Bandwidth Considerations for SMPs 16.1 Intr
- Control code
- 71353018
- Dimensions
- unknown
- Extent
- 1 online resource (xxxviii, 778 pages)
- Form of item
- online
- Isbn
- 9780471654711
- Media category
- computer
- Media MARC source
- rdamedia
- Media type code
-
- c
- Other control number
- 10.1002/0471732710
- Other physical details
- illustrations
- http://library.link/vocab/ext/overdrive/overdriveId
- 04563e26-ed8a-4991-84ab-19f271d7fc58
- Reproduction note
- Electronic reproduction.
- Specific material designation
- remote
- System control number
- (OCoLC)71353018
- System details
- Master and use copy. Digital master created according to Benchmark for Faithful Digital Reproductions of Monographs and Serials, Version 1. Digital Library Federation, December 2002.
Subject
- COMPUTERS -- Systems Architecture | Distributed Systems & Computing
- Electronic books
- Electronic data processing -- Distributed processing
- Electronic data processing -- Distributed processing
- Grid Computing
- High performance computing
- High performance computing
- Parallel processing (Electronic computers)
- Parallel processing (Electronic computers)
- Processamento eletrônico de dados
- Sistemas distribuídos
- Supercomputer
- Arquitetura e organização de computadores
Genre
Member of
Library Links
Embed
Settings
Select options that apply then copy and paste the RDF/HTML data fragment to include in your application
Embed this data in a secure (HTTPS) page:
Layout options:
Include data citation:
<div class="citation" vocab="http://schema.org/"><i class="fa fa-external-link-square fa-fw"></i> Data from <span resource="http://link.umsl.edu/portal/High-performance-computing--paradigm-and/2tCN2FAKzE4/" typeof="Book http://bibfra.me/vocab/lite/Item"><span property="name http://bibfra.me/vocab/lite/label"><a href="http://link.umsl.edu/portal/High-performance-computing--paradigm-and/2tCN2FAKzE4/">High-performance computing : paradigm and infrastructure, edited by Laurence T. Yang, Minyi Guo</a></span> - <span property="potentialAction" typeOf="OrganizeAction"><span property="agent" typeof="LibrarySystem http://library.link/vocab/LibrarySystem" resource="http://link.umsl.edu/"><span property="name http://bibfra.me/vocab/lite/label"><a property="url" href="http://link.umsl.edu/">University of Missouri-St. Louis Libraries</a></span></span></span></span></div>
Note: Adjust the width and height settings defined in the RDF/HTML code fragment to best match your requirements
Preview
Cite Data - Experimental
Data Citation of the Item High-performance computing : paradigm and infrastructure, edited by Laurence T. Yang, Minyi Guo
Copy and paste the following RDF/HTML data fragment to cite this resource
<div class="citation" vocab="http://schema.org/"><i class="fa fa-external-link-square fa-fw"></i> Data from <span resource="http://link.umsl.edu/portal/High-performance-computing--paradigm-and/2tCN2FAKzE4/" typeof="Book http://bibfra.me/vocab/lite/Item"><span property="name http://bibfra.me/vocab/lite/label"><a href="http://link.umsl.edu/portal/High-performance-computing--paradigm-and/2tCN2FAKzE4/">High-performance computing : paradigm and infrastructure, edited by Laurence T. Yang, Minyi Guo</a></span> - <span property="potentialAction" typeOf="OrganizeAction"><span property="agent" typeof="LibrarySystem http://library.link/vocab/LibrarySystem" resource="http://link.umsl.edu/"><span property="name http://bibfra.me/vocab/lite/label"><a property="url" href="http://link.umsl.edu/">University of Missouri-St. Louis Libraries</a></span></span></span></span></div>