Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/209936acmconferencesBook PagePublication PagesppoppConference Proceedingsconference-collections
PPOPP '95: Proceedings of the fifth ACM SIGPLAN symposium on Principles and practice of parallel programming
ACM1995 Proceeding
Publisher:
  • Association for Computing Machinery
  • New York
  • NY
  • United States
Conference:
PPoPP95: Principles & Practices of Parallel Programming Santa Barbara California USA July 19 - 21, 1995
ISBN:
978-0-89791-700-1
Published:
01 August 1995
Sponsors:
Reflects downloads up to 08 Feb 2025Bibliometrics
Abstract

No abstract available.

Skip Table Of Content Section
Article
Free
A model and compilation strategy for out-of-core data parallel programs

It is widely acknowledged in high-performance computing circles that parallel input/output needs substantial improvement in order to make scalable computers truly usable. We present a data storage model that allows processors independent access to their ...

Article
Free
pC++/streams: a library for I/O on complex distributed data sources

The design and implementation of portable, efficient, and expressive mechanisms for I/O on complex distributed data structures—such as found in adaptive parallel applications—is a challenging problem that we address in this paper.

We describe the design, ...

Article
Free
Parallel skeletons for structured composition

In this paper, we propose a straightforward solution to the problems of compositional parallel programming by using skeletons as the uniform mechanism for structured composition. In our approach parallel programs are constructed by composing procedures ...

Article
Free
Software caching and computation migration in Olden

The goal of the Olden project is to build a system that provides parallelism for general purpose C programs with minimal programmer annotations. We focus on programs using dynamic structures such as trees, lists, and DAGs. We demonstrate that providing ...

Article
Free
Distributed data access in AC

We have modified the C language to support a programming model based on a shared address space with physically distributed memory. With this model users can write programs in which the nodes of a massively parallel processor can access remote memory ...

Article
Free
Generating parallel code from object oriented mathematical models

For a long time efficient use of parallel computers has been hindered by dependencies introduced in software through low-level implementation practice. In this paper we present a programming environment and language called Object-Math (Object oriented ...

Article
Free
Flattening and parallelizing irregular, recurrent loop nests

Irregular loop nests in which the loop bounds are determined dynamically by indexed arrays are difficult to compile into expressive parallel constructs, such as segmented scans and reductions. In this paper, we describe a suite of transformations to ...

Article
Free
Efficient support for irregular applications on distributed-memory machines

Irregular computation problems underlie many important scientific applications. Although these problems are computationally expensive, and so would seem appropriate for parallel machines, their irregular and unpredictable run-time behavior makes this ...

Article
Free
High-level optimization via automated statistical modeling

We develop the use of statistical modeling for portable high-level optimizations such as data layout and algorithm selection. We build the models automatically from profiling information, which ensures robust and accurate models that reflect all aspects ...

Article
Free
Fuzzy array dataflow analysis

Exact array dataflow analysis can be achieved in the general case if the only control structures are do-loops and structural ifs, and if loop counter bounds and array subscripts are affine expressions of englobing loop counters and possibly some integer ...

Article
Free
A linear-time algorithm for computing the memory access sequence in data-parallel programs

Data-parallel languages, such as High Performance Fortran, are widely regarded as a promising means for writing portable programs for distributed-memory machines. Novel features of these languages call for the development of new techniques in both ...

Article
Free
An array operation synthesis scheme to optimize Fortran 90 programs

An increasing number of programming languages, such as Fortran 90 and APL, are providing a rich set of intrinsic array functions and array expressions. These constructs which constitute an important part of data parallel languages provide excellent ...

Article
Free
Parallel algorithms for image histogramming and connected components with an experimental study (extended abstract)

This paper presents efficient and portable implementations of two useful primitives in image processing algorithms, histogramming and connected components. Our general framework is a single-address space, distributed memory programming model. We use ...

Article
Free
Optimal mapping of sequences of data parallel tasks

Many applications in a variety of domains including digital signal processing, image processing and computer vision are composed of a sequence of tasks that act on a stream of input data sets in a pipelined manner. Recent research has established that ...

Article
Free
Compiler optimizations for eliminating barrier synchronization

This paper presents novel compiler optimizations for reducing synchronization overhead in compiler-parallelized scientific codes. A hybrid programming model is employed to combine the flexibility of the fork-join model with the precision and power of ...

Article
Free
Automatic alignment of array data and processes to reduce communication time on DMPPs

This paper investigates the problem of aligning array data and processes in a distributed-memory implementation. We present complete algorithms for compile-time analysis, the necessary program restructuring, and subsequent code-generation, and discuss ...

Article
Free
Data and computation transformations for multiprocessors

Effective memory hierarchy utilization is critical to the performance of modern multiprocessor architectures. We have developed the first compiler system that fully automatically parallelizes sequential programs and changes the original array layouts to ...

Article
Free
Reducing false sharing on shared memory multiprocessors through compile time data transformations

We have developed compiler algorithms that analyze explicitly parallel programs and restructure their shared data to reduce the number of false sharing misses. The algorithms analyze per-process shared data accesses, pinpoint the data structures that ...

Article
Free
Evaluating the locality benefits of active messages

A major challenge in fine-grained computing is achieving locality without excessive scheduling overhead. We built two J-Machine implementations of a fine-grained programming model, the Berkeley Threaded Abstract Machine. One implementation takes an ...

Article
Free
High performance synchronization algorithms for multiprogrammed multiprocessors

Scalable busy-wait synchronization algorithms are essential for achieving good parallel program performance on large scale multiprocessors. Such algorithms include mutual exclusion locks, reader-writer locks, and barrier synchronization. Unfortunately, ...

Article
Free
Cilk: an efficient multithreaded runtime system

Cilk (pronounced “silk”) is a C-based runtime system for multi-threaded parallel programming. In this paper, we document the efficiency of the Cilk work-stealing scheduler, both empirically and analytically. We show that on real and synthetic ...

Article
Free
Optimistic active messages: a mechanism for scheduling communication with computation

Low-overhead message passing is critical to the performance of many applications. Active Messages reduce the software overhead for message handling: messages are run as handlers instead of as threads, which avoids the overhead of thread management and ...

Contributors
  • Department of Computer Science and Engineering
  • University of Illinois Urbana-Champaign

Recommendations

Acceptance Rates

Overall Acceptance Rate 230 of 1,014 submissions, 23%
YearSubmittedAcceptedRate
PPoPP '211503121%
PPoPP '201212823%
PPoPP '191522919%
PPoPP '171322922%
PPoPP '141842815%
PPoPP '07652234%
PPoPP '03452044%
PPoPP '99791722%
PPOPP '97862630%
Overall1,01423023%