Article ID: CPE1020

Publisher John Wiley Sons, Ltd. Chichester, UK
Category Research Article
Article Title ShanghaiGrid: an Information Service Grid
Volume ID 18
Issue ID 1
Date 01 00 2006
DOI(URI) http://dx.doi.org/10.1002/cpe.1020
Article ID CPE1020
Author Name(s) Minglu Li1Min-You Wu2Ying Li3Jian Cao4Linpeng Huang5Qianni Deng6Xinhua Lin7Changjun Jiang8Weiqin Tong9Yadong Gui10Aoying Zhou11Xinhong Wu12Shui Jiang13
Author Email(s) mwu@sjtu.edu.cn1
Affiliation(s) Department of Computer Science and Engineering, Shanghai Jiao Tong University, Shanghai, People's Republic of China 1 2 3 4 5 6 7 Tongji University, Shanghai, People's Republic of China 8Shanghai University, Shanghai, People's Republic of China 9Shanghai Supercomputer Center, Shanghai, People's Republic of China 10Fudan University, Shanghai, People's Republic of China 11Shanghai Urban Transportation Information Center, Shanghai, People's Republic of China 12East China Institute of Computer Technology, Shanghai, People's Republic of China 13
Keyword(s) ShanghaiGrid, information Grid, metropolitan-area information services, Information Service Grid Toolkit, Grid application,
Abstract
The goal of the ShanghaiGrid is to provide information services to the people. It aims to construct a metropolitan-area information service infrastructure and establish an open standard for widespread upper-layer applications from both communities and the government. The Information Service Grid Toolkit and a typical application called the Traffic Information Grid are discussed in detail. Copyright © 2005 John Wiley & Sons, Ltd.

Article ID: CPE904

Publisher John Wiley Sons, Ltd. Chichester, UK
Category Research Article
Article Title Scheduling communication in multithreaded programs: experimental results
Volume ID 18
Issue ID 1
Date 01 00 2006
DOI(URI) http://dx.doi.org/10.1002/cpe.904
Article ID CPE904
Author Name(s) Vernon Rego1Juan Carlos Gomez2V. S. Sunderam3
Author Email(s) rego@cs.purdue.edu1
Affiliation(s) Department of Computer Sciences, Purdue University, West Lafayette, IN 47907, U.S.A. 1 2 Department of Mathematics and Computer Science, Emory University, Atlanta, GA 30322, U.S.A. 3
Keyword(s) user-level protocol, threads, scheduling, latency, interrupts, signals,
Abstract
When the critical path of a communication session between end points includes the actions of operating system kernels, there are attendant overheads. Along with other factors, such as functionality and flexibility, such overheads motivate and favor the implementation of communication protocols in user space. When implemented with threads, such protocols may hold the key to optimal communication performance and functionality. Based on implementations of reliable user-space protocols supported by a threads framework, we focus on our experiences with internal threads' scheduling techniques and their potential impact on performance. We present scheduling strategies that enable threads to do both application-level and communication-related processing. With experiments performed on a Sun SPARC-5 LAN environment, we show how different scheduling strategies yield different levels of application-processing efficiency, communication latency and packet-loss. This work forms part of a larger study on the implementation of multiple thread-based protocols in a single address space, and the benefits of coupling protocols with applications. Copyright © 2005 John Wiley & Sons, Ltd.

Article ID: CPE905

Publisher John Wiley Sons, Ltd. Chichester, UK
Category Research Article
Article Title Performance comparison of MPI and OpenMP on shared memory multiprocessors
Volume ID 18
Issue ID 1
Date 01 00 2006
DOI(URI) http://dx.doi.org/10.1002/cpe.905
Article ID CPE905
Author Name(s) Franck Cappello1G?raud Krawezik2
Author Email(s) fci@lri.fr1
Affiliation(s) LRI, Universit? de Paris Sud, Orsay, France 1 2
Keyword(s) MPI, OpenMP, performance evaluation, shared memory multiprocessors,
Abstract
When using a shared memory multiprocessor, the programmer faces the issue of selecting the portable programming model which will provide the best performance. Even if they restricts their choice to the standard programming environments (MPI and OpenMP), they have to select a programming approach among MPI and the variety of OpenMP programming styles. To help the programmer in their decision, we compare MPI with three OpenMP programming styles (loop level, loop level with large parallel sections, SPMD) using a subset of the NAS benchmark (CG, MG, FT, LU), two dataset sizes (A and B), and two shared memory multiprocessors (IBM SP3 NightHawk II, SGI Origin 3800). We have developed the first SPMD OpenMP version of the NAS benchmark and gathered other OpenMP versions from independent sources (PBN, SDSC and RWCP). Experimental results demonstrate that OpenMP provides competitive performance compared with MPI for a large set of experimental conditions. Not surprisingly, the two best OpenMP versions are those requiring the strongest programming effort. MPI still provides the best performance under some conditions. We present breakdowns of the execution times and measurements of hardware performance counters to explain the performance differences. Copyright © 2005 John Wiley & Sons, Ltd.

Article ID: CPE907

Publisher John Wiley Sons, Ltd. Chichester, UK
Category Research Article
Article Title Measuring and modelling the performance of a parallel ODMG compliant object database server
Volume ID 18
Issue ID 1
Date 01 00 2006
DOI(URI) http://dx.doi.org/10.1002/cpe.907
Article ID CPE907
Author Name(s) Norman W. Paton1G?raud Krawezik2Jim Smith3Paul Watson4
Author Email(s) norm@cs.man.ac.uk1
Affiliation(s) Department of Computer Science, University of Manchester, Oxford Road, Manchester M13 9PL, U.K. 1 2 Department of Computer Science, University of Newcastle upon Tyne, Newcastle upon Tyne NE1 7RU, U.K. 3 4
Keyword(s) object database, parallel databases, ODMG, benchmark, OQL, cost model,
Abstract
Object database management systems (ODBMSs) are now established as the database management technology of choice for a range of challenging data intensive applications. Furthermore, the applications associated with object databases typically have stringent performance requirements, and some are associated with very large data sets. An important feature for the performance of object databases is the speed at which relationships can be explored. In queries, this depends on the effectiveness of different join algorithms into which queries that follow relationships can be compiled. This paper presents a performance evaluation of the Polar parallel object database system, focusing in particular on the performance of parallel join algorithms. Polar is a parallel, shared-nothing implementation of the Object Database Management Group (ODMG) standard for object databases. The paper presents an empirical evaluation of queries expressed in the ODMG Query Language (OQL), as well as a cost model for the parallel algebra that is used to evaluate OQL queries. The cost model is validated against the empirical results for a collection of queries using four different join algorithms, one that is value based and three that are pointer based Copyright © 2005 John Wiley & Sons, Ltd.