Comparative Study of Message Passing and Shared Memory Parallel Programming Models in Neural Network Training

PDF Version Also Available for Download.

Description

It is presented a comparative performance study of a coarse grained parallel neural network training code, implemented in both OpenMP and MPI, standards for shared memory and message passing parallel programming environments, respectively. In addition, these versions of the parallel training code are compared to an implementation utilizing SHMEM the native SGI/CRAY environment for shared memory programming. The multiprocessor platform used is a SGI/Cray Origin 2000 with up to 32 processors. It is shown that in this study, the native CRAY environment outperforms MPI for the entire range of processors used, while OpenMP shows better performance than the other two ... continued below

Physical Description

592 Kilobytes pages

Creation Information

Vitela, J.; Gordillo, J.; Cortina, L & Hanebutte, U. December 14, 1999.

Context

This article is part of the collection entitled: Office of Scientific & Technical Information Technical Reports and was provided by UNT Libraries Government Documents Department to Digital Library, a digital repository hosted by the UNT Libraries. It has been viewed 49 times , with 10 in the last month . More information about this article can be viewed below.

Who

People and organizations associated with either the creation of this article or its content.

Sponsor

Publisher

Provided By

UNT Libraries Government Documents Department

Serving as both a federal and a state depository library, the UNT Libraries Government Documents Department maintains millions of items in a variety of formats. The department is a member of the FDLP Content Partnerships Program and an Affiliated Archive of the National Archives.

Contact Us

What

Descriptive information to help identify this article. Follow the links below to find similar items on the Digital Library.

Description

It is presented a comparative performance study of a coarse grained parallel neural network training code, implemented in both OpenMP and MPI, standards for shared memory and message passing parallel programming environments, respectively. In addition, these versions of the parallel training code are compared to an implementation utilizing SHMEM the native SGI/CRAY environment for shared memory programming. The multiprocessor platform used is a SGI/Cray Origin 2000 with up to 32 processors. It is shown that in this study, the native CRAY environment outperforms MPI for the entire range of processors used, while OpenMP shows better performance than the other two environments when using more than 19 processors. In this study, the efficiency is always greater than 60% regardless of the parallel programming environment used as well as of the number of processors.

Physical Description

592 Kilobytes pages

Source

  • High Performance Computing 2000, Washington, DC (US), 04/16/2000--04/20/2000

Language

Item Type

Identifier

Unique identifying numbers for this article in the Digital Library or other systems.

  • Report No.: UCRL-JC-136867
  • Grant Number: W-7405-Eng-48
  • Office of Scientific & Technical Information Report Number: 791052
  • Archival Resource Key: ark:/67531/metadc723216

Collections

This article is part of the following collection of related materials.

Office of Scientific & Technical Information Technical Reports

Reports, articles and other documents harvested from the Office of Scientific and Technical Information.

Office of Scientific and Technical Information (OSTI) is the Department of Energy (DOE) office that collects, preserves, and disseminates DOE-sponsored research and development (R&D) results that are the outcomes of R&D projects or other funded activities at DOE labs and facilities nationwide and grantees at universities and other institutions.

What responsibilities do I have when using this article?

When

Dates and time periods associated with this article.

Creation Date

  • December 14, 1999

Added to The UNT Digital Library

  • Sept. 29, 2015, 5:31 a.m.

Description Last Updated

  • May 6, 2016, 3:16 p.m.

Usage Statistics

When was this article last used?

Yesterday: 1
Past 30 days: 10
Total Uses: 49

Interact With This Article

Here are some suggestions for what to do next.

Start Reading

PDF Version Also Available for Download.

International Image Interoperability Framework

IIF Logo

We support the IIIF Presentation API

Vitela, J.; Gordillo, J.; Cortina, L & Hanebutte, U. Comparative Study of Message Passing and Shared Memory Parallel Programming Models in Neural Network Training, article, December 14, 1999; California. (digital.library.unt.edu/ark:/67531/metadc723216/: accessed May 25, 2018), University of North Texas Libraries, Digital Library, digital.library.unt.edu; crediting UNT Libraries Government Documents Department.