556 Matching Results

Search Results

Advanced search parameters have been applied.

Three Dimensional Holographic Archival Memory

Description: To address the DOE need for the storage and handling of terabyte of nuclear physics data, Physical Optics Corporation (POC) developed a new multi-terabit, 90° recording geometry 3D holographic archival optical memory storage and search system. In contrast to state-of-the-art memory approaches, 3DHAM handles the data through highly parallel optical processing in conjunction with highly redundant hologram multiplexing. The 3DHAM system advantages also stem from its unique 3D volume-recording medium, which theoretically has a significantly greater data density than diskbased systems.
Date: November 9, 2006
Creator: Gibbard, Bruce
Partner: UNT Libraries Government Documents Department

A Comparison of Three Voting Methods for Bagging with the MLEM2 Algorithm

Description: This paper presents results of experiments on some data sets using bagging on the MLEM2 rule induction algorithm. Three different methods of ensemble voting, based on support (a non-democratic voting in which ensembles vote with their strengths), strength only (an ensemble with the largest strength decides to which concept a case belongs) and democratic voting (each ensemble has at most one vote) were used. Our conclusions are that though in most cases democratic voting was the best, it is not significantly better than voting based on support. The strength voting was the worst voting method.
Date: March 17, 2010
Creator: Cohagan, Clinton; Grzymala-Busse, Jerzy W. & Hippe, Zdzislaw S.
Partner: UNT Libraries Government Documents Department

A Lightweight, High-performance I/O Management Package for Data-intensive Computing

Description: Our group has been working with ANL collaborators on the topic “bridging the gap between parallel file system and local file system” during the course of this project period. We visited Argonne National Lab -- Dr. Robert Ross’s group for one week in the past summer 2007. We looked over our current project progress and planned the activities for the incoming years 2008-09. The PI met Dr. Robert Ross several times such as HEC FSIO workshop 08, SC’08 and SC’10. We explored the opportunities to develop a production system by leveraging our current prototype to (SOGP+PVFS) a new PVFS version. We delivered SOGP+PVFS codes to ANL PVFS2 group in 2008.We also talked about exploring a potential project on developing new parallel programming models and runtime systems for data-intensive scalable computing (DISC). The methodology is to evolve MPI towards DISC by incorporating some functions of Google MapReduce parallel programming model. More recently, we are together exploring how to leverage existing works to perform (1) coordination/aggregation of local I/O operations prior to movement over the WAN, (2) efficient bulk data movement over the WAN, (3) latency hiding techniques for latency-intensive operations. Since 2009, we start applying Hadoop/MapReduce to some HEC applications with LANL scientists John Bent and Salman Habib. Another on-going work is to improve checkpoint performance at I/O forwarding Layer for the Road Runner super computer with James Nuetz and Gary Gridder at LANL. Two senior undergraduates from our research group did summer internships about high-performance file and storage system projects in LANL since 2008 for consecutive three years. Both of them are now pursuing Ph.D. degree in our group and will be 4th year in the PhD program in Fall 2011 and go to LANL to advance two above-mentioned works during this winter break. Since 2009, we have been collaborating ...
Date: June 22, 2011
Creator: Wang, Jun
Partner: UNT Libraries Government Documents Department

Department of Energy DE-FG02-05ER25686 Early Career Principal Investigator Award Final Report

Description: This project is researching FAST, a methodology to build very fast, cycleaccurate full system computer simulators and building the first set of such simulators and the first set of tools to help construct those simulators. The methodology relies on a functional model that is a fast, full-system but not cycle-accurate simulator coupled with a timing model that is models the micro-architectural structure and arbitration of a computer system, but not its functionality. The way FAST simulators differ from other simulators partitioned in the same way is that the interface between the functional and timing model is optimized to minimize the need for round-trip communication. The optimized communication enables FAST timing models to be implemented in an FPGA and the functional model in software while still achieving extremely high performance. Our general strategy is to design a methodology and then prototype an example using that methodology to ensure the methodology is sound before committing to it. This strategy ensures that the selected methodology has been tested and provides an early example of the output of that methodology.
Date: May 11, 2013
Creator: Derek, Chiou
Partner: UNT Libraries Government Documents Department

Architecture-Aware Algorithms for Scalable Performance and Resilience on Heterogeneous Architectures

Description: The goal of the Extreme-scale Algorithms & Software Institute (EASI) is to close the �application-architecture performance gap� by exploring algorithms and runtime improvements that will enable key science applications to better exploit the architectural features of DOE extreme-scale systems. For the past year of the project, our efforts at the University of Tennessee have concentrated on, and made significant progress related to, the following high-level EASI goals: � Develop multi-precision and architecture-aware implementations of Krylov, Poisson, Helmholtz solvers, and dense factorizations for heterogeneous multi-core systems; � Explore new methods of algorithm resilience, and develop new algorithms with these capabilities; � Develop runtime support for adaptable algorithms that are dealing with resilience, scalability; � Distribute the new algorithms and runtime support through widely used software packages; � Establish a strong outreach program to disseminate results, interact with colleagues and train students and junior members of our community.
Date: October 15, 2013
Creator: Dongarra, Jack
Partner: UNT Libraries Government Documents Department

Coordinated Fault Tolerance for High-Performance Computing

Description: Our work to meet our goal of end-to-end fault tolerance has focused on two areas: (1) improving fault tolerance in various software currently available and widely used throughout the HEC domain and (2) using fault information exchange and coordination to achieve holistic, systemwide fault tolerance and understanding how to design and implement interfaces for integrating fault tolerance features for multiple layers of the software stack—from the application, math libraries, and programming language runtime to other common system software such as jobs schedulers, resource managers, and monitoring tools.
Date: April 8, 2013
Creator: Dongarra, Jack; Bosilca, George & al., et
Partner: UNT Libraries Government Documents Department

The Harness Workbench: Unified and Adaptive Access to Diverse HPC Platforms (final report)

Description: In this project, we conducted preliminary research to create a flexible environment that encapsulates the knowledge of the application developers, site system administrators, and the vendors to assist application building and execution on HPC systems, in particular the DOE leadership computing platforms. The key research involved how to describe and use the knowledge from these varied sources to improve productivity of the end-user scientists, while creating a flexible and modular environment supporting all these features.
Date: November 7, 2012
Creator: Bosilca, George
Partner: UNT Libraries Government Documents Department

Minimizing System Noise Effects For Extreme-Scale Scientific Simulation Through Function Delegation

Description: The primary goal of the Minimizing System Noise Effects For Extreme-Scale Scientific Simulation through Function Delegation project is to eliminate or at best strongly minimize the impact of the noise introduced by the operating system, during large scale parallel applications runs. Collective communication operations are a basic building block for parallel programing models and scientific applications. These operations often dominate execution time of applications and tend to limit their scalability. In order to address this challenge, we evaluated different strategies to adapt the collective communications underlying topologies to the hardware architecture in order to provide increased levels of performance to the parallel applications.
Date: June 11, 2013
Creator: Dongarra, Jack J. & Bosilca, George
Partner: UNT Libraries Government Documents Department

Institute for Scalable Application Development Software

Description: Work by the University of Wisconsin as part of the DOE SciDAC CScADS includes the following accomplishments: � Research on tool componentization, with concentration on the: � InstructionAPI and InstructionSemanticsAPI � ParseAPI � DataflowAPI � Co-organized a series of high successful workshops with Prof. John Mellor-Crummey, Rice University, on Performance Tools for Petascale Computing, held in Snowbird, Utah and Lake Tahoe, California in July or August of 2007 through 2012. � Investigated the use of multicore in numerical libraries � Dyninst porting to 32- and 64bit Power/PowerPC (including BlueGene) and 32- and 64-bit Pentium platforms. � Applying our toolkits to advanced problems in binary code parsing associated with dealing with legacy and malicious code.
Date: November 14, 2012
Creator: Miller, Barton P.
Partner: UNT Libraries Government Documents Department

STATUS OF DATA RETRIEVAL SYSTEM AS OF APRIL 1, 1969

Description: Battelle-Northwest is developing a general-purpose data retrieval system. The system is planned to provide almost unrestricted flexibility in application and use, with no limitations on the quantity of data, except those imposed by hardware. This data retrieval system is being programmed in Fortran-V, for batch processing on a Univac-1108. The control, search-and-retrieval, query-input, output, and startup programs have been written in preliminary form, and are in the initial debugging and test stages. Current versions of these programs are presented.
Date: June 1, 1969
Creator: Reynolds, R. L.; Engel, R. L.; Toyooka, R. T. & Wells, E. L.
Partner: UNT Libraries Government Documents Department

An investigation of exploitation versus exploration in GBEA optimization of PORS 15 and 16 Problems

Description: It was hypothesized that the variations in time to solution are driven by the competing mechanisms of exploration and exploitation.This thesis explores this hypothesis by examining two contrasting problems that embody the hypothesized tradeoff between exploration and exploitation. Plus one recall store (PORS) is an optimization problem based on the idea of a simple calculator with four buttons: plus, one, store, and recall. Integer addition and store are classified as operations, and one and memory recall are classified as terminals. The goal is to arrange a fixed number of keystrokes in a way that maximizes the numerical result. PORS 15 (15 keystrokes) represents the subset of difficult PORS problems and PORS 16 (16 keystrokes) represents the subset of PORS problems that are easiest to optimize. The goal of this work is to examine the tradeoff between exploitation and exploration in graph based evolutionary algorithm (GBEA) optimization. To do this, computational experiments are used to examine how solutions evolve in PORS 15 and 16 problems when solved using GBEAs. The experiment is comprised of three components; the graphs and the population, the evolutionary algorithm rule set, and the example problems. The complete, hypercube, and cycle graphs were used for this experiment. A fixed population size was used.
Date: May 8, 2012
Creator: Koch, Kaelynn
Partner: UNT Libraries Government Documents Department

Paralization and check pointing of GPU applications through program transformation

Description: GPUs have emerged as a powerful tool for accelerating general-purpose applications. The availability of programming languages that makes writing general-purpose applications for running on GPUs tractable have consolidated GPUs as an alternative for accelerating generalpurpose applications. Among the areas that have bene#12;ted from GPU acceleration are: signal and image processing, computational uid dynamics, quantum chemistry, and, in general, the High Performance Computing (HPC) Industry. In order to continue to exploit higher levels of parallelism with GPUs, multi-GPU systems are gaining popularity. In this context, single-GPU applications are parallelized for running in multi-GPU systems. Furthermore, multi-GPU systems help to solve the GPU memory limitation for applications with large application memory footprint. Parallelizing single-GPU applications has been approached by libraries that distribute the workload at runtime, however, they impose execution overhead and are not portable. On the other hand, on traditional CPU systems, parallelization has been approached through application transformation at pre-compile time, which enhances the application to distribute the workload at application level and does not have the issues of library-based approaches. Hence, a parallelization scheme for GPU systems based on application transformation is needed. Like any computing engine of today, reliability is also a concern in GPUs. GPUs are vulnerable to transient and permanent failures. Current checkpoint/restart techniques are not suitable for systems with GPUs. Checkpointing for GPU systems present new and interesting challenges, primarily due to the natural di#11;erences imposed by the hardware design, the memory subsystem architecture, the massive number of threads, and the limited amount of synchronization among threads. Therefore, a checkpoint/restart technique suitable for GPU systems is needed. The goal of this work is to exploit higher levels of parallelism and to develop support for application-level fault tolerance in applications using multiple GPUs. Our techniques reduce the burden of enhancing single-GPU applications to support these features. ...
Date: November 15, 2012
Creator: Solano-Quinde, Lizandro Dami#19 & Laboratory], an
Partner: UNT Libraries Government Documents Department

Multicore Architecture-aware Scientific Applications

Description: Modern high performance systems are becoming increasingly complex and powerful due to advancements in processor and memory architecture. In order to keep up with this increasing complexity, applications have to be augmented with certain capabilities to fully exploit such systems. These may be at the application level, such as static or dynamic adaptations or at the system level, like having strategies in place to override some of the default operating system polices, the main objective being to improve computational performance of the application. The current work proposes two such capabilites with respect to multi-threaded scientific applications, in particular a large scale physics application computing ab-initio nuclear structure. The first involves using a middleware tool to invoke dynamic adaptations in the application, so as to be able to adjust to the changing computational resource availability at run-time. The second involves a strategy for effective placement of data in main memory, to optimize memory access latencies and bandwidth. These capabilties when included were found to have a significant impact on the application performance, resulting in average speedups of as much as two to four times.
Date: November 28, 2011
Creator: Srinivasa, Avinash
Partner: UNT Libraries Government Documents Department

Framework for Adaptable Operating and Runtime Systems: Final Project Report

Description: In this grant, we examined a wide range of techniques for constructing high-performance con#12;gurable system software for HPC systems and its application to DOE-relevant problems. Overall, research and development on this project focused in three specifc areas: (1) software frameworks for constructing and deploying con#12;gurable system software, (2) applcation of these frameworks to HPC-oriented adaptable networking software, (3) performance analysis of HPC system software to understand opportunities for performance optimization.
Date: February 1, 2012
Creator: Bridges, Patrick G.
Partner: UNT Libraries Government Documents Department

RATMAC PRIMER

Description: The language RATMAC is a direct descendant of one of the most successful structured FORTRAN languages, rational FORTRAN, RATFOR. RATMAC has all of the characteristics of RATFOR, but is augmented by a powerful recursive macro processor which is extremely useful in generating transportable FORTRAN programs. A macro is a collection of programming steps which are associated with a keyword. This keyword uniquely identifies the macro, and whenever it appears in a RATMAC program it is replaced by the collection of steps. This primer covers the language's control and decision structures, macros, file inclusion, symbolic constants, and error messages.
Date: October 1, 1980
Creator: Munn, R.J.; Stewart, J.M.; Norden, A.P. & Pagoaga, M. Katherine
Partner: UNT Libraries Government Documents Department

Scientific Data Services -- A High-Performance I/O System with Array Semantics

Description: As high-performance computing approaches exascale, the existing I/O system design is having trouble keeping pace in both performance and scalability. We propose to address this challenge by adopting database principles and techniques in parallel I/O systems. First, we propose to adopt an array data model because many scientific applications represent their data in arrays. This strategy follows a cardinal principle from database research, which separates the logical view from the physical layout of data. This high-level data model gives the underlying implementation more freedom to optimize the physical layout and to choose the most effective way of accessing the data. For example, knowing that a set of write operations is working on a single multi-dimensional array makes it possible to keep the subarrays in a log structure during the write operations and reassemble them later into another physical layout as resources permit. While maintaining the high-level view, the storage system could compress the user data to reduce the physical storage requirement, collocate data records that are frequently used together, or replicate data to increase availability and fault-tolerance. Additionally, the system could generate secondary data structures such as database indexes and summary statistics. We expect the proposed Scientific Data Services approach to create a “live” storage system that dynamically adjusts to user demands and evolves with the massively parallel storage hardware.
Date: September 21, 2011
Creator: Wu, Kesheng; Byna, Surendra; Rotem, Doron & Shoshani, Arie
Partner: UNT Libraries Government Documents Department

Web-based Tool Identifies and Quantifies Potential Cost Savings Measures at the Hanford Site - 14366

Description: The Technical Improvement system is an approachable web-based tool that is available to Hanford DOE staff, site contractors, and general support service contractors as part of the baseline optimization effort underway at the Hanford Site. Finding and implementing technical improvements are a large part of DOE’s cost savings efforts. The Technical Improvement dashboard is a key tool for brainstorming and monitoring the progress of submitted baseline optimization and potential cost/schedule efficiencies. The dashboard is accessible to users over the Hanford Local Area Network (HLAN) and provides a highly visual and straightforward status to management on the ideas provided, alleviating the need for resource intensive weekly and monthly reviews.
Date: January 9, 2014
Creator: Renevitz, Marisa J.; Peschong, Jon C.; Charboneau, Briant L. & Simpson, Brett C.
Partner: UNT Libraries Government Documents Department

PDP-7 HYBRID COMPUTER PROGRAM RECTANGULAR INTEGRATION SUBROUTINE

Description: The hybrid subroutine library contains three subroutines to be used in performing a rectangular integration, RINT$, CNP$, and ICL$. Subroutine RINT$ is used to perform the calculation of the integral and can be used in either synchronous or asynchronous mode. The purpose of CNP$ is for calculation {Delta}t , the time between integral calculations, for an asynchronous integration . Subroutine ICL$ is for initialization of the real-time clock and also contains the save and restore portions of the clock service subroutine.
Date: October 1, 1969
Creator: Gerhardstein,, L. H.
Partner: UNT Libraries Government Documents Department

PDP-7 HYBRID COMPUTER PROGRAM, VARIABLE MONITORING AND TRIP SYSTEM PROGRAM TRIP$

Description: TRIP$ provides the capability of monitoring any number of program variables comparing their values to some pre-determined trip point. If the trip point is reached a selected user subroutine will be entered continuously until the trip system is re-initialized.
Date: October 1, 1969
Creator: Worth, Grant A.
Partner: UNT Libraries Government Documents Department

DOE Hanford Network Upgrades and Disaster Recovery Exercise Support the Cleanup Mission Now and into the Future - 14303

Description: In 2013, the U.S. Department of Energy�s (DOE) Hanford Site, located in Washington State, funded an update to the critical network infrastructure supporting the Hanford Federal Cloud (HFC). The project, called ET-50, was the final step in a plan that was initiated five years ago called �Hanford�s IT Vision, 2015 and Beyond.� The ET-50 project upgraded Hanford�s core data center switches and routers along with a majority of the distribution layer switches. The upgrades allowed HFC the network intelligence to provide Hanford with a more reliable and resilient network architecture. The culmination of the five year plan improved network intelligence and high performance computing as well as helped to provide 10Gbps capable links between core backbone devices (10 times the previous bandwidth). These improvements allow Hanford the ability to further support bandwidth intense applications, such as video teleconferencing. The ET-50 switch upgrade, along with other upgrades implemented from the five year plan, have prepared Hanford�s network for the next evolution of technology in voice, video, and data. Hand-in-hand with ET-50�s major data center outage, Mission Support Alliance�s (MSA) Information Management (IM) organization executed a disaster recovery (DR) exercise to perform a true integration test and capability study. The DR scope was planned within the constraints of ET-50�s 14 hour datacenter outage window. This DR exercise tested Hanford�s Continuity of Operations (COOP) capability and failover plans for safety and business critical Hanford Federal Cloud applications. The planned suite of services to be tested was identified prior to the outage and plans were prepared to test the services ability to failover from the primary Hanford datacenter to the backup datacenter. The services tested were: � Core Network (backbone, firewall, load balancers) � Voicemail, � Voice over IP (VoIP) � Emergency Notification � Virtual desktops and; � Select set of production applications and ...
Date: November 7, 2013
Creator: Eckman, Todd J.; Hertzel, Ali K. & Lane, James J.
Partner: UNT Libraries Government Documents Department