- Scientific Computing and Data Management
- Distributed and Parallel Computing Systems
- Research Data Management Practices
- Semantic Web and Ontologies
- Advanced Database Systems and Queries
- Service-Oriented Architecture and Web Services
- Bioinformatics and Genomic Networks
- Power Systems and Technologies
- Genetics, Bioinformatics, and Biomedical Research
- Gene expression and cancer classification
- Biomedical Text Mining and Ontologies
- Metabolomics and Mass Spectrometry Studies
- Advanced Computational Techniques and Applications
- Data Visualization and Analytics
- Time Series Analysis and Forecasting
- Cloud Computing and Resource Management
- Big Data and Business Intelligence
- Web Data Mining and Analysis
- Spectroscopy Techniques in Biomedical and Chemical Research
- Data Management and Algorithms
- Advanced Proteomics Techniques and Applications
- Genomics and Phylogenetic Studies
- Real-Time Systems Scheduling
- Communication in Education and Healthcare
- Data Stream Mining Techniques
Pacific Northwest National Laboratory
2013-2024
Battelle
2018
San Diego Supercomputer Center
2017
University of California, San Diego
2017
Creative Commons
2017
Family Health International 360
2017
Rensselaer Polytechnic Institute
2016
Hôpital Louis Pradel
2009
University of Southampton
2007
Argonne National Laboratory
2007
Abstract The first Provenance Challenge was set up in order to provide a forum for the community understand capabilities of different provenance systems and expressiveness their representations. To this end, functional magnetic resonance imaging workflow defined, which participants had either simulate or run produce some representation, from identified queries be implemented executed. Sixteen teams responded challenge, submitted inputs. In paper, we present challenge queries, summarize...
Modern electric power distribution systems are data rich and include growing numbers of distributed energy resources automation. To take advantage automation manage penetrations resources, utilities need applications for planning operations that use all available may incorporate approaches to operate control. The industry would benefit from management based on a common platform makes each type interchangeable. This paper describes an approach enabling cost-effective development deployment...
Abstract Summary: The Bioinformatics Resource Manager (BRM) is a software environment that provides the user with data management, retrieval and integration capabilities. Designed in collaboration biologists, BRM simplifies mundane analysis tasks of merging microarray proteomic across platforms, facilitates users’ functional annotation interaction from public sources connectivity to visual analytic tools through reformatting for easy import or dynamic launching capability. developed using...
The collaboratory for multiscale chemical science (CMCS) is developing a powerful informatics-based approach to synthesizing information support systems-based research and applying it in of combustion research. An open source informatics toolkit being developed that addresses number issues core the emerging concept knowledge grids including provenance tracking lightweight federation data application resources into cross-scale flows. CMCS portal currently use by high-profile pilot groups...
Accurate and calibrated directional-hemispherical reflectance spectra of solids are important for both in situ remote sensing. Many the form powders or granules to measure their diffuse laboratory, it is often necessary place samples behind a transparent medium such as glass ultraviolet (UV), visible, near-infrared spectral regions. Using experimental methods simple optical model, we demonstrate that (fused quartz our case) leads artifacts values. We report observations measured values,...
One of the primary goals organizations in a free enterprise system is to be competitive. Although measures competitiveness may appear different for manufacturing and service or government religious organizations, order competitive, any organization must provide products services which customers clients are willing pay fair return price. In long run, system, measured by ability stay business protect organization's investments, earn on those ensure jobs future.
We propose an approach for improved reproducibility that includes capturing and relating provenance characteristics performance metrics. discuss two use cases: scientific of results in the Energy Exascale Earth System Model (E3SM—previously ACME) molecular dynamics workflows on HPC platforms. To capture persist data these workflows, we have designed developed Chimbuko ProvEn frameworks. captures enables detailed single workflow analysis. is a hybrid, queryable system storing analyzing...
As Raman spectroscopy continues to evolve, questions arise as the portability of data: dispersive versus Fourier transform, wavelength calibration, intensity and in particular frequency excitation laser. While concerns about fluorescence visible or ultraviolet, most modern (portable) systems use near-infrared lasers, many these are relatively close wavelength. We have investigated possibility porting reference data sets from one NIR system another: constructed a library consisting 145...
Challenges that make it difficult to find, share, and combine published data, such as data heterogeneity resource discovery, have led increased adoption of semantic standards publishing technologies. To more accessible, interconnected discoverable, some domains are being encouraged publish their Linked Data. Consequently, this trend greatly increases the amount web tools required process, store, interconnect. In attempting process manipulate large sets, -- ranging from simple text editors...
As high performance computing (HPC) infrastructures continue to grow in capability and complexity, so do the applications that they serve. HPC distributed-area (DAC) (e.g. grid cloud) users are looking increasingly toward workflow solutions orchestrate their complex application coupling, pre- post-processing needs. To end, US Department of Energy Integrated end-to-end Performance Prediction Diagnosis for Extreme Scientific Workflows (IPPD) project is currently investigating an integrated...
Abstract Workflows and data pipelines are becoming increasingly valuable to computational experimental sciences. These automated systems capable of generating significantly more within the same amount time compared their manual counterparts. Automatically capturing recording provenance annotation as part these workflows critical for management, verification, dissemination. We have been prototyping a workflow system, targeted at biological workflows, that extends our content management...
The evolving nature of electric power distribution systems is motivating the development advanced applications for utility planning, operation, management, and control. These model-based interact with multiple subsystems within software platforms referred to as management systems. Recently, Common Information Model has been used enable data model standardization across several entities. fidelity key leading-edge, scalable algorithms assessment emerging technologies such distributed energy...
Although drawings and graphs can express complex concepts, such methods can't be computationally analyzed. Researchers have developed visual approaches paradigms that scientists use to capture represent scientific models in a computational form they link integrate with data sets applications.
For Raman spectroscopy the ability to detect is often limited by existence and quality of reference library which field spectra are compared. Developing such databases labor- resource-intensive; typically generated data not transferred other instruments. Still considerations may exist for comparing at visible ultraviolet excitation wavelengths as resonance enhancement. However, common near-infrared 785, 830, 960, 1047 1064 nm where this normally a lesser concern, it logical consider whether...
The Belle II experiment at KEK is preparing for first collisions in 2017. Processing the large amounts of data that will be produced require conditions to readily available systems worldwide a fast and efficient manner straightforward both user maintainer. database was designed with goal: make it as easily maintainable possible. To this end, HEP-specific software tools were avoided much possible industry standard used instead. HTTP REST services selected application interface, which provide...