Why do we use HDF5? We moved to HDF5 for our simulation data in 2016 from using our own proprietary file format. HDF5 had been on our radar for some time and we spent a couple of years investigating it and other file formats before deciding which we should switch to. HDF5 met all the criteria we had at the time. Amongst the criteria were: performance in speed and size, an accepted standard for scientific data, being open source, providing additional tools....
Tobias Weinzierl, Durham University, UK, Sven Köppel, FIAS, Germany, Michael Bader, TUM, Germany, HDF Guest Bloggers
ExaHyPE develops a solver engine for hyperbolic differential equations solved on adaptive Cartesian meshes. It supports various HDF5 output formats.
Exascale computing is expected to allow scientists and engineers to simulate, and ultimately understand, wave phenomena with unprecedented accuracy over unprecedented time spans. To harvest the power of exascale machines, well-suited software however has to become available. ExaHyPE is a H2020 project writing a PDE solver engine—similar to a 3D computer game engine—that will allow groups with a decent CSE expertise to write their own solver for hyperbolic equation systems within a year. The resulting solver will scale to exascale.
This is made possible by the unique...
Internal compression is one of several powerful HDF5 features that distinguish HDF5 from other binary formats and make it very attractive for storing and organizing data. Internal HDF5 compression saves storage space and I/O bandwidth and allows efficient partial access to data. Chunk storage must be used when HDF5 compression is enabled....
Scot Martin, Harvard University, HDF Guest Blogger
HDF5 storage is really interesting. To me, its format has no fixed structure, but instead is based on introspection and discovery. Seems great to me; Mathematica has its origins first in artificial intelligence, so we ought to be able to do something here. Approaching twenty-two years with Mathematica and almost a “Hello, World!” ability in C, I decided to jump right in. Enter The HDF Group's P/Invoke for my salvation. Here’s how we make use of it in Mathematica:
Bang! Ready to go in Mathematica. Here’s a proof of concept for how it works:
(* The three symbols should have initial values so that there is *)
(* memory allocation when Mathematica interfaces with P/Invoke. *)
Christian Hoene, Symonics GmbH; and Piotr Majdak, Acoustics Research Institute; HDF Guest Bloggers
Spatial audio - 3D sound. Back in the ‘70’s, “dummy head” microphones were used to create spatial audio recordings. With headphones, one was able to listen to those recordings and marvel at the impressive spatial distribution of sounds – just like in real life.
[caption id="attachment_11132" align="aligncenter" width="624"] Displays the difference between listening to a real source and listening to realistic virtual sounds via headphones[/caption]
Nowadays, we have a much better understanding of the human binaural perception and we can even simulate spatial audio signals with the help of computers. Indeed, a modern virtual reality (VR) headset such as the Oculus Rift or Samsung Gear utilizes 3D audio to allow...
The HDF Server allows producers of complex datasets to share their results with a wide audience base. We used it to develop the Global Fire Emissions Database (GFED) Analysis Tool, a website which guides the user through our dataset. A simple webmap interface allows users to select an area of interest and produce data visualization charts. ...
Mark Miller, Lawrence Livermore National Laboratory, Guest Blogger
The HDF5 library has supported the I/O requirements of HPC codes at Lawrence Livermore National Labs (LLNL) since the late 90’s. In particular, HDF5 used in the Multiple Independent File (MIF) parallel I/O paradigm has supported LLNL code’s scalable I/O requirements and has recently been gainfully used at scales as large as 1,000,000 parallel tasks.
What is the MIF Parallel I/O Paradigm?
In the MIF paradigm, a computational object (an array, a mesh, etc.) is decomposed into pieces and distributed, perhaps unevenly, over parallel tasks. For I/O, the tasks are organized into groups and each group writes one file using round-robin exclusive access for the tasks in the group. Writes within groups are serialized but...
David Pearah, The HDF Group
Hello again, HDF User Community,
As I mentioned in my last blog post -- HDF: The Next 30 Years (Part 2) -- we're looking for ways to better engage our users, which includes providing better tools for you to get support from the HDF Community. We are looking for your input on three things: the HDF User Forum, Roadmap for HDF5, and Focus Groups - please take a few minutes to complete these short surveys and let us know what you think!
Discussion Board and Listserv Forum Feedback: LINK We currently provide a listserv as the primary way that folks post questions and feedback to the community, and we would like your feedback since we're looking at web-based tools to complement or...
David Pearah, The HDF Group
Hello HDF Community!
Thanks for the warm welcome into the HDF family: in my 4+ months as the new CEO, I've been blown away by your passion, diversity of interests and applications, and willingness to provide feedback on: 1. why you use HDF5?, and 2. how can HDF5 be improved? I also want to thank my predecessor Mike Folk for his invaluable and ongoing support.
The HDF community is growing fast: when I last checked, there are nearly 700 HDF5 projects in GitHub! I've had the privilege of connecting via phone/web with dozens of you over the past few months. Across all of my discussions, one piece of feedback came back loud and clear: The HDF Group needs to be more engaged with its users and help foster...