Reading and writing time series data¶
The TimeSeries
object includes read()
and
write()
methods to enable reading from and writing to files
respectively.
For example, to read from an ASCII file containing time and amplitude columns:
>>> from gwpy.timeseries import TimeSeries
>>> data = TimeSeries.read('my-data.txt')
TimeSeries.read()
will attempt to automatically identify the file
format based on the file extension and/or the contents of the file, however,
the format
keyword argument can be used to manually identify the input
file-format.
The read()
and write()
methods take
different arguments and keywords based on the input/output file format,
see Built-in file formats for details on reading/writing for
each of the built-in formats.
Automatic discovery of GW detector data¶
GW detector data¶
Gravitational-wave detector data, including all engineering diagnostic data
as well as the calibrated ‘strain’ data that are searched for GW signals,
are archived in GWF files stored at the
relevant observatory.
These data are available locally to authenticated users of the associated
computing centres (typically collaboration members), but are also
distributed using CVMFS and are available remotely using nds2
.
Access to these data is restricted to active collaboration members.
Additionally The Gravitational-Wave Open Science Centre (GWOSC) hosts publicly-accessible ‘open’ data, with event datasets made available at the same time as the relevant result publication and typically including ~1 hour of data around each published event detection, and bulk datasets with the entire observing run data available roughly 18 months after the end of the run.
GWOSC also hosts the Auxiliary Channel Three Hour Release, providing public access to
environmental sensor data around GW170814.
These data are freely accessible using nds2
.
Data discovery methods¶
Built-in file formats¶
ASCII¶
GWpy supports writing TimeSeries
(and FrequencySeries
) data to ASCII in a two-column time
and amplitude
format.
Reading¶
To read a TimeSeries
from ASCII:
>>> t = TimeSeries.read('data.txt')
See numpy.loadtxt()
for keyword argument options.
Writing¶
To write a TimeSeries
to ASCII:
>>> t.write('data.txt')
See numpy.savetxt()
for keyword argument options.
GWF¶
Additional dependencies: LDAStools.frameCPP
or framel
or lalframe
The raw observatory data are archived in .gwf
files, a custom binary format that efficiently stores the time streams and all necessary metadata, for more details about this particular data format, take a look at the specification document LIGO-T970130.
Reading¶
To read data from a GWF file, pass the input file path (or paths) and the name of the data channel to read:
>>> data = TimeSeries.read('HLV-HW100916-968654552-1.gwf', 'L1:LDAS-STRAIN')
Note
The HLV-HW100916-968654552-1.gwf
file is included with the GWpy source under /gwpy/testing/data/.
Reading a StateVector
uses the same syntax:
>>> data = StateVector.read('my-state-data.gwf', 'L1:GWO-STATE_VECTOR')
Multiple files can be read by passing a list of files:
>>> data = TimeSeries.read([file1, file2], 'L1:LDAS-STRAIN')
When reading multiple files, the nproc
keyword argument can be used to distribute the reading over multiple CPUs, which should make it faster:
>>> data = TimeSeries.read([file1, file2, file3, file4], 'L1:LDAS-STRAIN', nproc=2)
The above command will separate the input list of 4 file paths into two sets of 2 files, combining the results into a single TimeSeries
before returning.
The start
and end
keyword arguments can be used to downselect data to a specific [start, end)
time segment when reading:
>>> data = TimeSeries.read('HLV-HW100916-968654552-1.gwf', 'L1:LDAS-STRAIN', start=968654552.5, end=968654553)
Additionally, the following keyword arguments can be used:
Warning
These keyword arguments are only supported when using the
LDAStools.frameCPP
GWF API.
Keyword |
Type |
Default |
Usage |
---|---|---|---|
|
Apply ADC calibration when reading |
||
|
|
Reading multiple channels¶
To read multiple channels from one or more GWF files (rather than opening and closing the files multiple times), use the TimeSeriesDict
or StateVectorDict
classes, and pass a list of data channel names:
>>> data = TimeSeriesDict.read('HLV-HW100916-968654552-1.gwf', ['H1:LDAS-STRAIN', 'L1:LDAS-STRAIN'])
In this case the resample
and dtype
keywords can be given as a single value used for all data channels, or a dict
mapping an argument for each data channel name:
>>> data = TimeSeriesDict.read('HLV-HW100916-968654552-1.gwf', ['H1:LDAS-STRAIN', 'L1:LDAS-STRAIN'],
... resample={'H1:LDAS-STRAIN': 2048})
The above example will resample only the 'H1:LDAS-STRAIN'
TimeSeries
and will not modify that for 'L1:LDAS-STRAIN'
.
Note
A mix of TimeSeries
and StateVector
objects can be read by using only TimeSeriesDict
class, and casting the returned data to a StateVector
using view()
.
Writing¶
To write data held in any of the gwpy.timeseries
classes to a GWF file, simply use:
>>> data.write('output.gwf')
If the output file already exists it will be overwritten.
GWF library availability¶
(Last we checked…) The three GWF library interfaces are available on the following platforms:
Library |
Conda-forge package name |
Linux |
macOS |
Windows |
---|---|---|---|---|
|
Yes |
Yes |
No |
|
|
Yes |
Yes |
Yes |
|
|
Yes |
Yes |
No |
HDF5¶
GWpy allows storing data in HDF5 format files, using a custom specification for storage of metadata.
Warning
To read GWOSC data from HDF5, please see HDF5 (GWOSC).
Reading¶
To read TimeSeries
or StateVector
data held in HDF5 files pass the filename (or filenames) or the source, and the path of the data inside the HDF5 file:
>>> data = TimeSeries.read('HLV-HW100916-968654552-1.hdf', 'L1:LDAS-STRAIN')
As with GWF, the start
and end
keyword arguments can be used to downselect data to a specific [start, end)
time segment when reading:
>>> data = TimeSeries.read('HLV-HW100916-968654552-1.hdf', 'L1:LDAS-STRAIN', start=968654552.5, end=968654553)
Analogously to GWF, you can read multiple TimeSeries
from an HDF5 file via TimeSeriesDict.read()
:
>>> data = TimeSeriesDict.read('HLV-HW100916-968654552-1.hdf')
By default, all matching datasets in the file will be read, to restrict the output, specify the names of the datasets you want:
>>> data = TimeSeriesDict.read('HLV-HW100916-968654552-1.hdf', ['H1:LDAS-STRAIN', 'L1:LDAS-STRAIN'])
Writing¶
Data held in a TimeSeries
, TimeSeriesDict, `StateVector
, or StateVectorDict
can be written to an HDF5 file via:
>>> data.write('output.hdf')
The output argument ('output.hdf'
) can be a file path, an open h5py.File
object, or a h5py.Group
object, to append data to an existing file.
If the target file already exists, an IOError
will be raised, use overwrite=True
to force a new file to be written.
To write a TimeSeries
to an existing file, use append=True
:
>>> data.write('output.hdf', append=True)
To replace an existing dataset in an existing file, while preserving other data, use both append=True
and overwrite=True
:
>>> data.write('output.hdf', append=True, overwrite=True)
HDF5 (GWOSC)¶
GWOSC write data in HDF5 using a custom schema that is incompatible
with format='hdf5'
.
Reading¶
GWpy can read data from GWOSC HDF5 files using the format='hdf5.gwosc'
keyword:
>>> data = TimeSeries.read(
... "H-H1_GWOSC_16KHZ_R1-1187056280-4096.hdf5",
... format="hdf5.gwosc",
... )
By default, TimeSeries.read()
will return the contents of the
/strain/Strain
dataset, while StateVector.read()
will return those
of /quality/simple
.
As with regular HDF5, the start
and end
keyword arguments can be used
to downselect data to a specific [start, end)
time segment when reading.
WAV¶
Any TimeSeries
can be written to / read from a WAV file using TimeSeries.read()
:
Warning
No metadata are stored in the WAV file except the sampling rate, so any units or GPS timing information are lost when converting to/from WAV.
Reading¶
To read a TimeSeries
from WAV:
>>> t = TimeSeries.read('data.wav')
See scipy.io.wavfile.read()
for any keyword argument options.
Writing¶
To write a TimeSeries
to WAV:
>>> t.write('data.wav')
See scipy.io.wavfile.write()
for keyword argument options.