/NeuralSNR

A feed-forward neural network for fast SNR calculation

Primary LanguagePython

NeuralSNR

Gravitational-wave signal-to-noise interpolation via neural networks

Computing signal-to-noise ratios (SNRs) is one of the most common tasks in gravitational-wave data analysis. While a single SNR evaluation is generally fast, computing SNRs for an entire population of merger events could be time consuming. We compute SNRs for aligned-spin binary black-hole mergers as a function of the (detector-frame) total mass, mass ratio and spin magnitudes using selected waveform models and detector noise curves, then we interpolate the SNRs in this four-dimensional parameter space with a simple neural network (a multilayer perceptron). The trained network can evaluate 106 SNRs on a 4-core CPU within a minute with a median fractionalerror below 10−3. This corresponds to average speed-ups by factors in the range [120,7.5×104], depending on the underlying waveform model. Our trained network (and source code) is publicly available online, and it can be easily adapted to similar multidimensional interpolation problems.

This repo contains the network trained in arXiv:2007.10350. Please cite the paper if you make use of our code or of our pre-trained network.

The code is developed and maintained by Kaze Wong. Please open an issue on GitHub if you want to report bugs or make suggestions. For any other problem, feel free to contact me with kazewong@jhu.edu

The pytorch template used for training is from PyTorch Project Template

Prerequisites

This code depends on the following packages:

numpy, astropy, scipy, pytorch, h5py

Please also find relevant packages used in the pytorch project template repo.

Using pretrained model

Disclaimer: we encourage the use of our network only for prototyping purposes. For any publication-level calculation, we strongly recommend users to train their own network. The performance and accuracy of a neural network crictically depends on the training data, and there is no single set of assumptions fitting in any analysis. The goal of our paper is to encourage the community to use neural networks in more common GW tasks, not to answer particular scientific questions. SO TRAIN YOUR OWN NETWORK. You have been warned :).

Calling the pretrained model is pretty strightforward. As an example, let's call the SNR using the A+-PhenomD pretrained model.

import torch
model = torch.jit.load('./network/AplusDesign_IMRPhenomD.network',map_location='cpu')
print(model(torch.tensor([[30,1,0,0]]).float())

This would print the SNR of a 30-30 solar mass event with dimensionless spins equal to zero at 100 Mpc. The four inputs are (M_1,q,chi_1,chi_2). A more detailed tutorial of the usage of the pretrained network can be found in tutorial

Training your own network

We encourage users to train their own network. There are two steps in doing this:

  1. Preparing data
  2. Defining the training configuration

Preparing data

The users can use the script txt_to_train.py to prepare the training data needed. We need two files to do so, input.txt and output.txt. Each .txt should be in N rows and M columns, where N is the number of training sample points and M is the dimension of the input/output parameter space. txt_to_train.py will combine the two files into an .hdf5 which can then be used in training.

As an example, the user can enter the following command in the root directory of this repo:

python txt_to_train.py

This should output a file name test_random.hdf5 in the data directory.

Defining the training configuration

Once you have your data file set up, you also need to tell the computer where the relevant files are, and what network configuration you want to use. We have an example configuration file located in ~/training/config/test.json. Here are the lines you need to customize that are related to meta data

"data": "replaced this by your data path",
"output_model_path": "replaced this by your desire output path",
"exp_name": "replaced this by the tag you want to give to this training"
"cuda": false # set true if you want to use a GPU
"output_model": true # set true if you want to output a callable network
"load_checkpoint": true, # set true if you want to resume previous training session

Here are the lines related to the architechture of the network

"hidden_size": 32, number of neurons per layer
"hidden_layers": 3, number of layers
"learning_rate": 0.0001, step size of each gradient descent step
"input_size":4, number of input
"output_size": 1, number of output
"max_epoch": 100, maximum training epoch, basically how many loops the network is trained.

In general you would like to change input_size and output_size to fit your data structure. hidden_size and hidden_layers control how large is your model. In general larger models can capture more complicated behavior, but they take longer and need more data to train. max_epoch controls how many loops you want to train the network. You want to make sure this is large enough that your model has converged reasonably well.

Once you have a configuration files, you can run the following command in the training directory

python main.py config_path

This should start the training and output a network at the end of the training.

Training a neural network could be daunting if you have never done it before. Please ask for help you need any.