Simulation code and tutorial for BBHnet training data

Overview

Simulation Dataset for BBHnet

NOTE: OLD README, UPDATE IN PROGRESS

We generate simulation dataset to train BBHnet, our deep learning framework for detection of compact binary coalescene (CBC) gravitational-wave (GW) signals .

Example

To generate a noise dataset, simply run generateRealNoise.py:

python generateRealNoise.py -t0 1186729980 -t1 1186734086 -t0-psd 1186729980 -t1-psd 1186734086
    -fs 1024 -fmin 20 -o test_noise.h5

To also add CBC signals, enable the flag -S and add the prior distribution file in Bilby format with -p

python generateRealNoise.py -t0 1186729980 -t1 1186734086 -t0-psd 1186729980 -t1-psd 1186734086
    -fs 1024 -fmin 20 -S -p config/priors/nonspin_BBH.prior -o test_signal.h5

A full list of generateRealNoise.py arguments can be found below:

usage: generateRealNoise.py [-h] -t0 FRAME_START -t1 FRAME_STOP -t0-psd FRAME_START_PSD -t1-psd FRAME_STOP_PSD -o OUTFILE [-S]
                            [-fs SAMPLE_RATE] [-fmin HIGH_PASS] [-T SAMPLE_DURATION] [-dt TIME_STEP] [-p PRIOR_FILE]
                            [--correlation-shift CORRELATION_SHIFT] [--min-trigger MIN_TRIGGER] [--max-trigger MAX_TRIGGER]
                            [-s SEED]

optional arguments:
  -h, --help            show this help message and exit
  -t0 FRAME_START, --frame-start FRAME_START
                        starting GPS time of strain
  -t1 FRAME_STOP, --frame-stop FRAME_STOP
                        stopping GPS time of strain
  -t0-psd FRAME_START_PSD, --frame-start-psd FRAME_START_PSD
                        starting GPS time of strain for PSD estimation
  -t1-psd FRAME_STOP_PSD, --frame-stop-psd FRAME_STOP_PSD
                        stopping GPS time of strain for PSD estimation
  -o OUTFILE, --outfile OUTFILE
                        path to write output file in HDF5 format
  -S, --signal          Enable to add GW signal on top of background noise
  -fs SAMPLE_RATE, --sample-rate SAMPLE_RATE
                        sampling rate of strain
  -fmin HIGH_PASS, --high-pass HIGH_PASS
                        frequency of highpass filter
  -T SAMPLE_DURATION, --sample-duration SAMPLE_DURATION
                        duration in seconds of each sample
  -dt TIME_STEP, --time-step TIME_STEP
                        time step size in seconds between consecutive samples
  -p PRIOR_FILE, --prior-file PRIOR_FILE
                        path to prior config file. Required for signal simulation
  --correlation-shift CORRELATION_SHIFT
                        if given, also compute the correlation with given shift value
  --min-trigger MIN_TRIGGER
                        mininum trigger time w.r.t to sample. must be within [0, sample_duration]
  --max-trigger MAX_TRIGGER
                        maximum trigger time w.r.t to sample. must be within [0, sample_duration]
  -s SEED, --seed SEED  random seed for reproducibility

A Fast and Stable GAN for Small and High Resolution Imagesets - pytorch

A Fast and Stable GAN for Small and High Resolution Imagesets - pytorch The official pytorch implementation of the paper "Towards Faster and Stabilize

Bingchen Liu 455 Jan 08, 2023
Voice Gender Recognition

In this project it was used some different Machine Learning models to identify the gender of a voice (Female or Male) based on some specific speech and voice attributes.

Anne Livia 1 Jan 27, 2022
Out-of-Distribution Generalization of Chest X-ray Using Risk Extrapolation

OoD_Gen-Chest_Xray Out-of-Distribution Generalization of Chest X-ray Using Risk Extrapolation Requirements (Installations) Install the following libra

Enoch Tetteh 2 Oct 01, 2022
Self-training with Weak Supervision (NAACL 2021)

This repo holds the code for our weak supervision framework, ASTRA, described in our NAACL 2021 paper: "Self-Training with Weak Supervision"

Microsoft 148 Nov 20, 2022
Multi Task Vision and Language

12-in-1: Multi-Task Vision and Language Representation Learning Please cite the following if you use this code. Code and pre-trained models for 12-in-

Facebook Research 712 Dec 19, 2022
Towards Understanding Quality Challenges of the Federated Learning: A First Look from the Lens of Robustness

FL Analysis This repository contains the code and results for the paper "Towards Understanding Quality Challenges of the Federated Learning: A First L

3 Oct 17, 2022
Pytorch port of Google Research's LEAF Audio paper

leaf-audio-pytorch Pytorch port of Google Research's LEAF Audio paper published at ICLR 2021. This port is not completely finished, but the Leaf() fro

Dennis Fedorishin 80 Oct 31, 2022
🐤 Nix-TTS: An Incredibly Lightweight End-to-End Text-to-Speech Model via Non End-to-End Distillation

🐤 Nix-TTS An Incredibly Lightweight End-to-End Text-to-Speech Model via Non End-to-End Distillation Rendi Chevi, Radityo Eko Prasojo, Alham Fikri Aji

Rendi Chevi 156 Jan 09, 2023
[ICML 2021] “ Self-Damaging Contrastive Learning”, Ziyu Jiang, Tianlong Chen, Bobak Mortazavi, Zhangyang Wang

Self-Damaging Contrastive Learning Introduction The recent breakthrough achieved by contrastive learning accelerates the pace for deploying unsupervis

VITA 51 Dec 29, 2022
An implementation of shampoo

shampoo.pytorch An implementation of shampoo, proposed in Shampoo : Preconditioned Stochastic Tensor Optimization by Vineet Gupta, Tomer Koren and Yor

Ryuichiro Hataya 69 Sep 10, 2022
High dimensional black-box optimizer using Latent Action Monte Carlo Tree Search algorithm

LA-MCTS The code is based of paper Learning Search Space Partition for Black-box Optimization using Monte Carlo Tree Search. Component LA-MCTS has thr

Meta Research 18 Oct 24, 2022
Official implementation of NeurIPS 2021 paper "One Loss for All: Deep Hashing with a Single Cosine Similarity based Learning Objective"

Official implementation of NeurIPS 2021 paper "One Loss for All: Deep Hashing with a Single Cosine Similarity based Learning Objective"

Ng Kam Woh 71 Dec 22, 2022
RETRO-pytorch - Implementation of RETRO, Deepmind's Retrieval based Attention net, in Pytorch

RETRO - Pytorch (wip) Implementation of RETRO, Deepmind's Retrieval based Attent

Phil Wang 556 Jan 04, 2023
Symbolic Parallel Adaptive Importance Sampling for Probabilistic Program Analysis in JAX

SYMPAIS: Symbolic Parallel Adaptive Importance Sampling for Probabilistic Program Analysis Overview | Installation | Documentation | Examples | Notebo

Yicheng Luo 4 Sep 13, 2022
DIR-GNN - Discovering Invariant Rationales for Graph Neural Networks

DIR-GNN "Discovering Invariant Rationales for Graph Neural Networks" (ICLR 2022)

Ying-Xin (Shirley) Wu 70 Nov 13, 2022
Code for a seq2seq architecture with Bahdanau attention designed to map stereotactic EEG data from human brains to spectrograms, using the PyTorch Lightning.

stereoEEG2speech We provide code for a seq2seq architecture with Bahdanau attention designed to map stereotactic EEG data from human brains to spectro

15 Nov 11, 2022
Script that receives an Image (original) and a set of images to be used as "pixels" in reconstruction of the Original image using the set of images as "pixels"

picinpics Script that receives an Image (original) and a set of images to be used as "pixels" in reconstruction of the Original image using the set of

RodrigoCMoraes 1 Oct 24, 2021
Implementation of the 😇 Attention layer from the paper, Scaling Local Self-Attention For Parameter Efficient Visual Backbones

HaloNet - Pytorch Implementation of the Attention layer from the paper, Scaling Local Self-Attention For Parameter Efficient Visual Backbones. This re

Phil Wang 189 Nov 22, 2022
Website for D2C paper

D2C This is the repository that contains source code for the D2C Website. If you find D2C useful for your work please cite: @article{sinha2021d2c au

1 Oct 21, 2021
Official PyTorch implementation of "Physics-aware Difference Graph Networks for Sparsely-Observed Dynamics".

Physics-aware Difference Graph Networks for Sparsely-Observed Dynamics This repository is the official PyTorch implementation of "Physics-aware Differ

USC-Melady 46 Nov 20, 2022