# 1. Installing Maven & Pandas First, please install Java (JDK11) and Python 3 if they are not already. Next, make sure that Maven (for importing JGraphT) and Pandas(for data analysis) are installed. To install Maven on Ubuntu, type the following commands on terminal: sudo apt-get update sudo apt-get install maven For Pandas, type the following: pip3 install pandas ( sudo apt-get install python3-pip if pip is not installed already) # 2. Compilation Type the following to compile this project: mvn compile # 3. Running the code Below is the command for running tests for SNAP(DIMACS) and grid data. java -Xms24G -Xmx48G -Xmn36G -Xss1G -cp $CLASSPATHS shell.TestSNAP (the filename of data; just the name and not the path) (# of tests) (randomization seed) java -Xms32G -Xmx64G -Xmn48G -Xss1G -cp $CLASSPATHS shell.TestGrid (Maximum dimension) (dimension increment) [List of the values for k, space-separated] You may change the randomization seed (vertex selection) to assess reproducibility. (In our experiment, the seed was set to 2021.) For the data, check "src/SNAP(or DIMACS)". Output "test_result.csv" will be saved on "target" directory. Check if 'CLASSPATHS' is set properly. Please refer to " sample.sh " for examples & further details. #4. Obtaining average processing time and diversity First, move to the target directory. Then run get_averages.py python3 get_averages (.csv file name) [list of the values for k, space-separated. Optional parameter.]
Diverse graph algorithms implemented using JGraphT library.
Overview
[内测中]前向式Python环境快捷封装工具,快速将Python打包为EXE并添加CUDA、NoAVX等支持。
QPT - Quick packaging tool 快捷封装工具 GitHub主页 | Gitee主页 QPT是一款可以“模拟”开发环境的多功能封装工具,最短只需一行命令即可将普通的Python脚本打包成EXE可执行程序,并选择性添加CUDA和NoAVX的支持,尽可能兼容更多的用户环境。 感觉还可
[ICCV2021] Safety-aware Motion Prediction with Unseen Vehicles for Autonomous Driving
Safety-aware Motion Prediction with Unseen Vehicles for Autonomous Driving Safety-aware Motion Prediction with Unseen Vehicles for Autonomous Driving
Neural Scene Flow Fields for Space-Time View Synthesis of Dynamic Scenes
Neural Scene Flow Fields PyTorch implementation of paper "Neural Scene Flow Fields for Space-Time View Synthesis of Dynamic Scenes", CVPR 2021 [Projec
Towards Understanding Quality Challenges of the Federated Learning: A First Look from the Lens of Robustness
FL Analysis This repository contains the code and results for the paper "Towards Understanding Quality Challenges of the Federated Learning: A First L
MetaShift: A Dataset of Datasets for Evaluating Contextual Distribution Shifts and Training Conflicts (ICLR 2022)
MetaShift: A Dataset of Datasets for Evaluating Distribution Shifts and Training Conflicts This repo provides the PyTorch source code of our paper: Me
Neuralnetwork - Basic Multilayer Perceptron Neural Network for deep learning
Neural Network Just a basic Neural Network module Usage Example Importing Module
Dynamics-aware Adversarial Attack of 3D Sparse Convolution Network
Leaded Gradient Method (LGM) This repository contains the PyTorch implementation for paper Dynamics-aware Adversarial Attack of 3D Sparse Convolution
A Flow-based Generative Network for Speech Synthesis
WaveGlow: a Flow-based Generative Network for Speech Synthesis Ryan Prenger, Rafael Valle, and Bryan Catanzaro In our recent paper, we propose WaveGlo
Source code for our paper "Empathetic Response Generation with State Management"
Source code for our paper "Empathetic Response Generation with State Management" this repository is maintained by both Jun Gao and Yuhan Liu Model Ove
Unsupervised Feature Ranking via Attribute Networks.
FRANe Unsupervised Feature Ranking via Attribute Networks (FRANe) converts a dataset into a network (graph) with nodes that correspond to the features
Benchmark datasets, data loaders, and evaluators for graph machine learning
Overview The Open Graph Benchmark (OGB) is a collection of benchmark datasets, data loaders, and evaluators for graph machine learning. Datasets cover
Synthesize photos from PhotoDNA using machine learning 🌱
Ribosome Synthesize photos from PhotoDNA. See the blog post for more information. Installation Dependencies You can install Python dependencies using
Create Data & AI apps in 20 lines of code with Shimoku
Install with: pip install shimoku-api-python Start with: from os import getenv import shimoku_api_python.client as Shimoku
A large dataset of 100k Google Satellite and matching Map images, resembling pix2pix's Google Maps dataset.
Larger Google Sat2Map dataset This dataset extends the aerial ⟷ Maps dataset used in pix2pix (Isola et al., CVPR17). The provide script download_sat2m
Space Ship Simulator using python
FlyOver Basic space-ship simulator using python How to run? Just double click run.py What modules do i need? All modules that i currently using is bui
An Image compression simulator that uses Source Extractor and Monte Carlo methods to examine the post compressive effects different compression algorithms have.
ImageCompressionSimulation An Image compression simulator that uses Source Extractor and Monte Carlo methods to examine the post compressive effects o
Official source code to CVPR'20 paper, "When2com: Multi-Agent Perception via Communication Graph Grouping"
When2com: Multi-Agent Perception via Communication Graph Grouping This is the PyTorch implementation of our paper: When2com: Multi-Agent Perception vi
Learning to Communicate with Deep Multi-Agent Reinforcement Learning in PyTorch
Learning to Communicate with Deep Multi-Agent Reinforcement Learning This is a PyTorch implementation of the original Lua code release. Overview This
[ACL 2022] LinkBERT: A Knowledgeable Language Model 😎 Pretrained with Document Links
LinkBERT: A Knowledgeable Language Model Pretrained with Document Links This repo provides the model, code & data of our paper: LinkBERT: Pretraining
A Multi-modal Perception Tracker (MPT) for speaker tracking using both audio and visual modalities
MPT A Multi-modal Perception Tracker (MPT) for speaker tracking using both audio and visual modalities. Implementation for our AAAI 2022 paper: Multi-