AprilXiaoyanLiu / dpwgan

Differentially-private Wasserstein GAN implementation in PyTorch

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

DPWGAN

Build Status Contributor Covenant License

Code for a differentially private Wasserstein GAN implemented in PyTorch

Maintenance

This project is released as-is, and is not actively maintained by Civis Analytics.

Installation

This package requires Python >= 3.5

pip install -r requirements.txt
python setup.py install

For development, also install development requirements:

pip install -r dev-requirements.txt

Usage

Setting up, training, and generating from a DPWGAN:

gan = DPWGAN(generator, discriminator, noise_function)
gan.train(data)
synthetic_data = gan.generate(100)

generator and discriminator should be torch.nn modules, and noise_function should generate random data as input to the generator. As a simple example:

# simple generator module
generator = torch.nn.Sequential(
    torch.nn.Linear(noise_dim, hidden_dim),
    torch.nn.ReLU(),
    MultiCategoryGumbelSoftmax(hidden_dim, output_dims)
)

# simple discriminator module
discriminator = torch.nn.Sequential(
    torch.nn.Linear(sum(output_dims), hidden_dim),
    torch.nn.ReLU(),
    torch.nn.Linear(hidden_dim, 1)
)

# simple noise function (input to generator module)
def noise_function(n):
    return torch.randn(n, noise_dim)

The examples folder has two scripts to demonstrate setting up, training, and generating data with a DPWGAN with categorical data sets.

simple_example.py shows how to create a generator, discriminator, and noise function, and applies the DPWGAN to a toy categorical data set.

census_example.py applies a DPWGAN to ACS PUMS data. Use create_census_data.py to download the input data set.

If you want to print the training losses, set the logging level to logging.INFO:

import logging
logging.basicConfig(level=logging.INFO)

Model

This model is largely an implementation of the Differentially Private Generative Adversarial Network model from Xie, Lin, Wang, Wang, and Zhou (2018).

Calculating epsilon

You can use the compute_dp_sgd_privacy script in the TensorFlow Privacy library to calculate the privacy loss epsilon for a given data set and training regime. N corresponds to the number of data points, and noise_multiplier corresponds to sigma. batch_size and epochs are the same as in training. delta (the tolerance probability for privacy loss beyond epsilon) can be set according to your needs or optimized (the library currently uses 1e-6 as a reasonable default). The method also gives an estimate of α for (α, ε)-Rényi differential privacy.

Note that calculating the appropriate value for the weight_clip is non-trivial, and depends on the architecture of the discriminator. See section 3.4 of Xie, Lin, Wang, Wang, and Zhou (2018) for details, where c_p corresponds to weight_clip. sigma should be set to the product of sigma_n and c_g.

Contributing

See CONTRIBUTING.md for information about contributing to this project.

License

BSD 3-Clause

See LICENSE.txt for details.

About

Differentially-private Wasserstein GAN implementation in PyTorch

License:BSD 3-Clause "New" or "Revised" License


Languages

Language:Python 100.0%