Notice: The reproducibility variables underlying each score are classified using an automated LLM-based pipeline, validated against a manually labeled dataset. LLM-based classification introduces uncertainty; scores should be interpreted as estimates. Full accuracy metrics and methodology are described in [1]

Approximability of Discriminators Implies Diversity in GANs

Authors: Yu Bai, Tengyu Ma, Andrej Risteski

ICLR 2019 | Venue PDF | LLM Run Details

Reproducibility Variable Result LLM Response
Research Type Experimental Our preliminary experiments show that on synthetic datasets the test IPM is well correlated with KL divergence or the Wasserstein distance, indicating that the lack of diversity in GANs may be caused by the sub-optimality in optimization instead of statistical inefficiency.
Researcher Affiliation Academia Yu Bai Stanford University EMAIL Tengyu Ma Stanford University EMAIL Andrej Risteski MIT EMAIL
Pseudocode Yes Algorithm 1 Discriminator family with restricted approximability for degenerate manifold
Open Source Code No The paper does not provide concrete access to source code for the methodology described.
Open Datasets Yes We design synthetic datasets, set up suitable generators, and train GANs with either our theoretically proposed discriminator class with restricted approximability, or vanilla neural network discriminators of reasonable capacity. [...] We set the ground truth distribution to be a unit circle or a Swiss roll curve, sampled from Circle : (x, y) Uniform( (x, y) : x2 + y2 = 1 ) Swiss roll : (x, y) = (z cos(4πz), z sin(4πz)) : z Uniform([0.25, 1]).
Dataset Splits No The paper does not provide specific dataset split information (e.g., percentages, sample counts, or detailed methodology) for training, validation, or testing.
Hardware Specification No The paper does not provide specific hardware details (e.g., exact GPU/CPU models, memory amounts, or detailed computer specifications) used for running its experiments.
Software Dependencies No The paper mentions 'RMSProp optimizer (Tieleman & Hinton, 2012)' and the 'POT package' but does not specify version numbers for these or other software dependencies, which is required for reproducibility.
Experiment Setup Yes The generator architecture is 2-50-50-2, and the discriminator architecture is 2-50-50-1. We use the RMSProp optimizer (Tieleman & Hinton, 2012) as our update rule, the learning rates are 10 4 for both the generator and discriminator, and we perform 10 steps on the discriminator in between each generator step.