Generalization Analysis for Supervised Contrastive Representation Learning under Non-IID Settings
Authors: Nong Minh Hieu, Antoine Ledent
ICML 2025 | Venue PDF | Archive PDF | Plain Text | LLM Run Details
| Reproducibility Variable | Result | LLM Response |
|---|---|---|
| Research Type | Experimental | In this section, we describe the numerical experiments on both synthetic and open sourced datasets to empirically verify our main results. Specifically, we aim to provide empirical evidence to corroborate three hypotheses. ... We summarize our experiment results in Figure 3 and Figure 2. |
| Researcher Affiliation | Academia | 1School of Computing and Information Systems, Singapore Management University. Correspondence to: Nong Minh Hieu <EMAIL>. |
| Pseudocode | No | The paper describes methods and problem formulations using mathematical notation and text, but does not include any clearly labeled pseudocode or algorithm blocks. |
| Open Source Code | No | The paper does not contain any explicit statements about the release of source code, nor does it provide links to a code repository or mention code in supplementary materials. |
| Open Datasets | Yes | In this section, we describe the numerical experiments on both synthetic and open sourced datasets to empirically verify our main results. Specifically, we aim to provide empirical evidence to corroborate three hypotheses. ... In this section, we describe two experiments conducted on the MNIST dataset. |
| Dataset Splits | No | The paper mentions conducting experiments on the MNIST dataset and synthetic data, and details the number of tuples used for training (n=100 or n=10000, and M sub-sampled tuples). However, it does not specify how these datasets are split into training, validation, or test sets for evaluating model performance. |
| Hardware Specification | No | The paper describes training neural networks for experiments but does not provide specific hardware details such as GPU or CPU models, or cloud computing specifications. |
| Software Dependencies | No | The paper does not list any specific software dependencies or their version numbers required to reproduce the experiments. |
| Experiment Setup | No | The paper mentions using a 'shallow neural network (with a fixed number of layers L 2)' and training with '5 random weights initializations'. It also describes synthetic data generation parameters like 'fixed standard deviation of σ 0.1'. However, it does not provide key hyperparameters for training, such as learning rate, batch size, optimizer, or number of epochs, which are crucial for reproducing the experimental setup. |