The author proposes a method to generate synthetic data that maintains correlations from the original dataset while ensuring privacy. The approach aims to balance utility and disclosure levels effectively.
The author introduces a novel approach to generating synthetic data for system identification by leveraging knowledge transfer from similar systems. This method aims to enhance model generalization and robustness in scenarios with data scarcity.
Quality-Diversity Generative Sampling (QDGS) improves fairness and accuracy in training classifiers with balanced synthetic datasets.
Generative AI and Large Language Models are revolutionizing synthetic data generation, addressing data scarcity and privacy concerns while pushing the boundaries of AI development.
DataTune, a method to transform existing datasets into a format aligned with the requirements of target tasks, can significantly improve the quality and diversity of synthetically generated data compared to direct language model generation.