Evaluating Large Language Models' Ability to Generate Complex Structured Tabular Data
Large Language Models (LLMs) struggle to generate complex structured tabular data, despite their advanced text generation capabilities. This study introduces a comprehensive benchmark, STRUC-BENCH, to assess LLMs' performance on this task and proposes a structure-aware fine-tuning method to improve their abilities.