Core Concepts
Developing a dataset to enhance Large Language Models' proficiency in interpreting and generating Coq code, advancing automated theorem proving.
Abstract
The content introduces a dataset designed to improve Large Language Models (LLMs) in understanding and generating Coq code. It discusses the dataset's composition, creation methodology, and implications for machine learning in formal verification. Experiments show significant potential in enhancing accuracy and proof generation with models trained on this data. The dataset includes three tables: facts, propositions/proofs, and licensing information. Statistics on the dataset's tables are provided along with detailed experiments showcasing the model's capabilities.
Introduction:
Highlights limitations of LLMs in code optimization.
Proposes formal mathematical proofs as a solution.
Objectives:
Aims to refine ML models for formal theorem proving.
Prior Art:
Discusses existing datasets containing Coq source code.
Data Sources:
Details sources of Coq source files for the dataset.
Licenses:
Addresses licensing complexities within the dataset.
Dataset "coq-facts-props-proofs":
Describes the structure and preprocessing of the dataset.
Statistics:
Provides statistics on different tables within the dataset.
Experiments:
Showcases experiments fine-tuning models with the dataset.
Results:
Summarizes outcomes of fine-tuning LLMs with the dataset.
Outlook:
Explores future research directions using the dataset.
Quotes
"Addressing this gap, we present a comprehensive dataset specifically designed to enhance LLMs’ proficiency in interpreting and generating Coq code."
"Our primary aim is to facilitate the development of LLMs capable of generating syntactically correct and semantically meaningful Coq constructs."