The author presents a system to procedurally generate examples for the 400 tasks in the Abstraction and Reasoning Corpus (ARC). The key points are:
The generation process aims to cover a broad and diverse space of examples for each task, going beyond the constraints of the original examples. This is achieved by randomly sampling various parameters like grid dimensions, number of symbols and objects, etc.
For each task, a corresponding verifier function is implemented to ensure the generated examples are valid and follow the task's transformation logic. The verifiers are written in the ARC Domain Specific Language (DSL).
The generation process allows controlling the difficulty of examples by sampling parameters from specified ranges. This enables experiments on within-task generalization, e.g. testing if a model can solve more difficult examples from the same task after training on easier ones.
The author discusses limitations of the approach, noting that the generated examples may not always match the "true" underlying transformation logic of the tasks as perceived by humans. However, the verifiers provide a level of credibility to the generated data.
The generated data provides opportunities for fundamental experiments on sample-efficient learning, such as comparing model architectures and training algorithms, or exploring curriculum learning approaches.
다른 언어로
소스 콘텐츠 기반
arxiv.org
더 깊은 질문