Core Concepts
Proposing LILAC, a log parsing framework using LLMs with adaptive parsing cache, to enhance accuracy and efficiency in log parsing.
Abstract
The content introduces the LILAC framework for log parsing using Large Language Models (LLMs) with an adaptive parsing cache. It addresses challenges in log parsing by leveraging the in-context learning capability of LLMs and incorporating a novel adaptive parsing cache. The framework aims to improve accuracy, efficiency, and consistency in log message extraction.
- Log Parsing Challenges:
- Existing approaches compromised on complicated log data.
- Syntax-based parsers rely heavily on rules; semantic-based parsers lack training data.
- Introduction of LILAC:
- Utilizes Large Language Models (LLMs) for accurate log parsing.
- Features an ICL-enhanced parser and adaptive parsing cache.
- ICL-enhanced Parser:
- Hierarchical candidate sampling algorithm for diverse log messages.
- Demonstration selection based on similarity to queried logs.
- Adaptive Parsing Cache:
- Tree structure for efficient storage and retrieval of templates.
- Cache matching and updating operations ensure consistency and efficiency.
Stats
The recent emergence of powerful large language models (LLMs) demonstrates their vast pre-trained knowledge related to code and logging.
Extensive evaluation on public large-scale datasets shows that LILAC outperforms state-of-the-art methods by 69.5% in terms of the average F1 score of template accuracy.