Core Concepts
Lex-parse sensitivity to edit operations and alphabet ordering is analyzed using Fibonacci words, revealing tight upper and lower bounds.
Abstract
The content delves into the sensitivity analysis of lex-parse with respect to edit operations and alphabet ordering using Fibonacci words. The study provides insights into compression sensitivity and repetitiveness measures, showcasing the robustness of lex-parse. The analysis reveals tight upper and lower bounds for both edit operations and alphabet ordering sensitivity. Various structures and properties of strings, including Lyndon factorization and Fibonacci words, are explored to characterize the lex-parse structure.
-
Introduction
- Dictionary compression effectiveness for repetitive text collections.
- Sensitivity of compressors and repetitiveness measures.
-
Preliminaries
- Definitions and properties of strings, prefixes, substrings, and suffixes.
- Lyndon factorization and Fibonacci words.
-
Sensitivity of Lex-parse
- Sensitivity to edit operations and alphabet ordering.
- Analysis of compression sensitivity variants.
-
Upper Bounds
- Tight upper bounds for edit operations and alphabet ordering sensitivity.
-
Lower Bounds for Edit Operations
- Tight lower bounds using Fibonacci words.
-
Lower Bounds for Alphabet-Ordering
- Tight lower bounds using Fibonacci words.
-
Acknowledgments
- Support from JSPS KAKENHI Grant Numbers.
Stats
v(w2) ∈ O(b(w2) log(n/b(w2)))
v(w, ≺) ∈ O(b(w) log(n/b(w)))
Quotes
"The lex-parse of a string is a greedy left-to-right partitioning."
"Each phrase can be encoded by a pair (0, T[i]) or (ℓ, i′)."