A substring of a string is unbordered if its only border is the empty string. The study of unbordered substrings goes back to the paper of Ehrenfeucht and Silberger [Discr. Math 26 (1979)]. The main focus of their and subsequent papers was to elucidate the relationship between the longest unbordered substring and the minimal period of strings. In this paper, we consider the algorithmic problem of computing the longest unbordered substring of a string. The problem was introduced recently by G. Kucherov et al. [CPM (2015)], where the authors showed that the average-case running time of the simple, border-array based algorithm can be bounded by O(max{n, n2/σ4}) for σ being the size of the alphabet. (The worst-case running time remained O(n2).) Here we propose two algorithms, both presenting substantial theoretical improvements to the result of [11]. The first algorithm has O(n log n) average-case running time and O(n2) worst-case running time, and the second algorithm has O(n1.5) worst-case running time.
CITATION STYLE
Gawrychowski, P., Kucherov, G., Sach, B., & Starikovskaya, T. (2015). Computing the longest unbordered substring. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 9309, pp. 246–257). Springer Verlag. https://doi.org/10.1007/978-3-319-23826-5_24
Mendeley helps you to discover research relevant for your work.