Langages très simples générateurs
We give necessary and sufficient conditions for a language to be the language of finite words that occur infinitely many times in an infinite word.
We give necessary and sufficient conditions for a language to be the language of finite words that occur infinitely many times in an infinite word.
This paper consists of three parts. In the first part we prove a general theorem on the image of a language under a substitution, in the second we apply this to the special case when is the language of balanced words and in the third part we deal with recurrent Z-words of minimal block growth.
Tento článek je napsán u příležitosti udělení Abelovy ceny za rok 2021 László Lovászovi.
Since the huge database of patent documents is continuously increasing, the issue of classifying, updating and retrieving patent documents turned into an acute necessity. Therefore, we investigate the efficiency of applying Latent Semantic Indexing, an automatic indexing method of information retrieval, to some classes of patent documents from the United States Patent Classification System. We present some experiments that provide the optimal number of dimensions for the Latent Semantic Space and...
Text retrieval using Latent Semantic Indexing (LSI) with truncated Singular Value Decomposition (SVD) has been intensively studied in recent years. However, the expensive complexity involved in computing truncated SVD constitutes a major drawback of the LSI method. In this paper, we demonstrate how matrix rank approximation can influence the effectiveness of information retrieval systems. Besides, we present an implementation of the LSI method based on an eigenvalue analysis for rank approximation...
This paper reviews the existing literature on the combination of metaheuristics with machine learning methods and then introduces the concept of learnheuristics, a novel type of hybrid algorithms. Learnheuristics can be used to solve combinatorial optimization problems with dynamic inputs (COPDIs). In these COPDIs, the problem inputs (elements either located in the objective function or in the constraints set) are not fixed in advance as usual. On the contrary, they might vary in a predictable (non-random)...
The most common way of automatically learning Bayesian networks from data is the combination of a scoring metric, the evaluation of the fitness of any given candidate network to the data base, and a search procedure to explore the search space. Usually, the search is carried out by greedy hill-climbing algorithms, although other techniques such as genetic algorithms, have also been used.A recent metaheuristic, Ant Colony Optimisation (ACO), has been successfully applied to solve a great variety...