Upcoming competitions


Past competitions

In addition to formal grammatical inference, which focuses on formal proofs of learnability of classes of languages under certain conditions, there is also research in the area of empirical grammatical inference. In this field, the aim is to learn a particular grammar (not necessarily a class of grammars). Additional difficulties may be introduced, such as noise.

To test the practical possibilities (and limits) of empirical grammatical inference systems, several data sets, which often come from competitions are available as benchmarks.

  • TAYSIR (2023): The Transformers+RNN: Algorithms to Yield Simple and Interpretable Representations (TAYSIR) competition is an on-line challenge on extracting simpler models from already trained neural networks. These neural nets are trained on tasks involving sequences of symbols. Some of these tasks are artificial and some come from real world problems in domains like natural language processing (NLP), bio-informatics, software engineering and others. TAYSIR means “simplification” in Arabic.
  • The SPiCe (2016) was an on-line competition about guessing the next element in a sequence of symbols.Training datasets consist of whole sequences and the aim was to learn a model that allows the ranking of potential next symbols for a given prefix, that is, the most likely options for a single next symbol.
  • The PAutomaC challenge (2012) was about learning probability distributions from strings. The (artificial) data were generated using either Hidden Markov Chains (HMM) or (Deterministic or not) probability automaton (PA). (local mirror)
  • The Tenjinno competition (2006) looked at learning transducers from synthetic data sets motivated by Machine Translation. report
  • The Omphalos competition (2004) was a competition on context free grammars inference.

DFA Learning

The problem of learning a target DFA from labeled examples has been extensively studied in the literature for over 3 decades. A variety of symbolic, connectionist, and hybrid techniques have been proposed to address this difficult problem. A few benchmark datasets were available against which most of the new algorithms were tested.

  • The Abbadingo One Learning Competition (1996). Barak Pearlmutter and Kevin Lang posted a set of challenging DFA learning problems designed to allow researchers to test their favorite learning algorithms: this was the Abbadingo One Learning Competition. Although the problems were still artificially generated (i.e., the target DFAs were randomly generated), several reasearchers participated in the competition. The eventual winners came up with algorithms that were significant improvements over the existing methods for learning DFA.
  • The Gowachin DFA Learning Benchmark (1998). Following the success of Abbadingo One, Kevin Lang, Babak Pearlmutter, and François Coste have teamed up to launch the Gowachin Learning Competition. Users are allowed to generate their own problem (by specifying the size of the target DFA, the number of training examples, and the noise level).
  • The GECCO (2004?): Learning DFA from Noisy Samples Competition
  • The STAMINA Competition (2010): The competition was about learning regular languages with large alphabets.
  • The ZULU Competition (2010): Zulu is an active learning competition, where participants are to build algorithms that can learn deterministic finite automata (DFA) by making the smallest number of membership queries to the server/oracle.

Comments are closed.