As in ICCMA'17, the 2019 edition of the competition will feature seven main tracks, respectively on the complete (CO), preferred (PR), stable (ST), semi-stable (SST), stage (STG), grounded (GR), and Ideal (ID) semantics. Each of these tracks is composed of 4 (resp. 2 for single-status semantics, i.e., grounded and ideal) tasks, one for each reasoning task. Furthermore, four new dynamic tracks will be organized to test how solver behave on evolving frameworks.
Each solver participating in the competition can support, i.e. compete in, an arbitrary set of tasks. If a solver supports all the tasks of a track, it also automatically participates in the corresponding track.
Each solver will have 4GByte of RAM to compute the results of tasks in both the classical and dynamic track.
A task is a reasoning problem under a particular semantics. Following the same approach in ICCMA'15 and ICCMA'17, we consider four different problems:
for the seven semantics σ ∈ {CO, PR, ST, SST, STG, GR, ID}. For single-status semantics (GR and ID) only the problems SE and DC are considered (EE is equivalent to SE, and DS is equivalent to DC). Note that DC-CO and DC-PR are equivalent as well, but in order to allow the participation in the preferred track without implementing tasks on the complete semantics (or viceversa), we repeat the task.
The combination of problems with semantics amounts to a total number of 24 tasks. Each solver participating in a task will be queried with a fixed number of instances corresponding to the task with a timeout of 10 minutes each. For each instance, a solver gets
The terms correct and incorrect for the different reasoning problems are defined as follows.
Intuitively, a result is neither correct nor incorrect (and therefore gets 0 points) if (i) it is empty (e.g., the timeout was reached without answer) or (ii) it is not parsable (e.g., some unexpected error message).
The score of a solver for a particular task is the sum of points over all instances. The ranking of solvers for the task is then based on the scores in descending order. Ties are broken by the total time it took the solver to return correct results.
For semi-stable and stage semantics, we recall that those semantics coincide with stable for AFs that possess at least one stable extension.
All tasks for a particular semantics constitute a track. The ranking of solvers for a track is based on the sum of scores over all tasks of the track. Again, ties are broken by the total time it took the solver to return correct results. Note that in order to make sure that each task has the same impact on the evaluation of the track, all tasks for one semantics will have the same number of instances.
The winner of each track will be awarded.
The 2019 edition of ICCMA also features additional tracks to evaluate solvers on dynamic Dung's frameworks. The aim is to test those solvers dedicated to efficiently recompute a solution after a small change in the original abstract framework.
In this case, an instance consists of an initial framework (as for classical tracks) and an additional file storing a sequence of additions/deletions of attacks on the initial framework (at least 15 changes). This file will be provided through a simple text format, e.g., a sequence of +att(a,b) (attack addition) or -att(d,e) (attack deletion). The final single output needs to report the solution for the initial framework and as many outputs as the number of changes.
The four new dynamic tracks concern the following semantics and problems, for a total of 14 different tasks:Each solver participating in the competition can support, i.e. compete in, an arbitrary set of tasks. If a solver supports all the tasks of a track, it also automatically participates in the corresponding track.
A result is correct and complete if, for n changes, n+1 correct and complete results are given. The score for a correct and complete result is 1 as usual.
A partial (incomplete) result is considered correct if it gives less than n+1 answers, but each of the given answer is correct and complete (w.r.t. the corresponding static tasks). This rules holds for all the problems (SE, DC, DS, EE) in the dynamic track. A correct but incomplete result will score a value in (0, 1], depending on the rate of correct sub-solutions given. Exception: in case the considered dynamic task involves enumeration (i.e., EE), if the last solution a solver provides is correct but partial, then the whole answer will be evaluated as the last problem was not solved at all, considering the answer as partial and correct. A fraction of 1/n points, depending on the percentage of returned enumerated extensions will be assigned.
If any of the sub-solution is incorrect, then the overall output is considered incorrect (-5 points).
Otherwise, in case no answer is given, 0 points will be assigned (usually due to a timeout).
The timeout to compute an answer for the dynamic track is 5 minutes for each framework/change (half of the time in the classical track for a single instance). For example, considering n changes, a timeout of 5*(n+1) minutes is given: for 8 changes, a global timeout of 45 minutes.
In the final ranking, ties will be broken by the total time it took the solver to return correct results for all the considered frameworks (starting plus changes).
More information about the format of input files and the expected output is available at SolverRequirements.pdf.