Pseudo-Boolean Competition 2025: some details

Benchmarks selection

Last year, the steering committee decided of a random selection of instances, selecting (when possible) the same number of instances per domain. A domain is a set of similar instances. Basically, one directory in an archive equals one domain, but some directories must be merged into a single domain when they contain similar instances. At most 5% of the whole instance set can be selected per submitter. This is meant to avoid anyone to have too much influence on the competition results.

The classification obtained last year was kept (thanks again to Johannes Fichte). New instances submitted to PB25 were classified by the organizer according to the name of the directories.

The goal was to obtain between 300 and 600 instances in the DEC-LIN and OPT-LIN tracks. Given the number of domains, 10 instances were randomly selected per domain in all tracks, except in the DEC-LIN track where 15 instances per domain were selected. If a domain doesn't contain enough instances, they are all selected. Duplicate instances are removed. This test is performed by checking the md5 checksum of instances. This means that the check is purely syntactical. It doesn't identify that an instance where contraints are reordered or variables renamed can be the same as another instance.

The 5% limit per submitter was not hit except in the OPT-NLC track where few instances were available last year. In order to respect the selection limit, only 3 of the fresh instances were selected. Those instances will be more useful next year because the 5% limit won't be applied anymore.

The archive PB25-selection.tgz contains the source files used to perform the selection. The program was run on a system with glibc 2.40-26 (Fedora 41). The random seed used is 2025. The archive contains:

Selecting instances is a difficult subject. The selection made is not perfect and some mistakes were probably made. However, it is hopefully clear that the selection was automated, mainly random and largely independent of the submitted solvers.

Hardware used

The competition was run on the CRIL cluster. Two kind of nodes were available:

Resources limits