This repository has been archived on 2025-03-06. You can view files and clone it, but cannot push or open issues or pull requests.
dekker-phd-thesis/chapters/4_half_reif.tex

1023 lines
60 KiB
TeX

%************************************************
\chapter{Half Reification}\label{ch:half-reif}
%************************************************
\glsreset{half-reif}
\input{chapters/4_half_reif_preamble}
\paragraph{Reification - Text from rewriting}
\jip{Adjust with information in the background section. Is there anything specific to the \microzinc{}/\nanozinc{} system?}
Many constraint modelling languages, including \minizinc{}, not only enable \constraints{} to be globally enforced, but typically also support \constraints{} to be \emph{reified} into a Boolean variable.
Reification means that a variable \mzninline{b} is constrained to be true if and only if a corresponding constraint \mzninline{c(...)
} holds.
We have already seen reification in \cref{ex:4-absreif}: the truth of constraint \mzninline{abs(x) > y} was bound to a Boolean variable \mzninline{b1}, which was then used in a disjunction.
We say that the same constraint can be used in \emph{root context} as well as in a \emph{reified context}.
In \minizinc{}, almost all constraints can be used in both contexts.
However, reified constraints are often defined in the library in terms of complicated decompositions into simpler constraints, or require specialised algorithms in the target solvers.
In either case, it can be very beneficial for the efficiency of the generated \nanozinc\ program if we can detect that a reified constraint is in fact not required.
If a constraint is present in the root context, it means that it must hold globally.
If the same constraint is used in a reified context, it can therefore be replaced with the constant \mzninline{true}, avoiding the need for reification (or in fact any evaluation).
We can harness \gls{cse} to store the evaluation context when a constraint is added, and detect when the same constraint is used in both contexts.
Whenever a lookup in the \gls{cse} table is successful, action can be taken depending on both the current and stored evaluation context.
If the stored expression was in root context, then the constant \mzninline{true} can be used, independent of the current context.
Otherwise, if the stored expression was in reified context and the current context is reified, then the stored result variable can be used.
Finally, if the stored expression was in reified context and the current context is root context, then the previous result can be replaced by the constant \mzninline{true} (and all its dependencies removed) and the evaluation will proceed as normal with the root context constraint.
\begin{example}
Consider the fragment:
\begin{mzn}
function var bool: p(var int: x, var int: y) = q(x) /\ r(y);
constraint b0 <-> q(x);
constraint b1 <-> t(x,y);
constraint b1 <-> p(x,y);
\end{mzn}
If we process the top-level constraints in order we create a reified call to \mzninline{q(x)} for the original call.
Suppose processing the second constraint we discover \mzninline{t(x,y)} is \mzninline{true}, fixing \mzninline{b1}.
When we then process \mzninline{q(x)} in instantiation of the call \mzninline{p(x,y)}, we find it is the root context.
So \gls{cse} needs to set \mzninline{b0} to \mzninline{true}.
\end{example}
\minizinc{} distinguishes not only between root and reified contexts, but in addition recognises constraints in \emph{positive} contexts, also known as \emph{half-reified} \autocite{feydy-2011-half-reif}.
A full explanation and discussion of half-reification can be found in \cref{ch:half-reif}.
\jip{TODO:\ Reification should also be discussed in the background. How can we
make this more specific to how it works in \nanozinc{}.}
\section{An Introduction to Half Reification}
\label{sec:half-intro}
The complex expressions language used in \cmls{}, such as \minizinc{}, often require the use of \gls{reification} in the flattening process to reach a solver level constraint model.
If the Boolean expression \mzninline{pred(...)} is seen in a non-root context, then a new Boolean \variable{} \mzninline{b} is introduced to replace the expression, its \emph{control} \variable{}.
The flattener then enforces a \constraint{} \mzninline{pred_reif(...,b)}, which binds the \variable{} to be the \emph{truth-value} of the expression (\ie\ \mzninline{b <-> pred(...)}).
Feydy et al.\ show that although the usage of \gls{reification} in the flattening process is well-understood, it suffers from certain weaknesses:
\begin{enumerate}
\item Many \glspl{reification} are created in the rewriting of partial expressions to accommodate \minizinc{}'s relational semantics.
\item Propagators for the \glspl{reification} of global constraints are scarce.
\item A \gls{reification} sometimes provides too much information to its surrounding context, triggering propagators that will never be able to prune any values from a \gls{domain}.
\end{enumerate}
In constast, the authors introduce \gls{half-reif}.
\gls{half-reif} follows from the idea that in many cases it might be sufficient to use the logical implication of an expression, \mzninline{b -> pred(...)}, instead of the logical equivalence, \mzninline{b <-> pred(...)}.
Flattening with \gls{half-reif} is an approach that improves upon all these weaknesses of flattening with \emph{full} reification:
\begin{enumerate}
\item Flattening using \glspl{half-reif} naturally produces relational semantics.
\item Propagators for a \glspl{half-reif} can often be constructed by merely altering the implementation the regular \constraint{}.
\item The control \variables{} can limit the amount of triggered propagators that are known to be unable to prune any variables.
\end{enumerate}
Additionally, for many \solvers{} the decomposition of a \gls{reification} is more complex than a \gls{half-reif}.
We will show that the usage of \glspl{half-reif} can therefore lead to a reduction in \constraints{} in the solver level constraint model.
\Gls{half-reif} can be used instead of full \gls{reification} when the \gls{reification} can never be forced to be false.
We see this in, for example, a disjunction \(a \lor b\).
No matter the value of \(a\), setting the value of \(b\) to be true can never make the overall expression false.
\(b\) is thus never forced to be false.
This requirement follows from the difference between implication and logical equivalences.
Setting the left hand side of a implication to false, does not influence the value on the right hand side.
So if we know that this is never required in the overall expression, then using an implication instead of a logical equivalence (\ie{} a \gls{half-reif} instead of a full \gls{reification}) does not change the meaning of the constraint.
This property can be extended to include non-Boolean expressions.
Since Boolean expressions in \minizinc{} can be used in, for example, integer expressions, we can apply similar reasoning to these types of expressions.
\begin{example}
For example the left hand side of the constraint
\begin{mzn}
constraint count(x in arr)(x = 5) > 5;
\end{mzn}
\noindent{}is an integer expression that contains the Boolean expression \mzninline{x = 5}.
Since the increasing left hand side of the constraint will only ever help satisfy the constraint, the expression \mzninline{x = 5} will never forced to be false.
This means that we can half-reify the expression.
\end{example}
To systematically analyse whether Booelean expressions can be half-reified, we introduce extra distinctions in the context of expressions.
Before, we would merely distinguish between \rootc{} context and \emph{non-root} context.
Now, we will categorise the latter into:
\begin{description}
\item[\posc{} context] when an expression must reach \emph{at least} a certain value to satisfy its enclosing constraint.
The expression is never forced to take a lower value.
\item[\negc{} context] when an expression can reach \emph{at most} a certain value to satisfy its enclosing constraint.
The expression is never forced to take a higher value.
\item[\mixc{} context] when an expression must take an \emph{exact value}, be within a \emph{specified range} or when during flattening it cannot be determined whether the expression must be increased or decreased to satisfy the enclosing constraint.
\end{description}
As previously explained, \gls{half-reif} can be used for expressions in \posc{} context.
Although expressions in a \negc{} context cannot be directly half-reified, the negation of a expression in a \negc{} context can be half-reified.
Consider, for example, the constraint
\begin{mzn}
constraint b \/ not (x = 5);
\end{mzn}
The expression \mzninline{x = 5} is in a \negc{} context.
Although a \gls{half-reif} cannot be used directly, in some cases the solver can negate the expression which are then placed in a \posc{} context.
Our example can be transformed into:
\begin{mzn}
constraint b \/ x != 5;
\end{mzn}
The transformed expression, \mzninline{x != 5}, is now in a \posc{} context.
We can also speak of this process as ``pushing the negation inwards''.
Expressions in a \mixc{} context are in a position where \gls{half-reif} is impossible.
Only full \gls{reification} can be used for expressions in that are in this context.
This occurs, for example, when using an exclusive or expression in a constraint.
The value that one side must take directly depends on the value that the other side takes.
Each side can thus be forced to be true or false.
The \mixc{} context can also be used as a ``fall back'' context.
If it cannot be determined if an expression is in a \posc{} or \negc{} context, then it is always safe to say the expression is in a \mixc{} context.
\section{Propagation and Half Reification}%
\label{sec:half-propagation}
The tasks of a propagator for any constraint can logically be split into two tasks:
\begin{enumerate}
\item To \(check\) if the constraint can still be satisfied (and otherwise \emph{fail} the current state)
\item To \(prune\) values from the \glspl{domain} of \variables{} that would violate the constraint.
\end{enumerate}
When creating a propagator for the \gls{half-reif} of a constraint, it can be constructed from these two tasks.
The half-reified propagator is dependent on an additional argument \(b\), which is a Boolean \variable{}.
The Boolean \variable{} can be in three states, it can currently not have been assigned a value, it can be assigned \mzninline{true}, or it can be assigned \mzninline{false}.
Given \(b\), \(check\), and \(prune\), \cref{alg:half-prop} shows pseudo code for the propagation of the \gls{half-reif} of the constraint.
\begin{algorithm}
\KwIn{A function \(check\), that returns false when the constraint \(c\) cannot be satisfied, a function \(prune\), that eliminates values from \variable{} \glspl{domain} that violate the constraint \(c\), and a Boolean control \variable{} \(b\).
}
\KwResult{This pseudo code propagates the \gls{half-reif} of \(c\) (\ie{} \(b \implies\ c\)).}
\BlankLine{}
\If{\(b\) {\normalfont is unassigned} }{
\If{\(\neg{}check()\)}{
\(b \longleftarrow \) \mzninline{false}\;
}
}
\If{\(b = \) \mzninline{true}}{
\(prune()\)\;
}
\caption{\label{alg:half-prop} Propagation pseudo code for the \gls{half-reif} of a constraint \(c\), based on the propagator for \(c\).}
\end{algorithm}
Logically, the creation of propagators for \glspl{half-reif} can always follow this simple principle.
In practice, however, this is not always possible.
In some cases, propagators do not explicitly define \(check\) as a separate step.
Instead, this process can be implicit.
The propagator merely prunes the \glspl{domain} of the \variables{}.
When a \gls{domain} is found to be empty, then the propagator fails the current state.
It is not possible possible to construct the half-reified propagator from such an implicit \(check\) operation.
Instead a new explicit \(check\) method has to be devised to implement the propagator of the \gls{half-reif}.
A propagation engine gains certain advantages from \gls{half-reif}, but also may suffer certain penalties.
Half reification can cause propagators to wake up less frequently, since variables that are fixed to true by full reification will never be fixed by half reification.
This is advantageous, but a corresponding disadvantage is that variables that are fixed can allow the simplification of the propagator, and hence make its propagation faster.
When a full \gls{reification} is required and a full reification \mzninline{x <-> pred(...)} by using two half reified propagators, \mzninline{x -> pred(...)} and \mzninline{y \half \neg pred(...)}, and the constraint \mzninline{x <-> not y}.
This does not lose propagation strength.
For Booleans appearing in a positive context we can make the propagator \mzninline{y -> not pred(...)} run at the lowest priority, since it will never cause failure.
Similarly in negative contexts we can make the propagator \mzninline{b -> pred(...)} run at the lowest priority.
This means that Boolean variables are still fixed at the same time, but there is less overhead.
In \cref{sec:half-experiments} we assess the implementation of propagators for the \glspl{half-reif} of \mzninline{all_different} and \mzninline{element}.
Both propagators are designed and implemented in \gls{chuffed} according to the principle explained above.
\section{Decomposition and Half Reification}%
\label{sec:half-decomposition}
The use of \gls{half-reif} does not only offer a benefit when a propagator for the half-reified constraint is available.
It can also be beneficial in the decomposition of constraints.
Compared to full \gls{reification}, the decomposition of a \gls{half-reif} does not need to keep track of as much information.
In particular, this can be beneficial when the target \solver{} is a \gls{mip} or \gls{sat} solver.
The decompositions for these solver technologies often explicitly encode reified constraints using two implications.
If, however, a \gls{reification} is replaced by a \gls{half-reif}, then only one of these implication is required.
\begin{example}
Consider the \gls{reification} of the \constraint{} \mzninline{i <= 4} using the control variable \mzninline{b}, where \mzninline{i} can take values in the domain \mzninline{0..10}.
If the target solver is a \gls{mip} solver, then this \gls{reification} would be linearised.
It would take the following form:
\begin{mzn}
constraint i <= 10 - 6 * b; % b -> i <= 4
constraint i >= 5 - 5 * b; % not b -> i >= 5
\end{mzn}
Instead, if we could determine that the \constraint{} could be half-reified, then the linearisation could be simplified to only the first constraint.
\end{example}
The same principle can be applied all throughout the linearisation process.
Ultimately, \minizinc{}'s linearisation library rewrites most \glspl{reification} in terms of implied less than or equal to constraint.
For all these \glspl{reification}, its replacement by a \gls{half-reif} can remove half of the implications required for the \gls{reification}.
For \gls{sat} solvers, a decomposition for a \gls{half-reif} can be created from its regular decomposition.
Any constraint \(c\) will decompose into \gls{cnf}:
\[ c = \forall_{i} \exists_{j} lit_{ij} \]
The \gls{half-reif}, with control variable \(b\), could then be encoded as:
\[ b \implies c = \forall_{i} \neg b \lor \exists_{j} lit_{ij} \]
The transition from the \gls{cnf} of the regular constraint to its \gls{half-reif} only adds a single literal to each clause.
It is, however, not as straightforward to construct its full \gls{reification}.
In addition to the half-reified \gls{cnf}, a generic \gls{reification} would require the reverse implication, \(\neg b \implies \neg c\).
Based on the \gls{cnf} of \(c\), this would result in the following logical formula:
\[ \neg b \implies \neg c = \forall_{i} b \lor \neg \exists_{j} lit_{ij} \]
This formula, however, is no longer a direct set of clauses.
Rewriting this formula into \gls{cnf} would result in:
\[ \neg b \implies \neg c = \forall_{i,j} b \lor lit_{ij} \]
This adds a new binary clause for every literal in the original \gls{cnf}.
In general, many more clauses are needed to decompose a \gls{reification} compared to a \gls{half-reif}.
According to the principles above, decomposition libraries for the full \minizinc{} language have been implemented for \gls{mip} and \gls{sat} solvers.
In \cref{sec:half-experiments} we asses the effects when flattening with \gls{half-reif}.
\section{Context Analysis}%
\label{sec:half-context}
When taking into account the possible undefinedness of an expression, every expression in a \minizinc{} model has two different contexts: the context in which the expression itself occurs, its \emph{value context}, and the context in which the partiality of the expression is captured, its \emph{partiality context}.
As described in \cref{subsec:back-mzn-partial}, \minizinc{} uses relational semantics of partial values.
This means that if a function does not have a result, then its nearest enclosing Boolean expression is set to false.
In practice, this means that a condition that tests if the function will return a value is added to the nearest enclosing Boolean expression.
The \emph{partiality} context is the context in which this condition is placed.
The context of an expression cannot always be determined by merely considering \minizinc\ expressions top-down.
Expressions bound to a variable can be used multiple times in expressions that influence their context.
\begin{example}
Consider the following \minizinc\ fragment.
\begin{mzn}
constraint let {
var bool: x = pred(a, b, c);
} in y -> x /\ x -> z;
\end{mzn}
The predicate call \mzninline{pred(a, b, c)} is bound to the variable \mzninline{x}.
The call is in \posc\ context, and can be half-reified, if the variable \mzninline{x} is only used in \posc{} context.
Although this is the case in the left side of the conjunction, the other side uses \mzninline{x} in a \negc{} context.
This means that \mzninline{pred(a, b, c)} is in a \mixc{} context, and must fully reified.
\end{example}
Note that an alternative approach for this example would be to replace the identifier with its definition.
It would then be possible half-reify the call and the negation of the call.
Although this would increase the use of \gls{half-reif}, it should be noted that the propagation of these two \glspl{half-reif} would be equivalent to propagating the full \gls{reification} of the call.
In this scenario, we prefer to create the full \gls{reification} as it decreases the number of \variables{} and \constraint{} in the model.
\subsection{Automated analysis}%
\label{subsec:half-automated}
In the architecture introduced in \cref{ch:rewriting}, contexts of the expressions can be determined automatically.
The analysis is best performed during the compilation process from \minizinc{} to \microzinc{}, instead of during the \microzinc{} interpretation, since it requires knowledge about all usages of certain \variable{} at the same time.
Without loss of generality we can define the context analysis process for \microzinc{} models.
This has the advantage that the value and partiality have already been explicitly separated and no longer requires special handling.
We describe the context analysis performed on the \microzinc{} syntax in the form of inference rules.
The full set of rules appears in \cref{fig:half-analysis-expr,fig:half-analysis-it}.
Each rules describe how an expression is found in a context \(ctx\), above the line, changes the context of subordinate expressions, below the line.
The syntax \ctxeval{e}{ctx} is used to assert that the expression \(e\) is evaluated in the context \(ctx\).
We now specify two context transformations that will be used in further algorithms to transition between different contexts: \changepos{} and \changeneg{}.
The behaviour of these transformations is shown in \cref{fig:half-ctx-trans}.
\begin{figure*}
\begin{center}
\begin{tabular}{ccc}
\(
\begin{array}{lcl}
\changepos \rootc & = & \posc \\
\changepos \posc & = & \posc \\
\changepos \negc & = & \negc \\
\changepos \mixc & = & \mixc
\end{array}
\)
& ~ &
\(
\begin{array}{lcl}
\changeneg \rootc & = & \negc \\
\changeneg \posc & = & \negc \\
\changeneg \negc & = & \posc \\
\changeneg \mixc & = & \mixc
\end{array}
\)
\end{tabular}
\end{center}
\caption{\label{fig:half-ctx-trans} Definitions of the \changepos{} and \changeneg{} context transitions.}
\end{figure*}
\begin{figure*}
\centering
\begin{prooftree}
\hypo{\ctxeval{x}{ctx}}
\infer1[(Ident)]{\text{pushCtx}(x, ctx)}
\end{prooftree} \\
\bigskip
\begin{prooftree}
\hypo{\ctxeval{ident\texttt{(} e_{1}, \ldots, e_{n} \texttt{)}}{ctx}}
\hypo{\text{argCtx}(ident, ctx) = \tuple{ ctx'_{1}, \ldots, ctx'_{n}}}
\infer2[(Call)]{\ctxfunc{ident}{ctx},~\ctxeval{e_{1}}{ctx'_{1}},~\ldots,~ \ctxeval{e_{n}}{ctx'_{n}}}
\end{prooftree} \\
\bigskip
\begin{prooftree}
\hypo{\ctxeval{x\texttt{[}i\texttt{]}}{ctx}}
\infer1[(Access)]{\ctxeval{x}{\changepos{}ctx}}
\end{prooftree} \\
\bigskip
\begin{prooftree}
\hypo{\ctxeval{\texttt{if }b\texttt{ then } e_{1} \texttt{ else } e_{2} \texttt{ endif}}{ctx}}
\infer1[(ITE)]{\ctxeval{e_{1}}{\changepos{}ctx},~\ctxeval{e_{2}}{\changepos{}ctx}}
\end{prooftree} \\
\bigskip
\begin{prooftree}
\hypo{\ctxeval{\texttt{[}e~\texttt{|}~G\texttt{]}}{ctx}}
\infer1[(Compr)]{\ctxeval{e}{ctx}}
\end{prooftree} \\
\bigskip
\begin{prooftree}
\hypo{\ctxeval{\texttt{[}e_{1}, \ldots, e_{n}\texttt{]}}{ctx}}
\infer1[(Arr)]{\ctxeval{e_{1}}{ctx}, \ldots, \ctxeval{e_{n}}{ctx}}
\end{prooftree}
\caption{\label{fig:half-analysis-expr} Context inference rules for \microzinc\ expressions.}
\end{figure*}
\Cref{fig:half-analysis-expr} shows the inference rules for all \microzinc{} expressions, apart from let expressions.
The first rule, (Ident), is unique in the sense that the context of an identifier does not directly affects any sub-expressions.
Instead every context in which the identifier are found are collected and will be processed when evaluating the corresponding declaration.
Note that the presented inference rules do not have any explicit state object.
Instead, we introduce the functions ``pushCtx'' and ``collectCtx''.
These functions track and combine the contexts in which an value is used in an implicit global state.
Most changes in the context of \microzinc{} expressions stem from the consequent (Call) rule.
A call expression can change the context in which its arguments should be evaluated.
As an input to the inference process, a ``argCtx'' function is used to give the context of the arguments of a function, given the function itself and the context of the call.
A definition for this function for the \minizinc\ operators can be found in \cref{alg:arg-ctx}.\footnote{We use \minizinc\ operator syntax instead of \microzinc{} identifiers for brevity and clarity.}
Although a standard definition for the ``argMax'' function may cover the most common cases, it does not cover user-defined functions stemming from \minizinc{}.
To address this issue, we introduce the \minizinc{} annotations \mzninline{promise_ctx_monotone} and \mzninline{promise_ctx_antitone} to represent the operations \changepos{} and \changeneg{} respectively.
When a function argument is annotated with one of these annotations, the context of the argument in a call in context \(ctx\) is transformed with the operation coressponding to the annotation (\eg\ \(\changepos{}ctx\) or \(\changeneg{}ctx\)).
If a function argument is not annotated, then the argument is evaluated in \mixc context.
\begin{example}
Consider the following user-defined \minizinc\ implementation of a logical implication.
\begin{mzn}
predicate impli(
var bool: x ::promise_ctx_antitone,
var bool: y ::promise_ctx_monotone
) =
not x \/ y;
\end{mzn}
The annotations placed on the argument of the \mzninline{impli} function will apply the same context transformations as a direct \minizinc\ implication, as shown in \cref{alg:arg-ctx}.
In term of context analysis, this function now is equivalent to the \minizinc implication operator.
\end{example}
\begin{algorithm}
\KwIn{A \minizinc\ operator \(op\) and calling context \(ctx\)}
\KwOut{A tuple containing the contexts of the arguments \(\tuple{ctx_{1}, \ldots{}, ctx_{n}}\)}
\Switch{op}{
\Case{\mzninline{ not }}{
\Return{\tuple{\changeneg{}ctx}}
}
\Case{\mzninline{ \/ }, \mzninline{+ }}{
\Return{\tuple{\changepos{}ctx, \changepos{}ctx}}
}
\Case{\mzninline{ -> }, \mzninline{< }, \mzninline{<= }}{
\Return{\tuple{\changeneg{}ctx, \changepos{}ctx}}
}
\Case{\mzninline{ > }, \mzninline{>= }, \mzninline{- }}{
\Return{\tuple{\changepos{}ctx, \changeneg{}ctx}}
}
\Case{\mzninline{ <-> }, \mzninline{xor }, \mzninline{* }}{
\Return{\tuple{\mixc, \mixc}}
}
\Case{\mzninline{ /\ }}{
\Return{\tuple{ctx, ctx}}
}
\Other{
\Return{\tuple{\mixc, \ldots, \mixc}}
}
}
\caption{\label{alg:arg-ctx} Definition of the \emph{argCtx} function for \minizinc\ operators.}
\end{algorithm}
The context in which the result of a call expression is used must also be considered.
The (Call) rule, therefore, introduces the \ctxfunc{ident}{ctx} syntax.
This syntax is used to declare that the compiler must introduce a \microzinc\ function variant that flattens the function call to \(ident\) in the context \(ctx\).
This means that if \(ctx\) is \rootc{}, the compiler can use the the function as defined.
Otherwise, the compiler follows the following steps to try to introduce the most compatible variant of the function:
\begin{enumerate}
\item If a direct definition for \(ctx\) definition exists, then use this definition.
\begin{description}
\item[\posc] \glspl{half-reif} can be defined as \(ident\)\mzninline{_imp}.
\item[\negc] negations of \glspl{half-reif} can be defined as \(ident\)\mzninline{_imp_neg}.
\item[\mixc] \glspl{reification} can be defined as \(ident\)\mzninline{_reif}.
\end{description}
\item If \(ident\) is a \microzinc{} function with an expression body \(E\), then copy of the function can be made that is evaluated in the desired context:\ctxeval{E}{ctx}.
\item If \(ctx\) is \posc{} or \negc{}, then change \(ctx\) to \mixc{} and return to step 1.
\item Finally, if non of the earlier steps were successful, then the compilation fails.
Note that this can only occurr when a solver-level predicate does not have an \gls{reification}.
\end{enumerate}
The (Access) and (ITE) rules show the context inference for array access and if-then-else expressions respectively.
Their inner expressions are evaluated in \(\changepos{}ctx\).
The inner expressions cannot be simply be evaluated in \(ctx\), because it is not yet certain which expression will be chosen.
This is important for when \(ctx\) is \rootc{}, since we, at compile time, say which expression will hold globally.
We will revisit this issue in \cref{subsec:half-?root}.
Finally, the (Compr) and (Arr) rules show simple inference rules for array construction expressions.
If such an expression is evaluated in the context \(ctx\), then its members can be evaluated in the same context \(ctx\).
\begin{figure*}
\centering
\begin{prooftree}
\hypo{\ctxeval{\texttt{let \{ }I\texttt{ \} in } e}{ctx}}
\infer1[(Let)]{\ctxeval{e}{ctx}, \ctxitem{I} }
\end{prooftree} \\
\bigskip
\begin{prooftree}
\hypo{\ctxitem{I \texttt{; constraint } e }}
\infer1[(Con)]{\ctxeval{e}{\rootc},~\ctxitem{I}}
\end{prooftree} \\
\bigskip
\begin{prooftree}
\hypo{\ctxitem{I \texttt{; } T: x \texttt{ = } e }}
\hypo{\text{collectCtx}(x) = [ctx_{1}, \ldots, ctx_{n}]}
\infer2[(Assign)]{\ctxeval{x}{\text{join}([ctx_{1}, \ldots, ctx_{n}])},~\ctxitem{I}}
\end{prooftree} \\
\bigskip
\begin{prooftree}
\hypo{\ctxitem{I \texttt{; } \texttt{tuple(}\ldots\texttt{):}~x \texttt{ = (} e_{1}, \ldots, e_{n}\texttt{)}}}
\hypo{\text{collectCtx}(x) = \tuple{ctx_{1}, \ldots, ctx_{n}}}
\infer2[(TupC)]{\ctxeval{e_{1}}{ctx_{1}}, \ldots, \ctxeval{e_{n}}{ctx_{n}}, ~\ctxitem{I}}
\end{prooftree} \\
\bigskip
\begin{prooftree}
\hypo{\ctxitem{I \texttt{; (} T_{1}: x_{1}, \ldots, T_{n}: x_{n} \texttt{) = } e }}
\infer[no rule]1{\text{collectCtx}(x_{1}) = [ctx^{1}_{1}, \ldots, ctx^{1}_{k}], \ldots, \text{collectCtx}(x_{n}) = [ctx^{n}_{1}, \ldots, ctx^{n}_{l}]}
\infer1[(TupD)]{\ctxeval{e}{\tuple{\text{join}\left(\left[ctx^{1}_{1}, \ldots, ctx^{1}_{k}\right]\right), \ldots, \text{join}\left(\left[ctx^{n}_{1}, \ldots, ctx^{n}_{l}\right]\right)}},~\ctxitem{I}}
\end{prooftree} \\
\bigskip
\begin{prooftree}
\hypo{\ctxitem{I \texttt{; } T: x}}
\infer1[(Decl)]{}
\end{prooftree} \\
\bigskip
\begin{prooftree}
\hypo{\ctxitem{\epsilon{}}}
\infer1[(Item0)]{}
\end{prooftree}
\caption{\label{fig:half-analysis-it} Context inference rules for \microzinc\ let-expressions.}
\end{figure*}
\Cref{fig:half-analysis-it} shows the inference rules for let expressions and their inner items.
The first rule, (Let), propagates the context in which the expression is evaluated, \(ctx\), directly to the \mzninline{in}-expression.
Thereafter, the analysis will continue by iterating over its inner items.
This is described using the newly introduced syntax \ctxitem{I}.
Note that the \(ctx\) of the let-expression itself, is irrelevant for the analysis of its inner items.
The inference for \constraint{} items is described by the (Con) rule.
Since \microzinc{} implements strict semantics, the \constraint{} can be assumed to hold globally.
And, unlike \minizinc{}, the \constraint{} is not dependent on the context of the let-expression.
The \constraint{}'s expression is evaluated in \rootc context.
The (Assign) rule reasons about declarations that have a right hand side.
The expression that is assigned to the identifier must evaluated in the combined context of its usages.
As previously discussed, the contexts in which the identifier was used can be retrieved using the ``collectCtx'' function.
These contexts are then combined using a ``join'' function.
This function repeatedly applies the symmetric join operation described by \cref{fig:half-join}.
The right hand expression of the item is then evaluated in the resulting context.
\begin{figure*}
\begin{center}
\begin{tabular}{r | c c c c}
join & \rootc & \posc & \negc & \mixc \\
\hline
\rootc & \rootc & \rootc & \rootc & \rootc \\
\posc & \rootc & \posc & \mixc & \mixc \\
\negc & \rootc & \mixc & \negc & \mixc \\
\mixc & \rootc & \mixc & \mixc & \mixc \\
\end{tabular}
\end{center}
\caption{\label{fig:half-join} A table showing the result of joining two contexts.}
\end{figure*}
(TupC) and (TupD) handle the context inference during the construction and destructuring of tuples respectively.
The context of the individual members of tuples is tracked separately.
This means that individual members of a tuple, like the value and the partiality of a \minizinc{} expression, may be evaluated in different contexts.
Finally, the (Decl) and (Item0) rules describe two base cases in the inference.
The declaration item of a \variable{} does not further the context, and does not depend on it.
It merely triggers the creation of a new \variable{}.
The (Item0) rule is triggered when there are no more inner items in the let-expression.
\subsection{Potentially \rootc{}}%
\label{subsec:half-?root}
In the previous section, we briefly discussed the context transformations for the (Access) and (ITE) rules in \cref{fig:half-analysis-expr}.
Different from the rules described, when an array access or if-then-else expression is found in \rootc{} context, it often makes sense to evaluate its sub-expression in \rootc context as well.
It is, however, not always safe to do so.
\begin{example}
\label{ex:half-maybe-root}
For example, consider the following \microzinc{} fragment.
\begin{mzn}
constraint if b then F(x, y, z) else G(x, y, z) endif;
\end{mzn}
In this case, since only one side of the if-then-else expression is evaluated, the compiler can output calls to the \rootc{} variant of the functions.
This will enforce the constraint in the most efficient way.
Things, however, change when the situation gets more complex.
Consider the following \microzinc{} fragment.
\begin{mzn}
let {
var bool: p = F(x, y, z);
var bool: q = G(x, y, z);
constraint if b then p else q endif;
var bool: ret = bool_or(p, r);
} in ret;
\end{mzn}
One side of the if-then-else expression is not also used in a disjunction.
If \mzninline{b} evaluates to \mzninline{true}, then \mzninline{p} is evaluated in \rootc{} context, and \mzninline{p} can take the value \mzninline{true} in the disjunction.
Otherwise, \mzninline{q} is evaluated in \rootc{} context, and \mzninline{p} in the disjunction must be evaluated in \posc{} context.
It is, therefore, not safe to assume that all sides of the if-then-else expressions are evaluated in \rootc{} context.
\end{example}
Using the \changepos{} transformation for sub-expression contexts is safe, but it might place a large burden on the \solver{}.
The solver is likely to perform better when the direct constraint predicate is used.
To detect situation where the sub-expression are only used in an array access or if-then-else expression we introduce the \mayberootc{} context.
This context functions as a \emph{weak} \rootc{} context.
If it is joined with any other context, then it acts as \posc{}.
The extended join operation is shown in \cref{fig:half-maybe-join}.
Otherwise, it will act as a normal \rootc{} context.
\begin{figure*}
\begin{center}
\begin{tabular}{r | c c c c c}
join & \rootc & \mayberootc & \posc & \negc & \mixc \\
\hline
\rootc & \rootc & \rootc & \rootc & \rootc & \rootc \\
\mayberootc & \rootc & \mayberootc & \posc & \mixc & \mixc \\
\posc & \rootc & \posc & \posc & \mixc & \mixc \\
\negc & \rootc & \mixc & \mixc & \negc & \mixc \\
\mixc & \rootc & \mixc & \mixc & \mixc & \mixc \\
\end{tabular}
\end{center}
\caption{\label{fig:half-maybe-join} The join context operation extended with \mayberootc{}.}
\end{figure*}
\begin{figure*}
\centering
\begin{prooftree}
\hypo{\ctxeval{x\texttt{[}i\texttt{]}}{\rootc}}
\infer1[(Access-R)]{\ctxeval{x}{\mayberootc}}
\end{prooftree} \\
\bigskip
\begin{prooftree}
\hypo{\ctxeval{\texttt{if }b\texttt{ then } e_{1} \texttt{ else } e_{2} \texttt{ endif}}{\rootc}}
\infer1[(ITE-R)]{\ctxeval{e_{1}}{\mayberootc},~\ctxeval{e_{2}}{\mayberootc}}
\end{prooftree}
\caption{\label{fig:half-analysis-maybe-root} Updated context inference rules for \mayberootc{}.}
\end{figure*}
\Cref{fig:half-analysis-maybe-root} shows the additional inference rules for array access and if-then-else expressions.
Looking back at \cref{ex:half-maybe-root}, these additional rules and updated join operation will ensure that the first case will correctly use \rootc{} context calls.
For the second case, however, it detects that \mzninline{p} is used in both \posc{} and \mayberootc{} context.
Therefore, it will output the \posc{} call for the right hand side of \mzninline{p}, even when \mzninline{b} evaluates to \mzninline{true}.
At compile time, this is only correct context to use.
We will, however, discuss the dynamically adjusting of contexts during flattening in \cref{subsec:half-dyn-context}.
\section{Flattening and Half Reification}%
\label{sec:half-flattening}
During the flattening process the contexts assigned to the different expressions can be used directly to determine if and how a expression has to be reified.
\begin{example}
\label{ex:half-flatten}
\jip{TODO:\ Replace the previous example. It was too long and complex.}
\end{example}
A consequence of the use of \gls{half-reif}, shown in the example, is that it might form so called \emph{implication chains}.
This happens when the right hand side of an implication is half reifed and a new Boolean variable is created to represent the variable.
Instead, we could have directly posted the half-reified constraint using the left hand side of the implication as its control variable.
In \cref{subsec:half-compress} we present a new post-processing method, \emph{chain compression}, that can be used to eliminate these implication chains.
The flattening with \gls{half-reif} also interacts with some of the optimisation methods used during the flattening process.
Most importantly, \gls{half-reif} has to be considered when using \gls{cse} and \gls{propagation} might change the context of expression.
In \cref{subsec:half-cse} we will discuss how \gls{cse} can be adjusted to handle \gls{half-reif}.
Finally, in \cref{subsec:half-dyn-context} we will discuss how to context in which a expression is executed can be adjusted during the flattening process.
\subsection{Chain compression}%
\label{subsec:half-compress}
As shown in \cref{ex:half-flatten}, flattening with half reification will in many cases result in implication chains: \mzninline{b1 -> b2 /\ b2 -> c}, where \texttt{b2} has no other occurrences.
In this case the conjunction can be replaced by \mzninline{b1 -> c} and \texttt{b2} can be removed from the model.
The case shown in the example can be generalised to
\begin{mzn}
b1 -> b2 /\ forall(i in N)(b2 -> c[i])
\end{mzn}
\noindent{}which, if \texttt{b2} has no other usage in the instance, can be resolved to
\begin{mzn}
forall(i in N)(b1 -> c[i])
\end{mzn}
\noindent{}after which \texttt{b1} can be removed from the model.
An algorithm to remove these chains of implications is best visualised through the use of an implication graph.
An implication graph \(\tuple{V,E}\) is a directed graph where the vertices \(V\) represent the variables in the instance and an edge \(\tuple{x,y} \in E\) represents the presence of an implication \mzninline{x -> y} in the instance.
Additionally, for the benefit of the algorithm, a vertex is marked when it is used in other constraints in the constraint model.
The goal of the algorithm is now to identify and remove vertices that are not marked and have only one incoming edge.
\Cref{alg:half-compression} provides a formal specification of the chain compression method in pseudo code.
\begin{algorithm}
\KwIn{An implication constraint graph \(G=\tuple{V, E}\) and a set \(M
\subseteq{} V\) of vertices used in other constraints.}
\KwOut{An equisatisfiable graph \(G'=\tuple{V', E'}\) where chained
implications have been removed.}
\(V' \longleftarrow V\)\;
\(E' \longleftarrow E\)\;
\For{\(x \in V\)} {
\If{\(x \not\in M\) \textbf{ and }\(\left|\left\{\tuple{a,x}| \tuple{a,x} \in E\right\}\right| = 1\)}{
\For{\(\tuple{x, b} \in E\)}{
\(E' \longleftarrow E' \cup \{ \tuple{a,b} \} \)\;
\(E' \longleftarrow E' \backslash \{ \tuple{x,b} \} \)\;
}
\(E' \longleftarrow E' \backslash \{ \tuple{a,x} \} \)\;
\(V' \longleftarrow V' \backslash \{ x \} \)\;
}
}
\(G' \longleftarrow \tuple{V', E'}\)\;
\caption{\label{alg:half-compression} Implication chain compression algorithm}
\end{algorithm}
The algorithm can be further improved by considering implied conjunctions.
These can be split up into multiple implications:
\begin{mzn}
b -> forall(x in N)(x)
\end{mzn}
\noindent{}is equivalent to
\begin{mzn}
forall(x in N)(b -> x)
\end{mzn}
Adopting this transformation both simplifies a complicated constraint and possibly allows for the further compression of implication chains.
It should however be noted that although this transformation can result in an increase in the number of constraints, it generally increases the propagation efficiency.
To adjust the algorithm to simplify implied conjunctions more introspection from the \minizinc{} compiler is required.
The compiler must be able to tell if a variable is (only) a control variable of a reified conjunction and what the elements of that conjunction are.
In the case where a variable has one incoming edge, but it is marked as used in other constraint, we can now check if it is only a control variable for a reified conjunction and perform the transformation in this case.
\subsection{Common Sub-expression Elimination}%
\label{subsec:half-cse}
When using full \gls{reification}, all \glspl{reification} are stored in the \gls{cse} table.
This ensure that if we see the same expression is reified twice, then the resulting \variable{} would be reusing.
This avoids that the solver has to enforce the same functional relationship twice.
If the flattener uses \gls{half-reif}, in addition to full \gls{reification}, then \gls{cse} needs to ensure not just that the expressions are equivalent, but also that the context of the two expressions are compatible.
For example, if an expression was first found in a \posc{} context and later found in a \mixc{} context, then the resulting \gls{half-reif} from the first cannot be used for the second expression.
In general:
\begin{itemize}
\item The flattening result of a \posc{} context, a \gls{half-reif}, can only be reused if the same expression is again found in \posc{} context.
\item The flattening result of a \negc{} context, a \gls{half-reif} with its negation pushed inwards, can only be reused if the same expression is again found in \negc{} context.
\item The flattening result of a \mixc{} context, a \gls{reification}, can be reused in \posc{}, \negc{}, and \mixc{} context.
Since we assume that the result of a flattening an expression in \negc{} context pushes the negation inwards, the \gls{reification} does, however, need to be negated.
\item If the expression was already flattened in \rootc{} context, then any repeated usage of the expression can be assumed to take the value \mzninline{true} (or \mzninline{false} in \negc{} context).
\end{itemize}
When considering these compatibility rules, the result of flattening would be highly dependent on the order in which expressions are seen by the flattener.
It would always be better to encounter the expression in a context that results in a reusable expression, \eg{} \mixc{}, before seeing the same expression in another context, \eg{} \posc{}.
This avoids creating both a full \gls{reification} and a \gls{half-reif} of the same expression.
In the \microzinc{} interpreter, this problem is resolved by only keeping the result of their \emph{joint} context.
The context recorded in the \gls{cse} table and the found context are joint using the join operator, as described in \cref{fig:half-join}.
If this context is different from the recorded context, then the expression is re-evaluated in the joint context and its result kept in the \gls{cse} table.
All usages of the previously recorded result is replaced by the new result.
Because of dependency tracking of the constraints that define variables, we can be sure that all \variables{} and \constraints{} created in defining the earlier version are correctly removed.
Because the expression itself is changed when a negation is moved inwards, it may not always be clear when the same expression is used in both \posc{} and negc{} context.
This problem is solved by introducing a canonical form for expressions where negations can be pushed inwards.
In this form the result of flattening an expression and its negation are collected in the same place within the \gls{cse} table.
If it is found that for an expression that is about to be half reified there already exists an \gls{half-reif} for its negation, then we instead evaluate the expression in mixed context, reifying the expression and replacing the existing half reified expression.
This canonical form for expressions and their negations can also be used for the expressions in other contexts.
Using the canonical form we can now also be sure that we never create a full \gls{reification} for both an expression and its negation.
Instead, when one is created, the negation of the resulting \variable{} can directly be used as the result of reifying its negation.
Moreover, this mechanism also allows us to detect when an expression and its negation occur in \rootc{} context.
This is simple way to detect conflicts between \constraints{} and, by extend, prove that the constraint model is unsatisfiable.
Clearly, a \constraint{} and its negation cannot both hold at the same time.
\subsection{Dynamic Context Switching}%
\label{subsec:half-dyn-context}
In \cref{subsec:half-?root} we discussed the fact that the correct context of an expression might not be known when analysing a \microzinc{} model.
Its context depends on data that is only known at an instance level.
The same situation can be caused by \gls{propagation}.
\begin{example}
Consider the following \minizinc{} fragment
\begin{mzn}
var 1..4: x;
var 5..10: y;
var bool: b = x < y;
constraint b -> (2*x = y);
\end{mzn}
Since the domain of \mzninline{x} is strictly smaller than the domain of \mzninline{y}, propagation of \mzninline{b} will set it to the value \mzninline{true}.
This however means that the constraint is equivalent to
\begin{mzn}
constraint 2*x = y;
\end{mzn}
The linear constraint could be evaluated in \rootc{} context, instead of the \posc{} context that is detected by our context analysis.
\end{example}
The situation shown in the example is the most common change of context.
If the control \variable{} of a \gls{reification} is fixed, the context often changes to either \rootc{} or a negated \rootc{} context.
If, on the other hand, the control \variable{} of a \gls{half-reif} is fixed, then either the context becomes \rootc{} or the constraint already holds.
Since regular \constraints{} are strongly preferred over any form of \gls{reification}, it is important to dynamically pick the correct form at during the flattening process.
This problem can be solved by the compiler.
For each \gls{reification} and \gls{half-reif} the compiler should introduce another layer of decomposition.
In this layer, it checks its control \variable{}.
If the control \variable{} is already fixed, then it rewrites itself into its form in another context.
Otherwise, it behaves as it would have done normally.
The control \variable is thus used to communicate any change in context.
\begin{example}
Let's assume the compiler find a call to \mzninline{int_lin_le} in \posc{} context.
Instead of outputting the call to \mzninline{int_lin_le_imp} directly, it will instead output a call to \mzninline{_int_lin_le_imp}.
This predicate is then generated as follows:
\begin{mzn}
predicate _int_lin_le_imp(
array[int] of int: c,
array[int] of var int: x,
int: d,
var bool: b
) =
if is_fixed(b) then
if fix(b) then
int_lin_le(c, x, d)
else
true
endif
else
int_lin_le_reif(c, x, d, b)
endif;
\end{mzn}
This new predicate can then compiled using the normal methods.
\end{example}
\jip{Should we talk about \mixc{} that could have been \posc{}? I'm not sure we correctly do that one at the moment.}
\section{Experiments}
\label{sec:half-experiments}
We now present experimental evaluation of the presented \gls{half-reif} techniques.
First, to show the benefit of implementing propagators for half-reified constraint, we compare their performance against their decompositions.
To do this, we recreate two experiments presented by Feydy et al.\ in the original \gls{half-reif} paper in a modern \gls{cp} solver, \gls{chuffed}.
In the experiment, we use propagators implemented according to the principles described in this paper.
No new algorithm has been devised to perform the propagation.
The propagator of the original constraint is merely adjusted to influence and watch a control \variable{}.
Additionally, we assess the effects of automatically detecting and introducing \glspl{half-reif} during the flattening process in general.
We flatten and solve 200 \minizinc{} instances for several \solvers{} with and without the use of \gls{half-reif}.
We then analyse the trends in the generated \flatzinc{} models and performance in solving the instances.
A description of the used computational environment, \minizinc{} instances, and versioned software has been included in \cref{ch:benchmarks}.
\subsection{Half Reified Propagators}
\label{sec:half-exp-prop}
Our first experiment considers the \gls{qcp-max} quasi-group completion problem.
In this problem, we need to decide the value of an \((n \times n)\) matrix of integer \variables, with domains \mzninline{1..n}.
The aim of the problem is to create as many rows and columns where all \variables{} take a unique value.
In each instance certain values have already been fixed.
It is, thus, not always possible for all rows and columns to contain only distinct values.
In \minizinc{} counting the number of rows/columns with all different values can be accomplished by reifying the \mzninline{all_different} constraint.
Since the goal of the problem is to maximise the number of \mzninline{all_different} \constraints{} that hold, these constraints are never forced to be \mzninline{false}.
This means these constraints in a \posc{} context and can be half-reified.
\Cref{tab:half-qcp} shows the comparison of two solving configurations in \gls{chuffed} for the \gls{qcp-max} problem.
The results are grouped based on their size of the instance.
For each group we show the number of instances solved by the configuration and the average time used for this process.
The first configuration uses a newly created propagator for half-reified \mzninline{all_different} \constraints{}.
This propagator is an adjusted version from the existing bounds consistent \mzninline{all_different} propagator in chuffed.
The implementation of the propagator was already split into parts that \emph{check} the violation of the constraint and parts that \emph{prune} the \glspl{domain} of \variables{}.
Therefore, the transformation described in \cref{sec:half-propagation} can be directly applied.
Since \gls{chuffed} is a \gls{lcg} \solver{}, the explanations created by the propagator have to be adjusted as well.
These adjustments happen in a similar fashion to the adjustments of the general algorithm: explanations used for the violation of the \constraint{} can now be used to set the control variable to \mzninline{false} and the explanations given to prune a variable are appended by requirement that the control variable is \mzninline{true}.
The second configuration uses the following decomposition for the \mzninline{all_different} constraint.
\begin{mzn}
predicate all_different(array[int] of var int: x) =
forall(i,j in index_set(x) where i < j)(
x[i] != x[j]
);
\end{mzn}
The \mzninline{!=} constraints produced by this redefinition are reified.
Their conjunction, then represent the reification of the \mzninline{all_different} constraint.
\begin{table}
\begin{center}
\input{assets/table/half_qcp}
\caption{\label{tab:half-qcp} \gls{qcp-max} problems: number of solved instances and average time (in seconds) with a 300s timeout.}
\end{center}
\end{table}
The results in \cref{tab:half-qcp} show that the usage of the specialised propagator has a significant advantage over the use of the decomposition.
Although it only allows us to solve a single extra instance, there is a significant reduction in solving time for most instances.
Note that the qcp-15 instances are the only exception.
It appears that none of the instances in this group proved to be a real challenge to either method and we see similar solve times between the two methods.
For our second experiment we consider a variation on the prize collecting travelling salesman problem \autocite{balas-1989-pctsp} referred to as \emph{prize collecting path}.
In the problem we are given a graph with weighted edges, both positive and negative.
The aim of the problem is to find the optimal acyclic path from a given start node that maximises the weights on the path.
It is not required to visit every node.
In this experiment we can show how \gls{half-reif} can reduce the overhead of handling partial functions correctly.
The \minizinc{} model for this problem contains a unsafe array lookup \mzninline{pos[next[i]]}, where the domain of \mzninline{next[i]} is larger than the index set of \mzninline{pos}.
We compare safe decomposition of this \mzninline{element} constraint against a propagator of its \gls{half-reif}.
The decomposition creates a new variable that takes the value of the index only when it is within the index set of the array.
Otherwise, it will set its surrounding context to \mzninline{false}.
The \gls{half-reif} implicitly performs the same task by setting its control \variable{} to \mzninline{false} whenever the result of the \mzninline{element} constraint does not match the value of the index variable.
Again, for the implementation of the propagator of the \gls{half-reif} constraint we adjust the regular propagator as described above.
\begin{table}
\begin{center}
\input{assets/table/half_prize}
\caption{\label{tab:half-prize} Prize collecting paths: number of solved instances and average time (in seconds) and with a 300s timeout.}
\end{center}
\end{table}
The results of the experiment are shown in \cref{tab:half-prize}.
Although the performance on smaller instances is similar, the dedicated propagator consistently outperforms the usage of the decomposition.
The difference in performance becomes more pronounced in the bigger instances.
In the 32-4-8 group, we even see that usage of the propagator allows us to solve an additional three instances.
\subsection{Flattening with Half Reification}
\label{sec:half-exp-flatten}
The introduction of automated context analysis and introduction of \glspl{half-reif} allows us to evaluate the usage of \gls{half-reif} on a larger scale.
In our second experiment we assess its effects on the flattening and solving of the instances of the 2019 and 2020 \minizinc{} challenge \autocite{stuckey-2010-challenge,stuckey-2014-challenge}.
These experiments are conducted using the \gls{gecode} \solver{}, which has propagators for \glspl{half-reif} of many basic constraint, and the \minizinc{}'s linearisation library, which has been adapted to use \gls{half-reif} as earlier described.
The \minizinc{} instances are flattened using the \minizinc{} 2.5.5 flattener, which can enable and disable the usage of \gls{half-reif}.
The solving of the linearised models is tested using the \gls{cbc} and \gls{cplex} \gls{mip} \solvers{}.
\jip{TODO:\ Extend this section with the \gls{sat} results once they are run.}
\begin{table}
\begin{subtable}[b]{\linewidth}
\input{assets/table/half_flat_gecode}
\caption{\label{subtab:half-flat-gecode}\gls{gecode} library}
\end{subtable}
\begin{subtable}[b]{\linewidth}
\input{assets/table/half_flat_linear}
\caption{\label{subtab:half-flat-lin}Linearisation library}
\end{subtable}
\begin{subtable}[b]{\linewidth}
\input{assets/table/half_flat_sat}
\caption{\label{subtab:half-flat-bool}Booleanisation library}
\end{subtable}
\caption{\label{tab:half-flattening} Cumulative statistics of flattening all \minizinc{} instances from \minizinc{} Challenge 2019 \& 2020 (200 instances).}
\end{table}
Grouped by \solver{} library and whether \gls{half-reif} is used, \cref{tab:half-flattening} shows several cumulative figures from the flattening process of the \minizinc{} challenge.
These are:
\begin{itemize}
\item The number of \emph{constraints} in \flatzinc{}.
\item The number of \emph{\glspl{reification}} evaluated during the flattening process.
This includes both the \glspl{reification} that are decomposed and the \glspl{reification} that are present in the \flatzinc{}.
\item The number of \emph{\glspl{half-reif}} evaluated during the flattening process.
\item The number of \emph{implications removed} using the chain compression method.
\item The runtime of the flattening process (\ie{} the \emph{flattening time}).
\end{itemize}
The flattening statistics for the \gls{gecode} \solver{} library, shown in \cref{subtab:half-flat-gecode}, show significant changes in the resulting \flatzinc{}.
Although the total number of constraints remains stable, we see that well over half of all \glspl{reification} are replaced by \glspl{half-reif}.
This replacement happens mostly 1-for-1; the difference between the number of \glspl{half-reif} introduced and the number of \glspl{reification} reduced is only 20. In comparison, the number of implications removed by chain compression looks small, but this number is highly dependent on the \minizinc{} model.
In many models, no implications can be removed, but for some problems an implication is removed for every \gls{half-reif} that is introduced.
Finally, the overhead of the introduction of \gls{half-reif} and the newly introduced optimisation techniques is minimal.
The \Cref{subtab:half-flat-lin} paints an equally positive picture for the usage of \glspl{half-reif} for linearisation.
Since both \glspl{reification} and \glspl{half-reif} are decomposed during the flattening process, the usage of \gls{half-reif} is able to remove almost 7.5\% of the overall constraints.
The ratio of \glspl{reification} that is replaced with \glspl{half-reif} is not as high as \gls{gecode}.
This is caused by the fact that the linearisation process requires full \gls{reification} in the decomposition of many \gls{global} \constraints{}.
Similar to \gls{gecode}, the number of implications that is removed is dependent on the problem.
Lastly, the flattening time slightly increases for the linearisation process.
Since there are many more constraints, the introduced optimisation mechanisms have an slightly higher overhead.
\jip{TODO:\ The \gls{sat} statistics currently only include 2019. 2020 is still WIP.}
Finally, statistics for the flattening the instances is shown in \cref{subtab:half-flat-bool}.
Like linearisation, the usage of \gls{half-reif} significantly reduces number of constraints and \glspl{reification}.
Different, however, is that the booleanisation library is explicitly defined in terms of \glspl{half-reif}.Some constraints might manually introduce \mzninline{_imp} call as part of their definition.
Furthermore, the usage of chain compression does not seem to have any effect.
Since all \glspl{half-reif} are defined in terms of clauses, the implications normally removed using chain compression are instead aggregated into bigger clauses.
Surprisingly, the usage of \gls{half-reif} also reduces the flattening time as it reduces the workload.
\begin{table}
\input{assets/table/half_mznc}
\caption{\label{tab:half-mznc} Status overview of solving \minizinc{} Challenge 2019 \& 2020 with and without \gls{half-reif}.}
\end{table}
\Cref{tab:half-mznc} shows the results reported by the solvers. The \solver{} reports
\begin{itemize}
\item \emph{Unsatisfiable} when it proofs the instance does not have a solution,
\item \emph{Optimal solution} when it has found a solution and has proven it optimal,
\item \emph{Satisfied} when it has found a solution for the problem,
\item \emph{Unknown} when no solution is found,
\item and \emph{Error} when the \solver{} program crashes.
\end{itemize}
\noindent{}For \solver{} statuses that end the solving process before the time-out of 15 minutes we also show the average time.
The results shown is this table are very mixed.
For \gls{gecode}, the usage of \gls{half-reif} does not seem to impact its solving performance.
Although we would have hoped that propagators for \glspl{half-reif} would be more efficient and reduce the number of propagators scheduled in general.
Neither number of instances solved, nor the solving time required improved.
A single instance, however, is negatively impacted by the change; an optimal solution for this instance is no longer found.
We expect that this instance might have benefited from the increased Boolean propagation that is caused by full \gls{reification}.
Overall, these results do not show any significant positive or negative effects in \gls{gecode}'s performance when using \gls{half-reif}.
When using \gls{cplex} the usage of \gls{half-reif} is clearly a positive one.
Although it no longer proves the unsatisfiability of one instance and slightly increases the number of solver errors, an optimal solution is found for five more instances.
The same linearised instances when using the \gls{cbc} solver seem to have the opposite effect.
Eventhough it reduces the time required to prove that two instances are unsatisfiable, it can no longer find six optimal solutions.
These results are hard to explain.
In general, we would expect the reduction of constraints in a \gls{mip} instance would help the \gls{mip} solver.
However, we can imagine that the removed constraints might in some cases help the \gls{mip} solver.
An important technique used by \gls{mip} solvers is to detect certain pattern, such as cliques, during the pre-processing of the \gls{mip} instance.
Some patterns might only be detected when using a full \gls{reification}.
Furthermore, the performance of \gls{mip} solvers is often dependent on the order and form in which the constraints are given.
\jip{TODO:\ Is there a citation for this?} With the usage of the \gls{aggregation} and \gls{del-rew}, these can be exceedingly different when using \gls{half-reif}.
\jip{TODO: \gls{sat} number are still preliminary, but look optimistic.
Only one case where the solver time is severely impacted.}
% \section{Summary}
% \label{sec:half-summary}