text
stringlengths 6
128k
|
---|
# Deeply Coupled Auto-encoder Networks for
Cross-view Classification
Wen Wang, Zhen Cui, Hong Chang, Shiguang Shan, Xilin Chen
Institute of Computing Technology, Chinese Academy of Sciences, Beijing, China
{wen.wang, zhen.cui, hong.chang, shiguang.shan<EMAIL_ADDRESS>
(November 2013)
###### Abstract
The comparison of heterogeneous samples extensively exists in many
applications, especially in the task of image classification. In this paper,
we propose a simple but effective coupled neural network, called Deeply
Coupled Autoencoder Networks (DCAN), which seeks to build two deep neural
networks, coupled with each other in every corresponding layers. In DCAN, each
deep structure is developed via stacking multiple discriminative coupled auto-
encoders, a denoising auto-encoder trained with maximum margin criterion
consisting of intra-class compactness and inter-class penalty. This single
layer component makes our model simultaneously preserve the local consistency
and enhance its discriminative capability. With increasing number of layers,
the coupled networks can gradually narrow the gap between the two views.
Extensive experiments on cross-view image classification tasks demonstrate the
superiority of our method over state-of-the-art methods.
## 1 Introduction
Real-world objects often have different views, which might be endowed with the
same semantic. For example, face images can be captured in different poses,
which reveal the identity of the same object; images of one face can also be
in different modalities, such as pictures under different lighting condition,
pose, or even sketches from artists. In many computer vision applications,
such as image retrieval, interests are taken in comparing two types of
heterogeneous images, which may come from different views or even different
sensors. Since the spanned feature spaces are quite different, it is very
difficult to classify these images across views directly. To decrease the
discrepancy across views, most of previous works endeavored to learn view-
specific linear transforms and to project cross-view samples into a common
latent space, and then employed these newly generated features for
classification.
Though there are lots of approaches used to learn view-specific projections,
they can be divided roughly based on whether the supervised information is
used. Unsupervised methods such as Canonical Correlation Analysis (CCA)[14]
and Partial Least Square (PLS) [26] are employed to the task of cross-view
recognition. Both of them attempt to use two linear mappings to project
samples into a common space where the correlation is maximized, while PLS
considers the variations rather than only the correlation in the target space.
Besides, with use of the mutual information, a Coupled Information-Theoretic
Encoding (CITE) method is developed to narrow the inter-view gap for the
specific photo-sketch recognition task. And in [30], a semi-coupled dictionary
is used to bridge two views. All the methods above consider to reduce the
discrepancy between two views, however, the label information is not
explicitly taken into account. With label information available, many methods
were further developed to learn a discriminant common space For instance,
Discriminative Canonical Correlation Analysis (DCCA) [16] is proposed as an
extension of CCA. And In [22], with an additional local smoothness
constraints, two linear projections are simultaneously learnt for Common
Discriminant Feature Extraction (CDFE). There are also other such methods as
the large margin approach [8] and the Coupled Spectral Regression (CSR) [20].
Recently, multi-view analysis [27, 15] is further developed to jointly learn
multiple specific-view transforms when multiple views (usually more than 2
views) can be available.
Although the above methods have been extensively applied in the cross-view
problem, and have got encouraging performances, they all employed linear
transforms to capture the shared features of samples from two views. However,
these linear discriminant analysis methods usually depend on the assumption
that the data of each class agrees with a Gaussian distribution, while data in
real world usually has a much more complex distribution [33]. It indicates
that linear transforms are insufficient to extract the common features of
cross-view images. So it’s natural to consider about learning nonlinear
features.
A recent topic of interest in nonlinear learning is the research in deep
learning. Deep learning attempts to learn nonlinear representations
hierarchically via deep structures, and has been applied successfully in many
computer vision problems. Classical deep learning methods often stack or
compose multiple basic building blocks to yield a deeper structure. See [5]
for a recent review of Deep Learning algorithms. Lots of such basic building
blocks have been proposed, including sparse coding [19], restricted Boltzmann
machine (RBM) [12], auto-encoder [13, 6], etc. Specifically, the (stacked)
auto-encoder has shown its effectiveness in image denoising [32], domain
adaptation [7], audio-visual speech classification [23], etc.
As we all known, the kernel method, such as Kernel Canonical Correlation
Analysis(Kernel CCA) [1], is also a widely used approach to learn nonlinear
representations. Compared with the kernel method, deep learning is much more
flexible and time-saving because the transform is learned rather than fixed
and the time needed for training and inference process is beyond the limit of
the size of training set.
Inspired by the deep learning works above, we intend to solve the cross-view
classification task via deep networks. It’s natural to build one single deep
neural network with samples from both views, but this kind of network can’t
handle complex data from totally different modalities and may suffer from
inadequate representation capacity. Another way is to learn two different deep
neural networks with samples of the different views. However, the two
independent networks project samples from different views into different
spaces, which makes comparison infeasible. Hence, building two neural networks
coupled with each other seems to be a better solution.
In this work, we propose a Deeply Coupled Auto-encoder Networks(DCAN) method
that learns the common representations to conduct cross-view classification by
building two neural networks deeply coupled respectively, each for one view.
We build the DCAN by stacking multiple discriminative coupled auto-encoders, a
denoising auto-encoder with maximum margin criterion. The discriminative
coupled auto-encoder has a similar input corrupted and reconstructive error
minimized mechanism with the denoising auto-encoder proposed in [28], but is
modified by adding a maximum margin criterion. This kind of criterion has been
used in previous works, like [21, 29, 35], etc. Note that the counterparts
from two views are added into the maximum margin criterion simultaneously
since they both come from the same class, which naturally couples the
corresponding layer in two deep networks. A schematic illustration can be seen
in Fig.1.
The proposed DCAN is related to Multimodal Auto-encoders [23], Multimodal
Restricted Boltzmann Machines and Deep Canonical Correlation Analysis [3]. The
first two methods tend to learn a single network with one or more layers
connected to both views and to predict one view from the other view, and the
Deep Canonical Correlation Analysis build two deep networks, each for one
view, and only representations of the highest layer are constrained to be
correlated. Therefore, the key difference is that we learn two deep networks
coupled with each other in representations in each layer, which is of great
benefits because the DCAN not only learn two separate deep encodings but also
makes better use of data from the both two views. What’s more, these
differences allow for our model to handle the recognition task even when data
is impure and insufficient.
The rest of this paper is organized as follows. Section 2 details the
formulation and solution to the proposed Deeply Coupled Auto-encoder Networks.
Experimental results in Section 3 demonstrate the efficacy of the DCAN. In
section 4 a conclusion is given.
## 2 Deeply Coupled Auto-encoder Networks
In this section, we first present the basic idea. The second part gives a
detailed description of the discriminative coupled auto-encoder. Then, we
describe how to stack multiple layers to build a deep network. Finally, we
briefly describe the optimization of the model.
### 2.1 Basic Idea
Figure 1: An illustration of our proposed DCAN. The left-most and right-most
schematic show the structure of the two coupled network respectively. And the
schematic in the middle illustrates how the whole network gradually enhances
the separability with increasing layers, where pictures with solid line border
denote samples from view 1, those with dotted line border denote samples from
view 2, and different colors imply different subjects.
As shown in Fig.1, the Deeply Coupled Auto-encoder Networks(DCAN) consists of
two deep networks coupled with each other, and each one is for one view. The
network structures of the two deep networks are just like the left-most and
the right-most parts in Fig.1, where circles means the units in each layers
(pixels in a input image for the input layer and hidden representation in
higher layers), and arrows denote the full connections between adjacent
layers. And the middle part of Fig.1 illustrates how the whole network
projects samples in different views into a common space and gradually enhances
the separability with increasing layers.
The two deep networks are both built through stacking multiple similar coupled
single layer blocks because a single coupled layer might be insufficient, and
the method of stacking multiple layers and training each layer greedily has be
proved efficient in lots of previous works, such as those in [13, 6]. With the
number of layers increased, the whole network can compactly represent a
significantly larger set of transforms than shallow networks , and gradually
narrow the gap with the discriminative capacity enhanced.
We use a discriminative coupled auto-encoders trained with maximum margin
criterion as a single layer component. Concretely, we incorporate the
additional noises in the training process while maximizing the margin
criterion, which makes the learnt mapping more stable as well as discriminant.
Note that the maximum margin criterion also works in coupling two
corresponding layers. Formally, the discriminative coupled auto-encoder can be
written as follows:
$\displaystyle\quad\min_{f_{x},f_{y}}\quad L(X,f_{x})+L(Y,f_{y})$ (1)
$\displaystyle s.t.\quad
G_{1}(H_{x},H_{y})-G_{2}(H_{x},H_{y})\leq\varepsilon,$ (2)
where $X,Y$ denote inputs from the two views, and $H_{x},H_{y}$ denote hidden
representations of the two views respectively. $f_{x}:X\longrightarrow
H_{x},f_{y}:Y\longrightarrow H_{y}$ are the transforms we intend to learn, and
we denote the reconstructive error as $L(\cdot)$, and maximum margin criterion
as $G_{1}(\cdot)-G_{2}(\cdot)$, which are described detailedly in the next
subsection.$\varepsilon$ is the threshold of the maximum margin criterion.
### 2.2 Discriminative coupled auto-encoder
In the problem of cross-view, there are two types of heterogenous samples.
Without loss of generality, we denote samples from one view as
$X=[x_{1},\cdots,x_{n}]$ , and those from the other view as
$Y=[y_{1},\cdots,y_{n}]$, in which $n$ is the sample sizes. Noted that the
corresponding labels are known, and $H_{x},H_{y}$ denote hidden
representations of the two views we want to learn.
The DCAN attempts to learn two nonlinear transforms $f_{x}:X\longrightarrow
H_{x}$ and $f_{y}:Y\longrightarrow H_{y}$ that can project the samples from
two views to one discriminant common space respectively, in which the local
neighborhood relationship as well as class separability should be well
preserved for each view. The auto-encoder like structure stands out in
preserving the local consistency, and the denoising form enhances the
robustness of learnt representations. However, the discrimination isn’t taken
into consideration. Therefore, we modify the denoising auto-encoder by adding
a maximum margin criterion consisting of intra-class compactness and inter-
class penalty. And the best nonlinear transformation is a trade-off between
local consistency preserving and separability enhancing.
Just like the one in denoising auto-encoder, the reconstructive error
$L(\cdot)$ in Eq.(1) is formulated as follows:
$\displaystyle
L(X,\Theta)=\sum_{x\in{X^{p}}}{\mathbb{E}_{\tilde{x}\sim{P(\tilde{x}|x)}}}\|\hat{x}-x\|$
(3) $\displaystyle
L(Y,\Theta)=\sum_{y\in{Y^{p}}}{\mathbb{E}_{\tilde{y}\sim{P(\tilde{y}|y)}}}\|\hat{y}-y\|$
(4)
where $\mathbb{E}$ calculates the expectation over corrupted versions
$\tilde{X},\tilde{Y}$ of examples $X,Y$ obtained from a corruption process
$P(\tilde{x}|x),P(\tilde{y}|y)$.
$\Theta=\\{W_{x},W_{y},b_{x},b_{y},c_{x},c_{y}\\}$ specifies the two nonlinear
transforms $f_{x},f_{y}$ , where $W_{x},W_{y}$ is the weight matrix, and
$b_{x},b_{y},c_{x},c_{y}$ are the bias of encoder and decoder respectively,
and $\hat{X},\hat{Y}$ are calculated through the decoder process :
$\begin{split}\hat{X}=s(W_{x}^{T}H_{x}+c_{x})\\\
\hat{Y}=s(W_{y}^{T}H_{y}+c_{y})\end{split}$ (5)
And hidden representations $H_{x},H_{y}$ are obtained from the encoder that is
a similar mapping with the decoder,
$\begin{split}H_{x}=s(W_{x}\tilde{X}+b_{x})\\\
H_{y}=s(W_{y}\tilde{Y}+b_{y})\end{split}$ (6)
where $s$ is the nonlinear activation function, such as the point-wise
hyperbolic tangent operation on linear projected features, i.e.,
$s(x)=\frac{e^{ax}-e^{-ax}}{e^{ax}+e^{-ax}}$ (7)
in which $a$ is the gain parameter.
Moreover, for the maximum margin criterion consisting of intra-class
compactness and inter-class penalty, the constraint term
$G_{1}(\cdot)-G_{2}(\cdot)$ in Eq.(1) is used to realize coupling since
samples of the same class are treated similarly no matter which view they are
from.
Assuming $S$ is the set of sample pairs from the same class, and $D$ is the
set of sample pairs from different classes. Note that the counterparts from
two views are naturally added into $S,D$ since it’s the class rather than the
view that are considered.
Then, we characterize the compactness as follows,
$\displaystyle
G_{1}(H)=\frac{1}{2N_{1}}\sum\limits_{I_{i},I_{j}\in{S}}\|h_{i}-h_{j}\|^{2},$
(8)
where $h_{i}$ denotes the corresponding hidden representation of an input
$I_{i}\in{X\bigcap{Y}}$ and is a sample from either view 1 or view 2, and
$N_{1}$ is the size of $S$.
Meanwhile, the goal of the inter-class separability is to push the adjacent
samples from different classes far away, which can be formulated as follows,
$\displaystyle
G_{2}(H)=\frac{1}{2N_{2}}\sum\limits_{\tiny\begin{subarray}{c}I_{i},I_{j}\in{D}\\\
I_{j}\in{KNN(I_{i})}\end{subarray}}\|h_{i}-h_{j}\|^{2},$ (9)
where $I_{j}$ belongs to the $k$ nearest neighbors of $I_{i}$ with different
class labels, and $N_{2}$ is the number of all pairs satisfying the condition.
And the function of $G_{1}(H),G_{2}(H)$ is illustrated in the middel part of
Fig.1. In the projected common space denoted by $S$, the compactness term
$G_{1}(\cdot)$ shown by red ellipse works by pulling intra-class samples
together while the penalty term $G_{2}(\cdot)$ shown by black ellipse tend to
push adjacent inter-class samples away.
Finally, by solving the optimization problem Eq.(1), we can learn a couple of
nonlinear transforms $f_{x},f_{y}$ to transform the original samples from both
views into a common space.
### 2.3 Stacking coupled auto-encoder
Through the training process above, we model the map between original sample
space and a preliminary discriminant subspace with gap eliminated, and build a
hidden representation $H$ which is a trade-off between approximate
preservation on local consistency and the distinction of the projected data.
But since real-world data is highly complicated, using a single coupled layer
to model the vast and complex real scenes might be insufficient. So we choose
to stack multiple such coupled network layers described in subsection 2.2.
With the number of layers increased, the whole network can compactly represent
a significantly larger set of transforms than shallow networks, and gradually
narrow the gap with the discriminative ability enhanced.
Training a deep network with coupled nonlinear transforms can be achieved by
the canonical greedy layer-wise approach [12, 6]. Or to be more precise, after
training a single layer coupled network, one can compute a new feature $H$ by
the encoder in Eq.(6) and then feed it into the next layer network as the
input feature. In practice, we find that stacking multiple such layers can
gradually reduce the gap and improve the recognition performance (see Fig.1
and Section 3).
### 2.4 Optimization
We adopt the Lagrangian multiplier method to solve the objective function
Eq.(1) with the constraints Eq.(2) as follows:
$\begin{split}\min_{\Theta}\quad&\lambda(L(X,\Theta)+L(Y,\Theta))+(G_{1}(H)-G_{2}(H))+\\\
&\gamma(\frac{1}{2}\|W_{x}\|_{F}^{2}+\frac{1}{2}\|W_{y}\|_{F}^{2})\end{split}$
(10)
where the first term is the the reconstruction error, the second term is the
maximum margin criterion, and the last term is the shrinkage constraints
called the Tikhonov regularizers in [11], which is utilized to decrease the
magnitude of the weights and further to help prevent over-fitting. $\lambda$
is the balance parameter between the local consistency and empirical
separability. And $\gamma$ is called the weight decay parameter and is usually
set to a small value, e.g., 1.0e-4.
To optimize the objective function (10), we use back-propagation to calculate
the gradient and then employ the limited-memory BFGS (L-BFGS) method [24, 17],
which is often used to solve nonlinear optimization problems without any
constraints. L-BFGS is particularly suitable for problems with a large amount
of variables under the moderate memory requirement. To utilize L-BFGS, we need
to calculate the gradients of the object function. Obviously, the object
function in (10) is differential to these parameters $\Theta$, and we use
Back-propagation [18] method to derive the derivative of the overall cost
function. In our setting, we find the objective function can achieve as fast
convergence as described in [17].
## 3 Experiments
In this section, the proposed DCAN is evaluated on two datasets, Multi-PIE [9]
and CUHK Face Sketch FERET (CUFSF) [34, 31].
### 3.1 Databases
Multi-PIE dataset [9] is employed to evaluate face recognition across pose.
Here a subset from the 337 subjects in 7 poses
($-45^{\circ},-30^{\circ},-15^{\circ},0^{\circ},15^{\circ},30^{\circ},45^{\circ}$),
3 expression (Neutral,Smile, Disgust), no flush illumination from 4 sessions
are selected to validate our method. We randomly choose 4 images for each pose
of each subject, then randomly partition the data into two parts: the training
set with 231 subjects (i.e., $231\times 7\times 4=6468$ images) and the
testing set with the rest subjects.
CUHK Face Sketch FERET (CUFSF) dataset [34, 31] contains two types of face
images: photo and sketch. Total 1,194 images (one image per subject) were
collected with lighting variations from FERET dataset [25]. For each subject,
a sketch is drawn with shape exaggeration. According to the configuration of
[15], we use the first 700 subjects as the training data and the rest subjects
as the testing data.
### 3.2 Settings
All images from Multi-PIE and CUFSF are cropped into 64$\times$80 pixels
without any preprocess. We compare the proposed DCAN method with several
baselines and state-of-the-art methods, including CCA [14], Kernel CCA [1],
Deep CCA [3], FDA [4], CDFE [22], CSR [20], PLS [26] and MvDA [15]. The first
seven methods are pairwise methods for cross-view classification. MvDA jointly
learns all transforms when multiple views can be utilized, and has achieved
the state-of-the-art results in their reports [15].
The Principal Component Analysis (PCA) [4] is used for dimension reduction. In
our experiments, we set the default dimensionality as 100 with preservation of
most energy except Deep CCA, PLS, CSR and CDFE, where the dimensionality are
tuned in [50,1000] for the best performance. For all these methods, we report
the best performance by tuning the related parameters according to their
papers. Firstly, for Kernel CCA, we experiment with Gaussian kernel and
polynomial kernel and adjust the parameters to get the best performance. Then
for Deep CCA [3], we strictly follow their algorithms and tune all possible
parameters, but the performance is inferior to CCA. One possible reason is
that Deep CCA only considers the correlations on training data (as reported in
their paper) so that the learnt mode overly fits the training data, which thus
leads to the poor generality on the testing set. Besides, the parameter
$\alpha$ and $\beta$ are respectively traversed in [0.2,2] and [0.0001,1] for
CDFE, the parameter $\lambda$ and $\eta$ are searched in [0.001,1] for CSR,
and the reduced dimensionality is tuned for CCA, PLS, FDA and MvDA.
As for our proposed DCAN, the performance on CUFSF database of varied
parameters, $\lambda,k$, is shown in Fig.3. In following experiments, we set
$\lambda=0.2,\gamma=1.0e-4$, $k=10$ and $a=1$. With increasing layers, the
number of hidden neurons are gradually reduced by $10$, _i.e.,_ $90,80,70,60$
if four layers.
Method | Accuracy
---|---
CCA[14] | 0.698
KernelCCA[10] | 0.840
DeepCCA[3] | 0.599
FDA[4] | 0.814
CDFE[22] | 0.773
CSR[20] | 0.580
PLS[26] | 0.574
MvDA[15] | 0.867
DCAN-1 | 0.830
DCAN-2 | 0.877
DCAN-3 | 0.884
DCAN-4 | 0.879
Table 1: Evaluation on Multi-PIE database in terms of mean accuracy. DCAN-k
means a stacked k-layer network.
| $-45^{\circ}$ | $-30^{\circ}$ | $-15^{\circ}$ | $0^{\circ}$ | $15^{\circ}$ | $30^{\circ}$ | $45^{\circ}$
---|---|---|---|---|---|---|---
$-45^{\circ}$ | 1.000 | 0.816 | 0.588 | 0.473 | 0.473 | 0.515 | 0.511
$-30^{\circ}$ | 0.816 | 1.000 | 0.858 | 0.611 | 0.664 | 0.553 | 0.553
$-15^{\circ}$ | 0.588 | 0.858 | 1.000 | 0.894 | 0.807 | 0.602 | 0.447
$0^{\circ}$ | 0.473 | 0.611 | 0.894 | 1.000 | 0.909 | 0.604 | 0.484
$15^{\circ}$ | 0.473 | 0.664 | 0.807 | 0.909 | 1.000 | 0.874 | 0.602
$30^{\circ}$ | 0.515 | 0.553 | 0.602 | 0.604 | 0.874 | 1.000 | 0.768
$45^{\circ}$ | 0.511 | 0.553 | 0.447 | 0.484 | 0.602 | 0.768 | 1.000
(a) CCA, $Ave=0.698$
| $-45^{\circ}$ | $-30^{\circ}$ | $-15^{\circ}$ | $0^{\circ}$ | $15^{\circ}$ | $30^{\circ}$ | $45^{\circ}$
---|---|---|---|---|---|---|---
$-45^{\circ}$ | 1.000 | 0.878 | 0.810 | 0.756 | 0.706 | 0.726 | 0.737
$-30^{\circ}$ | 0.878 | 1.000 | 0.892 | 0.858 | 0.808 | 0.801 | 0.757
$-15^{\circ}$ | 0.810 | 0.892 | 1.000 | 0.911 | 0.880 | 0.861 | 0.765
$0^{\circ}$ | 0.756 | 0.858 | 0.911 | 1.000 | 0.938 | 0.759 | 0.759
$15^{\circ}$ | 0.706 | 0.808 | 0.880 | 0.938 | 1.000 | 0.922 | 0.845
$30^{\circ}$ | 0.726 | 0.801 | 0.861 | 0.759 | 0.922 | 1.000 | 0.912
$45^{\circ}$ | 0.737 | 0.757 | 0.765 | 0.759 | 0.845 | 0.912 | 1.000
(b) KernelCCA, $Ave=0.840$
| $-45^{\circ}$ | $-30^{\circ}$ | $-15^{\circ}$ | $0^{\circ}$ | $15^{\circ}$ | $30^{\circ}$ | $45^{\circ}$
---|---|---|---|---|---|---|---
$-45^{\circ}$ | 1.000 | 0.854 | 0.598 | 0.425 | 0.473 | 0.522 | 0.523
$-30^{\circ}$ | 0.854 | 1.000 | 0.844 | 0.578 | 0.676 | 0.576 | 0.566
$-15^{\circ}$ | 0.598 | 0.844 | 1.000 | 0.806 | 0.807 | 0.602 | 0.424
$0^{\circ}$ | 0.425 | 0.578 | 0.806 | 1.000 | 0.911 | 0.599 | 0.444
$15^{\circ}$ | 0.473 | 0.676 | 0.807 | 0.911 | 1.000 | 0.866 | 0.624
$30^{\circ}$ | 0.522 | 0.576 | 0.602 | 0.599 | 0.866 | 1.000 | 0.756
$45^{\circ}$ | 0.523 | 0.566 | 0.424 | 0.444 | 0.624 | 0.756 | 1.000
(c) DeepCCA, $Ave=0.599$
| $-45^{\circ}$ | $-30^{\circ}$ | $-15^{\circ}$ | $0^{\circ}$ | $15^{\circ}$ | $30^{\circ}$ | $45^{\circ}$
---|---|---|---|---|---|---|---
$-45^{\circ}$ | 1.000 | 0.847 | 0.754 | 0.686 | 0.573 | 0.610 | 0.664
$-30^{\circ}$ | 0.847 | 1.000 | 0.911 | 0.847 | 0.807 | 0.766 | 0.635
$-15^{\circ}$ | 0.754 | 0.911 | 1.000 | 0.925 | 0.896 | 0.821 | 0.602
$0^{\circ}$ | 0.686 | 0.847 | 0.925 | 1.000 | 0.964 | 0.872 | 0.684
$15^{\circ}$ | 0.573 | 0.807 | 0.896 | 0.964 | 1.000 | 0.929 | 0.768
$30^{\circ}$ | 0.610 | 0.766 | 0.821 | 0.872 | 0.929 | 1.000 | 0.878
$45^{\circ}$ | 0.664 | 0.635 | 0.602 | 0.684 | 0.768 | 0.878 | 1.000
(d) FDA, $Ave=0.814$
| $-45^{\circ}$ | $-30^{\circ}$ | $-15^{\circ}$ | $0^{\circ}$ | $15^{\circ}$ | $30^{\circ}$ | $45^{\circ}$
---|---|---|---|---|---|---|---
$-45^{\circ}$ | 1.000 | 0.854 | 0.714 | 0.595 | 0.557 | 0.633 | 0.608
$-30^{\circ}$ | 0.854 | 1.000 | 0.867 | 0.746 | 0.688 | 0.697 | 0.606
$-15^{\circ}$ | 0.714 | 0.867 | 1.000 | 0.887 | 0.808 | 0.704 | 0.579
$0^{\circ}$ | 0.595 | 0.746 | 0.887 | 1.000 | 0.916 | 0.819 | 0.651
$15^{\circ}$ | 0.557 | 0.688 | 0.808 | 0.916 | 1.000 | 0.912 | 0.754
$30^{\circ}$ | 0.633 | 0.697 | 0.704 | 0.819 | 0.912 | 1.000 | 0.850
$45^{\circ}$ | 0.608 | 0.606 | 0.579 | 0.651 | 0.754 | 0.850 | 1.000
(e) CDFE, $Ave=0.773$
| $-45^{\circ}$ | $-30^{\circ}$ | $-15^{\circ}$ | $0^{\circ}$ | $15^{\circ}$ | $30^{\circ}$ | $45^{\circ}$
---|---|---|---|---|---|---|---
$-45^{\circ}$ | 1.000 | 0.914 | 0.854 | 0.763 | 0.710 | 0.770 | 0.759
$-30^{\circ}$ | 0.914 | 1.000 | 0.947 | 0.858 | 0.812 | 0.861 | 0.766
$-15^{\circ}$ | 0.854 | 0.947 | 1.000 | 0.923 | 0.880 | 0.894 | 0.775
$0^{\circ}$ | 0.763 | 0.858 | 0.923 | 1.000 | 0.938 | 0.900 | 0.750
$15^{\circ}$ | 0.710 | 0.812 | 0.880 | 0.938 | 1.000 | 0.923 | 0.807
$30^{\circ}$ | 0.770 | 0.861 | 0.894 | 0.900 | 0.923 | 1.000 | 0.934
$45^{\circ}$ | 0.759 | 0.766 | 0.775 | 0.750 | 0.807 | 0.934 | 1.000
(f) MvDA, $Ave=0.867$
| $-45^{\circ}$ | $-30^{\circ}$ | $-15^{\circ}$ | $0^{\circ}$ | $15^{\circ}$ | $30^{\circ}$ | $45^{\circ}$
---|---|---|---|---|---|---|---
$-45^{\circ}$ | 1.000 | 0.872 | 0.819 | 0.730 | 0.655 | 0.708 | 0.686
$-30^{\circ}$ | 0.856 | 1.000 | 0.881 | 0.825 | 0.754 | 0.737 | 0.650
$-15^{\circ}$ | 0.807 | 0.874 | 1.000 | 0.869 | 0.865 | 0.781 | 0.681
$0^{\circ}$ | 0.757 | 0.854 | 0.896 | 1.000 | 0.938 | 0.858 | 0.790
$15^{\circ}$ | 0.688 | 0.777 | 0.854 | 0.916 | 1.000 | 0.900 | 0.823
$30^{\circ}$ | 0.708 | 0.735 | 0.788 | 0.834 | 0.918 | 1.000 | 0.916
$45^{\circ}$ | 0.719 | 0.715 | 0.697 | 0.752 | 0.832 | 0.909 | 1.000
(g) DCAN-1, $Ave=0.830$
| $-45^{\circ}$ | $-30^{\circ}$ | $-15^{\circ}$ | $0^{\circ}$ | $15^{\circ}$ | $30^{\circ}$ | $45^{\circ}$
---|---|---|---|---|---|---|---
$-45^{\circ}$ | 1.000 | 0.905 | 0.876 | 0.783 | 0.714 | 0.779 | 0.796
$-30^{\circ}$ | 0.927 | 1.000 | 0.954 | 0.896 | 0.850 | 0.825 | 0.730
$-15^{\circ}$ | 0.867 | 0.929 | 1.000 | 0.905 | 0.905 | 0.867 | 0.757
$0^{\circ}$ | 0.832 | 0.876 | 0.925 | 1.000 | 0.958 | 0.896 | 0.808
$15^{\circ}$ | 0.765 | 0.865 | 0.907 | 0.951 | 1.000 | 0.929 | 0.874
$30^{\circ}$ | 0.779 | 0.832 | 0.870 | 0.916 | 0.945 | 1.000 | 0.949
$45^{\circ}$ | 0.794 | 0.777 | 0.785 | 0.812 | 0.876 | 0.938 | 1.000
(h) DCAN-3, $Ave=0.884$
Table 2: Results of CCA, FDA [4], CDFE [22], MvDA [15] and DCAN on MultiPIE
dataset in terms of rank-1 recognition rate. DCAN-k means a stacked k-layer
network. Due to space limitation, the results of other methods cannot be
reported here, but their mean accuracies are shown in Table 1.
### 3.3 Face Recognition across Pose
First, to explicitly illustrate the learnt mapping, we conduct an experiment
on Multi-PIE dataset by projecting the learnt common features into a 2-D space
with Principal Component Analysis (PCA). As shown in Fig.2. The classical
method CCA can only roughly align the data in the principal directions and the
state-of-the-art method MvDA [15] attempts to merge two types of data but
seems to fail. Thus, we argue that linear transforms are a little stiff to
convert data from two views into an ideal common space. The three diagrams
below shows that DCAN can gradually separate samples from different classes
with the increase of layers, which is just as we described in the above
analysis.
Figure 2: After learning common features by the cross-view methods, we project
the features into 2-D space by using the principal two components in PCA. The
depicted samples are randomly chosen form Multi-PIE [9] dataset. The “$\circ$”
and “$+$” points come from two views respectively. Different color points
belong to different classes. DCAN-k is our proposed method with a stacked
k-layer neural network.
Next, we compare our methods with several state-of-the-art methods for the
cross-view face recognition task on Multi-PIE data set. Since the images are
acquired over seven poses on Multi-PIE data set, in total $7\times 6=42$
comparison experiments need to be conducted. The detailed results are shown in
Table 2,where two poses are used as the gallery and probe set to each other
and the rank-1 recognition rate is reported. Further, the mean accuracy of all
pairwise results for each methods is also reported in Table 1.
From Table 1, we can find the supervised methods except CSR are significantly
superior to CCA due to the use of the label information. And nonlinear methods
except Deep CCA are significantly superior to the nonlinear methods due to the
use of nonlinear transforms. Compared with FDA, the proposed DCAN with only
one layer network can perform better with 1.6% improvement. With increasing
layers, the accuracy of DCAN reaches a climax via stacking three layer
networks. The reason of the degradation in DCAN with four layers is mainly the
effect of reduced dimensionality, where 10 dimensions are cut out from the
above layer network. Obviously, compared with two-view based methods, the
proposed DCAN with three layers improves the performance greatly (88.4% vs.
81.4%). Besides, MvDA also achieves a considerably good performance by using
all samples from all poses. It is unfair to compare these two-view based
methods (containing DCAN) with MvDA, because the latter implicitly uses
additional five views information except current compared two views. But our
method performs better than MvDA, 88.4% vs. 86.7%. As observed in Table 2,
three-layer DCAN achieves a largely improvement compared with CCA,FDA,CDFE for
all cross-view cases and MvDA for most of cross-view cases. The results are
shown in Table 2 and Table 1.
### 3.4 Photo-Sketch Recognition
Method | Photo-Sketch | Sketch-Photo
---|---|---
CCA[14] | 0.387 | 0.475
KernelCCA[10] | 0.466 | 0.570
DeepCCA[3] | 0.364 | 0.434
CDFE[22] | 0.456 | 0.476
CSR[20] | 0.502 | 0.590
PLS[26] | 0.486 | 0.510
FDA[4] | 0.468 | 0.534
MvDA[15] | 0.534 | 0.555
DCAN-1 | 0.535 | 0.555
DCAN-2 | 0.603 | 0.613
DCAN-3 | 0.601 | 0.652
Table 3: Evluation on CUFSF database in terms of mean accuracy. DCAN-k means a
stacked k-layer network.
(a)
(b)
Figure 3: The performance with varied parameter values for our proposed DCAN.
The sketch and photo images in CUFSF [34, 31] are respectively used for the
gallery and probe set. (a) Varied $\lambda$ with fixed $k=10$. (b) Varied $k$
with fixed $\lambda=0.2$.
Photo-Sketch recognition is conducted on CUFSF dataset. The samples come from
only two views, photo and sketch. The comparison results are provided in Table
3. As shown in this table, since only two views can be utilized in this case,
MvDA degrades to a comparable performance with those previous two-view based
methods. Our proposed DCAN with three layer networks can achieve even better
with more than 6% improvement, which further indicates DCAN benefits from the
nonlinear and multi-layer structure.
Discussion and analysis: The above experiments demonstrate that our methods
can work very well even on a small sample size. The reasons lie in three
folds:
1. (1)
The maximum margin criterion makes the learnt mapping more discriminative,
which is a straightforward strategy in the supervised classification task.
2. (2)
Auto-encoder approximately preserves the local neighborhood structures.
For this, Alain et al. [2] theoretically prove that the learnt representation
by auto-encoder can recover local properties from the view of manifold. To
further validate that, we employ the first 700 photo images from CUFSF
database to perform the nonlinear self-reconstruction with auto-encoder. With
the hidden presentations, we find the local neighbors with 1,2,3,4,5 neighbors
can be preserved with the probability of 99.43%, 99.00%, 98.57%, 98.00% and
97.42% respectively. Thus, the use of auto-encoder intrinsically reduces the
complexity of the discriminant model, which further makes the learnt model
better generality on the testing set.
3. (3)
The deep structure generates a gradual model, which makes the learnt transform
more robust. With only one layer, the model can’t represent the complex data
very well. But with layers goes deeper, the coupled networks can learn
transforms much more flexible and hence can be allowed to handle more complex
data.
## 4 Conclusion
In this paper, we propose a deep learning method, the Deeply Coupled Auto-
encoder Networks(DCAN), which can gradually generate a coupled discriminant
common representation for cross-view object classification. In each layer we
take both local consistency and discrimination of projected data into
consideration. By stacking multiple such coupled network layers, DCAN can
gradually improve the learnt shared features in the common space. Moreover,
experiments in the cross-view classification tasks demonstrate the superior of
our method over other state-of-the-art methods.
## References
* [1] S. Akaho. A kernel method for canonical correlation analysis, 2006.
* [2] G. Alain and Y. Bengio. What regularized auto-encoders learn from the data generating distribution. arXiv preprint arXiv:1211.4246, 2012.
* [3] G. Andrew, R. Arora, J. Bilmes, and K. Livescu. Deep canonical correlation analysis.
* [4] P. N. Belhumeur, J. P. Hespanha, and D. J. Kriegman. Eigenfaces vs. fisherfaces: Recognition using class specific linear projection. IEEE Transactions on Pattern Analysis and Machine Intelligence, 19(7):711–720, 1997.
* [5] Y. Bengio, A. Courville, and P. Vincent. Representation learning: A review and new perspectives. 2013\.
* [6] Y. Bengio, P. Lamblin, D. Popovici, and H. Larochelle. Greedy layer-wise training of deep networks, 2007.
* [7] M. Chen, Z. Xu, K. Weinberger, and F. Sha. Marginalized denoising autoencoders for domain adaptation, 2012.
* [8] N. Chen, J. Zhu, and E. P. Xing. Predictive subspace learning for multi-view data: a large margin approach, 2010.
* [9] R. Gross, I. Matthews, J. Cohn, T. Kanade, and S. Baker. The cmu multi-pose, illumination, and expression (multi-pie) face database, 2007.
* [10] D. R. Hardoon, S. Szedmak, and J. Shawe-Taylor. Canonical correlation analysis: An overview with application to learning methods. Neural Computation, 16(12):2639–2664, 2004.
* [11] T. Hastie, R. Tibshirani, and J. J. H. Friedman. The elements of statistical learning, 2001.
* [12] G. E. Hinton, S. Osindero, and Y.-W. Teh. A fast learning algorithm for deep belief nets. Neural computation, 18(7):1527–1554, 2006.
* [13] G. E. Hinton and R. R. Salakhutdinov. Reducing the dimensionality of data with neural networks. Science, 313(5786):504–507, 2006.
* [14] H. Hotelling. Relations between two sets of variates. Biometrika, 28(3/4):321–377, 1936.
* [15] M. Kan, S. Shan, H. Zhang, S. Lao, and X. Chen. Multi-view discriminant analysis. pages 808–821, 2012.
* [16] T.-K. Kim, J. Kittler, and R. Cipolla. Discriminative learning and recognition of image set classes using canonical correlations. IEEE Transactions on Pattern Analysis and Machine Intelligence, 29(6):1005–1018, 2007.
* [17] Q. V. Le, J. Ngiam, A. Coates, A. Lahiri, B. Prochnow, and A. Y. Ng. On optimization methods for deep learning, 2011.
* [18] Y. LeCun, L. Bottou, G. B. Orr, and K.-R. Müller. Efficient backprop. In Neural networks: Tricks of the trade, pages 9–50. Springer, 1998\.
* [19] H. Lee, A. Battle, R. Raina, and A. Y. Ng. Efficient sparse coding algorithms, 2007.
* [20] Z. Lei and S. Z. Li. Coupled spectral regression for matching heterogeneous faces, 2009.
* [21] H. Li, T. Jiang, and K. Zhang. Efficient and robust feature extraction by maximum margin criterion. Neural Networks, IEEE Transactions on, 17(1):157–165, 2006.
* [22] D. Lin and X. Tang. Inter-modality face recognition. pages 13–26, 2006.
* [23] J. Ngiam, A. Khosla, M. Kim, J. Nam, H. Lee, and A. Y. Ng. Multimodal deep learning, 2011.
* [24] J. Nocedal and S. J. Wright. Numerical optimization, 2006.
* [25] P. J. Phillips, H. Wechsler, J. Huang, and P. J. Rauss. The feret database and evaluation procedure for face-recognition algorithms. Image and vision computing, 16(5):295–306, 1998.
* [26] A. Sharma and D. W. Jacobs. Bypassing synthesis: Pls for face recognition with pose, low-resolution and sketch, 2011.
* [27] A. Sharma, A. Kumar, H. Daume, and D. W. Jacobs. Generalized multiview analysis: A discriminative latent space, 2012.
* [28] P. Vincent, H. Larochelle, Y. Bengio, and P.-A. Manzagol. Extracting and composing robust features with denoising autoencoders, 2008\.
* [29] F. Wang and C. Zhang. Feature extraction by maximizing the average neighborhood margin. In Computer Vision and Pattern Recognition, 2007. CVPR’07. IEEE Conference on, pages 1–8. IEEE, 2007.
* [30] S. Wang, L. Zhang, Y. Liang, and Q. Pan. Semi-coupled dictionary learning with applications to image super-resolution and photo-sketch synthesis, 2012.
* [31] X. Wang and X. Tang. Face photo-sketch synthesis and recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence, 31(11):1955–1967, 2009.
* [32] J. Xie, L. Xu, and E. Chen. Image denoising and inpainting with deep neural networks, 2012.
* [33] S. Yan, D. Xu, B. Zhang, H.-J. Zhang, Q. Yang, and S. Lin. Graph embedding and extensions: a general framework for dimensionality reduction. IEEE Transactions on Pattern Analysis and Machine Intelligence, 29(1):40–51, 2007.
* [34] W. Zhang, X. Wang, and X. Tang. Coupled information-theoretic encoding for face photo-sketch recognition, 2011.
* [35] B. Zhao, F. Wang, and C. Zhang. Maximum margin embedding. In Data Mining, 2008. ICDM’08. Eighth IEEE International Conference on, pages 1127–1132. IEEE, 2008.
|
$\displaystyle=\mathbf{Q}_{n,ww}^{-1}\mathbf{q}_{n,wy}-\mathrm{E}\left(\mathbf{Q}_{n,ww}\right)^{-1}\mathrm{E}\left(\mathbf{q}_{n,wy}\right)$
$\displaystyle=\left[\mathbf{Q}_{n,ww}^{-1}-E\left(\mathbf{Q}_{n,ww}\right)^{-1}+E\left(\mathbf{Q}_{n,ww}\right)^{-1}\right]\left[\mathbf{q}_{n,wy}-\mathrm{E}\left(\mathbf{q}_{n,wy}\right)+\mathrm{E}\left(\mathbf{q}_{n,wy}\right)\right]-\mathrm{E}\left(\mathbf{Q}_{n,ww}\right)^{-1}\mathrm{E}\left(\mathbf{q}_{n,wy}\right)$
$\displaystyle=\left[\mathbf{Q}_{n,ww}^{-1}-E\left(\mathbf{Q}_{n,ww}\right)^{-1}\right]\left[\mathbf{q}_{n,wy}-\mathrm{E}\left(\mathbf{q}_{n,wy}\right)\right]+\left[\mathbf{Q}_{n,ww}^{-1}-\mathrm{E}\left(\mathbf{Q}_{n,ww}\right)^{-1}\right]\mathrm{E}\left(\mathbf{q}_{n,wy}\right)$
$\displaystyle\quad\quad\quad+\mathrm{E}\left(\mathbf{Q}_{n,ww}\right)^{-1}\left[\mathbf{q}_{n,wy}-\mathrm{E}\left(\mathbf{q}_{n,wy}\right)\right].$
Then,
$\displaystyle\left\|\hat{\mathbf{\phi}}-\mathbf{\phi}\right\|$
$\displaystyle\leq\left\|\mathbf{Q}_{n,ww}^{-1}-\mathrm{E}\left(\mathbf{Q}_{n,ww}\right)^{-1}\right\|\left\|\mathbf{q}_{n,wy}-\mathrm{E}\left(\mathbf{q}_{n,wy}\right)\right\|+\left\|\mathbf{Q}_{n,ww}^{-1}-\mathrm{E}\left(\mathbf{Q}_{n,ww}\right)^{-1}\right\|\left\|\mathrm{E}\left(\mathbf{q}_{n,wy}\right)\right\|$
$\displaystyle\quad\quad+\left\|\mathrm{E}\left(\mathbf{Q}_{n,ww}\right)^{-1}\right\|\left\|\mathbf{q}_{n,wy}-\mathrm{E}\left(\mathbf{q}_{n,wy}\right)\right\|.$
By Assumption 1(c), we have
$\left\|\mathbf{Q}_{n,ww}^{-1}-\mathrm{E}\left(\mathbf{Q}_{n,ww}\right)^{-1}\right\|=O_{p}\left(n^{-1/2}\right)$,
$\left\|\mathbf{q}_{n,wy}-\mathrm{E}\left(\mathbf{q}_{n,wy}\right)\right\|=O_{p}\left(n^{-1/2}\right)$,
and by Assumption 1(b),
$\left\|\mathrm{E}\left(\mathbf{q}_{n,wy}\right)\right\|$ and
$\left\|\mathrm{E}\left(\mathbf{Q}_{n,ww}\right)^{-1}\right\|$ are bounded.
Thus,
$\left\|\hat{\mathbf{\phi}}-\mathbf{\phi}\right\|=O_{p}\left(n^{-1/2}\right).$
(S.2.1)
To establish the asymptotic distribution of $\hat{\mathbf{\phi}}$, we first
note that
$\sqrt{n}\left(\hat{\mathbf{\phi}}-\mathbf{\phi}\right)=\mathbf{Q}_{n,ww}^{-1}\left(n^{-1/2}\sum_{i=1}^{n}\mathbf{w}_{i}\xi_{i}\right).$
By Assumption 3, we have
$\mathrm{var}\left(n^{-1/2}\sum_{i=1}^{n}\mathbf{w}_{i}\xi_{i}\right)=\frac{1}{n}\sum_{i=1}^{n}\mathrm{var}\left(\mathbf{w}_{i}\xi_{i}\right)=\frac{1}{n}\sum_{i=1}^{n}\mathrm{E}\left(\mathbf{w}_{i}\mathbf{w}_{i}^{\prime}\xi_{i}^{2}\right)\rightarrow\mathbf{V}_{w\xi}\succ
0.$
Note that $\xi_{i}=u_{i}+x_{i}v_{i}$, and $\mathbf{w}_{i}$ is distributed
independently of $u_{i}$ and $v_{i}$. Then
$\mathbf{w}_{i}\xi_{i}=\mathbf{w}_{i}\left(u_{i}+x_{i}v_{i}\right)=\mathbf{w}_{i}u_{i}+\left(\mathbf{w}_{i}x_{i}\right)v_{i},$
and by Minkowski’s inequality, for $r=2+\delta$ with $0<\delta<1$,
$\left[E\left\|\mathbf{w}_{i}\xi_{i}\right\|^{r}\right]^{1/r}\leq\left[E\left\|\mathbf{w}_{i}u_{i}\right\|^{r}\right]^{1/r}+\left[E\left\|\left(\mathbf{w}_{i}x_{i}\right)v_{i}\right\|^{r}\right]^{1/r}.$
Due to the independence of $u_{i}$ and $v_{i}$ from $\mathbf{w}_{i}$, we have
$\mathrm{E}\left(\left\|\mathbf{w}_{i}u_{i}\right\|^{r}\right)\leq
E\left\|\mathbf{w}_{i}\right\|^{r}E\left\|u_{i}\right\|^{r},\text{ and
}E\left\|\left(\mathbf{w}_{i}x_{i}^{\prime}\right)v_{i}\right\|^{r}\leq
E\left\|\mathbf{w}_{i}x_{i}\right\|^{r}E\left\|v_{i}\right\|^{r}.$
Also, $E\left\|\mathbf{w}_{i}x_{i}\right\|^{r}\leq
E\left\|\left(x_{i}^{2},x_{i}\mathbf{z}_{i}^{\prime}\right)^{\prime}\right\|^{r}\leq
E\left\|\mathbf{w}_{i}\mathbf{w}_{i}^{\prime}\right\|^{r}\leq
E\left\|\mathbf{w}_{i}\right\|^{2r}$, where $2<r<3$, and hence $2r<6$. By
Assumptions 1(a.ii) and 1(b.ii), we have
$\sup_{i}\mathrm{E}\left(\left\|\mathbf{w}_{i}\right\|^{6}\right)<C$,
$\sup_{i}\mathrm{E}\left(\left\|u_{i}\right\|^{3}\right)<C$, and
$\mathrm{E}\left(\left\|v_{i}\right\|^{3}\right)\leq\max_{1\leq k\leq
K}\left|b_{k}-\mathrm{E}\left(\beta_{i}\right)\right|^{3}<C.$ Then, we
verified that
$\sup_{i}\mathrm{E}\left(\left\|\mathbf{w}_{i}u_{i}\right\|^{r}\right)<C$, and
$E\left\|\left(\mathbf{w}_{i}x_{i}^{\prime}\right)v_{i}\right\|^{r}<C$, and
hence the Lyapunov condition that
$\sup_{i}\mathrm{E}\left(\left\|\mathbf{w}_{i}\xi_{i}\right\|^{r}\right)<C$,
where $r=2+\delta\in(2,3)$. By the central limit theorem for independent but
not necessarily identically distributed random vectors (see Pesaran (2015,
Theorem 18) or Hansen (2022, Theorem 6.5)), we have
$\frac{1}{\sqrt{n}}\sum_{i=1}^{n}\mathbf{w}_{i}\xi_{i}\rightarrow_{d}N(\mathbf{0},\mathbf{V}_{w\xi}),$
as $n\rightarrow\infty$, and by Assumption 1 and continuous mapping theorem,
$\sqrt{n}(\hat{\mathbf{\phi}}-\mathbf{\phi})\rightarrow_{d}N\left(\mathbf{0},\mathbf{Q}_{ww}^{-1}\mathbf{V}_{w\xi}\mathbf{Q}_{ww}^{-1}\right).$
We then turn to the consistent estimation of the variance matrix. By
Assumption 3, we have
$\displaystyle\left\|\hat{\mathbf{V}}_{w\xi}-\mathbf{V}_{w\xi}\right\|$
$\displaystyle=\left\|\frac{1}{n}\sum_{i=1}^{n}\mathbf{w}_{i}\mathbf{w}_{i}^{\prime}\hat{\xi}_{i}^{2}-\frac{1}{n}\sum_{i=1}^{n}\mathrm{E}\left(\mathbf{w}_{i}\mathbf{w}_{i}\xi_{i}^{2}\right)+\frac{1}{n}\sum_{i=1}^{n}\mathrm{E}\left(\mathbf{w}_{i}\mathbf{w}_{i}\xi_{i}^{2}\right)-\mathbf{V}_{w\xi}\right\|$
$\displaystyle\leq\left\|\frac{1}{n}\sum_{i=1}^{n}\mathbf{w}_{i}\mathbf{w}_{i}^{\prime}\xi_{i}^{2}-\frac{1}{n}\sum_{i=1}^{n}\mathrm{E}\left(\mathbf{w}_{i}\mathbf{w}_{i}\xi_{i}^{2}\right)\right\|+\left\|\frac{1}{n}\sum_{i=1}^{n}\mathrm{E}\left(\mathbf{w}_{i}\mathbf{w}_{i}\xi_{i}^{2}\right)-\mathbf{V}_{w\xi}\right\|$
$\displaystyle\quad\quad+\left\|\frac{1}{n}\sum_{i=1}^{n}\mathbf{w}_{i}\mathbf{w}_{i}^{\prime}\left(\hat{\xi}_{i}^{2}-\xi_{i}^{2}\right)\right\|$
$\displaystyle\leq\frac{1}{n}\sum_{i=1}^{n}\left\|\mathbf{w}_{i}\right\|^{2}\left|\hat{\xi}_{i}^{2}-\xi_{i}^{2}\right|+O_{p}(n^{-1/2}).$
(S.2.2)
Note that
$\hat{\xi}_{i}=\xi_{i}-\left(\hat{\mathbf{\phi}}-\mathbf{\phi}\right)^{\prime}\mathbf{w}_{i}$,
then
$\displaystyle\left|\hat{\xi}_{i}^{2}-\xi_{i}^{2}\right|$ $\displaystyle\leq
2\left|\xi_{i}\mathbf{w}_{i}^{\prime}\left(\hat{\mathbf{\phi}}-\mathbf{\phi}\right)\right|+\left(\hat{\mathbf{\phi}}-\mathbf{\phi}\right)^{\prime}\left(\mathbf{w}_{i}\mathbf{w}_{i}^{\prime}\right)\left(\hat{\mathbf{\phi}}-\mathbf{\phi}\right)$
$\displaystyle\leq
2\left|\xi_{i}\right|\left\|\mathbf{w}_{i}\right\|\left\|\hat{\mathbf{\phi}}-\mathbf{\phi}\right\|+\left\|\mathbf{w}_{i}\right\|^{2}\left\|\hat{\mathbf{\phi}}-\mathbf{\phi}\right\|^{2}.$
(S.2.3)
Combine (S.2.2) and (S.2.3), we have
$\left\|\hat{\mathbf{V}}_{w\xi}-\mathbf{V}_{w\xi}\right\|\leq
2\left(\frac{1}{n}\sum_{i=1}^{n}\left\|\mathbf{w}_{i}\right\|^{3}\left|\xi_{i}\right|\right)\left\|\hat{\mathbf{\phi}}-\mathbf{\phi}\right\|+\left(\frac{1}{n}\sum_{i=1}^{n}\left\|\mathbf{w}_{i}\right\|^{4}\right)\left\|\hat{\mathbf{\phi}}-\mathbf{\phi}\right\|^{2}+O_{p}\left(n^{-1/2}\right).$
(S.2.4)
By Hölder’s inequality,
$\frac{1}{n}\sum_{i=1}^{n}\left\|\mathbf{w}_{i}\right\|^{3}\left|\xi_{i}\right|\leq\left(\frac{1}{n}\sum_{i=1}^{n}\left\|\mathbf{w}_{i}\right\|^{4}\right)^{3/4}\left(\frac{1}{n}\sum_{i=1}^{n}\xi_{i}^{4}\right)^{1/4}.$
(S.2.5)
By Assumption 1(b.iii),
$n^{-1}\sum_{i=1}^{n}\left\|\mathbf{w}_{i}\right\|^{4}=O_{p}(1)$. By Minkowski
inequality,
$\displaystyle\left(\frac{1}{n}\sum_{i=1}^{n}\xi_{i}^{4}\right)^{1/4}$
$\displaystyle=\left(\frac{1}{n}\sum_{i=1}^{n}\left(u_{i}+x_{i}v_{i}\right)^{4}\right)^{1/4}\leq\left(\frac{1}{n}\sum_{i=1}^{n}u_{i}^{4}\right)^{1/4}+\left(\frac{1}{n}\sum_{i=1}^{n}x_{i}^{4}v_{i}^{4}\right)^{1/4}$
$\displaystyle\leq\left(\frac{1}{n}\sum_{i=1}^{n}u_{i}^{4}\right)^{1/4}+\max_{k}\left\\{\left|b_{k}-\mathrm{E}\left(\beta_{i}\right)\right|\right\\}\left(\frac{1}{n}\sum_{i=1}^{n}x_{i}^{4}\right)^{1/4}$
$\displaystyle=O_{p}(1),$
where the last inequality is from Assumptions 1(a.iii) and (b.iii) that
$n^{-1}\sum_{i=1}^{n}u_{i}^{4}=O_{p}\left(1\right)$, and
$n^{-1}\sum_{i=1}^{n}x_{i}^{4}\leq
n^{-1}\sum_{i=1}^{n}\left\|\mathbf{w}_{i}\right\|^{4}=O_{p}(1)$. Then we
verified in (S.2.5) that
$n^{-1}\sum_{i=1}^{n}\left\|\mathbf{w}_{i}\right\|^{3}\left|\xi_{i}\right|=O_{p}(1)$.
Then using the above results in (S.2.4), and noting from (S.2.1) that
$\left\|\hat{\mathbf{\phi}}-\mathbf{\phi}\right\|=O_{p}\left(n^{-1/2}\right)$,
we have
$\left\|\hat{\mathbf{V}}_{w\xi}-\mathbf{V}_{w\xi}\right\|=O_{p}\left(n^{-1/2}\right)$,
as required.
## Appendix S.3 Monte Carlo Simulation
### S.3.1 Results with $S=5$ and $S=6$
Tables S.1 and S.2 present the summary results corresponding to $S=5$ and
$S=6$, for the data generating processes described in Section 5.1. These
results show that adding more moments does not necessarily improve the
estimation accuracy but could be counter-productive.
Table S.1: Bias, RMSE and size of the GMM estimator for distributional
parameters of $\beta$ with $S=5$
DGP | Baseline | Categorical $x$ | Categorical $u$
---|---|---|---
Sample size $n$ | Bias | RMSE | Size | Bias | RMSE | Size | Bias | RMSE | Size
high variance: $\mathrm{var}\left(\beta_{i}\right)=0.25$
$\pi=0.5$ | 100 | 0.0308 | 0.1869 | 0.1021 | 0.0259 | 0.1986 | 0.1276 | 0.0106 | 0.1944 | 0.1050
1,000 | 0.0048 | 0.1235 | 0.1950 | 0.0054 | 0.1334 | 0.2112 | -0.0364 | 0.1638 | 0.2239
2,000 | -0.0006 | 0.0875 | 0.1641 | -0.0009 | 0.0962 | 0.1887 | -0.0238 | 0.1172 | 0.2059
5,000 | -0.0005 | 0.0484 | 0.1339 | -0.0001 | 0.0591 | 0.1602 | -0.0125 | 0.0740 | 0.1667
10,000 | -0.0002 | 0.0334 | 0.1152 | -0.0005 | 0.0373 | 0.1246 | -0.0080 | 0.0519 | 0.1386
100,000 | -0.0002 | 0.0096 | 0.0636 | 0.0001 | 0.0116 | 0.0738 | -0.0008 | 0.0174 | 0.0766
$\beta_{L}=1$ | 100 | 0.2234 | 0.4541 | 0.3205 | 0.1992 | 0.4777 | 0.2843 | 0.1780 | 0.5090 | 0.2519
1,000 | 0.0503 | 0.1609 | 0.3060 | 0.0475 | 0.1812 | 0.2963 | 0.0100 | 0.2024 | 0.2141
2,000 | 0.0265 | 0.1148 | 0.2501 | 0.0257 | 0.1262 | 0.2501 | 0.0088 | 0.1337 | 0.1905
5,000 | 0.0108 | 0.0606 | 0.1926 | 0.0130 | 0.0702 | 0.2042 | 0.0031 | 0.0803 | 0.1641
10,000 | 0.0054 | 0.0409 | 0.1408 | 0.0061 | 0.0456 | 0.1510 | 0.0008 | 0.0527 | 0.1338
100,000 | 0.0004 | 0.0114 | 0.0716 | 0.0006 | 0.0134 | 0.0790 | 0.0002 | 0.0184 | 0.0834
$\beta_{H}=2$ | 100 | -0.1956 | 0.5486 | 0.2448 | -0.1941 | 0.5638 | 0.2386 | -0.2029 | 0.5801 | 0.2269
1,000 | -0.0418 | 0.2080 | 0.3299 | -0.0414 | 0.2300 | 0.3384 | -0.0752 | 0.2583 | 0.3620
2,000 | -0.0264 | 0.1379 | 0.2799 | -0.0286 | 0.1554 | 0.2860 | -0.0529 | 0.1789 | 0.3048
5,000 | -0.0113 | 0.0696 | 0.2008 | -0.0116 | 0.0883 | 0.2170 | -0.0254 | 0.1038 | 0.2411
10,000 | -0.0053 | 0.0432 | 0.1502 | -0.0064 | 0.0520 | 0.1642 | -0.0156 | 0.0690 | 0.2002
100,000 | -0.0007 | 0.0113 | 0.0662 | -0.0004 | 0.0135 | 0.0764 | -0.0016 | 0.0209 | 0.0818
low variance: $\mathrm{var}\left(\beta_{i}\right)=0.15$
$\pi=0.3$ | 100 | 0.2214 | 0.2820 | 0.1063 | 0.2291 | 0.2942 | 0.1328 | 0.2212 | 0.2876 | 0.1221
1,000 | 0.0477 | 0.1746 | 0.2235 | 0.0605 | 0.1928 | 0.2430 | 0.0348 | 0.2039 | 0.2900
2,000 | 0.0217 | 0.1198 | 0.2020 | 0.0262 | 0.1331 | 0.2246 | -0.0080 | 0.1608 | 0.2822
5,000 | 0.0112 | 0.0709 | 0.1732 | 0.0154 | 0.0828 | 0.1956 | -0.0115 | 0.1072 | 0.2289
10,000 | 0.0063 | 0.0465 | 0.1588 | 0.0106 | 0.0576 | 0.1649 | -0.0075 | 0.0761 | 0.1890
100,000 | 0.0001 | 0.0130 | 0.0810 | 0.0014 | 0.0158 | 0.0882 | 0.0040 | 0.0280 | 0.0978
$\beta_{L}=0.5$ | 100 | 0.4245 | 0.5722 | 0.2938 | 0.4048 | 0.5818 | 0.2612 | 0.3827 | 0.6052 | 0.2278
1,000 | 0.1300 | 0.2692 | 0.3058 | 0.1300 | 0.2890 | 0.3057 | 0.0882 | 0.3673 | 0.1970
2,000 | 0.0763 | 0.1746 | 0.3147 | 0.0735 | 0.1903 | 0.2820 | 0.0149 | 0.2523 | 0.1964
5,000 | 0.0378 | 0.1018 | 0.2690 | 0.0410 | 0.1155 | 0.2695 | 0.0034 | 0.1417 | 0.1905
10,000 | 0.0202 | 0.0674 | 0.2344 | 0.0257 | 0.0822 | 0.2404 | 0.0013 | 0.0961 | 0.1690
100,000 | 0.0013 | 0.0184 | 0.0952 | 0.0026 | 0.0221 | 0.1042 | 0.0060 | 0.0347 | 0.1112
$\beta_{H}=1.345$ | 100 | -0.0646 | 0.3773 | 0.1781 | -0.0616 | 0.4058 | 0.1668 | -0.0564 | 0.4357 | 0.1688
1,000 | -0.0180 | 0.1523 | 0.2496 | -0.0119 | 0.1804 | 0.2615 | -0.0476 | 0.2022 | 0.2721
2,000 | -0.0104 | 0.1021 | 0.2375 | -0.0101 | 0.1147 | 0.2414 | -0.0381 | 0.1448 | 0.2830
5,000 | -0.0027 | 0.0549 | 0.1680 | -0.0016 | 0.0680 | 0.1936 | -0.0193 | 0.0927 | 0.2369
10,000 | -0.0001 | 0.0368 | 0.1458 | 0.0007 | 0.0438 | 0.1458 | -0.0115 | 0.0634 | 0.1976
100,000 | -0.0002 | 0.0102 | 0.0726 | 0.0005 | 0.0120 | 0.0688 | 0.0021 | 0.0214 | 0.0902
Notes: The data generating process is (5.1). high variance and low variance
parametrization are described in (5.2). “Baseline”, “Categorical $x$” and
“Categorical $u$” refer to DGP 1 to 3 as in Section 5.1. Generically, bias,
RMSE and size are calculated by
$R^{-1}\sum_{r=1}^{R}\left(\hat{\theta}^{(r)}-\theta_{0}\right)$,
$\sqrt{R^{-1}\sum_{r=1}^{R}\left(\hat{\theta}^{(r)}-\theta_{0}\right)^{2}}$,
and
$R^{-1}\sum_{r=1}^{R}\mathbf{1}\left[\left|\hat{\theta}^{(r)}-\theta_{0}\right|/\hat{\sigma}_{\hat{\theta}}^{(r)}>\mathrm{cv}_{0.05}\right]$,
respectively, for true parameter $\theta_{0}$, its estimate
$\hat{\theta}^{(r)}$, the estimated standard error of $\hat{\theta}^{(r)}$,
$\hat{\sigma}_{\hat{\theta}}^{(r)}$, and the critical value
$\mathrm{cv}_{0.05}=\Phi^{-1}\left(0.975\right)$ across $R=5,000$
replications, where $\Phi\left(\cdot\right)$ is the cumulative distribution
function of standard normal distribution.
Table S.2: Bias, RMSE and size of the GMM estimator for distributional
parameters of $\beta$ with $S=6$
DGP | Baseline | Categorical $x$ | Categorical $u$
---|---|---|---
Sample size $n$ | Bias | RMSE | Size | Bias | RMSE | Size | Bias | RMSE | Size
high variance: $\mathrm{var}\left(\beta_{i}\right)=0.25$
$\pi=0.5$ | 100 | 0.0337 | 0.1472 | 0.0456 | 0.0293 | 0.1645 | 0.0695 | 0.0227 | 0.1498 | 0.0469
1,000 | 0.0021 | 0.1405 | 0.2545 | 0.0015 | 0.1469 | 0.2543 | -0.0265 | 0.1635 | 0.2551
2,000 | 0.0008 | 0.1071 | 0.2614 | 0.0006 | 0.1185 | 0.2789 | -0.0201 | 0.1281 | 0.2732
5,000 | -0.0020 | 0.0661 | 0.2261 | -0.0016 | 0.0765 | 0.2518 | -0.0142 | 0.0836 | 0.2510
10,000 | -0.0005 | 0.0444 | 0.1844 | -0.0011 | 0.0505 | 0.2155 | -0.0093 | 0.0587 | 0.2323
100,000 | 0.0000 | 0.0097 | 0.0732 | 0.0000 | 0.0118 | 0.0912 | -0.0020 | 0.0178 | 0.1162
$\beta_{L}=1$ | 100 | 0.2226 | 0.4373 | 0.3341 | 0.2151 | 0.4658 | 0.3237 | 0.1879 | 0.4841 | 0.2896
1,000 | 0.0721 | 0.2081 | 0.4485 | 0.0780 | 0.2197 | 0.4318 | 0.0531 | 0.2283 | 0.3576
2,000 | 0.0443 | 0.1464 | 0.4056 | 0.0455 | 0.1609 | 0.4157 | 0.0342 | 0.1536 | 0.3271
5,000 | 0.0175 | 0.0806 | 0.3035 | 0.0203 | 0.0923 | 0.3341 | 0.0150 | 0.0933 | 0.2770
10,000 | 0.0092 | 0.0510 | 0.2350 | 0.0098 | 0.0594 | 0.2723 | 0.0081 | 0.0629 | 0.2403
100,000 | 0.0010 | 0.0114 | 0.0850 | 0.0013 | 0.0136 | 0.0982 | 0.0002 | 0.0186 | 0.1116
$\beta_{H}=2$ | 100 | -0.2495 | 0.5629 | 0.2563 | -0.2580 | 0.5681 | 0.2608 | -0.2589 | 0.5782 | 0.2248
1,000 | -0.0618 | 0.2530 | 0.4938 | -0.0686 | 0.2733 | 0.4867 | -0.0962 | 0.2814 | 0.4874
2,000 | -0.0334 | 0.1729 | 0.4454 | -0.0365 | 0.1951 | 0.4461 | -0.0625 | 0.2017 | 0.4643
5,000 | -0.0189 | 0.1010 | 0.3457 | -0.0203 | 0.1178 | 0.3638 | -0.0383 | 0.1223 | 0.3946
10,000 | -0.0080 | 0.0634 | 0.2670 | -0.0109 | 0.0732 | 0.3011 | -0.0246 | 0.0830 | 0.3347
100,000 | -0.0013 | 0.0114 | 0.0842 | -0.0012 | 0.0141 | 0.1070 | -0.0043 | 0.0220 | 0.1396
low variance: $\mathrm{var}\left(\beta_{i}\right)=0.15$
$\pi=0.3$ | 100 | 0.2374 | 0.2757 | 0.0591 | 0.2352 | 0.2816 | 0.0829 | 0.2330 | 0.2771 | 0.0801
1,000 | 0.1071 | 0.2107 | 0.2608 | 0.1114 | 0.2244 | 0.2775 | 0.0764 | 0.2158 | 0.2772
2,000 | 0.0702 | 0.1661 | 0.2994 | 0.0786 | 0.1815 | 0.3258 | 0.0242 | 0.1806 | 0.3291
5,000 | 0.0452 | 0.1101 | 0.3217 | 0.0519 | 0.1260 | 0.3466 | 0.0092 | 0.1263 | 0.3329
10,000 | 0.0300 | 0.0816 | 0.3060 | 0.0390 | 0.0933 | 0.3389 | 0.0108 | 0.0954 | 0.3161
100,000 | 0.0018 | 0.0164 | 0.1128 | 0.0041 | 0.0234 | 0.1482 | 0.0055 | 0.0298 | 0.1688
$\beta_{L}=0.5$ | 100 | 0.4146 | 0.5479 | 0.3137 | 0.4191 | 0.5636 | 0.2965 | 0.3844 | 0.5678 | 0.2532
1,000 | 0.2445 | 0.3459 | 0.4601 | 0.2436 | 0.3579 | 0.4561 | 0.2080 | 0.3872 | 0.3187
2,000 | 0.1663 | 0.2539 | 0.4809 | 0.1684 | 0.2620 | 0.4797 | 0.1108 | 0.2830 | 0.3203
5,000 | 0.0977 | 0.1648 | 0.4800 | 0.1051 | 0.1788 | 0.4938 | 0.0590 | 0.1731 | 0.3606
10,000 | 0.0613 | 0.1182 | 0.4230 | 0.0730 | 0.1315 | 0.4717 | 0.0417 | 0.1251 | 0.3667
100,000 | 0.0050 | 0.0242 | 0.1420 | 0.0086 | 0.0333 | 0.1808 | 0.0101 | 0.0386 | 0.1906
$\beta_{H}=1.345$ | 100 | -0.0817 | 0.3703 | 0.1601 | -0.0883 | 0.3842 | 0.1687 | -0.0806 | 0.4136 | 0.1614
1,000 | -0.0086 | 0.1726 | 0.3174 | -0.0144 | 0.1907 | 0.3295 | -0.0560 | 0.2029 | 0.3239
2,000 | 0.0022 | 0.1194 | 0.3267 | 0.0029 | 0.1368 | 0.3401 | -0.0395 | 0.1582 | 0.3736
5,000 | 0.0093 | 0.0722 | 0.2899 | 0.0099 | 0.0876 | 0.3254 | -0.0189 | 0.0998 | 0.3570
10,000 | 0.0092 | 0.0535 | 0.2642 | 0.0117 | 0.0601 | 0.2889 | -0.0076 | 0.0733 | 0.3141
100,000 | -0.0002 | 0.0116 | 0.0972 | 0.0012 | 0.0157 | 0.1326 | 0.0019 | 0.0220 | 0.1454
Notes: The data generating process is (5.1). high variance and low variance
parametrization are described in (5.2). “Baseline”, “Categorical $x$” and
“Categorical $u$” refer to DGP 1 to 3 as in Section 5.1. Generically, bias,
RMSE and size are calculated by
$R^{-1}\sum_{r=1}^{R}\left(\hat{\theta}^{(r)}-\theta_{0}\right)$,
$\sqrt{R^{-1}\sum_{r=1}^{R}\left(\hat{\theta}^{(r)}-\theta_{0}\right)^{2}}$,
and
$R^{-1}\sum_{r=1}^{R}\mathbf{1}\left[\left|\hat{\theta}^{(r)}-\theta_{0}\right|/\hat{\sigma}_{\hat{\theta}}^{(r)}>\mathrm{cv}_{0.05}\right]$,
respectively, for true parameter $\theta_{0}$, its estimate
$\hat{\theta}^{(r)}$, the estimated standard error of $\hat{\theta}^{(r)}$,
$\hat{\sigma}_{\hat{\theta}}^{(r)}$, and the critical value
$\mathrm{cv}_{0.05}=\Phi^{-1}\left(0.975\right)$ across $R=5,000$
replications, where $\Phi\left(\cdot\right)$ is the cumulative distribution
function of standard normal distribution.
### S.3.2 GMM Estimation of Moments of $\beta_{i}$
With the data generating processes described in Section 5.1, we report the
bias, RMSE and size of the GMM estimator for moments of $\beta_{i}$ in Table
S.3. The GMM estimator for moments of $\beta_{i}$ achieve better small sample
performance as compared to those for the distributional parameters
$\pi,\beta_{L}$ and $\beta_{H}$.
Table S.3: Bias, RMSE and size of the GMM estimator for moments of $\beta$
DGP | Baseline | Categorical $x$ | Categorical $u$
---|---|---|---
Sample size $n$ | Bias | RMSE | Size | Bias | RMSE | Size | Bias | RMSE | Size
high variance: $\mathrm{var}\left(\beta_{i}\right)=0.25$
$\mathrm{E}\left(\beta_{i}\right)=1.5$ | 100 | -0.0080 | 0.2262 | 0.1922 | -0.0117 | 0.2297 | 0.1940 | -0.0030 | 0.2418 | 0.1800
1,000 | -0.0029 | 0.0663 | 0.0936 | -0.0015 | 0.0673 | 0.0848 | -0.0037 | 0.0725 | 0.0804
2,000 | -0.0012 | 0.0431 | 0.0688 | -0.0015 | 0.0463 | 0.0700 | -0.0021 | 0.0494 | 0.0656
5,000 | -0.0003 | 0.0263 | 0.0566 | -0.0009 | 0.0276 | 0.0588 | -0.0013 | 0.0303 | 0.0622
10,000 | 0.0004 | 0.0183 | 0.0530 | -0.0001 | 0.0186 | 0.0498 | -0.0003 | 0.0206 | 0.0492
100,000 | 0.0000 | 0.0056 | 0.0434 | 0.0000 | 0.0058 | 0.0472 | 0.0000 | 0.0066 | 0.0514
$\mathrm{E}\left(\beta_{i}^{2}\right)=2.5$ | 100 | -0.0627 | 0.9082 | 0.3464 | -0.0826 | 0.8821 | 0.3166 | -0.0629 | 0.9459 | 0.3122
1,000 | -0.0300 | 0.2909 | 0.1518 | -0.0275 | 0.2837 | 0.1382 | -0.0362 | 0.3112 | 0.1512
2,000 | -0.0160 | 0.1751 | 0.0976 | -0.0188 | 0.1868 | 0.1074 | -0.0255 | 0.1900 | 0.1048
5,000 | -0.0067 | 0.0916 | 0.0658 | -0.0090 | 0.0993 | 0.0710 | -0.0124 | 0.1091 | 0.0754
10,000 | -0.0015 | 0.0580 | 0.0506 | -0.0036 | 0.0609 | 0.0530 | -0.0061 | 0.0704 | 0.0566
100,000 | -0.0005 | 0.0179 | 0.0462 | -0.0005 | 0.0185 | 0.0498 | -0.0011 | 0.0219 | 0.0542
$\mathrm{E}\left(\beta_{i}^{3}\right)=4.5$ | 100 | -0.2511 | 2.3755 | 0.3698 | -0.2990 | 2.3416 | 0.3424 | -0.2940 | 2.6179 | 0.3522
1,000 | -0.1155 | 0.7641 | 0.1734 | -0.1092 | 0.7613 | 0.1606 | -0.1478 | 0.8856 | 0.1904
2,000 | -0.0667 | 0.4683 | 0.1166 | -0.0745 | 0.5058 | 0.1234 | -0.1066 | 0.5485 | 0.1378
5,000 | -0.0290 | 0.2475 | 0.0800 | -0.0365 | 0.2696 | 0.0788 | -0.0507 | 0.3178 | 0.0942
10,000 | -0.0099 | 0.1559 | 0.0516 | -0.0163 | 0.1699 | 0.0602 | -0.0282 | 0.2088 | 0.0660
100,000 | -0.0020 | 0.0488 | 0.0462 | -0.0023 | 0.0515 | 0.0526 | -0.0052 | 0.0653 | 0.0520
low variance: $\mathrm{var}\left(\beta_{i}\right)=0.15$
$\mathrm{E}\left(\beta_{i}\right)=1.0915$ | 100 | 0.0165 | 0.1943 | 0.1618 | 0.0089 | 0.1983 | 0.1514 | 0.0169 | 0.2112 | 0.1416
1,000 | 0.0045 | 0.0577 | 0.0800 | 0.0042 | 0.0584 | 0.0702 | 0.0033 | 0.0655 | 0.0734
2,000 | 0.0019 | 0.0384 | 0.0594 | 0.0016 | 0.0410 | 0.0698 | 0.0010 | 0.0452 | 0.0632
5,000 | 0.0008 | 0.0243 | 0.0562 | 0.0003 | 0.0250 | 0.0540 | -0.0003 | 0.0283 | 0.0574
10,000 | 0.0007 | 0.0171 | 0.0502 | 0.0001 | 0.0175 | 0.0476 | 0.0000 | 0.0194 | 0.0442
100,000 | 0.0000 | 0.0052 | 0.0430 | 0.0000 | 0.0054 | 0.0476 | 0.0000 | 0.0062 | 0.0472
$\mathrm{E}\left(\beta_{i}^{2}\right)=1.3413$ | 100 | -0.0121 | 0.5119 | 0.2440 | -0.0280 | 0.5095 | 0.2330 | -0.0236 | 0.5724 | 0.2340
1,000 | -0.0061 | 0.1528 | 0.1232 | -0.0084 | 0.1566 | 0.1126 | -0.0163 | 0.1776 | 0.1246
2,000 | -0.0072 | 0.0973 | 0.0836 | -0.0080 | 0.1053 | 0.0922 | -0.0143 | 0.1154 | 0.0964
5,000 | -0.0037 | 0.0565 | 0.0658 | -0.0044 | 0.0603 | 0.0698 | -0.0088 | 0.0699 | 0.0720
10,000 | -0.0018 | 0.0381 | 0.0582 | -0.0027 | 0.0401 | 0.0590 | -0.0054 | 0.0476 | 0.0618
100,000 | -0.0004 | 0.0119 | 0.0496 | -0.0005 | 0.0125 | 0.0538 | -0.0009 | 0.0152 | 0.0506
$\mathrm{E}\left(\beta_{i}^{3}\right)=1.7407$ | 100 | -0.0759 | 0.9761 | 0.2806 | -0.0995 | 1.0052 | 0.2672 | -0.1277 | 1.2814 | 0.2718
1,000 | -0.0364 | 0.2925 | 0.1486 | -0.0396 | 0.3112 | 0.1456 | -0.0687 | 0.3973 | 0.1720
2,000 | -0.0297 | 0.1927 | 0.1040 | -0.0310 | 0.2126 | 0.1178 | -0.0526 | 0.2650 | 0.1324
5,000 | -0.0148 | 0.1141 | 0.0798 | -0.0168 | 0.1252 | 0.0860 | -0.0301 | 0.1619 | 0.0964
10,000 | -0.0078 | 0.0771 | 0.0654 | -0.0097 | 0.0846 | 0.0722 | -0.0188 | 0.1126 | 0.0828
100,000 | -0.0013 | 0.0242 | 0.0478 | -0.0016 | 0.0262 | 0.0554 | -0.0031 | 0.0360 | 0.0566
Notes: The data generating process is (5.1). $S=4$ is used. high variance and
low variance parametrization are described in (5.2). “Baseline”, “Categorical
$x$” and “Categorical $u$” refer to DGP 1 to 3 as in Section 5.1. Generically,
bias, RMSE and size are calculated by
$R^{-1}\sum_{r=1}^{R}\left(\hat{\theta}^{(r)}-\theta_{0}\right)$,
$\sqrt{R^{-1}\sum_{r=1}^{R}\left(\hat{\theta}^{(r)}-\theta_{0}\right)^{2}}$,
and
$R^{-1}\sum_{r=1}^{R}\mathbf{1}\left[\left|\hat{\theta}^{(r)}-\theta_{0}\right|/\hat{\sigma}_{\hat{\theta}}^{(r)}>\mathrm{cv}_{0.05}\right]$,
respectively, for true parameter $\theta_{0}$, its estimate
$\hat{\theta}^{(r)}$, the estimated standard error of $\hat{\theta}^{(r)}$,
$\hat{\sigma}_{\hat{\theta}}^{(r)}$, and the critical value
$\mathrm{cv}_{0.05}=\Phi^{-1}\left(0.975\right)$ across $R=5,000$
replications, where $\Phi\left(\cdot\right)$ is the cumulative distribution
function of standard normal distribution.
### S.3.3 Three Estimators of $\mathrm{E}\left(\beta_{i}\right)$
Table S.4 compares the finite sample performance of three estimators of
$\mathrm{E}\left(\beta_{i}\right)$ with the data generating processes
described in Section 5.1.
* •
The OLS estimator $\hat{\mathbf{\phi}}$ studied in Section 3.1
* •
The GMM estimator of $\mathrm{E}\left(\beta_{i}\right)$ with moment conditions
given by (3.7).
* •
$\widehat{\mathrm{E}\left(\beta_{i}\right)}=\hat{\pi}\hat{\beta}_{L}+\left(1-\hat{\pi}\right)\hat{\beta}_{H}$,
where $\hat{\pi},\hat{\beta}_{L},\hat{\beta}_{H}$ are the GMM estimators of
$\pi,\beta_{L},$ and $\beta_{H}$.
According to Table S.4, three estimators perform comparably well in terms of
bias and RMSE, whereas the OLS estimator, along with the standard error from
Theorem 3, controls size well when $n$ is small.
Table S.4: Bias, RMSE and size of three estimators for
$\mathrm{E}\left(\beta_{i}\right)$
DGP | Baseline | Categorical $x$ | Categorical $u$
---|---|---|---
Sample size $n$ | Bias | RMSE | Size | Bias | RMSE | Size | Bias | RMSE | Size
high variance: $\mathrm{E}\left(\beta_{i}\right)=1.5$,
$\mathrm{var}\left(\beta_{i}\right)=0.25$
OLS | 100 | -0.0024 | 0.2035 | 0.0966 | -0.0037 | 0.2035 | 0.0858 | -0.0042 | 0.2268 | 0.0920
1,000 | -0.0017 | 0.0669 | 0.0568 | -0.0002 | 0.0657 | 0.0540 | -0.0019 | 0.0738 | 0.0540
2,000 | -0.0008 | 0.0463 | 0.0512 | -0.0015 | 0.0475 | 0.0534 | -0.0010 | 0.0523 | 0.0522
5,000 | -0.0004 | 0.0301 | 0.0540 | -0.0008 | 0.0300 | 0.0546 | -0.0007 | 0.0335 | 0.0560
10,000 | 0.0002 | 0.0214 | 0.0508 | 0.0000 | 0.0212 | 0.0510 | 0.0000 | 0.0229 | 0.0456
100,000 | -0.0001 | 0.0066 | 0.0472 | 0.0000 | 0.0066 | 0.0460 | 0.0000 | 0.0075 | 0.0506
GMM | 100 | -0.0080 | 0.2262 | 0.1922 | -0.0117 | 0.2297 | 0.1940 | -0.0030 | 0.2418 | 0.1800
1,000 | -0.0029 | 0.0663 | 0.0936 | -0.0015 | 0.0673 | 0.0848 | -0.0037 | 0.0725 | 0.0804
2,000 | -0.0012 | 0.0431 | 0.0688 | -0.0015 | 0.0463 | 0.0700 | -0.0021 | 0.0494 | 0.0656
5,000 | -0.0003 | 0.0263 | 0.0566 | -0.0009 | 0.0276 | 0.0588 | -0.0013 | 0.0303 | 0.0622
10,000 | 0.0004 | 0.0183 | 0.0530 | -0.0001 | 0.0186 | 0.0498 | -0.0003 | 0.0206 | 0.0492
100,000 | 0.0000 | 0.0056 | 0.0434 | 0.0000 | 0.0058 | 0.0472 | 0.0000 | 0.0066 | 0.0514
$\hat{\pi}\hat{\beta}_{L}+(1-\hat{\pi})\hat{\beta}_{H}$ | 100 | -0.0087 | 0.2922 | 0.1961 | -0.1232 | 0.2347 | 0.1809 | -0.0037 | 0.2947 | 0.1894
1,000 | -0.0012 | 0.0648 | 0.0709 | -0.0237 | 0.0783 | 0.0665 | -0.0023 | 0.0713 | 0.0652
2,000 | -0.0004 | 0.0410 | 0.0556 | -0.0140 | 0.0537 | 0.0597 | -0.0015 | 0.0479 | 0.0558
5,000 | 0.0000 | 0.0259 | 0.0536 | -0.0063 | 0.0296 | 0.0546 | -0.0011 | 0.0299 | 0.0590
10,000 | 0.0004 | 0.0183 | 0.0526 | -0.0035 | 0.0205 | 0.0496 | -0.0003 | 0.0205 | 0.0488
100,000 | 0.0000 | 0.0056 | 0.0436 | -0.0006 | 0.0062 | 0.0472 | 0.0000 | 0.0066 | 0.0514
low variance: $\mathrm{E}\left(\beta_{i}\right)=1.0915$,
$\mathrm{var}\left(\beta_{i}\right)=0.15$
OLS | 100 | -0.0006 | 0.1829 | 0.0810 | -0.0023 | 0.1855 | 0.0766 | -0.0025 | 0.2094 | 0.0828
1,000 | -0.0005 | 0.0597 | 0.0610 | 0.0005 | 0.0590 | 0.0478 | -0.0006 | 0.0670 | 0.0542
2,000 | -0.0002 | 0.0408 | 0.0516 | -0.0007 | 0.0427 | 0.0606 | -0.0004 | 0.0475 | 0.0544
5,000 | -0.0002 | 0.0264 | 0.0530 | -0.0006 | 0.0266 | 0.0480 | -0.0005 | 0.0302 | 0.0538
10,000 | 0.0000 | 0.0189 | 0.0546 | -0.0002 | 0.0188 | 0.0486 | -0.0002 | 0.0208 | 0.0482
100,000 | -0.0001 | 0.0059 | 0.0474 | 0.0000 | 0.0059 | 0.0494 | 0.0000 | 0.0068 | 0.0508
GMM | 100 | -0.0121 | 0.5119 | 0.2440 | -0.0280 | 0.5095 | 0.2330 | -0.0236 | 0.5724 | 0.2340
1,000 | -0.0061 | 0.1528 | 0.1232 | -0.0084 | 0.1566 | 0.1126 | -0.0163 | 0.1776 | 0.1246
2,000 | -0.0072 | 0.0973 | 0.0836 | -0.0080 | 0.1053 | 0.0922 | -0.0143 | 0.1154 | 0.0964
5,000 | -0.0037 | 0.0565 | 0.0658 | -0.0044 | 0.0603 | 0.0698 | -0.0088 | 0.0699 | 0.0720
10,000 | -0.0018 | 0.0381 | 0.0582 | -0.0027 | 0.0401 | 0.0590 | -0.0054 | 0.0476 | 0.0618
100,000 | -0.0004 | 0.0119 | 0.0496 | -0.0005 | 0.0125 | 0.0538 | -0.0009 | 0.0152 | 0.0506
$\hat{\pi}\hat{\beta}_{L}+(1-\hat{\pi})\hat{\beta}_{H}$ | 100 | 0.0166 | 0.2392 | 0.1496 | 0.0063 | 0.2342 | 0.1412 | 0.0182 | 0.2432 | 0.1586
1,000 | 0.0078 | 0.0621 | 0.0827 | 0.0068 | 0.0615 | 0.0677 | 0.0064 | 0.0674 | 0.0693
2,000 | 0.0024 | 0.0388 | 0.0559 | 0.0021 | 0.0414 | 0.0672 | 0.0019 | 0.0454 | 0.0627
5,000 | 0.0009 | 0.0241 | 0.0554 | 0.0003 | 0.0247 | 0.0524 | 0.0001 | 0.0282 | 0.0548
10,000 | 0.0007 | 0.0170 | 0.0502 | 0.0002 | 0.0174 | 0.0478 | 0.0003 | 0.0193 | 0.0438
100,000 | 0.0000 | 0.0052 | 0.0430 | 0.0000 | 0.0054 | 0.0480 | 0.0004 | 0.0063 | 0.0494
Notes: The data generating process is (5.1). high variance and low variance
parametrization are described in (5.2). “Baseline”, “Categorical $x$” and
“Categorical $u$” refer to DGP 1 to 3 as in Section 5.1. Generically, bias,
RMSE and size are calculated by
$R^{-1}\sum_{r=1}^{R}\left(\hat{\theta}^{(r)}-\theta_{0}\right)$,
$\sqrt{R^{-1}\sum_{r=1}^{R}\left(\hat{\theta}^{(r)}-\theta_{0}\right)^{2}}$,
and
$R^{-1}\sum_{r=1}^{R}\mathbf{1}\left[\left|\hat{\theta}^{(r)}-\theta_{0}\right|/\hat{\sigma}_{\hat{\theta}}^{(r)}>\mathrm{cv}_{0.05}\right]$,
respectively, for true parameter $\theta_{0}$, its estimate
$\hat{\theta}^{(r)}$, the estimated standard error of $\hat{\theta}^{(r)}$,
$\hat{\sigma}_{\hat{\theta}}^{(r)}$, and the critical value
$\mathrm{cv}_{0.05}=\Phi^{-1}\left(0.975\right)$ across $R=5,000$
replications, where $\Phi\left(\cdot\right)$ is the cumulative distribution
function of standard normal distribution.
### S.3.4 Experiments with higher $\mathrm{var}\left(\beta_{i}\right)$
Following the data generating processes in Section 5.1, we increase the
variance of $\beta_{i}$ by considering the following two parametrizations:
$\left(\pi,\beta_{L},\beta_{H},\mathrm{E}\left(\beta_{i}\right),\mathop{\mathrm{v}ar}\left(\beta_{i}\right)\right)=\begin{cases}(0.3,0.5,6,4.35,6.3525),\\\
(0.3,0.5,10,7.15,18.9525).\end{cases}$ (S.3.1)
Table S.5 presents the results, which show that using larger values of
$\mathrm{var}\left(\beta_{i}\right)$ improves the small sample performance of
the GMM estimators.
Table S.5: Bias, RMSE and size of the GMM estimator for distributional
parameters of $\beta$
DGP | Baseline | Categorical $x$ | Categorical $u$
---|---|---|---
Sample size $n$ | Bias | RMSE | Size | Bias | RMSE | Size | Bias | RMSE | Size
$\mathrm{var}\left(\beta_{i}\right)=6.35$
$\pi=0.3$ | 100 | 0.0755 | 0.3014 | 0.1885 | 0.0628 | 0.2829 | 0.1601 | 0.0760 | 0.2967 | 0.1795
1,000 | -0.0113 | 0.1058 | 0.1485 | -0.0002 | 0.0882 | 0.1406 | -0.0092 | 0.1043 | 0.1509
2,000 | -0.0103 | 0.0646 | 0.1025 | -0.0016 | 0.0495 | 0.1072 | -0.0077 | 0.0598 | 0.1104
5,000 | -0.0026 | 0.0276 | 0.0718 | -0.0009 | 0.0197 | 0.0726 | -0.0021 | 0.0245 | 0.0742
10,000 | -0.0008 | 0.0095 | 0.0576 | -0.0005 | 0.0093 | 0.0608 | -0.0010 | 0.0099 | 0.0588
100,000 | -0.0002 | 0.0027 | 0.0490 | -0.0001 | 0.0026 | 0.0518 | -0.0002 | 0.0028 | 0.0504
$\beta_{L}=0.5$ | 100 | 2.7277 | 3.5109 | 0.2385 | 2.3640 | 3.2861 | 0.2207 | 2.6810 | 3.4783 | 0.2292
1,000 | 0.2951 | 1.1688 | 0.2743 | 0.1539 | 0.9017 | 0.2521 | 0.2473 | 1.1016 | 0.2725
2,000 | 0.0933 | 0.6394 | 0.1916 | 0.0460 | 0.5158 | 0.1988 | 0.0698 | 0.5904 | 0.1951
5,000 | 0.0159 | 0.2570 | 0.1236 | -0.0005 | 0.1786 | 0.1306 | 0.0066 | 0.2080 | 0.1225
10,000 | 0.0009 | 0.0607 | 0.0884 | -0.0005 | 0.0504 | 0.0998 | -0.0014 | 0.0585 | 0.0830
100,000 | 0.0000 | 0.0130 | 0.0572 | 0.0005 | 0.0135 | 0.0630 | -0.0003 | 0.0148 | 0.0622
$\beta_{H}=6$ | 100 | 0.1286 | 1.1700 | 0.0978 | 0.0482 | 1.1467 | 0.1057 | 0.1395 | 1.3662 | 0.0970
1,000 | 0.0031 | 0.2840 | 0.1320 | 0.0062 | 0.2695 | 0.1200 | 0.0043 | 0.3197 | 0.1382
2,000 | -0.0108 | 0.1392 | 0.0982 | 0.0007 | 0.1552 | 0.1094 | -0.0108 | 0.1519 | 0.1088
5,000 | -0.0041 | 0.0621 | 0.0746 | -0.0024 | 0.0608 | 0.0736 | -0.0054 | 0.0652 | 0.0794
10,000 | -0.0018 | 0.0340 | 0.0550 | -0.0012 | 0.0347 | 0.0678 | -0.0034 | 0.0386 | 0.0642
100,000 | -0.0003 | 0.0109 | 0.0530 | 0.0001 | 0.0107 | 0.0518 | -0.0006 | 0.0125 | 0.0588
$\mathrm{var}\left(\beta_{i}\right)=18.95$
$\pi=0.3$ | 100 | 0.0575 | 0.2896 | 0.1761 | 0.0530 | 0.2762 | 0.1524 | 0.0554 | 0.2889 | 0.1646
1,000 | -0.0136 | 0.1070 | 0.1217 | -0.0025 | 0.0892 | 0.1306 | -0.0110 | 0.1024 | 0.1369
2,000 | -0.0101 | 0.0650 | 0.0850 | -0.0032 | 0.0488 | 0.0969 | -0.0077 | 0.0610 | 0.0957
5,000 | -0.0027 | 0.0291 | 0.0668 | -0.0010 | 0.0217 | 0.0625 | -0.0023 | 0.0247 | 0.0713
10,000 | -0.0009 | 0.0122 | 0.0549 | -0.0005 | 0.0097 | 0.0600 | -0.0009 | 0.0100 | 0.0570
100,000 | -0.0002 | 0.0025 | 0.0480 | -0.0001 | 0.0024 | 0.0514 | -0.0002 | 0.0025 | 0.0484
$\beta_{L}=0.5$ | 100 | 4.5691 | 5.9597 | 0.2001 | 4.0139 | 5.6053 | 0.1750 | 4.4575 | 5.8827 | 0.1991
1,000 | 0.5104 | 1.8908 | 0.2327 | 0.2907 | 1.5133 | 0.2146 | 0.4062 | 1.7517 | 0.2522
2,000 | 0.1678 | 1.0260 | 0.1683 | 0.0929 | 0.8581 | 0.1714 | 0.1178 | 0.9144 | 0.1736
5,000 | 0.0292 | 0.3901 | 0.1069 | 0.0073 | 0.3040 | 0.1095 | 0.0186 | 0.3400 | 0.1036
10,000 | 0.0058 | 0.1638 | 0.0719 | 0.0014 | 0.0899 | 0.0834 | 0.0000 | 0.0919 | 0.0740
100,000 | 0.0000 | 0.0171 | 0.0572 | 0.0006 | 0.0171 | 0.0614 | -0.0004 | 0.0185 | 0.0576
$\beta_{H}=10$ | 100 | 0.0520 | 1.5471 | 0.0926 | -0.0530 | 1.4858 | 0.0944 | 0.0460 | 1.6879 | 0.0888
1,000 | -0.0078 | 0.4047 | 0.1185 | -0.0108 | 0.4158 | 0.1020 | -0.0100 | 0.4178 | 0.1195
2,000 | -0.0093 | 0.2058 | 0.0936 | -0.0005 | 0.2067 | 0.0975 | -0.0129 | 0.2546 | 0.0933
5,000 | -0.0037 | 0.0944 | 0.0727 | -0.0034 | 0.0922 | 0.0709 | -0.0052 | 0.0871 | 0.0709
10,000 | -0.0023 | 0.0512 | 0.0555 | -0.0010 | 0.0504 | 0.0684 | -0.0034 | 0.0529 | 0.0580
100,000 | -0.0005 | 0.0160 | 0.0522 | 0.0002 | 0.0154 | 0.0526 | -0.0007 | 0.0171 | 0.0560
Notes: The data generating process is (5.1). Parametrization are described in
(S.3.1). $S=4$ is used. “Baseline”, “Categorical $x$” and “Categorical $u$”
refer to DGP 1 to 3 as in Section 5.1. Generically, bias, RMSE and size are
calculated by
$R^{-1}\sum_{r=1}^{R}\left(\hat{\theta}^{(r)}-\theta_{0}\right)$,
$\sqrt{R^{-1}\sum_{r=1}^{R}\left(\hat{\theta}^{(r)}-\theta_{0}\right)^{2}}$,
and
$R^{-1}\sum_{r=1}^{R}\mathbf{1}\left[\left|\hat{\theta}^{(r)}-\theta_{0}\right|/\hat{\sigma}_{\hat{\theta}}^{(r)}>\mathrm{cv}_{0.05}\right]$,
respectively, for true parameter $\theta_{0}$, its estimate
$\hat{\theta}^{(r)}$, the estimated standard error of $\hat{\theta}^{(r)}$,
$\hat{\sigma}_{\hat{\theta}}^{(r)}$, and the critical value
$\mathrm{cv}_{0.05}=\Phi^{-1}\left(0.975\right)$ across $R=5,000$
replications, where $\Phi\left(\cdot\right)$ is the cumulative distribution
function of standard normal distribution.
### S.3.5 Experiments with three categories ($K=3)$
#### S.3.5.1 Data generating processes
We generate $y_{i}$ as
$y_{i}=\alpha+x_{i}\beta_{i}+z_{i1}\gamma_{1}+z_{i2}\gamma_{2}+u_{i},\text{
for }i=1,2,...,n,$ (S.3.2)
with $\beta_{i}$ distributed as in (2.2) with $K=3$,
$\beta_{i}=\begin{cases}\beta_{L},&\text{ w.p. }\pi_{L}\\\ \beta_{M},&\text{
w.p. }\pi_{M}\\\ \beta_{L},&\text{ w.p. }1-\pi_{L}-\pi_{M},\end{cases}$
where w.p. denotes “with probability”. The parameters take values
$\left(\pi_{L},\pi_{M},\beta_{L},\beta_{M},\beta_{H}\right)=\left(0.3,0.3,1,2,3\right)$.
Corresponding, the moments of $\beta_{i}$ are
$\left(\mathrm{E}\left(\beta_{i}\right),\mathrm{E}\left(\beta_{i}^{2}\right),\mathrm{E}\left(\beta_{i}^{3}\right),\mathrm{E}\left(\beta_{i}^{4}\right),\mathrm{E}\left(\beta_{i}^{5}\right)\right)=\left(2.1,5.1,13.5,37.5,107.1\right)$.
The remaining parameters are set as $\alpha=0.25$, and
$\mathbf{\gamma}=\left(1,1\right)^{\prime}$.
We first generate $\tilde{x}_{i}\sim\text{IID}\chi^{2}(2)$, and then set
$x_{i}=(\tilde{x}_{i}-2)/2$ so that $x_{i}$ has $0$ mean and unit variance.
The additional regressors, $z_{ij}$, for $j=1,2$ with homogeneous slopes are
generated as
$z_{i1}=x_{i}+v_{i1}\text{ and }z_{i2}=z_{i1}+v_{i2},$
with $v_{ij}\sim\text{IID }N\left(0,1\right)$, for $j=1,2$. The error term,
$u_{i}$, is generated as $u_{i}=\sigma_{i}\varepsilon_{i}$, where
$\sigma_{i}^{2}$ are generated as $0.5(1+\text{IID}\chi^{2}(1))$, and
$\varepsilon_{i}\sim\text{IID}N(0,1)$.
#### S.3.5.2 Results
Table S.6 reports the bias, RMSE and size of the GMM estimator for
distributional parameters and moments of $\beta_{i}$. The results are based on
$5,000$ replications and $S=6$. The results show that even larger sample sizes
are needed for the GMM estimators (both the moments of $\beta_{i}$ and its
distributional parameters) to achieve reasonable finite sample performance,
since higher order of moments are involved.
In additional to the results of jointly estimating distributional parameters
and moments of $\beta_{i}$ by GMM, Table S.7 reports the results of GMM
estimation of moments of $\beta_{i}$ up to order 3 using the moment conditions
as in the $K=2$ case where $S=4$ in the left panel, and the results of OLS
estimation of $\mathbf{\phi}$ in the right panel. These results show that we
are still able to obtain accurate estimation of lower order moments of
$\beta_{i}$ when the fourth and fifth moments of $\beta_{i}$ are not used,
confirming the lower information content of the higher order moments for
estimation of the lower order moments of $\beta_{i}$.
Table S.6: Bias, RMSE and size of the GMM estimator for distributional
parameters and moments of $\beta$ with $K=3$
| | Distribution of $\beta_{i}$ | | Moments of $\beta_{i}$
---|---|---|---|---
Sample size $n$ | | Bias | RMSE | Size | | Bias | RMSE | Size
100 | $\pi_{L}=0.3$ | -0.0405 | 0.1910 | 0.1319 | $\mathrm{E}(\beta_{i})=2.1$ | 0.1484 | 0.7471 | 0.6451
1,000 | -0.0417 | 0.1633 | 0.1915 | -0.0711 | 0.5415 | 0.6128
2,000 | -0.0383 | 0.1474 | 0.2354 | -0.1112 | 0.4408 | 0.5264
5,000 | -0.0299 | 0.1186 | 0.3098 | -0.0904 | 0.3712 | 0.4034
10,000 | -0.0209 | 0.0949 | 0.3371 | -0.0523 | 0.2740 | 0.2910
100,000 | -0.0074 | 0.0314 | 0.2295 | -0.0026 | 0.0400 | 0.0678
200,000 | -0.0050 | 0.0208 | 0.1917 | -0.0004 | 0.0202 | 0.0568
100 | $\beta_{M}=0.3$ | 0.2166 | 0.2995 | 0.0492 | $\mathrm{E}(\beta_{i}^{2})=5.1$ | 0.2841 | 2.8452 | 0.7223
1,000 | 0.1404 | 0.2378 | 0.1364 | -0.6374 | 1.9507 | 0.6456
2,000 | 0.1035 | 0.2117 | 0.1901 | -0.7163 | 1.7408 | 0.5472
5,000 | 0.0615 | 0.1645 | 0.2381 | -0.5478 | 1.4628 | 0.4472
10,000 | 0.0364 | 0.1292 | 0.2477 | -0.3391 | 1.1394 | 0.3432
100,000 | 0.0013 | 0.0322 | 0.1305 | -0.0209 | 0.2300 | 0.0932
200,000 | 0.0006 | 0.0185 | 0.1033 | -0.0046 | 0.1128 | 0.0620
100 | $\beta_{L}=1$ | 0.6881 | 1.1994 | 0.1110 | $\mathrm{E}(\beta_{i}^{3})=13.5$ | 0.4897 | 10.0757 | 0.7189
1,000 | 0.2588 | 0.7438 | 0.1994 | -2.7735 | 7.0573 | 0.6718
2,000 | 0.1096 | 0.5372 | 0.2607 | -2.9100 | 6.3988 | 0.5894
5,000 | 0.0205 | 0.4184 | 0.3426 | -2.1889 | 5.4307 | 0.5078
10,000 | 0.0070 | 0.2733 | 0.3360 | -1.3454 | 4.3382 | 0.4042
100,000 | -0.0064 | 0.0556 | 0.2213 | -0.0942 | 1.0263 | 0.1132
200,000 | -0.0047 | 0.0320 | 0.1775 | -0.0236 | 0.5035 | 0.0738
100 | $\beta_{M}=2$ | 0.1249 | 0.7256 | 0.0642 | $\mathrm{E}(\beta_{i}^{4})=37.5$ | 0.9092 | 35.1538 | 0.7235
1,000 | -0.1190 | 0.6298 | 0.1531 | -10.1071 | 24.1521 | 0.6944
2,000 | -0.1935 | 0.5762 | 0.2303 | -10.7108 | 21.5751 | 0.6268
5,000 | -0.1662 | 0.4777 | 0.3670 | -8.2675 | 18.7735 | 0.5464
10,000 | -0.1261 | 0.3703 | 0.4414 | -5.5310 | 15.4382 | 0.4406
100,000 | -0.0326 | 0.1175 | 0.2681 | -0.4433 | 3.5927 | 0.1240
200,000 | -0.0193 | 0.0682 | 0.2203 | -0.1114 | 1.6644 | 0.0810
100 | $\beta_{H}=3$ | 0.8514 | 3.1645 | 0.1064 | $\mathrm{E}(\beta_{i}^{5})=107.1$ | 2.4059 | 121.1286 | 0.6989
1,000 | 1.6632 | 4.5208 | 0.3124 | -34.0298 | 77.5508 | 0.7012
2,000 | 1.7929 | 4.6701 | 0.4000 | -35.4018 | 69.5876 | 0.6424
5,000 | 1.3425 | 4.0152 | 0.4539 | -27.3828 | 60.4373 | 0.5638
10,000 | 0.9637 | 3.3831 | 0.4333 | -18.1022 | 50.3990 | 0.4590
100,000 | 0.0474 | 0.8321 | 0.2046 | -1.5330 | 11.7796 | 0.1314
200,000 | 0.0033 | 0.3237 | 0.1573 | -0.4226 | 5.9529 | 0.0812
Notes: The data generating process is (S.3.2). Generically, bias, RMSE and
size are calculated by
$R^{-1}\sum_{r=1}^{R}\left(\hat{\theta}^{(r)}-\theta_{0}\right)$,
$\sqrt{R^{-1}\sum_{r=1}^{R}\left(\hat{\theta}^{(r)}-\theta_{0}\right)^{2}}$,
and
$R^{-1}\sum_{r=1}^{R}\mathbf{1}\left[\left|\hat{\theta}^{(r)}-\theta_{0}\right|/\hat{\sigma}_{\hat{\theta}}^{(r)}>\mathrm{cv}_{0.05}\right]$,
respectively, for true parameter $\theta_{0}$, its estimate
$\hat{\theta}^{(r)}$, the estimated standard error of $\hat{\theta}^{(r)}$,
$\hat{\sigma}_{\hat{\theta}}^{(r)}$, and the critical value
$\mathrm{cv}_{0.05}=\Phi^{-1}\left(0.975\right)$ across $R=5,000$
replications, where $\Phi\left(\cdot\right)$ is the cumulative distribution
function of standard normal distribution.
Table S.7: Bias, RMSE and size of estimation of $\phi$ and moments of
$\beta_{i}$ (using $S=4$) with $K=3$
| | Moments of $\beta_{i}$ ($S=4$) | | OLS Estimate $hat{\phi}i$
---|---|---|---|---
$n$ | | Bias | RMSE | Size | | Bias | RMSE | Size
100 | $\mathrm{E}(\beta_{i})=2.1$ | 0.0025 | 0.2867 | 0.2088 | $\mathrm{E}(\beta_{i})=2.1$ | -0.0031 | 0.2768 | 0.1042
1,000 | -0.0006 | 0.0821 | 0.1008 | -0.0008 | 0.0939 | 0.0588
2,000 | 0.0004 | 0.0537 | 0.0734 | 0.0000 | 0.0653 | 0.0550
5,000 | 0.0004 | 0.0323 | 0.0610 | -0.0008 | 0.0422 | 0.0506
10,000 | 0.0007 | 0.0224 | 0.0572 | -0.0001 | 0.0299 | 0.0510
100,000 | 0.0000 | 0.0069 | 0.0454 | -0.0001 | 0.0093 | 0.0462
200,000 | 0.0000 | 0.0050 | 0.0550 | 0.0000 | 0.0067 | 0.0498
100 | $\mathrm{E}(\beta_{i}^{2})=5.1$ | -0.1195 | 1.8290 | 0.3948 | $\gamma_{1}=1$ | -0.0020 | 0.1817 | 0.0604
1,000 | -0.0455 | 0.5965 | 0.1602 | 0.0000 | 0.0581 | 0.0474
2,000 | -0.0196 | 0.3454 | 0.0902 | 0.0001 | 0.0409 | 0.0474
5,000 | -0.0073 | 0.1630 | 0.0608 | -0.0001 | 0.0259 | 0.0494
10,000 | -0.0004 | 0.1028 | 0.0544 | -0.0004 | 0.0183 | 0.0518
100,000 | 0.0001 | 0.0311 | 0.0488 | -0.0001 | 0.0058 | 0.0490
200,000 | -0.0002 | 0.0217 | 0.0492 | -0.0001 | 0.0041 | 0.0490
100 | $\mathrm{E}(\beta_{i}^{3})=13.5$ | -0.7404 | 6.7772 | 0.4396 | $\gamma_{2}=1$ | 0.0011 | 0.1296 | 0.0672
1,000 | -0.3116 | 2.2732 | 0.1964 | 0.0000 | 0.0414 | 0.0570
2,000 | -0.1433 | 1.3285 | 0.1110 | 0.0000 | 0.0291 | 0.0478
5,000 | -0.0524 | 0.6468 | 0.0702 | -0.0001 | 0.0183 | 0.0506
10,000 | -0.0117 | 0.4052 | 0.0568 | 0.0002 | 0.0130 | 0.0526
100,000 | 0.0001 | 0.1236 | 0.0528 | 0.0001 | 0.0041 | 0.0494
200,000 | -0.0009 | 0.0850 | 0.0462 | 0.0000 | 0.0029 | 0.0542
Notes: The data generating process is (S.3.2). Generically, bias, RMSE and
size are calculated by
$R^{-1}\sum_{r=1}^{R}\left(\hat{\theta}^{(r)}-\theta_{0}\right)$,
$\sqrt{R^{-1}\sum_{r=1}^{R}\left(\hat{\theta}^{(r)}-\theta_{0}\right)^{2}}$,
and
$R^{-1}\sum_{r=1}^{R}\mathbf{1}\left[\left|\hat{\theta}^{(r)}-\theta_{0}\right|/\hat{\sigma}_{\hat{\theta}}^{(r)}>\mathrm{cv}_{0.05}\right]$,
respectively, for true parameter $\theta_{0}$, its estimate
$\hat{\theta}^{(r)}$, the estimated standard error of $\hat{\theta}^{(r)}$,
$\hat{\sigma}_{\hat{\theta}}^{(r)}$, and the critical value
$\mathrm{cv}_{0.05}=\Phi^{-1}\left(0.975\right)$ across $R=5,000$
replications, where $\Phi\left(\cdot\right)$ is the cumulative distribution
function of standard normal distribution.
### S.3.6 Experiments with idiosyncratic heterogeneity
In addition to the existing results, the following Monte Carlo experiment is
designed to examine the finite sample performance of the estimator under
different degrees of idiosyncratic heterogeneity. Following DGP 1 in Section
5.1, we generate $\tilde{x}_{i}\sim\text{IID}\chi^{2}(2)$, and then set
$x_{i}=(\tilde{x}_{i}-2)/2$. The additional regressors, $z_{ij}$, for $j=1,2$
with homogeneous slopes are generated as
$z_{i1}=x_{i}+v_{i1}\text{ and }z_{i2}=z_{i1}+v_{i2},$
with $v_{ij}\sim\text{IID }N\left(0,1\right)$, for $j=1,2$. The error term,
$u_{i}$, is generated as
$u_{i}=\begin{cases}\sigma_{i}\varepsilon_{i}+e_{i}&\text{if
}i=1,2,\cdots,\lfloor n^{\alpha}\rfloor\\\ \sigma_{i}\varepsilon_{i}&\text{if
}i=\lfloor n^{\alpha}\rfloor+1,\cdots,n\end{cases}$
where $\sigma_{i}^{2}$ are generated as $0.5(1+\text{IID}\chi^{2}(1))$,
$\varepsilon_{i}\sim\text{IID}N(0,1)$, and $e_{i}$ is the idiosyncratic
heterogeneity that is generated from the standard normal distribution and then
set to be fixed across Monte Carlo replications. Then in this case we have
$\left|n^{-1}\sum_{i=1}^{n}\mathrm{E}\left(u_{i}^{2}\right)-1\right|=\left|n^{-1}\sum_{i=1}^{\lfloor
n^{\alpha}\rfloor}e_{i}^{2}\right|\leq n^{-1}\sum_{i=1}^{\lfloor
n^{\alpha}\rfloor}\left|e_{i}^{2}\right|\leq\left(\max_{1\leq i\leq\lfloor
n^{\alpha}\rfloor}\left|e_{i}^{2}\right|\right)n^{\alpha-1}.$
Similar arguments can be made for $r=3$.
Following the same parametrization as in Section 5, we consider the degree of
heterogeneity $\alpha=0.25$, $0.4$, and $0.5$. The estimation results are
reported in Table S.8. The results are similar to that of the Baseline DGP as
reported in Table 3, which suggests that the GMM estimator is robust to
limited degrees of idiosyncratic heterogeneity.
Table S.8: Bias, RMSE and size of the GMM estimator for distributional
parameters of $\beta$
$\alpha$ | 0.25 | 0.40 | 0.50
---|---|---|---
Sample size $n$ | Bias | RMSE | Size | Bias | RMSE | Size | Bias | RMSE | Size
high variance: $\mathrm{var}\left(\beta_{i}\right)=0.25$
$\pi=0.5$ | 100 | 0.0292 | 0.2201 | 0.1957 | 0.0293 | 0.2177 | 0.1859 | 0.0297 | 0.2160 | 0.1609
1,000 | 0.0020 | 0.1273 | 0.1943 | 0.0039 | 0.1293 | 0.2047 | 0.0037 | 0.1356 | 0.2150
2,000 | 0.0014 | 0.0879 | 0.1585 | 0.0003 | 0.0812 | 0.1421 | 0.0020 | 0.0851 | 0.1455
5,000 | 0.0002 | 0.0440 | 0.0980 | 0.0010 | 0.0457 | 0.0982 | -0.0003 | 0.0445 | 0.0946
10,000 | -0.0007 | 0.0301 | 0.0764 | 0.0003 | 0.0304 | 0.0824 | -0.0001 | 0.0311 | 0.0910
100,000 | 0.0000 | 0.0098 | 0.0610 | 0.0000 | 0.0097 | 0.0536 | -0.0002 | 0.0096 | 0.0556
$\beta_{L}=1$ | 100 | 0.2027 | 0.5686 | 0.1807 | 0.1993 | 0.5706 | 0.1738 | 0.2007 | 0.5662 | 0.1712
1,000 | 0.0104 | 0.1711 | 0.2115 | 0.0136 | 0.1750 | 0.2156 | 0.0079 | 0.1827 | 0.2132
2,000 | 0.0094 | 0.1121 | 0.1741 | 0.0069 | 0.1025 | 0.1529 | 0.0087 | 0.1109 | 0.1593
5,000 | 0.0040 | 0.0543 | 0.1090 | 0.0052 | 0.0557 | 0.1136 | 0.0050 | 0.0546 | 0.1112
10,000 | 0.0023 | 0.0365 | 0.0856 | 0.0024 | 0.0365 | 0.0882 | 0.0025 | 0.0367 | 0.0922
100,000 | 0.0004 | 0.0116 | 0.0602 | 0.0005 | 0.0115 | 0.0604 | 0.0004 | 0.0115 | 0.0584
$\beta_{H}=2$ | 100 | -0.1947 | 0.5616 | 0.1307 | -0.1983 | 0.5545 | 0.1421 | -0.2094 | 0.5510 | 0.1358
1,000 | -0.0096 | 0.1720 | 0.1682 | -0.0078 | 0.1729 | 0.1710 | -0.0066 | 0.1802 | 0.1751
2,000 | -0.0060 | 0.1142 | 0.1445 | -0.0068 | 0.1066 | 0.1523 | -0.0070 | 0.1060 | 0.1405
5,000 | -0.0047 | 0.0530 | 0.1130 | -0.0037 | 0.0545 | 0.1110 | -0.0054 | 0.0559 | 0.1088
10,000 | -0.0031 | 0.0360 | 0.0922 | -0.0023 | 0.0370 | 0.0826 | -0.0024 | 0.0372 | 0.0896
100,000 | -0.0004 | 0.0116 | 0.0592 | -0.0003 | 0.0115 | 0.0546 | -0.0005 | 0.0114 | 0.0600
low variance: $\mathrm{var}\left(\beta_{i}\right)=0.15$
$\pi=0.3$ | 100 | 0.2132 | 0.2951 | 0.1851 | 0.2133 | 0.2912 | 0.1797 | 0.2132 | 0.2945 | 0.1716
1,000 | 0.0133 | 0.1591 | 0.1894 | 0.0125 | 0.1613 | 0.1872 | 0.0163 | 0.1637 | 0.1840
2,000 | -0.0051 | 0.1103 | 0.1619 | -0.0055 | 0.1048 | 0.1553 | -0.0027 | 0.1083 | 0.1559
5,000 | -0.0046 | 0.0599 | 0.1198 | -0.0029 | 0.0607 | 0.1070 | -0.0046 | 0.0620 | 0.1208
10,000 | -0.0038 | 0.0418 | 0.0900 | -0.0023 | 0.0418 | 0.0932 | -0.0022 | 0.0423 | 0.0930
100,000 | -0.0003 | 0.0132 | 0.0622 | -0.0003 | 0.0130 | 0.0576 | -0.0004 | 0.0127 | 0.0532
$\beta_{L}=0.5$ | 100 | 0.3935 | 0.6293 | 0.1959 | 0.3900 | 0.6353 | 0.1853 | 0.3917 | 0.6236 | 0.1811
1,000 | 0.0310 | 0.2598 | 0.1590 | 0.0357 | 0.2634 | 0.1589 | 0.0298 | 0.2653 | 0.1609
2,000 | 0.0025 | 0.1590 | 0.1539 | 0.0004 | 0.1478 | 0.1274 | 0.0025 | 0.1565 | 0.1459
5,000 | -0.0008 | 0.0849 | 0.1100 | 0.0018 | 0.0849 | 0.1122 | 0.0003 | 0.0854 | 0.1078
10,000 | -0.0001 | 0.0586 | 0.0922 | 0.0004 | 0.0586 | 0.0958 | 0.0012 | 0.0576 | 0.0918
100,000 | 0.0005 | 0.0183 | 0.0596 | 0.0002 | 0.0181 | 0.0582 | 0.0003 | 0.0177 | 0.0558
$\beta_{H}=1.345$ | 100 | -0.0463 | 0.4194 | 0.1128 | -0.0509 | 0.4224 | 0.1147 | -0.0489 | 0.4386 | 0.1239
1,000 | -0.0097 | 0.1428 | 0.1498 | -0.0106 | 0.1427 | 0.1523 | -0.0094 | 0.1486 | 0.1467
2,000 | -0.0107 | 0.0920 | 0.1443 | -0.0106 | 0.0917 | 0.1439 | -0.0093 | 0.0915 | 0.1389
5,000 | -0.0065 | 0.0492 | 0.1166 | -0.0056 | 0.0500 | 0.1092 | -0.0063 | 0.0532 | 0.1134
10,000 | -0.0045 | 0.0345 | 0.0910 | -0.0037 | 0.0344 | 0.0902 | -0.0035 | 0.0344 | 0.0900
100,000 | -0.0006 | 0.0108 | 0.0602 | -0.0004 | 0.0107 | 0.0572 | -0.0005 | 0.0105 | 0.0560
Notes: The data generating process is (S.3.2). high variance and low variance
parametrization are described in (5.2). $\alpha$ is the degree of
heterogeneity as in Remark 6. Generically, bias, RMSE and size are calculated
by $R^{-1}\sum_{r=1}^{R}\left(\hat{\theta}^{(r)}-\theta_{0}\right)$,
$\sqrt{R^{-1}\sum_{r=1}^{R}\left(\hat{\theta}^{(r)}-\theta_{0}\right)^{2}}$,
and
$R^{-1}\sum_{r=1}^{R}\mathbf{1}\left[\left|\hat{\theta}^{(r)}-\theta_{0}\right|/\hat{\sigma}_{\hat{\theta}}^{(r)}>\mathrm{cv}_{0.05}\right]$,
respectively, for true parameter $\theta_{0}$, its estimate
$\hat{\theta}^{(r)}$, the estimated standard error of $\hat{\theta}^{(r)}$,
$\hat{\sigma}_{\hat{\theta}}^{(r)}$, and the critical value
$\mathrm{cv}_{0.05}=\Phi^{-1}\left(0.975\right)$ across $R=5,000$
replications, where $\Phi\left(\cdot\right)$ is the cumulative distribution
function of standard normal distribution.
## Appendix S.4 Additional empirical results
In this section, we provide additional results for the empirical application.
In addition to the quadratic in experience in Section 6, we further consider
the following quartic in experience specification,
$\log\text{wage}_{i}=\alpha+\beta_{i}\text{edu}_{i}+\rho_{1}\text{exper}_{i}+\rho_{2}\text{exper}_{i}^{2}+\rho_{3}\text{exper}_{i}^{3}+\rho_{4}\text{exper}_{i}^{4}+\tilde{\mathbf{z}}_{i}^{\prime}\tilde{\mathbf{\gamma}}+u_{i},$
(S.4.1)
where
$\beta_{i}=\begin{cases}b_{L}&\text{w.p. }\pi,\\\ b_{H}&\text{w.p.
}1-\pi.\end{cases}$
Table S.9 and S.10 report the estimates of the distributional parameters of
$\beta_{i}$ and the estimates of $\mathbf{\gamma}$ with the specification
(S.4.1).
The estimates of parameter of interests with specification (S.4.1) are almost
the same as that with quadratic in experience specification (6.3), reported in
Table 5. The qualitative analysis and conclusion discussed in Section 6 remain
robust to adding third and fourth order powers of experi in the regressions.
Table S.9: Estimates of the distribution of the return to education with
specification (S.4.1) across two periods, 1973 - 75 and 2001 - 03, by years of
education and gender
| High School or Less | | Postsecondary Edu. | | All
---|---|---|---|---|---
| 1973 - 75 | 2001 - 03 | | 1973 - 75 | 2001 - 03 | | 1973 - 75 | 2001 - 03
| Both Male and Female
$\pi$ | 0.4841 | 0.5081 | | 0.4281 | 0.3576 | | 0.4689 | 0.3559
| (5274.3) | (0.0267) | | (0.0495) | (0.0089) | | (0.0534) | (0.0046)
$\beta_{L}$ | 0.0617 | 0.0392 | | 0.0627 | 0.0859 | | 0.0567 | 0.0658
| (5.9252) | (0.0013) | | (0.0035) | (0.0009) | | (0.0022) | (0.0004)
$\beta_{H}$ | 0.0628 | 0.0928 | | 0.1108 | 0.1397 | | 0.0938 | 0.1270
| (5.5919) | (0.0019) | | (0.0031) | (0.0007) | | (0.0023) | (0.0004)
$\beta_{H}/\beta_{L}$ | 1.0177 | 2.3645 | | 1.7675 | 1.6267 | | 1.6533 | 1.9299
| (7.1413) | (0.0400) | | (0.0629) | (0.0111) | | (0.0305) | (0.0076)
$\mathrm{E}\left(\beta_{i}\right)$ | 0.0623 | 0.0656 | | 0.0902 | 0.1205 | | 0.0764 | 0.1053
$\mathrm{var}\left(\beta_{i}\right)$ | 0.0005 | 0.0268 | | 0.0238 | 0.0258 | | 0.0185 | 0.0293
$n$ | 77,899 | 216,136 | | 33,733 | 295,683 | | 111,632 | 511,819
| Male
$\pi$ | 0.4835 | 0.4968 | | 0.4478 | 0.3007 | | 0.4856 | 0.3550
| n/a | (0.0394) | | (0.0676) | (0.0095) | | (0.0936) | (0.0052)
$\beta_{L}$ | 0.0648 | 0.0419 | | 0.0520 | 0.0733 | | 0.0553 | 0.0581
| n/a | (0.0019) | | (0.0047) | (0.0012) | | (0.0033) | (0.0005)
$\beta_{H}$ | 0.0651 | 0.0927 | | 0.0988 | 0.1321 | | 0.0875 | 0.1220
| n/a | (0.0026) | | (0.0041) | (0.0008) | | (0.0034) | (0.0005)
$\beta_{H}/\beta_{L}$ | 1.0048 | 2.2143 | | 1.9002 | 1.8015 | | 1.5816 | 2.1003
| n/a | (0.0495) | | (0.1124) | (0.0210) | | (0.0456) | (0.0124)
$\mathrm{E}\left(\beta_{i}\right)$ | 0.0649 | 0.0675 | | 0.0778 | 0.1144 | | 0.0719 | 0.0993
$\mathrm{var}\left(\beta_{i}\right)$ | 0.0002 | 0.0254 | | 0.0233 | 0.0269 | | 0.0161 | 0.0306
$n$ | 44,299 | 116,129 | | 20,851 | 144,138 | | 65,150 | 260,267
| Female
$\pi$ | 0.5000 | 0.5210 | | 0.4512 | 0.3849 | | 0.4733 | 0.3773
| (0.5611) | (0.0281) | | (0.0739) | (0.0167) | | (0.0870) | (0.0083)
$\beta_{L}$ | 0.0453 | 0.0352 | | 0.0804 | 0.0956 | | 0.0644 | 0.0762
| (0.0143) | (0.0016) | | (0.0050) | (0.0013) | | (0.0034) | (0.0006)
$\beta_{H}$ | 0.0724 | 0.0969 | | 0.1307 | 0.1449 | | 0.1032 | 0.1338
| (0.0169) | (0.0025) | | (0.0052) | (0.0011) | | (0.0040) | (0.0007)
$\beta_{H}/\beta_{L}$ | 1.5994 | 2.7540 | | 1.6252 | 1.5154 | | 1.6012 | 1.7564
| (0.1537) | (0.0666) | | (0.0551) | (0.0125) | | (0.0323) | (0.0084)
$\mathrm{E}\left(\beta_{i}\right)$ | 0.0588 | 0.0648 | | 0.1080 | 0.1260 | | 0.0848 | 0.1121
$\mathrm{var}\left(\beta_{i}\right)$ | 0.0136 | 0.0308 | | 0.0250 | 0.0240 | | 0.0193 | 0.0279
$n$ | 33,600 | 100,007 | | 12,882 | 151,545 | | 46,482 | 251,552
Notes: This table reports the estimates of the distribution of $\beta_{i}$
with the quartic in experience specification (S.4.1), using $S=4$ order
moments of $\text{edu}_{i}$. “Postsecondary Edu.” stands for the sub-sample
with years of education higher than 12 and “High School or Less” stands for
those with years of education less than or equal to 12.
$\mathrm{s.d.}\left(\beta_{i}\right)$ corresponds to the square root of
estimated $\mathrm{var}\left(\beta_{i}\right)$. $n$ is the sample size. “n/a”
is inserted when the estimates show homogeneity of $\beta_{i}$ and $\pi$ is
not identified and cannot be estimated.
Table S.10: Estimates of $\mathbf{\gamma}$ associated with control variables
$\mathbf{z}_{i}$ with specification (S.4.1) across two periods, 1973 - 75 and
2001 - 03, by years of education and gender, which complements Table S.9
| High School or Less | | Postsecondary Edu. | | All
---|---|---|---|---|---
| 1973 - 75 | 2001 - 03 | | 1973 - 75 | 2001 - 03 | | 1973 - 75 | 2001 - 03
| Both male and female
exper. | 0.0769 | 0.0526 | | 0.0817 | 0.0763 | | 0.0757 | 0.0603
| (0.0015) | (0.0009) | | (0.0029) | (0.0012) | | (0.0013) | (0.0007)
$\mathtt{exper.}^{2}$ | -0.0040 | -0.0020 | | -0.0045 | -0.0039 | | -0.0038 | -0.0024
| (0.0001) | (0.0001) | | (0.0003) | (0.0001) | | (0.0001) | (0.0001)
$\mathtt{exper.}^{3}$ ($\times 10^{5}$) | 9.2470 | 3.4329 | | 11.2100 | 8.9370 | | 8.3625 | 3.6521
| (0.4146) | (0.2882) | | (1.2538) | (0.4460) | | (0.3677) | (0.2412)
$\mathtt{exper.}^{4}$ ($\times 10^{5}$) | -0.0768 | -0.0236 | | -0.1074 | -0.0777 | | -0.0654 | -0.0169
| (0.0043) | (0.0031) | | (0.0158) | (0.0054) | | (0.0039) | (0.0027)
marriage | 0.0819 | 0.0700 | | 0.0728 | 0.0674 | | 0.0799 | 0.0718
| (0.0037) | (0.0020) | | (0.0060) | (0.0020) | | (0.0031) | (0.0014)
nonwhite | -0.1052 | -0.0808 | | -0.0486 | -0.0613 | | -0.0855 | -0.0719
| (0.0046) | (0.0024) | | (0.0088) | (0.0025) | | (0.0041) | (0.0018)
gender | 0.4146 | 0.2272 | | 0.2933 | 0.2008 | | 0.3854 | 0.2150
| (0.0029) | (0.0017) | | (0.0049) | (0.0018) | | (0.0025) | (0.0013)
$n$ | 77,899 | 216,136 | | 33,733 | 295,683 | | 111,632 | 511,819
| Male
exper. | 0.0823 | 0.0620 | | 0.0859 | 0.0780 | | 0.0825 | 0.0664
| (0.0020) | (0.0012) | | (0.0040) | (0.0018) | | (0.0017) | (0.0010)
$\mathtt{exper.}^{2}$ ($\times 10^{2}$) | -0.0039 | -0.0024 | | -0.0041 | -0.0036 | | -0.0037 | -0.0025
| (0.0002) | (0.0001) | | (0.0004) | (0.0002) | | (0.0001) | (0.0001)
$\mathtt{exper.}^{3}$ ($\times 10^{5}$) | 8.2014 | 4.3686 | | 9.2747 | 7.3170 | | 7.4306 | 3.6749
| (0.5321) | (0.3864) | | (1.7422) | (0.6709) | | (0.4700) | (0.3241)
$\mathtt{exper.}^{4}$ ($\times 10^{5}$) | -0.0650 | -0.0314 | | -0.0880 | -0.0582 | | -0.0552 | -0.0161
| (0.0054) | (0.0042) | | (0.0223) | (0.0081) | | (0.0049) | (0.0036)
marriage | 0.1493 | 0.1052 | | 0.1310 | 0.1234 | | 0.1421 | 0.1192
| (0.0056) | (0.0029) | | (0.0088) | (0.0031) | | (0.0048) | (0.0021)
nonwhite | -0.1362 | -0.1191 | | -0.1214 | -0.1040 | | -0.1309 | -0.1136
| (0.0064) | (0.0035) | | (0.0126) | (0.0039) | | (0.0057) | (0.0027)
$n$ | 44,299 | 116,129 | | 20,851 | 144,138 | | 65,150 | 260,267
| Female
exper. | 0.0713 | 0.0455 | | 0.0911 | 0.0782 | | 0.0729 | 0.0568
| (0.0022) | (0.0013) | | (0.0040) | (0.0016) | | (0.0019) | (0.0011)
$\mathtt{exper.}^{2}$ ($\times 10^{2}$) | -0.0044 | -0.0018 | | -0.0067 | -0.0045 | | -0.0045 | -0.0025
| (0.0002) | (0.0001) | | (0.0004) | (0.0002) | | (0.0002) | (0.0001)
$\mathtt{exper.}^{3}$ ($\times 10^{5}$) | 11.0325 | 3.4767 | | 19.6859 | 11.2858 | | 11.3406 | 4.4944
| (0.6649) | (0.4360) | | (1.7412) | (0.5915) | | (0.6095) | (0.3682)
$\mathtt{exper.}^{4}$ ($\times 10^{5}$) | -0.0974 | -0.0264 | | -0.1979 | -0.1046 | | -0.0969 | -0.0272
| (0.0071) | (0.0048) | | (0.0216) | (0.0071) | | (0.0066) | (0.0042)
marriage | -0.0078 | 0.0278 | | -0.0175 | 0.0168 | | -0.0082 | 0.0234
| (0.0048) | (0.0028) | | (0.0080) | (0.0026) | | (0.0041) | (0.0020)
nonwhite | -0.0714 | -0.0479 | | 0.0276 | -0.0291 | | -0.0356 | -0.0375
| (0.0065) | (0.0033) | | (0.0117) | (0.0033) | | (0.0057) | (0.0024)
$n$ | 33,600 | 100,007 | | 12,882 | 151,545 | | 46,482 | 251,552
Notes: This table reports the estimates of $\mathbf{\ \gamma}$ in (S.4.1).
“Postsecondary Edu.” stands for the sub-sample with years of education higher
than 12 and “High School or Less” stands for those with years of education
less than or equal to 12. The standard error of estimates of coefficients
associated with control variables are estimated based on Theorem 3 and
reported in parentheses. $n$ is the sample size.
## Appendix S.5 Computational algorithm
In this section, we describe the computational procedure used for estimation
of $\mathbf{\gamma}$, moments of $\beta_{i}$, and distributional parameters of
$\beta_{i}$.
1. 1.
Denote $\mathbf{w}_{i}=\left(x_{i},\mathbf{z}_{i}^{\prime}\right)^{\prime}$.
Compute the OLS estimator
$\left(\widehat{\mathrm{E}\left(\beta_{i}\right)}^{(0)},\widehat{\mathbf{\gamma}}^{\prime}\right)^{\prime}=\left(\frac{1}{n}\sum_{i=1}^{n}\mathbf{w}_{i}\mathbf{w}_{i}^{\prime}\right)^{-1}\left(\frac{1}{n}\sum_{i=1}^{n}\mathbf{w}_{i}^{\prime}y_{i}\right),$
and $\hat{\tilde{y}}_{i}=y_{i}-\mathbf{z}_{i}^{\prime}\widehat{\gamma}$.
2. 2.
For $r=2,3,\cdots,2K-1$, compute the sample version of the moment conditions
(2.8) and (2.9) in the main paper by replacing $\rho_{r,s}$ by
$n^{-1}\sum_{i=1}^{n}\hat{\tilde{y}}_{i}^{r}x_{i}^{s}$, and solving for
$\widehat{\mathrm{E}\left(\beta_{i}^{r}\right)}^{(0)}$ and
$\widehat{\sigma_{r}}^{(0)}$, recursively.
3. 3.
Use the initial estimates
$\left\\{\widehat{\mathrm{E}\left(\beta_{i}^{r}\right)}^{(0)}\right\\}_{r=1}^{2K-1}$
and $\left\\{\widehat{\sigma_{r}}^{(0)}\right\\}_{r=2}^{2K-1}$ to construct
the weighting matrix $\hat{\mathbf{A}}_{n}$ in (3.10) and compute the GMM
estimators
$\left\\{\widehat{\mathrm{E}\left(\beta_{i}^{r}\right)}^{(1)}\right\\}_{r=1}^{2K-1}$
and $\left\\{\widehat{\sigma_{r}}^{(1)}\right\\}_{r=2}^{2K-1}$ to compute the
moments of $\beta_{i}$ and $\sigma_{r}$. Iterate the GMM estimation one more
time with
$\left\\{\widehat{\mathrm{E}\left(\beta_{i}^{r}\right)}^{(1)}\right\\}_{r=1}^{2K-1}$
and $\left\\{\widehat{\sigma_{r}}^{(1)}\right\\}_{r=2}^{2K-1}$ as initial
estimates to obtain
$\left\\{\widehat{\mathrm{E}\left(\beta_{i}^{r}\right)}\right\\}_{r=1}^{2K-1}$
and $\left\\{\widehat{\sigma_{r}}\right\\}_{r=2}^{2K-1}$.
4. 4.
Solve
$\min_{\pi_{k},b_{k}}\left\\{\sum_{j=1}^{r}\left(\sum_{k=1}^{K}\pi_{k}b_{k}^{r}-\widehat{\mathrm{E}\left(\beta_{i}^{r}\right)}\right)^{2}\right\\}$
to get the initial estimates,
$\widehat{\mathbf{\theta}}^{(0)}=\left(\widehat{\mathbf{\pi}}^{(0)\prime},\widehat{\mathbf{b}}^{(0)\prime}\right)^{\prime}$.
5. 5.
Using
$\widehat{\mathbf{\theta}}^{(0)}=\left(\widehat{\mathbf{\pi}}^{(0)\prime},\widehat{\mathbf{b}}^{(0)\prime}\right)^{\prime}$
construct the weighting matrix $\hat{\mathbf{A}}_{n}$ and compute the GMM
estimator as
$\widehat{\mathbf{\theta}}^{(1)}=\left(\widehat{\mathbf{\pi}}^{(1)\prime},\widehat{\mathbf{b}}^{(1)\prime}\right)^{\prime}$
for $\mathbf{\theta}$. Iterate the GMM estimation one more time with
$\widehat{\mathbf{\theta}}^{(1)}=\left(\widehat{\mathbf{\pi}}^{(1)\prime},\widehat{\mathbf{b}}^{(1)\prime}\right)^{\prime}$
as initial estimates to obtain
$\widehat{\mathbf{\theta}}=\left(\widehat{\mathbf{\pi}}^{\prime},\widehat{\mathbf{b}}^{\prime}\right)^{\prime}$.
In the setup of the optimization problem for the optimization solver, imposing
the constraint $b_{1}<b_{2}<\cdots<b_{K}$ is important to improve the
numerical performance, particularly when $n$ is not sufficiently large (less
than $5,000$).
## References
* Hansen (2022) Hansen, E. B. (2022). Econometrics. Princeton University Press, Princeton.
* Pesaran (2015) Pesaran, M. H. (2015). Time Series and Panel Data Econometrics. Oxford University Press, Oxford.
|
# A Multi-Fluid Dust Module in Athena++: Algorithms and Numerical Tests
Pinghui Huang Institute for Advanced Study, Tsinghua University, Beijing
100084, People’s Republic of China Xue-Ning Bai Institute for Advanced Study,
Tsinghua University, Beijing 100084, People’s Republic of China Center for
Astrophysics, Department of Astronomy, Tsinghua University, Beijing 100084,
People’s Republic of China<EMAIL_ADDRESS><EMAIL_ADDRESS>
###### Abstract
We describe the algorithm, implementation and numerical tests of a multifluid
dust module in the Athena++ magnetohydrodynamic (MHD) code. The module can
accommodate an arbitrary number of dust species interacting with the gas via
aerodynamic drag (characterized by the stopping time), with a number of
numerical solvers. In particular, we describe two second-order accurate, two-
stage, fully-implicit solvers that are stable in stiff regimes including short
stopping time and high dust mass loading, and they are paired with the second-
order explicit van-Leer and Runge-Kutta gas dynamics solvers in Athena++,
respectively. Moreover, we formulate a consistent treatment of dust
concentration diffusion with dust back-reaction, which incorporates momentum
diffusion and ensures Galilean invariance. The new formulation and stiff drag
solvers are implemented to be compatible with most existing features of
Athena++, including different coordinate systems, mesh refinement, shearing-
box and orbital advection. We present a large suite of test problems,
including the streaming instability in linear and nonlinear regimes, as well
as local and global setting, which demonstrate that the code achieves the
desired performance. This module will be particularly useful for studies of
dust dynamics and planet formation in protoplanetary disks.
††software: astropy (Astropy Collaboration et al., 2013), Athena++ (Stone et
al., 2020), Mathematica (Wolfram, 1991)
## 1 Introduction
Protoplanetary disks (PPDs) are composed of gas and dust. Although sharing
only about 1% in mass, dust represents the fundamental building blocks of
planets, and it is primarily the thermal radiation from dust that makes PPDs
observable in continuum emission from infrared to millimeter wavelengths. Dust
is coupled with gas via aerodynamic drag, characterized by the stopping time.
Dust particles of small sizes have small stopping time and are strongly
coupled to the gas, while larger dust particles are more loosely coupled, and
hence do not necessarily trace the gas. This fact is not only important many
processes of planet formation, but also crucial for interpreting disk
observations.
The initial stage of planet formation involves dust growth and transport, both
of which are sensitive to disk structure and level of turbulence (Ormel &
Cuzzi, 2007; Birnstiel et al., 2010). In particular, disk turbulence leads to
dust diffusion (Cuzzi et al., 1993; Youdin & Lithwick, 2007; Carballido et
al., 2010; Zhu et al., 2015), which determines the thickness of the dust layer
in the vertical direction, as well as the mixing in the radial direction.
Additional “psudo-diffusion” can result from complex radial gas flow
structures due to, e.g., wind-driven accretion (Hu & Bai, 2021). Upon growing
to larger sizes, back-reaction from dust-to-gas leads to dust clumping due to
the streaming instability (SI, Goodman & Pindor, 2000; Youdin & Goodman,
2005), and subsequently planetesimal formation (Johansen et al., 2007). While
there has been a large number of further studies (e.g., Bai & Stone, 2010a, b;
Carrera et al., 2015; Simon et al., 2017; Yang et al., 2017; Li & Youdin,
2021), it is less clear how the SI interplay with more realistic gas dynamics
(see Johansen et al., 2011; Schäfer et al., 2020; Xu & Bai, 2022). Finally,
instead of planetesimal accretion, the growth of planetary cores by pebble
accretion has been identified to be more efficient towards higher core mass
(Ormel & Klahr, 2010; Lambrechts & Johansen, 2012). The efficiency of pebble
accretion again depends on disk structure and level of turbulence (e.g.
Morbidelli et al., 2015; Xu et al., 2017), and back-reaction from dust-to-gas
may destabilize the feeding zone (Fu et al., 2014; Pierens et al., 2019; Yang
& Zhu, 2020; Huang et al., 2020; Hsieh & Lin, 2020; Surville et al., 2020),
which requires careful study considering realistic gas dynamics in 3D.
Over the past decade, thanks to the advent of the Atacama Large
Millimeter/submillimeter Array (ALMA), as well as high-contrast imaging
techniques equipped in ground-based telescopes, the dramatically improved
resolution and sensitivity have led to the discovery of disk substructures
prevalent in PPDs, particularly in the form of rings and gaps, as well as
various forms of asymmetries (see Andrews, 2020, for a review). These features
are commonly interpreted as a consequence of planet-disk interaction, which
can open gaps (Bae et al., 2017; Dong et al., 2017, 2018), create vortices
(van der Marel et al., 2013; Zhu et al., 2014; Flock et al., 2015), drive
spirals (Dong et al., 2011b, a; Bae & Zhu, 2018a, b), etc. At millimeter/sub-
millimeter wavelength, the observed substructures reflect the distribution of
mm-sized dust particles, which likely substantially amplify substructures in
the gaseous disk because these particles are not strongly tied to gas and tend
to drift towards pressure maxima (Whipple, 1972; Weidenschilling, 1977).
Alternatively, a number of non-planet mechanisms have been identified which
lead to substructure formation, such as processes involving snow lines (Zhang
et al., 2015; Okuzumi et al., 2016; Owen, 2020) and MHD effects (Suriano et
al., 2018; Riols et al., 2020; Cui & Bai, 2021). Some of the mechanisms
requires active participation from dust itself due to its back-reaction
(Takahashi & Inutsuka, 2014, 2016; Tominaga et al., 2019, 2020). In all these
scenarios, it is crucial to co-evolve gas and dust in a self-consistent manner
to help constrain the physical mechanisms behind the observations.
Computationally, dust is commonly treated either as Lagrangian (super-)
particles, or as pressureless fluids. The particle methods have been
implemented in several MHD codes including Pencil (Johansen et al., 2007),
Athena (Bai & Stone, 2010c), FARGO-ADSG (Baruteau & Zhu, 2016) and PLUTO
(Mignone et al., 2019). It has also been naturally employed in smoothed
particle hydrodynamic (SPH) codes including PHANTOM (Price et al., 2018). One
major advantage of the Lagrangian treatment is being able to properly handle
particle crossing, more relevant for particles that are marginally or loosely
coupled to the gas, which is important for studying planetesimal formation by
the SI. On the other hand, it is generally difficult to handle the highly
stiff regime of extreme particle concentration (Bai & Stone, 2010c, but see
Yang & Johansen, 2016; Moseley et al., 2022), and achieving good load
balancing can be challenging for very large simulations (but see Johansen et
al., 2011). Moreover, it is common to treat the unspecified source of disk
turbulence as an effective viscosity in gas dynamic simulations. Doing so for
particles can be involved, especially if one were to further consider dust
back-reaction.
The alternative fluid treatment of dust is gaining popularity, such as in
PIERNIK (Hanasz et al., 2010a, b), MPI-AMRVAC (Porth et al., 2014; Xia et al.,
2018), LA-COMPASS (Li et al., 2005, 2009; Fu et al., 2014) and FARGO3D
(Benítez-Llambay & Masset, 2016; Benítez-Llambay et al., 2019). This approach
is more appropriate for relatively strongly coupled dust, as they quickly
respond to fluid motion to minimize particle crossing. As separate fluids are
co-located with gas in the computational domain, stiffness issues can be
overcome by designing fully-implicit schemes for the drag source term
simultaneously on gas and dust, and load balancing can be trivially satisfied.
Dust diffusion can be easily handled by incorporating a concentration
diffusion source term (Cuzzi et al., 1993; Youdin & Lithwick, 2007). Finally,
this approach is generalizable to further incorporate dust coagulation (Li et
al., 2019; Drazkowska et al., 2019; Li et al., 2020), so that one can self-
consistently compute the dust size distribution at every simulation cell.
In this paper, we describe the algorithm, implementation and numerical tests
of a multifluid dust module in the Athena++ MHD code (Stone et al., 2020). Our
development features a set of dust integrators, particularly two fully-
implicit integrators that can handle all stiff regimes while maintaining 2nd-
order accuracy, which improve upon previous works which were either explicit,
such as MPI-AMRVAC (Porth et al., 2014), FARGO-ADSG (Baruteau & Zhu, 2016) and
PHANTOM (Price et al., 2018), or implicit but only 1st-order accurate
(FARGO3D, Benítez-Llambay et al., 2019). With Athena++ being a Godunov MHD
code, our implementation naturally conserves total momentum and energy.
Moreover, we provide a consistent formulation of dust concentration diffusion,
and show that additional correction terms in the momentum equations of dust
are necessary to properly conserve total momentum and maintain Galilean
invariance. Implementing these terms yield physically sensible results in a
number of test problems.
The outline of this paper is as follows. In Section 2, we describe the
equations, our numerical schemes and implementations. In Section 3, we present
the benchmark tests, including collisions between gas and dust, dust diffusion
with or without momentum correction, linear and non-linear tests of the SI,
global curvilinear simulations of the SI, as well as static/adaptive mesh
refinement tests. Finally, we summarize and discuss our results in Section 4.
## 2 Numerical Scheme
In this section, we describe the basic equations including consistent
formulation of dust concentration diffusion, as well as the numerical schemes
and implementation of the multifluid dust module in Athena++.
### 2.1 General Equations (Conservative Form)
We start by presenting the full set of equations of gas and multifluid dust.
We use subscripts “d” and “g” to denote “dust” and “gas”. Let there be
$N_{\text{d}}$ dust species, each characterized by a stopping time
$T_{\text{s},n}$ representing the timescale they respond to gas drag, where we
use a label “$n$” for the $n$-th dust species. In conservative form, the
equations read:
$\frac{\partial\rho_{\text{g}}}{\partial
t}+\nabla\cdot\left(\rho_{\text{g}}\boldsymbol{v}_{\text{g}}\right)=0\ ,$ (1)
$\displaystyle\frac{\partial\left(\rho_{\text{g}}\boldsymbol{v}_{\text{g}}\right)}{\partial
t}$
$\displaystyle+\nabla\cdot\left(\rho_{\text{g}}\boldsymbol{v}_{\text{g}}\boldsymbol{v}_{\text{g}}+P_{\text{g}}\mathsf{I}+\boldsymbol{\Pi}_{\nu}\right)=$
(2)
$\displaystyle\rho_{\text{g}}\boldsymbol{f}_{\text{g,src}}+\sum^{N_{\text{d}}}_{n=1}\rho_{\text{d},n}\frac{\boldsymbol{v}_{\text{d},n}-\boldsymbol{v}_{\text{g}}}{T_{\text{s},n}}\
,$
$\displaystyle\frac{\partial E_{\text{g}}}{\partial
t}+\nabla\cdot\left[\left(E_{\text{g}}+P_{\text{g}}\right)\boldsymbol{v}_{\text{g}}+\boldsymbol{\Pi}_{\nu}\cdot\boldsymbol{v}_{\text{g}}\right]=\rho_{\text{g}}\boldsymbol{f}_{\text{g,src}}\cdot\boldsymbol{v}_{\text{g}}$
(3)
$\displaystyle+\sum^{N_{\text{d}}}_{n=1}\rho_{\text{d},n}\frac{\boldsymbol{v}_{\text{d},n}-\boldsymbol{v}_{\text{g}}}{T_{\text{s},n}}\cdot\boldsymbol{v}_{\text{g}}+\omega\sum^{N_{\text{d}}}_{n=1}\rho_{\text{d},n}\frac{\left(\boldsymbol{v}_{\text{d},n}-\boldsymbol{v}_{\text{g}}\right)^{2}}{T_{\text{s},n}}\
,$
$\frac{\partial\rho_{\text{d},n}}{\partial
t}+\nabla\cdot\left(\rho_{\text{d},n}\boldsymbol{v}_{\text{d},n}+\boldsymbol{\mathscr{F}}_{\text{dif},n}\right)=0\
,$ (4)
$\displaystyle\frac{\partial\rho_{\text{d},n}\left(\boldsymbol{v}_{\text{d},n}+\boldsymbol{v}_{\text{d,dif},n}\right)}{\partial
t}$
$\displaystyle+\nabla\cdot(\rho_{\text{d},n}\boldsymbol{v}_{\text{d},n}\boldsymbol{v}_{\text{d},n}+\boldsymbol{\Pi}_{\text{dif},n})=$
(5) $\displaystyle\rho_{\text{d},n}$
$\displaystyle\boldsymbol{f}_{\text{d,src},n}+\rho_{\text{d},n}\frac{\boldsymbol{v}_{\text{g}}-\boldsymbol{v}_{\text{d},n}}{T_{\text{s},n}}\
.$
There are $4N_{\text{d}}+5$ equations in total, where Equations (1) to (3) are
the gas continuity, momentum and energy equations, and Equations (4), (5) are
the continuity and momentum equations for the dust species, which are treated
as pressureless fluids (Garaud et al., 2004). In the above, $\rho$ is the
density, $\boldsymbol{v}$ is the velocity, $P_{\text{g}}$ is the gas pressure,
$\mathsf{I}$ is the identity tensor, and
$E_{\text{g}}=(\rho_{\text{g}}\boldsymbol{v}_{\text{g}}^{2})/2+P_{\text{g}}/(\gamma-1)$
is the total energy density of gas, with $\gamma$ being the adiabatic index.
Here we neglect magnetic fields, thermal conduction, etc., as they do not
directly couple to dust, and our dust fluid module is fully compatible with
these existing features.111We treat dust fluids as neutral, but future
extensions may incorporate dust charge, e.g., Hopkins & Squire, 2018.
We incorporate gas viscosity which mimics the presence of external turbulence,
described by the viscous stress tensor $\boldsymbol{\Pi}_{\nu}$:
$\Pi_{\nu,ij}=\rho_{\text{g}}\nu_{\text{g}}\left(\frac{\partial
v_{\text{g},i}}{\partial x_{\text{g},j}}+\frac{\partial
v_{\text{g},j}}{\partial
x_{\text{g},i}}-\frac{2}{3}\delta_{ij}\nabla\cdot\boldsymbol{v_{\text{g}}}\right)\
,$ (6)
where $\nu_{\text{g}}$ is the kinematic viscosity. Closely related to the gas
viscosity is a dust diffusivity $D_{\text{d},n}$, which leads to concentration
diffusion. We treat the diffusivity for each dust species as a free parameter
to be specified by the user, while they are usually prescribed as (Youdin &
Lithwick, 2007):
$D_{\text{d},n}=\frac{\nu_{\text{g}}}{1+\left(T_{\text{s},n}/T_{\text{g,eddy}}\right)^{2}}\
,$ (7)
where $T_{\rm g,eddy}$ is the turbulent eddy time of the external turbulence.
With this, the dust concentration diffusion flux, acting on the dust
continuity equation, is given by
$\boldsymbol{\mathscr{F}}_{\text{dif},n}\equiv-\rho_{\text{g}}D_{\text{d},n}\nabla\left(\frac{\rho_{\text{d},n}}{\rho_{\text{g}}}\right)=\rho_{\text{d},n}\boldsymbol{v}_{\text{d,dif},n}\
,$ (8)
which also gives the definition of the effective dust drift speed
$\boldsymbol{v}_{\text{d,dif},n}$ due to concentration diffusion. Associated
with this concentration diffusion, correction terms must be incorporated to
the dust momentum equation to ensure consistent momentum diffusion flux
(Tominaga et al., 2019) and Galilean invariance. The individual components of
the momentum diffusion flux tensor are given by
$\Pi_{\text{dif},n,ij}=v_{\text{d},n,j}\mathscr{F}_{\text{dif},n,i}+v_{\text{d},n,i}\mathscr{F}_{\text{dif},n,j}\
.$ (9)
Full derivations of the concentration diffusion terms will be presented in
Section 2.2.
The last terms of the right hand sides of the momentum equations (2) and (5)
correspond to the aerodynamic drag between gas and dust. Here we assume linear
drag law, where $T_{{\rm s},n}$ is independent of velocity. Additional two
source terms are added to the energy equation (3), which correspond to the
work done by the drag, and frictional heating. We have included a parameter
$\omega$ to control the level of frictional heating, being $0$ to be turned
off, and $1$ when all dissipation is deposited to the gas.222In reality, some
of the dissipation must lead to heating of the dust. If assuming gas and dust
should maintain the same temperature, one should assign
$\omega=c_{V,\text{g}}\rho_{\text{g}}/(c_{V,\text{g}}\rho_{\text{g}}+c_{V,\text{d}}\rho_{\text{d}})$,
where $c_{V,\text{g}}$ and $c_{V,\text{d}}$ are the heat capacity of gas and
dust, respectively, and
$\rho_{\text{d}}=\sum_{n=1}^{N_{\text{d}}}\rho_{\text{d},n}$.
Other external source terms are denoted by $\boldsymbol{f}_{\text{src}}$,
which may include stellar and/or planetary gravity in disk problems depending
on applications. They are implemented as explicit source terms added on the
momentum equations (2) and (5) following the standard in Athena++ (Stone et
al., 2020). Associated with them is a source term
$W\equiv\boldsymbol{f}_{\text{g,src}}\cdot\boldsymbol{v}_{\text{g}}$ in the
energy equation accounting for the work done by the source terms.
Note that our formulation does not contain an energy equation for dust, thus
does not ensure global energy conservation of the composite dust-gas system.
While this is not of overwhelming concern in typical applications, future
generalization to incorporate a dust energy equation is possible. At
algorithmic level, we thus aim at full momentum conservation, and implement
energy source terms to match the overall accuracy of the algorithm.
### 2.2 Consistent Formulation of Dust Concentration Diffusion
Here we derive dust fluid equations in the presence of turbulent diffusion,
following the procedures of Cuzzi et al. (1993) and Tominaga et al. (2019). We
use the Reynolds averaging technique with approximate closure relations to
properly account for the role of turbulence at sub-grid level while preserving
global conservation laws. In doing so, we are interested in the physics on
time (and potentially length) scales above those for the turbulence, and hence
any physical variable $A$ is decomposed into a time-averaged part
$\overline{A}$ and a fluctuating part $\Delta A$, i.e., $A=\overline{A}+\Delta
A$.
Without loss of generality, we focus on a single dust species and drop its
label $n$. We start from the standard dust fluid equations in conservation
form:
$\displaystyle\frac{\partial\rho_{\text{d}}}{\partial
t}+\nabla\cdot\left(\rho_{\text{d}}\boldsymbol{v}_{\text{d}}\right)$
$\displaystyle=0\ ,$ (10)
$\displaystyle\frac{\partial\rho_{\text{d}}\boldsymbol{v}_{\text{d}}}{\partial
t}+\nabla\cdot\left(\rho_{\text{d}}\boldsymbol{v}_{\text{d}}\boldsymbol{v}_{\text{d}}\right)$
$\displaystyle=\rho_{\text{d}}\frac{\boldsymbol{v}_{\text{g}}-\boldsymbol{v}_{\text{d}}}{T_{\text{s}}}\
.$
Taking averages to the continuity equation, we obtain
$\displaystyle\frac{\partial\overline{\rho_{\text{d}}}}{\partial
t}+\nabla\cdot\left(\overline{\rho}_{\text{d}}\overline{\boldsymbol{v}_{\text{d}}}+\overline{\Delta\rho_{\text{d}}\Delta\boldsymbol{v}_{\text{d}}}\right)=0\
.$ (11)
The extra term
$\overline{\Delta\rho_{\text{d}}\Delta\boldsymbol{v}_{\text{d}}}$, by
definition, corresponds to dust concentration diffusion flux (8)
$\overline{\Delta\rho_{\text{d}}\Delta\boldsymbol{v}_{\text{d}}}\equiv\boldsymbol{\mathscr{F}}_{\text{dif}}=\overline{\rho_{\text{d}}}\boldsymbol{v}_{\text{d},\text{dif}}\
.$ (12)
Next, by taking averages to the momentum equation, we obtain
$\displaystyle\frac{\partial\overline{\rho}_{\text{d}}\overline{\boldsymbol{v}_{\text{d},j}}}{\partial
t}+\frac{\partial\overline{\Delta\rho_{\text{d}}\Delta\boldsymbol{v}_{\text{d},j}}}{\partial
t}+\frac{\partial}{\partial
x_{i}}(\overline{\rho_{\text{d}}}\overline{\boldsymbol{v}_{\text{d},i}}\overline{\boldsymbol{v}_{\text{d},j}}$
(13)
$\displaystyle+\overline{\Delta\rho_{\text{d}}\Delta\boldsymbol{v}_{\text{d},i}}\overline{\boldsymbol{v}_{\text{d},j}}+\overline{\boldsymbol{v}_{\text{d},i}}\overline{\Delta\rho_{\text{d}}\Delta\boldsymbol{v}_{\text{d},j}}+\overline{\rho_{\text{d}}}\overline{\Delta\boldsymbol{v}_{\text{d},i}\Delta\boldsymbol{v}_{\text{d},j}})$
$\displaystyle=\overline{\rho_{\text{d}}}\frac{\overline{\boldsymbol{v}_{\text{g},j}}-\overline{\boldsymbol{v}_{\text{d},j}}}{T_{\text{s}}}+\frac{\overline{\Delta\rho_{\text{d}}\Delta\boldsymbol{v}_{\text{g},j}}-\overline{\Delta\rho_{\text{d}}\Delta\boldsymbol{v}_{\text{d},j}}}{T_{\text{s}}}\
.$
At this stage, it is often argued that one can drop the second term on the
left assuming the time-dependent diffusion flux is small compared to that of
the bulk flow (Cuzzi et al., 1993; Tominaga et al., 2019). However, our
analysis shows that this would violate Galilean invariance (see Appendix A,
and also numerical tests in Section 2.4.2), and hence it must be kept. The
second and third terms in the momentum flux can be reduced using the effective
dust drift velocity $\boldsymbol{v}_{\text{d,dif}}$, which leads to the
expression of momentum diffusion flux (9). We note that momentum conservation
does not necessarily requires the inclusion of momentum diffusion flux, but
this flux is important when considering angular momentum conservation in disk
problems (Tominaga et al., 2019), as well as for ensuring Galilean invariance.
For the last term in the momentum flux
$\overline{\rho}_{\text{d}}\overline{\Delta\boldsymbol{v}_{\text{d},i}\Delta\boldsymbol{v}_{\text{d},j}}$,
we may use the simple closure relation by Shariff & Cuzzi (2011) and Tominaga
et al. (2019) as
$\overline{\Delta\boldsymbol{v}_{\text{d},i}\Delta\boldsymbol{v}_{\text{d},j}}=\delta_{ij}c_{\text{s,d}}^{2}\
.$ (14)
where $c_{\text{s,d}}$ is the effective dust sound speed. This term can be
neglected in the multifluid approach (Garaud et al., 2004). The second term on
the right hand side is also neglected with the expectation that the standard
drag term dominates, as in Tominaga et al. (2019).
With all these considerations, we recover the dust momentum equation shown in
Section 2.1, here rewritten as
$\displaystyle\frac{\partial\rho_{\text{d}}\left(v_{\text{d},j}+v_{\text{d,dif},j}\right)}{\partial
t}$ $\displaystyle+\frac{\partial}{\partial
x_{i}}(\rho_{\text{d}}v_{\text{d},i}v_{\text{d},j}+\rho_{\text{d}}v_{\text{d,dif},i}v_{\text{d},j}$
(15) $\displaystyle+\rho_{\text{d}}v_{\text{d},i}v_{\text{d,dif},j})$
$\displaystyle=\rho_{\text{d}}\frac{v_{\text{g},j}-v_{\text{d},j}}{T_{\text{s}}}\
.$
where for notational convenience, we can drop the overline and interpret the
dust fluid quantities in the averaged sense. The presence of time-derivative
on $\rho_{d}\boldsymbol{v}_{\text{d,dif}}$ in the momentum equation is the
inevitable consequence of this averaging procedure. Missing this term would
lead to unphysical behaviors as we demonstrate in Section 3.2. Implementing
this term also requires special care, as will be discussed in Section 2.4.2.
### 2.3 Dust-Gas Drag Integrators
The drag term involves interactions between gas and all dust species. As a
special source term to both gas and dust, the drag integrator aims to solve
the following equation
$\displaystyle\frac{\partial\boldsymbol{M}}{\partial
t}=\begin{bmatrix}\sum^{N_{\text{d}}}_{n=1}\alpha_{n}\left(\boldsymbol{M}_{\text{d},n}-\epsilon_{n}\boldsymbol{M}_{\text{g}}\right)\\\
\alpha_{1}(\epsilon_{1}\boldsymbol{M}_{\text{g}}-\boldsymbol{M}_{\text{d},1})\\\
\alpha_{2}(\epsilon_{2}\boldsymbol{M}_{\text{g}}-\boldsymbol{M}_{\text{d},2})\\\
\vdots\\\
\alpha_{n}(\epsilon_{n}\boldsymbol{M}_{\text{g}}-\boldsymbol{M}_{\text{d},n})\\\
\end{bmatrix}\equiv\boldsymbol{f}_{\text{drag}}\left(\boldsymbol{M},\boldsymbol{W}\right)\
,$ (16)
with $\boldsymbol{f}_{\text{drag}}$ being the mutual drag force,
$\boldsymbol{M}\equiv[\boldsymbol{M}_{\text{g}},$
$\boldsymbol{M}_{\text{d,1}},\ \dots,\
\boldsymbol{M}_{\text{d},n}]^{\top}=[\rho_{\text{g}}\boldsymbol{v}_{\text{g}},\
\rho_{\text{d},1}\boldsymbol{v}_{\text{d,1}},\ \dots,\
\rho_{\text{d},n}\boldsymbol{v}_{\text{d},n}]^{\top}$ is the momentum vector
of gas and dust. The remaining variables are denoted as $\boldsymbol{W}$ given
by ($\boldsymbol{\epsilon},\boldsymbol{\alpha})$, where
$\boldsymbol{\epsilon}=\left[\epsilon_{1},\dots,\epsilon_{n}\right]\equiv[\rho_{\text{d},\
1}/\rho_{\text{g}},\ \dots,\ \rho_{\text{d},n}/\rho_{\text{g}}]$, and
$\boldsymbol{\alpha}=\left[\alpha_{1},\dots\alpha_{n}\right]\equiv[T_{\text{s},1}^{-1},\dots
T_{\text{s},n}^{-1}]$. They are treated as constant parameters in the
integrator.
The drag term is potentially stiff in two regimes. First, when the dust
stopping time $T_{s}$ is very small, and stiffness arises when $T_{S}<\Delta
t\equiv h$, the hydrodynamic time step. Second, when $\sum_{n}\epsilon_{n}\gg
1$, which arises when dust is strongly concentrated. The stiff regimes should
be handled by fully-implicit integrators for stability. We note that for
particle-based methods, handling the first regime of stiffness is relatively
straightforward (Bai & Stone, 2010c; Fung & Muley, 2019; Mignone et al.,
2019), whereas handling the second regime requires extra care, where one
either artificially reduces particle back-reaction (Bai & Stone, 2010c), or
sacrifice the time step (Li & Youdin, 2021), and more rigorous treatment
demands substantially more computational cost (Yang & Johansen, 2016). With
the fluid-treatment of dust, one can directly solve the above equation
implicitly, which automatically handles both stiffness regimes (Benítez-
Llambay et al., 2019). In doing so, we need to evaluate the Jacobian of
$\boldsymbol{f}_{\text{drag}}$, and for brevity we drop the subscript “drag”:
$\displaystyle\frac{\partial\boldsymbol{f}}{\partial\boldsymbol{M}}=\begin{bmatrix}-\sum^{N_{\text{d}}}_{n}\epsilon_{i}\alpha_{i}&\alpha_{1}&\alpha_{2}&\cdots&\alpha_{n}\\\
\epsilon_{1}\alpha_{1}&-\alpha_{1}&0&\cdots&0\\\
\epsilon_{2}\alpha_{2}&0&-\alpha_{2}&\cdots&0\\\
\vdots&\vdots&\vdots&\ddots&\vdots\\\
\epsilon_{n}\alpha_{n}&0&0&\cdots&-\alpha_{n}\\\ \end{bmatrix}\ .$ (17)
Note that for linear drag law, this Jacobian applies to individual dimensions
(which are independent of each other).
We have implemented a number of different drag integrators. Here we describe
the algorithms of fully-implicit integrators that we develop for achieving
numerical stability and towards higher-order accuracy. The implementation of
other simpler integrators, which are explicit or semi-implicit that are useful
for non-stiff problems, are described in Appendix C.
#### 2.3.1 First Order Fully-Implicit Method
We start from the standard backward Euler method, which is a single-stage
integrator to be combined with the Runge-Kutta 1 (RK1, or forward Euler) time
integrator in Athena++. Integrating from step $n$ to $n+1$, the format is
given by
$\displaystyle\boldsymbol{M}^{(n+1)}$
$\displaystyle=\boldsymbol{M}^{(n)}+h\boldsymbol{f}\left(\boldsymbol{M}^{(n+1)},\boldsymbol{W}^{(n)}\right)\
.$ (18)
Note that this numerical format guarantees momentum conservation. Substituting
$\displaystyle\boldsymbol{f}(\boldsymbol{M}^{(n+1)},\boldsymbol{W}^{(n)})$
$\displaystyle=\boldsymbol{f}(\boldsymbol{M}^{(n)},\boldsymbol{W}^{(n)})$ (19)
$\displaystyle+\frac{\partial\boldsymbol{f}}{\partial\boldsymbol{M}}|^{(n)}(\boldsymbol{M}^{(n+1)}-\boldsymbol{M}^{(n)})\
,$
we can update the momentum
$\Delta\boldsymbol{M}\equiv\boldsymbol{M}^{(n+1)}-\boldsymbol{M}^{(n)}$ by
$\Delta\boldsymbol{M}=\left(\mathsf{I}-h\frac{\partial\boldsymbol{f}}{\partial\boldsymbol{M}}|^{(n)}\right)^{-1}h\boldsymbol{f}\left(\boldsymbol{M}^{(n)},\boldsymbol{W}^{(n)}\right)\
,\\\ $ (20)
where $\mathsf{I}$ is the identity matrix and evaluating
$\Delta\boldsymbol{M}$ involves matrix inversion. This is the main integrator
implemented in FARGO3D (Benítez-Llambay et al., 2019), which makes the mutual
drag interaction unconditional stable, despite of being only 1st-order
accurate in time. With the simple form of the Jacobian (17), the matrix in the
backward Euler method can be solved efficiently on order $\sim O(N_{d})$
instead of $\sim O(N_{d}^{3})$ as in standard LU decomposition (Krapp &
Benítez-Llambay, 2020).
The energy source term on the gas has two parts. The first arises from the
work done by the drag force. To better preserve energy conservation, this term
should be implemented as the change in the gas kinetic energy due to gas drag:
$\Delta
E_{\text{g},1}=\Delta\boldsymbol{M}_{\text{g}}\cdot(\boldsymbol{v}^{(n)}_{\text{g}}+\boldsymbol{v}^{(n+1)}_{\text{g}})/2\
,$ (21)
The second part is from frictional heating (Marble, 1970; Laibe & Price, 2014;
Mignone et al., 2019), which is associated with the reduction of total kinetic
energy in the gas-dust system. This can be calculated by
$\Delta
E_{\text{g},2}=\Delta\boldsymbol{M}_{\text{g}}\cdot\frac{\boldsymbol{v}^{(n)}_{\text{g}}+\boldsymbol{v}^{(n+1)}_{\text{g}}}{2}+\sum_{n=1}^{N_{\text{d}}}\Delta\boldsymbol{M}_{\text{d},n}\cdot\frac{\boldsymbol{v}^{(n)}_{\text{d},n}+\boldsymbol{v}^{(n+1)}_{\text{d},n}}{2}\
,$ (22)
The source terms for the energy equation should thus be
$E_{\text{g}}^{(n+1)}=E_{\text{g}}^{(n)}+\Delta E_{\text{g},1}-\omega\Delta
E_{\text{g},2}\ .$ (23)
#### 2.3.2 Second Order Fully-Implicit Methods
Next we build two fully-implicit drag integrators to be combined with the van-
Leer 2 (VL2) and the Runge-Kutta 2 (RK2) time integrators in Athena++. We
refer to them as the “VL2-Implicit” and “RK2-Implicit” integrators,
respectively. Both integrators involve two stages. Here we describe their
implementation, while the derivation of the algorithm can be found in Appendix
B.
#### VL2-Implicit
Stage I: We apply the backward Euler method to update the system momenta from
step $n$ for half a time step $h/2$, denoted by a prime ′:
$\Delta\boldsymbol{M}^{\prime}=\left(\mathsf{I}-\frac{h}{2}\frac{\partial\boldsymbol{f}}{\partial\boldsymbol{M}}|^{(n)}\right)^{-1}\frac{h}{2}\boldsymbol{f}\left(\boldsymbol{M}^{(n)},\boldsymbol{W}^{(n)}\right)\
,\\\ $ (24)
Matrix inversion in this stage can be similarly achieved on order $O(N_{d})$.
The update in gas energy at this stage is exactly analogous to that in the
backward Euler method, which we do not repeat.
Stage II: the momentum is updated from step $n$ to $n+1$ using the following
$\displaystyle\Delta\boldsymbol{M}$
$\displaystyle=\boldsymbol{\Lambda}^{-1}\left(\mathsf{I}-\frac{h}{2}\frac{\partial\boldsymbol{f}}{\partial\boldsymbol{M}}\bigg{|}^{\prime}\right)h\boldsymbol{f}\left(\boldsymbol{M}^{(n)},\boldsymbol{W}^{\prime}\right)\
,$ (25)
where
$\boldsymbol{\Lambda}\equiv\mathsf{I}-\left(\mathsf{I}-\frac{h}{2}\frac{\partial\boldsymbol{f}}{\partial\boldsymbol{M}}\bigg{|}^{\prime}\right)h\frac{\partial\boldsymbol{f}}{\partial\boldsymbol{M}}|^{(n)}\
.$ (26)
Note that this matrix is more complex and should be inverted by LU
decomposition (Press et al., 1986). The update in gas energy has exactly the
same form as Equations (21) to (23), which we do not repeat.
#### RK2-Implicit
Stage I: We use the backward Euler method with time step $h$ to calculate the
momentum at step $n+1$, which is exactly the same as described in Section
2.3.1. We still denote the quantities at the end of this stage using a prime
′.
Stage II: The momentum at stage $n+1$ is:
$\displaystyle\Delta\boldsymbol{M}=\boldsymbol{\Lambda}^{-1}$
$\displaystyle\bigg{[}h\boldsymbol{f}(\boldsymbol{M}^{(n)},\boldsymbol{W}^{\prime})$
(27)
$\displaystyle+\bigg{(}\mathsf{I}-h\frac{\partial\boldsymbol{f}}{\partial\boldsymbol{M}}\bigg{|}^{\prime}\bigg{)}h\boldsymbol{f}(\boldsymbol{M}^{(n)},\boldsymbol{W}^{(n)})\bigg{]}\
,$
where
$\displaystyle\boldsymbol{\Lambda}$
$\displaystyle\equiv\mathsf{I}-h\frac{\partial\boldsymbol{f}}{\partial\boldsymbol{M}}\bigg{|}^{(n)}+\frac{h^{2}}{2}\frac{\partial\boldsymbol{f}}{\partial\boldsymbol{M}}\bigg{|}^{\prime}\frac{\partial\boldsymbol{f}}{\partial\boldsymbol{M}}\bigg{|}^{(n)}\
.$ (28)
Similarly, matrix inversion is solved by LU decomposition. The update in gas
energy also has exactly the same form as Equations (21) to (23), which we do
not repeat.
Note that this integration scheme is in essence the same as the fully-implicit
particle integrator in Bai & Stone (2010c).
#### 2.3.3 Coupling Explicit Hydrodynamic Integrators with Implicit Drag
Integrators
Special care must be taken when combining implicit integrators with the
explicit hydrodynamic integrators and source terms. When they are treated
separately, the combined algorithm would only be 1st-order accurate,333The
implicit-explicit Runge-Kutta schemes are viable choices (Pareschi & Russo,
2005), they usually involve more stages of integration than the order of
accuracy achieved, and do not necessarily match the existing hydrodynamic
integrators in Athena++. and the implicit drag integrator cannot maintain
exact equilibrium solutions.
To overcome these issues, we may consider the advection, diffusion and other
hydrodynamic source terms as an “add-on” to $\boldsymbol{f}$. In other words,
$\boldsymbol{f}$ in the drag algorithms above represents the combination of
the drag force ($\boldsymbol{f}_{\rm drag}$, treated implicitly), as well as
other explicit terms including advection and other source terms
$\boldsymbol{G}_{M}$ acting on the gas and dust momenta
$\boldsymbol{f}\equiv\boldsymbol{f}_{\rm
drag}(\boldsymbol{M},\boldsymbol{W})+\boldsymbol{G}_{M}(\boldsymbol{\overline{U}})\
,$ (29)
where $\boldsymbol{G}_{M}$ is expressed in terms of conserved variables
$\boldsymbol{U}$. By adding an overline on $\boldsymbol{U}$, we treat these
other explicit terms as known constant, readily obtained in the hydrodynamic
integrator. In the hydrodynamic integration from step 1 to step 2 over time
interval $\Delta t$, we estimate $\boldsymbol{G}_{M}$ to be the momentum
update from explicit terms
$\boldsymbol{G}_{M}(\boldsymbol{\overline{U}})=\frac{\boldsymbol{M}^{(2)}-\boldsymbol{M}^{(1)}}{\Delta
t}\ ,$ (30)
where $\boldsymbol{M}^{(2),(1)}$ represents the momenta before and after
explicit integration steps (advection, diffusion and other explicit source
terms). By treating this term as a constant, the Jacobian and the $\Lambda$
matrices described in the previous subsection remain unchanged.
Implementing the above requires extra storage to store the momentum updates,
and that we must finish all explicit steps in the hydrodynamic integration
before entering the drag integrator. We will show that our approach
successfully achieves 2nd-order accuracy when using VL2-Implicit and
RK2-Implicit integrators, and it also allows us to achieve exact equilibrium
solutions involving the drag force.
### 2.4 Integration of Multifluid Dust Equations
The integration of dust fluid is divided into several parts (advection,
diffusion, source terms and drag). Except for the drag term (described in the
previous subsection), the other terms are treated independently and
explicitly, and we describe their implementation in this subsection.
#### 2.4.1 General Procedures
Following the standard routine in Athena++, each integration time step is
divided into a number of stages depending on the time integrator employed (see
detailed descriptions in Stone et al., 2020). In each stage, the integration
procedures involve updating conserved variables based on primitive variables
by evolving the fluid equations by $dt$. Our multifluid dust module supports
Athena++ time integrators up to second order, including 1st-order Runge-Kutta
(RK1), 2nd-order Runge-Kutta (RK2) and van-Leer integrator (VL2).
For each dust species, the primitive ($\mathbf{W}_{\text{d}}$) and conserved
($\mathbf{U}_{\text{d}}$) variables are
$\mathbf{W}_{\text{d}}=\begin{bmatrix}\rho_{\text{d}}\\\
\boldsymbol{v}_{\text{d}}\\\ \end{bmatrix}\
,\quad\mathbf{U}_{\text{d}}=\begin{bmatrix}\rho_{\text{d}}\\\
\rho_{\text{d}}(\boldsymbol{v}_{\text{d}}+\boldsymbol{v}_{\text{d,dif}})\\\
\end{bmatrix}\ .$ (31)
Note that the presence of time derivative on
$\rho_{\text{d}}\boldsymbol{v}_{\text{d,dif}}$ in the momentum equation
suggests that the concentration diffusion flux should be considered as part of
the conserved dust momentum. The total momentum is thus
$\rho\boldsymbol{v}_{\text{g}}+\sum_{i}\rho_{\text{d},i}(\boldsymbol{v}_{\text{d},i}+\boldsymbol{v}_{\text{d,dif},i})$.
Integrating the bulk part of the dust fluid is very similar that of
hydrodynamics in Athena++. The main procedures involves the reconstruction of
primitive variables at cell interfaces, followed by solving a Riemann problem
to obtain the mass and momentum fluxes, after which we update the dust fluid
quantities from flux gradients. Same as in Athena++, the multifluid dust
module supports spatial reconstructions up to third order.
As pressureless fluids, the Riemann problem for dust fluids is greatly
simplified. In one-dimension along the $x$-direction, given the left/right
states $\mathbf{W}_{\text{d}}^{L/R}$, we provide the Riemann flux for
conserved variables as follows. The density flux reads
$F_{x}(\rho_{d})=\begin{cases}\rho_{\text{d}}^{L}v_{\text{d},x}^{L}&v_{\text{d},x}^{L}>0\
,\ v_{\text{d},x}^{R}>0\ ,\\\
\rho_{\text{d}}^{R}v_{\text{d},x}^{R}&v_{\text{d},x}^{L}<0\ ,\
v_{\text{d},x}^{R}<0\ ,\\\ 0&v_{\text{d},x}^{L}<0\ ,\ v_{\text{d},x}^{R}>0\
,\\\
\rho_{\text{d}}^{L}v_{\text{d},x}^{L}+\rho_{\text{d}}^{R}v_{\text{d},x}^{R}&v_{\text{d},x}^{L}>0\
,\ v_{\text{d},x}^{R}<0\ .\\\ \end{cases}$ (32)
Similar expressions hold for the momentum flux for all three directions.
Essentially, we use the upwind flux when the normal velocity in the L/R states
are the same, set the flux to be zero when the L/R normal velocities diverge,
and sum up the fluxes from the two sides when L/R normal velocities converge.
The last treatment reflects that as pressureless fluids, the flows on the two
sides can penetrate each other, just as particles.444Note that penetration is
still prohibited within each cell, where dust fluid velocities get well mixed.
Alternatively, one may set the flux to zero in this case. We do not find much
practical differences in test problems by using different Riemann solvers for
dust.
The implementation of other source terms on dust, such as stellar gravity and
source terms in shearing-box, as well as geometric source terms in cylindrical
and spherical coordinates, are same as that of gas, which are treated
explicitly.
#### 2.4.2 Dust Diffusion
The implementation of dust concentration diffusion starts by computing the
concentration diffusion flux according to Equation (8). The fluxes are
computed by standard finite differencing, and are located at cell interfaces.
Next, we calculate the momentum diffusion flux according to Equation (9). This
term contains two parts. The first part,
$v_{\text{d},n,j}\mathscr{F}_{\text{dif},n,i}$, describes the diffusion of the
$j$-momentum in the $i$-direction. At the implementation level, its value is
obtained by averaging from the upwind side based on the sign of the
concentration diffusion flux $\mathscr{F}_{\text{dif},n,i}$. The second part,
$v_{\text{d},n,i}\mathscr{F}_{\text{dif},n,j}$, represents the advection of
the $j$-diffusion flux in the $i$-direction. Its value is obtained by
averaging from the upwind side based on the sign of the advection velocity
$v_{\text{d},n,i}$. In addition, we note that in cylindrical/spherical
coordinates, we need to add extra diffusive geometric sources terms on the
momentum and energy equations (Skinner & Ostriker, 2010).
Finally, we compute the concentration diffusion momenta and compare to the
original concentration diffusion flux, from which we can estimate the
contribution from the
$\partial(\rho_{\text{d},n}\boldsymbol{v}_{\text{d,dif},n})/\partial t$ term.
The concentration diffusion momenta are stored in the cell center and are
averaged by the nearby face-centered concentration diffusion fluxes. We note
that although our formulation is Galilean invariant, it is not invariant to
machine precision at implementation level, but the incorporation of this term
is important to ensure approximate Galilean invariance in simulations.
### 2.5 Flow chart
Figure 1: Flow chart of a single integration stage of the multifluid dust
module in Athena++.
Figure 1 shows the flow chart of our multifluid dust module in Athena++, and
we summarize the main steps over one integration stage below.
Step 1: Backup the primitive variables for both gas and dust and calculate the
dust stopping time. The backed up primitive variables are used in the semi-
implicit and fully-implicit drag integrators to ensure higher-order accuracy
of the combined algorithm, as discussed in Section 2.3.3.
Step 2: Calculate the diffusion processes of gas and dust when applicable,
including viscosity, thermal conduction, and resistivity on the gas, and
concentration diffusion and momentum correction on dust fluids.
Step 3: Calculate the Riemann fluxes of both gas and dust, and integrate gas
and dust fluids by applying flux divergence. Send and receive flux corrections
when necessary for mesh refinement.
Step 4: Add explicit source terms on gas and dust, including geometric source
terms for curvilinear coordinates.
Step 5: Apply any of the drag integrators, and use the backed-up variables to
enhance the accuracy in implicit schemes.
Step 6: Do orbital advection when necessary (for disk problems).
Step 7: Send and receive boundary data, set boundary conditions, and do
prolongation/restriction for mesh refinement.
Step 8: Convert conserved variables to primitive variables. When the dust
momentum correction is turned on, the concentration diffusion flux calculated
by Step 2 will be subtracted from the dust momenta.
After finishing all stages of an integration cycle, we calculate the new time
step based on the Courant–Friedrichs–Lewy (CFL) condition for both gas and
dust. The dust CFL condition is set according to maximum dust velocity and
dust diffusion coefficient $D_{\text{d}}$ in the same way as gas velocity and
viscosity.
As a dust fluid module, it has fixed amount of floating point operations per
meshblock per integration cycle, as opposed to particle-based approaches.
Taking the advantage of the task-based execution model with excellent
scalability of Athena++, our dust fluid module primarily adds a fixed fraction
of computational cost. Such cost increases with $N_{d}$ non-linearly when
using higher-order fully implicit drag solvers due to the matrix inversion
whose cost scales as $O[(N_{d}+1)^{3}]$. In practice, we find that linear
scaling approximately applies for $N_{d}\lesssim 5$ and the cost of the drag
solver is no more than the cost from rest of the dust integration scheme for
$N_{d}\lesssim 10$. Further details about code performance are provided in
Appendix D.
## 3 Code Tests
In this section, we show benchmark numerical tests of our multifluid dust
module. They include the collisions between gas and dust, dust diffusion with
momentum correction, linear/non-linear streaming instability and
(static/adaptive) mesh refinement. We also follow the same dusty sound wave
and dusty shock tests in Section 3.2 and 3.3 of Benítez-Llambay et al. (2019).
To avoid repetitions, we show the test results of dusty sound wave and dusty
shock in Appendix F.1 and F.2. They demonstrate that our multi-fluid dust code
achieves full second-order accuracy when coupled with hydrodynamics, and it is
excellent at shock capturing.
### 3.1 Collisions
We start by conducting the 1D dust-gas collision test as a benchmark, similar
to Section 3.1 of Benítez-Llambay et al. (2019). We consider two dust species
with constant stopping time $T_{s,1},T_{s,2}$, and set three collision tests
named A, B and C. The gas and all dust species are homogeneous, each having
its own density ($\rho_{\text{g}},\rho_{\text{d},1},\rho_{\text{d},2}$) and
velocity ($v_{\text{g}},v_{\text{d},1},v_{\text{d},2}$). The system then
evolves under the mutual aerodynamic drag forces, characterized by two
eigenvalues $\lambda_{1}$, $\lambda_{2}$, in the form of
$v=v_{\text{COM}}+c_{1}\exp{(\lambda_{1}\;t)}+c_{2}\exp{(\lambda_{2}\;t)}\ ,$
(33)
where $v_{\text{COM}}$ is the center-of-mass velocity of the system. Their
initial conditions, as well as the associated coefficients and eigenvalues are
given by shown in Table 1, and we provide the calculation procedures in
Appendix E. The three tests are designed to test the non-stiff case (Test A),
the stiff case with small stopping time (Test B) and the stiff case with large
dust-to-gas ratios (Test C). These tests are conducted in 1D Cartesian
coordinates with a periodic boundary condition. We use the adiabatic equation
of state with the adiabatic index being $\gamma=1.4$ and an initial gas sound
speed is set as $c_{\text{s}}^{2}\equiv\gamma\frac{P}{\rho_{\text{g}}}=1.4$
for all three tests. We include the work and friction heating from drags in
the energy equation. We have tested eight drag integrators (Explicit:
“RK1-Explicit”, “RK2-Explicit”, “VL2-Explicit”; Semi-Implicit: “Trapezoid”,
“TrBDF2”; Fully-Implicit: “RK1-Implicit”, “RK2-Implicit” and “VL2-Implicit”),
and the main results are discussed below.
Table 1: The Initial Conditions, Eigenvalues and Coefficients of the Analytic
Solutions in the Collision Tests
Test | A | B | C
---|---|---|---
$\rho_{\text{g}}$ | 1 | 1 | 1
$v_{\text{g}}$ | 1 | 1 | 1
$\rho_{\text{d,1}}$ | 1 | 1 | 10
$v_{\text{d,1}}$ | 2 | 2 | 2
$T_{\text{s,1}}$ | 2 | 0.01 | 2
$\rho_{\text{d,2}}$ | 1 | 1 | 100
$v_{\text{d,2}}$ | 0.5 | 0.5 | 0.5
$T_{\text{s,2}}$ | 1 | 0.002 | 1
Coefficients1 | | |
$v_{\text{COM}}$ | 1.16666666666667 | 1.16666666666667 | 0.63963963963963
$\lambda_{1}$ | -0.63397459621556 | -141.742430504416 | -0.52370200744224
$\lambda_{2}$ | -2.36602540378444 | -1058.25756949558 | -105.976297992557
$c_{\text{g,1}}$ | -0.22767090063074 | -0.35610569612832 | -0.06458203330249
$c_{\text{g,2}}$ | 0.06100423396407 | 0.18943902946166 | 0.42494239366285
$c_{\text{d,1,1}}$ | 0.84967936855889 | 0.85310244713865 | 1.36237475791577
$c_{\text{d,1,2}}$ | -0.01634603522555 | -0.01976911380532 | -0.00201439755542
$c_{\text{d,2,1}}$ | -0.62200846792815 | -0.49699675101033 | -0.13559165545855
$c_{\text{d,2,2}}$ | -0.04465819873852 | -0.16966991565634 | -0.00404798418109
* 1
The analytic solutions of velocities of gas and dust are in the form of
Equation (33). The analytic solutions of gas energy can be obtained by
integrating the right hand side of the Equation (3) with $\omega=1$.
Figure 2: The top eight panels are the velocities and energy of gas and dust
in the collision Test B and C with the numerical time step $\Delta t=0.005$
and $\Delta t=0.05$. Solutions obtained by semi-implicit and fully-implicit
different drag integrators are shown. The black dashed lines represent the
analytic solutions. Note that explicit methods are unstable for our choice of
$\Delta t>1/|\lambda_{2}|$ (see Table 1), and the results are not shown. The
bottom three panels are the scaling of total relative error $\Delta
E_{\text{total}}$ with numerical time step $\Delta t$ for different drag
integrators (see legends) in the collision Tests A (left), B (middle) and C
(right).
The top eight panels of Figure 2 shows the temporal evolution of velocities
and energy of both gas and dust in Test B and C with numerical time steps
$\Delta t=0.005$ and $0.05$, respectively. The results are to be compared with
the analytic solution. Table 1 shows the largest eigenvalue
$|\lambda_{2}|\simeq 1058$ and $106$ for Test B and C. Therefore, the drag
terms become stiff when $\Delta t>1/|\lambda_{2}|\simeq 0.001$ and $0.01$ for
B and C. The explicit integrators are unstable in Test B and C with $\Delta
t=0.005$ and $0.05$. The semi-implicit methods are stable in the stiff drags,
however, the numerical updates oscillate around the analytic solutions
artificially, which is not unexpected as similar behavior was observed in Bai
& Stone (2010c). Our fully-implicit methods “RK1-Implicit”, “VL2-Implicit” and
“RK2-Implicit” handle these two stiff regimes (small stopping times and large
dust-to-gas ratios) very well, and the two second-order integrators are
clearly seen to be more accurate.
To test numerical convergence, we calculate the relative error $\Delta E$ as a
function of the numerical time step $\Delta t$ with different drag
integrators. The $\Delta E$ is calculate by:
$\Delta E\left(\Delta
t\right)=\frac{1}{t_{\text{max}}-t_{\text{min}}}\sum\frac{|U_{\text{num}}\left(\Delta
t\right)-U_{\text{ana}}|}{U_{\text{ana}}}\Delta t$ (34)
where $U$ represents momentum or the gas energy, subscripts “num” and “ana”
represent the numerical and analytic solutions, respectively. The scaling of
the total relative error $\Delta E_{\text{total}}=\Delta E_{\text{mom}}+\Delta
E_{\text{erg}}$ with different time steps $\Delta t$ is shown in the bottom
three panels of Figure 2 for different drag integrators. We vary $\Delta t$
from $10^{-4}$ to $10^{-1}$ which are non-stiff for Test A, but get
increasingly stiff for Test B and C. The total relative errors are calculated
between $t_{\text{min}}=0$ and $t_{\text{max}}=10$.
We see that in the non-stiff regime of Test A, all drag integrators achieve
first or second-order accuracy in time as desired, and there is no significant
difference between integrators of the same order. In the stiff regime of Test
B and C, we see that the error in explicit and semi-implicit integrators
diverge when $\Delta t\gtrsim(1/|\lambda_{2}|)$. The threshold for error
divergence is only slightly higher for semi-implicit integrators. The fully-
implicit integrators, on the other hand, achieve the desired level of accuracy
at small $\Delta t$, while remain stable at large $\Delta t$ for both tests.
Among them, the 2nd-order fully-implicit integrators show error levels that
are at least one order of magnitude smaller, and are the preferred choices
that we generally recommend.
Our drag integrators conserve total momentum of the dust-gas system by
construction. While not shown in the figures, we have verified that in all
three tests, total momentum is conserved to the fractional level of $\lesssim
10^{-14}$ (i.e., approaching machine precision) within the duration of the
simulations.
### 3.2 Momentum Correction in Dust Diffusion
Novel additions in our dust concentration diffusion formulation include the
time derivative of the $\rho_{\text{d}}\boldsymbol{v}_{\text{d,dif}}$ term and
the $\boldsymbol{\Pi}_{\text{d,dif}}$ in the dust momentum equation (5). We
refer to them as “momentum correction”. They reflect the fact that the
concentration diffusion flux carries momentum that back-react to the gas, and
that the new formulation is Galilean invariant. To test these aspects of our
implementation, we consider the following simple test problems without/with
momentum correction in 1D, 1.5D and 2D Cartesian coordinates, where $1.5D$
means 1D test in the presence of a transverse velocity.
#### 3.2.1 Initial Setups
We give the parameters of our tests in Table 2, with more details below. The
1D and 1.5D tests are carried in Cartesian coordinates, with 256 uniform
grids, covering $x\in[0,20]$. The gas has uniform initial density
$\rho_{\text{g0}}=1$ and a single dust species with constant stopping time
$T_{\text{s}}=10^{-2}$ is included with an initial Gaussian density profile:
$\rho_{\text{d0}}=A\exp{\left[-\frac{\left(x-x_{0}\right)^{2}}{2\sigma_{x}^{2}}\right]}+\rho_{\text{g0}}\
.$ (35)
where $A=5$, $x_{0}=10$ and $\sigma_{x}=2$ in both cases. We include gas
viscosity and dust diffusion with coefficients $\nu=D_{\text{d}}=1$. In the 1D
test, initial gas and dust velocities are zero, whereas in the 1.5D test, gas
and dust have constant transverse velocities
$v_{\text{g},y}=v_{\text{d},y}=1$.
In the 2D test, we use 2D Cartesian coordinates in the domain $x,y\in[0,20]$
with $256^{2}$ cells, and set the initial 2D Gaussian dust density profile in
the center of the box:
$\rho_{\text{d0,2D}}=A\exp{\left[-\frac{\left(x-x_{0}\right)^{2}}{2\sigma_{x}^{2}}-\frac{\left(y-y_{0}\right)^{2}}{2\sigma_{y}^{2}}\right]}+\rho_{\text{g0}}\
.$ (36)
where $y_{0}=10$ and $\sigma_{y}=2$, and the rest of parameters are same as in
the 1.5D test.
We note that when including momentum correction, there is an additional
contribution $v_{\text{d,dif}}$ to the conserved dust momentum. This leads to
two possible initial settings. One is to make the initial conserved momentum
to be zero. By the conversion relation (31), the primitive velocity thus
equals to $-v_{\text{d,dif}}$. Alternatively, one can choose the primitive
velocity to be zero (i.e., zero mean dust velocity), so that the conserved
momentum becomes $\rho v_{\text{d,dif}}$ (i.e., non-zero mean dust momentum).
This ambiguity reflects that the initial condition itself is physically
unrealistic to build up without involving additional source terms. As a test
problem, we choose the latter, which we consider to be physically more natural
and intuitive (the alternative choice would lead to different interpretable
outcomes that we omit here for brevity). Note that without momentum
correction, we set $v_{\text{d}}=0$ so that the two setups share the same
initial condition in primitive variables.
Table 2: Problem setup for 1D, 1.5D and 2D dust diffusion tests | Correction555The correction terms by default include both $\boldsymbol{\Pi}_{\text{dif}}$ and $\partial{(\rho_{\text{d}}\boldsymbol{v}_{\text{d,dif}})}/\partial{t}$. | $v_{\text{g},x}$ | $v_{\text{g},y}$ | $v_{\text{d},x}$ | $v_{\text{d},y}$
---|---|---|---|---|---
1D | No | 0 | 0 | 0 | 0
Yes | 0 | 0 | $v_{\text{d,dif},x}$666The diffusion velocity $v_{\text{d,dif}}$ is calculated by Equation (8). | 0
1.5D | No | 0 | 1 | 0 | 1
Yes | 0 | 1 | $v_{\text{d,dif},x}$ | 1
2D | No | 1 | 1 | 1 | 1
Yes777In the 2D cases, we have two corrections tests, one is with only $\boldsymbol{\Pi}_{\text{dif}}$, and the other is with both $\boldsymbol{\Pi}_{\text{dif}}$ and $\partial{(\rho_{\text{d}}\boldsymbol{v}_{\text{d,dif}})}/\partial{t}$. | 1 | 1 | 1+$v_{\text{d,dif},x,\text{2D}}$ | 1+$v_{\text{d,dif},y,\text{2D}}$
In all these tests, dust back-reaction is included. We use an isothermal
equation of state with sound speed $c_{\text{s,iso}}=1$, applying periodic
boundary conditions. We use the Piecewise Linear Method (PLM) spatial
reconstruction for both gas and dust, and a CFL number of 0.3. The
“VL2-Implicit” drag integrator is used to calculate the mutual drags.
#### 3.2.2 Results
Figure 3: 1D (top six panels) and 1.5D (bottom six panels) dust diffusion
tests without (“No Correction, NoCo”, dashed lines)/with (“Correction, Co”,
solid lines) momentum corrections. The first and the third rows show gas
density ($\rho_{\text{g}}$) and velocities ($v_{\text{g},x},\;v_{\text{g},y}$)
at different times ($t=0.0,\;1.0,\;3.0$ and $5.0$). Note that the gas density
and $x$-velocity in the first and third panels are identically one and zero in
the “NoCo” cases with these dashed lines embedded below the black solid line.
The second and the fourth rows show dust concentration
($\rho_{\text{d}}/\rho_{\text{g}}$) and dust velocities
($v_{\text{d},x},\;v_{\text{d},y}$) at different times. Similarly, the dust
$x$-velocity in the “NoCo” case is zero with the corresponding dashed lines
embedded below the black solid line. Figure 4: 2D dust diffusion tests without
correction (top eight panels), with only $\boldsymbol{\Pi}_{\text{dif}}$
correction (middle eight panels), and with both
$\boldsymbol{\Pi}_{\text{dif}}$ and
$\partial(\rho_{\text{d}}\boldsymbol{v}_{\text{d,dif}})/\partial t$
corrections (bottom eight panels). From left to right, the panels are at time
$t=0.0,\;1.0,\;3.0$ and $5.0$. The first, the third and the fifth rows are for
gas density ($\rho_{\text{g}}$), while the second, the fourth and the sixth
rows are for dust concentration ($\rho_{\text{d}}/\rho_{\text{g}}$). The black
(white) lines represent the velocity streamlines of gas (dust).
In the 1D tests shown in the top eight panels of Figure 3, we note that the
correction term $\boldsymbol{\Pi}_{\text{dif}}$ is zero, thus only the time
derivative term
$\frac{\partial{(\rho_{\text{d}}\boldsymbol{v}_{\text{d,dif}})}}{\partial{t}}$
is effective. When the correction is not included, dust proceeds as normal
concentration diffusion, whereas the gas is totally intact. This is unphysical
because turbulent mixing is a two-way process that not only mixes dust with
gas, but should also mix gas with dust. When the momentum correction is
included, we see that gas density exhibits a central deficit and two outside
bumps. This is essentially the outcome of gas being dragged by the outward
diffusion flow of dust, as can be seen in the central region of the middle
panel. The additional structures in the gas act to slow down dust
concentration diffusion, and we can see that without incorporating momentum
correction, dust diffuses more than 2 times more rapidly.
When adding a transverse velocity in the 1.5D test, one should expect
identical results as in 1D except for a velocity shift in the $y$-direction.
However, without momentum correction, we see in the two right panels in the
bottom of Figure 3 that the system develops artificial variations in $v_{y}$.
This is because concentration diffusion changes the dust density profile, but
without properly altering the momentum profile. When momentum corrections are
included, we see that dust and gas momenta are properly advected to ensure
Galilean invariance.
In the 2D tests shown in Figure 4, similar to the 1D case, incorporating
momentum correction leads to the physical behavior where the gas density
develops a dip in the center surrounded by an outside ring, which would be
absent without the correction. To test Galilean invariance, we add a
background advection velocity along the diagonal direction. We run two cases
with momentum correction, one includes only the
$\boldsymbol{\Pi}_{\text{dif}}$ term, and the other includes both the
$\boldsymbol{\Pi}_{\text{dif}}$ and
$\partial(\rho_{\text{d}}\boldsymbol{v}_{\text{d,dif}})/\partial t$ terms. We
see that the circular gas and dust density patterns are well preserved when
both terms are included, whereas artificial density structures show up without
the correction or when including only the $\Pi_{\text{dif}}$ term. These
artificial density patterns are due to artifacts in velocity variations
resulting from the violation of Galilean invariance. Therefore, both
correction terms are essential to ensure Galilean invariance. Overall, these
tests demonstrate the importance of properly incorporating the momentum
correction terms to for a consistent treatment of dust concentration
diffusion, especially when dust feedback is taken into account.
### 3.3 Streaming Instability
The streaming instability (SI) is a stringent test for two-way gas drag, and
its nonlinear evolution with strong dust clumping represents a further test of
code capability to handle sharp discontinuities. We adopt the linear tests
given by Youdin & Johansen (2007) and non-linear runs from Johansen & Youdin
(2007), which has become the standard test problem for codes with particle-
based treatment of dust (Bai & Stone, 2010c), and more recently for multifluid
dust as well (Benítez-Llambay et al., 2019). We will further extend the non-
linear tests to cylindrical coordinates (Section 3.3.4), and to incorporate
mesh refinement (Section 3.4.1).
#### 3.3.1 Shearing Box Equations and Equilibrium State
Most of our SI tests are carried out in the local shearing box framework,
which follows a local patch of a disk at some fiducial radius $r_{0}$ in the
corotating frame with orbital frequency $\Omega_{0}=\Omega\left(r_{0}\right)$
(Goldreich & Lynden-Bell, 1965; Hawley et al., 1995). The equations of gas and
dust are written in a Cartesian coordinate system $\left(x,y,z\right)$ for
radial, azimuthal and vertical directions with Coriolis and centrifugal source
terms (Stone & Gardiner, 2010). The unit vectors along this three directions
are denoted as $\left(\hat{i},\hat{j},\hat{k}\right)$, respectively. We do not
consider viscosity, diffusion, magnetic field and self-/vertical gravity.
Adopting an isothermal equation of state with isothermal sound speed
$c_{\text{s}}$, the continuity and momentum equations of gas and dust are:
$\displaystyle\frac{\partial\rho_{\text{g}}}{\partial
t}+\nabla\cdot\left(\rho_{\text{g}}\boldsymbol{v}_{\text{g}}\right)=0\ ,$ (37)
$\displaystyle\frac{\partial\left(\rho_{\text{g}}\boldsymbol{v}_{\text{g}}\right)}{\partial
t}+\nabla\cdot\left(\rho_{\text{g}}\boldsymbol{v}_{\text{g}}\boldsymbol{v}_{\text{g}}+P_{\text{g}}\mathsf{I}\right)=$
(38) $\displaystyle 2\rho_{\text{g}}q\Omega_{0}^{2}x\hat{\mathbf{i}}$
$\displaystyle-2\rho_{\text{g}}\Omega_{0}\hat{\mathbf{k}}\times\boldsymbol{v}_{\text{g}}-\sum^{n}_{k=1}\rho_{\text{d},k}\frac{\boldsymbol{v}_{\text{g}}-\boldsymbol{v}_{\text{d},k}}{T_{\text{s},k}}\
,$
$\displaystyle\frac{\partial\rho_{\text{d},k}}{\partial
t}+\nabla\cdot\left(\rho_{\text{d},k}\boldsymbol{v}_{\text{d},k}\right)=0\ ,$
(39)
$\displaystyle\frac{\partial\left(\rho_{\text{d},k}\boldsymbol{v}_{\text{d},k}\right)}{\partial
t}+\nabla\cdot$
$\displaystyle\left(\rho_{\text{d},k}\boldsymbol{v}_{\text{d},k}\boldsymbol{v}_{\text{d},k}\right)=$
(40) $\displaystyle
2\rho_{\text{d},k}q\Omega_{0}^{2}x\hat{\mathbf{i}}-2\rho_{\text{d},k}\Omega_{0}\hat{\mathbf{k}}$
$\displaystyle\times\boldsymbol{v}_{\text{d},k}+\rho_{\text{d},k}\frac{\boldsymbol{v}_{\text{g}}-\boldsymbol{v}_{\text{d},k}}{T_{\text{s},k}}\
.$
where $q=-d\ln{\Omega}/dr$ is the shear rate and $q=3/2$ for Keplerian disks
as we adopt.
Ignoring vertical gravity (i.e., unstratified disk), dust and gas can achieve
the so-called Nakagawa-Sekiya-Hayashi (NSH) equilibrium Nakagawa et al.
(1986). The equilibrium is a force balance between background pressure
gradient, centrifugal force, Coriolis force and mutual aerodynamic drags
between gas and dust in the horizontal plane. Here we consider a Keplerian
disk with angular speed $\Omega_{0}=\Omega_{\text{K}}$. In the absent of dust,
gas rotates slower than the Keplerian speed by a small amount $\eta
v_{K}\equiv\Pi c_{\text{s}}$ due to the background pressure gradient, and
$\Pi$ is $\lesssim 0.1$ under typical disk conditions. $\eta$ represents the
strength of radial gas pressure gradient:
$\eta\equiv\frac{1}{2}\frac{\ln P_{\text{g}}}{\ln
r}\left(\frac{h}{r}\right)^{2}=\frac{1}{2}\frac{\ln P_{\text{g}}}{\ln
r}\left(\frac{c_{\text{s}}}{v_{\text{k}}}\right)^{2}\ ,$ (41)
The original NSH solution considered a single dust species. Generalized to
multiple dust species with different stopping times, the velocities of gas and
dust in the multi-species equilibrium are given by (Benítez-Llambay et al.,
2019), and also see Tanaka et al. (2005):
$\displaystyle v_{\text{g0},x}$ $\displaystyle=2\eta
v_{\text{K}}\frac{\mathcal{A}}{\mathcal{A}+\mathcal{B}}\ ,\quad
v_{\text{d0},k,x}=\frac{v_{\text{g0},x}+2St_{k}v^{\prime}_{\text{g0},y}}{1+St_{k}^{2}}\
,$ (42) $\displaystyle v^{\prime}_{\text{g0},y}$ $\displaystyle=-\eta
v_{\text{K}}\frac{\mathcal{B}}{\mathcal{A}+\mathcal{B}}\ ,\quad
v^{\prime}_{\text{d0},k,y}=\frac{v^{\prime}_{\text{g0},y}-St_{k}v_{\text{g0},x}/2}{1+St_{k}^{2}}\
,$
where $St_{\text{k}}\equiv\Omega_{\text{K}}T_{\text{s},k}$ is the
dimensionless stopping time of k-th dust species, the $y-$ velocities with a
prime have Keplerian shear subtracted
$v^{\prime}_{\text{g}0,y}=v_{\text{g}0,y}+(3/2)\Omega_{K}x$,
$v^{\prime}_{\text{d}0,k,y}=v_{\text{d}0,k,y}+(3/2)\Omega_{K}x$, and
$\mathcal{A}=\sum_{k=1}^{n}\frac{\epsilon_{k}St_{k}}{1+St_{k}^{2}}\
,\quad\mathcal{B}=1+\sum_{k=1}^{n}\frac{\epsilon_{k}}{1+St_{k}^{2}}\ .$ (43)
In our numerical setup, we add an additional outward force
$\boldsymbol{f}\equiv 2\rho_{\text{g}}\eta v_{\text{K}}\Omega_{0}\hat{i}$ on
the gas to mimic the radial pressure gradient. Note that it differs from Bai &
Stone (2010c) and Benítez-Llambay et al. (2019), who add this force on the
dust component. The two approaches are equivalent except for a constant
velocity shift. Realize the exact analytic solution of the multi-species NSH
equilibrium is straightforward when using explicit integrators, and our
numerical implementation in Section 2.3.3 ensures that such exact solution can
be realized using the semi-implicit and implicit drag integrators as well.
#### 3.3.2 Linear SI Modes and Growth Rates
The NSH equilibrium is subject to the SI (Youdin & Goodman, 2005). The growth
rate $s$ of the SI is a function of initial dust-to-gas ratio (or metallicity)
$\epsilon_{0}\equiv\rho_{\text{d,0}}/\rho_{\text{g,0}}$, dimensionless
stopping time $St$ and two dimensionless wavenumbers $K_{x}\equiv k_{x}\eta
r_{0}=k_{x}\eta c_{\text{s}}/\Omega_{\text{K}}$ and $K_{z}\equiv k_{z}\eta
r_{0}=k_{z}\eta c_{\text{s}}/\Omega_{\text{K}}$:
$s=s(\epsilon_{0},St,K_{x},K_{z})$. In this work, we choose the Lin-A and
Lin-B tests in Table 1 of Youdin & Johansen (2007), which consist of gas and
one dust species, and the Lin-3 test in Section 3.5 of Benítez-Llambay et al.
(2019), which consists of gas and two dust species.
The numerical setups of the linear tests are similar to those of Youdin &
Johansen (2007), Bai & Stone (2010c) and Benítez-Llambay et al. (2019). The
numerical domain is a square box along $x$ and $z$ directions with
$-L_{x}/2\leq x\leq L_{x}/2$, $-L_{z}/2\leq z\leq L_{z}/2$, and
$L_{x}=L_{z}=1$. We choose $\Omega_{0}=1.0$ and $\eta
v_{\text{K}}=0.05\;c_{\text{s}}$. On top of the multi-species NSH equilibrium
from Equation (42), we add perturbations on densities and velocities of both
gas and dust of the following form, following Youdin & Johansen (2007):
$\displaystyle\delta\rho$
$\displaystyle=[\Re(\tilde{\rho})\cos\phi-\Im(\tilde{\rho})\sin\phi]\cos(k_{z}z)\
,$ (44) $\displaystyle\delta v_{x}$
$\displaystyle=[\Re(\tilde{v}_{x})\cos\phi-\Im(\tilde{v}_{x})\sin\phi]\cos(k_{z}z)\
,$ $\displaystyle\delta v_{y}$
$\displaystyle=[\Re(\tilde{v}_{y})\cos\phi-\Im(\tilde{v}_{y})\sin\phi]\cos(k_{z}z)\
,$ $\displaystyle\delta v_{z}$
$\displaystyle=[\Re(\tilde{v}_{z})\sin\phi+\Im(\tilde{v}_{z})\cos\phi]\sin(k_{z}z)\
.$
where $\phi\equiv k_{x}x-\omega t$, and density and velocity perturbations,
denoted by $\tilde{\rho}$, $\tilde{\boldsymbol{v}}$ for gas and individual
dust species, are given by the respective eigenvectors of the specific linear
modes, which are listed in Table 1 of Youdin & Johansen (2007) and Table 4 of
Benítez-Llambay et al. (2019).
In these tests, we fit one eigenmode in the simulation box. As our box size is
fixed, the sound speed $c_{\text{s}}$ is no longer a free parameter and it
depends on the dimensionless wavenumbers ($K_{x}$ and $K_{z}$). Because of
$K_{z}=K_{x}=k_{x}\eta r_{0}=2\pi\times 0.05c_{\text{s}}/\Omega_{0}$, we
obtain $c_{\text{s}}=K_{x}\Omega_{0}/(0.05\times 2\pi)$. In the Lin-A test,
$K_{x}=K_{z}=30$, thus we have $c_{\text{s,A}}=95.49296585$. Similar setup can
be done in Lin-B and Lin-3 tests, with $c_{\text{s,B}}=19.09859317$ in Lin-B
and $c_{\text{s,3}}=159.1549431$ in Lin-3. We use all 2nd-order drag
integrators, with PLM and Piecewise Parabolic Method (PPM) for spatial
reconstructions on both gas and dust, and the HLLE Riemann solver. Numerical
resolution spans from 8 to 256 cells per wavelength in all each directions,
with the CFL number being 0.3.
Figure 5: Fitted averaging growth rates (normalized as $s/\Omega_{0}$) by
measuring the mean kinetic energy $E_{\text{kinetic}}$ of gas and dust in
Lin-A, Lin-B and Lin-3 tests of SI. Results are shown as a function of the
number of grids per wavelength ($N/\lambda$). The analytic growth rates are
marked with blacked dotted lines. Different drag integrators are marked with
different colors. The solid (dashed) lines represent the runs with PPM (PLM)
spatial reconstruction.
We measure the growth rate of the kinetic energy $E_{\text{kinetic}}$ of both
gas and dust in the linear tests. Because the initial kinetic energy is
dominated by radial drift, we expect the temporal variation of kinetic energy
to grow as $\delta E_{\text{kinetic}}\propto\exp(s\;t)$. We fit the spatial
standard deviation $<\delta E_{\text{kinetic}}^{2}>^{\frac{1}{2}}$ over the
whole mesh as a function of time, and show the fitted growth rate
$s/\Omega_{0}$ in Figure 5 as a function of resolution in Lin-A, Lin-B and
Lin-3 tests. As these test problems are non-stiff, different drag integrators
generally show similar results. With PPM reconstruction, 16 cells per
wavelength is generally sufficient to accurately capture the growth in all
three tests. When using the PLM reconstructions, on the other hand, about
128-256 cells per wavelength is needed for similar accuracy, and the
requirement for Lin-B is the most stringent. When compared with the measured
growth rates in Table 5 of Benítez-Llambay et al. (2019), it appears that we
need more grid points to achieve similar accuracy in FARGO3D. On the other
hand, the level of accuracy that we achieve is similar to those obtained in
other finite volume method code, Athena (Bai & Stone, 2010c) and PLUTO
(Mignone et al., 2019), with PPM spatial reconstructions. We thus attribute
the difference primarily to the different nature of the base code. On the
other hand, we will show that our code show similar outcomes in the nonlinear
regime at a given resolution.
#### 3.3.3 The SI in the Non-Linear Regime
Figure 6: Dust densities of AB test (top) and BA test (bottom) with different
spatial resolutions at $t=40\;\Omega_{0}^{-1}$ for AB test and
$t=800\;\Omega_{0}^{-1}$ for BA test. Figure 7: The dust CDFs of AB test
(top) and BA test (bottom), similar to Figure 10 of Benítez-Llambay et al.
(2019). The left panels are the CDFs calculated by counting the number of
cells whose dust density exceeds certain threshold. The right panels are the
CDFs calculated by additional weighting by dust density. The color shaded
regions are the temporal standard deviations based on many snapshots (from
$30\;\Omega_{0}^{-1}$ to $40\;\Omega_{0}^{-1}$ for AB test, and from
$600\;\Omega_{0}^{-1}$ to $800\;\Omega_{0}^{-1}$ for BA test). The initial
dust densities are $\rho_{\text{d},0}=1.0$ and $\rho_{\text{d},0}=0.2$ for AB
and BA test, respectively. Table 3: Turbulence Properties of AB and BA test
with different resolutions
Run | $Ma_{x}$ | $Ma_{y}$ | $Ma_{z}$ | $\mathscr{Re}/\mathscr{Re}_{\text{NSH}}$ | $v_{\text{d,drift}}/v_{\text{d,drift,NSH}}$
---|---|---|---|---|---
AB-$128^{2}$ | $1.39(06)\times 10^{-2}$ | $8.97(55)\times 10^{-3}$ | $1.18(08)\times 10^{-2}$ | $2.15(22)$ | $1.74(07)$
AB-$256^{2}$ | $1.46(05)\times 10^{-2}$ | $9.30(49)\times 10^{-3}$ | $1.10(03)\times 10^{-2}$ | $2.56(10)$ | $2.03(08)$
AB-$512^{2}$ | $1.37(05)\times 10^{-2}$ | $7.32(39)\times 10^{-3}$ | $1.07(06)\times 10^{-2}$ | $2.66(12)$ | $2.19(07)$
AB-$1024^{2}$ | $1.24(02)\times 10^{-2}$ | $6.01(24)\times 10^{-3}$ | $8.95(28)\times 10^{-3}$ | $2.55(07)$ | $2.15(05)$
AB-$2048^{2}$ | $1.10(02)\times 10^{-2}$ | $4.96(17)\times 10^{-3}$ | $7.61(15)\times 10^{-3}$ | $2.38(02)$ | $2.07(02)$
AB-$4096^{2}$ | $1.07(01)\times 10^{-2}$ | $4.93(14)\times 10^{-3}$ | $7.26(14)\times 10^{-3}$ | $2.34(04)$ | $2.03(02)$
BA-$64^{2}$ | $1.03(15)\times 10^{-2}$ | $1.69(14)\times 10^{-2}$ | $3.92(33)\times 10^{-2}$ | $0.74(08)$ | $0.74(08)$
BA-$128^{2}$ | $1.46(23)\times 10^{-2}$ | $2.03(20)\times 10^{-2}$ | $4.92(26)\times 10^{-2}$ | $0.61(07)$ | $0.63(06)$
BA-$256^{2}$ | $1.78(11)\times 10^{-2}$ | $2.09(15)\times 10^{-2}$ | $4.91(45)\times 10^{-2}$ | $0.54(08)$ | $0.58(07)$
BA-$512^{2}$ | $1.59(21)\times 10^{-2}$ | $2.08(14)\times 10^{-2}$ | $5.02(16)\times 10^{-2}$ | $0.62(06)$ | $0.65(05)$
BA-$1024^{2}$ | $1.57(15)\times 10^{-2}$ | $2.07(07)\times 10^{-2}$ | $4.83(13)\times 10^{-2}$ | $0.58(04)$ | $0.62(04)$
BA-$2048^{2}$ | $1.61(15)\times 10^{-2}$ | $2.12(20)\times 10^{-2}$ | $4.95(24)\times 10^{-2}$ | $0.61(05)$ | $0.64(05)$
* 1
The number in parenthesis quotes the $1\sigma$ uncertainty of the last two
digits.
The non-linear SI runs are also carried in the 2D shearing box in $x-z$. We
select two non-linear tests of the SI, namely, AB and BA test from Johansen &
Youdin (2007). In AB test, the domain is $-\eta r_{0}\leq x\leq\eta r_{0}$,
$-\eta r_{0}\leq z\leq\eta r_{0}$, with parameters being $\epsilon_{0}=1.0$,
$St=0.1$. In BA test, the domain is $-20\;\eta r_{0}\leq x\leq 20\;\eta
r_{0}$, $-20\;\eta r_{0}\leq z\leq 20\;\eta r_{0}$, with parameters being
$\epsilon_{0}=0.2$ and $St=1.0$. The parameters and simulation setups are
similar to Table 1 of Johansen & Youdin (2007), Section 5 of Bai & Stone
(2010c) and Section 3.5.6 of Benítez-Llambay et al. (2019). We use the
“VL2-Implicit” drag integrator, the Roe Riemann solver, with PPM
reconstruction for the gas, and PLM reconstructions for the dust, in order to
more robustly follow the dramatic variation of dust density over space in the
non-linear stage with strong particle clumping. We use an isothermal equation
of state with sound speed is $c_{\text{s}}=1.0$, and the initial gas density
is $\rho_{\text{g}}=1.0$. We also set a density floor
$\rho_{\text{d,floor}}=10^{-6}$ on dust. Gas viscosity and dust diffusion are
not included. The simulations are initiated from the NSH equilibrium, on top
of which we add white-noise velocity perturbations with an amplitude of
$<A>\sim 0.02c_{\text{s}}$ on both gas and dust.
The saturated state of AB and BA tests, following the evolution of
$40\;\Omega_{0}^{-1}$ and $800\;\Omega_{0}^{-1}$ are shown in Figure 6, for
simulations with different resolutions. They are to be compared with with
Figure 8 and 9 of Benítez-Llambay et al. (2019). AB test is characterized by
the development of thin filaments and cavitation towards smaller scales at
higher resolution, well consistent with results in previous works (Johansen &
Youdin, 2007; Bai & Stone, 2010c; Benítez-Llambay et al., 2019). In BA test,
the system develops long dusty stripes and valleys nearly aligned with the $z$
direction and tilted towards the $x$ direction. Different from AB test, these
general features are similar at all resolutions, again consistent with
previous studies.
We further investigate the convergence of dust clumping by calculating the
cumulative dust density distributions (CDFs). Following the same procedures
described in Section 3.5.6 of Benítez-Llambay et al. (2019), there are two
ways of calculating the CDFs. One is based on the probability that local dust
density exceeds certain threshold, obtained by counting the number of cells
whose dust density exceeds the threshold. The other reflects the probability
where a particle resides in regions whose particle density exceeds a certain
threshold, obtained by weighting the first probability by local dust density.
We refer to the two CDFs as obtained by counting cell numbers, and counting
dust density, respectively. The overall results are shown in Figure 7.
The CDFs obtained by counting cell numbers are very similar to those obtained
by Benítez-Llambay et al. (2019) for both AB and BA. The CDFs of AB test
systematically vary with resolution at both the low-density and high-density
ends, in line with the non-convergent behaviors revealed in Figure 6. The CDFs
of BA test show convergence at the low-density end up to $P\sim 10^{-3}$,
while show more significant clumping at higher resolution (BA-$1024^{2}$ and
BA-$2048^{2}$).
Our CDFs by counting dust density, on the other hand, show more clumping than
that in FARGO3D. The clumping is also more significant than that obtained in
the particle module of Athena (see Figure 6 of Bai & Stone, 2010c). This might
be related to the higher-order drag integrators adopted here compared to
FARGO3D. We also note that in the original Athena code, some artificial
reduction of dust feedback was applied in strong dust clumps to alleviate the
stiffness in the system that is circumvented in our approach. We thus leave
this as an open issue. We also anticipate that our dust fluid module generally
finds most of applications in regimes with $St<1$, instead of $St\gtrsim 1$ as
in BA test.
We also examine the properties of gas turbulence triggered by the SI. We
calculate the turbulent Mach numbers along 3 directions ($x$, $y$ and $z$),
the Reynolds stress and mean radial drift velocities of dust fluids. The mach
number is calculated by
$Ma\equiv\sqrt{\langle(v_{\text{g}}-\overline{v_{\text{g}}})^{2}\rangle}/c_{\text{s}}$.
The Reynolds stress is calculated by:
$\mathscr{Re}\equiv\langle\rho_{\text{g}}v_{\text{g},x}(v_{\text{g},y}-v_{\text{K}})\rangle$.
The mean radial drift velocity is computed by dividing the mean dust momentum
over mean density. These properties are all spatially and time averaged,
indicated by angle brackets, based on many snapshots ($30\;\Omega_{0}^{-1}\sim
40\;\Omega_{0}^{-1}$ in AB test, $600\;\Omega_{0}^{-1}\sim
800\;\Omega_{0}^{-1}$ in BA test), and the results are shown in Table 3. These
diagnostic quantities are in broad agreement with the values obtained in
Johansen & Youdin (2007); Bai & Stone (2010c), all saturated into highly
subsonic, anisotropic turbulence, with enhanced radial drift and Reynolds
stress in AB test and reduced radial drift and Reynolds stress in BA test.
#### 3.3.4 Global Curvilinear Run of BA test
In order to test our multifluid module in curvilinear coordinates, we run a
global unstratified SI test in cylindrical coordinates $(r,\phi,z)$, similar
to earlier investigations (Kowalik et al., 2013; Mignone et al., 2019). We
choose to adopt the parameters close to BA test here, which is less demanding
in resolution and has better convergence properties. The computational domains
in three directions are $r\in[0.2,2.6]$, $\phi\in[0,2\pi]$ and
$z\in[-0.15,0.15]$. The numerical resolutions are $4096\times 4\times 512$
cells along $r$, $\phi$ and $z$, where we use a reduced $\phi-$resolution to
preserve axisymmetry.
Figure 8: Different snapshots of dust densities (scaled by $r^{-0.5}$) in the
global unstratified BA test in the cylindrical coordinate. From top to bottom,
the panels are at $t=100\;\Omega_{0}^{-1}$, $200\;\Omega_{0}^{-1}$,
$400\;\Omega_{0}^{-1}$ and $800\;\Omega_{0}^{-1}$, where $\Omega_{0}$ is the
orbital frequency at $r_{0}\equiv 1$. The strength of gas pressure gradient is
$\eta=0.0075$. The top and right axis are scaled by $\eta r_{0}$ for
reference. The effective resolution is the same as the BA-$512^{2}$ in Figure
6.
We set the central star mass $GM=1$, and set the gas radial density profile to
be $\rho_{\text{g}}(r)=\rho_{0}(r/r_{0})^{-0.5}$ with $\rho_{0}\equiv 1$ at
$r_{0}\equiv 1$. We adopt a vertically-isothermal equation of state with
$P(r)=c_{\text{s}}(r)^{2}\rho_{\text{g}}(r)$. The sound speed is chosen so
that the disk aspect ratio is $h/r=c_{s}/v_{K}=0.1$ at all radii, which gives
$c_{\text{s}}=0.1(r/r_{0})^{-0.5}$. From Equation (41), we obtain
$\eta=0.0075$ ($\Pi=0.075$). The effective resolutions along $r$ and $z$
directions are $12.8/\eta r_{0}$, same as BA-$512^{2}$. The initial dust
density follows that of the gas with a uniform dust-to-gas density ratio
$\epsilon_{0}=0.2$. The Stokes number of dust is
$St=T_{\text{s}}\Omega_{\text{K}}=1.0$. The initial velocities of both gas and
dust are set by the NSH equilibrium. Periodic boundary conditions are applied
along $\phi$ and $z$ directions. The radial boundary condition is fixed by the
NSH solution. To minimize of the unphysical wave reflection, we apply wave
damping zones near the inner and outer radial boundaries (de Val-Borro et al.,
2006, 2007), located at $0.20\leq r\leq 0.32$ and $2.44\leq r\leq 2.60$, where
gas and dust density and velocities (represented by $x$) are relaxed according
to:
$\frac{\mathrm{d}x}{\mathrm{~{}d}t}=-\frac{x-x_{\text{init}}}{\tau}R(r)\ ,$
(45)
where $x_{\text{init}}$ is the initial value and $\tau$ is damping rate. We
adopt $\tau=1$ in our simulation, and $R(r)$ is a smoothing parabolic function
that transiting from $0$ in the active zones to $1$ in the ghost zones. We
also use the orbital advection algorithm (Masset, 2000, 2002, FARGO algorithm)
to reduce truncation error. At the beginning, we add small velocity
perturbations in white noise with an amplitude of $0.02c_{\text{s}}$ to seed
the instability.
Figure 8 shows different snapshots of dust density in the global simulations
of this global BA test. We can clearly see the progressive development of the
SI from small radii to large radii, as the inner region has shorter dynamical
time. The simulation reaches saturated state over the entire domain after
about $t=600\;\Omega_{0}^{-1}$, where $\Omega_{0}=\Omega_{\text{K}}$ at
$r=r_{0}$, and we see the characteristic long dusty stripes and valleys with
the maximum of $\rho_{\text{d}}r^{0.5}$ significantly amplified by a factor of
$\gtrsim 100$.
More quantitatively, we have also examined the dust-to-gas ratio $\epsilon$
and the dust CDFs of the global run, and compared them with those from the
local BA-$512^{2}$ shown in Figure 9. The maximum $\epsilon$ in the global
simulation is higher than that in the local simulation by a small margin. Note
that due to the reduction of mean radial drift speed in BA test as the SI
saturates, and that the SI develops faster in the inner region than the outer
region, the mean dust-to-gas ratio $\epsilon$ in the global run increases with
time and reaches about $0.3$ (rather than the initial value of $0.2$) at
$t=800\;\Omega_{0}^{-1}$. This is likely the main cause of the stronger dust
clumping found in the global run. On the other hand, from the dust CDFs, we
see that the dust density distribution in the global and local runs converge
within the error bars at relatively high densities large
$\rho_{\text{threshold}}$ ($\gtrsim 5\;\rho_{\text{d}}$ both by counting
numbers and counting density), suggesting that dust clumping is well captured
in both local and global simulations. However, there are some deviations at
relatively small dust density. The cause of this deviation is unknown and may
require further investigations beyond the scope of this work, but we speculate
it may be related to a combination of higher pressure gradient $\Pi=0.075$
instead of $0.05$, higher mean $\epsilon$, and the global nature of the
simulation.
Figure 9: Left: The evolution of maximum and mean dust-to-gas ratio $\epsilon$
over time in the global and the local BA-$512^{2}$. The calculation of
$\epsilon$ in the global run are conducted over $r\in[0.4,2.4]$. Right: Dust
CDFs of the global and the local BA-$512^{2}$, similar to Figure 7. The dust
CDFs of the global run is calculated by the normalized dust density
$\rho_{\text{d}}r^{0.5}$ on $r\in[0.4,2.4]$. The color shaded regions are the
temporal standard deviations from $600\;\Omega_{0}^{-1}$ and
$800\;\Omega_{0}^{-1}$.
### 3.4 Mesh Refinement
In this subsection, we present additional tests to demonstrate the
compatibility of our multifluid dust module with static and adaptive mesh
refinement (SMR/AMR). Following the convention, here the root mesh is called
level 0, and each level of refinement doubles the resolution and is called
level 1, 2, etc.
#### 3.4.1 SMR Run of AB test
Figure 10: Snapshots of dust densities in the SMR test of AB test with two
levels of refinement. From top to bottom, from left to right, the panels are
at $t=5\;\Omega_{0}^{-1}$, $10\;\Omega_{0}^{-1}$, $15\;\Omega_{0}^{-1}$ and
$40\;\Omega_{0}^{-1}$. The edges of each meshblock are marked by black solid
lines, and meshblocks with different levels are labeled in the top-left panel.
Each meshblock contains $64^{2}$ cells, and there are 8, 16 and 64 meshblocks
in levels 0, 1 and 2.
We first rerun the AB test of the SI (see Section 3.3.3), but with two levels
of SMR. The resolution of the root mesh is $256^{2}$, and each level of
refinement doubles the resolution in the central region along the $z$
direction. Because AB test is sensitive to the amplitude of initial
perturbations, we use perturbations ten times smaller
($~{}0.002\;c_{\text{s}}$) than those in uniform runs. The results are shown
in Figure 10. As noted earlier, the outcome of AB test depends on resolutions.
Indeed, we see that the SI is first developed in the finest level and quickly
becomes nonlinear well within one orbital time, while the SI is developed more
slowly in coarser meshblocks, and it is not until after about $t\simeq
15\;\Omega_{0}^{-1}$ that the SI is fully developed in the entire domain. The
overall pattern in each refinement level closely resembles those shown in
Figure 6 with the same resolution ($256^{2}$ to $1024^{2}$), and there are no
abrupt features seen along coarse-fine meshblock boundaries, which testifies
the compatibility of our multifluid dust module with SMR in shearing box. We
have also examined the CDFs of this SMR run, and found that the CDFs at a
given level are largely consistent with the CDFs in the corresponding uniform-
level runs discussed earlier within the $1\sigma$ uncertainties.
#### 3.4.2 AMR Test of Kelvin-Helmholtz Instability
Figure 11: Snapshots of the KHI tests with 2 dust species with AMR. The top
(bottom) six panels are the cases without (with) dust feedback at $t=1.2$
($t=1.6$). The abbreviations of “nofb” and “fb” are for the cases without and
with feedback, respectively. From left to right, the panels are for gas, dust
species 1 and 2. The stopping times of dust are $T_{\text{s,1}}=10^{-2}$ and
$T_{\text{s,2}}=10^{-8}$. The first and the third rows are the gas and dust
densities from AMR runs with up to 3 refinement levels. The second and the
fourth rows are the relative differences between the AMR runs and the uniform
grid runs at a resolution matching the finest AMR refinement level of 3. The
edges of meshblocks in the AMR runs are also indicated in black solid lines.
We next conduct the standard test problem of the Kelvin-Helmholtz instability
(KHI) in Athena++ with AMR, exactly following the problem setup described in
Section 3.4.3 of Stone et al. (2020), but adding two dust species. The
resolution of the root mesh is $256^{2}$ with each meshblock size being
$8^{2}$, and the refinement condition is determined by
$g=\max{(|\partial_{x}v_{\text{g},y}|,|\partial_{y}v_{\text{g},x}|,|\partial_{x}v_{\text{d},n,y}|,|\partial_{y}v_{\text{d},n,x}|)}\
.\\\ $ (46)
which represents the maximum spatial velocity gradients in gas and all dust
fluids. Meshblocks with $g>0.01$ will be refined, and with $g<0.005$ will be
de-refined. We set a maximum of 3 refinement levels. We add two dust species
with stopping times $T_{\text{s,1}}=10^{-2}$, $T_{\text{s,2}}=10^{-8}$ but no
dust diffusion. We consider the cases without and with dust feedback, and the
dust-to-gas mass ratio for each species is set to unity. We use the
“RK2-Implicit” drag integrator, PLM reconstruction for both gas and dust, the
HLLC Riemann solver on gas, and a CFL number of 0.4. For comparison,
simulations with a uniform resolution of $2048^{2}$ (matching the finest
level) are also conducted.
Figure 11 shows the results of our dusty KHI tests with AMR. The first and the
third rows show gas and dust density patterns, while the second and the fourth
rows show the relative differences from the runs with uniform resolution. We
see that the strongly coupled dust with $T_{\text{s,2}}=10^{-8}$ shares
exactly the same density pattern as gas, as expected. The more marginally
coupled dust with $T_{\text{s,1}}=10^{-2}$, on the other hand, are depleted in
vortex centers, as they are relatively slow in response to the rapid vortical
motion to fill in the vortex eyes. We also see that the size of the vortices
are larger when feedback is included, as the inertia from more dust loading
would require more space for the KHI patterns to roll over.
We find there are no distinguishable differences between the AMR and uniform
grid runs, with relative differences of at most a few percent at vortex
centers in the first dust species with $T_{\text{s,1}}=10^{-2}$, due in part
to the low dust densities in there. The time step in our KHI tests is around
$5\times 10^{-5}$ in code units, which is much larger than the stopping time
of the second dust species ($T_{\text{s,2}}=10^{-8}$), making the drag
interaction highly stiff. The results again testify that our fully-implicit
methods are thus accurate and robust in these extremely stiff regimes with
AMR.
## 4 Summary and Discussion
In this paper, we describe the algorithm and implementation of a multifluid
dust module in Athena++, together with a suite of benchmark numerical tests.
The dust is treated as an arbitrary number of separate pressureless fluids,
each interacting with the gas via the aerodynamic gas drag, characterized by a
stopping time. Our development features two major advances.
First, we have provided a consistent formulation of dust concentration
diffusion. Dust concentration diffusion is commonly implemented as a diffusion
term in the dust continuity equation, which mimics the response to background
turbulence without explicitly simulating turbulence. This approach has been
shown to not conserve angular momentum in disk problems (Tominaga et al.,
2019). We further derive from a Reynolds averaging procedure the proper terms
that should be included in the momentum equation to ensure not only proper
momentum diffusion flux, but also Galilean invariance. The physically
meaningful behavior of dust concentration diffusion including dust feedback is
then illustrated from a simple test problem.
Second, we have developed two fully-implicit, second-order accurate drag
integrators, which naturally combine with the existing VL2 and RK2 integrators
in Athena++ to ensure 2nd-order accuracy in time for the composite system,
together with momentum conservation to machine precision. The integrators are
stable to highly stiff regimes not only in small dust stopping time, but also
in regimes of high dust mass loading. Our code is, to our knowledge, the first
to achieve the combination of these features. We have also implemented a
number of explicit and semi-implicit drag integrators for non-stiff
applications. In addition, we have incorporated frictional heating that can be
applied to any of the drag integrators.
The development of the multifluid dust module in Athena++, a higher-order
Godunov code, compliments the multifluid dust module in the widely used
FARGO3D code (Benítez-Llambay et al., 2019), which is Zeus-like (Stone &
Norman, 1992a, b). We conducted a large suite of code tests demonstrating code
performance, many in parallel to those done in (Benítez-Llambay et al., 2019).
In particular, we studied the SI from linear to nonlinear regimes, and the
results are generally in good agreement. We anticipate that the aforementioned
new features in our implementation represent more benefits, in addition to
better shock-capturing capabilities inherent to Godunov codes.
One of the main advantages in our multifluid dust module is its compatibility
with many of the existing functionalities and physics modules in Athena++. In
particular, our dust fluid module is compatible with static and adaptive mesh
refinement, curvilinear coordinate system including cylindrical and spherical
coordinates, shearing box and orbital advection, magnetic fields, diffusion
physics (viscosity, thermal conduction, non-ideal MHD), etc. The
implementation of the multifluid dust module thus enables a wide range of
applications involving dust dynamics, particularly related to the study of
physics, gas dynamics and observational signatures of protoplanetary disks and
planet formation, especially relevant to current and future disk observations
by ALMA, James Webb Space Telescope (JWST), Chinese Space Station Telescope
(CSST), the Next Generation Very Large Array (ngVLA) and the Square Kilometer
Array (SKA). We will also make this module publicly available in the near
future to benefit the broader astrophysical community.
There is still substantial room for further extensions of the multifluid dust
module, including dust coagulation/fragmentation (Drazkowska et al., 2019),
coupling with non-equilibrium radiative heating and cooling (Kamp & Dullemond,
2004), and self-gravity. Additionally, the drag term is exactly the same as
coupling among charged and neutral species in weakly ionized plasmas, e.g.,
O’Sullivan & Downes (2006), and the coupling term can be extremely stiff in
the strong coupling regime. Thus, our code can also be potentially extended to
accurately handle weakly ionized plasmas from multifluid to strong coupling
regimes. These directions will be considered in future works.
We thank Pablo Benítez-Llambay, Leonardo Krapp, Hui Li, Shengtai Li, Ruobing
Dong, Rixin Li, and Shangfei Liu for helpful discussions. This work is
supported by the National Key R&D Program of China No. 2019YFA0405100, and the
China Manned Space Project with NO. CMS-CSST-2021-B09.
## Appendix A Galilean invariance in momentum diffusion
In this Appendix, we prove the Galilean invariance of our dust concentration
diffusion formulation. In doing so, it would be much easier to recast the dust
momentum equation into an Euler-like equation. By applying the dust continuity
equation (4) and after some algebra, we arrive at
$\displaystyle\frac{\partial}{\partial
t}(\boldsymbol{v}_{\text{d}}+\boldsymbol{v}_{\text{d,dif}})+[(\boldsymbol{v}_{\text{d}}+\boldsymbol{v}_{\text{d,dif}})\cdot\nabla](\boldsymbol{v}_{\text{d}}+\boldsymbol{v}_{\text{d,dif}})=\frac{1}{\rho_{\text{d}}}\nabla\cdot(\rho_{\text{d}}\boldsymbol{v}_{\text{d,dif}}\boldsymbol{v}_{\text{d,dif}})+\frac{\boldsymbol{v}_{\text{g}}-\boldsymbol{v}_{\text{d}}}{T_{s}}\
,$ (A1)
Let us consider a different frame which moves at a constant speed
$\boldsymbol{v}_{0}$, where physical quantities are denoted with a prime ′.
Obviously, we have $\rho^{\prime}=\rho$,
$\boldsymbol{v}^{\prime}=\boldsymbol{v}-\boldsymbol{v}_{0}$, and
$\boldsymbol{v}^{\prime}_{\text{d,dif}}=\boldsymbol{v}_{\text{d,dif}}$. To
prove that the equation is Galilean invariant, it suffices to show that the
equations written in the new frame is exactly the same when expressed with
primed quantities.
With Equation (A1), the proof becomes quite straightforward. We first replace
$\boldsymbol{v}_{\text{d}}$ by
$\boldsymbol{v}^{\prime}_{\text{d}}+\boldsymbol{v}_{0}$,
$\boldsymbol{v}_{\text{g}}$ by
$\boldsymbol{v}^{\prime}_{\text{g}}+\boldsymbol{v}_{0}$, $\rho$ by
$\rho^{\prime}$, and $\boldsymbol{v}_{\text{d,dif}}$ by
$\boldsymbol{v}^{\prime}_{\text{d,dif}}$. We note that the right hand side
remains unchanged. Next, note that $\partial/\partial
t^{\prime}=\partial/\partial t+\boldsymbol{v}_{0}\cdot\nabla$, we can see that
terms proportional to $\boldsymbol{v}_{0}$ all cancel out, thus the form of
equation remain exactly the same as (A1) except all expressed in primed
quantities.
In the above discussion, we emphasize that if drop off the time derivative
term on $\rho_{\text{d}}\boldsymbol{v}_{\text{d,dif}}$, the Euler-like
equation would become
$\displaystyle\frac{\partial\boldsymbol{v}_{\text{d}}}{\partial
t}+[(\boldsymbol{v}_{\text{d}}+2\boldsymbol{v}_{\text{d,dif}})\cdot\nabla]\boldsymbol{v}_{\text{d}}+\frac{\boldsymbol{v}_{\text{d}}}{\rho_{\text{d}}}\nabla\cdot(\rho_{\text{d}}\boldsymbol{v}_{\text{d,dif}})=\frac{\boldsymbol{v}_{\text{g}}-\boldsymbol{v}_{\text{d}}}{T_{s}}\
.$ (A2)
which is very different from that of Equation (A1). In particular, going
through the same procedures, the presence of the third term on the left hand
side makes this equation violate the Galilean invariance.
## Appendix B Design of second order fully-implicit drag integrators
The VL2 implicit integrator is designed from the following format:
$\displaystyle\boldsymbol{M}^{(n+1)}-\boldsymbol{M}^{(n)}$
$\displaystyle=h\boldsymbol{f}\left(\boldsymbol{M}^{(n+\frac{1}{2})},\boldsymbol{W}^{(n+\frac{1}{2})}\right)$
(B1)
$\displaystyle=h\boldsymbol{f}\left[\boldsymbol{M}^{(n+1)}-\frac{h}{2}\boldsymbol{f}\left(\boldsymbol{M}^{(n+1)},\boldsymbol{W}^{(n+\frac{1}{2})}\right),\boldsymbol{W}^{(n+\frac{1}{2})}\right]$
$\displaystyle=h\boldsymbol{f}\left(\boldsymbol{M}^{(n+1)},\boldsymbol{W}^{(n+\frac{1}{2})}\right)-\frac{h^{2}}{2}\frac{\partial\boldsymbol{f}}{\partial\boldsymbol{M}}|^{(n+\frac{1}{2})}\boldsymbol{f}\left(\boldsymbol{M}^{(n+1)},\boldsymbol{W}^{(n+\frac{1}{2})}\right)$
$\displaystyle=h\boldsymbol{f}\left(\boldsymbol{M}^{(n)},\boldsymbol{W}^{(n+\frac{1}{2})}\right)+h\frac{\partial\boldsymbol{f}}{\partial\boldsymbol{M}}|^{(n)}\left(\boldsymbol{M}^{(n+1)}-\boldsymbol{M}^{(n)}\right)$
$\displaystyle\quad-\frac{h^{2}}{2}\frac{\partial\boldsymbol{f}}{\partial\boldsymbol{M}}|^{(n+\frac{1}{2})}\boldsymbol{f}\left(\boldsymbol{M}^{(n)},\boldsymbol{W}^{(n+\frac{1}{2})}\right)-\frac{h^{2}}{2}\frac{\partial\boldsymbol{f}}{\partial\boldsymbol{M}}|^{(n+\frac{1}{2})}\frac{\partial\boldsymbol{f}}{\partial\boldsymbol{M}}|^{(n)}\left(\boldsymbol{M}^{(n+1)}-\boldsymbol{M}^{(n)}\right)\
.$
This leads to the integration scheme (25) and (26). The quantities at
$n+\frac{1}{2}$ time step (denoted by a prime in Section 2.3.2) are obtained
from the first stage of the algorithm, and any first-order implicit
integration suffices (we use the backward Euler method).
The “RK2-Implicit” integrator is designed from the following format:
$\displaystyle\boldsymbol{M}^{(n+1)}-\boldsymbol{M}^{(n)}$
$\displaystyle=\frac{h}{2}\boldsymbol{f}\left(\boldsymbol{M}^{(n+1)},\boldsymbol{W}^{\prime}\right)+\frac{h}{2}\boldsymbol{f}\left(\boldsymbol{M}^{(n)},\boldsymbol{W}^{(n)}\right)$
(B2)
$\displaystyle=\frac{h}{2}\boldsymbol{f}\left(\boldsymbol{M}^{(n+1)},\boldsymbol{W}^{\prime}\right)+\frac{h}{2}\boldsymbol{f}\left[\boldsymbol{M}^{(n+1)}-h\boldsymbol{f}\left(\boldsymbol{M}^{(n+1)},\boldsymbol{W}^{(n)}\right),\boldsymbol{W}^{(n)}\right]$
$\displaystyle=\frac{h}{2}\boldsymbol{f}\left(\boldsymbol{M}^{(n+1)},\boldsymbol{W}^{\prime}\right)+\frac{h}{2}\left(\mathsf{I}-h\frac{\partial\boldsymbol{f}}{\partial\boldsymbol{M}}\bigg{|}^{\prime}\right)\boldsymbol{f}\left(\boldsymbol{M}^{(n+1)},\boldsymbol{W}^{(n)}\right)$
$\displaystyle=\frac{h}{2}\boldsymbol{f}\left(\boldsymbol{M}^{(n)},\boldsymbol{W}^{\prime}\right)+\frac{h}{2}\frac{\partial\boldsymbol{f}}{\partial\boldsymbol{M}}\bigg{|}^{(n)}\left(\boldsymbol{M}^{(n+1)}-\boldsymbol{M}^{(n)}\right)$
$\displaystyle\quad+\frac{h}{2}\left(\mathsf{I}-h\frac{\partial\boldsymbol{f}}{\partial\boldsymbol{M}}\bigg{|}^{\prime}\right)\boldsymbol{f}\left(\boldsymbol{M}^{(n)},\boldsymbol{W}^{(n)}\right)+\frac{h}{2}\left(\mathsf{I}-h\frac{\partial\boldsymbol{f}}{\partial\boldsymbol{M}}\bigg{|}^{\prime}\right)\frac{\partial\boldsymbol{f}}{\partial\boldsymbol{M}}\bigg{|}^{(n)}\left(\boldsymbol{M}^{(n+1)}-\boldsymbol{M}^{(n)}\right)\
.$
where the prime ′ on $W$ indicate the quantity is evaluated at the end of the
first stage (integrating to time step $n+1$ with a first-order implicit
scheme, i.e., backward Euler). This leads to the integration scheme (27) and
(28).
## Appendix C Explicit and Semi-implicit Drag Integrators
When the stopping time of dust is much larger than the numerical time step,
$T_{\text{s}}\gtrsim\delta t\equiv h$ and there is no strong dust mass
loading, the drag is in non-stiff regime. Here we have also implemented a
number of standard explicit and semi-implicit drag integrators in Athena++. We
omit the trivial implementation of the forward Euler method (RK1), and 2nd-
order methods are described below. For all methods, the energy equation is
updated in each stage in a way analogous to Equations (21) to (23), which we
again omit here.
### C.1 Second Order Explicit Methods
Here we document the two explicit integrators following the VL2 and RK2
integrators in Athena++, termed “VL2-Explicit” and “RK2-Explicit” in this
paper. Note that the explicit integrators usually requires the time step
$h<T_{\text{s}}$ for all dust species. The momentum update is as follows.
VL2-Explicit: We first update the system for half a time step, followed by a
full update using the midpoint values:
$\displaystyle\boldsymbol{M}^{(n+\frac{1}{2})}$
$\displaystyle=\boldsymbol{M}^{(n)}+\frac{h}{2}\boldsymbol{f}\left(\boldsymbol{M}^{(n)},\boldsymbol{W}^{(n)}\right)\
,$ (C1) $\displaystyle\boldsymbol{M}^{(n+1)}$
$\displaystyle=\boldsymbol{M}^{(n)}+h\boldsymbol{f}\left(\boldsymbol{M}^{(n+\frac{1}{2})},\boldsymbol{W}^{(n+\frac{1}{2})}\right)\
.$
RK2-Explicit: it first provides an estimate after a time step $h$ denoted by
′, followed by a correction:
$\displaystyle\boldsymbol{M}^{\prime}$
$\displaystyle=\boldsymbol{M}^{(n)}+h\boldsymbol{f}\left(\boldsymbol{M}^{(n)},\boldsymbol{W}^{(n)}\right)\
,$ (C2) $\displaystyle\boldsymbol{M}^{(n+1)}$
$\displaystyle=\frac{1}{2}\left(\boldsymbol{M}^{(n)}+\boldsymbol{M}^{\prime}\right)+\frac{1}{2}h\boldsymbol{f}\left(\boldsymbol{M}^{\prime},\boldsymbol{W}^{\prime}\right)\
.$
### C.2 Second Order Semi-implicit Methods
Here we present two semi-implicit methods, which are more robust than the
explicit methods.
Trapezoid (Crank-Nicholson Method)
The trapezoid method is derived from
$\displaystyle\boldsymbol{M}^{(n+1)}$
$\displaystyle=\boldsymbol{M}^{(n)}+\frac{1}{2}\left[h\boldsymbol{f}\left(\boldsymbol{M}^{(n)},\boldsymbol{W}^{(n)}\right)+h\boldsymbol{f}\left(\boldsymbol{M}^{(n+1)},\boldsymbol{W}^{(n)}\right)\right],$
(C3)
$\displaystyle=\boldsymbol{M}^{(n)}+h\boldsymbol{f}\left(\boldsymbol{M}^{(n)},\boldsymbol{W}^{(n)}\right)+\frac{h}{2}\frac{\partial\boldsymbol{f}}{\partial\boldsymbol{M}}|^{(n)}\left(\boldsymbol{M}^{(n+1)}-\boldsymbol{M}^{(n)}\right),$
$\displaystyle\Rightarrow\boldsymbol{M}^{(n+1)}$
$\displaystyle=\boldsymbol{M}^{(n)}+\left(\mathsf{I}-\frac{h}{2}\frac{\partial\boldsymbol{f}}{\partial\boldsymbol{M}}|^{(n)}\right)^{-1}h\boldsymbol{f}\left(\boldsymbol{M}^{(n)},\boldsymbol{W}^{(n)}\right).$
The first stage ′ of “Trapezoid” is updated by backward Euler method with $h$,
so as to be compatible with “RK2-Explicit”.
Trapezoid Backward Differentiation Formula 2 (TrBDF2)
In the TrBDF2, the momentum at the middle stage $n+\frac{1}{2}$ is calculated
by the trapezoid method with time step $h/2$, so as to be compatible with
“VL2-Explicit”. Then $\boldsymbol{M}^{(n+1)}$ is updated by backward
Differentiation Formula 2 (BDF2) method at the stage $n+1$.
$\displaystyle\boldsymbol{M}^{(n+\frac{1}{2})}$
$\displaystyle=\boldsymbol{M}^{(n)}+\left(\mathsf{I}-\frac{h}{4}\frac{\partial\boldsymbol{f}}{\partial\boldsymbol{M}}|^{(n)}\right)^{-1}\frac{h}{2}\boldsymbol{f}\left(\boldsymbol{M}^{(n)},\boldsymbol{W}^{(n)}\right),$
(C4) $\displaystyle\boldsymbol{M}^{(n+1)}$
$\displaystyle=\frac{4}{3}\boldsymbol{M}^{(n+\frac{1}{2})}-\frac{1}{3}\boldsymbol{M}^{(n)}+\frac{1}{3}\left(\mathsf{I}-h\frac{\partial\boldsymbol{f}}{\partial\boldsymbol{M}}|^{(n)}\right)^{-1}h\boldsymbol{f}\left(\boldsymbol{M}^{(n)},\boldsymbol{V}^{(n)}\right).$
## Appendix D Performance
Figure 12: Code performance measured by CPU time per cell in micron second,
shown as a function of total number of species (gas and $\mathbf{N_{d}}$
species of dust).
The two higher-order fully-implicit drag integrators VL2-Implicit and
RK2-Implicit involve solving the inverse of an $(N_{d}+1)\times(N_{d}+1)$
matrix. The cost of matrix inversion is $O[(N_{d}+1)^{3}]$. Moreover, it also
takes $O[(N_{d}+1)^{3}]$ to handle matrix multiplications, such as in
computing $\boldsymbol{\Lambda}$ in Equations (26) and (28). Such operations
would make the calculations increasingly expensive as $N_{d}$ increases, and
can become a bottleneck at sufficiently large $N_{d}$.
Here we measure code performance as a function of $N_{d}$ from the NSH
equilibrium test in shearing box. The test is run on a Intel Xeon Gold 6132
CPU with 28 cores. We use all the cores with 28 meshblocks, so that we occupy
the entire CPU (and hence its cache) to mimic more realistic situation in
large-scale simulations (note that communications in Athena++ are mostly
hidden thanks its use of tasklist and performance is more sensitive to catch
use). We use the HLLE Riemann solver for gas and the PLM reconstruction for
both gas and dust. We measure the performance in terms of the time spent to
update a single cell by an individual core. The results for different drag
integrators are shown in Figure 12, as a function of total number of species
($N_{d}+1$), which are further compared to results with gas drag turned off.
For explicit and semi-implicit integrators, we see that the drag integrators
add to very limited computational cost relative to the no-drag case. In
particular, the semi-implicit integrators that involve two matrix inversion
operations remain computationally efficient thanks to the fast analytical
solver (Krapp & Benítez-Llambay, 2020) that reduces the cost to $O(N_{d}+1)$.
The total cost increases linearly with $N_{d}$ for $N_{d}\lesssim 12$, but
gets slightly nonlinear at larger $N_{d}$. We speculate it is likely due to
heavier memory use that reduces cache performance.
For the two fully implicit solvers, we manage to improve the performance by
using the fast matrix inversion at the first integration stage, yet the more
complex matrix computation and inversion at the second stage substantially
increases the computational cost. This cost increases non-linearly with
$N_{d}$. It is relatively negligible for $N_{d}\lesssim 5$, and remains to be
minor compared to the rest of the dust integrator for $N_{d}\lesssim 10$, but
becomes rather significant for larger $N_{d}$.
## Appendix E Solutions to the Collision Tests
The mutual drags between gas and $n$ dust species can be written in the matrix
form:
$\displaystyle\frac{\partial\boldsymbol{M}}{\partial
t}=\boldsymbol{A}\boldsymbol{M}=\begin{bmatrix}-\sum^{n}_{i}\epsilon_{i}\alpha_{i}&\alpha_{1}&\alpha_{2}&\cdots&\alpha_{n}\\\
\epsilon_{1}\alpha_{1}&-\alpha_{1}&0&\cdots&0\\\
\epsilon_{2}\alpha_{2}&0&-\alpha_{2}&\cdots&0\\\
\vdots&\vdots&\vdots&\ddots&\vdots\\\
\epsilon_{n}\alpha_{n}&0&0&\cdots&-\alpha_{n}\\\
\end{bmatrix}\begin{bmatrix}\boldsymbol{M}_{\text{g}}\\\
\boldsymbol{M}_{\text{d},1}\\\ \boldsymbol{M}_{\text{d},2}\\\ \vdots\\\
\boldsymbol{M}_{\text{d},n}\\\ \end{bmatrix}\ .$ (E1)
The analytic solution of the momentum vector must take the form
$\boldsymbol{M(t)}=\sum_{0}^{n}\boldsymbol{c}_{i}\exp{(\lambda_{i}\;t)}$,
where $\boldsymbol{c}_{i}$ are the coefficient for each momentum component
determined by initial condition and $\lambda_{i}$ are the eigenvalue. The key
is to solve the eigensystem
$\boldsymbol{A}\boldsymbol{M}=\lambda\boldsymbol{M}$, yielding eigenvalues
$\lambda_{i}$ and eigenvectors $\boldsymbol{M}_{i}$. The coefficients
$\boldsymbol{c}_{i}$ are obtained by decomposing the initial condition into
the eigenvectors.
Because of momentum conservation, i.e., the drag force acted on the gas equals
to the sum of drag forces acted on dust, the matrix $A$ has an eigenvalue
$\lambda_{0}=0$ corresponding to bulk motion, and hence the coefficient
$c_{0}=v_{\text{COM}}$ is the velocity of center of mass (COM). When multiple
dust species share the same stopping time, the eigensystem can be greatly
simplified (see Table 1 of Benítez-Llambay et al. (2019)), but this is no
longer true in the general case. We use Mathematica (Wolfram, 1991) to
calculate the eigenvalues and the rest of the coefficients in Table 1.
## Appendix F Additional Numerical Tests
In this Appendix, we present additional code tests that largely reproduce the
dusty sound wave test and the dusty shock test in Benítez-Llambay et al.
(2019) to demonstrate our code performance.
### F.1 Dusty Sound Wave
To demonstrate that our multifluid dust module achieves second order accuracy
when combined with the hydrodynamic solver, we conduct the dusty sound wave
test by exactly following Section 3.2 of Benítez-Llambay et al. (2019),
originally proposed by Laibe & Price (2011, 2012). We use the PLM spatial
reconstruction, isothermal equation of state with the HLLE gas Riemann solver,
and consider both the “VL2-Implicit” and “RK2-Implict” drag integrators. The
tests are conducted in 1D, starting from a resolution of $N=64$ cells, and we
double the resolution at a time until reaching a resolution of $N=512$ cells.
We have conducted simulations for both single-species and multi-species cases,
and found excellent agreement with analytical theory. For brevity, we show in
Figure 13 the time evolution of the normalized dust and gas velocities for the
single species dust case, showing that our numerical solution perfectly
matches the analytical solution. Moreover, we measure the root mean square of
the L1 norms $[\sum_{n}(\sum|U_{n}-U_{n,{\rm ana}}|/N)^{2}]^{1/2}$ after one
wave period, similar to the approach in the linear wave test in Athena++ Stone
et al. (2020), where $U_{n}$ and $U_{n,{\rm ana}}$ are the numerical and
analytical solutions of the $n$-th variable. The variables include gas density
and velocity, and dust density and velocity, all in normalized units. We see
in the right panel of Figure 13 that our code clearly achieves second order
convergence.
Figure 13: Results from the dusty sound wave test. The left and the middle
panels are the time evolution of normalized density and normalized velocities
of gas and dust located at $x=0$, to be compared with Figure 3 of Benítez-
Llambay et al. (2019). The solid lines represent the analytical solutions, and
the triangle and circle markers are the numerical results from test runs with
256 cells. The right panel demonstrates the numerical convergence, where we
calculate the mean L1-error after one wave period, and show how it varies with
the number of grid cells using the “VL2-Implicit” and “RK2-Implicit” drag
integrators.
### F.2 Dusty Shock
Being a Godunov code, Athena++ has excellent shock capturing properties that
we demonstrate using the generalized dusty shock test presented in Section 3.3
of Benítez-Llambay et al. (2019), which is generalized from Lehmann & Wardle
(2018). We follow the same procedures and adopt identical parameters as in
Benítez-Llambay et al. (2019) to conduct two simulations with one and three
dust species (two and four species in total) on 400 grid points. PLM spatial
reconstruction and “VL2-Implicit” drag integrator are used for these tests.
The results are shown in Figure 14, which is to be compared with Figure 5 of
Benítez-Llambay et al. (2019) side by side. Note that the shocks are in steady
state thus we focus on the overall shock profile rather than the specific
shock locations. Because of dust drag, the dust profile near the shock is
connected by a precursor that is accurately reproduced, similar to that in
FARGO3D. On the other hand, Athena++ very captures the discontinuity in the
gas within neighboring cells, as opposed to $\sim 4$ cells in FARGO3D.
Figure 14: Normalized velocities (top) and densities (bottom) of dusty wave
tests with 1 (left) and 3 (right) dust species, to be directly compared with
Figure 5 of Benítez-Llambay et al. (2019). The gas profile is shown in red,
while the other colors correspond to the dust species, with the insets showing
the zoomed in profiles across the shock. Analytical solutions are shown in
solid lines, while numerical results are shown in filled circles.
## References
* Andrews (2020) Andrews, S. M. 2020, ARA&A, 58, 483, doi: 10.1146/annurev-astro-031220-010302
* Astropy Collaboration et al. (2013) Astropy Collaboration, Robitaille, T. P., Tollerud, E. J., et al. 2013, A&A, 558, A33, doi: 10.1051/0004-6361/201322068
* Bae & Zhu (2018a) Bae, J., & Zhu, Z. 2018a, ApJ, 859, 118, doi: 10.3847/1538-4357/aabf8c
* Bae & Zhu (2018b) —. 2018b, ApJ, 859, 119, doi: 10.3847/1538-4357/aabf93
* Bae et al. (2017) Bae, J., Zhu, Z., & Hartmann, L. 2017, ApJ, 850, 201, doi: 10.3847/1538-4357/aa9705
* Bai & Stone (2010a) Bai, X.-N., & Stone, J. M. 2010a, ApJ, 722, 1437, doi: 10.1088/0004-637X/722/2/1437
* Bai & Stone (2010b) —. 2010b, ApJ, 722, L220, doi: 10.1088/2041-8205/722/2/L220
* Bai & Stone (2010c) —. 2010c, ApJS, 190, 297, doi: 10.1088/0067-0049/190/2/297
* Baruteau & Zhu (2016) Baruteau, C., & Zhu, Z. 2016, MNRAS, 458, 3927, doi: 10.1093/mnras/stv2527
* Benítez-Llambay et al. (2019) Benítez-Llambay, P., Krapp, L., & Pessah, M. E. 2019, ApJS, 241, 25, doi: 10.3847/1538-4365/ab0a0e
* Benítez-Llambay & Masset (2016) Benítez-Llambay, P., & Masset, F. S. 2016, ApJS, 223, 11, doi: 10.3847/0067-0049/223/1/11
* Birnstiel et al. (2010) Birnstiel, T., Dullemond, C. P., & Brauer, F. 2010, A&A, 513, A79, doi: 10.1051/0004-6361/200913731
* Carballido et al. (2010) Carballido, A., Cuzzi, J. N., & Hogan, R. C. 2010, MNRAS, 405, 2339, doi: 10.1111/j.1365-2966.2010.16653.x
* Carrera et al. (2015) Carrera, D., Johansen, A., & Davies, M. B. 2015, A&A, 579, A43, doi: 10.1051/0004-6361/201425120
* Cui & Bai (2021) Cui, C., & Bai, X.-N. 2021, MNRAS, 507, 1106, doi: 10.1093/mnras/stab2220
* Cuzzi et al. (1993) Cuzzi, J. N., Dobrovolskis, A. R., & Champney, J. M. 1993, Icarus, 106, 102, doi: 10.1006/icar.1993.1161
* de Val-Borro et al. (2007) de Val-Borro, M., Artymowicz, P., D’Angelo, G., & Peplinski, A. 2007, A&A, 471, 1043, doi: 10.1051/0004-6361:20077169
* de Val-Borro et al. (2006) de Val-Borro, M., Edgar, R. G., Artymowicz, P., et al. 2006, MNRAS, 370, 529, doi: 10.1111/j.1365-2966.2006.10488.x
* Dong et al. (2017) Dong, R., Li, S., Chiang, E., & Li, H. 2017, ApJ, 843, 127, doi: 10.3847/1538-4357/aa72f2
* Dong et al. (2018) —. 2018, ApJ, 866, 110, doi: 10.3847/1538-4357/aadadd
* Dong et al. (2011a) Dong, R., Rafikov, R. R., & Stone, J. M. 2011a, ApJ, 741, 57, doi: 10.1088/0004-637X/741/1/57
* Dong et al. (2011b) Dong, R., Rafikov, R. R., Stone, J. M., & Petrovich, C. 2011b, ApJ, 741, 56, doi: 10.1088/0004-637X/741/1/56
* Drazkowska et al. (2019) Drazkowska, J., Li, S., Birnstiel, T., Stammler, S. M., & Li, H. 2019, ApJ, 885, 91, doi: 10.3847/1538-4357/ab46b7
* Flock et al. (2015) Flock, M., Ruge, J. P., Dzyurkevich, N., et al. 2015, A&A, 574, A68, doi: 10.1051/0004-6361/201424693
* Fu et al. (2014) Fu, W., Li, H., Lubow, S., Li, S., & Liang, E. 2014, ApJ, 795, L39, doi: 10.1088/2041-8205/795/2/L39
* Fung & Muley (2019) Fung, J., & Muley, D. 2019, ApJS, 244, 42, doi: 10.3847/1538-4365/ab45f3
* Garaud et al. (2004) Garaud, P., Barrière-Fouchet, L., & Lin, D. N. C. 2004, ApJ, 603, 292, doi: 10.1086/381385
* Goldreich & Lynden-Bell (1965) Goldreich, P., & Lynden-Bell, D. 1965, MNRAS, 130, 125, doi: 10.1093/mnras/130.2.125
* Goodman & Pindor (2000) Goodman, J., & Pindor, B. 2000, Icarus, 148, 537, doi: 10.1006/icar.2000.6467
* Hanasz et al. (2010a) Hanasz, M., Kowalik, K., Wóltański, D., & Pawłaszek, R. 2010a, in EAS Publications Series, Vol. 42, EAS Publications Series, ed. K. Gożdziewski, A. Niedzielski, & J. Schneider, 275–280, doi: 10.1051/eas/1042029
* Hanasz et al. (2010b) Hanasz, M., Kowalik, K., Wóltański, D., Pawłaszek, R., & Kornet, K. 2010b, in EAS Publications Series, Vol. 42, EAS Publications Series, ed. K. Gożdziewski, A. Niedzielski, & J. Schneider, 281–285, doi: 10.1051/eas/1042030
* Hawley et al. (1995) Hawley, J. F., Gammie, C. F., & Balbus, S. A. 1995, ApJ, 440, 742, doi: 10.1086/175311
* Hopkins & Squire (2018) Hopkins, P. F., & Squire, J. 2018, MNRAS, 479, 4681, doi: 10.1093/mnras/sty1604
* Hsieh & Lin (2020) Hsieh, H.-F., & Lin, M.-K. 2020, MNRAS, 497, 2425, doi: 10.1093/mnras/staa2115
* Hu & Bai (2021) Hu, Z., & Bai, X.-N. 2021, MNRAS, 503, 162, doi: 10.1093/mnras/stab542
* Huang et al. (2020) Huang, P., Li, H., Isella, A., et al. 2020, ApJ, 893, 89, doi: 10.3847/1538-4357/ab8199
* Johansen et al. (2011) Johansen, A., Klahr, H., & Henning, T. 2011, A&A, 529, A62, doi: 10.1051/0004-6361/201015979
* Johansen et al. (2007) Johansen, A., Oishi, J. S., Mac Low, M.-M., et al. 2007, Nature, 448, 1022, doi: 10.1038/nature06086
* Johansen & Youdin (2007) Johansen, A., & Youdin, A. 2007, ApJ, 662, 627, doi: 10.1086/516730
* Kamp & Dullemond (2004) Kamp, I., & Dullemond, C. P. 2004, ApJ, 615, 991, doi: 10.1086/424703
* Kowalik et al. (2013) Kowalik, K., Hanasz, M., Wóltański, D., & Gawryszczak, A. 2013, MNRAS, 434, 1460, doi: 10.1093/mnras/stt1104
* Krapp & Benítez-Llambay (2020) Krapp, L., & Benítez-Llambay, P. 2020, Research Notes of the American Astronomical Society, 4, 198, doi: 10.3847/2515-5172/abc7be
* Laibe & Price (2011) Laibe, G., & Price, D. J. 2011, MNRAS, 418, 1491, doi: 10.1111/j.1365-2966.2011.19291.x
* Laibe & Price (2012) —. 2012, MNRAS, 420, 2345, doi: 10.1111/j.1365-2966.2011.20202.x
* Laibe & Price (2014) —. 2014, MNRAS, 440, 2136, doi: 10.1093/mnras/stu355
* Lambrechts & Johansen (2012) Lambrechts, M., & Johansen, A. 2012, A&A, 544, A32, doi: 10.1051/0004-6361/201219127
* Lehmann & Wardle (2018) Lehmann, A., & Wardle, M. 2018, MNRAS, 476, 3185, doi: 10.1093/mnras/sty450
* Li et al. (2005) Li, H., Li, S., Koller, J., et al. 2005, ApJ, 624, 1003, doi: 10.1086/429367
* Li et al. (2009) Li, H., Lubow, S. H., Li, S., & Lin, D. N. C. 2009, ApJ, 690, L52, doi: 10.1088/0004-637X/690/1/L52
* Li & Youdin (2021) Li, R., & Youdin, A. N. 2021, ApJ, 919, 107, doi: 10.3847/1538-4357/ac0e9f
* Li et al. (2020) Li, Y.-P., Li, H., Li, S., et al. 2020, ApJ, 892, L19, doi: 10.3847/2041-8213/ab7fb2
* Li et al. (2019) Li, Y.-P., Li, H., Ricci, L., et al. 2019, ApJ, 878, 39, doi: 10.3847/1538-4357/ab1f64
* Marble (1970) Marble, F. E. 1970, Annual Review of Fluid Mechanics, 2, 397, doi: 10.1146/annurev.fl.02.010170.002145
* Masset (2000) Masset, F. 2000, A&AS, 141, 165, doi: 10.1051/aas:2000116
* Masset (2002) Masset, F. S. 2002, A&A, 387, 605, doi: 10.1051/0004-6361:20020240
* Mignone et al. (2019) Mignone, A., Flock, M., & Vaidya, B. 2019, ApJS, 244, 38, doi: 10.3847/1538-4365/ab4356
* Morbidelli et al. (2015) Morbidelli, A., Lambrechts, M., Jacobson, S., & Bitsch, B. 2015, Icarus, 258, 418, doi: 10.1016/j.icarus.2015.06.003
* Moseley et al. (2022) Moseley, E. R., Teyssier, R., & Draine, B. T. 2022, arXiv e-prints, arXiv:2204.07681. https://arxiv.org/abs/2204.07681
* Nakagawa et al. (1986) Nakagawa, Y., Sekiya, M., & Hayashi, C. 1986, Icarus, 67, 375, doi: 10.1016/0019-1035(86)90121-1
* Okuzumi et al. (2016) Okuzumi, S., Momose, M., Sirono, S.-i., Kobayashi, H., & Tanaka, H. 2016, ApJ, 821, 82, doi: 10.3847/0004-637X/821/2/82
* Ormel & Cuzzi (2007) Ormel, C. W., & Cuzzi, J. N. 2007, A&A, 466, 413, doi: 10.1051/0004-6361:20066899
* Ormel & Klahr (2010) Ormel, C. W., & Klahr, H. H. 2010, A&A, 520, A43, doi: 10.1051/0004-6361/201014903
* O’Sullivan & Downes (2006) O’Sullivan, S., & Downes, T. P. 2006, MNRAS, 366, 1329, doi: 10.1111/j.1365-2966.2005.09898.x
* Owen (2020) Owen, J. E. 2020, MNRAS, 495, 3160, doi: 10.1093/mnras/staa1309
* Pareschi & Russo (2005) Pareschi, L., & Russo, G. 2005, Journal of Scientific Computing, 25, 129, doi: 10.1007/BF02728986
* Pierens et al. (2019) Pierens, A., Lin, M. K., & Raymond, S. N. 2019, MNRAS, 488, 645, doi: 10.1093/mnras/stz1718
* Porth et al. (2014) Porth, O., Xia, C., Hendrix, T., Moschou, S. P., & Keppens, R. 2014, ApJS, 214, 4, doi: 10.1088/0067-0049/214/1/4
* Press et al. (1986) Press, W. H., Flannery, B. P., & Teukolsky, S. A. 1986, Numerical recipes. The art of scientific computing
* Price et al. (2018) Price, D. J., Wurster, J., Tricco, T. S., et al. 2018, PASA, 35, e031, doi: 10.1017/pasa.2018.25
* Riols et al. (2020) Riols, A., Lesur, G., & Menard, F. 2020, A&A, 639, A95, doi: 10.1051/0004-6361/201937418
* Schäfer et al. (2020) Schäfer, U., Johansen, A., & Banerjee, R. 2020, A&A, 635, A190, doi: 10.1051/0004-6361/201937371
* Shariff & Cuzzi (2011) Shariff, K., & Cuzzi, J. N. 2011, ApJ, 738, 73, doi: 10.1088/0004-637X/738/1/73
* Simon et al. (2017) Simon, J. B., Armitage, P. J., Youdin, A. N., & Li, R. 2017, ApJ, 847, L12, doi: 10.3847/2041-8213/aa8c79
* Skinner & Ostriker (2010) Skinner, M. A., & Ostriker, E. C. 2010, ApJS, 188, 290, doi: 10.1088/0067-0049/188/1/290
* Stone & Gardiner (2010) Stone, J. M., & Gardiner, T. A. 2010, ApJS, 189, 142, doi: 10.1088/0067-0049/189/1/142
* Stone & Norman (1992a) Stone, J. M., & Norman, M. L. 1992a, ApJS, 80, 753, doi: 10.1086/191680
* Stone & Norman (1992b) —. 1992b, ApJS, 80, 791, doi: 10.1086/191681
* Stone et al. (2020) Stone, J. M., Tomida, K., White, C. J., & Felker, K. G. 2020, ApJS, 249, 4, doi: 10.3847/1538-4365/ab929b
* Suriano et al. (2018) Suriano, S. S., Li, Z.-Y., Krasnopolsky, R., & Shang, H. 2018, MNRAS, 477, 1239, doi: 10.1093/mnras/sty717
* Surville et al. (2020) Surville, C., Mayer, L., & Alibert, Y. 2020, arXiv e-prints, arXiv:2009.04775. https://arxiv.org/abs/2009.04775
* Takahashi & Inutsuka (2014) Takahashi, S. Z., & Inutsuka, S.-i. 2014, ApJ, 794, 55, doi: 10.1088/0004-637X/794/1/55
* Takahashi & Inutsuka (2016) —. 2016, AJ, 152, 184, doi: 10.3847/0004-6256/152/6/184
* Tanaka et al. (2005) Tanaka, H., Himeno, Y., & Ida, S. 2005, ApJ, 625, 414, doi: 10.1086/429658
* Tominaga et al. (2019) Tominaga, R. T., Takahashi, S. Z., & Inutsuka, S.-i. 2019, ApJ, 881, 53, doi: 10.3847/1538-4357/ab25ea
* Tominaga et al. (2020) —. 2020, ApJ, 900, 182, doi: 10.3847/1538-4357/abad36
* van der Marel et al. (2013) van der Marel, N., van Dishoeck, E. F., Bruderer, S., et al. 2013, Science, 340, 1199, doi: 10.1126/science.1236770
* Weidenschilling (1977) Weidenschilling, S. J. 1977, MNRAS, 180, 57, doi: 10.1093/mnras/180.2.57
* Whipple (1972) Whipple, F. L. 1972, in From Plasma to Planet, ed. A. Elvius, 211
* Wolfram (1991) Wolfram, S. 1991, Mathematica: a system for doing mathematics by computer
* Xia et al. (2018) Xia, C., Teunissen, J., El Mellah, I., Chané, E., & Keppens, R. 2018, ApJS, 234, 30, doi: 10.3847/1538-4365/aaa6c8
* Xu & Bai (2022) Xu, Z., & Bai, X.-N. 2022, ApJ, 924, 3, doi: 10.3847/1538-4357/ac31a7
* Xu et al. (2017) Xu, Z., Bai, X.-N., & Murray-Clay, R. A. 2017, ApJ, 847, 52, doi: 10.3847/1538-4357/aa8620
* Yang & Johansen (2016) Yang, C.-C., & Johansen, A. 2016, ApJS, 224, 39, doi: 10.3847/0067-0049/224/2/39
* Yang et al. (2017) Yang, C. C., Johansen, A., & Carrera, D. 2017, A&A, 606, A80, doi: 10.1051/0004-6361/201630106
* Yang & Zhu (2020) Yang, C.-C., & Zhu, Z. 2020, MNRAS, 491, 4702, doi: 10.1093/mnras/stz3232
* Youdin & Johansen (2007) Youdin, A., & Johansen, A. 2007, ApJ, 662, 613, doi: 10.1086/516729
* Youdin & Goodman (2005) Youdin, A. N., & Goodman, J. 2005, ApJ, 620, 459, doi: 10.1086/426895
* Youdin & Lithwick (2007) Youdin, A. N., & Lithwick, Y. 2007, Icarus, 192, 588, doi: 10.1016/j.icarus.2007.07.012
* Zhang et al. (2015) Zhang, K., Blake, G. A., & Bergin, E. A. 2015, ApJ, 806, L7, doi: 10.1088/2041-8205/806/1/L7
* Zhu et al. (2015) Zhu, Z., Stone, J. M., & Bai, X.-N. 2015, ApJ, 801, 81, doi: 10.1088/0004-637X/801/2/81
* Zhu et al. (2014) Zhu, Z., Stone, J. M., Rafikov, R. R., & Bai, X.-n. 2014, ApJ, 785, 122, doi: 10.1088/0004-637X/785/2/122
|
# Learning to Navigate Intersections
with Unsupervised Driver Trait Inference
Shuijing Liu, Peixin Chang, Haonan Chen,
Neeloy Chakraborty, and Katherine Driggs-Campbell S. Liu, P. Chang, H. Chen,
N. Chakraborty and K. Driggs-Campbell are with the Department of Electrical
and Computer Engineering at the University of Illinois at Urbana-Champaign.
emails<EMAIL_ADDRESS>material is
based upon work supported by the National Science Foundation under Grant No.
2143435.
###### Abstract
Navigation through uncontrolled intersections is one of the key challenges for
autonomous vehicles. Identifying the subtle differences in hidden traits of
other drivers can bring significant benefits when navigating in such
environments. We propose an unsupervised method for inferring driver traits
such as driving styles from observed vehicle trajectories. We use a
variational autoencoder with recurrent neural networks to learn a latent
representation of traits without any ground truth trait labels. Then, we use
this trait representation to learn a policy for an autonomous vehicle to
navigate through a T-intersection with deep reinforcement learning. Our
pipeline enables the autonomous vehicle to adjust its actions when dealing
with drivers of different traits to ensure safety and efficiency. Our method
demonstrates promising performance and outperforms state-of-the-art baselines
in the T-intersection scenario. For code implementation and videos, please
visit https://github.com/Shuijing725/VAE_trait_inference.
## I Introduction
To successfully navigate through uncontrolled intersections, autonomous
vehicles must carefully reason about how to interact with different types of
human drivers [1]. It is important for the vehicle to infer the traits of
human drivers, such as the propensity for aggression or cooperation, and
adjust its strategies accordingly [2, 3]. Inspired by recent advancements of
unsupervised learning and deep reinforcement learning, we propose a novel
pipeline to learn a representation of traits of other drivers, which is used
for autonomous navigation in uncontrolled intersections.
Trait inference is challenging yet essential for the navigation of autonomous
vehicles for two reasons. First, the environment is not fully observable to
the ego vehicle, since each traffic participant runs its own policy
individually and has its own internal states. The ego vehicle needs to
interpret the hidden states such as driving styles and intents of other agents
to understand future behaviors that may influence planning [2, 3]. Second,
uncontrolled intersections are less structured since traffic lights and stop
signs are not present to coordinate agent behaviors. The traffic participants
implicitly interact and negotiate with each other, making the environment
complex and potentially dangerous [4, 5, 6, 7]. By inferring the traits of
other drivers, the ego vehicle can be cautious when other drivers are
aggressive or irrational and bold when they are passive or cooperative,
improving both the safety and efficiency of interactive navigation.
To address the above problems, Morton et al learns a latent representation of
driver traits, which is fed into a feedforward policy to produce multimodal
behaviors [8]. However, the feedforward policy only considers current states
and actions which are not sufficient to fully express long-term properties of
drivers such as traits. As a result, this representation fails to distinguish
between different traits. Ma et al classifies driver traits with supervised
learning to aid navigation in intersections [3] but has the following two
problems. First, the trait labels are expensive to obtain and usually do not
exist in most real driving datasets [9, 1]. Second, the navigation policy is
trained with ground truth trait labels instead of predicted traits. When the
trait classifier and the policy are combined in testing, intermediate and
cascading errors cause severe performance degradation.
Figure 1: The T-intersection scenario in left-handed traffic. The goal of the
ego car (yellow) is to take a right turn and merge into the upper lane without
colliding with other cars. The conservative car (blue) yields to the ego car
while the aggressive cars (red) do not yield.
In this paper, we study the same uncontrolled T-intersection navigation
problem as in [3], which is shown in Fig. 1. Before entering the
T-intersection, the ego vehicle needs to infer the latent driving styles of
other drivers to determine whether they are willing to yield to the ego
vehicle. Based on the inferred driving style, the ego vehicle must intercept
the drivers who will yield to achieve the goal.
We seek to create a pipeline that first learns a representation of driver
traits in an unsupervised way, and then uses the trait representation to
improve navigation in the T-intersection. In the first stage, we encode the
trajectories of other drivers to a latent representation using a variational
autoencoder (VAE) with recurrent neural networks (RNN). Since trajectory
sequences reveal richer information about driver traits than single states,
the VAE+RNN network effectively learns to distinguish between different traits
without any explicit trait labels. In the second stage, we use the latent
representations of driver traits as inputs to the ego vehicle’s navigation
policy, which is trained with model-free reinforcement learning (RL). With the
inferred traits, the ego vehicle is able to adjust its decisions when dealing
with different drivers, which leads to improved performance. Since the RL
policy is trained with the learned representations instead of ground truth
labels, our pipeline is much less sensitive to cascading errors from the trait
inference network.
We present the following contributions: (1) We propose a novel unsupervised
approach to learn a representation of driver traits with a VAE+RNN network;
(2) We use the learned representation to improve the navigation of an
autonomous vehicle through an uncontrolled T-intersection; (3) Our trait
representation and navigation policy exhibit promising results and outperform
previous works.
This paper is organized as follows: We review related works in Section II. We
formalize the problem and propose our method in Section III. Experiments and
results are discussed in Section IV. We conclude the paper in Section V.
## II Related Works
### II-A Driver internal state estimation
Driver internal state estimation can be divided into intent estimation and
trait estimation [10]. Intent estimation often uses methods such as
probabilistic graph model and unparameterized belief tracker to predict the
future maneuvers of other drivers [6, 11, 12], which can then inform
downstream planning for the ego driver [13, 6, 12]. Trait estimation infers
the properties of drivers such as driving styles, preferences, fatigue, and
level of distraction [10]. Some works distinguish between distracted and
attentive drivers for behavior prediction and cooperative planning [14, 15].
Driving style recognition has been addressed with both unsupervised and
supervised learning methods, which we will discuss in detail below [16, 17, 8,
3].
Morton et al propose a method that first encodes driving trajectories with
different driving styles to a latent space. Then, the latent encodings and the
current driver states are fed into a feedforward policy that produces
multimodal actions [8]. The encoder and the policy are optimized jointly.
However, since the feedforward policy only considers the relations between
current states and actions, the joint optimization encourages the encoder to
encode short-term information of the trajectories such as accelerations while
ignoring the persistent properties such as traits. In contrast, since our
method uses an RNN decoder to reconstruct the trajectories, the encoder must
encode the trait information. Ma et al propose a graph neural network that
classifies the driving styles with supervised learning, which requires large
amounts of labeled data and is difficult to scale in reality [3]. In addition,
since the definitions of trait properties vary by individuals and cultures,
manually labeled trait information will likely be noisy and inconsistent. In
contrast, our method is unsupervised and does not need any trait labels.
### II-B Representation learning for sequential data
Contrastive learning is widely used to learn representations from sequential
data such as videos and pedestrian trajectories [18, 19, 20]. However, the
performance of contrastive learning is sensitive to the quality of negative
samples, and finding efficient negative sampling strategies remains an open
challenge [21].
Another category of representation learning methods is VAE and its variants
[22, 23]. Bowman et al introduce an RNN-based VAE to model the latent
properties of sentences [24], which inspires us to learn the traits of drivers
from their trajectories. Conditional VAEs (CVAE) are widely used in pedestrian
and vehicle trajectory predictions since discrete latent states can represent
different behavior modes such as braking and turning [25, 26, 27, 28]. While
these behavior modes change frequently, the driver traits that we aim to learn
are persistent with each driver [10]. Also, the goal of these CVAEs is to
generate multimodal trajectory predictions while, to the best of our
knowledge, our work is the first to infer driver traits with VAE for
autonomous driving.
### II-C Autonomous driving in uncontrolled intersections
Autonomous navigation through uncontrolled intersections is well studied and
has had many successful demonstrations [29, 6, 30, 3]. Some heuristic methods
use a time-to-collision (TTC) threshold to decide when to cross [31, 30].
However, the TTC models assume constant velocity for the surrounding vehicles,
which ignores the interactions and internal states of drivers. Also, the TTC
models can be overly cautious and cause unnecessary delays [29]. Another line
of works formulates the problem as a partially observable Markov decision
process (POMDP), which accounts for the uncertainties and partial
observability in the intersection scenario but is computationally expensive to
solve [6, 32, 33]. RL-based methods use neural networks as function
approximators to learn driving policies. Isele et al learns to navigate in
occluded intersections using deep Q-learning [29]. Ma et al focuses on
navigation in a T-Intersection where the drivers exhibit different traits [3].
However, the navigation policy is trained with ground truth traits and only
uses the inferred traits in testing. Thus, the performance of the RL policy is
sensitive to the intermediate errors from the trait inference module. In
contrast, our method trains the RL policy directly with inferred trait
representations, which eliminates the problems caused by the intermediate
errors.
## III Methodology
In this section, we first present our unsupervised approach to represent
driver traits from unlabeled driving trajectories with a VAE+RNN network.
Then, we discuss how we use the inferred traits to improve the navigation
policy.
Figure 2: The network architectures. (a) The VAE+RNN network for trait
representation learning. The fully connected layers before and after GRUs are
eliminated for clarity. The dice represents the reparameterization trick. The
start-of-sequence state is denoted by $\langle SOS\rangle$. (b) The navigation
policy network. The weights of the encoder (blue) is fixed and only the yellow
part is trained with RL. We use $[\bullet\bullet]$ to denote concatenation.
For illustration purposes, we assume that the inferred trait $z_{1},...,z_{n}$
is updated at time $t$.
### III-A Preliminaries
Consider a T-intersection environment with an ego vehicle and $n$ surrounding
vehicles. The number of surrounding vehicles $n$ may change at any timestep
$t$. Suppose that all vehicles move in a 2D Euclidean space. Let ${o}_{0}^{t}$
be the state of the ego vehicle and ${o}_{i}^{t}$ be the observable state of
the $i$-th surrounding vehicle at time $t$, where $i\in\\{1,...,n\\}$. The
state of the ego vehicle ${o}_{0}^{t}$ consists of the position
$(p_{x},p_{y})$ and the velocity $(v_{x},v_{y})$ of the vehicle. The
observable state of each surrounding vehicle ${o}_{i}^{t}$ consists of its
position $(p_{x}^{i},p_{y}^{i})$. In contrast to the previous work [3],
${o}_{i}^{t}$ does not include other vehicles’ velocities because they are
hard to accurately measure without special facilities and algorithms in the
real world [34, 35]. In addition, each surrounding vehicle has a latent state
$z_{i}$ that indicates the aggressiveness (i.e., the trait) of the $i$-th
driver. We assume that the driving style of each driver $z_{i}$ does not
change throughout an episode. The positions of surrounding vehicles
${o}_{1}^{t},...,{o}_{n}^{t}$ are observable to the ego vehicle, while the
latent states $z_{1},...,z_{n}$ are not.
### III-B Trait representation learning
#### III-B1 Data collection
For each vehicle in the T-Intersection, the vehicle in front of it has the
most direct influence on its behaviors. For this reason, in the trait
representation learning stage, we define the observable state of each vehicle
to be $x=(\Delta p_{x},\Delta p_{x,f})$, where $\Delta p_{x}$ is the
horizontal offset from the vehicle’s starting position in the trajectory, and
$\Delta p_{x,f}$ is the horizontal displacement of the vehicle from its front
vehicle. Then, the trajectory of each driver is a sequence of states
$\mathbf{x}=[x^{1},...,x^{l}]$, where $l$ is the length of the trajectory.
We only keep the longitudinal state information because all vehicles move in
horizontal lanes and the lateral states are not insightful in this setting,
except indicating which lane the vehicle is in. We rotate the trajectories so
that all trajectories in the dataset are aligned in the same direction. Thus,
the lane and the directional information is indistinguishable within the
trajectory data, allowing the network to focus on learning the trait
difference instead of other differences between vehicles.
To collect the trajectory data, we run a simulation of the T-intersection
scenario without the presence of the ego car and record the trajectories of
all surrounding vehicles, which are controlled by the Intelligent Driver Model
(IDM) [36]. Learning trait representations from this dataset allows the ego
car to infer the traits from the trajectories of other drivers before deciding
to intercept or wait. The dataset is denoted as
$\\{\mathbf{x}_{j}\\}_{j=1}^{N}$, where $N$ is the total number of
trajectories.
#### III-B2 Network architecture
We use the collected dataset to train the VAE+RNN network to learn a
representation of traits, as shown in Fig. 2a. The VAE network consists of an
encoder, which compresses a trajectory $\mathbf{x}$ to a distribution of a
latent trait vector $z$, and a decoder, which reconstructs the trajectory from
the latent vector $z$. Both the encoder and the decoder are gated recurrent
unit (GRU) networks since GRUs are more computationally efficient than long
short-term memory networks (LSTM).
Given a trajectory $\mathbf{x}=[x^{1},...,x^{l}]$, the encoder GRU first
applies a non-linear embedding layer $f_{\textrm{encoder}}$ to each state
$x^{t}$ and then feeds the embedded features to the GRU cell:
$h^{t}_{e}=\mathrm{GRU}\left(h^{t-1}_{e},f_{\textrm{encoder}}(x^{t})\right)$
(1)
where $h^{t}_{e}$ is the hidden state of the encoder GRU at time
$t\in\\{1,...,l\\}$. After the entire trajectory is fed through the encoder
GRU, we take the last hidden state $h^{l}_{e}$ as the encoded latent feature
of the trajectory $\mathbf{x}$ and apply fully connected layers $f_{\mu}$ and
$f_{\sigma}$ to get the Gaussian parameters of the latent driving style
$z\in\mathbb{R}^{2}$ in a two-dimensional latent space:
$\mu=f_{\mu}(h^{t}_{e}),\quad\sigma_{i}=f_{\sigma}(h^{t}_{e}).$ (2)
Finally, we use the reparameterization trick to sample $z$ from
$\mathcal{N}(\mu,\sigma)$ for efficient learning:
$z=\mu+\epsilon\sigma,\epsilon\sim\mathcal{N}(0,I)$.
In the decoding stage, since the driving style of each driver does not change
over time, we treat the latent state $z$ as part of the vehicle state instead
of the initial hidden state of decoder GRU. To this end, at each timestep $t$,
we concatenate the reconstructed state $\hat{x}^{t-1}$ from the last timestep
and the latent state $z$ from the encoder. Then, we embed the joint states
using $f_{\textrm{decoder}}$, feed the embeddings into the next decoder GRU
cell, and apply another embedding $g_{\textrm{decoder}}$ to the next hidden
state $h^{t}_{d}$, which outputs the next reconstructed state $\hat{x}^{t}$:
$\displaystyle h^{t}_{d}$
$\displaystyle=\mathrm{GRU}\left(h^{t-1}_{d},f_{\textrm{decoder}}([\hat{x}^{t-1},z])\right)$
(3) $\displaystyle\hat{x}^{t}$
$\displaystyle=g_{\textrm{decoder}}(h^{t}_{d}).$
In the first timestep, we use a special start-of-sequence (SOS) state, which
is similar to the start-of-sequence symbol in natural language processing to
reconstruct $\hat{x}^{1}$ [37]. The process in Eq. 3 repeats until we
reconstruct the whole trajectory
$\hat{\mathbf{x}}=[\hat{x}^{1},...,\hat{x}^{l}]$.
The objective for training our VAE+RNN network is
$\mathcal{L}=\beta
D_{KL}\left(\mathcal{N}(\mu,\sigma)||\mathcal{N}(0,I)\right)+||\mathbf{x}-\hat{\mathbf{x}}||_{2}$
(4)
where $D_{KL}$ is the Kullback–Leibler (KL) divergence. The first term
regularizes the distribution of the latent trait $z$ to be close to a prior
with a standard normal distribution. The second term is the reconstruction
loss and measures the $L2$ error of the reconstructed trajectories from the
original trajectories. The two terms are summed with a weight $\beta$.
By optimizing Eq. 4, our network learns latent encodings that represent the
trait of each trajectory without any ground truth trait labels. Note that we
also make no assumptions on the number of trait classes or the semantic
meanings of the trait classes. Thus, our network has the potential to
generalize to real trajectory datasets with more complex traits.
### III-C Navigation policy learning
We model the T-intersection scenario as a POMDP, defined by the tuple
$\langle\mathcal{S},\mathcal{A},\mathcal{T},\mathcal{R},\mathcal{O},\mathcal{P},\gamma\rangle$.
Suppose that there are $n$ surrounding vehicles at timestep $t$. We use
${o}_{t}=[{o}_{0}^{t},{o}_{1}^{t},...,{o}_{n}^{t}]\in\mathcal{O}$ to denote
the observations of the ego vehicle, where $\mathcal{O}$ is the observation
space. Let ${u}^{t}_{i}=[{o}^{t}_{i},z_{i}]\in\mathbb{R}^{4}$ be the state of
the $i$-th surrounding vehicle. Since the ego vehicle is influenced by all
surrounding vehicles, we use
$s_{t}=[{o}^{t}_{0},{u}^{t}_{1},...,{u}^{t}_{n}]\in\mathcal{S}$ to denote the
state of the POMDP, where $\mathcal{S}$ is the state space. And
$\mathcal{P}:\mathcal{S}\rightarrow\mathcal{O}$ is the set of conditional
observation probabilities.
At each timestep $t$, the ego vehicle chooses the desired speed for the low-
level controller $a_{t}\in\mathcal{A}$ according to its policy
$\pi(a_{t}|s_{t})$, where $\mathcal{A}=\\{0,0.5,3\\}\mathrm{m/s}$ is the
action space. In return, the agent receives a reward $r_{t}$ and transits to
the next state $s_{t+1}$ according to an unknown state transition
$\mathcal{T}(\cdot|s_{t},a_{t})$. Meanwhile, all other vehicles also take
actions according to their policies and move to the next states with unknown
state transition probabilities. The episode continues until $t$ exceeds the
maximum episode length $T$, the ego vehicle reaches its goal, or the ego
vehicle collides with any other vehicle. The goal of the agent is to maximize
the expected return, $R_{t}=\mathbb{E}[\sum^{T}_{i=t}\gamma^{i-t}r_{i}]$,
where $\gamma$ is a discount factor. The value function $V^{\pi}(s)$ is
defined as the expected return starting from $s$, and successively following
policy $\pi$.
To handle the unknown transitions and environment complexity, we train our
policy network illustrated in Fig. 2b using model-free deep RL. During RL
training, we fix the trainable weights of the encoder. For every $l$
timesteps, we feed the trajectories of the surrounding vehicles over the past
$l$ timesteps to the encoder and infer the driving style $z_{i}^{t}$ of each
driver $i$. To improve efficiency, we only update the latent states of the
drivers in the lanes that the ego car has not passed yet, as shown in Fig. 5.
Since the ego car is not allowed to move backward, the latent states of the
drivers that the ego car has already passed or the drivers that have already
yielded to the ego car have no effect on the ego car’s decisions and thus are
not updated anymore.
Our policy network is a GRU with an attention module. We concatenate the state
of each driver ${u}^{t}_{i}$ with the state of the ego vehicle ${o}_{0}$ to
obtain ${q}_{i}^{t}=[{u}^{t}_{i},{o}_{0}^{t}]$, where
${q}_{i}^{t}\in\mathbb{R}^{8}$ and $i\in\\{1,...,n\\}$. We feed each
concatenated state ${q}_{i}^{t}$ into an attention module which assigns
attention weights to each surrounding vehicle. Specifically, we embed
${q}_{i}^{t}$ using a multi-layer perceptron (MLP) denoted as $f_{emb}$ and
calculate the mean of the embeddings of each surrounding vehicle:
$m^{t}=\frac{1}{n}\sum^{n}_{i=1}{e}_{i}^{t},\quad{e}_{i}^{t}=f_{emb}({q}_{i}^{t})$
(5)
where ${m}^{t}$ is the resulting mean. The weighted feature of each
surrounding vehicle ${c}_{i}^{t}$ is calculated by
${c}_{i}^{t}=\alpha_{i}^{t}\cdot{e}_{i}^{t},\quad\alpha_{i}=f_{attn}([{e}_{i}^{t},{m}^{t}])$
(6)
where $\alpha_{i}^{t}\in\mathbb{R}$ is the attention score for the $i$-th
vehicle, and $f_{attn}$ is another MLP. We then concatenate the sum of
$c_{1}^{t},...,c_{n}^{t}$ with the state of the ego vehicle ${o}_{0}^{t}$ and
feed the result to a GRU:
$h^{t}_{\pi}=\mathrm{GRU}\left(h^{t-1}_{\pi},\left[\sum^{n}_{i=1}c_{i}^{t},o_{0}^{t}\right]\right)$
(7)
Finally, the hidden state of the GRU $h^{t}_{\pi}$ is fed to a fully connected
layer to obtain the value $V(s_{t})$ and the policy $\pi(a_{t}|s_{t})$. We use
Proximal Policy Optimization (PPO), a model-free policy gradient algorithm,
for policy and value function learning [38].
Figure 3: Visualizations of the latent representations of two driver traits.
(a) Our method. (b) The method by Morton et al. (c) The original processed
trajectories corresponding to the labeled latent vectors in (a) and (b). The
$x$-axis is the horizontal displacement in meters. Each triangle marker
indicates the position of a car at each timestep and the markers become darker
over time. Denser triangles indicate smaller velocities. The blue and red
trajectories indicate conservative and aggressive cars respectively. The brown
trajectories indicate the front cars. If the front car goes out of the
boundary before the trajectory ends, the brown trajectories will be shorter
than the red or blue trajectories, such as #1 and #2.
## IV Experiments and results
In this section, we first describe the simulation environment for training. We
then present our experiments and results for trait representation and
navigation policy.
### IV-A Simulation environment
Fig. 5 shows our simulation environment adapted from [3]. At the beginning of
an episode, the surrounding vehicles are randomly placed in a two-way street
with opposite lanes and we assume that they never take turns or change lanes.
The number of surrounding vehicles varies as vehicles enter into or exit from
the T-intersection. We assume that all cars can always be detected and
tracked. The surrounding vehicles are controlled by IDM [36]. Conservative
drivers vary their front gaps from the preceding vehicles between $0.5m$ and
$0.7m$ and have the desired speed of $2.4m/s$. Aggressive drivers vary their
front gaps between $0.3m$ and $0.5m$ and have the desired speed of $3m/s$. If
the ego car moves in front of other cars, conservative drivers will yield to
the ego car, while aggressive drivers will ignore and collide with the ego
car.
The ego car starts at the bottom of the T-intersection. The goal of the ego
car is to take a right turn to merge into the upper lane without colliding
with other cars. The ego car is controlled by a longitudinal PD controller
whose desired speed is set by the RL policy network. The right-turn path of
the ego car is fixed to follow the traffic rule. The ego car also has a safety
checker that clips the magnitude of its acceleration if it gets dangerously
close to other cars.
Let $S_{goal}$ be the set of goal states, where the ego vehicle successfully
makes a full right-turn, and $S_{fail}$ be the set of failure states, where
the ego vehicle collides with other vehicles. Let
$r_{speed}(s)=0.05\times\lVert v_{ego}\rVert_{2}$ be a small reward on the
speed of the ego vehicle and $r_{step}=-0.0013$ be a constant penalty that
encourages the ego vehicle to reach the goal as soon as possible. The reward
function is defined as
$\begin{split}\begin{gathered}r(s,a)=\begin{cases}2.5,&\text{if }s\in
S_{goal}\\\ -2,&\text{if }s\in S_{fail}\\\
r_{speed}(s)+r_{step},&\text{otherwise}.\end{cases}\end{gathered}\end{split}$
(8)
### IV-B Trait inference
#### IV-B1 Baseline
We compare the latent representation of our method with the method proposed by
Morton et al [8].
TABLE I: Testing results of simple supervised classifiers with learned latent representations Method | Accuracy
---|---
Ours | 98.08%
Morton et al. | 60.22%
The encoder of the baseline is the same as our encoder except that the
baseline takes the longitudinal acceleration at each timestep as an extra
input. The policy network of the baseline is a 4-layer multilayer perceptron
(MLP) network that imitates the IDM policy from the trajectory dataset.
#### IV-B2 Training and evaluation
Our dataset contains approximately $696,000$ trajectories from both classes,
and the train/test split ratio is $2$:$1$. We train both methods for $1000$
epochs with a decaying learning rate $5\times 10^{-4}$. The weight of the KL
divergence loss $\beta$ is $5\times 10^{-8}$ for both methods.
To better understand the learned latent representation and how it provides
trait information to the navigation policy, we fix the trainable weights of
the encoders, train linear support vector classifiers using the inferred
latent states as inputs, and record the testing accuracy of the classifiers.
To visualize the learned representations, we feed a set of testing
trajectories into the encoder and visualize their latent representations.
#### IV-B3 Results and interpretability
In Table I, the supervised classifier with our latent representation achieves
much higher classification accuracy than that of Morton et al. Together with
Fig. 3a, we show that our representation successfully separates the vehicle
trajectories with different traits in most cases. For example, in Fig. 3c, #1,
#5, and #6 are in the aggressive cluster, while #2, #7, and #8 are in the
conservative cluster. The overlapped region encodes the trajectories with
ambiguous traits, such as very short trajectories (#3) and the trajectories
with ambiguous front gaps. For example, the average front gap of #4 is between
the aggressive #6 and the conservative #8. Besides the binary traits, our
latent representation also captures other meaningful information. First, the
trajectories of the cars whose front cars go out of the border, such as #1 and
#2, are mapped into the fan-shaped peripherals of the two clusters
respectively. Second, in the central regions of the two clusters, the
trajectories with larger average speeds are mapped in the lower left part (#5
and #7), while those with smaller average speeds are in the upper right part
(#6 and #8).
From Table I and Fig. 3b, the baseline suffers from severe model collapse and
the representation fails to separate the two traits. The reason is that the
MLP policy only considers current state-action pairs, which encourages the
encoder to only encode features within short time windows such as
accelerations while ignoring the properties of the trajectories such as
traits. Despite the model collapse, the baseline still learns some meaningful
representations. For example, the trajectories with uniform speeds together
such as #5 and #7 and forms separate clusters for the decelerating
trajectories such as #6. Therefore, we conclude that compared with single
states, trajectories exhibit richer information about traits and are better
suited for trait representation learning.
### IV-C Navigation with inferred traits
#### IV-C1 Baselines and ablations
We use the following two baselines: (1) The pipeline proposed by Ma et al,
which trains a supervised trait predictor and an RL policy with binary ground
truth trait labels separately and combines them at test time [3]; (2) We use
the latent representation by Morton et al to train a policy network using the
same method described in Sec. III-C. In addition, we use an RL policy trained
with ground truth labels as an oracle, and another RL policy trained without
any trait information as a naïve model. To examine the effectiveness of the
attention mechanism, we train an ablated model of our method without the
attention module. For a fair comparison, the architectures of all policy
networks are kept the same.
#### IV-C2 Training
We run three experiments with different proportions of two types of drivers,
as shown in Fig. 4. We train the policy networks for all methods and ablations
for $1\times 10^{7}$ timesteps with a decaying learning rate $1\times
10^{-4}$. To accelerate and stabilize training, we run twelve instances of the
simulation environment in parallel for collecting the ego car’s experiences.
At each policy update, 30 steps of 6 episodes are used. For Ma et al, we
pretrain a trait classification network with a $96\%$ testing accuracy and use
the classifier to infer traits for the RL policy.
#### IV-C3 Evaluation
We test all models with $500$ random unseen test cases. We measure the
percentage of success, collision, and timeout episodes as the evaluation
criteria.
Figure 4: Success, timeout, and collision rates w.r.t. different driver trait
distributions. $P(\textrm{conservative})$ is the probability for each
surrounding driver to be conservative. The task difficulty increases as
$P(\textrm{conservative})$ decreases. The numbers on the bars indicate the
percentages of the corresponding bars. Figure 5: Qualitative result of our
method. The ego car is in yellow, the conservative cars are in blue, and the
aggressive cars are in red. As mentioned in Sec. III-C, the latent traits of
the light blue and light red cars are updated periodically, while those of the
dark blue and dark red cars are not updated and stay the same as before.
#### IV-C4 Results
As shown in Fig. 4, the success rates of our method are closest to the oracle
who has access to true trait labels in all experiments, with an average
difference of $2\%$. We believe the main reason is that our trait
representation effectively captures the trait differences of the surrounding
drivers and aids the decision-making in RL. The performance gap between the
oracle and our method is caused by the drivers with ambiguous traits, as well
as the fact that the learned representation is noisier than the true labels.
Although the model with no labels can implicitly infer traits to some extent
in RL training, our policy is able to utilize the existing trait
representation and focuses more on the decision-making of the ego vehicle,
which leads to better navigation performance. Fig. 5 shows a successful
episode of our method, where the ego car waits until a conservative car
appears, cuts in the front of the conservative cars when passing both lanes,
and completes the right turn.
Compared with our model, the model by Morton et al has a lower success rate
especially in more challenging experiment settings when
$P(\textrm{conservative})$ is smaller. The reason is that the latent
representation does not distinguish between different traits and only provides
very limited useful information to RL. This implies that the policy still
needs to implicitly infer the traits while being distracted by the latent
representation, which negatively affects the performance.
For Ma et al, the trait classifier and the RL policy both have good
performance when tested separately. However, when the two modules are combined
together, intermediate and cascading errors significantly lower the success
rates. The performance drop is due to the distribution shift between the true
traits and inferred traits. Since the policy is trained with true traits, it
fails easily whenever the trait classifier makes a small mistake. Moreover, Ma
et al requires trait labels, but our trait representation is learned without
any labels and still outperforms Ma et al in navigation.
### IV-D Attention vs. no attention
The second from the rightmost graph in Fig. 4 shows that the removal of the
attention module results in $3\%\sim 14\%$ lower success rates. Since the
attention module assigns different weights to the cars with different traits
and in different positions, the policy is able to focus on the cars which have
a bigger influence on the ego car, which validates the necessity of the
attention mechanism.
## V Conclusions and future work
We propose a novel pipeline that encodes the trajectories of drivers to a
latent trait representation with a VAE+RNN network. Then, we use the trait
representation to improve the navigation of an autonomous vehicle through an
uncontrolled T-intersection. The trait representation is learned without any
explicit supervision. Our method outperforms baselines in the navigation task
and the trait representation shows interpretability. Possible directions to
explore in future work include (1) validating our method with real driving
trajectory data, (2) generalizing our model to more sophisticated driver
internal states and more navigation tasks, and (3) incorporating occlusions
and limited sensor range of the ego vehicle to close the gap between the
simulation and the real world.
## Acknowledgements
We thank Xiaobai Ma for thoughtful discussions and for providing the baseline
code. We thank Zhe Huang and Tianchen Ji for feedback on paper drafts.
## References
* [1] W. Zhan, L. Sun, D. Wang, H. Shi, A. Clausse, M. Naumann, J. Kümmerle, H. Königshof, C. Stiller, A. de La Fortelle, and M. Tomizuka, “INTERACTION Dataset: An INTERnational, Adversarial and Cooperative moTION Dataset in Interactive Driving Scenarios with Semantic Maps,” _arXiv preprint arXiv: 1910.03088_ , 2019.
* [2] Z. N. Sunberg, C. J. Ho, and M. J. Kochenderfer, “The value of inferring the internal state of traffic participants for autonomous freeway driving,” in _American Control Conference (ACC)_ , 2017, pp. 3004–3010.
* [3] X. Ma, J. Li, M. J. Kochenderfer, D. Isele, and K. Fujimura, “Reinforcement learning for autonomous driving with latent state inference and spatial-temporal relationships,” in _IEEE International Conference on Robotics and Automation (ICRA)_ , 2021.
* [4] “Traffic safety facts 2019,” National Highway Traffic Safety Administration, Tech. Rep. DOT HS 813 141, 2021.
* [5] B. Qian, H. Zhou, F. Lyu, J. Li, T. Ma, and F. Hou, “Toward collision-free and efficient coordination for automated vehicles at unsignalized intersection,” _IEEE Internet of Things Journal_ , vol. 6, no. 6, pp. 10 408–10 420, 2019\.
* [6] W. Song, G. Xiong, and H. Chen, “Intention-aware autonomous driving decision-making in an uncontrolled intersection,” _Mathematical Problems in Engineering_ , vol. 2016, 2016.
* [7] S. Liu, P. Chang, W. Liang, N. Chakraborty, and K. Driggs-Campbell, “Decentralized structural-rnn for robot crowd navigation with deep reinforcement learning,” in _IEEE International Conference on Robotics and Automation (ICRA)_ , 2021, pp. 3517–3524.
* [8] J. Morton and M. J. Kochenderfer, “Simultaneous policy learning and latent state inference for imitating driver behavior,” in _IEEE International Conference on Intelligent Transportation Systems (ITSC)_ , 2017, pp. 1–6.
* [9] R. Krajewski, J. Bock, L. Kloeker, and L. Eckstein, “The highd dataset: A drone dataset of naturalistic vehicle trajectories on german highways for validation of highly automated driving systems,” in _IEEE International Conference on Intelligent Transportation Systems (ITSC)_ , 2018, pp. 2118–2125.
* [10] K. Brown, K. Driggs-Campbell, and M. J. Kochenderfer, “A taxonomy and review of algorithms for modeling and predicting human driver behavior,” _arXiv preprint arXiv:2006.08832_ , 2020.
* [11] C. Dong, J. M. Dolan, and B. Litkouhi, “Intention estimation for ramp merging control in autonomous driving,” in _IEEE Intelligent Vehicles Symposium (IV)_ , 2017, pp. 1584–1589.
* [12] H. Bai, S. Cai, N. Ye, D. Hsu, and W. S. Lee, “Intention-aware online pomdp planning for autonomous driving in a crowd,” in _IEEE International Conference on Robotics and Automation (ICRA)_ , 2015, pp. 454–460.
* [13] K. Driggs-Campbell, V. Govindarajan, and R. Bajcsy, “Integrating intuitive driver models in autonomous planning for interactive maneuvers,” _IEEE Transactions on Intelligent Transportation Systems_ , vol. 18, no. 12, pp. 3461–3472, 2017.
* [14] K. Driggs-Campbell, V. Shia, and R. Bajcsy, “Improved driver modeling for human-in-the-loop vehicular control,” in _IEEE International Conference on Robotics and Automation (ICRA)_ , 2015, pp. 1654–1661.
* [15] D. Sadigh, N. Landolfi, S. S. Sastry, S. A. Seshia, and A. D. Dragan, “Planning for cars that coordinate with people: leveraging effects on human actions for planning and active information gathering over human internal state,” _Autonomous Robots_ , vol. 42, no. 7, pp. 1405–1426, 2018.
* [16] C. M. Martinez, M. Heucke, F.-Y. Wang, B. Gao, and D. Cao, “Driving style recognition for intelligent vehicle control and advanced driver assistance: A survey,” _IEEE Transactions on Intelligent Transportation Systems_ , vol. 19, no. 3, pp. 666–676, 2017.
* [17] W. Dong, T. Yuan, K. Yang, C. Li, and S. Zhang, “Autoencoder regularized network for driving style representation learning,” in _International Joint Conference on Artificial Intelligence (IJCAI)_ , 2017, p. 1603–1609.
* [18] X. Wang and A. Gupta, “Unsupervised learning of visual representations using videos,” in _International Conference on Computer Vision (ICCV)_ , 2015, pp. 2794–2802.
* [19] I. Misra, C. L. Zitnick, and M. Hebert, “Shuffle and learn: unsupervised learning using temporal order verification,” in _European Conference on Computer Vision (ECCV)_ , 2016, pp. 527–544.
* [20] Y. Liu, Q. Yan, and A. Alahi, “Social nce: Contrastive learning of socially-aware motion representations,” in _International Conference on Computer Vision (ICCV)_ , 2021.
* [21] J.-B. Grill, F. Strub, F. Altché, C. Tallec, P. Richemond, E. Buchatskaya, C. Doersch, B. Avila Pires, Z. Guo, M. Gheshlaghi Azar, B. Piot, K. Kavukcuoglu, R. Munos, and M. Valko, “Bootstrap your own latent - a new approach to self-supervised learning,” in _Advances in Neural Information Processing Systems (NeurIPS)_ , vol. 33, 2020, pp. 21 271–21 284.
* [22] D. P. Kingma and M. Welling, “Auto-encoding variational bayes,” in _International Conference on Learning Representations(ICLR)_ , 2014.
* [23] K. Sohn, H. Lee, and X. Yan, “Learning structured output representation using deep conditional generative models,” in _Advances in Neural Information Processing Systems (NeurIPS)_ , vol. 28, 2015, pp. 3483–3491.
* [24] S. R. Bowman, L. Vilnis, O. Vinyals, A. M. Dai, R. Jozefowicz, and S. Bengio, “Generating sentences from a continuous space,” _arXiv preprint arXiv:1511.06349_ , 2015.
* [25] T. Salzmann, B. Ivanovic, P. Chakravarty, and M. Pavone, “Trajectron++: Dynamically-feasible trajectory forecasting with heterogeneous data,” in _European Conference on Computer Vision (ECCV)_ , 2020, pp. 683–700.
* [26] B. Ivanovic, K. Leung, E. Schmerling, and M. Pavone, “Multimodal deep generative models for trajectory prediction: A conditional variational autoencoder approach,” _IEEE Robotics and Automation Letters_ , vol. 6, no. 2, pp. 295–302, 2020.
* [27] X. Feng, Z. Cen, J. Hu, and Y. Zhang, “Vehicle trajectory prediction using intention-based conditional variational autoencoder,” in _IEEE International Conference on Intelligent Transportation Systems (ITSC)_ , 2019, pp. 3514–3519.
* [28] E. Schmerling, K. Leung, W. Vollprecht, and M. Pavone, “Multimodal probabilistic model-based planning for human-robot interaction,” in _IEEE International Conference on Robotics and Automation (ICRA)_ , 2018, pp. 3399–3406.
* [29] D. Isele, R. Rahimi, A. Cosgun, K. Subramanian, and K. Fujimura, “Navigating occluded intersections with autonomous vehicles using deep reinforcement learning,” in _IEEE International Conference on Robotics and Automation (ICRA)_ , 2018, pp. 2034–2039.
* [30] A. Cosgun, L. Ma, J. Chiu, J. Huang, M. Demir, A. M. Anon, T. Lian, H. Tafish, and S. Al-Stouhi, “Towards full automated drive in urban environments: A demonstration in gomentum station, california,” in _IEEE Intelligent Vehicles Symposium (IV)_ , 2017, pp. 1811–1818.
* [31] M. M. Minderhoud and P. H. Bovy, “Extended time-to-collision measures for road traffic safety assessment,” _Accident Analysis & Prevention_, vol. 33, no. 1, pp. 89–97, 2001.
* [32] M. Bouton, A. Cosgun, and M. J. Kochenderfer, “Belief state planning for autonomously navigating urban intersections,” in _IEEE Intelligent Vehicles Symposium (IV)_ , 2017, pp. 825–830.
* [33] S. Brechtel, T. Gindele, and R. Dillmann, “Probabilistic decision-making under uncertainty for autonomous driving using continuous pomdps,” in _IEEE International Conference on Intelligent Transportation Systems (ITSC)_ , 2014, pp. 392–399.
* [34] K.-H. Lee, “A study on distance measurement module for driving vehicle velocity estimation in multi-lanes using drones,” _Applied Sciences_ , vol. 11, no. 9, p. 3884, 2021.
* [35] K. Han, E. Lee, M. Choi, and S. B. Choi, “Adaptive scheme for the real-time estimation of tire-road friction coefficient and vehicle velocity,” _IEEE/ASME Transactions on mechatronics_ , vol. 22, no. 4, pp. 1508–1518, 2017.
* [36] A. Kesting, M. Treiber, and D. Helbing, “Enhanced intelligent driver model to access the impact of driving strategies on traffic capacity,” _Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences_ , vol. 368, no. 1928, pp. 4585–4605, 2010.
* [37] E. Grefenstette, K. M. Hermann, M. Suleyman, and P. Blunsom, “Learning to transduce with unbounded memory,” _Advances in Neural Information Processing Systems (NeurIPS)_ , vol. 28, pp. 1828–1836, 2015.
* [38] J. Schulman, F. Wolski, P. Dhariwal, A. Radford, and O. Klimov, “Proximal policy optimization algorithms,” _arXiv preprint arXiv:1707.06347_ , 2017\.
|
Miyachi’s Theorem For the Quaternion Fourier Transform
Youssef El Haoui1,111Corresponding author., Said Fahlaoui1
1Department of Mathematics and Computer Sciences, Faculty of Sciences, Equipe
d’Analyse Harmonique et Probabilités, University Moulay Ismail, BP 1120,
Zitoune, Meknes, Morocco
E-MAIL<EMAIL_ADDRESS><EMAIL_ADDRESS>
###### Abstract
The quaternion Fourier transform (QFT) satisfies some uncertainty principles
similar to the Euclidean Fourier transform. In this paper, we establish
Miyachi’s theorem for this transform and consequently generalize and prove the
analogue of Hardy’s theorem and Cowling-Price uncertainty principle in the QFT
domain.
Key words: Quaternion Fourier transform, Miyachi’s theorem.
## 1 Introduction
Uncertainty principle (UP) is an important tool in harmonic anlysis; it states
that a nonzero function and its Fourier transform cannot both be very rapidly
decreased. UP has implications in different areas like quantum physics,
information processing, signal analysis, etc.
In signal analysis, it gives in general a lower bound for the simultaneous
localization of signals in phase and frequency spaces. There are many
advantageous ways to get the statement about localization precise; examples
include theorems of Hardy theorem [11], Beurling [13], and Miyachi [17] which
interpreted differently the localisation, as sharp pointwise estimates of a
signal and its Fourier transform. More precisely Miyachi’s theorem asserts
that if $f$ is a measurable function on $\mathbb{R}$ satisfying :
$\displaystyle e^{{\alpha x}^{2}}f\in
L^{1}({{\mathbb{R}}})+L^{\infty}({{\mathbb{R}}}),$
and
$\displaystyle\int_{{\mathbb{R}}}{{\log}^{+}}\left(\frac{\left|{\hat{f}}(y)e^{{\frac{{\pi}^{2}}{\alpha}y^{2}}}\right|}{\rho}\right)\
{\hbox{d}}y<\infty,$
for some positive constants $\alpha$ and $\rho$, where
$\displaystyle\log^{+}(x)={\left\\{\begin{array}[]{ll}\log(x),&{}\text{if
}x>1.\\\ 0&{}\text{otherwise}.\end{array}\right.}$
and $\hat{f}$ stands for the classical Fourier transform of $f$,
then $f$ is a constant multiple of the Gaussian $e^{{-\alpha x}^{2}}$.
Miyachi’s theorem has been extended in several different directions in recent
years, including extensions to Dunkl transform [5], Clifford–Fourier
transform[9], and much more generally, to nilpotent lie groups [1] and
Heisenberg motion groups [2].
The quaternion Fourier transform (QFT) is a non-trivial extension of the real
and complex classical Fourier transform to the algebra of the quaternions.
Since the quaternion multiplication is non-commutative, there are three types
of the QFT depending on which side multiplication of the kernel is done, that
is the so-called left-sided, right-sided and the two-sided QFT.
The QFT was introduced at first by Ell [10] for the analysis linear time-
invariant partial differential systems and then applied in color image
processing.
Later, Bülow [3] investigated the important properties of the two-sided QFT
for real signals and applied it to signal and image processing. Furthermore,
several uncertainty principles have been formulated for the quaternion Fourier
transform.
In [15], the authors generalized a component-wise UP for the right-sided QFT.
The directional UP related to the two-sided QFT was proposed in [12].
Recently, in [4], the authors established logarithmic UP associated with the
QFT. Meanwhile, Mawardi [16] obtained the connection between the QFT and
quantum mechanics and then established the modified UP (full UP) for the two-
sided QFT.
Our contribution to these developments is that we propose a new UP for the
QFT, namely Miyachi’s UP.
So far, no such uncertainty principle for the QFT (one-sided or two-sided) had
been established. In our previous works, other UPs: Heisenberg, Hardy[7], and
Beurling[8], have been extended for the two-sided QFT. Also, we derived in [8]
the UPs of Cowling-Price and Hardy using the extension of Beurling theorem in
a quaternion framework. Here, we will obtain, in a different way, by the main
result of Miyachi, the same UPs of Cowling-Price and Hardy in QFT domain. The
techniques used here are also applicable for the left-sided and the right-
sided QFT as well.
Our paper is organized as follows. In Sect. 2, we review some basic notions
and notations related to the quaternion algebra. In Sect. 3, we recall the
definition and some results for the quaternion Fourier transform useful in the
sequel. In Sect. 4, we prove Miyachi’s theorem for the quaternion Fourier
transform, and provide an extension of certain UPs to the quaternion Fourier
transform domain. In Sect. 5, we conclude the paper.
## 2 The Algebra of Quaternions
In order to extend complex numbers to a four-dimensional algebra, the Irish W.
R. Hamilton invented in 1843 the quaternion algebra $\mathbb{H}$.
Any quaternion $q\in\mathbb{H}$ can be expressed by
$q=q_{0}+\bm{i}q_{1}+\bm{j}q_{2}+\bm{k}q_{3};\
q_{0},q_{1},q_{2},q_{3}\in\mathbb{R},$
where $\bm{i},\bm{j},\bm{k}$ satisfy Hamilton’s rules
$\displaystyle\bm{i}^{2}=\bm{j}^{2}=\bm{k}^{2}=-1,\
\bm{i}\bm{j}=-\bm{j}\bm{i}=\bm{k},$ $\bm{j}\bm{k}=-\bm{k}\bm{j}=\bm{i};\
\bm{k}\bm{i}=-\bm{i}\bm{k}=\bm{j}.$
Quaternions are isomorphic to the Clifford algebra ${Cl}_{(0,2)}$ of
${\mathbb{R}}^{(0,2)}$:
$\mathbb{H}\cong Cl_{(0,2)}.$ (2.1)
We define the conjugation of $q\in\mathbb{H}$ by:
$\overline{q}\mathrel{\overset{\makebox[0.0pt]{\mbox{\tiny
def}}}{=}}q_{0}-\bm{i}q_{1}-\bm{j}q_{2}-\bm{k}q_{3}$
and its modulus $|q|_{Q}$ is defined by
${|q|}_{Q}\mathrel{\overset{\makebox[0.0pt]{\mbox{\tiny
def}}}{=}}\sqrt{q\overline{q}}=\sqrt{q^{2}_{0}+q^{2}_{1}+q^{2}_{2}+q^{2}_{3}}.$
Particularly, when $q=q_{0}$ is a real number, the module ${|q|}_{Q}$ reduces
to the ordinary Euclidean module $\left|q\right|=\sqrt{q^{2}_{0}}$. Also, we
observe that for $\bm{x}\in{\mathbb{R}}^{2},\
{\left|\bm{x}\right|}_{Q}=\left|\bm{x}\right|,$ where $\left|.\right|\ $is the
Euclidean norm $\left|(x_{1},x_{2})\right|^{2}=x^{2}_{1}+x^{2}_{2}$
Moreover, for arbitrary $p,q\in\mathbb{H}$ the following identity holds
${|pq|}_{Q}={|p|}_{Q}{|q|}_{Q}.$
Clearly, the inverse of $0\neq q\in\mathbb{H}$ is defined by :
$q^{-1}=\frac{\overline{q}}{|q|^{2}_{Q}}.$
which shows that $\mathbb{H}$ is a normed division algebra.
Due to (2.1), we recall the following properties:
if $\bm{x}$ is a vector in ${Cl}_{(0,2)}$, then
${\left|\bm{x}\right|}^{2}=-\bm{x}^{2}.$ (2.2)
Let (,) be the inner product on ${\mathbb{R}}^{(0,2)}$; for 2.3rs $\bm{x}$ and
$\bm{y}$ we have
$(\bm{x},\bm{y})~{}\mathrel{\overset{\makebox[0.0pt]{\mbox{\tiny
def}}}{=}}\sum^{2}_{l=1}{x_{l}}y_{l}=-\frac{1}{2}(\bm{x}\bm{y}+\bm{y}\bm{x}).$
(2.3)
In this paper, we will study the quaternion-valued signal
$f:\mathbb{R}^{2}\to\mathbb{H}$, which can be written in this form
$f=f_{0}+\bm{i}f_{1}+\bm{j}f_{2}+\bm{k}f_{3},$
with $f_{m}~{}:{\mathbb{R}^{2}}\to\ {\mathbb{R}}\ for\ m=0,1,2,3.$
We introduce the Banach spaces
$L^{p}\left({\mathbb{R}}^{2},\mathbb{H}\right)$, $1\leq p\leq\infty$,
$L^{p}\left({\mathbb{R}}^{2},\mathbb{H}\right)\mathrel{\overset{\makebox[0.0pt]{\mbox{\tiny
def}}}{=}}\
\\{f|f:\mathbb{R}^{2}\rightarrow\mathbb{H},{\left|f\right|}_{p,Q}\mathrel{\overset{\makebox[0.0pt]{\mbox{\tiny
def}}}{=}}({\int_{{\mathbb{R}}^{2}}{{\left|f(\bm{x})\right|}^{p}_{Q}\
d\bm{x}}})^{\frac{1}{p}}<\infty\\},1\leq p<\infty,$
$L^{\infty}\left({\mathbb{R}}^{2},\mathbb{H}\right)\mathrel{\overset{\makebox[0.0pt]{\mbox{\tiny
def}}}{=}}\
\\{f|f:\mathbb{R}^{2}\rightarrow\mathbb{H},{\left|f\right|}_{\infty,Q}\mathrel{\overset{\makebox[0.0pt]{\mbox{\tiny
def}}}{=}}ess\
{sup}_{\bm{x}\in\mathbb{R}^{2}}{\left|f(\bm{x})\right|}_{Q}<\infty\\}.$
where $d\bm{x}=dx_{1}dx_{2},$ refers to the usual Lebesgue measure in
$\mathbb{R}^{2}.$
If $f\ {\in L}^{\infty}\left({\mathbb{R}}^{2},\mathbb{H}\right)$ is
continuous, then
${\left|f\right|}_{\infty,Q}={sup}_{{\bm{x}\in{\mathbb{R}}}^{2}}{{\left|f(\bm{x})\right|}_{Q}}$
Furthermore, we define naturally the two following Banach spaces
$L^{1}(\mathbb{R}^{2},\mathbb{H})\cap
L^{\infty}(\mathbb{R}^{2},\mathbb{H})=\\{f|f\in
L^{1}(\mathbb{R}^{2},\mathbb{H})\ \textnormal{and}\ f\in
L^{\infty}(\mathbb{R}^{2},\mathbb{H})\\},$
$L^{1}(\mathbb{R}^{2},\mathbb{H})+L^{\infty}(\mathbb{R}^{2},\mathbb{H})=\\{f=f_{1}+f_{2},f_{1}\in
L^{1}(\mathbb{R}^{2},\mathbb{H}),\ f_{2}\in
L^{\infty}(\mathbb{R}^{2},\mathbb{H})\\}.$
We denote by ${\mathcal{S}}(\mathbb{R}^{2},\mathbb{H})$ the quaternion
Schwartz test function space, i.e., the set $C^{\infty}$ of smooth functions
$f$, from ${{\mathbb{R}}}^{2}$ to $\mathbb{H}$, given by
$\mathcal{S}(\mathbb{R}^{2},\mathbb{H})\mathrel{\overset{\makebox[0.0pt]{\mbox{\tiny
def}}}{=}}\\{f\in
C^{\infty}(\mathbb{R}^{2},\mathbb{H}):{sup}_{\bm{x}\in\mathbb{R}^{2},\
{|\alpha|\leq
n}}{({\left(1+\left|\bm{x}\right|^{m}\right)}\partial^{\mathbf{\alpha}}{\left|f(\bm{x})\right|}_{Q})}<\infty,\
m,n\in{\mathbb{N}}\\},$
where $\partial^{\mathbf{\alpha}}\mathrel{\overset{\makebox[0.0pt]{\mbox{\tiny
def}}}{=}}\frac{\partial^{|\alpha|}}{\partial^{\alpha_{1}}_{x_{1}}\partial^{\alpha_{2}}_{x_{2}}}$,
$|\alpha|=\alpha_{1}+\alpha_{2}$ for a multi-index
$\alpha=({\alpha}_{1},{\alpha}_{2})\in\mathbb{N}^{2}$.
## 3 Quaternion Fourier Transform
In this section, we review the definition and some properties of the two-sided
QFT.
###### Definition 3.1.
Let$\ f$ in $L^{1}\left({\mathbb{R}}^{2},\mathbb{H}\right)$. Then, the two-
sided quaternion Fourier transform of the function $f$ is given by
$\mathcal{F}\\{f\\}(\bm{\xi})\mathrel{\overset{\makebox[0.0pt]{\mbox{\tiny
def}}}{=}}\int_{{\mathbb{R}}^{2}}{e^{-\bm{i}{2\pi\xi}_{1}x_{1}}}f(\bm{x})e^{-\bm{j}{2\pi\xi}_{2}x_{2}}d\bm{x},$
(3.1)
where $\bm{\xi},\bm{x}\in{\mathbb{R}}^{2}.$
###### Lemma 3.2.
Inverse QFT [3, Thm. 2.5]
If $f\in L^{2}\left(\mathbb{R}^{2},\mathbb{H}\right),and\
\mathcal{F}\\{f\\}\in L^{1}\left(\mathbb{R}^{2},\mathbb{H}\right)$, then the
two-sided QFT is an invertible transform and its inverse is given by
$f(\bm{x})=\int_{\mathbb{R}^{2}}{e^{\bm{i}2\pi{\xi}_{1}x_{1}}}\mathcal{F}\\{f\\}(\bm{\xi})e^{\bm{j}{2\pi\xi}_{2}x_{2}}d\bm{\xi},\
\ d\bm{\xi}=d\xi_{1}d\xi_{2}.$
###### Lemma 3.3.
Scaling property
Let $\alpha$ be a positive scalar constant; then, the two-sided QFT of
${f}_{\alpha}\left({\bm{x}}\right)={f}(\alpha{\bm{x}})$ becomes
$\mathcal{F}\left\\{f_{\alpha}\right\\}(\bm{\xi})={\left(\frac{1}{\alpha}\right)}^{2}\mathcal{F}\left\\{f\right\\}(\frac{1}{\alpha}\bm{\xi}).$
(3.2)
Proof. Equation (3.1) gives
$\displaystyle\mathcal{F}\\{{{f}}_{\alpha}\\}(\bm{\xi})=\int_{{\mathbb{R}}^{2}}{e^{-\bm{i}{2\pi\xi}_{1}x_{1}}}\
f(\alpha\bm{x})\ e^{-\bm{j}{2\pi\xi}_{2}x_{2}}d\bm{x}.$
We substitute $\bm{y}$ for $\alpha\bm{x}$ and get
$\displaystyle\mathcal{F}\\{{{f}}_{\alpha}\\}(\bm{\xi})$ $\displaystyle=$
$\displaystyle{\left(\frac{1}{\alpha}\right)}^{2}\int_{{\mathbb{R}}^{2}}{e^{-\bm{i}2\pi(\frac{1}{\alpha}{\xi}_{1})y_{1}}}f(\bm{y})e^{-\bm{j}2\pi(\frac{1}{\alpha}{\xi}_{2})y_{2}}d\bm{y}$
$\displaystyle=$
$\displaystyle{\left(\frac{1}{\alpha}\right)}^{2}\mathcal{F}\left\\{f\right\\}(\frac{1}{\alpha}\bm{\xi}).$
The next lemma states that the QFT of a Gaussian quaternion function is
another quaternion Gaussian quaternion function.
###### Lemma 3.4.
QFT of a Gaussian quaternion function.
Consider a two-dimensional Gaussian quaternion function $f$ given by
$f(\bm{x})=qe^{-(\alpha_{1}x_{1}^{2}+\alpha_{2}x_{2}^{2})},$
where $q=q_{0}+iq_{1}+jq_{2}+kq_{3}$ is a constant quaternion, and
$\alpha_{1},\alpha_{2}$ are positive real constants.
Then
$\mathcal{F}\\{f\\}(\bm{\xi})=q\frac{\pi}{\sqrt{\alpha_{1}\alpha_{2}}}e^{-\frac{\pi^{2}}{\alpha_{1}}\xi_{1}^{2}-\frac{\pi^{2}}{\alpha_{2}}\xi_{2}^{2}}.$
(3.3)
where $\bm{x},\bm{\xi}\in{\mathbb{R}}^{2}.$
Proof. Let $g$ be defined by
$g(\bm{x})\mathrel{\overset{\makebox[0.0pt]{\mbox{\tiny
def}}}{=}}e^{-(\alpha_{1}x_{1}^{2}+\alpha_{2}x_{2}^{2})},$ we have
$\mathcal{F}\\{f\\}(\bm{\xi})=q_{0}\mathcal{F}\\{g\\}(\bm{\xi})+\bm{i}q_{1}\mathcal{F}\\{g\\}(\bm{\xi})+q_{2}\mathcal{F}\\{g\\}(\bm{\xi})\bm{j}+q_{3}\bm{i}\mathcal{F}\\{g\\}(\bm{\xi})\bm{j},$
where we used the $\mathbb{R}$-linearity of the QFT and the properties
$\mathcal{F}\\{\bm{i}h\\}=\bm{i}\mathcal{F}\\{h\\},\mathcal{F}\\{\bm{j}h\\}=\mathcal{F}\\{h\\}\bm{j},$
and $\mathcal{F}\\{\bm{k}h\\}=\bm{i}\mathcal{F}\\{h\\}\bm{j}$ for $h$ real-
valued function.
On the other hand, we have
$\displaystyle\mathcal{F}\\{g\\}(\bm{\xi})$ $\displaystyle=$
$\displaystyle\int_{{\mathbb{R}}^{2}}{e^{-\bm{i}{2\pi\xi}_{1}x_{1}}}g(\bm{x})e^{-\bm{j}{2\pi\xi}_{2}x_{2}}d\bm{x}$
$\displaystyle=$
$\displaystyle\int_{\mathbb{R}}e^{-\alpha_{1}x_{1}^{2}}{e^{-\bm{i}{2\pi\xi}_{1}x_{1}}}dx_{1}\int_{\mathbb{R}}e^{-\alpha_{2}x_{2}^{2}}e^{-\bm{j}{2\pi\xi}_{2}x_{2}}dx_{2}$
$\displaystyle=$
$\displaystyle\sqrt{\frac{\pi}{\alpha_{1}}}e^{-\frac{\pi^{2}}{\alpha_{1}}\xi_{1}^{2}}\sqrt{\frac{\pi}{\alpha_{2}}}e^{-\frac{\pi^{2}}{\alpha_{2}}\xi_{2}^{2}}$
$\displaystyle=$
$\displaystyle\frac{\pi}{\sqrt{\alpha_{1}\alpha_{2}}}e^{-\frac{\pi^{2}}{\alpha_{1}}\xi_{1}^{2}-\frac{\pi^{2}}{\alpha_{2}}\xi_{2}^{2}}.$
This completes the proof of Lemma 3.4.∎
###### Lemma 3.5.
[8, Lemma 3.11]
Let $f:\mathbb{R}^{2}\to\mathbb{R}$ be of the form
$f\left(\bm{x}\right)=P(\bm{x})e^{-\pi\alpha{\left|\bm{x}\right|}^{2}},$
where $P$ is a polynomial and $\alpha>0,$
Then
$\mathcal{F}\\{f\\}(\bm{\xi}{\rm)}=\ Q(\bm{\xi})\
e^{-\frac{\pi}{\alpha}{\left|\bm{\xi}\right|}^{2}},$
where $Q$ is a quaternion polynomial with $degP=degQ.$
## 4 Miyachi’s theorem
In this section, we prove Miyachi’s theorem for the quaternion Fourier
transform. For this, we need the following technical lemma of the complex
analysis.
###### Lemma 4.1.
[5, Lemma 1]
Let $h$ be an entire function on ${\mathbb{C}}^{2}$ such that
${\left|h\left(\bm{z}\right)\right|}\leq{Ae}^{B{\left|Re(\bm{z})\right|}^{2}}\
for\ all\ \bm{z}\in{\mathbb{C}}^{2}$
and
$\int_{{\mathbb{R}}^{2}}{{\log}^{+}}\left({\left|h\left(\bm{y}\right)\right|}\right)d\bm{y}<\infty,$
for some positive constants $A$ and $B$.
Then $h$ is a constant function.
###### Theorem 4.2.
(Miyachi’s Theorem).
Let $\alpha$, $\beta>0$. Suppose that $f$ is a measurable function such that
$e^{{\alpha\left|\bm{x}\right|}^{2}}f\in
L^{1}({\mathbb{R}}^{2},\mathbb{H})+L^{\infty}({\mathbb{R}}^{2},\mathbb{H}),$
(4.1)
and
$\int_{{\mathbb{R}}^{2}}{{\log}^{+}}(\frac{{\left|\mathcal{F}\left\\{f\right\\}(\bm{y})e^{{\beta\left|\bm{y}\right|}^{2}}\right|}_{Q}}{\rho})\
d\bm{y}<\infty,$ (4.2)
for some $\rho,\ 0<\rho<+\infty$
Then, three cases can occur:
1. (i)
If $\alpha\beta>{\pi}^{2},\ $ then $\ f=0$ almost everywhere.
2. (ii)
If $\ \alpha\beta={\pi}^{2},\ $then
$f\left(\bm{x}\right)=Ce^{{-\alpha\left|\bm{x}\right|}^{2}},$ where $C$ is a
constant quaternion.
3. (iii)
$If\ \alpha\beta<{\pi}^{2},\ $then there exist infinitely many functions
satisfying (4.1) and (4.2).
Proof.
1. (i)
We first prove the result for the case $\alpha\beta={\pi}^{2}.$
By scaling, we can assume that $\alpha=\beta=\pi.$
Indeed, let $g\left(\bm{x}\right)=f(\sqrt{\frac{\pi}{\alpha}}\ \bm{x})$; then,
by Lemma 3.3 we obtain
$\mathcal{F}\left\\{g\right\\}\left(\bm{t}\right)=\frac{\alpha}{\pi}\
\mathcal{F}\left\\{f\right\\}(\sqrt{\frac{\alpha}{\pi}}\bm{t}),\ \ \
\bm{t}\in\mathbb{R}^{2}.$ (4.3)
In addition, we get by (4.1)
$\displaystyle
e^{{\pi\left|\bm{x}\right|}^{2}}g(\bm{x})=e^{{\pi\left|\bm{x}\right|}^{2}}f(\sqrt{\frac{\pi}{\alpha}}\
\bm{x})\in
L^{1}({\mathbb{R}}^{2},\mathbb{H})+L^{\infty}({\mathbb{R}}^{2},\mathbb{H}).$
Also, we have
$\displaystyle\int_{\mathbb{R}^{2}}{{\log}^{+}}\left(\frac{{\left|\mathcal{F}\left\\{f\right\\}\left(\bm{y}\right)e^{{\beta\left|\bm{y}\right|}^{2}}\right|}_{Q}}{\rho}\right)d\bm{y}$
$\displaystyle=$
$\displaystyle\frac{\alpha}{\pi}\int_{{\mathbb{R}}^{2}}{{\log}^{+}}\left(\frac{{\left|\mathcal{F}\left\\{f\right\\}\left(\sqrt{\frac{\alpha}{\pi}}\bm{t}\right)e^{{\frac{\alpha\beta}{\pi}\left|\bm{t}\right|}^{2}}\right|}_{Q}}{\rho}\right)d\bm{t}$
$\displaystyle\overset{\makebox{\mbox{\eqref{revision}}}}{=}$
$\displaystyle\frac{\alpha}{\pi}\int_{{\mathbb{R}}^{2}}{{\log}^{+}}\left(\frac{{\left|\mathcal{F}\left\\{g\right\\}\left(\bm{t}\right)e^{{\pi\left|\bm{t}\right|}^{2}}\right|}_{Q}}{\rho^{{}^{\prime}}}\right)d\bm{t}<\infty,$
where ${\rho}^{{}^{\prime}}=\frac{\alpha}{\pi}\rho.$
If the result is shown for $\alpha=\beta=\pi,$
then $g\left(\bm{x}\right)=Ce^{{-\pi\left|\bm{x}\right|}^{2}}$, and thus,
$f\left(\bm{x}\right)=\ g\left(\sqrt{\frac{\alpha}{\pi}}\bm{x}\right)=C\
e^{{-\alpha\left|\bm{x}\right|}^{2}}.$
Now, we assume that $\alpha=\beta=\pi:$
Applying the same method as in [7, Thm. 5.3], by complexifying the variable
$\bm{z}=\bm{a}+i_{{\mathbb{C}}}\bm{b},$
where $\bm{a}=(a_{1},a_{2}),\bm{b}=(b_{1},b_{2})\in\mathbb{R}^{2}$, and we
note by $i_{{\mathbb{C}}}$ the complex number which satisfies
$i^{2}_{\mathbb{C}}=-1.$
We have
${\left|\bm{z}\right|}^{2}_{Q}={\left|\bm{a}\right|}^{2}+{\left|\bm{b}\right|}^{2}={\left|\bm{z}\right|}^{2},$
where $\left|.\right|\ $is the Euclidean norm in ${\mathbb{C}}^{2}.$
Let
$\displaystyle w\left({\bm{x}}\right)$ $\displaystyle=$
$\displaystyle{{e}}^{\pi({\left|{{a}}_{1}\right|^{2}{+}\left|{{b}}_{1}\right|^{2})}}{{e}}^{\pi({\left|{{a}}_{2}\right|^{2}{+}\left|{{b}}_{2}\right|^{2})}}{{e}}^{{-}\pi({\left|x_{1}\right|-(\left|{{a}}_{1}\right|{+}\left|{{b}}_{1}\right|))}^{2}}{{e}}^{{-}\pi({\left|x_{2}\right|-(\left|{{a}}_{2}\right|{+}\left|{{b}}_{2}\right|))}^{2}}$
$\displaystyle=$ $\displaystyle
e^{\pi{\left|{z}\right|}^{2}}{{e}}^{{-}\pi({\left|x_{1}\right|-(\left|{{a}}_{1}\right|{+}\left|{{b}}_{1}\right|))}^{2}}{{e}}^{{-}\pi({\left|x_{2}\right|-(\left|{{a}}_{2}\right|{+}\left|{{b}}_{2}\right|))}^{2}}.$
Clearly, $w$ belongs to $L^{1}(\mathbb{R}^{2},\mathbb{H})\cap
L^{\infty}(\mathbb{R}^{2},\mathbb{H}).$
By assumption, $e^{{\pi\left|\bm{x}\right|}^{2}}f$ belongs to
$L^{1}$(${{\mathbb{R}}}^{2}{,\mathbb{H})+}L^{\infty}$(${{\mathbb{R}}}^{2}{,\mathbb{H})}.$
As
$\mathcal{F}\left\\{f\right\\}\left(\bm{z}\right)=\int_{{{\mathbb{R}}}^{2}}{e^{-2\pi\bm{i}x_{1}{(}a_{1}{+}i_{\mathbb{C}}b_{1})}f\left(\bm{x}\right)e^{-2\pi\bm{j}{\bm{x}}_{2}(a_{2}+i_{\mathbb{C}}b_{2})}d\bm{x}},$
we have
$\displaystyle\left|\mathcal{F}\left\\{f\right\\}\left(\bm{z}\right)\right|_{Q}$
$\displaystyle\leq$
$\displaystyle\int_{{{\mathbb{R}}}^{2}}{{\left|f\left(\bm{x}\right)\right|}_{Q}}e^{2\pi(\left|x_{1}a_{1}\right|+\left|x_{1}b_{1}\right|{+}\left|x_{2}a_{2}\right|+\left|x_{2}b_{2}\right|)}d\bm{x}$
$\displaystyle=$
$\displaystyle\int_{\mathbb{R}^{2}}{{\left|e^{{\pi\left|\bm{x}\right|}^{2}}f\left({\bm{x}}\right)\right|}_{Q}\
}{w}\left({\bm{x}}\right)d\bm{x}.$
Hence, $\mathcal{F}\left\\{{f}\right\\}\left(\bm{z}\right)$ is well defined,
and is an entire function on ${\mathbb{C}}^{2}.$
Furthermore, by (4.1) there exists $u\in
L^{1}({{\mathbb{R}}}^{2}{,\mathbb{H})}$ and $v\in
L^{\infty}({{\mathbb{R}}}^{2}{,\mathbb{H})}$ such that
$e^{{\pi\left|\bm{x}\right|}^{2}}f\left({\bm{x}}\right)=u\left({\bm{x}}\right)+v\left({\bm{x}}\right),$
Using the triangle inequality and the linearity of the integral we get
${\left|\mathcal{F}\left\\{f\right\\}\left(\bm{z}\right)\right|}_{Q}\leq\int_{\mathbb{R}^{2}}{{\left|u\left({\bm{x}}\right)\right|}_{Q}}{w}\left({\bm{x}}\right)d{\bm{x}}+\int_{\mathbb{R}^{2}}{{\left|v\left({\bm{x}}\right)\right|}_{Q}\
}{w}\left({\bm{x}}\right)d{\bm{x}}.$
Then according to the Hôlder’s inequality, we have
${\left|\mathcal{F}\left\\{f\right\\}\left(z\right)\right|}_{Q}\leq{\left|u\right|}_{1,Q}{\left|w\right|}_{\infty,Q}+{\left|v\right|}_{\infty,Q}{\left|w\right|}_{1,Q},$
Since
$\int_{\mathbb{R}}{e^{-\pi({\left|t\right|+m)}^{2}}}dt\leq 2,\ \
\textnormal{where}\ m\in{\mathbb{R}},$
we obtain
${\left|w\right|}_{1,Q}\leq 4e^{\pi{\left|\bm{z}\right|}^{2}},\
\textnormal{and}\ {\left|w\right|}_{\infty,Q}\leq
e^{\pi{\left|\bm{z}\right|}^{2}}.$
Then
$\displaystyle{\left|\mathcal{F}\left\\{f\right\\}\left(\bm{z}\right)\right|}_{Q}$
$\displaystyle\leq$
$\displaystyle{e^{\pi{\left|\bm{z}\right|}^{2}}\left|u\right|}_{1,Q}+4\
e^{\pi{\left|\bm{z}\right|}^{2}}{\left|v\right|}_{\infty,Q}$
$\displaystyle\leq$ $\displaystyle K\ e^{\pi{\left|\bm{z}\right|}^{2}},$
where $K$ is a positive constant independent of $z.$
Now, let
$h(\bm{z})=e^{-\pi
z^{2}}|\mathcal{F}\left\\{f\right\\}\left(\bm{z}\right)|_{Q},\
\textnormal{for}\ \bm{z}\in{\mathbb{C}}^{2},$
then, $h$ is an entire function.
By (2.2) and (2.3), we have
$\bm{z}^{2}=\left(\bm{a}+i_{\mathbb{C}}\bm{b}\right)\left(\bm{a}+i_{\mathbb{C}}\bm{b}\right)=-{\left|\bm{a}\right|}^{2}+{\left|\bm{b}\right|}^{2}-2i_{\mathbb{C}}(\bm{a},\bm{b}),$
then
${\left|e^{-\pi\bm{z}^{2}}\right|}_{Q}\leq
e^{\pi{\left|\bm{a}\right|}^{2}}e^{-\pi{\left|\bm{b}\right|}^{2}},$
where we used ${\left|e^{2\pi i_{\mathbb{C}}(\bm{a},\bm{b})}\right|}_{Q}=1$.
As a result
$\left|h\left(z\right)\right|\leq{Ke}^{\pi{\left|a\right|}^{2}}e^{-\pi{\left|b\right|}^{2}}e^{\pi{\left|a\right|}^{2}}e^{\pi{\left|b\right|}^{2}},$
$=Ke^{2\pi{\left|a\right|}^{2}}.$ (4.4)
On the other hand, by (2.2) and by assumption
$\int_{{\mathbb{R}}^{2}}{{\log}^{+}}\left(\frac{\left|h\left(\bm{y}\right)\right|}{\rho}\right)d\bm{y}=\int_{{\mathbb{R}}^{2}}{{\log}^{+}}\left(\frac{{\left|e^{\pi{\left|\bm{y}\right|}^{2}}\
\mathcal{F}\left\\{f\right\\}\left(\bm{y}\right)\right|}_{Q}}{\rho}\right)d\bm{y}<\infty.$
(4.5)
Then by (4.4),(4.5) and by applying Lemma 4.1 to the function
$h\left(\bm{y}\right){/}\rho$, we deduce that $h\left(\bm{y}\right){=const.}$
i.e
$|\frac{\mathcal{F}\left\\{f\right\\}\left(\bm{y}\right)}{e^{-\pi{\left|\bm{y}\right|}^{2}}}|_{Q}=\
const.$
Therefore
$\mathcal{F}\left\\{f\right\\}\left(\bm{y}\right)=Ce^{-\pi{\left|\bm{y}\right|}^{2}},$
where $C$ is a constant quaternion.
Then by Lemmas 3.2 and 3.4, we have
$f\left(\bm{x}\right)=C\ e^{-\pi{\left|\bm{x}\right|}^{2}}.$
2. (ii)
If $\alpha\beta>{\pi}^{2}.$
Let $g\left(\bm{x}\right)=f(\sqrt{\frac{\pi}{\alpha}}\ \bm{x})$, a simple
calculation shows that
$\displaystyle\int_{\mathbb{R}^{2}}{{\log}^{+}}\left(\frac{{\left|\mathcal{F}\left\\{g\right\\}\left(t\right)e^{{\pi\left|t\right|}^{2}}\right|}_{Q}}{\rho^{{}^{\prime}}}\right)dt$
$\displaystyle<$
$\displaystyle\frac{\pi}{\alpha}\int_{{\mathbb{R}}^{2}}{{\log}^{+}}\left(\frac{{\left|\mathcal{F}\left\\{f\right\\}\left(\bm{y}\right)e^{{\beta\left|\bm{y}\right|}^{2}}\right|}_{Q}}{\rho}\right)d\bm{y}$
$\displaystyle<$ $\displaystyle\infty,\ \ \ \ \ \ (\textnormal{by}\
\eqref{4.2})$
where ${\rho}^{{}^{\prime}}=\frac{\alpha}{\pi}\ \rho.$
Then, according to the first case
$g\left(\bm{x}\right)=C\ e^{-\pi{\left|\bm{x}\right|}^{2}},$
where $C$ is a constant quaternion.
Consequently
${f}\left({\bm{x}}\right)=C\ e^{-\alpha{\left|\bm{x}\right|}^{2}}.$
Hence, by Lemma 3.4 we get
$\mathcal{F}\left\\{f\right\\}\left(\bm{y}\right)=Ce^{-\frac{{\pi}^{2}}{\alpha}{\left|\bm{y}\right|}^{2}}.$
Refering to (4.2), $C$ must be zero.
3. (iii)
For the final case, $\alpha\beta<{\pi}^{2}$
Let ${f\left(\bm{x}\right)=\varphi}_{k,l}(\bm{x})\
e^{-\pi\gamma{\left|\bm{x}\right|}^{2}}$ with
$\frac{\alpha}{\pi}$$<$$\gamma$$<$$\frac{\pi}{\beta},$
where ${{\\{\varphi}_{k,l}\\}}_{k,l\in{\mathbb{N}}}$ is a basis of
${\mathcal{S}}({\mathbb{R}}^{2},\mathbb{H})$, which is defined by
${\varphi}_{k,l}\left(x_{1},x_{2}\right)\mathrel{\overset{\makebox[0.0pt]{\mbox{\tiny
def}}}{=}}{\varphi}_{k}\left(x_{1}\right){\varphi}_{l}\left(x_{2}\right),$
for $\left(x_{1},x_{2}\right)\in\mathbb{R}^{2},$
and
${\varphi}_{k}\left(x\right)=\frac{{(-1)}^{k}}{k!}e^{\pi
x^{2}}\frac{d^{k}}{{dx}^{k}}(e^{-{2}\pi\bm{x}^{2}}),x\in\mathbb{R}.$
It is important to see that${{\\{\varphi}_{k,l}\\}}_{k,l\in\mathbb{N}}$ is a
basis of ${\mathcal{S}}(\mathbb{R}^{2},\mathbb{H})$ (see [6, 7]).
We have
$\displaystyle e^{\alpha{\left|\bm{x}\right|}^{2}}f$ $\displaystyle=$
$\displaystyle
e^{\alpha{\left|\bm{x}\right|}^{2}}{\varphi}_{k,l}\left(\bm{x}\right)e^{-\pi\gamma{\left|\bm{x}\right|}^{2}}$
$\displaystyle=$
$\displaystyle{\varphi}_{k,l}\left(\bm{x}\right)e^{(\alpha-\pi\gamma){\left|\bm{x}\right|}^{2}}\in
L^{1}(\mathbb{R}^{2}{,\mathbb{H})+}L^{\infty}(\mathbb{R}^{2},\mathbb{H}).$
Lemma 3.5 implies
$\mathcal{F}\left\\{f\right\\}\left(\bm{y}\right)=Q(\bm{y})\
e^{-\frac{\pi}{\gamma}{\left|\bm{y}\right|}^{2}},$
where $Q$ is a quaternion polynomial.
Then, since $\beta<\frac{\pi}{\alpha}$,
$\displaystyle\int_{\mathbb{R}^{2}}{{\log}^{+}}\left(\frac{{\left|\mathcal{F}\left\\{f\right\\}\left(\bm{y}\right)e^{{\beta\left|\bm{y}\right|}^{2}}\right|}_{Q}}{\rho}\right)d\bm{y}$
$\displaystyle=$
$\displaystyle\int_{\mathbb{R}^{2}}{{\log}^{+}}\left(\frac{{\left|Q(\bm{y})\
e^{{(\beta-\frac{\pi}{\gamma})\left|\bm{y}\right|}^{2}}\right|}_{Q}}{\rho}\right)d\bm{y}$
$\displaystyle<$ $\displaystyle\infty.$
This completes the proof of Theorem 4.2.∎
In the following, we illustrate the effectiveness of Theorem 4.2, by giving an
example, and derive two generalizations of uncertainty principle associated
with the QFT.
###### Example 4.3.
Consider $\alpha,\beta$ two positive numbers with $\alpha\beta=\pi^{2},$ and
the quaternion Gaussian function
$f(\bm{x})=qe^{-\alpha|\bm{x}|^{2}},$
where $q=q_{0}+\bm{i}q_{1}+\bm{j}q_{2}+\bm{k}q_{3}$ is a constant quaternion.
Obviously, we have
$e^{\alpha|\bm{x}|^{2}}f=q\in L^{\infty}(\mathbb{R}^{2},\mathbb{H})\subset
L^{1}(\mathbb{R}^{2},\mathbb{H})+L^{\infty}(\mathbb{R}^{2},\mathbb{H}).$
Thus, $f$ satisfies condition (4.1).
By lemma 3.4, we get
$\mathcal{F}\\{f\\}(\bm{y})=q\frac{\pi}{\alpha}e^{-\beta|\bm{y}|^{2}}.$
Then
$\int_{{\mathbb{R}}^{2}}\log^{+}(\frac{|\mathcal{F}\\{f\\}(\bm{y})|_{Q}e^{\beta|\bm{y}|^{2}}}{\rho})d\bm{y}=\int_{{\mathbb{R}}^{2}}\log^{+}(\frac{|q|_{Q}\frac{\pi}{\alpha}}{\rho})d\bm{y}<\infty,$
whenever $\rho>|q|_{Q}\ \frac{\pi}{\alpha}.$
###### Corollary 4.4.
Hardy’s uncertainty principle for the QFT.
Let $\alpha$ and $\beta$ be both positive constants. Suppose $f$ be in
$L^{1}(\mathbb{R}^{2},\mathbb{H})$ with
1. (i)
$|f(\bm{x})|^{2}<Ce^{-\alpha|\bm{x}|^{2}}.$
2. (ii)
$|\mathcal{F}\\{f\\}(\bm{y})|^{2}<C^{{}^{\prime}}e^{-\beta|\bm{y}|^{2}}.$
for some constants $C>0$ and $C^{\prime}>0$. Then, three cases can occur :
1. 1.
If $\alpha\beta>\pi^{2},$ $f=0$ almost everywhere on $\mathbb{R}^{2}.$
2. 2.
If $\alpha\beta=\pi^{2},$ then $f$ is a constant quaternion multiple of
$e^{-\alpha|\bm{x}|^{2}}$.
3. 3.
If $\alpha\beta<\pi^{2},$ there are infinitely many linearly independent
functions satisfying both conditions $(i)$ and $(ii).$
Proof. Immediately using the decay condition $(i)$ one has
$fe^{\alpha|\bm{x}|^{2}}\in L^{\infty}(\mathbb{R}^{2},\mathbb{H}).$ Hence $f$
verifies condition (4.1) of Theorem 4.2.
Moreover, for $\rho>0$ we have
$\frac{|\mathcal{F}\\{f\\}(\bm{y})|_{Q}e^{\beta|\bm{y}|^{2}}}{\rho}\leq\frac{C^{{}^{\prime}}}{\rho}.$
Thus
$\log^{+}(\frac{|\mathcal{F}\\{f\\}(\bm{y})|_{Q}e^{\beta|\bm{y}|^{2}}}{\rho})\leq\underbrace{\log^{+}(\frac{C^{{}^{\prime}}}{\rho})}_{=0},$
whenver $\rho>C^{{}^{\prime}}.$
So condition (4.2) of Theorem 4.2 is verified. Then, direct application of
Theorem 4.2 enables us to achieve the proof.
###### Corollary 4.5.
Cowling-Price’s uncertainty principle for the QFT.
Let $\alpha$ and $\beta$ be positive real numbers, $1\leq p,q\leq\infty$ such
that $min(p,q)$ is finite, and let $f$ are a square integrable quaternion-
valued function satisfying the following decay conditions Suppose $f$ be in
$L^{1}(\mathbb{R}^{2},\mathbb{H})$
1. (i)
$\int_{\mathbb{R}^{2}}(|f(\bm{x})|_{Q}e^{\alpha|\bm{x}|^{2}})^{p}d\bm{x}<\infty.$
2. (ii)
$\int_{\mathbb{R}^{2}}(|\mathcal{F}\\{f\\}(\bm{y})|_{Q}e^{\beta|\bm{y}|^{2}})^{q}d\bm{y}<\infty.$
Then the three following conclusions hold:
1. 1.
$f=0$ almost everywhere whenever $\alpha\beta>\pi^{2}.$
2. 2.
If $\alpha\beta=\pi^{2},$ then $f$ is a constant quaternion multiple of
$e^{-\alpha|\bm{x}|^{2}}$.
3. 3.
If $\alpha\beta<\pi^{2},$ there are infinitely many linearly independent
functions satisfying both conditions $(i)$ and $(ii).$
Proof. According to (i), we get $fe^{\alpha|\bm{x}|^{2}}\in
L^{p}(\mathbb{R}^{2},\mathbb{H})$, and using the fact that $L^{p}\subset
L^{1}+L^{\infty}$ we obtain that $f$ fulfills condition (4.1) of Theorem 4.2.
Furthermore, based on the inequality
$\displaystyle\log^{+}(x)\leq x\ \ \textnormal{for}\ x\in\mathbb{R}_{+},$
we can easily see that $f$ satisfies the second condition (4.2).
Therefore, by applying Theorem 4.2 we conclude the proof.
## 5 Conclusions
In this paper, based on some obtained results of the two-sided QFT and one
technical lemma of the complex analysis, a generalization of Miyachi’s
uncertainty principle associated with the QFT was proposed. Consequently, two
variants of this UP were provided, namely the theorems of Hardy and Cowling-
Price. The extension of these qualitative UPs to the quaternionic algebra
framework shows that a quaternionic 2D signal $f$ and its QFT cannot both
simultaneously decrease very rapidly.
The QFT has proved to be very significant tool for applications in color image
processing[14], quantum mechanics, engineering, signal processing, optics,
etc. Apart from their importance to pure mathematics, our results are also
relevant to applied mathematics and signal processing.
## References
* [1] A. Baklouti, S. Thangavelu, Variants of Miyachi’s theorem for nilpotent lie groups. J. Aust. Math. Soc. 88, 1–17 (2010). https://doi.org/10.1017/S144678870900038X
* [2] A. Baklouti and S. Thangavelu, Hardy and Miyachi theorems for Heisenberg motion groups, Nagoya Math. J. 229, 1–20 (2016). https://doi.org/10.1017/nmj.2016.58
* [3] T. Bülow, Hypercomplex spectral signal representations for the processing and analysis of images, Ph.D. Thesis, Institut für Informatik und Praktische Mathematik, University of Kiel, Germany (1999)
* [4] L.P. Chen, K.I. Kou, M.S. Liu, Pitt’s inequality and the uncertainty principle associated with the quaternion Fourier transform. J. Math. Anal. Appl. 423, 681–700 (2015)
* [5] F. Chouchene, R. Daher, T. Kawazoe, H. Mejjaoli, Miyachi’s theorem for the Dunkl transform. Integral Transform. Spec. Funct. 22, 167–173 (2011)
* [6] H. De Bie, New techniques for two-sided quaternion Fourier transform. In: Procedings of AGACSE (2012).
* [7] Y. El Haoui, S. Fahlaoui, The Uncertainty principle for the two-sided quaternion Fourier transform. Mediterr. J. Math. (2017). https://doi.org/10.1007/s00009-017-1024-5
* [8] Y. El Haoui, S. Fahlaoui, Beurling’s theorem for the quaternion Fourier transform. J. Pseudo-Differ. Oper. Appl. (2019). https://doi.org/10.1007/s11868-019-00281-7
* [9] J. El Kamel, R. Jday, Uncertainty Principles for the Clifford–Fourier Transform, Adv. Appl. Clifford Algebras (2017) https://doi.org/10.1007/s00006-017- 0791-1
* [10] T.A. Ell, Quaternion-Fourier transforms for analysis of two-dimensional linear time-invariant partial differential systems. In: Proceeding of the 32nd Conference on Decision and Control, San Antonio, Texas, pp. 1830–1841 (1993).
* [11] G.H. Hardy, A theorem concerning Fourier transform, J. Lond. Math. Soc. 8, 227–231 (1933)
* [12] E. Hitzer, Directional uncertainty principle for quaternion Fourier transform. Adv. Appl. Clifford Algebras 20, 271–284 (2010)
* [13] L. Hörmander, A uniqueness theorem of Beurling for Fourier transform pairs. Ark. Math. 2, 237–240 (1991)
* [14] K.M. Hosny, Y.M. Khedr, W.I. Khedr et al., Robust color image hashing using quaternion polar complex exponential transform for image authentication. Circuits Syst. Signal Process. 37, 5441 (2018). https://doi.org/10.1007/s00034-018-0822-8
* [15] B. Mawardi, E. Hitzer, A. Hayashi, R. Ashino, An uncertainty principle for quaternion fourier transform. Comput. Math. Appl. 56, 2398–2410 (2008)
* [16] B. Mawardi, A modified uncertainty principle for two-sided quaternion Fourier transform. Adv. Appl. Clifford Algebras 26(2), 513–527 (2016)
* [17] A. Miyachi, A generalization of theorem of Hardy, Harmonic AnalysisSeminar held at Izuna-gaoka, Shizuoka-Ken, Japon, pp. 44–51 (1997)
|
# On The Temporal Evolution of Particle Production in $f(T)$ Gravity
Sanjay Mandal 0000-0003-2570-2335 Department of Mathematics,
Birla Institute of Technology and Science-Pilani,
Hyderabad Campus, Hyderabad-500078, India
<EMAIL_ADDRESS>P.K. Sahoo 0000-0003-2130-8832 Department of
Mathematics,
Birla Institute of Technology and Science-Pilani,
Hyderabad Campus, Hyderabad-500078, India
<EMAIL_ADDRESS>
###### Abstract
The thermodynamical study of the universe allow particle production in
modified $f(T)$ ($T$ is the torsion scalar) theory of gravity within a flat
FLRW framework for line element. The torsion scalar $T$ plays the same role as
the Ricci scalar $R$ in the modified theories of gravity. We derived the
$f(T)$ gravity models by taking $f(T)$ as the sum of $T$ and an arbitrary
function of $T$ with three different arbitrary function. We observe that the
particle production describes the accelerated expansion of the universe
without a cosmological constant or any unknown “quintessence” component. Also,
we discussed the supplementary pressure, particle number density and particle
production rate for three cases.
###### keywords:
Modified $f(T)$ gravity; Particle creation; Thermodynamics
Received (22 July 2020)Accepted (05 October 2020)
PACS Nos.: 04.50.kd
## 1 Introduction
At the beginning of the 19th century, the General Theory of Relativity brought
the revolution to the modern cosmology proposed by Albert Einstein. The
Riemannian space-time formulates this theory based on the Levi-Civita
connection, a torsion-free, and metric compatibility connection. It also helps
us to understand the geodesic structure of the Universe. Later on, it faced
problems like fine-tuning, cosmic co-incidence, initial singularity,
cosmological constant, and flatness [1], since modern cosmology is growing by
a prominent number of accurate observations. Besides, the cosmological
observation, such as type Ia supernovae [2, 3], cosmic microwave background
(CMB) radiation [4, 5], large scale structure [6, 7], baryon acoustic
oscillations [8], and weak lensing [9] confirms that currently, our Universe
is going through an accelerated expansion phase that happens because of the
highly negative pressure produced by the unknown form of matter and energy,
called dark energy and dark matter. To overcome the above issues, researchers
started to modify the Einstein’s theory of relativity, and they ended up with
several modified theories of gravity such as $f(R)$ gravity [10], $f(R,T)$
gravity [11], $f(T)$ gravity [12], $f(Q)$ gravity [13], $f(Q,T)$ gravity [14],
etc. As a result, cosmologists found many interesting results such as Yousaf
et al., studied the self-gravitating structures [15], gavastars [16], Sahoo et
al. [17] studied the wormhole geometry, bouncing cosmology, and accelerated
expansion of the universe using modified theories of gravity. The main
advantage of the modified theories of gravity is that it successfully
describes the late-time cosmic acceleration and the early time inflation. In
the early stage of the Universe, there is a possibility of particle creation.
In this study, we are focusing on particle production in the teleparallel
gravity.
The $f(T)$ theories of gravity are the generalization of the Teleparallel
Equivalent of General Relativity (TEGR), where $T$ is the torsion scalar [18].
TEGR was first presented by Einstein. In the context of $f(T)$ theories, the
Reimann-Cartan space-time requires the torsional curvature to vanish.
Furthermore, this type of space-time is constructed by the Weitzenböck
connection [19, 20]. TEGR is equivalent to GR; the reason is that both cases’
action is the same except the surface term in TEGR. But, the physical
interpretation is different from each other. The construction of gravitational
Lagrangian in TEGR formulation was done in [21, 22]. TEGR is Lorentz invariant
theories, whereas the modified $f(T)$ theories are not Lorentz invariant.
Also, the motion equations in $f(T)$ gravity are not necessarily Lorentz
invariant, because the certainty is that this theory’s property explains the
recent interests [23, 24, 25, 26]. Moreover, the modification of TEGR was
motivated by the $f(R)$ gravity theory. In the teleparallel gravity theory, we
use the Weitzenböck connection instead of the Levi-Civita connection, which
uses $f(R)$ gravity to vanish the non-zero torsion curvature. Here, we would
like to mention that the $f(T)$ gravity does not need the Equivalence
principle because the Weitzenböck connection describes its gravitational
interaction. It is a simple modified theory compared to other modified
theories because the torsion scalar $T$ contains only the first-order
derivatives of the vierbeins. In contrast, Ricci scalar $R$ contains the
second-order derivatives of the metric tensors. Recently, Mandal et al.,
studied the acceleration expansion of the Universe using the parametrization
technique with presuming exponential and logarithmic form of $f(T)$ [27] in
$f(T)$ gravity. They also studied a complete cosmological scenario of the
Universe in $f(T)$ gravity, where they discussed the difference between
General Relativity and Teleparallel gravity [28] in $f(T)$ gravity. M. Sharif
and S. Rani studied the dynamical instability ranges in Newtonian as well as
post-Newtonian regimes considering power-law $f(T)$ model with anisotropic
fluid in $f(T)$ gravity [29]. Cai et al., studied the matter bounce cosmology
using perturbation technique and they found a scale-invariant power spectrum,
which is consistent with cosmological observations in $f(T)$ gravity [30]. In
[31], the wormhole solutions with non-commutative geometry have been studied
assuming power-law $f(T)$ model and a particular shape function in
teleparallel gravity. Inflationary universe studied using power-law $f(T)$
function and logamediate scale factor in [32], and constant-roll inflation
studied in [33].
In the early stage of the universe, the possibility of particle creation has
been discussed for curved space-time by Schrodinger [34], Dewitt [35], Imamura
[36]. Later, the first ever particle creation was treated by an external
gravitational field by Parkar [37, 38]. In flat space-time, the unique vacuum
state is identifying by the guidance of Lorentz invariance. Moreover, we do
not have Lorentz symmetry in curved space-time. In general, there are more
than one vacuum state exists in a curved space-time. Therefore, the particle
creation idea becomes open to discuss, but it’s physical interpretation
becomes more difficult [39, 40]. The interaction between the dynamical
external gradients causes the particle creation from the vacuum. The particle
creation produces negative pressure, so it is considered to explain the
accelerated expansion of the universe and got some unexpected outcomes. Also,
it might play the role of unknown gradients of the universe. In [41, 42]
studied the particle creation with SNe Ia data. Singh [43], and Singh and
Beesham [44, 45] studied the particle creation with some kinematical tests in
FLRW cosmology. The continuous creation of particle predicts the assumptions
of standard Big Bang cosmology.
The thermodynamical study of black hole gives the fundamental relation between
thermodynamics and gravitation [46, 47, 48, 49, 50]. In GR, the relation
between the entropy and the horizon area with the Einstein equation derives
from the Clausius relation in thermodynamics [51, 52]. This idea is also used
for other theories, mainly, the generalized thermodynamics laws and modified
theories of gravity which are derived from the GR [53, 54]. Among the modified
theories of gravity, $f(R)$ gravity got more attention on this framework.
Thereby, one can obtained the gravitational field equation through the non-
equilibrium feature of thermodynamics by using the Clausius approach. There
are some work have been done in the thermodynamics of particle creation in
$f(T)$ gravity theory [56, 57, 55, 58, 50].
In this work, we study the theoretical significance of particle creation in
$f(T)$ gravity theory considering a flat FRW model. Assuming $f(T)$ as the sum
of torsion scalar $T$ and an arbitrary function of torsion scalar $T$, we
studied the thermodynamics of particle creation with $f(T)=0$ is a simple
teleparallel gravity, $f(T)=A(-T)^{q}$ as power law gravity and
$f(T)=A(1-e^{-qT})$ as exponential gravity. After that we discussed the
behaviour of supplementary pressure $p_{c}$, particle number density $n$, and
the particle creation rate $\psi$ for three models. Also we compared the
effect of the cosmological pressure $p_{m}$ with the supplementary pressure
$p_{c}$ for different values of equation of state parameter $\omega$ on
particle creation.
This work is organised as follows. In Sec. 2, we discussed the thermodynamics
of particle creation, which is followed by the overview of $f(T)$ gravity and
it’s field equations in Sec. 3. In Sec. 4, we discussed three $f(T)$ gravity
models. Finally, the results are summarized in Sec. 5
## 2 Thermodynamics of particle creation
If we assume the total number of particles in the universe to be conserved,
the laws of thermodynamics can be expressed as
$dQ=d(\rho_{m}V)+p_{m}dV$ (1)
and
$TdS=p_{m}dV+d(\rho_{m}V)$ (2)
where $p_{m}$, $\rho_{m}$, $V$, $T$ and $S$ denote respectively the
cosmological pressure, density, volume, temperature and entropy. Also, $dQ$
represent the heat exchange in the time interval $dt$. From (1) and (2), we
further obtain,
$dQ=TdS$ (3)
Eq. (3) reflects the fact that the entropy is a conserved quantity, since for
lose adiabatic system $dQ=0$. We now consider a scenario in which the total
number of particles in the universe is not constant. Under this condition, Eq
(1) gets modified to [59]
$dQ=d(\rho_{m}V)+p_{m}dV+(h/n)d(nV)$ (4)
where $N=nV$, $n$ being the number density of the particles and
$h=(p_{m}+\rho_{m})$ the enthalpy per unit volume of the system. For an
adiabatic system where $dQ=0$, (4) reads [59]
$d(\rho_{m}V)+p_{m}dV=(h/n)d(nV)$ (5)
In [59], the authors stated that in cosmology this change in the total number
of particles in the universe can be understood as a transformation of
gravitational field energy to the matter.
For an open thermodynamic system, Eq. (5) can be expressed as [59]
$d(\rho_{m}V)=-\left(p_{m}+p_{c}\right)dV$ (6)
where
$p_{c}=-(h/n)(dN/dV)$ (7)
represents supplementary pressure associated with the creation of particles
[59]. Note that negative $p_{c}$ indicate production of particles whereas
positive $p_{c}$ implies particle annihilation and finally for $p_{c}=0$ the
total number of particles is constant. Using Eq (2) and (5), it can also be
shown that [59]
$S=S_{0}\left(\frac{N}{N_{0}}\right)$ (8)
where $S_{0}$ and $N_{0}$ represent current values of these quantities.
Additionally, we assume the particles follow a barotropic equation of state
and therefore can be written as
$p_{m}=(\omega)\rho_{m}$ (9)
where $-1\leq\omega\leq 1$ is the EoS parameter. The number density of
particles is related to the density $\rho_{{}_{m}}$ as [45]
$n=n_{0}\left(\frac{\rho_{m}}{\rho_{0}}\right)^{\frac{1}{1+\omega}}$ (10)
where $\rho_{0}\geq 0$ and $n_{0}\geq 0$ are the present values of density and
particle number density respectively.
We now consider the matter creation rate to be defined as [60]
$\psi(t)=3\beta nH$ (11)
where $0\leq\beta\leq 1$ is assumed to be a constant and $\psi(t)$ represent
the rate of particle creation and has a dimension of $t^{-1}$. $\psi$ can
either be positive or negative depending on the creation or annihilation of
particles. $\psi=0$ indicate particle number being conserved in the universe.
For cosmological matter following barotropic equation of state (Eq. 9), the
supplementary pressure $p_{c}$ can be expressed as [59]
$p_{c}=-\beta(\omega+1)\rho_{m}$ (12)
## 3 Overview of $f(T)$ Gravity
Let us consider the extension of Einstein-Hilbert Lagrangian of $f(T)$ theory
of gravity (which is similar to $f(R)$ gravity extension from the Ricci scalar
$R$ to $R+f(R)$ in the action), namely the teleparallel gravity term $T$ to
$T+f(T)$, where $f(T)$ is an arbitrary function of $T$ as
$S=\frac{1}{16\pi G}\int[T+f(T)]ed^{4}x,$ (13)
where $e=det(e^{i}_{\mu})=\sqrt{-g}$ and $G$ is the gravitational constant.
Assume $k^{2}=8\pi G=M_{p}^{-1}$, where $M_{p}$ is the Planck mass.The
gravitational field is defined by the torsion one as
$T^{\gamma}_{\mu\nu}\equiv
e^{\gamma}_{i}(\partial_{\mu}e^{i}_{\nu}-\partial_{\nu}e^{i}_{\mu}).$ (14)
The contracted form of the above torsion tensor is
$T\equiv\frac{1}{4}T^{\gamma\mu\nu}T_{\gamma\mu\nu}+\frac{1}{2}T^{\gamma\mu\nu}T_{\nu\mu\gamma}-T^{\gamma}_{\gamma\mu}T^{\nu\mu}_{\nu}.$
(15)
By the variation of the total action $S+L_{m}$, here $L_{m}$ is the matter
Lagrangian gives us the field equation for $f(T)$ gravity as
$e^{-1}\partial_{\mu}(ee^{\gamma}_{i}S^{\mu\nu}_{\gamma})(1+f_{T})-(1+f_{T})e^{\lambda}_{i}T^{\gamma}_{\mu\lambda}S^{\nu\mu}_{\gamma}\\\
+e^{\gamma}_{i}S^{\mu\nu}_{\gamma}\partial_{\mu}(T)f_{TT}+\frac{1}{4}e^{\nu}_{i}[T+f(T)]=\frac{k^{2}}{2}e^{\gamma}_{i}T^{(M)\nu}_{\gamma},$
(16)
where $f_{T}=df(T)/dT$, $f_{TT}=d^{2}f(T)/dT^{2}$, the ”superpotential “
tensor $S^{\mu\nu}_{\gamma}$ written in terms of cotorsion
$K^{\mu\nu}_{\gamma}=-\frac{1}{2}(T^{\mu\nu}_{\gamma}-T^{\nu\mu}_{\alpha}-T^{\mu\nu}_{\alpha})$
as
$S^{\mu\nu}_{\gamma}=\frac{1}{2}(K^{\mu\nu}_{\gamma}+\delta^{\mu}_{\gamma}T^{\alpha\nu}_{\alpha}-\delta^{\nu}_{\gamma}T^{\alpha\mu}_{\alpha})$
and $T^{(M)\nu}_{\gamma}$ represents the energy-momentum tensor to the matter
Lagrangian $L_{m}$. Now we consider a flat FLRW universe with the metric as
$ds^{2}=dt^{2}-a^{2}(t)dx^{\mu}dx^{\nu},$ (17)
where $a(t)$ is the scale factor, which gives us
$e^{i}_{\mu}=diag(1,a,a,a).$ (18)
Using equation (18) into the field equation (16), we get the modified field
equation as follows
$H^{2}=\frac{8\pi G}{3}\rho_{m}-\frac{f}{6}+\frac{Tf_{T}}{3},$ (19)
$\dot{H}=-\left[\frac{4\pi G(\rho_{m}+p_{m}+p_{c})}{1+f_{T}+2Tf_{TT}}\right],$
(20)
where $H\equiv\dot{a}/a$ be the Hubble parameter and ”dot“ represents the
derivative with respect to $t$. Here, $\rho_{m}$ and $p_{m}$ be the energy
density and pressure of the matter content, $p_{c}$ be the supplementary
pressure. Also, we have used
$T=-6H^{2},$ (21)
which holds for a FLRW Universe according to equation (15).
## 4 $f(T)$ gravity models
In this section we shall investigate the temporal evolution of particle
production in radiation ($\omega=1/3$) and dust universe ($\omega=0$) for
various $f(T)$ gravity models with model parameters constrained from
cosmological observations related to gravitational baryogenesis.
For the purpose of analysis, we shall assume a power law evolution of scale
factor of the form
$a(t)=a_{0}t^{\left[\frac{2}{3(1+\omega)}\right]}$ (22)
where $a_{0}>0$ is a constant.
### 4.1 Simple Teleparallel gravity
In simple teleparallel equivalent of general relativity [61], where $f(T)=0$ ,
for a universe composed of perfect fluid, the field equations (19) and (20)
becomes
$H^{2}=\frac{8\pi G}{3}\rho_{m}$ (23) $\dot{H}=-4\pi G(\rho_{m}+p_{m}+p_{c})$
(24)
Substituting (22) in (23), we obtain the expression of density $\rho_{m}$ as
$\rho_{m}=\frac{4}{3}\left(\frac{1}{t^{2}(1+\omega)^{2}}\right)$ (25)
The expression of supplementary pressure $p_{c}$, particle number density $n$
and particle creation rate $\psi$ are obtained respectively as
$p_{c}=\frac{4}{3}\left(\frac{\beta(1+\omega)}{t^{2}(1+\omega)^{2}}\right)$
(26)
$n=\left[\frac{4}{3}\left(\frac{1}{t^{2}(1+\omega)^{2}}\right)\right]^{\frac{1}{(1+\omega)}}$
(27)
$\psi=3\beta\left[\frac{2}{3t(1+\omega)}\right]\times\left[\frac{4}{3}\left(\frac{1}{t^{2}(1+\omega)^{2}}\right)\right]^{\frac{1}{(1+\omega)}}$
(28)
Figure 1: The behaviour of supplementary pressure $p_{c}$ with respect to
cosmic time $t$ for $\omega=\frac{1}{3},\omega=0$ and $\beta=1$. Figure 2: The
behaviour of particle number density $n$ with respect to cosmic time $t$ for
$\omega=\frac{1}{3},\omega=0$ and $\beta=1$. Figure 3: The behaviour of
particle creation rate $\psi$ with respect to cosmic time $t$ for
$\omega=\frac{1}{3},\omega=0$ and $\beta=1$.
### 4.2 Power Law Gravity
The power law model of Bengochea and Ferraro [62] reads
$f(T)=A(-T)^{q}$ (29)
where $A$ is a constant and $q>1$. In [63], the authors reported viable
baryon-to-entropy ratio for $A=-10^{-7}\texttt{or}-10^{-6}$ and $q\gtrsim
4.8$. However, other values of the model parameters could also yield viable
estimates of baryon-to-entropy ratio. Nonetheless, we restrict ourselves to
the values $A=-10^{-7}$ and $q=5$ for the present analysis. Substituting (22)
and (29) in (19) and (20), the expression of density $\rho_{m}$ reads
$\rho_{m}=A6^{(q-1)}(1-18q)t^{-2q}\left[\frac{2}{3(1+\omega)}\right]^{2q}$
(30)
The expression of supplementary pressure $p_{c}$, particle number density $n$
and particle creation rate $\psi$ for the power law gravity are obtained
respectively as
$p_{c}=-\beta(1+\omega)A6^{(q-1)}(1-18q)t^{-2q}\left[\frac{2}{3(1+\omega)}\right]^{2q}$
(31)
$n=\left[A6^{(q-1)}(1-18q)t^{-2q}\left[\frac{2}{3(1+\omega)}\right]^{2q}\right]^{1/(1+\omega)}$
(32)
$\psi=3\beta\left[\frac{2}{3t(1+\omega)}\right]\times\left[A6^{(q-1)}(1-18q)t^{-2q}\left[\frac{2}{3(1+\omega)}\right]^{2q}\right]^{1/(1+\omega)}$
(33)
Figure 4: The behaviour of supplementary pressure $p_{c}$ with respect to
cosmic time $t$ for $\omega=\frac{1}{3},\omega=0$ and $\beta=1,q=5,A=-10^{-7}$
Figure 5: The behaviour of particle number density $p_{c}$ with respect to
cosmic time $t$ for $\omega=\frac{1}{3},\omega=0$ and
$\beta=1,q=5,A=-10^{-7}$. Figure 6: The behaviour of particle creation rate
$\psi$ with respect to cosmic time $t$ for $\omega=\frac{1}{3},\omega=0$ and
$\beta=1,q=5,A=-10^{-7}$.
### 4.3 Exponential Gravity
The exponential $f(T)$ model given in [64] reads
$f(T)=A(1-e^{-qT})$ (34)
where $A$ and $q$ are model parameters. In [63] the authors reported a wide
range of values of $A$ and $q$ for which a viable baryon-to-entropy ratio
could be realized. However, we shall work with $A=1$ and $q=10^{-10}$ as these
values were used in [63] to fit the baryon-to-entropy ratio with observations.
Substituting (22) and (34) in (19) and (20), the expression of density
$\rho_{m}$ reads
$\rho_{m}=\left[\frac{18A\left[\frac{2}{3(1+\omega)}\right]^{2}qe^{\frac{6\left[\frac{2}{3(1+\omega)}\right]^{2}q}{t^{2}}}}{t^{2}}\right]$
(35)
The expression of supplementary pressure $p_{c}$, particle number density $n$
and particle creation rate $\psi$ for the exponential gravity are obtained
respectively as
$p_{c}=-\beta(1+\omega)\left[\frac{18A\left[\frac{2}{3(1+\omega)}\right]^{2}qe^{\frac{6\left[\frac{2}{3(1+\omega)}\right]^{2}q}{t^{2}}}}{t^{2}}\right]$
(36)
$n=\left[\frac{18A\left[\frac{2}{3(1+\omega)}\right]^{2}qe^{\frac{6\left[\frac{2}{3(1+\omega)}\right]^{2}q}{t^{2}}}}{t^{2}}\right]^{1/(1+\omega)}$
(37)
$\psi=3\beta\left[\frac{2}{3t(1+\omega)}\right]\times\left[\left(\frac{18A\left[\frac{2}{3(1+\omega)}\right]^{2}qe^{\frac{6\left[\frac{2}{3(1+\omega)}\right]^{2}q}{t^{2}}}}{t^{2}}\right)^{1/(1+\omega)}\right]$
(38)
Figure 7: The behaviour of supplementary pressure $p_{c}$ with respect to
cosmic time $t$ for $\omega=\frac{1}{3},\omega=0$ and
$\beta=1,q=10^{-10},A=1$. Figure 8: The behaviour of particle number density
$n$ with respect to cosmic time $t$ for $\omega=\frac{1}{3},\omega=0$ and
$\beta=1,q=10^{-10},A=1$. Figure 9: The behaviour of particle creation rate
$\psi$ with respect to cosmic time $t$ for $\omega=\frac{1}{3},\omega=0$ and
$\beta=1,q=10^{-10},A=1$.
## 5 Discussion of Outcomes and Conclusions
In this article, we have studied the thermodynamics of an open system with
particle creation of a flat FLRW universe in $f(T)$ theory of gravity. We have
constructed three cosmological models by assuming suitable functions for
$f(T)$ as $f(T)=0,f(T)=A(-T)^{q},f(T)=A(1-e^{-qT})$ and the particle creation
rate $\psi$. To analyze our models, we have considered the power law evolution
of the scale factor and studied the behaviour of physical quantities (i.e. the
supplementary pressure $p_{c}$, particle number density $n$, and the particle
creation rate $\psi$) through their graphical representations with respect to
cosmic time $t$ and some fixed values of $\beta$ in various phase of the
evolution of the universe. And, details of our cosmological models discussed
in the following.
In our simple Teleparallel gravity model, we have considered the minimal
coupling between matter and geometry. In Fig. 3, profiles of $\psi$ have been
shown. From Fig. 3, one can easily observe that the rate of particle creation
is high in the early time and tends to zero when $t$ tends to infinity. But,
the number of particle in the universe increases with cosmic time $t$ shown in
Fig. 2. The supplementary pressure $p_{c}$ has higher negative which shows
that the particle production is high during the early stage and tends to zero
when $t$ tends infinity, in Fig. 1. From this model we have concluded that the
evolution of the universe depends on the contribution of the particle
production.
In power law gravity and exponential gravity models, we have considered the
non-minimal coupling between matters. The profiles of $\psi,n$ and $p_{c}$
have been shown for the corresponding models. In Fig. 6,9, the particle
creation rate $\psi$ is high in the early stage and it tends to zero as cosmic
time $t$ tends to infinity. Also, the particle number density n in Fig. 5,8
goes to zero as cosmic time $t$ goes to infinity which concluded that the
expansion rate overcomes the rate particle creation as the supplementary
pressure in Fig. 4,7 is negative throughout the evolution of the universe in
different phases. The density parameters in three models shows that the
universe is open in the presence of particle creation in $f(T)$ theory of
gravity.
In summary, we have studied the cosmological models with particle production
in $f(T)$ theory of gravity to explore the current accelerated phenomenon of
the universe. We have found that the particle creation produces negative
pressure which may derive the accelerated expansion of the universe and play
the role of unknown matter called “dark energy” in $f(T)$ theory of gravity.
We may expect that the particle creation process be a constraint for the
unexpected observational outcomes. The new fact about this article is that the
particle creation is studied by the thermodynamics approach in $f(T)$ theory
of gravity.
Acknowledgements S.M. acknowledges Department of Science & Technology (DST),
Govt. of India, New Delhi, for awarding Junior Research Fellowship (File No.
DST/INSPIRE Fellowship/2018/IF180676). PKS acknowledges DST, New Delhi, India
for providing facilities through DST-FIST lab, Department of Mathematics,
BITS-Pilani, Hyderabad Campus where a part of this work was done. The authors
thank S. Bhattacharjee for stimulating discussions. We are very much grateful
to the honorable referee and the editor for the illuminating suggestions that
have significantly improved our work in terms of research quality and
presentation.
## References
* [1] V. Sahni, A. Starobinsky, Int. J. Mod. Phys. D 9, 373 (2000); T. Padmanabhan, Phys. Rep. 380, 235 (2003).
* [2] S. Perlmutter et al., Astrophys. J. 517, 565 (1999).
* [3] A. G. Riess et al., Astron. J. 116, 1009 (1998).
* [4] D. N. Spergel et al., Astrophys. J. Suppl. 148, 175 (2003).
* [5] E. Komatsu et al., Astrophys. J. Suppl. 192, 18 (2011).
* [6] M. Tegmark et al., Phys. Rev. D 69, 103501 (2004).
* [7] U. Seljak et al., Phys. Rev. D 71, 103515 (2005).
* [8] D. J. Eisenstein et al., Astrophys. J. 633, 560 (2005).
* [9] B. Jain, A. Taylor, Phys. Rev. Lett. 91, 141302 (2003).
* [10] S. Nojiri, S. D. Odintsov, Int. J. Geom. Meth. Mod. Phys. 4, 115 (2007).
* [11] T. Harko, F. S. N. Lobo, S. Nojiri, S. D. Odintsov, Phys. Rev. D 84, 024020 (2011).
* [12] Y. Cai et al., Rep. Prog. Phys. 79, 106901 (2016).
* [13] J. B. Jimenez, L. Heisenberg, T. Koivisto, Phys. Rev. D, 98 (2018) 044048.
* [14] Y. Xu et al., Eur. Phys. J. C., 79, 708 (2019).
* [15] Z. Yousaf etal., Int. J. Mod. Phys. D 29, 2050061 (2020); Z. Yousaf, Phys. Scripta 95, 075307 (2020); M. Z. Bhatti et al., Phys. Rev. D 101, 104029 (2020); Z. Yousaf, et al., Eur. Phys. J. Plus 135, 397 (2020).
* [16] M. Z. Bhatti, et al., Phys. Dark Univ. 29, 100561 (2020); Z. Yousaf et al., Phys. Dark Univ. 28, 100527 (2020).
* [17] P. H. R. S. Moraes, P. K. Sahoo, Phys. Rev. D 96, 044038 (2017); S. Bhattacharjee, P. K. Sahoo, Phys. Dark Univ. 28, 100537 (2020); S. Mandal, et al., Phys. Rev. D 102, 024057 (2020); S. Arora, P. K. Sahoo, Phys. Scripta 95, 095003 (2020).
* [18] Y. F. Cai et al.,Rept. Prog. Phys. 79, 106901 (2016).
* [19] A. Weitzenböck, Invariantentheorie, Noordhoff, Gronningen, (1923).
* [20] U. K. B. Vinckers et al., arXiv:2009.04353 [gr-qc].
* [21] J. W. Maluf, J. Math. Phys. 35, 335 (1994).
* [22] V.C. de Andrade, J. G. Pereira, Phys. Rev. D 56, 4689 (1997).
* [23] N. Tamanini, C. G. Boehmer, Phys. Rev. D 86, 044009 (2012).
* [24] X. H. Meng, Y. b. Wang, Eur. Phys. J. C 71, 1755 (2011).
* [25] R. Ferraro, F. Fiorini, Phys. Rev. D 84, 083518 (2011).
* [26] Á. de la Cruz-Dombriz et al., JCAP 1412, 048 (2014).
* [27] S. Mandal et al., Phys. Dark Univ. 28, 100551 (2020).
* [28] S. Mandal, P.K. Sahoo, Eur. Phys. J. Plus 135, 706 (2020).
* [29] M. Sharif, Shamaila Rani, Mon. Not. R. Astron. Soc. 440, 2255 (2014).
* [30] Y. F. Cai et al., Class. Quantum Grav. 28, 215011 (2011).
* [31] M. Sharif, Shamaila Rani, Phys. Rev. D 88, 123501 (2013).
* [32] K. Rezazadeh et al., Astrophys. J. 836, 228 (2017).
* [33] A. Awad et al., JCAP 2018, 026 (2018).
* [34] E. Schrodinger, Physica 6, 899 (1939).
* [35] B. S. DeWitt, Phys. Rev. 90, 357 (1953).
* [36] T. Imamura, Phys. Rev. 118, 1430, (1960).
* [37] L. Parker, Phys. Rev. Lett. 21, 562, (1968).
* [38] L. Parker, Phys. Rev. 183, 1057 (1969).
* [39] N. D. Birrell, P. C. W. Davies, “Quantum fields in curved space”, Cambridge University Press, (1982).
* [40] L. H. Ford, gr-qc/9707062.
* [41] W. Zimdahl et al., Phys. Rev. D 64, 063501 (2001).
* [42] Y. Qiang, T. Jiezhang, Y. Ze-Long, Astrophys. Space Sci. 311, 407 (2007).
* [43] C.P. Singh, Astrophys. Space Sci. 338, 411 (2012).
* [44] C.P. Singh, A. Beesham, Astrophys. Space Sci. 336, 469 (2011).
* [45] C.P. Singh, A. Beesham, Int. J. Theor. Phys. 51, 3951 (2012).
* [46] J. M. Bardeen, B. Carter, S. W. Hawking, Commun. Math. Phys. 31, 161 (1973).
* [47] J. D. Bekenstein, Phys. Rev. D 7, 2333 (1973).
* [48] S. W. Hawking, Commun. Math. Phys. 43, 199 (1975).
* [49] G. W. Gibbons, S. W. Hawking, Phys. Rev. D 15, 2738 (1977).
* [50] K. Bamba, C. Q. Geng, JCAP 1111, 008 (2011).
* [51] T. Jacobson, Phys. Rev. Lett. 75, 1260 (1995).
* [52] E. Elizalde, P. J. Silva, Phys. Rev. D 78, 061501 (2008).
* [53] R. Miaoa, M. Lib, Y. Miaoc, JCAP 11, 033 (2011).
* [54] K. Karami, A. Abdolmaleki, JCAP 04, 007 (2012).
* [55] C. P. Singh, V. Singh, arXiv:1408.0633 [gr-qc].
* [56] M.R. Setare, N. Mohammadipour, JCAP 01, 015 (2013).
* [57] K. Bamba, R. Myrzakulov, S. Nojiri, S. D. Odintsov, Phys. Rev. D 85, 104036 (2012).
* [58] I. G. Salako et al., JCAP 060, 1475 (2013).
* [59] V. Singh, C. P. Singh, Int. J. Theor. Phys. 55, 1257 (2016).
* [60] J.A.S. Lima, A.S. Germano, L.R.W. Abramo, Phys. Rev. D 53, 4287 (1996).
* [61] A. Einstein, Riemann-Geometrie mit Aufrechterhaltung des Begriffes des Fernparallelismus (Verlag der Akademie der Wissenschaften, Berlin, 1928), p.217; Neue Moglichkeit fur eine einheitliche Feldtheorie von Gravitation und Elektizitat (Verlag der Akademie der Wissenschaften, Berlin, 1928), p. 224; K. Hayashi & T. Shirafuji, Phys. Rev. D 19, 3524 (1979); K. Hayashi & T. Shirafuji, Phys. Rev. D 24, 3312 (1981); R. Aldrovandi & J.G. Pereira, Teleparallel Gravity: An Introduction (Springer, Dordrecht, Netherlands, 2013); J.W. Maluf, Ann. Phys. (Berlin) 525, 339 (2013).
* [62] G.R. Bengochea, R. Ferraro, Phys. Rev. D 79, 124019 (2009).
* [63] V.K. Oikonomou, E. N. Saridakis, Phys. Rev. D 94, 124005 (2016).
* [64] S. Nesseris, S. Basilakos, E.N. Saridakis, L. Perivolaropoulos, Phys. Rev. D 88, 103010 (2013).
|
# Self-Supervised Path Consistency Learning for HOI Detection
Jihwan Park1,2 SeungJun Lee1 Hwan Heo1 Hyeong Kyu Choi1 Hyunwoo J. Kim1,
1Department of Computer Science and Engineering, Korea University 2Kakao Brain
{jseven7071, lapal0413, gjghks950, imhgchoi<EMAIL_ADDRESS>
<EMAIL_ADDRESS>corresponding author.
# DP-Aug: Decoding Path Augmentation for Transformers in HOI Detection
Jihwan Park1,2 SeungJun Lee1 Hwan Heo1 Hyeong Kyu Choi1 Hyunwoo J. Kim1,
1Department of Computer Science and Engineering, Korea University 2Kakao Brain
{jseven7071, lapal0413, gjghks950, imhgchoi<EMAIL_ADDRESS>
<EMAIL_ADDRESS>corresponding author.
# Consistency Learning via Decoding Path Augmentation
for Transformers in Human Object Interaction Detection
Jihwan Park1,2 SeungJun Lee1 Hwan Heo1 Hyeong Kyu Choi1 Hyunwoo J. Kim1,
1Department of Computer Science and Engineering, Korea University 2Kakao Brain
{jseven7071, lapal0413, gjghks950, imhgchoi<EMAIL_ADDRESS>
<EMAIL_ADDRESS>corresponding author.
###### Abstract
Human-Object Interaction detection is a holistic visual recognition task that
entails object detection as well as interaction classification. Previous works
of HOI detection has been addressed by the various compositions of subset
predictions, _e.g_., Image $\rightarrow$ HO $\rightarrow$ I, Image
$\rightarrow$ HI $\rightarrow$ O. Recently, transformer based architecture for
HOI has emerged, which directly predicts the HOI triplets in an end-to-end
fashion (Image $\rightarrow$ HOI). Motivated by various inference paths for
HOI detection, we propose cross-path consistency learning (CPC), which is a
novel end-to-end learning strategy to improve HOI detection for transformers
by leveraging augmented decoding paths. CPC learning enforces all the possible
predictions from permuted inference sequences to be consistent. This simple
scheme makes the model learn consistent representations, thereby improving
generalization without increasing model capacity. Our experiments demonstrate
the effectiveness of our method, and we achieved significant improvement on
V-COCO and HICO-DET compared to the baseline models. Our code is available at
https://github.com/mlvlab/CPChoi.
## 1 Introduction
Human-Object Interaction (HOI) detection is a holistic visual recognition task
that includes detecting individual objects as <human, object>, while properly
classifying the type of <interaction>. Previous HOI detectors [15, 52, 31, 49]
were mainly built on object detection models. They commonly extend CNN-based
object detectors [45, 34, 42] with an additional head for interaction
classification, _e.g_., humans and objects are detected first, and their
interaction is associated subsequently.
To alleviate the high computation cost of such two-stage HOI detection
methods, one-stage models [52, 22, 33] have been proposed for faster
detection. These models perform interaction prediction and object detection in
parallel. They compensate for their lower performance with auxiliary
predictions for the HOI subsets, _i.e_., auxiliary predictions for subset
<human, interaction> or <object, interaction> may help HOI prediction through
post-processing. However, these works demand different network architectures
for each auxiliary prediction, due to strict disciplines for each network’s
input. Hence, to introduce flexibility, transformer-based architectures [23,
46, 6, 48] have recently been adopted for HOI detection. They reformulate the
HOI detection problem as a direct set prediction building on DETR [4].
(a) cycle consistency
(b) cross-task consistency
(c) cross-path consistency
Figure 1: Comparison on the variants of consistencies. The black line refers
to the main task function $f_{\mathcal{X}\mathcal{Y}}$, and the red, blue
lines refer to the pair of tasks trained to be consistent with each other. (a)
Cycle consistency enforces the composite function of
$f_{\mathcal{Y}\mathcal{X}}\circ f_{\mathcal{X}\mathcal{Y}}$ to be consistent
with identity function $f_{\mathcal{X}\mathcal{X}}$. (b) Cross-task
consistency requires an auxiliary pretrained network
$f_{\mathcal{Y}\mathcal{Y_{\text{a}}}}$, represented in dashed lines, to give
consistent outputs across tasks. (c) Cross-path consistency does not require
task-specific pretrained networks. The output of main task function
$f_{\mathcal{X}\mathcal{Y}}$ should be consistent with the composition of the
outputs from sub-task functions $f_{\mathcal{X}\mathcal{Y_{\text{s}}}}$ and
$f_{\mathcal{Y}_{\text{s}}\mathcal{Y}}\circ
f_{\mathcal{X}\mathcal{Y}_{\text{s}}}$.
Motivated by various inference paths in HOI detectors, we propose a simple yet
effective method to train HOI transformers. We augment the decoding paths with
respect to the possible prediction sequences of HOI triplets. Then, with the
cascade structure of transformers, an input query is sequentially decoded into
auxiliary sub-task outputs and the final output. The stage of each augmented
paths stage shares a decoder, in a multi-task learning fashion. We further
improve our method to leverage the augmented decoding paths by enforcing the
outputs from the various paths to be consistent. Accordingly, we propose
Cross-Path Consistency (CPC) Learning, which aims to predict HOI triplets
regardless of inference sequences.
Similar to cross-task consistency [55], cross-path consistency retains
inference path invariance. However, cross-path consistency learning does not
require additional pre-trained networks. In contrast to cross-task
consistency, which demands an auxiliary network to train the main task
$\mathcal{X}\rightarrow\mathcal{Y}$ (Figure 1-(b)), cross-path consistency
defines an auxiliary domain $\mathcal{Y}_{s}$ in between $\mathcal{X}$ and
$\mathcal{Y}$ (Figure 1-(c)). In other words, the main task
$\mathcal{X}\rightarrow\mathcal{Y}$ (_i.e_., Image$\rightarrow$HOI) is divided
into subtasks $\mathcal{X}\rightarrow\mathcal{Y}_{s}$ and
$\mathcal{Y}_{s}\rightarrow\mathcal{Y}$ (_e.g_.,
Image$\rightarrow$HO$\rightarrow$I). The main task function
$f_{\mathcal{X}\mathcal{Y}}$ is then trained by enforcing its output and the
composition of sub-task predictions to be consistent. Moreover, cross-path
consistency learning is temporarily adopted for training only.
Our training strategy can be generalized to any transformer based
architecture, and can be applied in an end-to-end method. Extensive
experiments show that HOI transformers trained with CPC learning strategy
achieves substantial improvements in two popular HOI detection benchmarks:
V-COCO and HICO-DET. The contribution of this work can be summarized as the
followings:
* •
We propose Cross-Path Consistency (CPC) learning, which is a novel end-to-end
learning strategy to improve transformers for HOI detection leveraging various
inference paths. In this learning scheme, we use Decoding-Path Augmentation to
generate various inference paths which are compositions of subtasks with a
shared decoder for effective training.
* •
Our training scheme achieves substantial improvements on V-COCO and HICO-DET
without increasing model capacity and inference time.
## 2 Related Works
### 2.1 Human Object Interaction Detection
Human-Object Interaction (HOI) detection has been proposed in [16]. Later,
human-object detectors have been improved using human or instance appearance
and their spatial relationship [15, 25, 12]. On the other hand, graph-based
approaches [44, 49, 11, 51] have been proposed to clarify the action between
the <human, object> pair.
HOI detection models based on only visual cues often suffer from the lack of
contextual information. Thus, recent works utilize external knowledge to
improve the quality of HOI detection. Human pose information extracted from
external models [3, 7, 19, 28] or linguistic priors and knowledge graph models
show meaningful improvement in performance [58, 36, 43, 54, 18, 14, 31, 57,
37].
Since the majority of the previous works are based on two-stage methods with
slower inference time, attempts for faster HOI detection by introducing simple
end-to-end multi-layer perceptrons [17], or directly detecting interaction
points [52, 33], or union regions [22, 20, 30] have been suggested.
### 2.2 Transformers in Computer Vision
Transformer has become the state-of-the-art method in many computer vision
tasks. In image classification, [9] has shown competitive performance on
ImageNet without any convolution layers. DeiT [48] applied knowledge
distillation to data-efficiently train the vision transformer. To extract
multi-scale image features, Swin Transformer [38] proposed shifted window
based self-attention modules that effectively aggregate small patches to
increase the receptive field. In the object detection task, DETR [4] has
proposed an end-to-end framework eliminating the need for hand-designed
components. DETR’s bipartite matching loss between the predicted set and the
ground truth labels enables direct set prediction at inference. Recently,
DETR’s late convergence problem has been tackled in [62, 40, 13].
Inspired by DETR, transformer-based HOI (Human-Object Interaction) detectors
[23, 46, 6, 63, 8] have been recently proposed. HOI transformer models have
two types of structure, one decoder model and the two decoder model. The one-
decoder model which follows the structure of DETR [4] predicts triplets from
the output of a single decoder. QPIC [46] and HoiT [63] are one-decoder models
that output <human, object, interaction> triplets directly with multiple
interaction detection heads. Two-decoder models use two transformer decoders
to output distinctive targets. For instance, HOTR [23] and AS-NET [6] are
composed of an instance decoder that outputs object and an interaction decoder
that outputs interaction. In contrast to previous works that are trained with
a single inference path, our model learns with the augmented decoding paths.
Also, our framework can be applied to any transformer-based model. More
explanation of HOI transformers are in Section 3.1.
### 2.3 Consistency Learning in Vision
Consistency constraints applied to many computer vision topics have been
extensively studied. In semi-supervised learning, consistency regularization
is widely used to train the model to be invariant to input noise. Label
consistency methods [27, 53, 41, 47] augment or perturb an input image and
apply consistency loss between model predictions. CDS [21] explored object
detection in a semi-supervised setting with classification and localization
consistency regularization. Also, consistency regularization in cyclic form is
commonly used in generative models [61], image matching [60, 59], temporal
correspondence [10], and in many other domains.
#### Comparison with Consistency Learning
Our consistency training scheme is relevant to cross-task consistency learning
[55]. Cross-task consistency learning is based on inference-path invariance,
where the predictions should be consistent regardless of the inference paths.
As shown in Figure 1 (b), cross-task consistency learning uses an auxiliary
task $\mathcal{Y}\rightarrow\mathcal{Y}_{a}$ to train the main task function
$f_{\mathcal{X}\mathcal{Y}}$, _i.e_., given $x$ from the query domain, and $y$
from target domain $\mathcal{Y}$, predictions of
$f_{\mathcal{Y}\mathcal{Y}_{a}}\circ f_{\mathcal{X}\mathcal{Y}}(x)$ and
$f_{\mathcal{Y}\mathcal{Y}_{a}}(y)$ are expected to be consistent. Different
from cross-task consistency, our cross-path consistency learning (Figure 1
(c)) trains the main task function $f_{\mathcal{X}\mathcal{Y}}$ by enforcing
the prediction of target domain $\mathcal{Y}$ of $f_{\mathcal{X}\mathcal{Y}}$
and $f_{\mathcal{Y}_{s}\mathcal{Y}}\circ f_{\mathcal{X}\mathcal{Y}_{s}}$,
where auxiliary domain $\mathcal{Y}_{s}$ is decomposed from the target domain
$\mathcal{Y}$, to be consistent. Also, while cross-task consistency learning
requires the mapping function $f_{\mathcal{Y}\mathcal{Y}_{a}}$ to be
pretrained to avoid suboptimal training with the noisy estimator, cross-path
consistency learning does not demand any task-specific pre-trained networks
since the auxiliary domain $\mathcal{Y}_{s}$ is part of the target domain
$\mathcal{Y}$. Details for our framework is described in section 3.2.
## 3 Method
In this section, we present our novel end-to-end training strategy for
Transformers with cross-path consistency in Human-Object Interaction
Detection. The training strategy includes 1) augmenting the decoding path and
2) consistency regularization between predictions of multiple decoding paths.
Before discussing our training strategy, we briefly summarize transformers in
Human-Object Interaction detection.
### 3.1 Transformer in HOI detection
HOI transformers are commonly extended upon DETR [4], which is composed of a
CNN backbone followed by the encoder-decoder architecture of Transformer [1].
The CNN backbone first extracts a locally aggregated feature map
$f\in\mathbb{R}^{H^{\prime}\times W^{\prime}\times D}$ from input image
$x\in\mathbb{R}^{H\times W\times 3}$. Then, the feature map $f$ is passed into
the encoder to globally aggregate features via the self-attention mechanism,
resulting in the encoded feature map $X\in\mathbb{R}^{H^{\prime}\times
W^{\prime}\times D}.$ At a decoding stage, a decoder takes learnable query
embeddings ${\color[rgb]{0,0,0}q}\in\mathbb{R}^{N\times D}$ and outputs
$e\in\mathbb{R}^{N\times D}$ by interacting with encoded feature map $X$
through cross-attention. The outputs are converted to final HOI predictions
(_i.e_., human, object, interaction) by read-out functions, which are
generally feed-forward networks.
Training Transformers for detection entails matching between predictions and
ground truth labels since Transformers provide detections as set predictions.
To compute losses, the Hungarian algorithm [26] is used to associate
detections with ground truth labels. The predictions unmatched with ground
truth labels are considered as no object or no interactions. In general, HOI
transformers can be categorized into two groups based on human/object
localization schemes. [46, 63] directly predict the box coordinates of human
and object from an HOI prediction. But this causes problems that human or
object can be redundantly predicted by multiple query embeddings and the
localizations of the same object often differ across HOI triplet predictions.
To address these problems, [23, 6] propose parallel architectures to perform
interaction detection separately from object detection.
### 3.2 Decoding-Path Augmentation
(a)
(b)
Figure 2: Cross-path consistency for HOI detection. (a) Main task path
${\color[rgb]{1,0,0}\mathcal{P}_{1}}$ should be consistent with each augmented
path. _e.g_. path ${\color[rgb]{0,0,1}\mathcal{P}_{2}}$. (b) Augmented paths
should be consistent with one another. _e.g_. path
${\color[rgb]{0,0,1}\mathcal{P}_{2}}$ and
${\color[rgb]{1,0,0}\mathcal{P}_{3}}$. Figure 3: The overall process of
Cross-Path Consistency Learning. The encoded image features are passed into
the shared decoder with multiple inference paths
$\\{\mathcal{P}_{1},...,\mathcal{P}_{k-1},\mathcal{P}_{k}\\}$. Each path is
augmented based on the decoding-path augmentation to generate various
sequences of inference paths (see Section 3.2). To avoid clutter, we visualize
only the main path $\mathcal{P}_{1}$ and an augmented path $\mathcal{P}_{k}$.
The main path $\mathcal{P}_{1}$ consists of a single decoding stage, and the
augmented path $\mathcal{P}_{k}$ is a composition of decoding stages; all $f$
blocks share parameters. Given queries $q$ a learnable position embeddings,
each decoder extracts output embeddings denoted as $e_{1,1}$, $e_{k,1}$, and
$e_{k,2}$. Then, each of the output embeddings is fed into the readout
function FFN to predict each HOI element _i.e_. <human, object, interaction>.
With Cross-Path Consistency Learning (Section 3.3), all the outputs supervised
with the same ground truth label are trained to be consistent regardless of
their inference paths. Cross-Matching is used to match the queries that are
considered to be consistent by leveraging ground truth label. Along with the
supervision loss $\mathcal{L}_{\mathbf{sup}}^{k}$ for all paths
$\mathcal{P}_{k}$, cross-path consistency loss $\mathcal{L}_{\textbf{CPC}}$ is
added to our final loss.
We observe that HOI detection can be achieved by various sequences of
predictions. For instance, CNN-based HOI detection models [15, 5, 11, 17]
first detect instances (human and object) and then predict interactions
between the instances, _i.e_., $x\rightarrow\text{HO}\rightarrow\text{I}$,
where $x$ is an input image and H, O, I are predictions for human, object,
interaction, respectively. On the other hand, the HOI Transformers by [23, 46,
6, 63] directly predict HOI triplets, _i.e_., $x\rightarrow\text{HOI}$.
Inspired by Cross-Task Consistency [56] and this observation, we propose
decoding-path augmentation to generate various decoding paths (or prediction
paths) and impose consistency regularization. Decoding-path augmentation for
Transformers in HOI detection can be easily achieved by partially decoded HOI
predictions. Furthermore, sharing decoders across paths is beneficial in terms
of knowledge sharing.
In our experiments, we consider four decoding paths as follows:
$\begin{split}&\;\mathcal{P}_{1}=x\rightarrow\text{HOI}\\\
&\begin{rcases*}\mathcal{P}_{2}=x\rightarrow\text{HO}\rightarrow\text{I}\\\
\mathcal{P}_{3}=x\rightarrow\text{HI}\ \rightarrow\text{O}\\\
\mathcal{P}_{4}=x\rightarrow\text{OI}\
\rightarrow\text{H}\end{rcases*}Augmented.\end{split}$ (1)
Each decoding stage of path $\mathcal{P}_{k}$ can be written as:
$\begin{split}e_{{k,1}}&=f(e_{k,0}+q_{k,1},\ X),\\\
e_{{k,2}}&=f(e_{k,1}+q_{k,2},\ X),\end{split}$ (2)
where ${q}_{k,j},\ e_{k,j}$ denote learnable query and output embeddings on
$k^{\text{th}}$ path at $j^{\text{th}}$ decoding stage. The decoder $f$ is
shared across all paths and stages. The $e_{k,0}$ above is dummy output
embeddings set to zeros since there is no 0-th stage, see Figure 3. Each
decoding stage and path use a separate readout function FFN to translate the
output embeddings into HOI instance predictions. For example, on
$\mathcal{P}_{2}:x\rightarrow\text{HO}\rightarrow\text{I}$, at stage 1
$e_{2,1}$ is read out by $\text{FFN}^{\mathcal{P}_{2}}_{h}$ and
$\text{FFN}_{o}^{\mathcal{P}_{2}}$ to predict bounding boxes of human and
object respectively. Prediction for HOI element $m\in\\{h,o,act\\}$ in each
$k^{th}$ path at $j^{th}$ decoding stage can be written as
$\hat{y}_{k}^{m}=\text{FFN}^{\mathcal{P}_{k}}_{m}\left(e_{k,j}\right)$.
### 3.3 Cross-Path Consistency Learning
We now present our Cross-Path Consistency Learning framework (CPC) that
imposes consistency regularization between predictions from different decoding
paths as shown in Figure 2. Learning with CPC leads better generalization
without any additional data or labels.
#### Cross-Path Consistency.
We explain our consistency learning scheme with an exemplary case of main path
$\mathcal{P}_{1}$ and augmented path $\mathcal{P}_{2}$ given as
$\begin{split}&\mathcal{P}_{1}:x\rightarrow\text{HOI}\\\
&\mathcal{P}_{2}:x\rightarrow\text{HO}\rightarrow\text{I}.\end{split}$ (3)
Here, the main path $\mathcal{P}_{1}$ is the HOI transformers’ original
inference path. In path $\mathcal{P}_{2}$, human and object detection logits
$\hat{y}^{h}_{2}$ and $\hat{y}^{o}_{2}$ are obtained reading out $e_{2,1}$,
which is the output embeddings on path 2 at stage 1. Then, the interaction
logit $\hat{y}^{act}_{2}$ is obtained after another subsequent decoder pass
defined as $f_{2,2}$. The corresponding inference scheme of $\mathcal{P}_{2}$
can be written in more formal terms:
$\begin{split}\hat{y}_{2}^{h}&=\text{FFN}_{h}^{\mathcal{P}_{2}}(f_{2,1}(X))\\\
\hat{y}_{2}^{o}&=\text{FFN}_{o}^{\mathcal{P}_{2}}(f_{2,1}(X))\\\
\hat{y}_{2}^{act}&=\text{FFN}_{act}^{\mathcal{P}_{2}}(f_{2,2}\circ
f_{2,1}(X))\\\ \end{split}$ (4)
In (4), input arrays for $f$ other than feature map $X$ were omitted for
simplicity.
With the predictions, we impose regularization to make the outputs from path
$\mathcal{P}_{1}$ and path $\mathcal{P}_{2}$ consistent. Note that HOI
detections from $\mathcal{P}_{2}$ consist of both final and intermediate
decoder outputs. To this end, we define the loss function
$\mathcal{L}_{\mathcal{P}_{1}\mathcal{P}_{2}}$ by aggregating losses from
multiple augmented paths to enforce consistency. The loss function is given
as:
$\begin{split}\mathcal{L}_{\mathcal{P}_{1}\mathcal{P}_{2}}=\lambda_{h}\cdot\mathcal{L}_{h}\big{(}\hat{y}_{1}^{h},\hat{y}_{2}^{h}\big{)}+\lambda_{o}\cdot\mathcal{L}_{o}\big{(}\hat{y}_{1}^{o},\hat{y}_{2}^{o}\big{)}\\\
+\lambda_{act}\cdot\mathcal{L}_{act}\big{(}\hat{y}_{1}^{act},\hat{y}_{2}^{act}\big{)},\end{split}$
(5)
where $\hat{y}^{h}_{1}$, $\hat{y}^{o}_{1}$ and $\hat{y}^{act}_{1}$ are the
output from the main path $\mathcal{P}_{1}$ and $\lambda$ are the loss
weights. In our experiments, softmax-type outputs use Jensen-Shannon
divergence (JSD) for consistency loss to give loss to each path symmetrically,
while outputs followed by sigmoid, _e.g_., box regression, multi-label action
classes, take the Mean-Squared Error loss. More details on type-specific loss
functions are in the supplement.
In the case of other path pairs, loss is computed in the same manner. The
final loss should thus incorporate all possible pairs. Then, the cross-path
consistency (CPC) loss can be written as:
$\mathcal{L}_{\textbf{CPC}}=\frac{1}{S}\sum_{(k,k^{\prime})\in\mathcal{K}}\mathcal{L}_{\mathcal{P}_{k}\mathcal{P}_{k^{\prime}}}$
(6)
where $\mathcal{K}$ denotes the set of all possible path pairs, and $S$ refers
to the size of set $\mathcal{K}$, _i.e_. the number of path combinations.
#### Cross Matching.
Cross-path consistency learning encourages outputs from different paths to be
consistent. However, since the outputs from a path are given as a set, we
first need to resolve correspondence to specify the pairs of predictions to
enforce consistency. We present cross matching, a simple method that tags each
instance with its corresponding ground truth label. The instances tagged with
the same label are paired to compute consistency loss. On the other hand, if
an instance is not matched with any of the paths’ output, we simply exclude
the instance from consistency learning treating it as no object or no
interaction. Our cross-path consistency loss is introduced below.
Let $\sigma_{k}(i)$ denote the index of the ground truth label that matches
the $i^{th}$ query in the $k^{th}$ path. We define
$\sigma_{k}^{-1}\left(n\right)$ as the query index of path $\mathcal{P}_{k}$
which is matched with the ground truth index $n$. To avoid clutter, we use
$\tilde{\sigma}_{k,n}$ as a shorthand notation for
$\sigma_{k}^{-1}\left(n\right)$. The outputs from different paths with the
same ground-truth label should be consistent. For example,
$\hat{y}^{m}_{k,{\tilde{\sigma}_{k,n}}}$ and
$\hat{y}^{m}_{k,{\tilde{\sigma}_{k^{\prime},n}}}$ which are predictions for
$m$ from $\mathcal{P}_{k}$ and $\mathcal{P}_{k^{\prime}}$ with the same
ground-truth index $n$ should be consistent.
Cross-path consistency loss between output predictions from $\mathcal{P}_{k}$
and $\mathcal{P}_{k^{\prime}}$ with the same ground-truth with index $n$ is
defined as follows:
$\begin{split}\mathcal{L}_{\mathcal{P}_{k}\mathcal{P}_{k^{\prime}}}^{n}=&\hskip
5.69054pt\lambda_{h}\cdot\mathcal{L}_{h}\big{(}\hat{y}^{h}_{k,\tilde{\sigma}_{k,n}},\hat{y}^{h}_{k^{\prime},\tilde{\sigma}_{k^{\prime},n}}\big{)}\\\
&+\lambda_{o}\cdot\mathcal{L}_{o}\big{(}\hat{y}^{o}_{k,\tilde{\sigma}_{k,n}},\hat{y}^{o}_{k^{\prime},\tilde{\sigma}_{k^{\prime},n}}\big{)}\\\
&+\lambda_{act}\cdot\mathcal{L}_{act}\big{(}\hat{y}^{act}_{k,\tilde{\sigma}_{k,n}},\hat{y}^{act}_{k^{\prime},\tilde{\sigma}_{k^{\prime},n}}\big{)}\end{split}$
(7)
#### Final Loss
The final cross-path consistency loss for all $\mathcal{P}_{k}$ is derived as,
$\mathcal{L}_{\textbf{CPC}}=\frac{1}{{S}\cdot\mathcal{N}}\sum_{n=1}^{\mathcal{N}}\sum_{(k,k^{\prime})\in\mathcal{K}}\mathcal{L}_{\mathcal{P}_{k}\mathcal{P}_{k^{\prime}}}^{n}$
(8)
where $\mathcal{N}$ is the number of ground truth labels. Then, the final form
of our training loss $\mathcal{L}$ is defined by
$\mathcal{L}=\sum_{k}\mathcal{L}_{\mathbf{sup}}^{k}+w(t)\cdot\mathcal{L}_{\textbf{CPC}},$
(9)
where $\mathcal{L}_{\mathbf{sup}}^{k}$ is the supervision loss for each path
$\mathcal{P}_{k}$ and $w(t)$ is a ramp-up function [27, 2, 47] for stable
training. Our overall framework is illustrated in Figure 3.
## 4 Experiments
In this section, we empirically evaluate the effectiveness of our cross-path
consistency learning with HOI transformers. Our experiments are conducted on
public HOI detection benchmark datasets: V-COCO and HICO-DET. We first briefly
introduce the datasets and provide implementation details. Our extensive
experiments demonstrate that our training strategy renders significant
improvement on the baseline models without additional parameters or inference
time.
### 4.1 Dataset
#### V-COCO
[16] is a subset of the COCO dataset [35] which contains 5,400 trainval images
and 4,946 test images. V-COCO is annotated with 29 common action classes. For
evaluation of the V-COCO dataset, we report the mAP metric over 25
interactions for two scenarios, The first scenario includes a prediction of
occluded objects and is evaluated with respect to AP${}_{\text{role1}}$. On
the other hand, the second scenario does not contain such cases, and
performance is measured in AP${}_{\text{role2}}$.
#### HICO-DET
[5] is a subset of the HICO dataset which has more than 150K annotated
instances of human-object pairs in 47,051 images (37,536 for training and
9,515 for testing). It is annotated with 600 <interaction, object> instances.
There are 80 unique object types, identical to the COCO object categories, and
117 unique interaction verbs. For evaluation of the HICO-DET, we report the
mAP over three different set categories: (1) all 600 HOI categories in HICO
(Full), (2) 138 HOI categories with less than 10 training samples (Rare), and
(3) 462 HOI categories with more than 10 training samples (Non-Rare).
### 4.2 Implementation Details
#### Training
In our experiment, QPIC [46] and HOTR [23] were used as the baseline for the
HOI transformer respectively. During training, we initialize the network with
pretrained DETR [4] on MS-COCO with a Resnet-50 backbone. For all decoding
paths, parameters of the model are shared except for stage-wise queries and
feedforward networks.
All our experiments using consistency regularization are trained for 90 epochs
and the learning rate is decayed at the 60-th epoch by a factor of 0.1. As an
exception, HOTR is trained up to 50 epochs and the learning rate is decayed at
epoch 30 by a factor of 0.1 for HICO-DET. Following the original training
schemes in QPIC and HOTR, we freeze the encoder and backbone for HOTR, whereas
unfreeze those for QPIC. We use the AdamW [39] optimizer with a batch size of
16, and the initial learning rates for the transformer and backbone parameters
are set to $10^{-4}$ and $10^{-5}$ respectively, and weight decay is set to
$10^{-4}$. All experiments are trained on 8 V100 GPUs.
We re-implement the result of QPIC and HOTR on V-COCO [16] since our
reproduction results are quite different from the official ones in the paper.
For a fair comparison, all the loss coefficients overlapping between baselines
and our training strategy are identical to the ones reported in the paper [23,
46]. Details for hyperparameters relevant to our training strategy are
reported in the supplementary material.
#### Inference
We mainly use $\mathcal{P}_{1}$ ( $x\rightarrow\text{HOI}$ ) for inference to
compare with the baseline models without increasing the number of parameters.
Also, we report the results of other inference paths in our ablation studies.
### 4.3 Comparison with HOI transformer
We evaluate the effectiveness of our method compared to the existing HOI
transformers. All experiments are reported with the main path
$\mathcal{P}_{1}$ that infers HOI triplets by a single decoding stage
($x\rightarrow\text{HOI}$) which is identical to the original HOI transformer.
As shown in Table 1, our CPC training strategy significantly outperforms on
two baselines, HOTR [23] and QPIC [46]. In the V-COCO dataset, the experiment
shows improvement in performance by a considerable margin of 0.9 mAP for QPIC
in AP${}_{\text{role1}}$, and 1.8 mAP for HOTR. For AP${}_{\text{role2}}$,
QPIC and HOTR gain improvement by 0.9 mAP and 1.9 mAP respectively, similar to
that of AP${}_{\text{role1}}$.
In the HICO-DET dataset, our CPC learning with HOTR and QPIC outperforms all
the evaluation categories of HICO-DET, except negligible degradation in the
Non-Rare category on HOTR. Results on rare class on the HICO-DET are improved
by a significant margin of 1.29 mAP and 5.5 mAP for QPIC and HOTR
respectively. In both models, we observe a more prominent performance
improvement in the Rare category. This supports that our training strategy
performs well on rarely seen examples. Our strategy improves the conventional
HOI transformer models.
| V-COCO | HICO-DET
---|---|---
Method | AP${}_{\text{role1}}$ | AP${}_{\text{role2}}$ | Full | Rare | Non-Rare
QPIC | 62.2* | 64.5* | 29.07 | 21.85 | 31.23
QPIC + ours | 63.1 | 65.4 | 29.63 | 23.14 | 31.57
HOTR | 59.8* | 64.9* | 25.10 | 17.34 | 27.42
HOTR + ours | 61.6 | 66.8 | 26.16 | 22.84 | 27.15
Table 1: Comparison of our training strategy with vanilla HOI transformers on V-COCO and HICO-DET. * signifies our results reproduced with the official implementation codes of QPIC and HOTR. Method | Backbone | AP${}_{\text{role1}}$ | AP${}_{\text{role2}}$
---|---|---|---
CNN-based HOI Detection Model
InteractNet [15] | R50-FPN | 40.0 | 48.0
iCAN [12] | R50 | 45.3 | 52.4
TIN [32] | R50 | 47.8 | -
RPNN [58] | R50 | - | 47.5
Verb Embd. [54] | R50 | 45.9 | -
PMFNet [50] | R50-FPN | 52.0 | -
PastaNet [31] | R50-FPN | 51.0 | 57.5
VCL [20] | R50 L | 48.3 | -
UniDet [22] | R50-FPN | 47.5 | 56.2
DRG [11] | R50-FPN | 51.4 | -
FCMNet [36] | R50 | 53.1 | -
ConsNet [37] | R50-FPN | 53.2 | -
PDNet [57] | R50-FPN | 53.3 | -
IDN [30] | R50 | 53.3 | 60.3
GPNN [44] | R152 | 44.0 | -
IPNet [52] | H.G.104 | 51.0 | -
VSGNet [49] | R152 | 51.8 | 57.0
PDNet [57] | Res152 | 52.2 | -
ACP [24] | Res152 | 53.0 | -
Transformer-based HOI Detection Model
HoiT [63] | R101 | 52.9 | -
AS-Net [6] | R50 | 53.9 | -
HOTR [23] | R50 | 55.2 | 64.4
HOTR+ Ours | R50 | 61.6 | 66.8
QPIC [46] | R50 | 58.8 | 61.0
QPIC+ Ours | R50 | 63.1 | 65.4
Table 2: Comparison of performances on the V-COCO test set. AP${}_{\text{role1}}$ and AP${}_{\text{role2}}$ denotes performances under Scenario 1 and Scenario 2 in V-COCO respectively. | Default
---|---
Method | Detector | Backbone | Extra | Full | Rare | Non Rare
CNN-based HOI Detection Model | | | |
InteractNet [15] | COCO | R50-FPN | ✗ | 9.94 | 7.16 | 10.77
iCAN [12] | COCO | R50 | S | 14.84 | 10.45 | 16.15
TIN [32] | COCO | R50 | S+P | 17.03 | 13.42 | 18.11
RPNN [58] | COCO | R50 | P | 17.35 | 12.78 | 18.71
PMFNet [50] | COCO | R50-FPN | S+P | 17.46 | 15.65 | 18.00
No-Frills HOI [17] | COCO | R152 | S+P | 17.18 | 12.17 | 18.68
UnionDet [22] | COCO | R50-FPN | ✗ | 14.25 | 10.23 | 15.46
DRG [11] | COCO | R50-FPN | S+L | 19.26 | 17.74 | 19.71
VCL [20] | COCO | R50 | S | 19.43 | 16.55 | 20.29
FCMNet [36] | COCO | R50 | S+P | 20.41 | 17.34 | 21.56
ACP [24] | COCO | R152 | S+P | 20.59 | 15.92 | 21.98
DJ-RN [29] | COCO | R50 | S+V | 21.34 | 18.53 | 22.18
ConsNet [37] | COCO | R50-FPN | S+L | 22.15 | 17.12 | 23.65
PastaNet [31] | COCO | R50 | S+P+L | 22.65 | 21.17 | 23.09
IDN [30] | COCO | R50 | S | 23.36 | 22.47 | 23.63
GPNN [44] | COCO | R152 | ✗ | 13.11 | 9.41 | 14.23
IPNet [52] | COCO | HourGlass104 | ✗ | 19.56 | 12.79 | 21.58
VSGNet [49] | COCO | R152 | S | 19.80 | 16.05 | 20.91
PD-Net [57] | COCO | R152 | S+P+L | 20.81 | 15.90 | 22.28
Transformer-based HOI Detection Model | | | |
HoiT [63] | HICO-DET | R50 | ✗ | 23.46 | 16.91 | 25.41
AS-Net [6] | HICO-DET | R50 | ✗ | 28.87 | 24.25 | 30.25
HOTR [23] | HICO-DET | R50 | ✗ | 25.10 | 17.34 | 27.42
HOTR+ Ours | HICO-DET | R50 | ✗ | 26.16 | 22.84 | 27.15
QPIC [46] | HICO-DET | R50 | ✗ | 29.07 | 21.85 | 31.23
QPIC+ Ours | HICO-DET | R50 | ✗ | 29.63 | 23.14 | 31.57
Table 3: Performance comparison in HICO-DET. For the Detector, COCO means that the detector is trained on COCO, while HICO-DET means that the detector is first trained on COCO and then fine-tuned on HICO-DET. The each letter in Extra column stands for S: Interaction Patterns (Spatial Correlations), P: Pose, L: Linguistic Priors, V: Volume. Method | Share Dec. | CPC | $\mathcal{P}_{1}$ | $\mathcal{P}_{2}$ | $\mathcal{P}_{3}$ | $\mathcal{P}_{4}$ | Average
---|---|---|---|---|---|---|---
QPIC | ✓ | ✓ | 63.1 | 63.3 | 63.1 | 63.0 | 63.13 $\pm$ 0.05†
| ✓ | 62.4 | 62.9 | 60.8 | 59.4 | 61.38 $\pm$ 1.38
✓ | | 60.7 | 60.7 | 59.9 | 58.1 | 59.85 $\pm$ 1.06
HOTR | ✓ | ✓ | 61.6 | 61.5 | 61.6 | 61.6 | 61.58 $\pm$ 0.02†
| ✓ | 61.2 | 61.6 | 61.1 | 60.6 | 61.13 $\pm$ 0.36
✓ | | 60.6 | 60.6 | 61.2 | 60.6 | 60.75 $\pm$ 0.13
Table 4: Ablation Study on our learning strategies. Ablation results on
shared decoder (Share Dec.), and Cross-Path Consistency (CPC) are
demonstrated. For main path $\mathcal{P}_{1}$, and each augmented path
$\mathcal{P}_{2}$, $\mathcal{P}_{3}$, $\mathcal{P}_{4}$, their performances
are reported measured in mAP. They are evaluated on the V-COCO test set with
respect to Scenario 1. The best performances for each path are highlighted in
bold, and $\dagger$ refers to the case where the least standard deviation is
observed.
### 4.4 Comparison with State-of-the-Art Methods
In Table 2 and Table 3, we compare previous HOI detection methods with ours.
As demonstrated in the tables, our training strategy achieves the best
performance among its peers. Table 2 shows the result on V-COCO dataset in
both AP${}_{\text{role1}}$ and AP${}_{\text{role2}}$. In the V-COCO dataset,
our method achieves outstanding performance of 63.1 mAP in
AP${}_{\text{role1}}$ and 66.8 mAP in AP${}_{\text{role2}}$. Also, the results
on the HICO-DET dataset in Table 3 show that our CPC further improves the
state-of-the-art models (_e.g_., HOTR, and QPIC) in the default setting
achieving 26.16 mAP and 29.63 mAP, respectively.
### 4.5 Ablation Study
We further discuss the effectiveness of our framework through a series of
ablation studies. We first provide a path-wise analysis for our cross-path
consistency learning method. The effect of our training technique components
was tested on each path to validate our method. Subsequently, we analyze the
impact of the number of augmented paths on the main task performance. We
experimentally prove the validity of our method by demonstrating the
correlation between the number of paths and performance.
#### Efficiency of CPC.
Table 4 presents ablation experiment results for all inference paths,
$\mathcal{P}_{1}$, $\mathcal{P}_{2}$, $\mathcal{P}_{3}$, and
$\mathcal{P}_{4}$. Path $\mathcal{P}_{1}$ is the main path, which we aim to
boost performance with the rest of the augmented paths. We try ablating
decoder sharing or cross-path consistency regularization one at a time to
confirm each component’s contribution to our training strategy. Note that all
of our experiments are conducted with the encoder block shared across paths.
When our CPC training strategies are applied, QPIC and HOTR achieve an mAP of
63.1, and 61.6 on main path $\mathcal{P}_{1}$. When the decoder parameters are
not shared, performance degradation in path $\mathcal{P}_{1}$ was observed for
both baselines; a 0.7 mAP drop for QPIC, and 0.4 mAP drop for HOTR. On the
other hand, when CPC regularization is left out while decoder parameters are
shared, performance of QPIC and HOTR decreased by a large margin of 2.4 mAP
and 1.0 mAP each. In terms of overall performance across all paths, the
average mAP showed a similar trend for each experiment condition. The overall
results support that our learning strategy improves generalization of base
architectures, and boosts performance by sharing knowledge throughout paths
and stages.
Interestingly, the standard deviation of all performances dramatically
increases without both components. With unshared decoders, deviation increases
by 1.33 for QPIC and 0.35 for HOTR. Also, when CPC regularization is removed,
deviation increases by 1.01 for QPIC and 0.11 for HOTR. This implies that our
training strategy with shared decoder and CPC leads to more stable training as
well as consistent representations.
#### Impact of Augmented Paths.
We explore how the number of augmented paths affects the performance of the
main path $\mathcal{P}_{1}$ in V-COCO benchmark. Starting from
$\mathcal{P}_{1}$, the augmented paths are gradually added with respect to mAP
of Scenario 1 from Table 5, where each path is independently trained with
default settings with no training techniques applied. We leverage the
augmented path with better performance first, as performance of each model
will serve as a lower bound for the ensemble of paths. Specifically, as shown
in Table 5, both HOTR and QPIC showed better performance in the order of
$\mathcal{P}_{1}$, $\mathcal{P}_{2}$, $\mathcal{P}_{3}$, and
$\mathcal{P}_{4}$, when trained independently.
We compare the four cases where the augmented paths are gradually added in the
corresponding order; _i.e_., $\mathcal{P}_{1}$,
$\mathcal{P}_{1}+\mathcal{P}_{2}$,
$\mathcal{P}_{1}+\mathcal{P}_{2}+\mathcal{P}_{3}$, and
$\mathcal{P}_{1}+\mathcal{P}_{2}+\mathcal{P}_{3}+\mathcal{P}_{4}$. As shown in
Figure 4, performance is gradually improved as augmented paths are added. The
ablation study evidences that regardless of each path performance, taking
advantage of more paths bolsters the learning capability of our main task, and
its performance builds up as the number of augmented paths increases.
Method | $\mathcal{P}_{1}$ | $\mathcal{P}_{2}$ | $\mathcal{P}_{3}$ | $\mathcal{P}_{4}$ | Average
---|---|---|---|---|---
QPIC | 62.2 | 61.9 | 61.7 | 60.4 | 61.55 $\pm$ 0.69
HOTR | 59.8 | 59.5 | 59.0 | 58.9 | 59.3 $\pm$ 0.37
Table 5: Path-wise results on V-COCO.
(a) QPIC
(b) HOTR
Figure 4: Ablation on the number of augmented paths. As the number of
augmented paths increases, main task performance increases accordingly.
## 5 Conclusion
We propose end-to-end Cross-Path Consistency learning for Human-Object
Interaction detection. Through decoding-path augmentation, various decoder
paths are generated which predict HOI triplets in permuted sequences. Then,
consistency regularization is applied across paths to enforce the predictions
to be consistent. Parameter sharing and cross-matching were introduced as well
to enhance learning.
Our method is conceptually simple, and can be applied to a wide range of
transformer architectures. Also, it does not require additional model capacity
nor inference time. The substantial improvements on V-COCO and HICO-DET
support our method’s efficacy in various HOI detection tasks. Through further
empirical studies, its capabilities to improve generalization and to encourage
consistent representations are approved.
#### Acknowledgements
This work was partly supported by Institute of Information & communications
Technology Planning & Evaluation (IITP) grant funded by the Korea government
(MSIT) (No.2021-0-02312, Efficient Meta-learning Based Training Method and
Multipurpose Multi-modal Artificial Neural Networks for Drone AI),
(IITP-2022-2020-0-01819, the ICT Creative Consilience program); ETRI grant
(22ZS1200, Fundamental Technology Research for Human-Centric Autonomous
Intelligent System); and KakaoBrain corporation.
## References
* [1] Vaswani A., Shazeer N., Parmar N., Uszkoreit J., Jones L., Gomez A.N., Kaiser L., and Polosukhin I. Attention is all you need. In NeurIPS, 2017.
* [2] David Berthelot, Nicholas Carlini, Ian Goodfellow, Nicolas Papernot, Avital Oliver, and Colin Raffel. Mixmatch: A holistic approach to semi-supervised learning. In NeurIPS, 2019.
* [3] Zhe Cao, Tomas Simon, Shih-En Wei, and Yaser Sheikh. Realtime multi-person 2d pose estimation using part affinity fields. In CVPR, 2017.
* [4] Nicolas Carion, Francisco Massa, Gabriel Synnaeve, Nicolas Usunier, Alexander Kirillov, and Sergey Zagoruyko. End-to-end object detection with transformers. In ECCV, 2020.
* [5] Yu-Wei Chao, Yunfan Liu, Xieyang Liu, Huayi Zeng, and Jia Deng. Learning to detect human-object interactions. In WACV, 2018.
* [6] Mingfei Chen, Yue Liao, Si Liu, Zhiyuan Chen, Fei Wang, and Chen Qian. Reformulating hoi detection as adaptive set prediction. In CVPR, 2021.
* [7] Yilun Chen, Zhicheng Wang, Yuxiang Peng, Zhiqiang Zhang, Gang Yu, and Jian Sun. Cascaded pyramid network for multi-person pose estimation. In CVPR, 2018.
* [8] Qi Dong, Zhuowen Tu, Haofu Liao, Yuting Zhang, Vijay Mahadevan, and Stefano Soatto. Visual relationship detection using part-and-sum transformers with composite queries. In ICCV, 2021.
* [9] Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, et al. An image is worth 16x16 words: Transformers for image recognition at scale. In ICLR, 2020.
* [10] Debidatta Dwibedi, Yusuf Aytar, Jonathan Tompson, Pierre Sermanet, and Andrew Zisserman. Temporal cycle-consistency learning. In CVPR, 2019.
* [11] Chen Gao, Jiarui Xu, Yuliang Zou, and Jia-Bin Huang. DRG: dual relation graph for human-object interaction detection. In ECCV, 2020.
* [12] Chen Gao, Yuliang Zou, and Jia-Bin Huang. ican: Instance-centric attention network for human-object interaction detection. In WACV, 2018.
* [13] Peng Gao, Minghang Zheng, Xiaogang Wang, Jifeng Dai, and Hongsheng Li. Fast convergence of detr with spatially modulated co-attention. In ICCV, 2021.
* [14] Nikolaos Gkanatsios, Vassilis Pitsikalis, Petros Koutras, Athanasia Zlatintsi, and Petros Maragos. Deeply supervised multimodal attentional translation embeddings for visual relationship detection. In ICCV, 2019.
* [15] Georgia Gkioxari, Ross Girshick, Piotr Dollár, and Kaiming He. Detecting and recognizing human-object interactions. In CVPR, 2018.
* [16] Jitendra Gupta, Saurabh Malik. Visual semantic role labeling. In CVPR, 2015.
* [17] Tanmay Gupta, Alexander Schwing, and Derek Hoiem. No-frills human-object interaction detection: Factorization, layout encodings, and training techniques. In ICCV, 2019.
* [18] Tanmay Gupta, Kevin Shih, Saurabh Singh, and Derek Hoiem. Aligned image-word representations improve inductive transfer across vision-language tasks. In ICCV, 2017.
* [19] Kaiming He, Georgia Gkioxari, Piotr Dollár, and Ross Girshick. Mask r-cnn. In ICCV, 2017.
* [20] Zhi Hou, Xiaojiang Peng, Yu Qiao, and Dacheng Tao. Visual compositional learning for human-object interaction detection. In ECCV, 2020.
* [21] Jisoo Jeong, Seungeui Lee, Jeesoo Kim, and Nojun Kwak. Consistency-based semi-supervised learning for object detection. In NeurIPS, 2019.
* [22] Bumsoo Kim, Taeho Choi, Jaewoo Kang, and Hyunwoo J. Kim. Uniondet: Union-level detector towards real-time human-object interaction detection. In ECCV, 2020.
* [23] Bumsoo Kim et al. Hotr: End-to-end human-object interaction detection with transformers. In CVPR, 2021.
* [24] Dong-Jin Kim, Xiao Sun, Jinsoo Choi, Stephen Lin, and In So Kweon. Detecting human-object interactions with action co-occurrence priors. In ECCV, 2020.
* [25] Alexander Kolesnikov, Alina Kuznetsova, Christoph H Lampert, and Vittorio Ferrari. Detecting visual relationships using box attention. In ICCV, 2019.
* [26] Harold W Kuhn. The hungarian method for the assignment problem. NRL, 1955.
* [27] Samuli Laine and Timo Aila. Temporal ensembling for semi-supervised learning. In ICLR, 2017.
* [28] Jiefeng Li, Can Wang, Hao Zhu, Yihuan Mao, Hao-Shu Fang, and Cewu Lu. Crowdpose: Efficient crowded scenes pose estimation and a new benchmark. In CVPR, 2019.
* [29] Yong-Lu Li, Xinpeng Liu, Han Lu, Shiyi Wang, Junqi Liu, Jiefeng Li, and Cewu Lu. Detailed 2d-3d joint representation for human-object interaction. In CVPR, 2020.
* [30] Yong-Lu Li, Xinpeng Liu, Xiaoqian Wu, Yizhuo Li, and Cewu Lu†. Hoi analysis: Integrating and decomposing human-object interaction. In NeurIPS, 2020.
* [31] Yong-Lu Li, Liang Xu, Xinpeng Liu, Xijie Huang, Yue Xu, Shiyi Wang, Hao-Shu Fang, Ze Ma, Mingyang Chen, and Cewu Lu. Pastanet: Toward human activity knowledge engine. In CVPR, 2020.
* [32] Yong-Lu Li, Siyuan Zhou, Xijie Huang, Liang Xu, Ze Ma, Hao-Shu Fang, Yanfeng Wang, and Cewu Lu. Transferable interactiveness knowledge for human-object interaction detection. In CVPR, 2019.
* [33] Yue Liao, Si Liu, Fei Wang, Yanjie Chen, Chen Qian, and Jiashi Feng. Ppdm: Parallel point detection and matching for real-time human-object interaction detection. In CVPR, 2020.
* [34] Tsung-Yi Lin, Piotr Dollár, Ross Girshick, Kaiming He, Bharath Hariharan, and Serge Belongie. Feature pyramid networks for object detection. In CVPR, 2017.
* [35] Tsung-Yi Lin, Michael Maire, Serge Belongie, James Hays, Pietro Perona, Deva Ramanan, Piotr Dollár, and C Lawrence Zitnick. Microsoft coco: Common objects in context. In ECCV, 2014.
* [36] Yang Liu, Qingchao Chen, and Andrew Zisserman. Amplifying key cues for human-object-interaction detection. In ECCV, 2020.
* [37] Ye Liu, Junsong Yuan, and Chang Wen Chen. Consnet: Learning consistency graph for zero-shot human-object interaction detection. In ACMMM, 2020.
* [38] Ze Liu, Yutong Lin, Yue Cao, Han Hu, Yixuan Wei, Zheng Zhang, Stephen Lin, and Baining Guo. Swin transformer: Hierarchical vision transformer using shifted windows. In ICCV, 2021.
* [39] Ilya Loshchilov and Frank Hutter. Decoupled weight decay regularization. In ICLR, 2019.
* [40] Depu Meng, Xiaokang Chen, Zejia Fan, Gang Zeng, Houqiang Li, Yuhui Yuan, Lei Sun, and Jingdong Wang. Conditional detr for fast training convergence. In ICCV, 2021.
* [41] Takeru Miyato, Shin-ichi Maeda, Masanori Koyama, and Shin Ishii. Virtual adversarial training: a regularization method for supervised and semi-supervised learning. TPAMI, 2018.
* [42] Alejandro Newell, Kaiyu Yang, and Jia Deng. Stacked hourglass networks for human pose estimation. In ECCV, 2016.
* [43] Julia Peyre, Ivan Laptev, Cordelia Schmid, and Josef Sivic. Detecting unseen visual relations using analogies. In ICCV, 2019.
* [44] Siyuan Qi, Wenguan Wang, Baoxiong Jia, Jianbing Shen, and Song-Chun Zhu. Learning human-object interactions by graph parsing neural networks. In ECCV, 2018.
* [45] Shaoqing Ren, Kaiming He, Ross Girshick, and Jian Sun. Faster r-cnn: towards real-time object detection with region proposal networks. TPAMI, 2016.
* [46] Masato Tamura et al. Qpic: Query-based pairwise human-object interaction detection with image-wide contextual information. In CVPR, 2021.
* [47] Antti Tarvainen and Harri Valpola. Mean teachers are better role models: Weight-averaged consistency targets improve semi-supervised deep learning results. In NeurIPS, 2017.
* [48] Hugo Touvron, Matthieu Cord, Matthijs Douze, Francisco Massa, Alexandre Sablayrolles, and Hervé Jégou. Training data-efficient image transformers & distillation through attention. In ICML, 2021.
* [49] Oytun Ulutan, ASM Iftekhar, and BS Manjunath. Vsgnet: Spatial attention network for detecting human object interactions using graph convolutions. In CVPR, 2020.
* [50] Bo Wan, Desen Zhou, Yongfei Liu, Rongjie Li, and Xuming He. Pose-aware multi-level feature network for human object interaction detection. In ICCV, 2019.
* [51] Hai Wang, Wei shi Zheng, and Ling Yingbiao. Contextual heterogeneous graph network for human-object interaction detection. In ECCV, 2020.
* [52] Tiancai Wang, Tong Yang, Martin Danelljan, Fahad Shahbaz Khan, Xiangyu Zhang, and Jian Sun. Learning human-object interaction detection using interaction points. In CVPR, 2020.
* [53] Qizhe Xie, Zihang Dai, Eduard Hovy, Thang Luong, and Quoc Le. Unsupervised data augmentation for consistency training. In NeurIPS, 2020.
* [54] Bingjie Xu, Yongkang Wong, Junnan Li, Qi Zhao, and Mohan S Kankanhalli. Learning to detect human-object interactions with knowledge. In CVPR, 2019.
* [55] Amir R Zamir et al. Robust learning through cross-task consistency. In CVPR, 2020.
* [56] Ying Zhang, Tao Xiang, Timothy M Hospedales, and Huchuan Lu. Deep mutual learning. In CVPR, 2018.
* [57] Xubin Zhong, Changxing Ding, Xian Qu, and Dacheng Tao. Polysemy deciphering network for robust human-object interaction detection. In ECCV, 2020.
* [58] Penghao Zhou and Mingmin Chi. Relation parsing neural network for human-object interaction detection. In ICCV, 2019.
* [59] Tinghui Zhou, Philipp Krahenbuhl, Mathieu Aubry, Qixing Huang, and Alexei A Efros. Learning dense correspondence via 3d-guided cycle consistency. In CVPR, 2016.
* [60] Tinghui Zhou, Yong Jae Lee, Stella X. Yu, and Alexei A. Efros. Flowweb: Joint image set alignment by weaving consistent, pixel-wise correspondences. In CVPR, 2015.
* [61] Jun-Yan Zhu, Taesung Park, Phillip Isola, and Alexei A Efros. Unpaired image-to-image translation using cycle-consistent adversarial networks. In ICCV, 2017.
* [62] Xizhou Zhu, Weijie Su, Lewei Lu, Bin Li, Xiaogang Wang, and Jifeng Dai. Deformable detr: Deformable transformers for end-to-end object detection. In ICLR, 2020.
* [63] Cheng Zou, Bohan Wang, Yue Hu, Junqi Liu, Qian Wu, Yu Zhao, Boxun Li, Chenguang Zhang, Chi Zhang, Yichen Wei, et al. End-to-end human object interaction detection with hoi transformer. In CVPR, 2021.
|
\geq \mu_o \left( B \cap \mathcal O^+_o(g_no,r) \right)
\geq \mu_o(B)
\end{equation*}
By <ref>, $\mathcal O^+_o(g_no,r) \cap \Lambda$ is contained in $\mathcal O_o(g_no,r+20\alpha)$.
Since $\nu_o$ gives full measure to $\Lambda$, we get
\begin{equation*}
< \mu_o(B)
\leq \nu_o \left(\mathcal O^+_o(g_no,r) \right)
\leq \nu_o \left(\mathcal O^+_o(g_no,r) \cap \Lambda\right)
\leq \nu_o \left(\mathcal O_o(g_no,r+20\alpha)\right).
\end{equation*}
Recall that $\dist o{g_no}$ diverges to infinity.
Hence the above inequality contradicts the Shadow Lemma (<ref>).
Let us prove <ref>.
It follows from the construction that $\mu$ is $G$-invariant, $\norm{\mu_o} = 1$, and $\mu_x \ll \mu_y$, for every $x,y \in X$.
Hence we are only left to prove that $\mu$ is quasi-conformal.
Let $x,y \in X$.
We define two auxiliary maps as follows.
\begin{equation*}
\begin{array}{ccccccc}
\bar X & \to & \R & \quad \text{and} \quad\ & \bar X & \to & \R \\
c & \mapsto &\displaystyle \inf_{c' \sim c} c'(x,y) && c & \mapsto &\displaystyle \sup_{c' \sim c} c'(x,y)
\end{array}
\end{equation*}
We denote them by $c \mapsto \beta^-_c(x,y)$ and $c \mapsto \beta^+_c(x,y)$ respectively.
As $\bar X$ is separable, one checks that these maps are $\mathfrak R$-measurable.
Let $B$ be a saturated Borel subset.
Using the conformality of $\nu$ we have
\begin{equation*}
\nu_x(B)
\leq \int \mathbb 1_B(c) e^{-\omega_Gc(x,y)} d\nu_y(c)
\leq \int \mathbb 1_B(c) e^{-\omega_G\beta^-_c(x,y)} d\nu_y(c).
\end{equation*}
Since $B$ is saturated and $c \mapsto \beta^-_c(x,y)$ is $\mathfrak R$-measurable, we get
\begin{equation*}
\mu_x(B)
\leq \int \mathbb 1_B(c) e^{-\omega_G\beta^-_c(x,y)} d\mu_y(c).
\end{equation*}
This inequality holds for every $B \in \mathfrak R$.
\begin{equation*}
\frac {d\mu_x}{d\mu_y} (c) \leq e^{-\omega_G \beta^-_c(x,y)}, \quad \mu\text{-a.e.}
\end{equation*}
In the same way, we obtain a lower bound for the Radon-Nikodym derivative with $\beta^+_c(x,y)$ in place of $\beta^-_c(x,y)$.
By <ref>, for $\mu$-almost every $c \in \bar X$, we have
\begin{equation*}
c(x,y) -20\alpha \leq \beta^-_c(x,y) \leq \beta^+_c(x,y) \leq c(x,y) + 20\alpha.
\end{equation*}
Hence $\mu$ is quasi-conformal.
Point <ref> now follows from <ref> and the quasi-conformality.
§.§ More applications
Assume that the action of $G$ on $X$ is divergent.
For every infinite normal subgroup of $G$ we have
\begin{equation*}
\omega_N > \frac 12 \omega_G.
\end{equation*}
Let $Q = G/N$ and $\omega_Q$ be the growth rate of $Q$ on $X/N$.
According to <ref> we have
\begin{equation*}
\omega_N + \frac 12 \omega_Q \geq \omega_G.
\end{equation*}
Since the map $X \to X/N$ is $1$-Lipschitz, $\omega_Q \leq \omega_G$.
\begin{equation*}
\omega_N \geq \frac 12 \omega_G.
\end{equation*}
Suppose now that, contrary to our claim, $\omega_G = 2 \omega_N$.
We choose
* a $G$-invariant, $\omega_G$-conformal density $\nu = (\nu_x)$ and
* an $N$-invariant, $\omega_N$-conformal density $\nu' = (\nu'_x)$ such that the action of $N$ on $(\bar X, \mathfrak B, \nu'_o)$ is ergodic.
We write $\mu$ and $\mu'$ for their respective restrictions to the reduced horocompactification $(\bar X, \mathfrak R)$.
In particular, the action of $N$ on $(\bar X, \mathfrak R, \mu'_o)$ is ergodic.
We claim that $\mu_0$ is absolutely continuous with respect to $\mu'_0$.
According to <ref>, $(G, \nu')$ satisfies the Shadow Lemma for some parameters $(\epsilon, r_0) \in \R_+^* \times \R_+$.
By <ref>, there exists $\alpha,r \in \R^*_+$ such that $\nu_o$ gives full measure to $\Lambda = \Lambda_{\rm ctg} (G, o, \alpha,r)$.
Without loss of generality, we can assume that $r \geq r_0$.
For simplicity we set $r' = r + 16\alpha$ and write $\Lambda'$ for $\Lambda_{\rm ctg} (G, o, \alpha,r)$.
Let $B$ be a saturated subset contained in $\Lambda_{\rm ctg} (G, o,\alpha,r)$.
Let $V$ be an open set containing $B$.
Observe that $(B \cap \Lambda)^+$ is contained in $\Lambda'$ (<ref>)
Fix $L > r' +16\alpha$.
Using <ref> with $(B \cap \Lambda)^+$, we build a subset $S \subset \mathcal T(\alpha, L)$ such that
\begin{equation*}
B \cap \Lambda \subset (B \cap \Lambda)^+ \subset \bigcup_{g \in S}\mathcal O_o(go,r') \subset V.
\end{equation*}
According to <ref>, there is a subset $S^*$ of $S$ such that
* the collection $\left( \mathcal O_o(go,r')\right)_{g \in S^*}$ is pairwise disjoint, and
* $B\cap \Lambda$ is covered by $\left( \mathcal O_o(go,r'+42\alpha)\right)_{g \in S^*}$
Since $\nu$ gives full measure to $\Lambda$, we have $\nu_o(B) = \nu_o(B \cap \Lambda)$.
Using <ref> with the density $\nu$ we get
\begin{equation*}
\nu_o(B)
\leq \sum_{g \in S^*} \nu_o(\mathcal O_o(go,r'+42\alpha))
\leq e^{2\omega_G(r'+42\alpha)} \sum_{g \in S^*} e^{-\omega_G \dist o{go}}.
\end{equation*}
Recall that for every $g \in G$, we have $\norm{\nu'_{go}} \geq e^{-\omega_N \dist o{go}}$.
Since $\omega_G = 2 \omega_N$, we obtain
\begin{equation*}
\nu_o(B)
\leq e^{2\omega_G(r'+42\alpha)} \sum_{g \in S^*} \norm{\nu'_{go}} e^{-\omega_N \dist o{go}}.
\end{equation*}
Using now the Shadow Principle with the density $\nu'$ we obtain
\begin{equation*}
\nu_o(B)
\leq C \sum_{g \in S^*} \nu'_o(\mathcal O_o(go,r'))
\leq C \nu'_o(V),
\quad \text{where} \quad
C = \frac 1\epsilon e^{2\omega_G(r'+42\alpha)}
\end{equation*}
does not depend on $B$.
This inequality holds for every open subset $V$ containing $B$, hence $\nu_o(B) \leq C\nu'_o(B)$, i.e. $\mu_o(B) \leq C \mu'_o(B)$.
This completes the proof of our claim.
Denote by $f$ the Radon-Nikodym derivative $f = d\mu_o / d \mu'_o$.
Both $\mu$ and $\mu'$ are $N$-invariant.
Hence the set $A = \set{c \in \bar X}{f(c) > 0}$ is $N$-invariant.
Note that $\mu'_o(A) > 0$.
Indeed otherwise $\mu_o$ would be the zero measure.
Since the action of $N$ on $(\bar X, \mathfrak R, \mu'_o)$ is ergodic, we get $\mu'_o(A) = 1$.
Hence $\mu_o$ and $\mu'_o$ are in the same class of measures.
Since $\mu_o$ is $G$-invariant, $\mu'_o$ is $G$-quasi-invariant.
We assumed that $\mu'_o$ is ergodic for the action of $N$.
It follows from <ref> that $\mu'$ is almost fixed by $G$.
Thus $\omega_N \geq \omega_G$ by <ref>.
This contradicts our assumption and completes the proof.
Let $H \subset G$ be a subgroup which is not virtually cyclic and contains a contracting element.
Let $\nu = (\nu_x)$ be an $H$-invariant, $\omega_H$-conformal density and $\mu = (\mu_x)$ its restriction to the reduced horocompactification $(\bar X, \mathfrak R)$.
Assume that the action of $H$ is divergent.
If $\mu$ is almost fixed by $G$, then $(G, \nu)$ satisfies the Shadow Principle.
According to <ref>, there are $\alpha, r_0 \in \R^*_+$ such that $\nu$ gives full measure to $\Lambda_{\rm ctg}(H,o,\alpha,r_0)$.
Proceeding as in the proof of <ref>, we show that for every $g \in G$ and $r \in \R_+$,
\begin{equation}
\label{eqn: shadow lemma almost-fixed reduced}
\nu_o\left(\mathcal O_o(go,r)\right)
\geq \norm{\nu_{go}} e^{-\omega_H \dist o{go}}\nu^g_o\left(\mathcal O_{g^{-1}o}(o,r)\right).
\end{equation}
Choose now $r \geq r_0$ and $g \in G$.
We denote by $\mathcal O^+_{g^{-1}o}(o,r)$ the saturation of the shadow $\mathcal O_{g^{-1}o}(o,r)$, which is mesurable by <ref>.
According to <ref>,
\begin{equation*}
\mathcal O^+_{g^{-1}o}(o,r) \cap \Lambda_{\rm ctg}(H,o,\alpha, r) \subset \mathcal O_{g^{-1}o}(o,r+20\alpha).
\end{equation*}
Recall that $\nu$ gives full measure to $\Lambda_{\rm ctg}(H,o,\alpha, r_0)$, thus to $\Lambda_{\rm ctg}(H,o,\alpha, r)$ as well.
Since $\mu$ is almost fixed by $G$ we have
\begin{align*}
\nu^g_o\left(\mathcal O_{g^{-1}o}(o,r+20\alpha)\right)
\geq \mu^g_o\left(\mathcal O^+_{g^{-1}o}(o,r) \right)
& \geq \epsilon \mu_o\left(\mathcal O^+_{g^{-1}o}(o,r) \right) \\
& \geq \epsilon \nu_o\left(\mathcal O_{g^{-1}o}(o,r) \right),
\end{align*}
where $\epsilon \in \R_+^*$ does not depend on $g$ and $r$.
Combined with (<ref>) it shows that for every $r \geq r_0$, for every $g \in G$, we have
\begin{equation*}
\nu_o\left(\mathcal O_o(go,r +20\alpha)\right)
\geq \epsilon \norm{\nu_{go}} e^{-\omega_H \dist o{go}}\nu_o\left(\mathcal O_{g^{-1}o}(o,r)\right).
\end{equation*}
According to our assumption, $H$ is not virtually cyclic and contains a contracting element.
The conclusion now follows from <ref> applied with the group $H$ and the set $\mathcal D_0 = \{ \nu\}$..
Let $H$ be a commensurated subgroup of $G$.
If the action of $H$ on $X$ is divergent, then the following holds.
Any $H$-invariant, $\omega_H$-conformal density is $G$-almost invariant when restricted to the reduced horocompactification $(\bar X, \mathfrak R)$.
$\omega_H = \omega_G$.
The action of $G$ on $X$ is divergent.
Let $\nu = (\nu_x)$ be an $H$-invariant, $\omega_H$-conformal density.
We denote by $\mu = (\mu_x)$ its restriction to the reduced horocompactification $(\bar X, \mathfrak R)$.
Let $g \in G$.
By definition of commensurability the intersection $H_0 = H^g \cap H$ has finite index in $H$.
In particular, $H_0$ is divergent and $\omega_{H_0} = \omega_H$.
Recall that $\nu^g$ is the image of $\nu$ under the right action of $g \in G$.
It is an $H^g$-invariant, $\omega_H$-conformal density, thus an $H_0$-invariant, $\omega_{H_0}$-conformal density.
Similarly $\nu$ is an $H_0$-invariant, $\omega_{H_0}$-conformal density.
Since $H_0$ is divergent, <ref> tells us that
* there is $C \in \R_+$ such that $\mu^g \leq C \mu$,
* the action of $H^g \cap H$ on $(\bar X, \mathfrak R, \mu_o)$ is ergodic.
Note that $C$ depends a priori on $H_o$ and thus on $g$.
Nevertheless, it still proves that $\mu$ is $G$-quasi-invariant.
Consequently $\mu$ is $C_0$-almost fixed by $G$, for some $C_0 \in \R_+^*$ (<ref>).
We deduce from <ref> that $(G, \nu)$ satisfies the Shadow Principle.
Point <ref> now follows from <ref>.
Recall that $\mathcal P_H(s) \leq \mathcal P_G(s)$, for every $s \in \R_+$.
Since the action of $H$ on $X$ is divergent, $\mathcal P_G(s)$ diverges at $s = \omega_H = \omega_G$.
Hence the action of $G$ on $X$ is divergent as well, which proves <ref>.
We already know that $\mu$ is almost-fixed by $G$, so that the map $\chi \colon G \to \R$ sending $g$ to $\ln \norm{\mu_{go}}$ is a quasi-morphism (<ref>).
We are left to prove that $\mu$ is actually $G$-almost invariant, i.e. $\chi$ is bounded.
Recall that $(G, \nu)$ satisfies the Shadow Principle.
It follows from <ref> that the critical exponent of the series
\begin{equation*}
\sum_{g \in G} e^{\chi(g)}e^{-s\dist o{go}} = \sum_{g \in G} \norm{\nu_{go}} e^{-s\dist o{go}}
\end{equation*}
is exactly $\omega_H$.
Hence $\omega_{-\chi} = \omega_\chi = \omega_H$.
Note also that, since $\nu$ is $H$-invariant, $\chi(hg) = \chi(g)$ for every $h \in H$ and $g \in G$.
Using <ref>, with the quasi-morphism $- \chi$, we produce an $H$-invariant, $\omega_H$-conformal density $\nu^* = (\nu^*_x)$ satisfying the following additional property:
there is $C_1 \in \R_+^*$ such that for every $g \in G$, for every $x \in X$, we have
\begin{equation*}
\frac 1{C_1} \nu_x \leq e^{\chi(g)} {g^{-1}}_\ast \nu_{gx} \leq C_1 \nu_x.
\end{equation*}
Denote by $\mu^*$ its restriction to the reduced horocompactification $(\bar X, \mathfrak R)$.
According to <ref><ref>, there is $C_2 \in \R_+$ such that $\mu \leq C_2 \mu^*$.
Recall that $\mu$ is $C_0$-almost fixed by $G$.
Consequently for every $g \in G$, we have
\begin{equation*}
\leq C_0 \left( {g^{-1}}_\ast \mu_{go}\right)
\leq C_0C_2 \left( {g^{-1}}_\ast\mu^*_{go} \right)
\leq C_0C_1C_2 \left(e^{-\chi(g)}\mu^*_o\right).
\end{equation*}
Since $\mu_o$ and $\mu^*_o$ are probability measures, $\chi$ is bounded, whence the result.
As we noticed in the introduction, every finite index and every normal subgroup of $G$ is commensurated.
More generally, consider a subgroup $H$ of $G$ and $N$ a normal subgroup of $G$.
If $H$ and $N$ are commensurable (i.e. $H\cap N$ has finite index in both $H$ and $N$) then $H$ is commensurated.
However there are plenty of other examples.
Here is a construction suggested by Uri Bader [4].
Consider the free group $\free 2$ and morphism $\phi \colon \free 2 \to M$ where $M$ is a topological group.
Let $K$ be an open compact subgroup of $M$.
Then $H = \phi^{-1}(K)$ is commensurated.
Now if $\phi$ has dense image, then $H$ is commensurable with a normal subgroup of $\free 2$ if and only if $K$ is commensurable with a normal subgroup of $M$.
Consider now for instance a prime $p$ and a morphism $\phi \colon \free 2 \to {\rm PSL}_2(\Q_p)$ with dense image.
Then $\phi^{-1}({\rm PSL}_2(\Z_p))$ provides an example of a commensurated subgroup of $\free 2$ that is not commensurable with a normal subgroup of $\free 2$.
§ STRONGLY POSITIVELY RECURRENT ACTIONS
§.§ Definition
Let $G$ be a group acting properly, by isometries on a proper, geodesic, metric space $X$.
Given a compact subset $K \subset X$, we define a subset $G_K \subset G$ as follows:
an element $g \in G$ belongs to $G_K$, if there exist $x,y \in K$ and a geodesic $\gamma$ joining $x$ to $gy$ such that the intersection $\gamma \cap G K$ is contained in $K \cup gK$.
Although $G_K$ is not a subgroup of $G$, its exponential growth rate $\omega(G_K, X)$ is defined in the same way as for the one of $G$.
The entropy at infinity of the action of $G$ on $X$ is
\begin{equation*}
\omega_\infty(G, X) = \inf_K \omega(G_K, X),
\end{equation*}
where $K$ runs over all compact subsets of $X$.
The action of $G$ on $X$ is strongly positively recurrent (or statistically convex co-compact) if $\omega_\infty(G, X) < \omega(G,X)$.
We refer the reader to [45, 19] for examples of strongly positively recurrent actions in the context of hyperbolic geometry.
Arzhantseva, Cashen and Tao <cit.> also observed that the work of Eskin, Mirzakani, and Rafi <cit.> implies that the action of the mapping class group on the Teichmüller space endowed with the Teichmüller metric is strongly positively recurrent.
§.§ Divergence
If the action of $G$ on $X$ is strongly positively recurrent, then it is divergent.
The statement was proved by Yang [51].
We give here an alternative approach in the spirit of Schapira-Tapie [45].
The idea is to build a $G$-invariant, $\omega_G$-conformal density which gives positive measure to the radial limit set.
Indeed, according to <ref>, this will imply that the action of $G$ on $X$ is divergent.
As we explained in <ref> this part of <ref> does not require that $G$ contains a contracting element.
First, we give a description of the complement of the radial limit set.
To that end we introduce some notations.
Given a compact subset $K \subset X$ and $\epsilon \in \R_+^*$, we denote by $A_{K, \epsilon}$ the set of all cocycles $c \in \partial X$ with the following property:
there is a point $x \in K$ such that for every $\epsilon$-quasi-gradient ray $\gamma \colon \R_+ \to X$ for $c$ starting at $x$, for every $u \in G$, if the intersection $\gamma \cap uK$ is non-empty, then $d(K, uK) \leq 1$.
The radial set of $G$ satisfies the following inclusion
\begin{equation*}
\partial X \setminus \Lambda_{\rm rad} (G)
\subset \bigcap_{K \subset X} G \left( \bigcup_{\epsilon > 0} A_{K, \epsilon}\right),
\end{equation*}
where $K$ runs over all compact subsets of $X$.
The proof is by contraposition.
Consider a cocycle $c \in \partial X$ that is not in the set
\begin{equation*}
\bigcap_{K \subset X} G \left( \bigcup_{\epsilon > 0} A_{K, \epsilon}\right).
\end{equation*}
There is a compact subset $K \subset X$ such that for every $g \in G$ and $\epsilon > 0$, the cocycle $c$ does not belong to $gA_{K, \epsilon}$.
Fix $\epsilon \in (0, 1)$ and $x_0 \in K$.
In addition we let $g_0 = 1$.
We are going to build by induction, a sequence of points $x_1, x_2\dots$ in $X$, a sequence of elements $g_1, g_2,\dots$ in $G$, and a sequence of rays $\gamma_1, \gamma_2,\dots$, such that for every $i \in \N\setminus\{0\}$ the following holds.
* $x_i$ belongs to $g_iK$.
* $c(x_0,x_i) \geq i /2$.
* For every $i \in \N\setminus \{0\}$, the path $\gamma_i$ is a $2^{-i}\epsilon$-quasi-gradient ray of $c$ starting at $x_{i-1}$ and passing through $x_i$.
Let $i \in \N$.
Assume that $x_i \in X$, $g_i \in G$ have been defined.
By assumption $c$ does not belong to the set
\begin{equation*}
g_iA_{K, 2^{-(i+1)}\epsilon}.
\end{equation*}
Hence there exists a $2^{-(i+1)}\epsilon$-quasi-gradient ray $\gamma_{i+1} \colon \R_+ \to X$ for $c$ starting at $x_i$ and an element $u_i \in G$ such that $\gamma_{i+1} \cap g_iu_iK$ is non-empty and $d(g_iK, g_iu_iK) > 1$.
We let $g_{i+1} = g_iu_i$ and denote by $x_{i+1}$ a point in $\gamma_{i+1} \cap g_iu_iK$.
Since $x_i \in g_iK$ and $x_{i+1} \in g_iu_iK$, we have $\dist {x_i}{x_{i+1}} > 1$.
However $\gamma_{i+1}$ is a quasi-gradient line.
\begin{equation*}
c(x_i, x_{i+1}) \geq \dist {x_i}{x_{i+1}} - 2^{-(i+1)}\epsilon \geq 1/2.
\end{equation*}
Using the induction hypothesis, we get
\begin{equation*}
c(x_0, x_{i+1}) \geq c(x_0, x_i) + c(x_i, x_{i+1}) \geq (i+1)/2.
\end{equation*}
Consequently $x_{i+1}$, $g_{i+1}$, and $\gamma_{i+1}$ satisfy the announced properties.
Note that the sequence $(x_i)$ is unbounded.
Indeed otherwise $c(x_0, x_i)$ should be bounded as well.
Thus we can build an infinite path $\gamma$ by concatenating the restriction of each $\gamma_i$ between $x_{i-1}$ and $x_i$.
It follows from the construction that $\gamma$ is an $\epsilon$-quasi-gradient line for $c$, see <ref>.
Moreover $\gamma$ intersects $g_iK$ for every $i \in \N$.
One proves using the triangle inequality that $c$ belongs to the radial limit set.
Let $K \subset X$ be a compact subset and $\epsilon \in \R_+^*$.
For every compact subset $F \subset X$, we define $U_{K, \epsilon}(F)$ to be the set of cocycles $b \in \bar X$ for which there is a cocycle $c \in A_{K, \epsilon}$ satisfying $\norm[F]{b - c} < \epsilon$.
Observe that $U_{K,\epsilon}(F)$ is an open subset of $\bar X$ containing $A_{K, \epsilon}$.
Let $K \subset X$ be a compact set and $\epsilon \in \R_+^*$.
Fix a base point $o \in K$.
There exist $r \in \R_+$ and a finite subset $S \subset G$, such that for every $T \geq \epsilon$, if $F$ stands for the closed ball of radius $T + 2r$ centered at $o$, then
\begin{equation*}
U_{K, \epsilon}(F) \cap Go \subset S \left(\bigcup_{\substack{k \in G_K \\ \dist o{ko} \geq T}} \mathcal O_o(ko,r)\right).
\end{equation*}
Since the action of $G$ on $X$ is proper, the set
\begin{equation*}
S = \set{u \in G}{d(K, uK) \leq 1}
\end{equation*}
if finite.
We fix $r > 2 \diam K + 1$.
Let $T \geq \epsilon$ and $F$ be the closed ball of radius $R = T+2r$ centered at $o$.
Let $g \in G$ such that $go$ belongs to $U_{K, \epsilon}(F)$.
We write $b = \iota(go)$ for the corresponding cocycle.
By definition, there is $c \in A_{K, \epsilon}$ such that $\norm[F]{b-c} < \epsilon$.
Observe first that $\dist o{go} > R - \epsilon$.
Indeed the map $x \mapsto b(x,go)$ admits a global minimum at $go$, while there exists a $c$-gradient line starting at $go$.
We cannot have at the same time $\dist o{go} \leq R - \epsilon$ and $\norm[F]{b-c} < \epsilon$.
In particular, $g \notin S$.
Since $c \in A_{K, \epsilon}$, there exists $x \in K$, such that for every $\epsilon$-quasi-gradient ray $\gamma \colon \R_+ \to X$ for $c$, starting at $x$, if $\gamma$ intersects $uK$ for some $u \in G$, then $u \in S$.
Consider now a geodesic $\alpha\colon \intval 0\ell \to X$ from $x$ to $go$.
We denote by $s \in \intval 0\ell$, the largest time such that the point $y = \alpha(s)$ belongs to $SK$.
We now denote by $t \in \intval s\ell$, the smallest time such that the point $z = \alpha(t)$ lies in $hK$ for some $h \in G\setminus S$ (such a time $t$ exists since $\alpha(\ell)$ belongs to $gK$).
It follows from the construction that $h$ can be written $h = uk$ with $u \in S$ and $k \in G_K$.
Moreover $y \in uK$.
Observe that $\gro y{go}z = 0$, while $\dist y{uo} \leq r/2$ and $\dist z{uko}\leq r/2$.
The triangle inequality yields $\gro {uo}{go}{uko} \leq r$, i.e. $go$ belongs to $u\mathcal O_o(ko, r)$.
We are left to prove that $\dist o{ko} \geq T$.
By construction $\dist o{uo} \leq r$.
Thanks to the triangle inequality, it suffices to show that $\dist oz \geq R$.
Assume on the contrary that $\dist oz < R$.
In particular, both $x$ and $z$ belong to $F$.
Since $b$ and $c$ differ by at most $\epsilon$ on $F$, we get that $c(x,z) \geq \dist xz - \epsilon$.
Hence any geodesic from $x$ to $z$ is an $\epsilon$-quasi-gradient arc for $c$.
If we concatenate this path with a gradient ray for $c$ starting at $z$, we obtain an $\epsilon$-quasi-gradient ray for $c$ starting at $x$ and intersecting $hK$ with $d(K, hK) > 1$.
This contradicts the fact that $c$ belongs to $A_{K, \epsilon}$, and completes the proof.
If the action of $G$ on $X$ is strongly positively recurrent, then there is a $G$-invariant, $\omega_G$-conformal density which gives full measure to the radial limit set $\Lambda_{\rm rad}(G)$.
By definition, there is a compact subset $K \subset X$ such that $\omega_{G_K} < \omega_G$.
We fix once and for all a base point $o \in K$.
The argument relies on Patterson's construction recalled in the proof of <ref> with $H = G$ and $\chi$ the trivial morphism.
In particular, $\mathcal Q(s)$ stands for the weighted Poincaré series defined in (<ref>).
For every $s > \omega_G$, we consider the density $\nu^s = (\nu_x^s)$ defined as in (<ref>).
As we explained there is a sequence $(s_n)$ converging to $\omega_G$ from above such that $\nu^{s_n}$ converges to a $G$-invariant, $\omega_G$-conformal density $\nu$ supported on $\partial X$.
Let $\eta > 0$ such that $\omega_{G_K} < \omega_G - \eta$.
The weight $\theta$ used to construct $\nu$ is slowing increasing.
More precisely, according to <ref> there exists $t_0$ such that for every $t \geq t_0$ and $u \in \R_+$ we have $\theta(t + u) \leq e^{\eta u}\theta(t)$.
Let $\epsilon > 0$.
Let $r \in \R_+$ and $S \subset G$ be the data provided by <ref> applied with $K$ and $\epsilon$.
For every $T \in \R_+$, we write $F_T$ for the closed ball of radius $R =T+2r$ centered at $o$.
Let $s > \omega_G$ and $T \geq \max\{t_0, \epsilon\}$.
In view of <ref>, we have
\begin{equation*}
\nu^s_o\left(U_{K,\epsilon}(F_T)\right)
\leq \card S \sum_{\substack{k \in G_K \\ \dist o{ko} \geq T}} \nu^s_o\left(\mathcal O_o(ko, r) \right).
\end{equation*}
Let us estimate the measures of the shadows in the sum.
Let $k \in G_K$, such that $\dist o{ko} \geq T$.
Any element $g \in G$ such that $go \in \mathcal O_o(ko,r)$ can be written $g = ku$ with $u \in G$ and
\begin{equation*}
\dist o{ko} + \dist o{uo} - 2r \leq \dist o{go} \leq \dist o{ko} + \dist o{uo}.
\end{equation*}
Unfolding the definition of $\nu^s$, we get
\begin{equation}
\label{eqn: spr gives full measure}
\nu^s_o\left(\mathcal O_o(ko, r) \right)
\leq \frac{e^{2sr}e^{-s \dist o{ko}}}{\mathcal Q(s)}\sum_{u \in G} \theta(\dist o{go}) e^{-s \dist o{uo}}.
\end{equation}
Observe that if $\dist o{uo} \geq t_0$, then it follows from our choice of $t_0$ that
\begin{equation*}
\theta\left(\dist o{go}\right)
\leq \theta\left( \dist o{ko} + \dist o{uo}\right)
\leq e^{\eta\dist o{ko}} \theta\left(\dist o{uo}\right).
\end{equation*}
Otherwise, since $\dist o{ko}\geq T \geq t_0$, we have
\begin{equation*}
\theta\left(\dist o{go}\right)
\leq \theta\left(t_0 + \dist o{uo}+ \dist o{ko} -t_0 \right)
\leq e^{\eta\dist o{ko}} \theta\left(t_0\right).
\end{equation*}
We break the sum in (<ref>) according to the length of $u$ and get
\begin{equation*}
\nu^s_o\left(\mathcal O_o(ko, r) \right)
\leq \frac{e^{2sr}e^{-(s-\eta) \dist o{ko}}}{\mathcal Q(s)} \left[\theta(t_0)\Sigma_1(s) + \Sigma_2(s)\right],
\end{equation*}
\begin{align*}
\Sigma_1(s) & = \sum_{\substack{u \in G \\ \dist o{uo} \leq t_0}} e^{-s \dist o{uo}}, \quad \text{and}\quad \\
\Sigma_2(s) & = \sum_{\substack{u \in G \\ \dist o{uo} > t_0}} \theta(\dist o{uo}) e^{-s \dist o{uo}}.
\end{align*}
Note that $\Sigma_1(s)$ is a finite sum that does not depend on $k$, while $\Sigma_2(s)$ is the remainder of the series $\mathcal Q(s)$.
\begin{equation*}
\nu^s_o\left(\mathcal O_o(ko, r) \right)
\leq e^{2sr}
\left[ \frac {\theta(t_0)}{\mathcal Q(s)} \Sigma_1(s) + 1\right] e^{-(s-\eta) \dist o{ko}},
\end{equation*}
Summing over all long elements $k \in G_K$, we get
\begin{equation*}
\nu^s_o\left(U_{K,\epsilon}(F_T)\right)
\leq \card S e^{2sr}
\left[ \frac {\theta(t_0)}{\mathcal Q(s)} \Sigma_1(s) + 1\right]
\sum_{\substack{k \in G_K \\ \dist o{ko} \geq T}} e^{-(s-\eta) \dist o{ko}}.
\end{equation*}
Note that $\Sigma_1(s)$ is bounded, while $\mathcal Q(s)$ diverges to infinity.
Since $U_{K,\epsilon}(F_T)$ is an open subset of $\bar X$, we can pass to the limit and get
\begin{equation*}
\nu_o\left(U_{K,\epsilon}(F_T)\right)
\leq \card S e^{2\omega_Gr}
\sum_{\substack{k \in G_K \\ \dist o{ko} \geq T}} e^{-(\omega_G-\eta) \dist o{ko}}.
\end{equation*}
The sum corresponds to the remainder of the Poincaré series of $G_K$ at $s = \omega_G - \eta$.
However $\omega_G - \eta > \omega_{G_K}$.
Hence this series converges, and its reminder tends to zero when $T$ approaches infinity.
Consequently, for every $\epsilon > 0$,
\begin{equation*}
\nu_o\left(\bigcap_{T \geq 0}U_{K,\epsilon}(F_T)\right) = 0.
\end{equation*}
By construction the set $A_{K, \epsilon}$ is contained in $U_{K, \epsilon}(F_T)$ for every $T \in \R_+$.
It follows from <ref> that
\begin{equation*}
\partial X \setminus \Lambda_{\rm rad} (G)
\subset G \left( \bigcup_{\epsilon > 0} \bigcap_{T \geq 0}U_{K, \epsilon}(F_T)\right).
\end{equation*}
Since $G$ is countable we conclude that $\nu_o(\partial X \setminus \Lambda_{\rm rad}(G)) = 0$.
Recall that $\nu_o$ is supported on $\partial X$, thus $\nu_o(\Lambda_{\rm rad}(G)) = 1$.
[1]
G. Arzhantseva, C. H. Cashen, and J. Tao.
Growth tight actions.
Pacific Journal of Mathematics, 278(1):1–49, Sept. 2015.
[2]
G. N. Arzhantseva and C. H. Cashen.
Cogrowth for group actions with strongly contracting elements.
Ergodic Theory Dynam. Systems, 40(7):1738–1754, 2020.
[3]
U. Bader and A. Furman.
Some ergodic properties of metrics on hyperbolic groups.
arXiv 1707.02020, July 2017.
[4]
U. Bader.
Commensurated subgroups of a free group, May 2018.
[5]
M. Bestvina, K. Bromberg, and K. Fujiwara.
Constructing group actions on quasi-trees and applications to mapping
class groups.
Publ. Math., Inst. Hautes Étud. Sci., 122:1–64, 2015.
[6]
M. Bestvina and K. Fujiwara.
A characterization of higher rank symmetric spaces via bounded
Geometric and Functional Analysis, 19(1):11–40, 2009.
[7]
M. R. Bridson and A. Haefliger.
Metric Spaces of Non-Positive Curvature, volume 319 of Grundlehren Der Mathematischen Wissenschaften [Fundamental Principles of
Mathematical Sciences].
Springer-Verlag, Berlin, 1999.
[8]
R. Brooks.
The fundamental group and the spectrum of the Laplacian.
Commentarii Mathematici Helvetici, 56(4):581–598, 1981.
[9]
M. Burger and S. Mozes.
CAT(-1)-spaces, divergence groups and their commensurators.
Journal of the American Mathematical Society, 9(1):57–93,
[10]
C. Champetier.
Cocroissance des groupes à petite simplification.
The Bulletin of the London Mathematical Society,
25(5):438–444, 1993.
[11]
R. Charney and H. Sultan.
Contracting boundaries of CAT(0) spaces.
Journal of Topology, 8(1):93–117, 2015.
[12]
J. M. Cohen.
Cogrowth and amenability of discrete groups.
Journal of Functional Analysis, 48(3):301–309, 1982.
[13]
M. Coornaert.
Mesures de Patterson-Sullivan sur le bord d'un espace hyperbolique
au sens de Gromov.
Pacific Journal of Mathematics, 159(2):241–270, 1993.
[14]
M. Coornaert, T. Delzant, and A. Papadopoulos.
Géométrie et théorie des groupes, volume 1441 of Lecture Notes in Mathematics.
Springer-Verlag, Berlin, 1990.
[15]
M. Coornaert and A. Papadopoulos.
Horofunctions and symbolic dynamics on Gromov hyperbolic groups.
Glasgow Mathematical Journal, 43(3):425–456, 2001.
[16]
M. Cordes.
Morse Boundaries of Proper Geodesic Metric Spaces.
arXiv 1502.04376, Feb. 2015.
[17]
R. Coulon.
Growth of periodic quotients of hyperbolic groups.
Algebraic & Geometric Topology, 13(6):3111–3133, 2013.
[18]
R. Coulon, F. Dal'bo, and A. Sambusetti.
Growth gap in hyperbolic groups and amenability.
Geometric and Functional Analysis, 28(5):1260–1320, 2018.
[19]
R. Coulon, R. Dougall, B. Schapira, and S. Tapie.
Twisted Patterson-Sullivan measures and applications to amenability
and coverings.
arXiv 1809.10881, to appear in Memoirs of the AMS., Sept. 2018.
[20]
R. Coulon.
Ergodicity of the geodesic flow for groups with a contracting
arXiv 2303.01390, 2023.
[21]
F. Dahmani, D. Futer, and D. T. Wise.
Growth of quasiconvex subgroups.
Math. Proc. Cambridge Philos. Soc., 167(3):505–530, 2019.
[22]
A. Eskin, M. Mirzakhani, and K. Rafi.
Counting closed geodesics in strata.
Invent. Math., 215(2):535–607, 2019.
[23]
P. Eymard.
Moyennes invariantes et représentations unitaires.
Lecture Notes in Mathematics, Vol. 300. Springer-Verlag, Berlin-New
York, 1972.
[24]
V. Gerasimov and L. Potyagailo.
Quasiconvexity in relatively hyperbolic groups.
J. Reine Angew. Math., 710:95–135, 2016.
[25]
R. I. Grigorchuk.
Symmetric random walks on discrete groups.
Akademiya Nauk SSSR i Moskovskoe Matematicheskoe Obshchestvo.
Uspekhi Matematicheskikh Nauk, 32(6(198)):217–218, 1977.
[26]
R. I. Grigorchuk.
Symmetrical random walks on discrete groups.
In Multicomponent random systems, pages 285–325. Dekker, New
York, 1980.
[27]
M. Gromov.
Hyperbolic groups.
In Essays in group theory, pages 75–263. Springer, New York,
New York, 1987.
[28]
B. M. Gurevich and S. V. Savchenko.
Thermodynamic formalism for symbolic Markov chains with a countable
number of states.
Akademiya Nauk SSSR i Moskovskoe Matematicheskoe Obshchestvo.
Uspekhi Matematicheskikh Nauk, 53(2(320)):3–106, 1998.
[29]
J. Jaerisch and K. Matsuzaki.
Growth and cogrowth of normal subgroups of a free group.
Proceedings of the American Mathematical Society,
145(10):4141–4149, 2017.
[30]
G. Knieper.
On the asymptotic geometry of nonpositively curved manifolds.
Geom. Funct. Anal., 7(4):755–782, 1997.
[31]
G. Knieper.
The uniqueness of the measure of maximal entropy for geodesic flows
on rank 1 manifolds.
Ann. Math. (2), 148(1):291–314, 1998.
[32]
S. Kochen and C. Stone.
A note on the Borel-Cantelli lemma.
Ill. J. Math., 8:248–251, 1964.
[33]
J. Li and D. T. Wise.
No growth-gaps for special cube complexes.
Groups Geom. Dyn., 14(1):117–135, 2020.
[34]
G. Link.
Hopf-Tsuji-Sullivan dichotomy for quotients of Hadamard spaces
with a rank one isometry.
Discrete Contin. Dyn. Syst., 38(11):5577–5613, 2018.
[35]
G. A. Margulis.
Certain applications of ergodic theory to the investigation of
manifolds of negative curvature.
Akademija Nauk SSSR. Funkcional nyi Analiz i ego Prilo zenija,
3(4):89–90, 1969.
[36]
K. Matsuzaki and Y. Yabuki.
The Patterson-Sullivan measure and proper conjugation for
Kleinian groups of divergence type.
Ergodic Theory Dyn. Syst., 29(2):657–665, 2009.
[37]
K. Matsuzaki, Y. Yabuki, and J. Jaerisch.
Normalizer, divergence type, and Patterson measure for discrete
groups of the Gromov hyperbolic space.
Groups Geom. Dyn., 14(2):369–411, 2020.
[38]
Y. N. Minsky.
Quasi-projections in Teichmüller space.
J. Reine Angew. Math., 473:121–136, 1996.
[39]
D. Murray.
Topology and dynamics of the contracting boundary of cocompact $\rm
CAT(0)$ spaces.
Pacific J. Math., 299(1):89–116, 2019.
[40]
S. J. Patterson.
The limit set of a Fuchsian group.
Acta Mathematica, 136(3-4):241–273, 1976.
[41]
Y. Qing and K. Rafi.
Sublinearly Morse boundary I: CAT(0) spaces.
Adv. Math., 404:Paper No. 108442, 51, 2022.
[42]
Y. Qing, K. Rafi, and G. Tiozzo.
Sublinearly morse boundary ii: Proper geodesic spaces.
arXiv 2011.03481, 2020.
[43]
T. Roblin.
Un Théorème de Fatou Pour les Densités Conformes Avec
Applications Aux Revêtements Galoisiens en Courbure Négative.
Israel Journal of Mathematics, 147(1):333–357, Dec. 2005.
[44]
O. M. Sarig.
Thermodynamic formalism for null recurrent potentials.
Israel Journal of Mathematics, 121:285–311, 2001.
[45]
B. Schapira and S. Tapie.
Regularity of entropy, geodesic currents and entropy at infinity.
Ann. Sci. Éc. Norm. Supér. (4), 54(1):1–68, 2021.
[46]
A. G. Shukhov.
On the dependence of the growth exponent on the length of the
defining relation.
65(4):612–618, 1999.
[47]
A. Sisto.
On metric relative hyperbolicity.
arXiv:1210.8081, 2012.
[48]
A. Sisto.
Contracting elements and random walks.
J. Reine Angew. Math., 742:79–114, 2018.
[49]
D. Sullivan.
Related aspects of positivity in Riemannian geometry.
Journal of Differential Geometry, 25(3):327–351, 1987.
[50]
W. Yang.
Growth tightness for groups with contracting elements.
Mathematical Proceedings of the Cambridge Philosophical
Society, 157(2):297–319, 2014.
[51]
W. Yang.
Statistically convex-cocompact actions of groups with contracting
Int. Math. Res. Not. IMRN, (23):7259–7323, 2019.
[52]
W. Yang.
Genericity of contracting elements in groups.
Math. Ann., 376(3-4):823–861, 2020.
[53]
W. Yang.
Conformal dynamics at infinity for groups with contracting elements.
arXiv 2208.04861, August 2022.
Rémi Coulon
Université de Bourgogne, CNRS
IMB - UMR 5584
F-21000 Dijon, France
|
# Luminous Late-time Radio Emission from Supernovae Detected by the Karl G.
Jansky Very Large Array Sky Survey (VLASS)
Michael C. Stroh Center for Interdisciplinary Exploration and Research in
Astrophysics (CIERA) and Department of Physics and Astronomy, Northwestern
University, Evanston, IL 60201, USA Giacomo Terreran Center for
Interdisciplinary Exploration and Research in Astrophysics (CIERA) and
Department of Physics and Astronomy, Northwestern University, Evanston, IL
60201, USA Las Cumbres Observatory, 6740 Cortona Drive, Suite 102, Goleta, CA
93117-5575, USA Deanne L. Coppejans Center for Interdisciplinary Exploration
and Research in Astrophysics (CIERA) and Department of Physics and Astronomy,
Northwestern University, Evanston, IL 60201, USA Department of Physics,
University of Warwick, Gibbet Hill Road, Coventry CV4 7AL, UK Joe S. Bright
Center for Interdisciplinary Exploration and Research in Astrophysics (CIERA)
and Department of Physics and Astronomy, Northwestern University, Evanston, IL
60201, USA Department of Astronomy, University of California, Berkeley, CA
94720, USA Raffaella Margutti Center for Interdisciplinary Exploration and
Research in Astrophysics (CIERA) and Department of Physics and Astronomy,
Northwestern University, Evanston, IL 60201, USA Department of Astronomy,
University of California, Berkeley, CA 94720, USA Michael F. Bietenholz
Department of Physics and Astronomy, York University, Toronto, M3J 1P3,
Ontario, Canada Hartebeesthoek Radio Observatory, P.O. Box 443, Krugersdorp,
1740, South Africa Fabio De Colle Instituto de Ciencias Nucleares,
Universidad Nacional Autónoma de México, A. P. 70-543 04510 D.F. Mexico
Lindsay DeMarchi Center for Interdisciplinary Exploration and Research in
Astrophysics (CIERA) and Department of Physics and Astronomy, Northwestern
University, Evanston, IL 60201, USA Rodolfo Barniol Duran Department of
Physics and Astronomy, California State University, Sacramento, 6000 J Street,
Sacramento, CA 95819, USA Danny Milisavljevic Department of Physics and
Astronomy, Purdue University, 525 Northwestern Avenue, West Lafayette, IN
47907, USA Kohta Murase Department of Physics, The Pennsylvania State
University, University Park, PA 16802, USA Department of Astronomy &
Astrophysics, The Pennsylvania State University, University Park, PA 16802,
USA Center for Mulitmessenger Astrophysics, Institute for Gravitation and the
Cosmos, The Pennsylvania State University, University Park, PA 16802, USA
Center for Gravitational Physics, Yukawa Institute for Theoretical Physics,
Kyoto University, Sakyo-ku, Kyoto 606-8502, Japan Kerry Paterson Center for
Interdisciplinary Exploration and Research in Astrophysics (CIERA) and
Department of Physics and Astronomy, Northwestern University, Evanston, IL
60201, USA Wendy L. Williams Leiden Observatory, Leiden University, PO Box
9513, NL-2300 RA Leiden, the Netherlands
###### Abstract
We present a population of 19 radio-luminous supernovae (SNe) with emission
reaching $L_{\nu}{\sim}10^{26}-10^{29}\,\rm{erg\,s^{-1}Hz^{-1}}$ in the first
epoch of the Very Large Array Sky Survey (VLASS) at $2-4$ GHz. Our sample
includes one long Gamma-Ray Burst, SN 2017iuk/GRB171205A, and 18 core-collapse
SNe detected at $\approx(1-60)$ years after explosion. No thermonuclear
explosion shows evidence for bright radio emission, and hydrogen-poor
progenitors dominate the sub-sample of core-collapse events with spectroscopic
classification at the time of explosion (79%). We interpret these findings
into the context of the expected radio emission from the forward shock
interaction with the circumstellar medium (CSM). We conclude that these
observations require a departure from the single wind-like density profile
(i.e., $\rho_{\rm{CSM}}\propto r^{-2}$) that is expected around massive stars
and/or a departure from a spherical Newtonian shock. Viable alternatives
include the shock interaction with a detached, dense shell of CSM formed by a
large effective progenitor mass-loss rate $\dot{M}\sim(10^{-4}-10^{-1})$ M⊙
yr-1 (for an assumed wind velocity of $1000\,\rm{km\,s^{-1}}$); emission from
an off-axis relativistic jet entering our line of sight; or the emergence of
emission from a newly-born pulsar-wind nebula. The relativistic SN 2012ap that
is detected 5.7 and 8.5 years after explosion with $L_{\nu}{\sim}10^{28}$ erg
s-1 Hz-1 might constitute the first detections of an off-axis jet+cocoon
system in a massive star. However, none of the VLASS-SNe with archival data
points are consistent with our model off-axis jet light curves. Future multi-
wavelength observations will distinguish among these scenarios. Our VLASS
source catalogs, which were used to perform the VLASS cross matching, are
publicly available athttps://doi.org/10.5281/zenodo.4895112 (catalog
10.5281/zenodo.4895112).
Core-collapse supernovae, Radio transient sources, Sky surveys, Very Large
Array
††software: APLpy (Robitaille & Bressert, 2012; Robitaille, 2019), Astropy
(Astropy Collaboration et al., 2013, 2018), BOXFIT (van Eerten et al., 2012),
Pandas (Wes McKinney, 2010), PyBDSF (Mohan & Rafferty, 2015), Q3C (Koposov &
Bartunov, 2006), SExtractor (Bertin & Arnouts, 1996)
## 1 Introduction
Radio observations of stellar explosions in the years-to-decades after stellar
demise constitute a probe of the physical properties of the fastest ejecta in
the explosion (i.e. their velocity and energy), and of the environment at
large distances of $r\geq 10^{17}\,\rm{cm}$ (i.e., the density of the
circumstellar medium, CSM), e.g., Chevalier & Fransson (2017). There are three
main sources of bright non-thermal synchrotron radio emission in SNe at $t\geq
1$ yr: (i) the deceleration of the forward shock in a dense environment (e.g.,
Chevalier, 1998; Chevalier & Fransson, 2006); (ii) emission from an off-axis
relativistic jet entering our line of sight (e.g., Granot et al., 2002); (iii)
emergence of emission from a newly-formed pulsar-wind nebula (PWN; Slane,
2017). Late-time radio observations of cosmic explosions can thus reveal a
complex mass-loss history of the stellar progenitors in the years leading up
to core-collapse; they can reveal jet-driven explosions similar to long Gamma-
Ray Bursts (GRBs) that launched a jet that was misaligned with our line of
sight; or they can reveal the energetics and properties of the compact object
remnant. However, most SNe are not observed at radio wavelengths at very late
times. For example, in the sample of 294 SNe observed at $\sim 5-8$ GHz
compiled by Bietenholz et al. (2021), only 87 were observed at more than 1000
days post explosion and of these only 28 were detected111This sample is one of
the largest compilations of radio observations of SNe. We note that it is not
a complete sample (see Bietenholz et al. 2021), so these numbers are likely
underestimates. One important selection effect is that the peak frequency of
the radio emission from SNe declines with time, so lower frequency
observations would yield a higher detection rate.. As a result, the late-time
radio emission from SNe constitutes a poorly explored region of the phase
space (Figure 2 in Bietenholz et al. 2021).
Here we present a sample of 19 radio-luminous SNe detected in the first epoch
of the Very Large Array Sky Survey (VLASS, Lacy et al., 2020) carried out by
the Karl G. Jansky Very Large Array (VLA). VLASS is a successor and
complementary survey to the National Radio Astronomy Observatory (NRAO) VLA
Sky Survey (NVSS, Condon et al., 1998) and Faint Images of the Radio Sky at
Twenty centimeters (FIRST, Becker et al., 1995) surveys. The survey is
conducted at $2$ to $4$ GHz and is split into three distinct epochs, each
scanning the full survey region (declination $\delta{>}{-}40^{\circ}$) with an
${\approx}\,32$ month observing cadence reaching an RMS noise of
$120\mu$Jy/beam per epoch. Exploring ‘Hidden Explosions and Transient Events’
is a key VLASS science theme, and Hallinan et al. (2020) demonstrated the
synergy between VLASS and newly discovered transients with the detection of
the type II SN 2019xhb in the 2nd VLASS observing epoch 202 days after
discovery. Importantly, by scanning the northern sky, VLASS offers the
opportunity to perform a systematic and unbiased survey of the late-time radio
emission from the tens of thousands of previously reported SNe. With the
exception of SN 2017iuk/GRB171205A, the sample of radio luminous SNe presented
in this paper was imaged at an epoch corresponding to $\approx(1-60)$ years
post explosion.
This paper is organized as follows. In Section 2, we describe our methodology
for identifying optically detected SNe in the VLASS sample, how we filtered
out potential spurious detections, and present our list of VLASS detected SNe.
Section 3 discusses physical processes that could produce the bright radio
emission associated with the SNe. Finally, in Section 4 we summarize the
conclusions.
## 2 A Sample of SNe with luminous late-time radio emission
We created a catalog consisting of all publicly-announced, optical SNe by
combining the Bright Supernovae222http://rochesterastronomy.org/snimages, Open
SNe333https://sne.space, and Transient Name Server444https://www.wis-tns.org
catalogs. We included all optical SNe detected prior to 2020-01-01 leading to
an initial sample of $\approx$70,000 unique SNe. Two independently generated
VLASS detection catalogs were produced using the ${\approx}35,000$ VLASS epoch
1 quick look images provided by NRAO.
### 2.1 PyBDSF detections
The complete VLASS epoch 1 was processed using the Python Blob Detector and
Source Finder version 1.9.1 (PyBDSF, Mohan & Rafferty, 2015). We created a
local catalog consisting of all detections using a source detection threshold
of 5$\sigma$ (thresh_pix=5.0), threshold for islands of 3$\sigma$
(thresh_isl=3.0), and fixing source components to be Gaussian with major axis,
minor axis, and position angle equal to the synthesised beam shape from the
respective VLASS observations (fix_to_beam=True). The values of thresh_pix and
thresh_isl are used to calculate pixel islands of significant emission. The
thresh_pix parameter is used to identify significant pixels (where the pixel
value is greater than mean + $\texttt{thresh\\_pix}\times\sigma$, where
$\sigma$ is the rms noise of the image) used for fitting, with the mean map
calculated using a box of pixel size and pixel step size either calculated
within PyBDSF or set manually by the user. The fitting region is then extended
based on the value of thresh_isl such that all pixels greater than mean +
$\texttt{thresh\\_isl}\times\sigma$ which are adjacent (including diagonally)
to a significant pixel identified using thresh_pix are included in the fitting
region or ‘island’ (if this process causes islands to overlap they are
combined). Multiple Gaussians are then fit to the island in order to best
describe the source (the collection of Gaussians within an island) and we
fixed the Gaussians to the shape of the synthesized beam.
This VLASS-PyBDSF catalog resulted in 3 752 214 sources, similar to the 3 381
277 VLASS sources cataloged by the Canadian Initiative for Radio Astronomy
Data Analysis (CIRADA, Gordon et al., 2020), and more than the 2 232 726
sources in the VLASS Quicklook Catalog (Bruzewski et al., 2021). Our VLASS-
PyBDSF catalog is publicly available athttps://doi.org/10.5281/zenodo.4895112
(catalog 10.5281/zenodo.4895112).
Figure 1: Field of SN 2012ap in the Pan-STARRS1 r-band (left) and at 3 GHz
with VLASS (right). The red lines indicate the optical SN position, and each
image notes the date of each observation. The optical image clearly shows the
host galaxy of SN 2012ap, but no emission from the host is found in the
optical image near the optical SN position. No radio emission is found in the
field near SN 2012ap except near the optical SN position. The VLASS image was
taken 2065 days since explosion (Milisavljevic et al., 2014).
### 2.2 Source extractor detections
We created an initial list of possible VLASS source detections using Source
Extractor version 2.25.0 (SExtractor, Bertin & Arnouts, 1996). SExtractor is
not optimized for radio imaging analysis, but through trial and error with
fields where transients were known to exist, we settled on a requirement of at
least 5 contiguous pixels (detect_minarea) to be above the low
detect_threshold=5$\sigma$. The VLASS-SExtractor catalog contained 9 652 665
possible source detections. Most of these detections are likely spurious, as
evidenced by the number of sources detected by the previous section. We note
that SExtractor runs at least an order of magnitude faster through VLASS quick
look images than PyBDSF, so it may be preferred for studies where a large
number of possible bogus detections are acceptable. Our VLASS-SExtractor
catalog is publicly available athttps://doi.org/10.5281/zenodo.4895112
(catalog 10.5281/zenodo.4895112).
### 2.3 Identifying supernovae for cross matching
We chose low detection thresholds in order to minimize the chance that we may
miss potential SNe cross matches on our initial pass. In order to reduce the
number of spurious matches, we required that the potential SNe cross matches
must be detected by both VLASS-SExtractor and VLASS-PyBDSF catalogs. We cross-
matched the locations of the optical SNe with the VLASS-PyBDSF and VLASS-
SExtractor catalogs using a $5^{\prime\prime}$ angular separation. This
separation is ${\approx}2$ times the average VLASS beam size, and helps
account for a lack of positional precision in SN discovery reports. Using our
initial list of ${\approx}70,000$ SNe, ${\approx}1600$, and ${\approx}1400$
SNe have potential PyBDSF, and SExtractor cross-matches, respectively. By
requiring that a source must have cross-matches in both the VLASS-PyBDSF and
VLASS-SExtractor catalogs, we have ${\approx}1300$ potential cross-matches in
VLASS. We further required that the VLASS observation must have occurred
_after_ the SN discovery date.
The possible VLASS-SNe were visually inspected to ensure that the VLASS-SNe
detections are real and are not due to radio imaging artifacts (see for
example Figure 1). We also rejected SNe when the location of the radio source
broadly overlapped with that of the galactic nucleus. After visual inspection,
only ${\approx}100$ potential VLASS-SNe detections remained.
### 2.4 Multiwavelength cross matching
In order to filter out known radio sources, we rejected associations that had
counterparts in the NVSS or FIRST catalogs prior to their explosion date. We
removed VLASS-SNe near active galactic nuclei (AGN) identified by Assef et al.
(2018), who cataloged probable AGN in the Wide-field Infrared Survey Explorer
(WISE, Wright et al., 2010) AllWISE data release (Cutri et al., 2021).
D’Abrusco et al. (2013) examined the chance of probability for spurious
associations between a sample of NVSS detected blazars and AllWISE. They
calculated the number of additional cross-matches between their NVSS blazar
catalog and AllWISE $\Delta N_{t}$ as the cross-matching radius increases.
Similarly, the number of additional spurious cross-matches per increasing
cross-matching radius, $\Delta N_{r}$, was calculated by adding a random
offset to the NVSS blazar positions. D’Abrusco et al. (2013) found that
$\Delta N_{r}>\Delta N_{t}$ for cross-matching radii above
$3.3^{\prime\prime}$, thus a cross-matching radius of $3.3^{\prime\prime}$ can
be considered a cross-matching between point-like VLA sources and AllWISE. For
the infrared AGN cross matching, we adopted the $3.3^{\prime\prime}$ angular
search radius suggested by D’Abrusco et al. (2013).
Possibly misidentified AGN were also removed by cross matching our VLASS-SNe
candidates against the Chandra Source Catalog v2.0 (CSC 2.0, Evans et al.,
2020a), the most recent XMM-Newton X-ray source data release (4XMM-DR10, Webb
et al., 2020), and the 2nd Swift-XRT point source catalog (2SXPS, Evans et
al., 2020b). The error in the X-ray position is generally greater than the
astrometric uncertainties in the VLASS positions (see the Appendix of
Bruzewski et al., 2021), thus for X-ray catalog cross-matching, we removed
sources within the 1$\sigma$ X-ray error region in the respective X-ray
catalog. We ensured that no VLASS-SNe candidates were rejected by targeted SNe
follow-up observations.
We further inspected the VLASS-SNe candidates within the 424 square degrees
covered so far by the LOFAR Two-metre Sky Survey (LoTSS Data Release 1,
Shimwell et al., 2019), to ensure that none were classified as AGN based on
any of their $150$ MHz radio morphology or luminosity, or based on their
multi-wavelength cross-identifications (Williams et al., 2019). While only two
of the candidates lie within this area, both were detected as star-forming
galaxies. Further releases of LoTSS over the Northern sky will enable further
such comparisons.
This multiwavelength filtering procedure leads to a sample of 19 core-collapse
SNe with associated VLASS emission. The final VLASS-SNe sample is listed in
Table 1. All VLASS-SNe also have counterparts in the CIRADA VLASS catalogue,
and SN 2017hcb is the only source without a cross-match in the VLASS Quicklook
Catalog. Interestingly, we note that no thermonuclear explosion (i.e. Ia-like)
passed the criteria above, in spite of it largely dominating the initial
optical SN sample. The lack of type Ia SNe in the sample is consistent with
the lack of radio emission associated with type Ia (e.g., Chomiuk et al.,
2016), but the lack chance coincidence matches may be evidence of the strength
of our multiwavelength filtering described above. Of the 14 VLASS detected SNe
with early-time spectroscopic classification, 13 (93%) and 12 (86%) are
detected at ${>}10^{2}$ and ${>}10^{3}$ days post explosion, respectively. SNe
with hydrogen poor progenitors at the time of explosion make-up the majority
11 (79%) of the sources with early-time spectroscopic classification.
Remarkably, we find that SN 2012ap, one of the only two known SNe with
relativistic ejecta without a GRB (Margutti et al., 2014a; Chakraborti et al.,
2015; Milisavljevic et al., 2015a), shows evidence for bright radio emission
years after explosion, and is a member of the sample. The very nearby SN
2017iuk/GRB171205A, at redshift $z{=}0.0368$ (de Ugarte Postigo et al., 2017),
was detected in the first VLASS epoch less than 60 days following the Swift
trigger.
### 2.5 Final flux densities
In addition to a complete and consistent processing of VLASS epoch 1 (as
described in Section 2.1) we performed an optimized manual analysis on each of
the fields containing the sources listed in Table 1. The default significance
parameters used for this analysis were thresh_isl=5.0 and thresh_pix=5.0, with
an adaptive region used to calculate the RMS and mean maps
(adaptive_rms_box=True), and components fixed to be the same shape as the
synthesised beam (fix_to_beam=True). The island and pixel threshold values
were adjusted on a per-field basis depending on e.g. bright source artifacts
or extended host structure; however, we require
$\texttt{thresh\\_pix}\geq\texttt{thresh\\_isl}$. Adaptive RMS calculation
ensures that region size near bright sources is reduced compared to regions
devoid of bright emission, properly accounting for elevated RMS levels
resulting from imaging artifacts (an issue in a number of the VLASS fields).
For fields with particularly strong artifacts we set the RMS box size and step
size manually such that the noise map captured the variation caused by the
artifacts. In the cases where the SN emission formed part of an extended
emission complex from the host galaxy we set fix_to_beam=False in order to
better model the emitting region. We ran PyBDSF in interactive mode
(interactive=True) and manually inspected the result of the island and source
detection, adjusting our significance threshold and the size of the region
used to calculate the RMS noise to improve the source fitting.
For fields with extreme imaging artifacts around bright sources, we manually
set the RMS box size and disabled adaptive region sizing. In a handful of
cases, emission from the transient was part of a larger emission complex
(radio emission from the host galaxies) and the emission island was better
described using Gaussians with unconstrained shapes. We list the results of
our fitting in Table 1, and note any deviations from the default parameters.
Additionally we analyzed each of the target fields and list the flux densities
and position of the sources in the first half of the second VLASS epoch (i.e.
epoch 2.1). In the cases of SN 2017iuk/GRB171205A and SDSS-II SN 8524, the
source is no longer detected in the second epoch, so instead we list a
$3\sigma$ upper limit. We compare the luminosities and timescales of the VLASS
detected SNe to historical radio light curves in Figure 2.
Figure 2: VLASS-SNe detections in the context of H-rich SNe (red), H-poor SNe
(blue) SNe and long GRBs (gray). A number of the VLASS observations were taken
at a later stage than SNe are typically observed and detected at radio
wavelengths, and show brighter emission than would be expected at this epoch.
Notably, the H-poor VLASS-SNe were observed years after the radio emission (at
$\gtrsim$1 GHz) from this class typically fades. Archival radio light curves
for VLASS detected SNe are included: SN 1986J ($3-5$ GHz; Bietenholz & Bartel,
2017), SN 2003bg (4.86 GHz; Soderberg et al., 2006a), SN 2004C (4.9 GHz;
DeMarchi in prep.), SN 2004dk ($3-5$ GHz; Wellons et al., 2012;
Balasubramanian et al., 2021), PTF11qcj ($3-4$ GHz; Palliyaguru et al., 2019a;
Corsi et al., 2014), SN 2012ap ($3$ GHz extrapolation based on radio SED
modeling; Chakraborti et al., 2015), SN 2012au ($3-4$ GHz; Kamble et al.,
2014a, Terreran et al. in prep.), SN 2014C (7.1 GHz; Margutti et al., 2017),
and SN 2016coi (3 GHz; Terreran et al., 2019). The archival radio observations
of SNe are from Bietenholz et al. (2021), and the archival long GRBs are from
Chandra & Frail (2012). Most archival historical light curves are at 8.6 GHz,
as the 3 GHz light curves are not well sampled. From the VLASS detected
sample, SDSS-II SN 8524 is not included since it has neither a known host
galaxy nor redshift, thus a luminosity cannot be calculated. The H-rich and
H-poor designations are inferred from the spectrum near the time of explosion.
The upper x-axis provides a reference distance scale for a fiducial normal SN
shock velocity of $0.05$c with no deceleration. This figure highlights the
presence of two groups of H-rich SNe in the radio phase space, with IIn SNe
belonging to the group with luminous radio emission years after explosion (see
e.g. Bietenholz et al., 2021).
## 3 Powering luminous late-time radio emission
Winds from massive stars enhance and shape the density of their immediate
surroundings (e.g., Smith, 2014). Radio emission from stellar explosions is
normally associated with an interaction between the fastest SN ejecta (i.e.
the blastwave) and the wind-shaped CSM. As the forward shock propagates
through the CSM, the electrons are accelerated, creating a bell-shaped non-
thermal synchrotron spectrum (Chevalier & Fransson, 2017). The radio spectrum
is characterized by a peak frequency, $\nu_{pk}$, that cascades to lower
frequencies as the blastwave expands (e.g., Chevalier 1998; Chevalier &
Fransson 2006). For synchrotron self-absorption (SSA) dominated spectra,
Chevalier (1998) suggests $\nu_{pk}\propto R^{2/7}B^{9/7}$, and the spectral
peak flux $F_{pk}\propto R^{19/7}B^{19/7}$, where $R$ is the forward shock
radius and $B$ is the post-shock magnetic field. The optically thin flux
density at $\nu>\nu_{pk}$ scales as $F_{\nu,thin}{\propto}\,\nu^{-(p-1)/2}$
(where $p$ is the power-law index of the electron distribution,
$N_{e}(\gamma_{e}){\propto}\,\gamma_{e}^{-p}$, and $\gamma_{e}$ is the
electron Lorentz factor) and the optically thick spectrum at $\nu<\nu_{pk}$ is
described as $F_{\nu,thick}\propto\nu^{5/2}$.
During the SN interaction phase with a “wind density profile” environment
expected around massive stars ($\rho_{\rm{CSM}}{\propto}r^{-2}$), the self-
similar solutions by Chevalier (1982) apply and the shock radius evolves with
time as $R\propto t^{\frac{n-3}{n-2}}$, where the density in the outer layers
of the stellar progenitor has been parametrized as $\rho_{\rm SN}\propto
r^{-n}$. In the limit of no evolution of the shock microphysical parameters
(e.g., the fraction of post-shock energy in magnetic fields and relativistic
electrons, $\epsilon_{B}$ and $\epsilon_{e}$), and using $n\approx 10$ as
appropriate for compact massive stars (Chevalier & Fransson, 2006; Matzner &
McKee, 1999), the equations above imply $\nu_{pk}\approx t^{-1}$ and
$F_{pk}\approx$ constant. Since radio SNe typically show $p\approx 3$ (or
$F_{\nu,thin}\propto\nu^{-1}$), $L_{\nu,pk}\leq
10^{28}\,\rm{erg\,s^{-1}Hz^{-1}}$ and $\nu_{pk}\leq 10$ GHz at ${<}0.1$ year
after explosion, the prediction of this single wind model is $v_{pk}{\ll}1$
GHz and a luminosity $<10^{27}\,\rm{erg\,s^{-1}Hz^{-1}}$ in the VLASS bandpass
at the current epoch (which corresponds to ${>}10^{3}$ days since explosion,
Figure 2). We conclude that our sample of VLASS SNe require a deviation from a
single-wind model.
In the remainder of this section, we discuss three alternative explanations:
(i) interaction of the SN shock with dense shells of CSM (Section 3.1); (ii)
emission from an off-axis relativistic jet entering our line of sight (Section
3.2); (iii) emergence of emission from a PWN (Section 3.3).
### 3.1 Dense detached CSM shells in the local SN environment
VLASS SNe show a level of radio emission comparable to the most luminous type
IIn SNe (Fig. 2). We place the VLASS-SNe into the phase space of radio
observables $\nu_{pk}$, $L_{\nu,pk}$ and peak time $t_{pk}$ in Figure 3, where
we calculated lines of constant shock velocity $v_{sh}$ and mass-loss
$\dot{M}$ rate following the standard formulation of SSA radio emission from a
blast wave during the interaction phase with a wind-like environment (e.g.,
Chevalier, 1998; Chevalier & Fransson, 2006; Soderberg et al., 2005, 2012).
Equipartition of energy between the relativistic electrons, protons and
magnetic field, i.e. $\epsilon_{e}=\epsilon_{B}=1/3$, where $\epsilon_{e}$ is
the fraction of thermal energy stored in electrons, and $\epsilon_{B}$ is the
fraction of magnetic energy relative to the thermal energy leads to a solid
lower limit on the mass-loss rate parameter $\dot{M}$ for a given wind
velocity ($v_{w}$), where $\rho_{CSM}=\frac{\dot{M}}{4\pi v_{w}r^{2}}$. We
present our results for both $\epsilon_{e}=\epsilon_{B}=1/3$ and for
$\epsilon_{e}=0.1$ and $\epsilon_{B}=0.01$ Our discussion below focuses on our
fiducial case of $\epsilon_{e}=0.1$ and $\epsilon_{B}=0.01$. All $\dot{M}$
values quoted are for a wind velocity $v_{w}=10^{3}\,\rm{km\,s^{-1}}$. A few
considerations follow from Figure 3:
Figure 3: Location of the VLASS SNe (filled red squares) in the phase space
of radio observables. A blue outline marks the VLASS SNe with an H-poor
spectrum at the time of explosion. Black filled circles: GRB-SNe. Grey filled
circles: H-stripped SNe from radio observations typically acquired at
$\lesssim 100$ days since explosion. We assume $p=3$ and the shock
microphysics indicated in the title of each plot. Black dashed lines: lines of
constant shock velocity assuming SSA only. Orange dashed lines: lines of
constant mass-loss rate, here calculated for an assumed wind velocity of
$10^{3}$ km/s. The red filled squares show what the properties of the VLASS
SNe would be in the case that the emission peaked in the VLASS band at the
time of the observations (see Section 3.1). Red open squares: location of the
VLASS SNe for an optically thin spectrum $L_{\nu}\propto\nu^{-1}$, assuming
that the $\nu_{pk}$ of the SSA spectrum is below the VLASS band at $\approx
0.3$ GHz. The VLASS object that crosses the $v=c$ line is SN
2017iuk/GRB171205A. VLASS Memo 13 report an ${\sim}10\%$ overestimate in flux
densities from the VLASS epoch 1 QuickLook data. No appreciable difference is
found in this analysis when applying a 10% correction to these figures.
References: Soderberg et al. (2012) and references therein.
* •
If $\nu_{pk}\gtrsim\nu_{obs}$ (where $\nu_{obs}$ is the frequency of the VLASS
observations), then VLASS-SNe require very dense environments with an
effective $\dot{M}\gtrsim 0.1\,\rm{M_{\odot}year^{-1}}$, which is
significantly larger than the typical $\dot{M}$ inferred for non-type-IIn SN
progenitors that comprise the majority of our sample (Smith 2014). In absolute
terms, the inferred $\dot{M}$ would compete with the most extreme mass-loss
rates invoked for evolved massive stars.
* •
A lower $\nu_{pk}<\nu_{obs}$ would bring the VLASS-SNe in line with the lower
$\dot{M}\sim 10^{-4}-10^{-3}\,\rm{M_{\odot}year^{-1}}$ that are typical of
massive stars. The empty squares of Figure 3 show the location of VLASS-SNe
for $\nu_{pk}=0.3$ GHz as an example. However, the lower $\nu_{pk}$ would also
lead to shock velocities $v_{sh}$ $\geq 0.1$c and
$L_{\nu,pk}>10^{28}\,\rm{erg\,s^{-1}Hz^{-1}}$, implying that VLASS-SNe would
constitute a class of radio SNe as luminous as long GRBs and with mildly
relativistic shocks at ${>}10^{3}$ days (and likely faster at earlier times).
Since earlier-time radio follow up of some VLASS-SNe indicated “normal” SN
shock speeds of $\sim 0.1-0.2c$ at a few months post-explosion (e.g., SN
2012au in Kamble et al. 2014b), it is clear that the relativistic ejecta
scenario cannot explain the entire VLASS-SNe sample unless the relativistic
ejecta is highly collimated (i.e. a jet) and pointing away from our line of
sight at early times (i.e. off-axis). We further explore the relativistic
ejecta scenario in Section 3.2.
Mass-loss rates $\dot{M}{\gtrsim}0.1\,\rm{M_{\odot}year^{-1}}$, _if_ sustained
until the time of explosion, would lead to very prominent type-IIn like
spectroscopic features at earlier times for all the VLASS-SNe, which were not
observed for the majority of the sample. Earlier radio observations of some
targets also pointed to significantly lower $\dot{M}\approx
10^{-5}\,\rm{M_{\odot}year^{-1}}$ (e.g., SNe 2004dk, 2012au and 2012ap;
Wellons et al., 2012; Kamble et al., 2014b; Chakraborti et al., 2015) at the
smaller radii probed at those epochs $r{\lesssim}5\times 10^{16}\,\rm{cm}$.
The emerging picture is that at least some VLASS-SNe exploded in a low-density
bubble surrounded by a shell of dense material at $r{\sim}v_{sh}\delta
t=(v_{sh}/10^{4}\,\rm{km\,s^{-1}})(\delta t/8000\,\rm{days}){\approx}0.5$ pc,
consistent with the findings from the multi-wavelength monitoring of SNe
2003bg, 2004C, 2004dk, 2014C, and PTF11qcj (Soderberg et al. 2006b, Margutti
et al. 2017, Pooley et al. 2019, Corsi et al. 2014, Palliyaguru et al. 2019b,
Murase et al. 2019, Brethauer et al. 2020, Balasubramanian et al. 2021,
DeMarchi in prep.). For VLASS-SNe from H-poor stellar progenitors (which
interestingly dominate the sample), these overdensities might represent the
shedding of their H-rich envelope in the centuries before core-collapse.
Optical spectroscopy at the time of the radio re-brightenings of SN 2003bg, SN
2004dk, SN 2014C and PTF11qcj confirmed the later appearance of H features in
the spectra (Soderberg et al., 2006b; Pooley et al., 2019; Milisavljevic et
al., 2015b; Palliyaguru et al., 2019b), consistent with this scenario.
Potential theoretical explanations of this phenomenology include the
interaction of faster Wolf-Rayet winds with the slower winds of the red
supergiant phase coupled with a shorter-than expected Wolf-Rayet phase;
envelope ejection due to binary interaction; or mass shedding due to gravity-
wave powered mass loss (e.g., Smith 2014, Zhao & Fuller 2020, Wu & Fuller
2021).
### 3.2 Off-axis relativistic jets
Figure 4: VLASS-SNe 3 GHz light curves for SN 2005ha, SN 2012ap, and SDSS-II
SN 12882. The 3 GHz SN 2012ap archival light curve is included using the model
from Chakraborti et al. (2015). The SN 2012ap light curve is shown along with
the 10 top-hat jet models in our grid that best fit the VLASS only light curve
(gray dashed lines) and the 10 models that best fit the combined VLASS and
archival light curve (gray dotted lines). SN 2005ha and SDSS-II SN 12882 are
the only multi-epoch detected VLASS-SNe that are consistent with off-axis jet
models from our grid (i.e. $\chi^{2}<1$). The solid black lines represent the
off-axis jet models that are consistent with SN 2005ha and SDSS-II SN 12882
light curves. SN 2005ha and SDSS-II SN 12882 had limited spectroscopic follow-
up and were not classified as H-rich or H-poor (e.g., Marsden 2005). We show
this figure as a proof of concept, but we note that the models that best fit
the SN 2005ha and SDSS-II SN 12882 light curves have high isotropic kinetic
energies of $E_{iso}=10^{55}$ erg, corresponding to beaming-corrected energies
of $E=3\times 10^{53}$ and $1\times 10^{54}$ erg for SN 2005ha and SDSS-II SN
12882, respectively, and are likely unrealistic.
Off-axis jets can result in bright synchrotron emission that peaks years after
explosion (e.g., Granot et al., 2002, 2018). The emission from off-axis jets
enters our line of sight as the jet decelerates in the ambient medium and
relativistic beaming becomes less severe (Rhoads, 1997; Sari et al., 1999).
The fraction of stellar explosions that are jet driven is still unclear (e.g.,
Corsi & Lazzati 2021; Corsi et al. 2016; Bietenholz et al. 2014; Soderberg et
al. 2006c). Successful relativistic jet have been so far associated with
broad-line type-Ic SNe accompanying cosmological GRBs, while partially
successful and partially failed jets have been proposed to be powering low-
luminosity GRBs and relativistic SNe, respectively (e.g., for observations see
Margutti et al. 2014a, and for theory see Morsony et al. 2007; Lazzati et al.
2012). While observations of energetic H-stripped SNe point at a continuum of
jet properties from normal Ibc SNe to GRB/SNe (e.g., Xu & Wei, 2008; Mazzali
et al., 2008; Margutti et al., 2014b; Corsi & Lazzati, 2021), no bona fide
off-axis jet has ever been associated with a SN without a GRB detection. In
this context it is particularly interesting to note that the relativistic SN
2012ap, which is one of only two known relativistic SNe without a GRB
counterpart (Chakraborti et al., 2015; Soderberg et al., 2010), is detected by
VLASS observations 5.7 and 8.5 years after explosion. Thus SN 2012ap is a
clear candidate for an off-axis jet driving late-time emission.
To determine whether the detected VLASS emission is associated with off-axis
jets, we generated a set of synthetic 3 GHz jet afterglow light-curves with
Boxfit v2 (van Eerten et al., 2012). Boxfit assumes a top-hat jet structure,
i.e. a jet with energy uniformly distributed within $\theta\leq\theta_{jet}$
and $E=0$ for $\theta>\theta_{jet}$. We assumed a wind-like CSM density
profile. We explored the parameter space with a grid of parameter values
defined as follows: isotropic-equivalent jet kinetic energies of
$E_{iso}=[10^{50},10^{51},10^{52},10^{53},10^{54},10^{55}]\,\rm{erg}$; jet
opening angles of $\theta_{jet}=[5^{\circ},15^{\circ},30^{\circ}]$; off-axis
angle $\theta_{obs}=[30^{\circ},60^{\circ},90^{\circ}]$ from our line of
sight; mass-loss rates of
$\dot{M}=[10^{-8},10^{-7},10^{-6},10^{-5},10^{-4}10^{-3}]$ $M_{\odot}$ year-1
for $v_{w}=1000\,\rm{km\,s^{-1}}$; shock micro-physical parameters
$\epsilon_{e}=0.1$, $\epsilon_{B}=[0.001,0.01]$, $p=[2,2.5,3]$. Finally we
used the $\chi^{2}$ as a metrics to evaluate the agreement between the models
and the VLASS data of Figure 2.
None of the VLASS-SNe with archival (i.e. pre-VLASS) data points are
consistent with our model off-axis jet light curves. We find that the
synthetic models that best approximate the VLASS data of SDSS-II SN 12882, SN
2002hi, SN 2005ha, SN 2009fi, and SN 2012cc with $\chi^{2}\lesssim 1$ have
large off-axis angles $\theta_{obs}\geq 60^{\circ}$, large densities
corresponding to $\dot{M}>10^{-5}$ $M_{\odot}$ year-1, $p\sim 3$,
$\epsilon_{B}=0.01$ coupled with large $E_{iso}\geq 10^{54}\,\rm{erg}$ and
large jet angle $\theta_{jet}\geq 15^{\circ}$. The values of these parameters
is driven by the large radio luminosities of the sample at late times, and
imply extremely large beaming-corrected jet energies $3\times
10^{52}-10^{54}\,\rm{erg}$. While we show some examples of top-hat off-axis
jet light-curve consistent with the VLASS data in Figure 4, we consider this
top-hat jet scenario unlikely because of the large jet energies needed and the
fact that only SNe with a sparse data set can be fitted. We consider
alternative jet models and environments below.
We start by noting that in SN 1965G, SN 2004C, SN 2005ha, SN 2012ap, SN
2012at, and SDSS-II SN 12882, the radio flux density remains nearly constant
over ${\sim}2-3$ years between two VLASS epochs. SN 1986J also has nearly
constant radio flux densities (see Figure 2); however, it has only been
observed in a single VLASS epoch. Numerical simulations of GRB jets
propagating through a stratified media show that nearly flat, wide peaks are
obtained only if the jet propagates through a wind-profile medium with
$\rho\propto r^{-2}$ (see Figure 1 in Granot et al., 2018). Jets propagating
through a uniform density environment have a much more narrow peak (e.g., as
seen in the GRB 170817A afterglow Margutti & Chornock 2020) and are ruled out
by our observations. GRB 170817A also clearly showed that relativistic jets
can have angular structure (i.e. the jet is not necessarily top-hat; see e.g.,
Nakar 2020 and references therein).
The propagation of relativistic GRB jets through a massive Wolf-Rayet
progenitor star leads to the production of extended wide-angle outflows known
as cocoons, with masses ${\approx}10^{-2}-10^{-1}$ M⊙ and energies
${\approx}10^{50}-10^{51}$ ergs (see, e.g., Lazzati & Begelman, 2005; Nakar &
Piran, 2017; De Colle et al., 2021) possibly observationally identified in SN
2017iuk/GRB171205A (Izzo et al., 2019). Once the jet breaks out of the star,
the cocoon engulfs the star and expands nearly spherically into the
environment (see e.g. Figure 3 of De Colle et al. 2021). The cocoon initially
expands with relativistic velocities (corresponding to Lorentz factors
${\sim}2-10$), but later decelerates to mildly relativistic velocities at
${\sim}10^{16}$ cm (see Figure 2 in De Colle et al., 2018). Particle
acceleration through the shock cocoon itself will lead to a bright afterglow.
While GRB jets are collimated and enter into the observer line of sight only
at late times, the cocoon radio emission should be detectable at early times
by observers located at nearly all angles (as beaming effects are much less
important in the slower moving cocoon material) and would thus be able to
explain the larger radio fluxes of pre-VLASS observations. The predicted
early-time radio emission (De Colle et al., 2018, 2021) is similar to that
observed in relativistic SNe 2009bb and 2012ap (Soderberg et al., 2010;
Margutti et al., 2014a; Chakraborti et al., 2015). Several SNe in our sample,
including SN 2012ap, have been observed at early times (${\sim}$days to a
month after explosion), but only SN 2012ap showed mildly relativistic material
consistent with the expectations from the cocoon model. SN 2012ap is the only
VLASS-SN for which a cocoon and off-axis relativistic jet is a viable
explanation. The largely uncollimated, mildly relativistic cocoon would be
responsible for the early emission. The late-time VLASS emission would be
powered by the off-axis relativistic jet. In this case, SN 2012ap would
represent the first evidence of a cocoon and jet system from a massive stellar
explosion. Future multifrequency observations will test this scenario. Thus,
with the exception of SN 2012ap, we find that the late-time VLASS emission is
unlikely to be caused by relativistic jets.
### 3.3 Emergence of emission from a pulsar wind nebula
Another candidate for the cause of late-time radio emission from SNe is the
presence of a PWN (e.g., Gaensler & Slane 2006; Slane 2017). Core-collapse
SNe, which comprise the totality of our sample, are expected to leave a
compact remnant. If a fast rotating NS is left behind, it can feed a steady
highly-energetic wind of relativistic particles into the SN ejecta, and this
“bubble” of relativistic particles is referred to as a PWN. As this wind
interacts with the slower SN ejecta, a termination shock forms and high-energy
photon emission heats and ionizes the surrounding SN ejecta. Shortly after the
explosion, the emission is absorbed by the dense ejecta (e.g., Metzger et al.,
2014; Murase et al., 2015, 2016, 2021). Over time, as the ejecta expands and
the optical depth decreases, the PWN emission becomes observable. No SN has
unambiguously shown the transition from ejecta-dominated emission to PWN-
dominated emission. Recently, there have been hints towards the detection of a
PWN associated with SN 1987A. This suggestion is due to non-thermal emission
in the hard X-rays (Greco et al., 2021), and from the radio detection of a
warm dust concentration at the center of the remnant (Cigan et al., 2019);
however, alternative mechanisms to explain the emission cannot be ruled out.
Beyond SN 1987A, two young SNe have been suggested to harbor PWNe (SN 1986J
and SN 2012au), and both are in our VLASS-SNe sample.
The presence of a PWN energizing the ejecta in a young SN has been proposed to
explain the anomalous state of high ionization inferred from optical
spectroscopy of the H-stripped energetic SN 2012au ${\approx}6$ years after
explosion by Milisavljevic et al. (2018). The spectra of this transient
acquired $\approx$7 years after explosion were dominated by forbidden oxygen
lines with velocities of ${\approx}2300$ km s-1. Oxygen resides in the inner
part of the SN ejecta, thus one explanation for this emission is the presence
of a pulsar that ionizes the internal material (Milisavljevic et al., 2018).
The lack of narrow hydrogen in the early spectra of SN 2012au suggests a
different powering mechanism than CSM-ejecta interaction, and supports the
scenario of ionization by a pulsar as the origin of the emission.
Bietenholz et al. (2002) suggested the late-time radio emission from SN 1986J
is evidence of a PWN. SN 1986J showed a broad radio SED $7-16$ years after
explosion with a spectrum at $\nu>10$ GHz which evolved from thin to thick
(i.e. an inverted radio spectrum). However, observed SEDs of evolved PWNe are
relatively flat, with typical spectral indices between $-0.3$ and $0.0$. In
contrast, SN 1986J has an SED that peaked at ${\approx}20$ GHz, with an
absorbed optically thick region, and an optically thin spectral index of
$-0.76$. SN 2012au has a similarly shaped SED at $8$ years post-explosion
(Terreran et al. in prep.). The bell-like synchrotron SEDs produced by CSM-
interaction of the SN shock wave peak below GHz frequencies on these time-
scales. The observed radio spectrum of SN 1986J and SN 2012au is also unusual
for evolved PWNe, but we emphasize that the spectral properties of nascent
PWNe that are a few years old are not observationally well constrained. From a
theoretical perspective, we expect the radiative electrons to be in the “fast-
cooling” regime, which can lead to radio spectra similar to those observed
(e.g., Murase et al., 2016; Omand et al., 2018; Murase et al., 2021). A young
PWN is expected to be smaller in size than the SN ejecta, thus one can
distinguish between the shock interaction and a PWN with very long baseline
interferometry.
Interestingly, both SN 1986J and SN 2012au are in the VLASS-SNe sample. If the
bright radio emission is confirmed to be powered by PWNe, the associated PWNe
would be the two youngest discovered to date. No forming PWN has been observed
in the Milky Way or the Magellanic Clouds. The youngest known galactic PWN,
Kes 75, has an estimated age of $480\pm 40$ years (Reynolds et al., 2018), and
thus little is known about PWN properties in the years to decades after the SN
explosion.
## 4 Conclusions
We present evidence for a population of 19 radio luminous SNe ($L_{\nu}\sim
10^{26}-10^{29}$ erg s-1 Hz-1 at 3 GHz) ${\approx}1-60$ years after explosion
found in the first epoch of the VLASS. This is part of the radio phase space
of stellar explosions that has not been systematically explored so far. Our
filtering procedure leveraged multiwavelength catalogs to remove potential AGN
contaminants, and other known radio sources leading to a sample that is
entirely comprised of core-collapse SNe and surprisingly dominated by stellar
explosions with hydrogen stripped progenitors at the time of collapse. Our
main result is that the large radio luminosities at these late stages of
evolution require deviation from the traditional single wind mass-loss
scenario and/or spherical shock assumption. Potential alternatives include the
following:
1. 1.
Initial expansion of the SN shock into a lower-density bubble, followed by
strong shock interaction with a sharp density increase (i.e. a “bubble” plus
detached shell CSM structure). This dense shell might be connected to the
shedding of the H-rich stellar envelope in the centuries before core-collapse
through mass-loss mechanisms that have yet to be observationally identified.
VLASS SNe are as luminous as the most luminous radio SNe IIn few yrs post
explosion, which indicates CSM densities at large radii from the progenitors
that are comparable to those inferred for SNe IIn.
2. 2.
While top-hat relativistic jets viewed off-axis are unlikely to provide an
adequate explanation due to the under-prediction of the pre-VLASS radio
observations of most elements of the sample, relativistic jets with structure
are not ruled out. SN 2012ap, which showed evidence for an uncollimated mildly
relativistic outflow at $\delta t<40$ days, is the primary candidate for being
the first jet+cocoon system in a massive star observed off-axis, which may
signal that relativistic SNe are “cocoons” observed early on.
3. 3.
The final alternative is the emergence of a PWN. The VLASS-SNe sample includes
SN 1986J and SN 2012au, the two young SNe that have previously been suggested
to have PWNe powered late-time radio emission.
The VLA Sky Survey provides an unprecedented and unbiased window into the
variable radio sky, combining the large survey area of the Northern VLA Sky
Survey with the depth and angular resolution of the Faint Images of the Radio
Sky survey. These features, and the planned multiple field visits, are
particularly useful for the discovery and study of extragalaxtic transients,
where the angular resolution (and higher frequency, 3 GHz vs. 1 GHz) minimizes
confusion by the host galaxies of transients of interest. Planned
interferometers such as the Next Generation VLA (Carilli et al., 2015) and the
Square Kilometer Array (Dewdney et al., 2009) will expand our ability to study
the variable radio sky with increased depth. The VLASS is complimented by
other surveys and serendipitous transient discovery programs being carried out
with SKA pathfinder instruments such as ASKAP (VAST; Murphy et al. 2013),
Westerbork (Apertif; Adams & van Leeuwen 2019), MeerKAT (ThunderKAT; Fender et
al. 2016), LOFAR (LoTSS; Shimwell et al. 2017), which encompass a range of
frequencies and angular resolutions while providing access to the southern
sky.
Follow-up with multiwavelength observations including radio spectral energy
distributions, and optical spectroscopy will help constrain the mechanisms
responsible for the bright radio emission of our VLASS-SN sample. We will
present the multiwavelength follow-up of the VLASS detected SNe sample in
future papers.
We thank the referee for providing constructive comments. We also thank Seth
Bruzewski for providing astrometric corrections to the VLASS Quicklook epoch 1
data. This work is supported by the Heising-Simons Foundation under grant
#2018-0911 (PI: Margutti). R.M. acknowledges support by the NSF under grants
AST-1909796 and AST-1944985. F.D.C. acknowledges support from the UNAM-PAPIIT
grant AG100820. R.B.D. acknowledges support from National Science Foundation
(NSF) under grant 1816694. The National Radio Astronomy Observatory is a
facility of the National Science Foundation operated under cooperative
agreement by Associated Universities, Inc. This publication makes use of data
products from the Wide-field Infrared Survey Explorer, which is a joint
project of the University of California, Los Angeles, and the Jet Propulsion
Laboratory/California Institute of Technology, and NEOWISE, which is a project
of the Jet Propulsion Laboratory/California Institute of Technology. WISE and
NEOWISE are funded by the National Aeronautics and Space Administration. This
research has made use of data obtained from the Chandra Source Catalog,
provided by the Chandra X-ray Center (CXC) as part of the Chandra Data
Archive. This research has made use of data obtained from the 4XMM XMM-Newton
serendipitous source catalogue compiled by the 10 institutes of the XMM-Newton
Survey Science Centre selected by ESA. This work made use of data supplied by
the UK Swift Science Data Centre at the University of Leicester. This research
was supported in part through the computational resources and staff
contributions provided for the Quest high performance computing facility at
Northwestern University which is jointly supported by the Office of the
Provost, the Office for Research, and Northwestern University Information
Technology. Development of the BOXFIT code was supported in part by NASA
through grant NNX10AF62G issued through the Astrophysics Theory Program and by
the NSF through grant AST-1009863. This research made use of APLpy, an open-
source plotting package for Python (Robitaille & Bressert, 2012; Robitaille,
2019).
## Appendix A VLASS detected SNe
We present in Table 1 the sample of SNe detected in the VLASS data set. The
second column lists whether the progenitor was hydrogen rich or hydrogen poor
at the time of explosion. Flux densities and positions were derived as
described in Section 2.5, with any deviations from the default procedure given
in the ‘Notes’ column. The list of VLASS SNe is ordered by increasing right
ascension. Bruzewski et al. (2021) calculated the astrometric corrections
required to align the VLASS Quicklook epoch 1 data with the Gaia catalog, and
we list the coordinates that include these astrometric corrections. The
positional errors include the uncertainties from PyBDSF fits and the
astrometric corrections (Bruzewski private communication) added in quadrature.
We applied the Bruzewski et al. (2021) derived corrections and astrometric
uncertainties for the VLASS epoch 2.1 observations. We note that applying the
VLASS epoch 1 uncertainties may overstate the positional uncertainties in the
second epoch observations, since the second VLASS epoch will have likely
benefited from studying the systematic uncertainties in the first epoch (see
e.g. VLASS Memo 13555The VLASS Project Memo Series is listed at
https://go.nrao.edu/vlass-memos.).
We report the PyBDSF flux density errors which are purely statistical. There
are known flux density offsets in the VLASS Quicklook images, as detailed in
VLASS Memo 13 and the CIRADA Catalogue User Guide. The detection type defines
the nature of the source structure, where “S” indicates a single Gaussian that
is the only source in the island, “C” indicates a single source in an island
with other sources, and “M” indicates multiple Gaussian source. The angular
separation lists the distance between the listed VLASS position and the
optical position. For the SN classifications, “Pec” and “BL” stand for
peculiar and broad-lined, respectively. Four SNe (20% of the sample) had
limited follow-up leaving the classification unknown, but they are believed to
be core-collapse SNe (i.e. SN 1965G, SN 2005ha, SDSS-II SN 8524, and SDSS-II
SN 12882).
Table 1: Supernovae detected in VLASS epoch 1
| | VLASS | | Angular |
---|---|---|---|---|---
Name | Progenitor | R.A. | Decl. | Flux Density | Detection | Obs. Date | Luminosity | Separation | Classification
| [H-rich/poor] | [hh:mm:ss.ss] | [dd:mm:ss.ss] | [mJy] | [S, M or C] | [MJD] | [erg/s/Hz] | [′′] |
SN 1986J | ? | 02:22:31.293(15) | $+$42:19:57.5(3) | $1.3\pm 0.2$ | C | 58588 | $(1.6\pm 0.2)\times 10^{26}$ | $0.56$ | IIn
SN 2017hcb | H-poor | 02:36:23.756(14) | $+$31:42:36.2(3) | $0.5\pm 0.2$ | M | 58569 | $(3.0\pm 1.2)\times 10^{27}$ | $0.99$ | Ib
SDSS-II SN 12882 | ? | 03:03:49.977(9) | $-$00:12:14.3(3) | $1.7\pm 0.4$ a | S | 58103 | $(2.8\pm 0.7)\times 10^{28}$ | $0.32$ | ?
| | 03:03:49.975(10) | $-$00:12:14.2(3) | $1.6\pm 0.3$ a | S | 59078 | $(2.6\pm 0.5)\times 10^{28}$ | $0.41$ |
SN 2003bg | H-poor | 04:10:59.436(6) | $-$31:24:50.2(3) | $4.1\pm 0.2$ | S | 58663 | $(2.83\pm 0.14)\times 10^{27}$ | $0.61$ | IcPecBL
SN 2012at | H-poor | 04:54:52.783(7) | $-$37:19:16.9(3) | $2.0\pm 0.2$ | S | 58153 | $(2.5\pm 0.2)\times 10^{27}$ | $0.41$ | Ic
| | 04:54:52.786(8) | $-$37:19:17.4(3) | $1.9\pm 0.2$ | S | 59155 | $(2.3\pm 0.2)\times 10^{27}$ | $0.70$ |
SN 2012ap | H-poor | 05:00:13.734(5) | $-$03:20:51.4(3) | $4.0\pm 0.3$ | S | 58027 | $(8.9\pm 0.7)\times 10^{27}$ | $0.25$ | IcBL
| | 05:00:13.738(5) | $-$03:20:51.6(3) | $4.5\pm 0.3$ | S | 59078 | $(1.00\pm 0.07)\times 10^{28}$ | $0.23$ |
SN 2005ha | ? | 06:21:49.110(6) | $+$00:21:56.2(3) | $2.2\pm 0.2$ | S | 58123 | $(3.7\pm 0.3)\times 10^{27}$ | $1.23$ | ?
| | 06:21:49.106(6) | $+$00:21:56.0(3) | $1.9\pm 0.3$ | S | 59048 | $(3.2\pm 0.5)\times 10^{27}$ | $1.09$ |
SN 2002hi | H-rich | 07:19:54.127(9) | $+$17:58:18.5(3) | $2.1\pm 0.3$ | S | 58572 | $(1.7\pm 0.2)\times 10^{29}$ | $0.72$ | IIn
SN 2017iuk | H-poor | 11:09:39.519(5) | $-$12:35:18.5(3) | $4.8\pm 0.2$ | S | 58150 | $(1.39\pm 0.06)\times 10^{29}$ | $0.24$ | IcBL
| | $\cdots$ | $\cdots$ | $<0.45$ | $\cdots$ | 59133 | $<1.3\times 10^{28}$ | $\cdots$ |
SN 2004C | H-poor | 11:27:29.80(2) | $+$56:52:47.9(3) | $4.2\pm 0.6$ b | S | 58020 | $(2.8\pm 0.4)\times 10^{27}$ | $0.57$ | Ic
| | 11:27:29.77(2) | $+$56:52:47.9(3) | $5.3\pm 0.9$ | S | 59064 | $(3.5\pm 0.6)\times 10^{27}$ | $0.79$ |
SN 1965G | ? | 12:11:54.049(5) | $+$24:06:58.5(3) | $7.7\pm 0.4$ | S | 58082 | $(1.15\pm 0.06)\times 10^{28}$ | $2.59$ | ?
| | 12:11:54.045(5) | $+$24:06:58.4(3) | $7.8\pm 0.3$ | S | 59099 | $(1.17\pm 0.05)\times 10^{28}$ | $2.52$ |
SN 2012cc | H-rich | 12:26:56.829(9) | $+$15:02:45.6(3) | $2.3\pm 0.4$ | S | 58590 | $(1.10\pm 0.19)\times 10^{27}$ | $0.36$ | II
SN 2012au | H-poor | 12:54:52.257(5) | $-$10:14:50.5(3) | $4.5\pm 0.3$ | S | 58553 | $(3.0\pm 0.2)\times 10^{27}$ | $1.16$ | Ib
PTF11qcj | H-poor | 13:13:41.480(9) | $+$47:17:56.8(3) | $6.8\pm 0.2$ | S | 58561 | $(1.12\pm 0.03)\times 10^{29}$ | $0.44$ | IcBL
SN 2009fi | H-rich | 14:06:05.757(6) | $+$11:47:13.6(3) | $2.2\pm 0.2$ | S | 58611 | $(1.17\pm 0.11)\times 10^{28}$ | $0.88$ | IIb
SN 2004dk | H-poor | 16:21:48.872(4) | $-$02:16:17.6(3) | $6.3\pm 0.2$ | S | 58624 | $(3.34\pm 0.11)\times 10^{27}$ | $0.75$ | Ib
SDSS-II SN 8524 | ? | 21:29:23.354(6) | $+$00:56:42.9(3) | $1.7\pm 0.2$ | S | 58023 | $\cdots$ c | $1.18$ | ?
| | $\cdots$ | $\cdots$ | $<0.5$ | $\cdots$ | 59049 | $\cdots$ c | $\cdots$ |
SN 2016coi | H-poor | 21:59:04.127(8) | $+$18:11:10.8(3) | $1.8\pm 0.2$ | S | 58604 | $(4.7\pm 0.5)\times 10^{26}$ | $0.35$ | IcBL
SN 2014C | H-poor | 22:37:05.601(6) | $+$34:24:31.5(3) | $29.0\pm 0.3$ | S | 58642 | $(7.91\pm 0.08)\times 10^{27}$ | $0.49$ | Ibd
Note. — a) The region surrounding SDSS-II SN 12882 is contaminated by radial
artifacts from quasar PB 6989, so flux density could be less reliable.
b) The region surrounding SN 2004C is contaminated by radial artifacts from
NVSS J112731$+$565240, so the flux density could be less reliable. SN 2004C is
clearly part of an extended emission complex, which is not detected with
PyBDSF unless the island and detection threshold are both lowered.
There is a clear point source at the location of SN 2004C which we associate
with the transient, but the flux density here is likely unreliable.
c) SDSS-II SN 8524 has no known redshift, thus a luminosity cannot be
calculated.
d) SN 2014C was initially classified as a type Ib but was later classified as
a type IIn.
## References
* Adams & van Leeuwen (2019) Adams, E. A. K., & van Leeuwen, J. 2019, Nature Astronomy, 3, 188, doi: 10.1038/s41550-019-0692-4
* Assef et al. (2018) Assef, R. J., Stern, D., Noirot, G., et al. 2018, ApJS, 234, 23, doi: 10.3847/1538-4365/aaa00a
* Astropy Collaboration et al. (2013) Astropy Collaboration, Robitaille, T. P., Tollerud, E. J., et al. 2013, A&A, 558, A33, doi: 10.1051/0004-6361/201322068
* Astropy Collaboration et al. (2018) Astropy Collaboration, Price-Whelan, A. M., SipHocz, B. M., et al. 2018, aj, 156, 123, doi: 10.3847/1538-3881/aabc4f
* Balasubramanian et al. (2021) Balasubramanian, A., Corsi, A., Polisensky, E., Clarke, T. E., & Kassim, N. E. 2021, arXiv e-prints, arXiv:2101.07348. https://arxiv.org/abs/2101.07348
* Becker et al. (1995) Becker, R. H., White, R. L., & Helfand, D. J. 1995, ApJ, 450, 559, doi: 10.1086/176166
* Bertin & Arnouts (1996) Bertin, E., & Arnouts, S. 1996, A&AS, 117, 393, doi: 10.1051/aas:1996164
* Bietenholz & Bartel (2017) Bietenholz, M. F., & Bartel, N. 2017, ApJ, 851, 7, doi: 10.3847/1538-4357/aa960b
* Bietenholz et al. (2021) Bietenholz, M. F., Bartel, N., Argo, M., et al. 2021, The Astrophysical Journal, 908, 75, doi: 10.3847/1538-4357/abccd9
* Bietenholz et al. (2002) Bietenholz, M. F., Bartel, N., & Rupen, M. P. 2002, ApJ, 581, 1132, doi: 10.1086/344251
* Bietenholz et al. (2014) Bietenholz, M. F., De Colle, F., Granot, J., Bartel, N., & Soderberg, A. M. 2014, MNRAS, 440, 821, doi: 10.1093/mnras/stu246
* Brethauer et al. (2020) Brethauer, D., Margutti, R., Milisavljevic, D., & Bietenholz, M. 2020, Research Notes of the American Astronomical Society, 4, 235, doi: 10.3847/2515-5172/abd252
* Bruzewski et al. (2021) Bruzewski, S., Schinzel, F. K., Taylor, G. B., & Petrov, L. 2021, ApJ, 914, 42, doi: 10.3847/1538-4357/abf73b
* Carilli et al. (2015) Carilli, C. L., McKinnon, M., Ott, J., et al. 2015, arXiv e-prints, arXiv:1510.06438. https://arxiv.org/abs/1510.06438
* Chakraborti et al. (2015) Chakraborti, S., Soderberg, A., Chomiuk, L., et al. 2015, ApJ, 805, 187, doi: 10.1088/0004-637X/805/2/187
* Chandra & Frail (2012) Chandra, P., & Frail, D. A. 2012, ApJ, 746, 156, doi: 10.1088/0004-637X/746/2/156
* Chevalier (1982) Chevalier, R. A. 1982, ApJ, 259, 302, doi: 10.1086/160167
* Chevalier (1998) —. 1998, ApJ, 499, 810, doi: 10.1086/305676
* Chevalier & Fransson (2006) Chevalier, R. A., & Fransson, C. 2006, ApJ, 651, 381, doi: 10.1086/507606
* Chevalier & Fransson (2017) Chevalier, R. A., & Fransson, C. 2017, Thermal and Nonthermal Emission from Circumstellar Interaction, ed. A. W. Alsabti & P. Murdin (Cham: Springer International Publishing), 1–63, doi: 10.1007/978-3-319-20794-0_34-1
* Chomiuk et al. (2016) Chomiuk, L., Soderberg, A. M., Chevalier, R. A., et al. 2016, ApJ, 821, 119, doi: 10.3847/0004-637X/821/2/119
* Cigan et al. (2019) Cigan, P., Matsuura, M., Gomez, H. L., et al. 2019, ApJ, 886, 51, doi: 10.3847/1538-4357/ab4b46
* Condon et al. (1998) Condon, J. J., Cotton, W. D., Greisen, E. W., et al. 1998, AJ, 115, 1693, doi: 10.1086/300337
* Corsi & Lazzati (2021) Corsi, A., & Lazzati, D. 2021, New A Rev., 92, 101614, doi: 10.1016/j.newar.2021.101614
* Corsi et al. (2014) Corsi, A., Ofek, E. O., Gal-Yam, A., et al. 2014, ApJ, 782, 42, doi: 10.1088/0004-637X/782/1/42
* Corsi et al. (2016) Corsi, A., Gal-Yam, A., Kulkarni, S. R., et al. 2016, ApJ, 830, 42, doi: 10.3847/0004-637X/830/1/42
* Cutri et al. (2021) Cutri, R. M., Wright, E. L., Conrow, T., et al. 2021, VizieR Online Data Catalog, II/328
* D’Abrusco et al. (2013) D’Abrusco, R., Massaro, F., Paggi, A., et al. 2013, ApJS, 206, 12, doi: 10.1088/0067-0049/206/2/12
* De Colle et al. (2018) De Colle, F., Kumar, P., & Aguilera-Dena, D. R. 2018, ApJ, 863, 32, doi: 10.3847/1538-4357/aad04d
* De Colle et al. (2021) De Colle, F., Kumar, P., & Hoeflich, P. 2021, arXiv e-prints, arXiv:2105.09376. https://arxiv.org/abs/2105.09376
* de Ugarte Postigo et al. (2017) de Ugarte Postigo, A., Izzo, L., & Kann, D. A. 2017, Transient Name Server Classification Report, 2017-1387, 1
* Dewdney et al. (2009) Dewdney, P. E., Hall, P. J., Schilizzi, R. T., & Lazio, T. J. L. W. 2009, IEEE Proceedings, 97, 1482, doi: 10.1109/JPROC.2009.2021005
* Evans et al. (2020a) Evans, I. N., Primini, F. A., Miller, J. B., et al. 2020a, in American Astronomical Society Meeting Abstracts, Vol. 235, American Astronomical Society Meeting Abstracts #235, 154.05
* Evans et al. (2020b) Evans, P. A., Page, K. L., Osborne, J. P., et al. 2020b, ApJS, 247, 54, doi: 10.3847/1538-4365/ab7db9
* Fender et al. (2016) Fender, R., Woudt, P. A., Corbel, S., et al. 2016, in MeerKAT Science: On the Pathway to the SKA, 13. https://arxiv.org/abs/1711.04132
* Gaensler & Slane (2006) Gaensler, B. M., & Slane, P. O. 2006, ARA&A, 44, 17, doi: 10.1146/annurev.astro.44.051905.092528
* Gordon et al. (2020) Gordon, Y. A., Boyce, M. M., O’Dea, C. P., et al. 2020, Research Notes of the American Astronomical Society, 4, 175, doi: 10.3847/2515-5172/abbe23
* Granot et al. (2018) Granot, J., De Colle, F., & Ramirez-Ruiz, E. 2018, MNRAS, 481, 2711, doi: 10.1093/mnras/sty2454
* Granot et al. (2002) Granot, J., Panaitescu, A., Kumar, P., & Woosley, S. E. 2002, ApJ, 570, L61, doi: 10.1086/340991
* Greco et al. (2021) Greco, E., Miceli, M., Orlando, S., et al. 2021, ApJ, 908, L45, doi: 10.3847/2041-8213/abdf5a
* Hallinan et al. (2020) Hallinan, G., Mooley, K., Dong, D., et al. 2020, The Astronomer’s Telegram, 14020, 1
* Izzo et al. (2019) Izzo, L., de Ugarte Postigo, A., Maeda, K., et al. 2019, Nature, 565, 324, doi: 10.1038/s41586-018-0826-3
* Kamble et al. (2014a) Kamble, A., Soderberg, A. M., Chomiuk, L., et al. 2014a, ApJ, 797, 2, doi: 10.1088/0004-637X/797/1/2
* Kamble et al. (2014b) —. 2014b, ApJ, 797, 2, doi: 10.1088/0004-637X/797/1/2
* Koposov & Bartunov (2006) Koposov, S., & Bartunov, O. 2006, in Astronomical Society of the Pacific Conference Series, Vol. 351, Astronomical Data Analysis Software and Systems XV, ed. C. Gabriel, C. Arviset, D. Ponz, & S. Enrique, 735
* Lacy et al. (2020) Lacy, M., Baum, S. A., Chandler, C. J., et al. 2020, PASP, 132, 035001, doi: 10.1088/1538-3873/ab63eb
* Lazzati & Begelman (2005) Lazzati, D., & Begelman, M. C. 2005, ApJ, 629, 903, doi: 10.1086/430877
* Lazzati et al. (2012) Lazzati, D., Morsony, B. J., Blackwell, C. H., & Begelman, M. C. 2012, ApJ, 750, 68, doi: 10.1088/0004-637X/750/1/68
* Margutti & Chornock (2020) Margutti, R., & Chornock, R. 2020, arXiv e-prints, arXiv:2012.04810. https://arxiv.org/abs/2012.04810
* Margutti et al. (2014a) Margutti, R., Milisavljevic, D., Soderberg, A. M., et al. 2014a, ApJ, 797, 107, doi: 10.1088/0004-637X/797/2/107
* Margutti et al. (2014b) —. 2014b, ApJ, 780, 21, doi: 10.1088/0004-637X/780/1/21
* Margutti et al. (2017) Margutti, R., Kamble, A., Milisavljevic, D., et al. 2017, ApJ, 835, 140, doi: 10.3847/1538-4357/835/2/140
* Marsden (2005) Marsden, B. G. 2005, Central Bureau Electronic Telegrams, 260, 4
* Matzner & McKee (1999) Matzner, C. D., & McKee, C. F. 1999, ApJ, 510, 379, doi: 10.1086/306571
* Mazzali et al. (2008) Mazzali, P. A., Valenti, S., Della Valle, M., et al. 2008, Science, 321, 1185, doi: 10.1126/science.1158088
* Metzger et al. (2014) Metzger, B. D., Vurm, I., Hascoët, R., & Beloborodov, A. M. 2014, MNRAS, 437, 703, doi: 10.1093/mnras/stt1922
* Milisavljevic et al. (2018) Milisavljevic, D., Patnaude, D. J., Chevalier, R. A., et al. 2018, ApJ, 864, L36, doi: 10.3847/2041-8213/aadd4e
* Milisavljevic et al. (2014) Milisavljevic, D., Margutti, R., Crabtree, K. N., et al. 2014, ApJ, 782, L5, doi: 10.1088/2041-8205/782/1/L5
* Milisavljevic et al. (2015a) Milisavljevic, D., Margutti, R., Parrent, J. T., et al. 2015a, ApJ, 799, 51, doi: 10.1088/0004-637X/799/1/51
* Milisavljevic et al. (2015b) Milisavljevic, D., Margutti, R., Kamble, A., et al. 2015b, ApJ, 815, 120, doi: 10.1088/0004-637X/815/2/120
* Mohan & Rafferty (2015) Mohan, N., & Rafferty, D. 2015, PyBDSF: Python Blob Detection and Source Finder. http://ascl.net/1502.007
* Morsony et al. (2007) Morsony, B. J., Lazzati, D., & Begelman, M. C. 2007, ApJ, 665, 569, doi: 10.1086/519483
* Murase et al. (2019) Murase, K., Franckowiak, A., Maeda, K., Margutti, R., & Beacom, J. F. 2019, ApJ, 874, 80, doi: 10.3847/1538-4357/ab0422
* Murase et al. (2015) Murase, K., Kashiyama, K., Kiuchi, K., & Bartos, I. 2015, ApJ, 805, 82, doi: 10.1088/0004-637X/805/1/82
* Murase et al. (2016) Murase, K., Kashiyama, K., & Mészáros, P. 2016, MNRAS, 461, 1498, doi: 10.1093/mnras/stw1328
* Murase et al. (2021) Murase, K., Omand, C. M. B., Coppejans, D. L., et al. 2021, arXiv e-prints, arXiv:2105.05239. https://arxiv.org/abs/2105.05239
* Murphy et al. (2013) Murphy, T., Chatterjee, S., Kaplan, D. L., et al. 2013, PASA, 30, e006, doi: 10.1017/pasa.2012.006
* Nakar (2020) Nakar, E. 2020, Phys. Rep., 886, 1, doi: 10.1016/j.physrep.2020.08.008
* Nakar & Piran (2017) Nakar, E., & Piran, T. 2017, ApJ, 834, 28, doi: 10.3847/1538-4357/834/1/28
* Omand et al. (2018) Omand, C. M. B., Kashiyama, K., & Murase, K. 2018, MNRAS, 474, 573, doi: 10.1093/mnras/stx2743
* Palliyaguru et al. (2019a) Palliyaguru, N. T., Corsi, A., Frail, D. A., et al. 2019a, ApJ, 872, 201, doi: 10.3847/1538-4357/aaf64d
* Palliyaguru et al. (2019b) —. 2019b, ApJ, 872, 201, doi: 10.3847/1538-4357/aaf64d
* Pooley et al. (2019) Pooley, D., Wheeler, J. C., Vinkó, J., et al. 2019, ApJ, 883, 120, doi: 10.3847/1538-4357/ab3e36
* Reynolds et al. (2018) Reynolds, S. P., Borkowski, K. J., & Gwynne, P. H. 2018, ApJ, 856, 133, doi: 10.3847/1538-4357/aab3d3
* Rhoads (1997) Rhoads, J. E. 1997, ApJ, 487, L1, doi: 10.1086/310876
* Robitaille (2019) Robitaille, T. 2019, APLpy v2.0: The Astronomical Plotting Library in Python, doi: 10.5281/zenodo.2567476
* Robitaille & Bressert (2012) Robitaille, T., & Bressert, E. 2012, APLpy: Astronomical Plotting Library in Python, Astrophysics Source Code Library. http://ascl.net/1208.017
* Sari et al. (1999) Sari, R., Piran, T., & Halpern, J. P. 1999, ApJ, 519, L17, doi: 10.1086/312109
* Shimwell et al. (2017) Shimwell, T. W., Röttgering, H. J. A., Best, P. N., et al. 2017, A&A, 598, A104, doi: 10.1051/0004-6361/201629313
* Shimwell et al. (2019) Shimwell, T. W., Tasse, C., Hardcastle, M. J., et al. 2019, A&A, 622, A1, doi: 10.1051/0004-6361/201833559
* Slane (2017) Slane, P. 2017, Pulsar Wind Nebulae, ed. A. W. Alsabti & P. Murdin (Cham: Springer International Publishing), 2159–2179, doi: 10.1007/978-3-319-21846-5_95
* Smith (2014) Smith, N. 2014, Annual Review of Astronomy and Astrophysics, 52, 487, doi: 10.1146/annurev-astro-081913-040025
* Soderberg et al. (2006a) Soderberg, A. M., Chevalier, R. A., Kulkarni, S. R., & Frail, D. A. 2006a, ApJ, 651, 1005, doi: 10.1086/507571
* Soderberg et al. (2006b) —. 2006b, ApJ, 651, 1005, doi: 10.1086/507571
* Soderberg et al. (2005) Soderberg, A. M., Kulkarni, S. R., Berger, E., et al. 2005, ApJ, 621, 908, doi: 10.1086/427649
* Soderberg et al. (2006c) Soderberg, A. M., Nakar, E., Berger, E., & Kulkarni, S. R. 2006c, ApJ, 638, 930, doi: 10.1086/499121
* Soderberg et al. (2010) Soderberg, A. M., Chakraborti, S., Pignata, G., et al. 2010, Nature, 463, 513, doi: 10.1038/nature08714
* Soderberg et al. (2012) Soderberg, A. M., Margutti, R., Zauderer, B. A., et al. 2012, ApJ, 752, 78, doi: 10.1088/0004-637X/752/2/78
* Terreran et al. (2019) Terreran, G., Margutti, R., Bersier, D., et al. 2019, ApJ, 883, 147, doi: 10.3847/1538-4357/ab3e37
* van Eerten et al. (2012) van Eerten, H., van der Horst, A., & MacFadyen, A. 2012, ApJ, 749, 44, doi: 10.1088/0004-637X/749/1/44
* Webb et al. (2020) Webb, N. A., Coriat, M., Traulsen, I., et al. 2020, A&A, 641, A136, doi: 10.1051/0004-6361/201937353
* Wellons et al. (2012) Wellons, S., Soderberg, A. M., & Chevalier, R. A. 2012, ApJ, 752, 17, doi: 10.1088/0004-637X/752/1/17
* Wes McKinney (2010) Wes McKinney. 2010, in Proceedings of the 9th Python in Science Conference, ed. Stéfan van der Walt & Jarrod Millman, 56 – 61, doi: 10.25080/Majora-92bf1922-00a
* Williams et al. (2019) Williams, W. L., Hardcastle, M. J., Best, P. N., et al. 2019, A&A, 622, A2, doi: 10.1051/0004-6361/201833564
* Wright et al. (2010) Wright, E. L., Eisenhardt, P. R. M., Mainzer, A. K., et al. 2010, AJ, 140, 1868, doi: 10.1088/0004-6256/140/6/1868
* Wu & Fuller (2021) Wu, S., & Fuller, J. 2021, ApJ, 906, 3, doi: 10.3847/1538-4357/abc87c
* Xu & Wei (2008) Xu, C. Y., & Wei, D. M. 2008, Acta Astronomica Sinica, 49, 387
* Zhao & Fuller (2020) Zhao, X., & Fuller, J. 2020, MNRAS, 495, 249, doi: 10.1093/mnras/staa1097
|
# HEIGHTS AND SINGULAR MODULI OF DRINFELD MODULES
Zhenlin Ran
###### Abstract
We prove that there are only finitely many singular moduli of rank 2 Drinfeld
modules that are units. In particular, we develop some techniques of heights
of Drinfeld modules to approach it.
###### Contents
1. 1 Introduction
2. 2 Minimal models of Drinfeld $A$-modules
3. 3 Heights
4. 4 Arithmetic on quadratic fundamental domain
5. 5 Bounding $h(J)$
6. A More on Complex Multiplication
## 1 Introduction
A singular modulus is the $j$-invariant of a CM elliptic curve over the
complex numbers. It is well-known that a singular modulus is an algebraic
integer. In 2015, Habegger proved the following theorem [10]:
###### Theorem 1.1.
At most finitely many singular moduli are algebraic units.
His proof employs many classical tools from Diophantine geometry. The main
idea of his proof is to bound the Weil height of a unitary singular modulus.
In particular, assuming the singular modulus is a unit, an upper bound of its
Weil height could be obtained by applying an equidistribution theorem from
Clozel and Ullmo [8, Section 2.3]. This is the biggest difficulty of the
entire proof as the number of singular moduli with big Galois orbits is hard
to control.
In this paper, we consider a function field analogue and prove the same result
for Drinfeld $\mathbb{F}_{q}[t]$-modules.
Let $C$ be a smooth, projective and geometrically irreducible curve over a
finite field $\mathbb{F}_{q}$. Fix a closed point $\infty\in C$ and let $A$ be
the ring of functions regular outside $\infty$. A Drinfeld $A$-module over
some scheme $S$ over $A$ is a pair $(\mathbb{G}_{a,\mathcal{L}},\phi)$ such
that $\mathbb{G}_{a,\mathcal{L}}$ is a line bundle over $S$ and $\phi$ is a
ring homomorphism from $A$ to $\text{End}(\mathbb{G}_{a,\mathcal{L}})$ with
some extra conditions (cf. Definition 2.1). It is well-known that Drinfeld
$A$-modules of rank 2 are the analogue of elliptic curves. Most of the
concepts and results of elliptic curves could be found for Drinfeld
$A$-modules of rank 2. For example, we can define singular modulus of Drinfeld
$\mathbb{F}_{q}[t]$-modules in the same way as elliptic curves. The main
theorem of this paper is:
###### Theorem 1.2.
Let $q$ be odd. There are only finitely many singular moduli of rank 2
Drinfeld $\mathbb{F}_{q}[t]$-modules that are algebraic units.
The strategy of proving our main theorem follows that of Habbeger. As we
pointed out earlier, the key tool that Habegger uses to control the number of
Galois orbits of quadratic imaginary numbers close to the roots of the
$j$-function is an equidistribution theorem from Clozel and Ullmo, which
enables him to obtain an upper bound for the Weil height of a unitary singular
modulus. However, this idea does not work well for our case since to the best
of our knowledge, there are not any equidistribution results for Drinfeld
$A$-modules like the one of Clozel and Ullmo. Instead, our idea to address
this issue is to study the arithmetic of quadratic imaginary points. Though
our method for the case of Drinfeld $\mathbb{F}_{q}[t]$-modules is somehow
elementary, we can show that there is at most one quadratic imaginary point in
a certain small neighbourhood of a root of the $j$-function (cf. Proposition
4.5). Thus, we could also obtain an upper bound for the Weil height of unitary
singular moduli of Drinfeld $\mathbb{F}_{q}[t]$-modules.
On the other hand, Habegger also gives a lower bound for the Weil height of
singular moduli that grows faster than the upper bound he obtained. Many tools
for the case of elliptic curves were already known while the analogues for
Drinfeld $A$-modules are not available. We follow Habegger’s strategy and
prove some analogous results for the case of Drinfeld modules, which will lead
us to a lower bound for the Weil height of singular moduli for Drinfeld
$\mathbb{F}_{q}[t]$-modules. In Drinfeld $A$-modules, the analogue of Faltings
height is Taguchi height which was introduced by Taguchi in [20] for the case
of finite characteristic and in [21] for the case of generic characteristic.
In particular, we obtain an analogous result of Nakkajima and Taguchi for
Drinfeld $\mathbb{F}_{q}[t]$-modules, which gives an explicit description for
the variation of the Taguchi heights of rank 2 Drinfeld
$\mathbb{F}_{q}[t]$-modules under isogeny, where one Drinfeld
$\mathbb{F}_{q}[t]$-module of rank $2$ has CM by arbitrary order and the other
one has CM by the maximal order. From this point on, we can obtain the
variation of the graded heights of the same Drinfeld
$\mathbb{F}_{q}[t]$-modules. As the graded height of Drinfeld $A$-modules is
the generalization of the Weil height of $j$-invariants for Drinfeld
$A$-modules, we thus obtain a lower bound for the Weil height of singular
moduli by applying a theorem of Wei where he proves the Colmez conjecture for
Drinfeld $A$-modules.
More recently Bilu, Habegger and Kühne proved the stronger result that there
are no singular moduli that are algebraic units in the case of elliptic curves
[7]. Their approach is the same as Habegger’s while the method in [7] is
effective. What makes Habegger’s proof ineffective is the equidistribution
theorem he applied. Compared to this work, a totally different approach using
Gross-Zagier, Gross-Kohnen-Zagier and their generalizations was given by Li
[15]. Li could deduce the same result of [7] as a special case of [15, Theorem
1.1].
This paper is organized as follows:
* •
In Section 2, we recall the concept of Drinfeld $A$-modules over arbitrary
$A$-schemes and the minimal models of Drinfeld $A$-modules. In particular, we
study some properties of the minimal models.
* •
In Section 3, three different types of heights are introduced and their
relationships are discussed. In particular, we study the variation of heights
of Drinfeld $A$-modules under isogenies.
* •
In Section 4, we study the number of quadratic imaginary points near the root
of the $j$-function of Drinfeld $A$-modules.
* •
In Section 5, we bound the Weil height of $j$-invariants and prove the main
theorem.
* •
The appendix is written to give some results related to CM Drinfeld
$A$-modules that could not be found in some common literature. In particular,
this section is devoted to the proof of Proposition 3.14.
Acknowledgements The author is very grateful to his advisor Florian Breuer for
advising him such an interesting project and for being supportive in all
aspects throughout his PhD study. He would also thank Marc Hindry and Urs
Hartl for inviting him to visit Paris and Münster respectively where the
author had delightful and helpful discussions with them and learnt mathematics
from them. He also appreciate Fabien Pazuki and Fu-Tsun Wei for helpful
discussions. Many thanks go to Philipp Habegger for his suggestion of working
on Lemma 5.7.
## 2 Minimal models of Drinfeld $A$-modules
In this section, we review the definition of Drinfeld $A$-modules and the
associated minimal models. In particular, we prove an analogue of a result of
Néron models of abelian varieties.
Let $C$ be a smooth, projective and geometrically irreducible curve over
$\mathbb{F}_{q}$, where $\mathbb{F}_{q}$ is a finite field with $q$ elements.
Let $\infty\in C$ be a closed point. We set
$A:=\Gamma(C\backslash\\{\infty\\},\mathcal{O}_{C})$ to be the ring of
functions regular outside $\infty$. We fix $k$ to be the field of fractions of
$A$. Let $M_{k}$ denote the set of all places of $k$. To each place $v\in
M_{k}$ we associate an absolute value $|\cdot|_{v}$ as follows:
$|x|_{v}=|q|^{-\deg(v)v(x)},\ \forall x\in k.$
Let $k_{\infty}$ denote the completion of $k$ with respect to $\infty$ and
$\mathbb{C}_{\infty}$ denote the completion of an algebraic closure of
$k_{\infty}.$
Throughout this paper, we denote by $\log$ the logarithm function of base $q$
and assume that $q$ is odd.
###### Definition 2.1.
Let $S$ be a scheme over $\text{Spec}(A)$ with structure morphism
$i:S\rightarrow\text{Spec}(A)$. A Drinfeld $A$-module over $S$ is a pair
$\textbf{E}=(\mathbb{G}_{a,\mathcal{L}},\phi)$, where $\mathcal{L}$ is an
invertible sheaf over $S$ and $\phi$ is a homomorphism from $A$ to
$\text{End}(\mathbb{G}_{a,\mathcal{L}})$ such that
1. (1)
$\partial\circ\phi=i^{\\#}$, where $i^{\\#}:A\rightarrow\mathcal{O}_{S}(S)$
and $\partial$ is the natural homomorphism of taking the constant term of
$\text{End}(\mathbb{G}_{a,\mathcal{L}})$.
2. (2)
for any $0\neq a\in A$, the morphism $\phi(a)$ is finite, and at any point of
$S$ its degree is $>1$ for some $a\in A$.
###### Remark 2.2.
1. (1)
If $\varphi\in\text{End}(\mathbb{G}_{a,\mathcal{L}})$ then
$\varphi=\sum_{n\geq 0}a_{n}\tau_{p}^{n}$, where
$a_{n}\in\Gamma(S,\mathcal{L}^{1-p^{n}})$ and $\tau_{p}$ is the relative
$p$-Frobenius. So $D\circ\phi$ lands in $\mathcal{O}_{S}(S)$.
2. (2)
We denote the subalgebra of $\mathbb{F}_{q}$-linear endomorphisms of
$\text{End}(\mathbb{G}_{a,\mathcal{L}})$ by
$\text{End}_{q}(\mathbb{G}_{a,\mathcal{L}})$. The elements of
$\text{End}_{q}(\mathbb{G}_{a,\mathcal{L}})$ are of the form
$\sum_{n}a_{n}\tau_{q}^{n}$ for $a_{n}\in\Gamma(S,\mathcal{L}^{1-q^{n}})$.
3. (3)
For any $\varphi=\sum_{n\geq
0}a_{n}\tau_{p}^{n}\in\text{End}(\mathbb{G}_{a,\mathcal{L}})$, the sum is
locally finite [14, Remark 1.2.4].
4. (4)
If $S$ is connected, then there exists an integer $r>0$ such that
$\deg(\phi(a))=|a|_{\infty}^{r}$ [14, Proposition 2.2.2]. The integer $r$ is
called the rank of the Drinfeld module E.
5. (5)
A homomorphism (resp. isogeny) from $(\mathbb{G}_{a,\mathcal{L}},\phi)$ to
$(\mathbb{G}_{a,\mathcal{M}},\psi)$ is a (resp. finite) homomorphism
$f:\mathbb{G}_{a,\mathcal{L}}\rightarrow\mathbb{G}_{a,\mathcal{M}}$ such that
$f\circ\phi(a)=\psi(a)\circ f$ for all $a\in A$.
We abbreviate $\phi(a)$ as $\phi_{a}$ for $a\in A$. If $S$ is the spectrum of
a field, then the line bundle on $S$ is unique in which case we specify a
Drinfeld module over $S$ only by $\phi$. In this paper, we assume for any
Drinfeld module $(\mathbb{G}_{a,\mathcal{L}},\phi)$, $\phi$ is $q$-linear and
$\tau:=\tau_{q}$.
###### Definition 2.3.
(Taguchi) Let $S$ be an integral normal scheme of finite type over $A$ with
function field $F$. Let $\phi$ be a Drinfeld $A$-module over $F$. A model
$\mathscr{M}=(\mathbb{G}_{a,\mathcal{L}},\varphi,f)$ of $\phi$ over $S$ is an
$A$-module scheme $\textbf{E}=(\mathbb{G}_{a,\mathcal{L}},\varphi)$ over $S$
such that $f:\textbf{E}\times_{S}\text{Spec}(F)\rightarrow\phi$ is an
isomorphism of Drinfeld modules over $F$. A model $\mathscr{M}$ of $\phi$ over
$S$ is minimal if given any other model
$\mathscr{N}=(\mathbb{G}_{a,\mathcal{L}^{\prime}},\varphi^{\prime},f^{\prime})$,
there exists a unique homomorphism $\mathscr{N}\rightarrow\mathscr{M}$ which
induces an isomorphism on the generic fibre compatible with $f$ and
$f^{\prime}$.
###### Proposition 2.4.
([21, Proposition 2.2]) Let $S$ and $F$ be as in Definition 2.3, and we
further assume $S$ is a scheme on which the two concepts of Weil divisors and
Cartier divisors coincide. Then there exists a minimal model over $S$ of
$\phi$.
###### Remark 2.5.
1. (1)
If $\phi$ has a minimal model, then the minimal model is unique up to
isomophism;
2. (2)
By checking on fibres we see that every model of $\phi$ over $S$ is smooth
over $S$.
3. (3)
By [21, Remark (4), p. 299], each model is isomorphic to a standard one. To
avoid confusion with the term standard Drinfeld module, we call the standard
model the normalized model. That is, a normalized model is a model whose
generic fibre is exactly the given Drinfeld module over $F$.
Let $\phi$ be a Drinfeld $A$-module of rank $r$ over $F$ which is a field of
finite degree over $k$, and let $R$ be the integral closure of $A$ in $F$.
Thus we can identify the invertible sheaves over $\text{Spec}(R)$ with the
fractional ideals of $R$ in $F$. For any $a\in A$, we write:
$\phi_{a}=a\tau^{0}+\cdots+\Delta_{a}\tau^{r\deg(a)}.$
Let $\mathscr{M}=(\mathbb{G}_{a,\mathcal{L}},\varphi)$ be the normalized
minimal model of $\phi$ over $R$. Then
$\Delta_{a}\in\mathcal{L}^{1-q^{r\deg(a)}}$. Moreover, if $\phi$ has good
reduction at a place $v$ of $R$, then
$v(\Delta_{a})=v(\mathcal{L}^{1-q^{r\deg(a)}})$ [21, Example, p. 301], where
$v(\mathcal{I})$ is the ramification index of a fractional ideal $\mathcal{I}$
at the prime corresponding to $v$.
###### Remark 2.6.
In this paper we do not distinguish invertible sheaves over $\text{Spec}(R)$
and fractional ideals of $R$ in $F$. We use them interchangeably without
causing any ambiguity since we only work on $\text{Spec}(R).$ Moreover, We do
not distinguish prime ideals of $R$ and valuations on $F$. That is, if we take
a prime ideal $v\in\text{Spec}(R)$, by $v(I)$ we mean the ramification index
of the fractional ideal $I$ at $v$.
###### Proposition 2.7.
Let $\phi$ be as above.
1. 1.
If $\phi$ has everywhere good reduction on $R$, then the minimal model
$\mathscr{M}$ of $\phi$ over $R$ is a Drinfeld module.
2. 2.
If $\phi$ has everywhere good reduction on $R$, $\phi^{\prime}$ is another
Drinfeld module over $F$ and $f:\phi\rightarrow\phi^{\prime}$ is an isogeny of
Drinfeld modules over $F$, then $f$ is an isogeny of their normalized minimal
models.
###### Proof.
To prove the first statement, we are left to show that $\phi_{a}$ is finite
for each $0\neq a\in A$. The above argument implies that
$v(\Delta_{a})=v(\mathcal{L}^{1-q^{r\deg(a)}})$ for every place $v$ on $R$.
This means that $\Delta_{a}$ corresponds to an isomorphism in
$\text{End}(\mathbb{G}_{a,\mathcal{L}})$. By [14, Proposition 1.2.6] we see
$\phi_{a}$ is finite for any $a\in A$. So the first statement is true.
To prove the second statement, we first write:
$\phi^{\prime}_{a}=a\tau^{0}+\cdots+\Delta_{a}^{\prime}\tau^{r\deg(a)},\
\forall a\in A;\quad f=f_{0}\tau^{0}+\cdots+f_{n}\tau^{n}.$
For any $a\in A$ we have $f\phi_{a}=\phi_{a}^{\prime}f$. By comparing the
coefficients we get
$f_{n}\Delta_{a}^{q^{n}}=\Delta_{a}^{\prime}f_{n}^{q^{r\deg(a)}}.$ (1)
Let
$\mathscr{M}^{\prime}=(\mathbb{G}_{a,\mathcal{L}^{\prime}},\varphi^{\prime})$
be the normalized minimal model of $\phi^{\prime}$. Then $f$ extends uniquely
to a morphism from $\mathscr{M}$ to $\mathscr{M}^{\prime}$ whose generic fibre
is $f$ [21, Proposition 2.5]. We therefore denote the two morphisms by $f$
interchangeably without causing any ambiguity. Since $\phi$ has good reduction
everywhere, so does $\phi^{\prime}$. The above argument again implies
$v(\Delta_{a}^{\prime})=v(\mathcal{L}^{\prime 1-r\deg(a)})$. From (1) we see
for every place $v$ on $R$
$v(f_{n})=v(\mathcal{L}^{\prime})-q^{n}v(\mathcal{L})=v(\mathcal{L}^{\prime}\mathcal{L}^{-q^{n}}).$
Thus by the same argument in the proof of the first statement, $f$ is finite,
hence an isogeny. ∎
###### Remark 2.8.
This proposition actually indicates an analogue of a well-known result that if
an abelian variety over a number field has good reduction everywhere then its
Néron model is an abelian scheme, and moreover, an isogeny between abelian
varieties with everywhere good reductions extends to a finite flat
homomorphism between their Néron models. In our case, the flatness of $f$ is
in consequence of the finiteness since a homomorphism between two line bundles
is quasi-finite if and only if it is flat [14, Proposition 1.2.5].
## 3 Heights
In this section, we study Taguchi heights, Weil heights and graded heights. In
particular, we calculate the variation of Taguchi heights and graded heights
of rank 2 Drinfeld $\mathbb{F}_{q}[t]$-modules under an isogeny.
Taguchi heights
Let $A,k,k_{\infty}$ and $\mathbb{C}_{\infty}$ be as in Section 2. In [21],
Taguchi introduced his so-called differential heights of Drinfeld $A$-modules
which are now called Taguchi heights. He defines this concept in the case when
the lattices associated to Drinfeld $A$-modules are free. Wei generalizes
Taguchi’s definition for arbitrary Drinfeld $A$-modules [23, Section 5]. We
copy the following definition from Wei. For more details, the reader could
refer to [22, Section 4] or [23, Remark 2.10].
###### Definition 3.1.
Let $\Lambda$ be an $A$-lattice of rank $r$ in $\mathbb{C}_{\infty}$, and let
$\mathcal{O}_{\infty}$ be the ring of $\infty$-adic integers in $k_{\infty}$.
Choose an orthogonal $k_{\infty}$-basis $\\{\lambda_{i}\\}_{i=1}^{r}$ of
$k_{\infty}\otimes\Lambda$ such that:
1. (1)
$\lambda_{i}\in\Lambda$ for $1\leq i\leq r$;
2. (2)
$|a_{1}\lambda_{1}+\cdots+a_{r}\lambda_{r}|_{\infty}=\max\\{|a_{i}\lambda_{i}|_{\infty}:1\leq
i\leq r\\}$ for all $a_{1},...,a_{r}\in k_{\infty}$;
3. (3)
$k_{\infty}\otimes\Lambda=\Lambda+(\mathcal{O}_{\infty}\lambda_{1}+\cdots+\mathcal{O}_{\infty}\lambda_{r})$.
The covolume $D_{A}(\Lambda)$ of the $A$-lattice $\Lambda$ is defined as
follows:
$D_{A}(\Lambda):=q^{1-g_{k}}\cdot\left(\frac{\prod_{i=1}^{r}|\lambda_{i}|_{\infty}}{\\#\left(\Lambda\cap\left(\mathcal{O}_{\infty}\lambda_{1}+\cdots+\mathcal{O}_{\infty}\lambda_{r}\right)\right)}\right)^{\frac{1}{r}}=\left(\frac{\prod_{i=1}^{r}|\lambda_{i}|_{\infty}}{\\#(\Lambda/(A\lambda_{1}+\cdots+A\lambda_{r}))}\right)^{\frac{1}{r}},$
where $g_{k}$ is the genus of the field $k$.
Let $F/k$ be a finite field extension and $R$ be the integral closure of $A$
in $F$. We may assume that $F\subset\mathbb{C}_{\infty}$ and for any infinite
place $w$ of $F$ we fix $F_{w}\subset\mathbb{C}_{\infty}$. A metrized line
bundle $(\mathcal{L},\|\cdot\|)$ on $\text{Spec}(R)$ is a projective
$R$-module $\mathcal{L}$ of rank $1$, together with norms
$\|\cdot\|_{w}:\mathcal{L}\otimes_{R}F_{w}\rightarrow\mathbb{R}$
for all infinite places $w$ of $F$. The degree $\deg(\mathcal{L},\|\cdot\|)$
of a metrized line bundle $(\mathcal{L},\|\cdot\|)$ on $R$ is
$\deg(\mathcal{L},\|\cdot\|):=\log\\#(\mathcal{L}/lR)-\sum_{w|\infty}\epsilon_{w}\log\|l\|_{w}$
for some $l\in\mathcal{L}$, and $\epsilon_{w}$ is the local degree at $w$. It
is independent of the choice of $l$ by the product formula. We note that by
taking a norm we implicitly indicate an extension of the absolute value
$|\cdot|_{\infty}$ on $k$. In this above equation, the extension of absolute
value is taken as the one remaining unchanged on $k$ while our normalization
below is different. The reader should note only in this definition we take the
infinite absolute value on $F$ such that it remains the same on $k$.
Let $\phi$ be a Drinfeld $A$-module of rank $r$ over $F$ and
$\mathscr{M}=(\mathbb{G}_{a,\mathcal{L}},\varphi,f)$ the minimal model of
$\phi$ over $R$, where $\textbf{E}=(\mathbb{G}_{a,\mathcal{L}},\varphi)$ is an
$A$-module scheme and $\mathcal{L}$ is an invertible sheaf over
$\text{Spec}(R)$. Since $\mathbb{G}_{a,\mathcal{L}}\rightarrow\text{Spec}(R)$
is smooth of finite type and relative dimension $1$, we see
$\Omega_{\mathbb{G}_{a,\mathcal{L}}/R}^{1}$ is locally free of rank $1$. Let
$e:\text{Spec}(R)\rightarrow\mathbb{G}_{a,\mathcal{L}}$ be the unit section,
then we set
$\omega_{\textbf{E}/R}:=e^{*}(\Omega_{\mathbb{G}_{a,\mathcal{L}}/R}^{1}).$
Thus $\omega_{\textbf{E}/R}\cong\mathcal{L}^{-1}$ which is the inverse of
$\mathcal{L}$ in $\text{Pic}(R)$. Without causing any ambiguity, we treat
$\omega_{\textbf{E}/R}$ as a rank $1$ projective $R$-module. Let $w$ be an
infinite place of $F$ and $\textbf{E}_{w}$ be the Drinfeld module over $F_{w}$
by extension of scalars $R\rightarrow F_{w}$. Let $\Lambda_{w}$ be the
corresponding $A$-lattice of rank $d$ in $\mathbb{C}_{\infty}$ ([9, Theorem
4.6.9] or [14, Proposition 2.1.5]). If $x$ is the coordinate function of
$\mathbb{G}_{a}/F_{w}$, then $dx$ is a generator of
$\omega_{\textbf{E}_{w}/F_{w}}(\cong\mathcal{L}^{-1}\otimes_{R}F_{w})$. We put
a metric $\|\cdot\|_{w}$ on $\omega_{\textbf{E}_{w}/F_{w}}$ by
$\|dx\|:=D_{A}(\Lambda_{w}).$
###### Definition 3.2.
Let $\phi$ be a Drinfeld $A$-module of rank $r$ over $F$ and
$\mathscr{M}=(\mathbb{G}_{a,\mathcal{L}},\varphi,f)$ be its minimal model over
$R$, where $\textbf{E}=(\mathbb{G}_{a,\mathcal{L}},\varphi)$ is an $A$-module
scheme over $R$. The Taguchi height of $\phi$ over $F$ is
$h_{\text{Tag}}(\phi/F):=\frac{1}{[F:k]}\deg(\omega_{\textbf{E}/R},\|\cdot\|),$
where the metric $\|\cdot\|$ is given as above.
It is obvious the Taguchi height of a Drinfeld $A$-module $\phi$ depends on
the choice of the field $F$. However, it will remain unchanged when taking a
finite field extension of $F$ if $\phi$ has everywhere stable reduction over
$F$. Since every Drinfeld $A$-modules has everywhere potential stable
reduction, we can define the stable Taguchi height of $\phi$ to be
$h_{\text{Tag}}^{\text{st}}(\phi):=h_{\text{Tag}}(\phi/F^{\prime}),$
where $F^{\prime}$ is a field of finite degree over $F$ on which $\phi$ has
everywhere stable reduction. The following isogeny lemma is standard.
###### Lemma 3.3.
Let $f:\phi_{1}\rightarrow\phi_{2}$ be an isogeny of Drinfeld $A$-modules over
$F$ with everywhere stable reduction. Then we have:
$h_{\text{Tag}}^{\text{st}}(\phi_{2})-h_{\text{Tag}}^{\text{st}}(\phi_{1})=\frac{1}{r}\log|\deg(f)|-\frac{1}{[F:k]}\log\\#(R/D_{f}),$
where $D_{f}$ is the different of $f$ (cf. [21, section 5.4] or [18, section
1.3]).
###### Remark 3.4.
Let $G$ be the kernel of the induced homomorphism of the minimal models of
$\phi_{1}$ and $\phi_{2}$. If $f:\phi_{1}\rightarrow\phi_{2}$ is an isogeny of
Drinfeld $A$-modules over $F$ with everywhere good reduction, then by
Proposition 2.7 we see $f$ induce an isogeny on the minimal models. In this
case, according to [13, equation 4.9.6], $D_{f}$ is the absolute different of
$G$ [17, Apendice, Définition 8].
If $\phi$ is a Drinfeld $A$-module of rank $r$ over $F$, we set:
$v(\phi):=-\min_{a\in A-\\{0\\}}\min_{i}\left\\{\frac{v(a_{i})}{q^{i}-1}:1\leq
i\leq r\deg(a)\right\\},$
where the $a_{i}$’s are coefficients in
$\phi_{a}=a\tau^{0}+\sum_{i=1}^{r\deg(a)}a_{i}\tau^{i}.$
###### Lemma 3.5.
Let $f:\phi_{1}\rightarrow\phi_{2}$ be an isogeny of Drinfeld $A$-modules over
$F$ with everywhere good reduction on $R$, and let $\mathscr{M}_{1}$ and
$\mathscr{M}_{2}$ be the normalized minimal models over $R$ of $\phi_{1}$ and
$\phi_{2}$ respectively. If $v\in\text{Spec}(R)$ is a finite place, then
$v(D_{f})=v(f_{0})+v(\phi_{1})-v(\phi_{2}),$
where $f_{0}=\partial(f)$ is the coefficient of the linear term.
###### Proof.
It will suffice to prove for the local cases, i.e. we may assume that $R$ is a
discrete valuation ring. Suppose $v$ is the valuation on $R$. Let us first
look at the case when the normalized minimal models
$\mathscr{M}_{1}=(\text{Spec}(R[X]),\varphi_{1},\text{Id})$ and
$\mathscr{M}_{2}=(\text{Spec}(R[Y]),\varphi_{2},\text{Id})$. We use $f$ to
denote the isogeny between $\text{Spec}(R[X])$ and $\text{Spec}(R[Y])$. We
denote $f^{\\#}:R[Y]\rightarrow R[X]$ the corresponding homomorphism of rings.
Thus we have
$f^{\\#}(Y)=f_{0}X+f_{1}X^{q}+\cdots+f_{n}X^{q^{n}}\in R[X].$
The kernel $G$ of $f$ is then given by $\text{Spec}(R[X]/(f^{\\#}(Y))$. By
[13, Equations 4.9.5, 4.9.6], the absolute different of $G$ is $(f_{0})\subset
R$. Thus $D_{f}=(f_{0})$ (Remark 3.4). By the construction of minimal models
[21, Proposition 2.2], we note that in this case $v(\phi_{1})=v(\phi_{2})=0$.
Hence our claim is true in this case.
To prove the general cases, let $\mathbb{G}_{a,\mathcal{L}_{i}}$ be the line
bundle of the normalized minimal model of $\phi_{i}$ for $i=1,2$. We assume
$\mathcal{L}_{i}=(a_{i})$ to be a fractional ideal for some $a_{i}\in F$. Thus
we have
$f_{j}\in(a_{2}a_{1}^{-q^{j}}).$
In particular, $f_{0}=a_{2}a_{1}^{-1}b$ for some $b\in R$. Now apply the same
argument above with the variables $X$ replaced by $a_{1}^{-1}X$ and $Y$
replaced by $a_{2}^{-1}Y$, we see $D_{f}=(b)=(f_{0}a_{1}a_{2}^{-1})$.
Therefore we have
$v(D_{f})=v(f_{0})+v(a_{1})-v(a_{2}).$
This proves our claim. ∎
Logarithmic heights
The absolute value $|\cdot|_{\infty}$ on $k$ naturally extends to a unique
absolute value on $\mathbb{C}_{\infty}$, which we denote $|\cdot|$. If $F/k$
is a finite field extension and $w$ is a place of $F$ lying over $v\in M_{k}$,
we normalize the absolute value associated to $w$ as
$|y|_{w}=|\text{N}_{F_{w}/k_{v}}(y)|_{v}^{\frac{1}{[F:k]}},\ \forall y\in F.$
Since $k$ has degree of imperfection 1 (see the Remark 3.6 below), for any
place $v\in M_{k}$ we have:
$F\otimes_{k}k_{v}\cong\prod_{w|v}F_{w}.$ (2)
Let $M_{F}$ be the set of places $w$ normalized as above. By $(2)$ we have
following two properties:
* •
Product formula: For every $y\in F$, $\sum_{w\in M_{F}}\log|y|_{w}=0.$
* •
Extension formula: $[F:k]=\sum_{w|v}[F_{w}:k_{v}].$
###### Remark 3.6.
If $F$ is a field of characteristic $p\neq 0$, by degree of imperfection of
$F$ we mean the number $n$ such that $[F:F^{p}]=p^{n}$. The isomorphism in
$(2)$ holds in a more general case of simple extension. A theorem from Becker
and MacLane [1, Theorem 6] tells us any finite extension $L/F$ can be
generated by at most $\max\\{1,n\\}$ elements.
Let $\overline{k}$ be the algebraic closure of $k$ in $\mathbb{C}_{\infty}$
and we denote $\mathbb{P}^{n}(\overline{k})$ the $n$-dimensional projective
space over $\overline{k}$. If
$\textbf{x}=(x_{0}:\cdots:x_{n})\in\mathbb{P}^{n}(\overline{k})$ and $F$ is a
finite extension of $k$ containing these coordinates, then the Weil height of
x is:
$h(\textbf{x}):=\sum_{w\in M_{F}}\max_{j}\log|x_{j}|_{w}.$
As in the number field case, this definition is independent of the choice of
both the field $F$ and the coordinates. The definition of Weil heights of
points in affine space is naturally obtained by embedding the affine space to
a projective space. In particular, for $x\in\overline{k}$ and $F$ a finite
extension of $k$ containing $x$ we have
$h(x)=\sum_{w\in M_{F}}\log\max\\{1,|x|_{w}\\}=\sum_{w\in
M_{F}}\log^{+}|x|_{w}.$
The proof of the following lemma is similar to [2, Proposition 1.6.6].
###### Lemma 3.7.
Let $\alpha\in\overline{k}$ of degree $d$ and $f(X)$ be the minimal polynomial
of $\alpha$ over $A$ with leading coefficient $a_{d}$ and roots $\alpha_{j}$,
$j=1,...,d$. Then
$dh(\alpha)=\log|a_{d}|+\sum_{j=1}^{d}\log^{+}|\alpha_{j}|.$
###### Definition 3.8.
Let $\phi$ be a Drinfeld $A$-module of rank $r$ over $F$. The global graded
degree $h_{G}(\phi)$ (resp. local graded degree $h_{G}^{w}(\phi/F)$ at a place
$w$ over $F$) of $\phi$ is
$h_{G}(\phi):=\frac{1}{[F:k]}\sum_{w\in M_{F}}\deg(w)w(\phi)\
\left(\text{resp}.\ h_{G}^{w}(\phi/F):=\frac{\deg(w)w(\phi)}{[F:k]}\right).$
###### Remark 3.9.
1. (1)
It is obvious the global graded height of $\phi$ does not depend on the choice
of the field $F$ and it is invariant under isomorphisms, while a local graded
height of $\phi$ will not satisfy such properties.
2. (2)
The global graded height is a direct interpretation of “finite” Taguchi height
(cf. [20, Definition 2.3]).
3. (3)
Let $\phi$ be a Drinfeld $\mathbb{F}_{q}[t]$-module of rank $r$ over $F$. Then
it is characterised by:
$\phi_{t}=t\tau^{0}+g_{1}\tau+\cdots+g_{r}\tau^{r},\ g_{i}\in F,g_{r}\neq 0.$
Let $m=\text{lcm}\\{q-1,...,q^{r}-1\\}$. We set
$J:=(j_{1}:\cdots:j_{r})\in\mathbb{P}^{r-1}(\overline{k})$ where
$j_{i}=g_{i}^{m/(q^{i}-1)},\ \text{for}\ i=1,...,r.$
If $r=2$, then $j_{\phi}:=j_{1}/j_{2}$ is the $j$-invariant of the Drinfeld
$A$-module $\phi$. It plays the same role as the $j$-invariant of elliptic
curves. The global graded height of $\phi$ is then given by:
$h_{G}(\phi)=\sum_{w\in M_{F}}\max_{1\leq i\leq
r}\log|g_{i}|_{w}^{1/(q^{i}-1)}.$
Thus the global graded height coincides the one in [5, equation 6]. It is
obvious that $mh_{G}(\phi)=h(J)$.
###### Proposition 3.10.
Let $\phi$ be a Drinfeld $A$-module of rank $r$ over $F$ such that $F/k$ is a
separable extension. For any $\sigma\in\text{Gal}(k^{\text{sep}}/k)$, we
denote by $\sigma(\phi)$ the Drinfeld $A$-module obtained by acting $\sigma$
on the coefficients of a Drinfeld $A$-module $\phi$. Then we have
$h_{G}(\phi)=h_{G}(\sigma(\phi)).$
###### Proof.
By Remark 3.9 (1) we may assume $F/k$ is a Galois extension so that for any
$\sigma\in\text{Gal}(F/k)$ the Drinfeld $A$-module $\sigma(\phi)$ is defined
over $F$. For any places $v\in M_{k}$ and $w\in M_{F}$ such that $w$ lies over
$v$, we see $w\circ\sigma$ is again a place lying over $v$. Thus $\sigma$
permutes the places lying over $v$. By a result of algebraic number theory,
$w$ and $w\circ\sigma$ have the same degree. Therefore
$h_{G}(\sigma(\phi))=\sum_{v\in M_{k}}\sum\limits_{\begin{subarray}{c}w\in
M_{F}\\\ w|v\end{subarray}}\deg(w\circ\sigma)w\circ\sigma(\phi)=\sum_{v\in
M_{k}}\sum\limits_{\begin{subarray}{c}w\in M_{F}\\\
w|v\end{subarray}}\deg(w)w(\phi)=h_{G}(\phi).$
∎
###### Theorem 3.11.
Let $f:\phi_{1}\rightarrow\phi_{2}$ be an isogeny of Drinfeld $A$-modules over
$F$ with everywhere good reduction on $\text{Spec}(R)$. Then we have:
$h_{\text{Tag}}^{\text{st}}(\phi_{2})-h_{\text{Tag}}^{\text{st}}(\phi_{1})=\frac{1}{r}\log|\deg(f)|-\sum\limits_{\begin{subarray}{c}w|\infty\\\
w\in
M_{F}\end{subarray}}\log|f_{0}|_{w}+h_{G}^{\text{fin}}(\phi_{2})-h_{G}^{\text{fin}}(\phi_{1}),$
where $f_{0}$ is the linear coefficient of $f$ and for $i=1,2$,
$h_{G}^{\text{fin}}(\phi_{i})$ is the sum of the local graded heights running
over all finite places of $F$.
###### Proof.
By applying Lemma 3.3 and Lemma 3.5 we obtain:
$h_{\text{Tag}}^{\text{st}}(\phi_{2})-h_{\text{Tag}}^{\text{st}}(\phi_{1})=\frac{1}{r}\log|\deg(f)|-\frac{1}{[F:k]}\sum\limits_{\begin{subarray}{c}w\in
M_{F}\\\
w\nmid\infty\end{subarray}}\deg(w)(w(f_{0})+w(\phi_{1})-w(\phi_{2})).$ (3)
By applying the product formula, we get:
$\sum\limits_{\begin{subarray}{c}w\in M_{F}\\\
w\nmid\infty\end{subarray}}\deg(w)w(f_{0})=-\sum\limits_{\begin{subarray}{c}w\in
M_{F}\\\ w|\infty\end{subarray}}\deg(w)w(f_{0}).$ (4)
Under our normalization, we have
$\log|f_{0}|_{w}=\frac{-\deg(w)w(f_{0})}{[F:k]}.$ (5)
Now substitute (5) and (4) to (3) we obtain our formula. ∎
Heights of Drinfeld $A$-modules with complex multiplication
We first recall the CM theory for Drinfeld modules. Let $\phi$ be a Drinfeld
$A$-module of rank $r$ over $\mathbb{C}_{\infty}$. We say $\phi$ has complex
multiplication if the ring of endomorphisms $\mathcal{O}:=\text{End}(\phi)$ is
a projective $A$-module of rank $r$, and $K:=\mathcal{O}\otimes_{A}k$ is
called the CM field of $\phi$. In this case, $K/k$ is an imaginary extension
of degree $r$. Here by an imaginary extension we mean there is only one place
of $K$ extending the infinite place $\infty$ of $k$. As with the case of
abelian varieties, there is also a standard theory of complex multiplication
for Drinfeld modules:
###### Theorem 3.12.
(Main Theorem of Complex Multiplication) Let $\phi$ be a Drinfeld $A$-module
of rank $r$ over $\mathbb{C}_{\infty}$ with complex multiplication. Let
$\mathcal{O}$ be the ring of endomorphisms of $\phi$ and $K$ be the CM field.
The following statements are true:
1. 1.
There is a finite extension $H_{\mathcal{O}}/K$ such that
$\text{Gal}(H_{\mathcal{O}}/K)\cong\text{Pic}(\mathcal{O})$ via the Artin map.
The field $H_{\mathcal{O}}$ is the ring class field of $\mathcal{O}$. The
prime $\infty$ of $K$ splits completely in $H_{\mathcal{O}}$, and
$\mathcal{O}$ is unramified outside $\mathcal{C}$ which is the conductor of
$\mathcal{O}$, i.e. the largest common ideal of $\mathcal{O}$ and
$\mathcal{O}_{K}$.
2. 2.
$\phi$ has good reduction at every finite place of $H_{\mathcal{O}}$.
3. 3.
If $r=2$ and $A=\mathbb{F}_{q}[t]$, then the $j$-invariant $j_{\phi}$ of
$\phi$ is integral over $A$ and $H_{\mathcal{O}}=K(j_{\phi})$.
The reader could find a proof to the above statements in [11], as well as a
complete treatment of theory of complex multiplications for Drinfeld
$A$-modules.
Suppose $\phi$ is a Drinfeld $A$-module of rank $r$ with CM by an order
$\mathcal{O}$ in a CM field $K$. We denote by $\text{Pr}(\mathcal{O})$ the
monoid of proper fractional ideals of $\mathcal{O}$ quotient by principal
ideals. It is then obvious
$\text{Pic}(\mathcal{O})\subset\text{Pr}(\mathcal{O})$ and
$\text{Pic}(\mathcal{O})$ has a natural action on $\text{Pr}(\mathcal{O})$.
Since $\phi$ has CM by $\mathcal{O}$, its associated lattice is isomorphic to
a proper ideal $I_{\phi}$ of $\mathcal{O}$.
###### Lemma 3.13.
Let $\phi_{1}$ and $\phi_{2}$ be two Drinfeld $A$-modules of rank $r$ with CM
by the same order $\mathcal{O}$, and $I_{1}$ (resp. $I_{2}$) be a proper ideal
of $\mathcal{O}$ such that the associated lattice of $\phi_{1}$ (resp.
$\phi_{2}$) is isomorphic to $I_{1}$ (resp. $I_{2}$). If $I_{1}$ and $I_{2}$
are in the same orbit of $\text{Pr}(\mathcal{O})$ under the action of
$\text{Pic}(\mathcal{O})$, then $h_{G}(\phi_{1})=h_{G}(\phi_{2})$. In
particular, if $\text{Pr}(\mathcal{O})=\text{Pic}(\mathcal{O})$ then all the
Drinfeld $A$-modules with CM by $\mathcal{O}$ have the same graded height.
###### Proof.
Without loss of generality, we assume $\phi_{i}$ has associated lattice
$I_{i}$ where $i=1,2$. We choose an invertible ideal
$J\in\text{Pic}(\mathcal{O})$ such that $I_{1}=J^{-1}\cdot I_{2}$. Thus
$I_{1}$ is homothetic to the lattice associated to $J*\phi_{2}$ (cf. [11,
Proposition 5.10 and Equation (5.18)]). On the other hand, $J*\phi_{2}$ is
isomorphic to a Drinfeld $A$-module $\phi_{2}^{\prime}$ obtained by a Galois
action on the coefficients of $\phi_{2}$ [23, Theorem A.1 (2)]. We note that
we can always choose suitable $I_{1}$ and $I_{2}$ to make $J$ integral so that
our argument makes sense. Now by Remark 3.9 (1) and Proposition 3.10 we
complete our proof. ∎
From now on, we always fix $A=\mathbb{F}_{q}[t]$. A proof of the following
result for elliptic curves is due to Nakkajima and Taguchi [16]. We only make
a few arguments here to adapt their formula to Drinfeld $A$-modules.
###### Proposition 3.14.
Let $\phi_{1}$ and $\phi_{2}$ be two Drinfeld $A$-modules of rank 2 with CM by
$\mathcal{O}_{K}$ and $\mathcal{O}$ respectively, where $K$ is an imaginary
quadratic field and $\mathcal{O}_{K}$ (resp. $\mathcal{O}$) is a maximal
(resp. arbitrary) order. We write $\mathcal{O}=A+f_{0}\mathcal{O}_{K}$ for
some $f_{0}\in A$. If $F/K$ is a finite field extension such that both
$\phi_{1}$ and $\phi_{2}$ are defined over $F$ with everywhere good reduction
then
$h_{\text{Tag}}^{\text{st}}(\phi_{2})-h_{\text{Tag}}^{\text{st}}(\phi_{1})=\frac{1}{2}\log|f_{0}|-\frac{1}{2}\sum_{v|f_{0}}\deg(v)e_{f_{0}}(v),$
where $v$ runs over all monic prime factors of $f_{0}$ and for
$l:=q^{\deg(v)}$
$e_{f_{0}}(v)=\frac{(1-\chi(v))(1-l^{-v(f_{0})})}{(l-\chi(v))(1-l^{-1})},$
and $\chi(v)=1$ if $v$ splits in $K$; $\chi(v)=0$ if $v$ ramifies in $K$;
$\chi(v)=-1$ if $v$ is inert in $K$.
###### Proof.
First we note that the Taguchi height of rank 2 Drinfeld $A$-modules with CM
does not depend on the choice of lattice that analytically generates the
corresponding Drinfeld $A$-module. So we may assume that $\phi_{1}$ is given
by $\mathcal{O}_{K}$ and $\phi_{2}$ is given by $\mathcal{O}$, and an isogeny
$f:\phi_{1}\rightarrow\phi_{2}$ given by
$f_{0}\mathcal{O}_{K}\subset\mathcal{O}$.
For the case when $\chi(v)=1$, from Theorem A.2 we see $\phi_{1}$ has ordinary
reduction at any place lying over $v$. Applying the same argument from [16,
Proposition 4] we get $e_{f_{0}}(v)=0$. The reduction process in the argument
for Drinfeld $A$-modules is given by Theorem A.4. For the case of
supersingular reduction, the argument for our case is exactly the same as [16,
Section 2.2] with only one modification that we take $l=q^{\deg(v)}$. ∎
###### Corollary 3.15.
Assume the same conditions as in Proposition 3.14. The following formula is
true
$h_{G}(\phi_{2})-h_{G}(\phi_{1})=\log|f_{0}|-\frac{1}{2}\sum_{v|f_{0}}\deg(v)e_{f_{0}}(v)+h_{G}^{\infty}(\phi_{2}^{\prime})-h_{G}^{\infty}(\phi_{1}^{\prime}),$
where $\phi_{1}^{\prime}$ is the Drinfeld $A$-module given by the lattice
$\mathcal{O}_{K}$ and $\phi_{2}^{\prime}$ is given by $\mathcal{O}$, and
$h_{G}^{\infty}(\phi_{i}^{\prime})$ is the sum of local graded heights of
$\phi_{i}^{\prime}$ at infinite places for $i=1,2$.
###### Proof.
By Lemma 3.13 we can choose $\phi_{i}$ to be $\phi_{i}^{\prime}$, $i=1,2.$ It
is then a trivial consequence of Theorem 3.11 and Proposition 3.14. ∎
## 4 Arithmetic on quadratic fundamental domain
We assume our Drinfeld $A$-module $\phi$ has rank 2 with CM for the rest this
paper. Let $\Omega:=\mathbb{C}_{\infty}\backslash k_{\infty}$ be the Drinfeld
upper-half plane so that $\text{PGL}_{2}(A)\backslash\Omega$ are the
$\mathbb{C}_{\infty}$-points of the coarse moduli space of rank 2 Drinfeld
$A$-modules over $\mathbb{C}_{\infty}$. As in the case of elliptic curves,
there are bijections:
$\text{PGL}_{2}(A)\backslash\Omega\xrightleftharpoons{\quad\quad}\\{\text{Lattices
of rank 2 in }\mathbb{C}_{\infty}/\cong\\}\xrightleftharpoons{\quad\
\quad}\mathbb{C}_{\infty}.$
Thus we obtain a natural $j$-function$:\Omega\rightarrow\mathbb{C}_{\infty}$.
The set of $j$-invariants of rank 2 Drinfeld $A$-modules with CM is precisely
the image of the $j$-function at imaginary quadratic arguments.
Unfortunately, the Drinfeld upper-half plane doesn’t have a good geometry as
the Poincaré upper-half plane of complex numbers does. However, we can still
define the quadratic fundamental domain.
###### Definition 4.1.
(cf. [4, Definition 3.4]) The quadratic fundamental domain is
$\displaystyle\mathcal{D}=\\{z\in\Omega$ $\displaystyle:z\text{ satisfies an
equation of the form }az^{2}+bz+c=0,$ (6) $\displaystyle\quad\text{where
}a,b,c\in A,\ a\text{ is monic, }|b|<|a|\leq|c|\text{ and}$
$\displaystyle\quad\text{gcd}(a,b,c)=1\\}.$
Any rank 2 lattice corresponding to a CM Drinfeld module is homothetic to
$\Lambda_{z}$ for some $z\in\mathcal{D}_{K}$, where $\Lambda_{z}$ denotes the
lattice generated by $z$ and 1, and $\mathcal{D}_{K}:=\mathcal{D}\cap K$ for
some quadratic imaginary extension $K/k$ in $\mathbb{C}_{\infty}$. Unlike the
case of elliptic curves, such $z$ is not necessarily unique.
###### Proposition 4.2.
([3, Proposition 1.2.1]) Let $q$ be odd and $K$ be a quadratic extension of
$k$. Then $K$ is a Kummer extension and can be written in the form
$K=k(\sqrt{\delta})$ for some square-free $\delta\in A$. Let $m=\deg(\delta)$.
Then we have
1. 1.
The place $\infty$ ramifies in $K$ if and only if $m$ is odd;
2. 2.
The place $\infty$ is inert in $K$ if and only if $m$ is even and the leading
coefficient of $\delta$ is not a square in $\mathbb{F}_{q}$;
3. 3.
The place $\infty$ splits in $K$ if and only if $m$ is even and the leading
coefficient of $\delta$ is a square in $\mathbb{F}_{q}$.
Let $\delta\in A$ be a polynomial of odd degree or even degree with the
leading coefficient not being a square in $\mathbb{F}_{q}$, and let
$\sqrt{\delta}\in\overline{k}$ be a root of $X^{2}-\delta$. The field
$K:=k(\sqrt{\delta})$ is a quadratic imaginary field by Proposition 4.2. Let
$\mathcal{O}_{K}$ be its maximal order. If $\mathcal{O}\subset\mathcal{O}_{K}$
is a suborder of discriminant $\delta$, then there exsits some $f\in A$ such
that $\mathcal{O}=A+f\mathcal{O}_{K}$ and such $f$ is called the conductor of
$\mathcal{O}$. The discriminant $\delta_{0}$ of $\mathcal{O}_{K}$ is called
the fundamental discriminant of $K$ and $K=k(\sqrt{\delta_{0}})$. The
discriminant of $\mathcal{O}$ is $\delta=4f^{2}\delta_{0}$.
We denote $T_{\delta}$ the set of triples $(a,b,c)$ with $a,b,c\in A$ such
that $b^{2}-4ac=\delta$ and satisfying (6). For $(a,b,c)\in T_{\delta}$ we set
$z(a,b,c)=\frac{-b+\sqrt{\delta}}{2a}\in K=k(\sqrt{\delta}).$
The map $(a,b,c)\mapsto j(z(a,b,c))$ is a bijection from $T_{\delta}$ to the
Galois conjugates of $j(z)$.
###### Lemma 4.3.
If $\infty$ ramifies in $K$, then there doesn’t exist any
$z\in\mathcal{D}_{K}$ such that $z$ is in the open ball of radius 1 of the
points $u$ for any $u\in\mathbb{F}_{q^{2}}\backslash\mathbb{F}_{q}$.
###### Proof.
For some $u\in\mathbb{F}_{q^{2}}\backslash\mathbb{F}_{q}$, we assume there
exists $z\in\mathcal{D}_{K}$ such that $|z-u|<1$. In this case, we have
$|z|=1$ as $|u|=1$. Besides, there exists a triple $(a,b,c)$ satisfying (6)
such that $b^{2}-4ac=\delta$ and $z$ is root of the equation:
$aX^{2}+bX+c=0.$
We write $K=k(\sqrt{\delta_{0}})$, where $\delta_{0}\in A$ is square-free. By
Proposition 4.2 and since $K$ is a quadratic imaginary extension of $k$, we
see $\delta_{0}$ is either of odd degree or of even degree with leading
coefficients not a square in $\mathbb{F}_{q}$. Also we have
$|az^{2}+bz|=|c|.$
Since $|bz|=|b|<|a|=|az^{2}|$, we have $|a|=|az^{2}+bz|=|c|$, which implies
$a$ and $c$ have the same degree and $b$ has degree less than $a$ and $c$.
Therefore $\delta$ has even degree. Since $\delta=4f^{2}\delta_{0}$, we see
$\delta_{0}$ has even degree, hence with leading coefficients not a square in
$\mathbb{F}_{q}$. This is equivalant to saying that $\infty$ is inert in $K$.
∎
###### Remark 4.4.
If $\delta$ has even degree with leading coefficient not a square in
$\mathbb{F}_{q}$, then $\sqrt{\delta}\in\mathbb{F}_{q^{2}}((\frac{1}{t})).$
Moreover, if there exists some $z\in\mathcal{D}_{K}$ with discriminant
$\delta$ such that $|z-u|<1$, then $\delta$ has leading coefficient $4u^{2}$.
This also suggests such $z$ can be close only to those
$u\in\mathbb{F}_{q^{2}}\backslash\mathbb{F}_{q}$ such that
$u^{2}\in\mathbb{F}_{q}$.
###### Proposition 4.5.
Let $u\in\mathbb{F}_{q^{2}}\backslash\mathbb{F}_{q}$. The number of
$(a,b,c)\in T_{\delta}$ such that $|z(a,b,c)-u|<\sqrt{|\delta|}^{-1}$ is at
most 1.
###### Proof.
We first notice that $\delta$ has degree $>0$. If not, then
$z\in\mathbb{F}_{q^{2}}\backslash\mathbb{F}_{q}$. This contradicts the fact
that $\mathcal{D}_{K}\cap\mathbb{F}_{q^{2}}=\emptyset$. So $\delta$ has
positive degree, which implies $\sqrt{|\delta|}\geq 1$. Thus we have
$|z-u|<\sqrt{|\delta|}^{-1}\leq 1$. By the proof of Lemma 4.3, we have
$\delta=\alpha_{2e}t^{2e}+\cdots+\alpha_{0},\ \alpha_{i}\in\mathbb{F}_{q}\
\text{for }i=0,...,2e\ \text{and $e$ is a positive integer.}$
Thus we have
$\sqrt{\delta}=\lambda_{e}t^{e}+\cdots+\lambda_{0}+\lambda_{-1}t^{-1}+\lambda_{-2}t^{-2}\cdots$
with coefficients in $\mathbb{F}_{q^{2}}$. By identifying
$(\sqrt{\delta})^{2}=\delta$ we obtain:
$\displaystyle\alpha_{2e}$ $\displaystyle=\lambda_{e}^{2};$
$\displaystyle\alpha_{2e-1}$
$\displaystyle=\lambda_{e}\lambda_{e-1}+\lambda_{e-1}\lambda_{e};$
$\displaystyle\vdots$ $\displaystyle\alpha_{e}$
$\displaystyle=\lambda_{e}\lambda_{0}+\lambda_{e-1}\lambda_{1}+\cdots+\lambda_{1}\lambda_{e-1}+\lambda_{0}\lambda_{e};$
$\displaystyle\vdots$ $\displaystyle\alpha_{0}$
$\displaystyle=\lambda_{e}\lambda_{-e}+\lambda_{e-1}\lambda_{-(e-1)}+\cdots+\lambda_{-(e-1)}\lambda_{e-1}+\lambda_{-e}\lambda_{e}.$
First notice $\lambda_{e}=2u\neq 0$ because $|z-u|<1$. We first claim that
$(2u)^{-1}\lambda_{i}\in\mathbb{F}_{q}$ when $i=0,1,...,e$.
Our claim is trivial when $i=e$. We proceed by induction and suppose it’s true
for $\lambda_{e},...,\lambda_{n}$ when $0<n\leq e$. From the equations above
we have
$\alpha_{e+n-1}=\lambda_{e}\lambda_{n-1}+\lambda_{e-1}\lambda_{n}+\cdots+\lambda_{n}\lambda_{e-1}+\lambda_{n-1}\lambda_{e}.$
By multiplying $(2u)^{-2}$ on both sides we obtain
$(2u)^{-2}\alpha_{e+n-1}=(2u)^{-1}\lambda_{e}(2u)^{-1}\lambda_{n-1}+\cdots+(2u)^{-1}\lambda_{n-1}(2u)^{-1}\lambda_{e}.$
Since all terms other than $(2u)^{-1}\lambda_{e}(2u)^{-1}\lambda_{n-1}$ are in
$\mathbb{F}_{q}$ and $\lambda_{e}=2u$, we have
$(2u)^{-1}\lambda_{n-1}\in\mathbb{F}_{q}$.
Now recall $z=\frac{-b+\sqrt{\delta}}{2a}$ with triple $(a,b,c)\in T_{\delta}$
and $\delta=b^{2}-4ac$. Then $|z-u|<\sqrt{|\delta|}^{-1}$ is equivalent to
$\deg(a)-\deg(\sqrt{\delta}-b-2au)>\deg(\sqrt{\delta})=\deg(a).$
So we have $\deg(\sqrt{\delta}-b-2au)<0$. Suppose
$a=\sum_{i=0}^{e}a_{i}t^{i},\ b=\sum_{i=0}^{e}b_{i}t^{i},\ \text{with all
$a_{i},b_{i}\in\mathbb{F}_{q}$}.$
Thus for all $i=0,...,e$ we have $\lambda_{i}-2a_{i}u=b_{i}$. Since
$(2u)^{-1}\lambda_{i}\in\mathbb{F}_{q}$, we see $b_{i}=0$ and
$a_{i}=(2u)^{-1}\lambda_{i}$ for all $i=0,...,e$. Thus, $a,b,c$ are completely
determined by $|z(a,b,c)-u|<\sqrt{|\delta|}^{-1}.$ ∎
## 5 Bounding $h(J)$
We prove our main theorem in this section.
Upper bound on $h(J)$
Let $\phi$ be a CM Drinfeld $A$-module of rank 2 over $\mathbb{C}_{\infty}$
and $J$ be its $j$-invariant of degree $d$ over $A$. Let
$\mathcal{O}=\text{End}(\phi)$ and $K=\mathcal{O}\otimes_{A}k$. We denote
$J=J_{1},...,J_{d}$ all the Galois conjugates of $J$ and $z_{1},...,z_{d}$ the
corresponding points in $\mathcal{D}_{K}$. Then for each $i$, we have $z_{i}$
satisfying the equation:
$a_{i}X^{2}+b_{i}X+c_{i}=0,\ (a_{i},b_{i},c_{i})\in T_{\delta}$
and $b_{i}^{2}-4a_{i}c_{i}=\delta$ for some $\delta\in A$ that is the
discriminant of $\mathcal{O}$. In this subsection we prove:
###### Proposition 5.1.
Assuming the notations above and $J$ is an algebraic unit, we have
$h(J)\leq\frac{2}{d}(q+1)\log\sqrt{|\delta|}+O(q),$
where $O(q)$ is some constant depending on $q$.
We first fix some notations. Let $u$ be a point such that
$u\in\mathbb{F}_{q^{2}}\backslash\mathbb{F}_{q}$ and
$|z|_{A}=\inf_{a\in A}|z-a|,$ $|z|_{i}=\inf_{x\in k_{\infty}}|z-x|.$
###### Lemma 5.2.
([3, Proposition 3.2.5]) If $z\in\mathcal{D}_{K},$ then
$|z|_{i}=|z|_{A}=|z|\geq 1$.
###### Lemma 5.3.
For each $z_{i}$, we have $h(z_{i})\leq\log\sqrt{|\delta|}.$
###### Proof.
Proof is similar to [10, Lemma 5]. ∎
###### Lemma 5.4.
([6, Lemma 2.6.9]) Suppose $z\in\Omega$ such that $|z|_{A}>q^{-1}$. If
$u\in\mathbb{F}_{q^{2}}\backslash\mathbb{F}_{q}$ and $|z-u|<q^{-1}$, then
there exists $\zeta\in\mathbb{C}_{\infty}$ with $|\zeta|<1$ such that
$\displaystyle j(z)$
$\displaystyle=t^{q}u^{-2}(1-u^{q-1})^{-2}(z-u)^{q+1}(1+\zeta),$
$\displaystyle|j(z)|$ $\displaystyle=q^{q}|z-u|^{q+1}.$
Let $h(J)$ be the Weil height of $J$. Therefore, by Lemma 3.7 we have
$dh(J)=\sum_{i=1}^{d}\log^{+}|J_{i}|=\sum_{i=1}^{d}\log^{+}|j(z_{i})|.$
If we assume $J$ is an algebraic unit, then pick a
$u\in\mathbb{F}_{q^{2}}\backslash\mathbb{F}_{q}$ we have:
$\displaystyle dh(J)=dh(J^{-1})$
$\displaystyle=\sum_{i=1}^{d}\log^{+}|j(z_{i})|^{-1}$
$\displaystyle=\sum_{|z_{i}-u|<\sqrt{|\delta|}^{-1}}\log^{+}|j(z_{i})|^{-1}+\sum_{\sqrt{|\delta|}^{-1}\leq|z_{i}-u|\leq
q^{-1}}\log^{+}|j(z_{i})|^{-1}+dO(q).$ (7)
###### Lemma 5.5.
Assuming the notations above, we have
$\sum_{|z_{i}-u|<\sqrt{|\delta|}^{-1}}\log^{+}|j(z_{i})|^{-1}\leq
2(q+1)\log\sqrt{|\delta|}-q.$
###### Proof.
Since $z_{i}\in\mathcal{D}_{K}$, we have $|z|_{A}\geq 1$ by Lemma 5.2. Thus
according to Lemma 5.4 we have
$\log|j(z_{i})|^{-1}=-q-(q+1)\log|z_{i}-u|.$ (8)
Let $L=k(z_{i},u)$, then $[L:k]=2$ if
$z_{i}\in\mathbb{F}_{q}(t)(u)=\mathbb{F}_{q^{2}}(t)$ and $[L:k]=4$ otherwise.
In the first case, there is only one $w\in M_{L}$ lying over $\infty\in
M_{k}$. Therefore we have $|z_{i}-u|_{w}=|z_{i}-u|$ under our normalization.
In the second case, there are two places in $M_{L}$ lying over $\infty\in
M_{k}$, and we take $w$ to be one of the two. Thus we have
$|z_{i}-u|=|z_{i}-u|_{w}^{2}$. Either way, we find
$\log|z_{i}-u|\leq 2\sum_{w}\log^{+}|z_{i}-w|_{w}=2h(z_{i}-u).$
Note the fact that $h(\alpha)=h(1/\alpha)$ for any $\alpha\in\overline{k}$ and
we have
$\log|z_{i}-u|\geq-2h(z_{i}-u)\geq-2(h(z_{i})+h(u))=-2h(z_{i}).$
Now substitute this inequality to (8) and apply Lemma 5.3 we obtain
$\log|j(z_{i})|^{-1}\leq 2(q+1)\log\sqrt{|\delta|}-q.$
Since the number of such $z_{i}$ is at most one by Proposition 4.5, we
conclude. ∎
Now we are ready to prove Proposition 5.1.
###### Proof.
We are left to estimate
$\sum_{\sqrt{|\delta|}^{-1}\leq|z_{i}-u|\leq q^{-1}}\log^{+}|j(z_{i})|^{-1}.$
From (8) and $\sqrt{|\delta|}^{-1}\leq|z_{i}-u|$ we have
$\sum_{\sqrt{|\delta|}^{-1}\leq|z_{i}-u|\leq
q^{-1}}\log^{+}|j(z_{i})|^{-1}\leq\left(-q+(q+1)\log\sqrt{|\delta|}\right)\cdot\sum_{\sqrt{|\delta|}^{-1}\leq|z_{i}-u|\leq
q^{-1}}1$
* •
If there exists some $z_{i}$ such that $|z_{i}-u|<\sqrt{|\delta|}^{-1}$, then
we have
$\sum_{\sqrt{|\delta|}^{-1}\leq|z_{i}-u|\leq q^{-1}}1\leq d-1;$
* •
If there doesn’t exist $z_{i}$ such that $|z_{i}-u|<\sqrt{|\delta|}^{-1}$,
then
$\sum_{\sqrt{|\delta|}^{-1}\leq|z_{i}-u|\leq q^{-1}}1\leq d.$
Combining with Lemma 5.5, we are done. ∎
Lower bound on $h(J)$
###### Lemma 5.6.
Let $\phi^{0}$ be a Drinfeld $A$-module of rank 2 with CM by the maximal order
$\mathcal{O}_{K}$ in an imaginary quadratic field $K$. We denote the genus of
$K$ by $g_{K}$. Then we have:
$h_{\text{Tag}}^{\text{st}}(\phi^{0})\geq\left(\frac{1}{2}-\frac{1}{\sqrt{q}+1}\right)g_{K}-\frac{5q-3}{4(q-1)}.$
###### Proof.
By the equation of stable Taguchi height from [23, Section 5.2.1], we get
$h_{\text{Tag}}^{\text{st}}(\phi^{0})=\frac{g_{K}}{2}\log
q_{K}+\frac{1}{2}\left(\log q_{\infty}-\log
q_{K}\right)-1+\frac{\gamma_{K}}{2\ln q},$
where $q_{K}$ is the cardinality of the field of constants in $K$,
$q_{\infty}$ is the cardinality of the residue field of $K_{\infty}$, and
$\gamma_{K}$ is the Euler-Kronecker constant of $K$ [12, Equation (0.2)]. We
remind the reader that the definition of stable Taguchi height in [23,
Equation 5.1] is a multiple of our stable Taguchi height by the constant $\ln
q$. From [12, Equation 1.4.6] we get
$\frac{\gamma_{K}}{2\ln q}\geq\frac{-g_{K}}{\sqrt{q}+1}+\frac{q-3}{4(q-1)}.$
Now using the fact $q\leq q_{K},q_{\infty}\leq q^{2}$ we get the lower bound.
∎
###### Lemma 5.7.
Let $e_{f_{0}}(v)$ be as in Proposition 3.14 and $f_{0}\in A$. Then we have
$\frac{1}{2}\sum_{v|f_{0}}\deg(v)e_{f_{0}}(v)\leq\frac{9}{4}\log\log|f_{0}|+C_{q},$
where $v$ runs through all the monic prime factors of $f_{0}$, and $C_{q}$ is
a computable constant depending on $q$.
###### Proof.
First we need a Mertens-type formula for function field, i.e. the following
inequality:
$\sum_{|v|\leq x}\frac{\log|v|}{|v|}\leq\log x+O(1),\text{where $v$'s are
monic prime polynomials}.$
To see this, we notice that
$\sum_{|v|\leq x}\frac{\log|v|}{|v|}=\sum_{i=1}^{n:=\lfloor\log
x\rfloor}\frac{i}{q^{i}}\cdot a_{i},$
where $a_{i}$ is the number of monic prime polynomials of degree $i$. By [19,
Theorem 2.2] we obtain
$\sum_{i=1}^{n}\frac{i}{q^{i}}\cdot
a_{i}=\sum_{i=1}^{n}\left(1+O({q^{-i/2}})\right)\leq\log x+O(1).$
Recall that
$e_{f_{0}}(v)=\frac{(1-\chi(v))(1-l^{-v(f_{0})})}{(l-\chi(v))(1-l^{-1})},\text{
where }l=|v|.$
Note that $|l|\geq 3$ and $\chi(v)\in\\{-1,0,1\\}$. Thus we get
$e_{f_{0}}(v)\leq\frac{2}{l-l^{-1}}\leq\frac{9}{4l}.$
Thus we have
$\sum_{v|f_{0}}\deg(v)e_{f_{0}}(v)\leq\frac{9}{4}\sum_{v|f_{0}}\frac{\log
l}{l}=\frac{9}{4}\left(\sum\limits_{\begin{subarray}{c}v|f_{0}\\\
|v|\leq\log|f_{0}|\end{subarray}}\frac{\log
l}{l}+\sum\limits_{\begin{subarray}{c}v|f_{0}\\\
|v|>\log|f_{0}|\end{subarray}}\frac{\log l}{l}\right).$
We have proven
$\sum\limits_{\begin{subarray}{c}v|f_{0}\\\
|v|\leq\log|f_{0}|\end{subarray}}\frac{\log l}{l}\leq\log\log|f_{0}|+O(1).$
For the other term, we have
$\sum\limits_{\begin{subarray}{c}v|f_{0}\\\
|v|>\log|f_{0}|\end{subarray}}\frac{\log
l}{l}\leq\frac{\log\log|f_{0}|}{\log|f_{0}|}\cdot\sum\limits_{\begin{subarray}{c}v|f_{0}\\\
|v|>\log|f_{0}|\end{subarray}}1.$
Using the product formula we see there are at most $\log|f_{0}|$ monic prime
factors of $f_{0}$. As [19, Theorem 2.2] is effective, all the constant terms
are summed up to a computable constant $C_{q}$. ∎
###### Proposition 5.8.
Let $J$ be a singular modulus of rank 2 Drinfeld $A$-module with corresponding
discriminant $\delta$ with conductor $f_{0}$. There exists some computable
constant $C_{q}$ with respect to $q$ such that
$h(J)\geq(q^{2}-1)\left(\frac{1}{2}-\frac{1}{\sqrt{q}+1}\right)\log\sqrt{|\delta|}+\left(\frac{1}{2}+\frac{1}{\sqrt{q}+1}\right)\log|f_{0}|-\frac{9}{4}\log\log|f_{0}|-C_{q}.$
###### Proof.
We recall that for $K=k(\sqrt{\delta_{0}})$ with $\delta_{0}$ square free, the
genus $g_{K}$ of $K$ is given by (cf. [4, Section 3])
$g_{K}=\begin{cases}\frac{\log|\delta_{0}|-1}{2}&\text{if $\deg(\delta_{0})$
is odd,}\\\ \frac{\log|\delta_{0}|-2}{2}&\text{if $\deg(\delta_{0})$ is
even.}\end{cases}$
We input a result from [5, equation (23)], which says that
$|h_{G}^{\infty}(\phi^{\prime})-h_{G}^{\infty}(\phi)|\leq\frac{q}{q-1}-\frac{q^{r}}{q^{r}-1},$
where $\phi$ and $\phi^{\prime}$ are two isogenous Drinfeld $A$-modules of
rank $r$. Using Corollary 3.15, Lemma 5.6, Lemma 5.7 and the facts that
$h_{G}(\phi)\geq h_{\text{Tag}}^{\text{st}}(\phi)$ and
$h(J)=(q^{2}-1)h_{G}(\phi)$ we complete our proof. ∎
###### Remark 5.9.
1. (1)
We note that [5, Equation (23)] holds true only for reduced Drinfeld modules
[5, definition before Lemma 4.2]. However, Lemma 3.13 ensures in the rank 2
case we can always choose the graded height of the Drinfeld module with
associated lattice being the CM order, hence reduced.
2. (2)
One can also take $q\geq 3$ to make our statement independent of $q$, i.e.
$h(J)\geq
4(2-\sqrt{3})\log\sqrt{|\delta|}+\frac{1}{2}\log|f_{0}|-\frac{9}{4}\log\log|f_{0}|-C_{3}.$
Now we are ready to prove our main theorem. We restate our theorem:
###### Theorem 5.10.
Let $q$ be odd. There are only finitely many singular moduli of rank 2
Drinfeld $A$-modules that are algebraic units.
###### Proof.
Assume $J$ is a unitary singular modulus. There are finitely many singular
moduli with bounded degree. So we can choose $d$ large enough in Proposition
5.1 such that
$\frac{2}{d}(q+1)\log\sqrt{|\delta|}<(q^{2}-1)\left(\frac{1}{2}-\frac{1}{\sqrt{q}+1}\right)\log\sqrt{|\delta|}.$
We note that
$\left(\frac{1}{2}+\frac{1}{\sqrt{q}+1}\right)\log|f_{0}|-\frac{9}{4}\log\log|f_{0}|\geq
0$. Therefore, by Proposition 5.1 and Proposition 5.8 we find a constant upper
bound for $\log\sqrt{|\delta|}$. Lemma 5.3 implies that this is also a
constant upper bound for $h(z_{i})$ for $z_{i}$. We note that $z_{i}$ has
degree 2. Thus the Northcott theorem implies our theorem. ∎
## Appendix A More on Complex Multiplication
We use notations as in section 2. This appendix is mainly devoted for the
proof of Proposition 3.14. Actually the results stated here are already known
for elliptic curves. The results below may be already known to many experts.
Because of a dearth of literature for Drinfeld modules, the details are worked
out here for the convenience of the reader.
Let $\phi_{1}$ and $\phi_{2}$ both be rank $r$ Drinfeld $A$-module over
$\mathbb{C}_{\infty}$ with complex multiplication. Let $F/k$ be a finite field
extension such that both $\phi_{1}$ and $\phi_{2}$ are defined over $F$ with
everywhere good reduction. Let $R$ be the integral closure of $A$ in $F$ and
denote $\mathscr{M}_{1}$, $\mathscr{M}_{2}$ the minimal model over $R$ of
$\phi_{1}$, $\phi_{2}$ respectively. If
$f:\mathscr{M}_{1}\rightarrow\mathscr{M}_{2}$ is an isogeny, then it induces
an isogeny of Drinfeld modules after taking reduction at a prime
$v\in\text{Spec}(R)$.
###### Lemma A.1.
Let $\phi_{1}^{v}$ and $\phi_{2}^{v}$ be Drinfeld modules over $\mathbf{k}(v)$
obtained by taking reduction on $\mathscr{M}_{1}$ and $\mathscr{M}_{2}$
respectively at $v$, where $\mathbf{k}(v)$ is the residue field at $v$. Let
$\text{Hom}_{F}(\phi_{1},\phi_{2})$ denote the set of isogenies over $F$
between $\phi_{1}$ and $\phi_{2}$, similarly for
$\text{Hom}_{\mathbf{k}(v)}(\phi_{1}^{v},\phi_{2}^{v})$. Then there is a
canonical injection of $A$-modules:
$\text{Hom}_{F}(\phi_{1},\phi_{2})\xhookrightarrow{}\text{Hom}_{\mathbf{k}(v)}(\phi_{1}^{v},\phi_{2}^{v}).$
###### Proof.
It is easy to check the map is a morphism of $A$-modules. By Proposition 2.7
and [21, Proposition 2.5] we obtain
$\text{Hom}_{F}(\phi_{1},\phi_{2})=\text{Hom}_{R}(\mathscr{M}_{1},\mathscr{M}_{2}).$
Let $f\in\text{Hom}_{R}(\mathscr{M}_{1},\mathscr{M}_{2})$ be an isogeny. Then
it is finite, which implies it has leading coefficient that is non-zero after
reduction at $v$. This proves the injectivity. ∎
From now on, we assume $A=\mathbb{F}_{q}[t]$ and $\phi$ is a rank $2$ Drinfeld
$A$-module over $\mathbb{C}_{\infty}$ with CM by the maximal order
$\mathcal{O}_{K}$, where $K\subset\mathbb{C}_{\infty}$ is a quadratic
imaginary field over $k$. Further we assume $\phi$ is obtained through the
$A$-lattice $\mathcal{O}_{K}$ and $\phi$ is defined over $F$. Let $P\in A$ be
a prime element.
###### Theorem A.2.
For any place $v\in\text{Spec}(R)$, the reduction of $\phi$ at $v$ is ordinary
if and only if $v\cap k$ splits in $\mathcal{O}_{K}$.
###### Proof.
Let $P=v\cap k$ and $\bar{\pi}$ be the Frobenius morphism of $\phi^{v}$. From
Lemma A.1 we see
$K=\text{End}_{F}(\phi_{1})\otimes_{A}k\xhookrightarrow{}\text{End}_{\mathbf{k}(v)}(\phi^{v})\otimes_{A}k:=D.$
By [9, Proposition 4.12.17] we deduce that $\phi$ is ordinary at $v$ if and
only if $K=\text{End}_{\mathbf{k}(v)}(\phi^{v})\otimes_{A}k$. This is
equivalent to saying that
$\mathcal{O}_{K}=\text{End}_{\mathbf{k}(v)}(\phi^{v})$. We can embed $A$ into
$\text{End}_{\mathbf{k}(v)}(\phi^{v})$ via the Drinfeld module $\phi^{v}$. Let
$E:=k(\bar{\pi})$. Then there is only one prime $\mathscr{P}$ of $E$
containing $\bar{\pi}$ and $\mathscr{P}$ lies over $P$ [9, Theorem 4.12.8].
If $\phi$ has ordinary reduction at $v$, then $E\cong K$ as $\bar{\pi}\notin
k$. Again by [9, Proposition 4.12.17] there are more than one primes of $E$
lying over $P$. Thus $P$ splits in $\mathcal{O}_{K}$. Next we show the other
way around. First, we write
$P\mathcal{O}_{K}=\mathcal{P}\mathcal{P}^{\prime}$. Assume the reduction of
$\phi$ at $v$ is supersingular. So it is a consequence that
$\dim_{k}D=r^{2}=4$. Since $\phi$ has good reduction at $v$, $\phi^{v}$ has
rank 2 over $\mathbf{k}(v)$. Thus we have
$2=\text{rank}(\phi^{v})=t\cdot[E:k]$, where $t$ is an integer such that
$t^{2}=\dim_{E}D$. As $\dim_{k}D=4=\dim_{E}D\cdot[E:k]$, we have $t=2$.
Therefore, $E=k$. In particular, $\bar{\pi}\in A$. In this case, it is clear
that $\mathscr{P}=(P)\subset A$. On the other hand, we can obtain a Drinfeld
$\mathcal{O}_{K}$-module $\psi$ over $F$ by extending $\phi$ to
$\text{End}_{F}(\phi)$. By taking reduction at $v$ again, we obtain a Drinfeld
$\mathcal{O}_{K}$-module $\psi^{v}$ over $\mathbf{k}(v)$. It is trivial
$\bar{\pi}$ is the Frobenius element of $\psi^{v}$. As $\bar{\pi}\in A\subset
K$, there is only one prime ideal of $K$ containing $\bar{\pi}$. However,
$\bar{\pi}\in P\mathcal{O}_{K}=\mathcal{P}\mathcal{P}^{\prime}$. This is a
contradiction. ∎
###### Corollary A.3.
If $P\mathcal{O}_{K}=\mathscr{P}\mathscr{P}^{\prime}$ where $\mathscr{P}$ and
$\mathscr{P}^{\prime}$ are prime ideals of $\mathcal{O}_{K}$ both lying over
$P$, then for any place $v\in\text{Spec}(R)$ over $P$ the natural morphism
$\text{End}_{F}(\phi)\rightarrow\text{End}_{\mathbf{k}(v)}(\phi^{v})$ is an
isomorphism.
Reduction process
Let $\mathscr{M}$ be the minimal model of $\phi$. We set
$\mathscr{M}[P]:=\text{Ker}(\phi_{P}:\phi\rightarrow\phi)$. We suppose
moreover that $P\mathcal{O}_{K}=\mathscr{P}\mathscr{P}^{\prime}$. Then it is
easy to see:
$\mathscr{M}[P](\bar{A})=\mathscr{M}[P](\mathbb{C}_{\infty})\cong\mathcal{O}_{K}/P\mathcal{O}_{K}=\mathscr{P}/P\mathcal{O}_{K}\oplus\mathscr{P}^{\prime}/P\mathcal{O}_{K}\cong\mathcal{O}_{K}/\mathscr{P}\oplus\mathcal{O}_{K}/\mathscr{P}^{\prime}.$
There is a natural morphism
$\theta:\mathscr{M}[P](\bar{A})\rightarrow\phi^{v}[P](\overline{\mathbb{F}_{q}})$
by taking reduction at $v\in\text{Spec}(R)$ such that $v$ lies over $P$.
###### Theorem A.4.
If we assume further that
$\pi\in\mathscr{P}\subset\mathcal{O}_{K}=\text{End}_{F}(\phi)=\text{End}_{R}(\mathscr{M})$
is the lifting of the Frobenius element $\bar{\pi}$, then $\theta$ is a
surjection, and the kernel of $\theta$ is isomorphic to
$\mathcal{O}_{K}/\mathscr{P}$.
###### Proof.
Since $P\mathcal{O}_{K}=\mathscr{P}\mathscr{P}^{\prime}$, by Theorem A.2
$\phi^{v}[P](\overline{\mathbb{F}_{q}})$ is non-trivial and finite. We embed
$A$ into $\text{End}_{\mathbf{k}(v)}(\phi^{v})=\mathcal{O}_{K}$ via
$\phi^{v}$. As an $\mathcal{O}_{K}$-module, we have
$\phi^{v}[P](\overline{\mathbb{F}_{q}})\cong\mathcal{O}_{K}/I$ for some proper
ideal $I\subset\mathcal{O}_{K}$. Therefore, we have
$\phi^{v}_{P}\cdot\mathcal{O}_{K}/I=0$, which implies $P\in I$. It is clear
$\\#\\{\phi^{v}[P](\overline{\mathbb{F}_{q}})\\}<\\#\\{\mathscr{M}[P](\mathbb{C}_{\infty})\\}.$
So either $I=\mathscr{P}$ or $I=\mathscr{P}^{\prime}$. Since $\bar{\pi}$ acts
on $\phi^{v}[P](\overline{\mathbb{F}_{q}})$ non-trivially, we see
$I=\mathscr{P}^{\prime}$. Therefore, the kernel of $\theta$ is
$\mathscr{P}^{\prime}/P\mathcal{O}_{K}$ that is isomorphic to
$\mathcal{O}_{K}/\mathscr{P}$. ∎
###### Remark A.5.
1. (1)
If we identify
$\mathcal{O}_{K}=\text{End}_{F}(\phi)=\text{End}_{R}(\mathscr{M})$, then
$\mathscr{P}$ is the collection of isogenies whose reduction has linear
coefficient 0.
2. (2)
Another approach to Theorem A.4 using canonical subgroup of Drinfeld modules
has be shown to the author by Urs Hartl. The two approaches essentially have
the same core.
## References
* [1] M. Becker and S. Maclane, The minimum number of generators for inseparable algebraic extensions, Bulletin of the American Mathematical Society, 46 (1940), pp. 182–186.
* [2] E. Bombieri and W. Gubler, Heights in Diophantine Geometry, New Mathematical Monographs, Cambridge University Press, 2006.
* [3] F. Breuer, Sur la conjecture d’André-Oort et courbes modulaires de Drinfeld, PhD thesis, Université Denis Diderot, 2002.
* [4] F. Breuer, The André-Oort conjecture for products of Drinfeld modular curves, Journal für die reine und angewandte Mathematik, 2005 (2005), pp. 115–144.
* [5] F. Breuer, F. Pazuki, and M. H. Razafinjatovo, Heights and isogenies of Drinfeld modules, Acta Arithmetica, 197 (2021), pp. 111 – 128.
* [6] M. Brown, Singular moduli and supersingular moduli of Drinfeld modules., Inventiones Mathematicae, 110 (1992), pp. 419–440.
* [7] Y. Buli, P. Habegger, and L. Kühne, No singular modulus is a unit, International Mathematics Research Notice, 2020 (2020), pp. 10005–10041.
* [8] L. Clozel and E. Ullmo, Équidistribution des points de Hecke, Contributions to automorphic forms, geometry, and number theory, John Hopkins Univ. Press, Baltimore, MD, 2004, pp. 193–254.
* [9] D. Goss, Basic Structures of Function Field Arithmetic, Springer-Verlag Berlin Heidelberg, 1998.
* [10] P. Habegger, Singular moduli that are algebraic units, Algebra and Number Theory, 9 (2015), pp. 1515 – 1524.
* [11] D. Hayes, Explicit class field theory in global function fields, Studies in algebra and number theory, Academic Press, New York-London, 1979, pp. 173–217.
* [12] Y. Ihara, On the Euler-Kronecker constants of global fields and primes with small norms, Algebraic Geometry and Number Theory: In Honor of Vladimir Drinfeld’s 50th Birthday, Birkhäuser Boston, Boston, MA, 2006, pp. 407–451.
* [13] L. Illusie, Déformations de groupes de Barsotti-Tate, Astérisque, 127 (1985), pp. 151 – 198.
* [14] T. Lehmkuhl, Compactification of the Drinfeld Modular Surfaces, vol. 197, Memoirs of the American Mathematical Society, 2009.
* [15] Y. Li, Singular units and isogenies between CM elliptic curves, Compositio Mathematica, 157 (2021), p. 1022–1035.
* [16] Y. Nakkajima and Y. Taguchi, A generalization of the Chowla-Selberg formula, Journal für die reine und angewandte Mathematik, 419 (1991), pp. 119 – 124.
* [17] M. Raynaud, Schémas en groupes de type (p,…,p), Bulletin de la Société Mathématique de France, 102 (1974), pp. 241–280.
* [18] , Hauteurs et isogénies, Astérisque, 127 (1985), pp. 199 – 234.
* [19] M. Rosen, Number Theory in Function Fields, Graduate Texts in Mathematics, Springer New York, 2002.
* [20] Y. Taguchi, Semisimplicity of the Galois represebtations attached to Drinfeld modules over fields of “finite characteristics”, Duke Mathematical Journal, 62 (1991), pp. 593–599.
* [21] , Semi-simplicity of Galois represebtations attached to Drinfeld modules over fields of “infinite characteristics”, Journal of Number Theory, 44 (1993), pp. 292–314.
* [22] F.-T. Wei, Kronecker limit formula over global function fields, American Journal of Mathematics, 139 (2017), pp. 1047 – 1084.
* [23] F.-T. Wei, On Kronecker terms over global function fields, Inventiones Mathematicae, 220 (2020), pp. 847–907.
Zhenlin Ran, University of Newcastle, Australia
E-mail address<EMAIL_ADDRESS>
|
=18pt plus1pt
CHAPTER: INTRODUCTION
This work is about the geometry of moduli spaces of vortices and antivortices
on a Riemann surface $\surface$. We are interested mostly in the
gauged $O(3)$ Sigma
where the fields are represented by a connection
$\gp$ and a section $\hf$ of a fibre bundle
with fibres
diffeomorphic to $\mathbb{P}^1$, the Riemann sphere.
We say $\hf$ is a Higgs field with target the Riemann sphere.
Static solutions of the field equations modulo gauge equivalence
form the moduli space of vortices and antivortices, each solution is
determined by the cores of the fields:
the preimages of the north pole (vortex points) and the south pole (antivortex
points). It can be proved the total number of the cores is enumerable and if
$\surface$ is compact, it is finite. We will assume without
loss of generality this is the
case, even though $\surface$ can be the complex plane. The dynamics
of slowly varying fields can be described by geodesic motion of curves
on the moduli space [Manton, 1982] with
a metric called the $\Lsp^2$ metric. This metric is Kähler and well
understood for the moduli space of vortices of the Ginzburg-Landau
functional, in which case it is known that the moduli space is a complete
metric space and if the ambient surface is compact the moduli space is also
compact, hence of finite volume.
The $O(3)$ Sigma model we will study is
asymmetric, vortices and antivortices have different
effective mass, moreover,
the existence of two types of cores
means vortices and
antivortices cannot coalesce, therefore, a natural question is if
the moduli space is still complete.
Another question we address is how the asymmetry
affects the volume of the moduli space. These questions were addressed for the
symmetric case in the
reference [Romão and Speight, 2020]. The techniques used in the reference however
do not apply in general, we developed analytical tools to
extend the results to the asymmetric case.
Later, we add a Chern-Simons deformation to the model and
describe the change in the dynamics of the fields on the moduli space.
The deformation is tuned by means of a deformation constant
$\kappa$ which we assume small.
It turns out that the dynamics of the theory is described by
geodesic motion perturbed with a connection term proportional to
$\kappa$, i.e. a term
dependent on the velocity of the cores. Our model resembles the model of Kim and
Lee [Kim and Lee, 2002] with the difference that the target is the
sphere and there are two types of cores to consider.
It is well known for several related models with Chern-Simons deformations
that multiple solutions of the field equations occur.
We study the problem of existence and multiplicity of
solutions to the field equations of the deformed
$O(3)$ Sigma model, the main result is that even though
multiple solutions of the equations can exist,
there is a minimal deformation, such that
no matter which configuration of vortices and antivortices on the
moduli space we choose, we can find exactly one
solution close to the undeformed solution of the
$O(3)$ Sigma model.
We conclude with a description of the chapters of the thesis.
In chapter <ref> we describe the ideas of localization in
abstract terms. Our approach is general and suits equally well
Ginzburg-Landau vortices as well as the $O(3)$ Sigma model, with the benefit
that it
makes clear what we mean by adding a Chern-Simons term.
We also present
analytical results that are common to other parts of the next
In chapter <ref> we focus on the $O(3)$ Sigma
model on the
euclidean plane. We study asymmetric vortex-antivortex pairs,
supporting our
analysis with numerical evidence of the behaviour of colliding
vortex-antivortex pairs.
We compute the metric on the moduli space of vortex-antivortex
pairs numerically
and use this computations to study the scattering of
approaching cores.
The main result is theorem <ref>
says that the moduli space is incomplete.
In chapter <ref> we move to a compact ambient
The main results are the incompleteness of the moduli space of
vortex-antivortex pairs,
theorem <ref>,
and the computation of the volume of
the moduli space for the round sphere and for flat tori in
theorem <ref>, confirming
a general conjecture by Romão-Speight [Romão and Speight, 2020]
in these cases.
Chapter <ref> is devoted to the study of Chern-Simons
deformations on compact surfaces.
We prove the existence of multiple solutions for small
deformations of the $O(3)$ Sigma model if the number of vortices and
antivortices is different and find bounds for
the deformation constant.
We also solve the
field equations numerically on the sphere for two configurations of
vortices and antivortices at antipodal positions.
The main result is theorem <ref>,
describing the behaviour of the solutions to the field
We finalise the chapter
applying the localization technique to vortices of the Ginzburg-Landau
model and vortices/antivortices of the $O(3)$ Sigma model, both with a
Chern-Simons deformation. We found that dynamics is deviated from
geodesic motion by a connection term consistent with
previous results of
Kim-Lee [Kim and Lee, 2002] and Collie-Tong [Collie and Tong, 2008],
and compared our result with theirs.
CHAPTER: PRELIMINARIES
This chapter is for basic definitions and results of field theory
that we will use in the successive. To study the geometry
of the moduli space of vortices we need several analytical tools, this
chapter is intended to be a bridge between field theory and analysis.
In section <ref> we introduce the $O(3)$ Sigma model, which
will play a central role all along the thesis.
In section <ref> we discuss a localization formula
for the $O(3)$ Sigma model, we compute a metric for the moduli space
of vortices and antivortices, the $\spL^2$ metric,
prove that
it is Kähler.
Section <ref> is about the analytic properties of
the Taubes equation, this is the elliptic PDE that guarantees the
existence of the moduli space of vortices and antivortices. Several
theorems of analysis are introduced in this section to keep them
collected in the same place for further reference.
In subsection <ref> we prove that the
solution to the Taubes equation depends differentiably on the
position of the vortices and antivortices.
In section <ref> we state less known theorems of functional
analysis about compact non-linear operators that we will need later.
ASYMMETRIC VORTEX-ANTIVORTEX SYSTEMS IN THE EUCLIDEAN PLANE
CHAPTER: ASYMMETRIC VORTEX-ANTIVORTEX PAIRS ON A COMPACT SURFACE
In this chapter we study vortex-antivortex systems on a compact surface.
We aim to prove that the moduli space is
incomplete and to compute the volume of the moduli space for the
round sphere and flat tori. On a general compact domain, the problem
of the statistical mechanics of Ginzburg-Landau vortices was
addressed by Manton [Manton, 1993] and
by Manton-Nasir [Manton and Nasir, 1999].
As shown
in [Manton, 1993], it can be described if we know the volume
of the moduli space. For
the abelian $O(3)$ Sigma model however, the problem of the volume of the moduli
space is constrained by the fact that vortices and antivortices cannot
coalesce, however, computing the volume is necessary for the partition function
of a gas of BPS vortices
[Romão and Speight, 2020, Manton and Nasir, 1999, Manton, 1993]. There is a
conjectured formula for the volume by Speight and Romão that depends on
topological data, the
volume of the domain,
$\tau$ and the size of the sets $\vset$, $\avset$ of
core positions [Romão and Speight, 2020]. The content of the chapter is as follows.
In section <ref>, we prove that the Taubes equation has
exactly one solution for any $\tau \in (-1, 1)$.
The main result of section <ref>
is theorem <ref>
which asserts that the moduli
space of vortex-antivortex pairs is incomplete. We prove the theorem
after proving several lemmas necessary to bound the derivatives of solutions
to the Taubes equation.
In section <ref> we compute the volume of
the moduli space of vortex-antivortex pairs for the round sphere and
flat tori and compare our results with the conjecture.
§ EXISTENCE OF VORTICES
In this section we will prove the existence of solutions to the
Taubes equation
on a compact surface. In [Sibner et al., 2000]
Sibner-Signer-Yang proved existence and
uniqueness of solutions of the gauged $O(3)$ Sigma model on a compact manifold
for $\tau = 0$.
We prove the following generalisation of their results.
On any compact
Riemann surface there exists exactly one
solution $u$ to the
Taubes equation (<ref>), provided the
\begin{equation}
\label{eq:vav-size-constraint}
- \frac{1 + \tau}{2\pi}\abs{\surface}
k_+ - k_-
\frac{1 - \tau}{2\pi}\abs{\surface}
\end{equation}
holds. Moreover,
$u$ is of class $C^2$ except for the core positions.
We prove the theorem at the end of the section. The inequality
(<ref>) is a Bradlow type restriction
[Bradlow, 1990], constraining the relative number of vortices and
antivortices on a compact surface.
It arises naturally from the second Bogomolny
equation (<ref>), since the total magnetic flux is,
\begin{align}
2\pi(k_+ - k_-) &=
\int_{\surface} B \nonumber\\
&= \int_\surface \lproduct{N,\hf}\,\vol -
\tau\,|\surface|,
\end{align}
where $N$ is the north pole section on the target sphere
and hence $\lproduct{N,\phi}\in [-1, 1]$, it follows
that (<ref>) is a necessary condition for a pair
$(\hf, \gp)$ of a field and a connection to be a solution to the Bogomolny
We will define
the function $\Fstable: \reals \to \reals$,
\begin{equation}
\Fstable(t) = 2 \left(
\frac{e^t - 1}{e^t + 1} + \tau
\right),
\end{equation}
and the constant,
\begin{equation}
\label{eq:fstable-pm-infty}
\Fstable^{\pm\infty} = 2(\pm 1 + \tau),
\end{equation}
in order to simplify notation in the proof of theorem <ref>.
Let us define
$\Fstable_0: \reals \to \reals$ as the function,
\begin{equation}
\label{eq:param-fstable-sphere}
\begin{aligned}[b]
\Fstable_0(t) &= 2\brk(\frac{e^t - 1}{e^t + 1} + \tau) + \frac{4\pi(k_+ -
k_-)}{\abs\surface }\\
&= \frac{4 e^t}{e^t + 1} - C_0,
\end{aligned}
\end{equation}
where the constant $C_0$ is,
\begin{align}
C_0 = 2(1 - \tau) - \frac{4\pi}{\abs\surface}(k_+ - k_-).
\end{align}
For a given configuration of non-coalescent vortices, recall the function
$v : \surface \to \reals\cup\set{\pm\infty}$, defined on
equation (<ref>), if $u$ is the solution of the Taubes equation, and
we define $\tilde h = u - v$, then
the regularized Taubes equation on a compact surface,
equation (<ref>), is equivalent to,
\begin{align}
\label{eq:regular-taubes-bps}
-\laplacian \tilde h = \Fstable_0(v + \tilde h).
\end{align}
Equation (<ref>) shows why Bradlow's bound is
necessary: If a smooth solution exists, by the divergence
theorem a necessary condition for $C_0$ is,
\begin{equation}
\label{eq:abstract-c0-cond}
C_0 = \frac{1}{|\surface|}\,\int_\surface\frac{4\,e^{v + \tilde h}}
{e^{v + \tilde h} + 1}\,\vol \in [0, 4],
\end{equation}
Bradlow's bound is equivalent to
(<ref>). Let
\begin{equation}
\label{eq:Xsp-def}
\Xsp = \left\{
u \in \Hsp^1(\surface) \,:\, \int_\surface u \,\vform = 0
\right\}
\end{equation}
be the subspace of Sobolev's space $\Hsp^1(\surface)$ of functions of zero
average. Since $\surface$ is compact, $\Hsp^1(\surface)$ can be
decomposed as
\begin{equation}
\label{eq:cann-decomp}
\Hsp^1(\surface) = \Xsp \oplus \reals.
\end{equation}
Any $h \in\Hsp^1(\surface)$ can be decomposed as a pair
$(u, \ctilde) \in \Xsp\times\reals$, such that $h = u +
\ctilde$. Hence, $u$ is a solution to the equation,
\begin{equation}
\label{eq:htilde}
-\laplacian u = \Fstable_0(v + u + \ctilde).
\end{equation}
We will use
Leray-Schauder theory to prove existence of solutions to the Taubes
equation as in the proof of Sibner et
al. [Sibner et al., 2000] for $\tau = 0$. Given $\htilde \in \Xsp$, the
\begin{equation}
\label{eq:integral-eqn}
c \mapsto \int_{\surface} \Fstable_0(v + \htilde + c) \,\vol,
\end{equation}
is a well defined, monotonous, continuous function. By Bradlow's
bound, there exists a unique number
$\ctilde$ such that
\begin{equation}
\label{eq:ctilde-defs}
\int_{\surface} \Fstable_0(v + \htilde + \ctilde) \,\vol = 0.
\end{equation}
The function $ \mathcal{C}: \Xsp \to \reals$, $\mathcal{C}(\tilde h) = \tilde
sequentially continuous in
We will highlight the steps different from [Sibner et al., 2000] in the
general case. If $\htilde_n \wto \htilde_0$ in $\Xsp$, then
$\htilde_n$ is a bounded sequence in $\Xsp$, and by the Rellich lemma,
after passing to a sub-sequence if necessary, we can assume
$\htilde_n \to \htilde_0$ in $\Lsp^p$ for $p \geq 1$. Let $\ctilde_n =
\ctilde(\htilde_n)$, $\ctilde_0 = \ctilde(\htilde_0)$ and assume
towards a contradiction that $\ctilde_n$ does not converge to
$\ctilde_0$. In this case we can assume the existence of a constant
$\epsilon_0$ such that,
\begin{equation}
\label{eq:cn-conv-contradiction}
\abs{\ctilde_n - \ctilde_0} \geq \epsilon_0,
\end{equation}
for all $n$. We claim the sequence $\left\{\ctilde_n\right\}$
is bounded. Assume the contrary, after passing to a sub-sequence if
necessary, we can assume the limit
$\ctilde_n \to \infty$. Let $K$ be any bound
for $\Fstable_0$. By Egorov's theorem [Lieb, 1997]
and the strong convergence in $\Lsp^p$,
there exists a measurable set $\surface_{\epsilon}$ and a constant
$K_{\epsilon}$, such that $\abs{\surface_{\epsilon}} < \epsilon
K^{-1}$, the sequence $\htilde_n$ converges uniformly to $\htilde_0$
in $\surface \setminus \surface_{\epsilon}$ and $\abs{\htilde_n} \leq
K_{\epsilon}$ in $\surface\setminus\surface_{\epsilon}$.
On the one hand,
the equality
\begin{equation}
\int_{\surface\setminus\surface_{\epsilon}}\Fstable_0(v +
\ctilde_n + \htilde_n) \,\vol =
-\int_{\surface_{\epsilon}}\Fstable_0(v + \ctilde_n +
\htilde_n) \,\vol,
\end{equation}
\begin{equation}
\label{eq:egorov-integral-impl}
\abs*{\int_{\surface\setminus\surface_{\epsilon}}\Fstable_0(v +
\ctilde_n + \htilde_n) \,\vol } \leq \epsilon,
\end{equation}
and on the other hand, by monotony of $\Fstable_0$,
\begin{equation}
\label{eq:monotony-fstable0-integral}
\int_{\surface\setminus\surface_{\epsilon}}\Fstable_0(v + \ctilde_n
- K_{\epsilon}) \,\vol \leq
\int_{\surface\setminus\surface_{\epsilon}}\Fstable_0(v + \ctilde_n
+ \htilde_n) \,\vol.
\end{equation}
Taking the limit as $n \to \infty$, from these two equations
we have,
\begin{equation}
\label{eq:int-finty-surface-minus-eps}
(\Fstable^{\infty} - C_0)(\abs\surface - \abs{\surface_{\epsilon}}) \leq \epsilon.
\end{equation}
\begin{equation}
\label{eq:fstable-inf-bound-contradiction}
(\Fstable^{\infty} - C_0) \abs\surface \leq \epsilon +
K\abs{\surface_{\epsilon}} < 2\epsilon,
\end{equation}
a contradiction since $\epsilon$ is arbitrary. A similar
argument shows $\ctilde_n$ is bounded below. Therefore,
$\ctilde_n$ is a bounded sequence of real numbers. By the
Bolzano-Weierstrass theorem, we can assume towards a contradiction
$\ctilde_n \to \ctilde$, but $\ctilde \neq \ctilde_0$
by (<ref>). Let
\begin{equation}
\label{eq:alpha-def}
\alpha = \abs*{
\int_{\surface}\Fstable_0(v + \htilde_0 + \ctilde) \,\vol
} > 0,
\end{equation}
bearing in mind the definition of $\htilde_n$,
\begin{multline}
\alpha = \abs*{
\int_{\surface}\Fstable_0(v + \htilde_0 + \ctilde)
- \Fstable_0(v + \htilde_n + \ctilde_n)
\,\vol
\leq
\sup_{t \in \reals}{\set{\Fstable'(t)}}\cdot\left(
\abs{\ctilde - \ctilde_n}\cdot\abs{\surface} + C\,\norm{\htilde_0 -
\htilde_n}_0 \cdot \abs{\surface}^{1/2}
\right) \to 0.\label{eq:alpha-bound}
\end{multline}
Hence $\alpha = 0$, a
contradiction. Therefore (<ref>) is
false and $\ctilde_n \to \ctilde_0$. This proves the lemma.
Let us consider the operator $T: \Xsp \to \Xsp$, mapping each
$\htilde\in\Xsp$ to the weak solution $H \in \Xsp$ of the equation
\begin{equation}
\label{eq:T-op}
-\laplacian H = \Fstable_0(v + \ctilde + \htilde).
\end{equation}
Given that $\int_\surface \Fstable_0(v + \ctilde + \htilde)\,\vol = 0$,
existence of a weak $\Hsp^1$ solution to (<ref>) is a well established
analysis fact <cit.>, moreover, any two weak
solutions to the equation differ by a constant, by taking $H \in \Xsp$ we
guarantee it is unique.
Recall a compact operator is an operator that maps bounded sequences
to sequences with convergent subsequences. We aim to use Schäfer's
alternative, theorem <ref>, to prove $T$
has a fixed point.
The operator $T:\Xsp \to \Xsp$ is compact in the strong topology of
$\Xsp$ as a subspace of $\Hsp^1(\surface)$.
Let $\{\tilde h_n\} \subset \Xsp$ be a bounded sequence, after
passing to a subsequence if necessary, we can assume $\htilde_n \wto
\htilde_0$ in $\Xsp$ and strongly in $\Lsp^2$. Let $H_n = T\htilde_n$,
$n \geq 0$, by lemma <ref> $\ctilde_n \to
\ctilde_0$. Moreover,
\begin{align}
\nonumber
\norm{\grad H_n - \grad H_0}_{\Lsp^2}^2
&= \int_{\surface} (H_n - H_0)\, \laplacian (H_n - H_0)\,\vol\\
\nonumber
&= \int_{\surface} (H_n - H_0)\left(\Fstable(v + \ctilde_n + \htilde_n)
- \Fstable(v + \ctilde_0 + \htilde_0)\right) \,\vol\\
\nonumber
&\leq \sup_{t \in \reals} \set{F'(t)}\,\int_{\surface} \left(
\abs{\ctilde_n - \ctilde_0} + \abs{\htilde_n - \htilde_0}\right)
\abs{H_n - H_0}\,\vol\\
&\leq \sup_{t \in \reals} \set{F'(t)}\, \left(
\abs{\ctilde_n - \ctilde_0}\cdot\abs{\surface}^{1/2} + \norm{\htilde_n -
\htilde_0}_{\Lsp^2}
\right) \norm{H_n - H_0}_{\Lsp^2}.
\end{align}
The last inequality is a consequence of the Cauchy-Schwarz
inequality. By the Poincaré inequality, there are constants $C_1$,
$C_2$ such that
\begin{equation}
\label{eq:Hn-H0-X-convergence}
\norm{H_n - H_0}_{\Hsp^1} \leq C_1\abs{\ctilde_n - \ctilde_0}
+ C_2\norm{\htilde_n - \htilde_0}_{\Lsp^2} \to 0.
\end{equation}
This proves compactness of $T$.
Let us consider the set
\begin{equation}
\label{eq:wt-def}
S = \set{\htilde \in \Xsp\, : \,
\exists\, t \in [0, 1]\;s.t.\;
\htilde = t\cdot
\end{equation}
If $\htilde \in S$, then it is a solution of the equation,
\begin{equation}
\label{eq:htilde-t-sol}
\laplacian \htilde = t \Fstable_0(v + \ctilde + \htilde),
\end{equation}
where $\tilde c = \mathcal C(\tilde h)$ was defined on
lemma <ref>.
By the Cauchy-Schwarz inequality,
\begin{equation}
\begin{aligned}
\norm{\grad\htilde_t}_{\Lsp^2}^2 &= \lproduct{\tilde h_t, \laplacian
\tilde h_t}
\leq C
\int_{\surface}\abs{\htilde_t} \,\vol
\leq C\,
\abs{\surface}^{1/2}\,\norm{\htilde_t}_{\Lsp^2}.
\end{aligned}
\end{equation}
By the Poincaré
inequality we conclude the existence of a constant $C$ such that
\begin{equation}
\label{eq:ht-uniform-bound}
\norm{\htilde_t}_{\Hsp^1} \leq C.
\end{equation}
Since $S$ is bounded, by Schäfer's alternative there is a fixed point
$\htilde$ of $T$. Let $h = \htilde + \ctilde$, where $\ctilde = \mathcal
C(\tilde h)$, then $h$ is a weak
solution to the regularised Taubes equation. By the elliptic
estimates $h$ is also a strong solution in $\Hsp^2$. We follow a
bootstrap argument to prove $h \in C^2$: By Sobolev's embedding we
know $h$ is continuous, hence $h \in \Lsp^p$ for any $p \geq
1$. By (<ref>) and the elliptic estimates $h
\in \Wsp^{2,p}$ for some $p > 2$, once more by Sobolev's embedding $h
\in C^1$. Let $u = h + v$, the derivative $d h \in
\Gamma(T^{*}\surface)$ is a weak solution of the linearized equation,
\begin{align}
-\laplacian\,d h = \frac{4\,e^u}{(e^u + 1)^2}\,d h +
\frac{4\,e^u}{(e^u + 1)^2}\,d v.
\end{align}
The potential function $e^u(e^u + 1)^{-2}$ is continuous and with
zeros of the same order than the singularities of $d v$ at the cores,
hence $\laplacian(d h) \in \Lsp^p$, $p > 2$. Since $d h$ is
continuous, it is also an $\Lsp^p$ form. By the elliptic estimates and
Sobolev's embedding we conclude $h \in C^2$. Since $\Fstable$
is monotonous, $h$ is unique by the strong maximum
principle. Finally, $u$ is the necessarily unique solution to the
Taubes equation.
§ INCOMPLETENESS OF THE MODULI SPACE
In [Romão and Speight, 2020] Romão and Speight prove that the moduli space of
vortex-antivortex pairs on the sphere is incomplete. In this section we extend
their result to general $\tau$ on a compact manifold. In order to prove this,
we find bounds for
the derivatives $\partial_{z_j}\nabla h_\epsilon$ on a holomorphic chart,
where the cores are at positions $z_1, z_2$.
Let $\mu = \log\, (1 - \tau) - \log\, (1 + \tau)$, first we prove a pair of
technical lemmas.
Let $\diag$ be the diagonal set of $\surface \times \surface$ and
let $\set{\vb x_n} \subset \moduli^{1,1}(\surface)$ be a
sequence such that $\vb x_n \to \vb x \in \diag$ in the product metric.
Let $\tilde h_n$ be the
solution of the regular Taubes equation corresponding to each
$\vb x_n$, then
$\tilde h_n \wto \mu$ in $\Hsp^1$ and $\tilde h_n \to \mu$
strongly in $\Lsp^2$.
Let $v_n = v_{\vb x_n}$ for
each point $\vb x_n$ in the given sequence. Let us
decompose each
solution to the regular Taubes equation as $\tilde h_n = u_n +
\tilde c_n \in \Xsp \oplus \reals$. We claim the sequence
$\set{ \tilde c_n}$ is bounded. Assume towards a contradiction
$\tilde c_n \to \infty$. Notice that in the vortex-antivortex
case the functions $\fnF$ and $\fnF_0$ coincide. We know that,
\begin{align}
-\laplacian u_n = \fnF( u_n + \tilde c_n +
\fnv_n).\label{eq:lap-h0n}
\end{align}
By the standard elliptic estimates, there is a constant $C$ such
\begin{align}
\norm{u_n}_{\Hsp^2} \leq C \norm{\laplacian u_n}_{\Lsp^2}.
\end{align}
Since $\fnF$ is a
bounded function, $\set{u_n}$ is bounded in $\Hsp^2$ and by
Sobolev's embedding also in $C^0$.
Assume $\vb x = (x_*, x_*)$ and notice that,
\begin{align}
|v_n(x)| = 4\pi |G(x, x_{1n}) - G(x, x_{2n})|,
\end{align}
where $\vb x_n = (x_{1n}, x_{2n})$,
since $G(x,y)$ is continuous away of the diagonal set,
$\fnv_n(x) \to 0$ for $x \neq x_*$, whence,
we also have the convergence,
\begin{align}
\fnF( u_n + \tilde c_n +
\fnv_n) \to 2(1 + \tau),
\end{align}
pointwise almost everywhere. Applying the dominated convergence theorem
equation (<ref>),
\begin{align}
\int_{\surface} \fnF( u_n + \tilde c_n +
\fnv_n) \,\vform = 0 \to 2(1 +
\tau)\abs\surface,
\end{align}
a contradiction. If $\tilde c_n \to -\infty$ a similar argument
holds. Therefore the sequence of averages $\tilde c_n$ is bounded,
$\{\tilde h_n\}$ is bounded in $C^0$.
Hence, the sequence is also bounded in $\Lsp^p$ for any
positive $p$. By the elliptic estimate
\begin{align}
\norm{\tilde h_n}_{\Hsp^2} \leq C
\pbrk{\norm{\laplacian \tilde
h_n}_{\Lsp^2} + \norm{\tilde h_n}_{\Lsp^2}},
\end{align}
$\{\tilde h_n\}$ is also bounded in $\Hsp^1$. By the Alaoglu
and Rellich theorems, after passing to a subsequence if
necessary, we can assume $\tilde h_n \wto h_{*} \in \Hsp^1$ and
strongly in $\Lsp^2$. We claim that $h_*$ is the constant
function $\mu$. To see this, let $\varphi \in \Hsp^1$. From the
regularized Taubes equation we have,
\begin{align}
\lproduct{\tilde h_n, \varphi}_{\Hsp^1}
&= \lproduct{\tilde h_n,
\varphi}_{\Lsp^2} + \lproduct{\grad \tilde h_n, \grad
\varphi}_{\Lsp^2},\nonumber \\
&= \lproduct{\tilde h_n,
\varphi}_{\Lsp^2} + \lproduct{\laplacian \tilde h_n,
\varphi}_{\Lsp^2}\nonumber \\
&= \lproduct{\tilde h_n,
\varphi}_{\Lsp^2} - \lproduct{\fnF(\tilde h_n + \fnv_n),
\varphi}_{\Lsp^2}.
\end{align}
Since $\tilde h_n \to h_{*}$ strongly in $\Lsp^2$, after passing
to a subsequence if necessary, we can assume
$\tilde h_n\to h_{*}$ pointwise almost everywhere. By the weak
convergence of $\tilde h_n$ in $\Hsp^1$, together with the
strong convergence in $\Lsp^2$ and the dominated convergence
\begin{align}
\lproduct{h_{*}, \varphi}_{\Hsp^1} &= \lim\,\lproduct{\tilde h_n,
\varphi}_{\Hsp^1}\nonumber\\
&= \lim \lproduct{\tilde h_n,
\varphi}_{\Lsp^2} - \lim \lproduct{\fnF(\tilde h_n + \fnv_n),
\varphi}_{\Lsp^2}\nonumber\\
&= \lproduct{h_{*},\varphi}_{\Lsp^2} - \lproduct{\fnF(h_{*}),
\varphi}_{\Lsp^2}.
\end{align}
From this equation, we infer
\begin{align}
\lproduct{\grad h_{*}, \grad\varphi}_{\Lsp^2}
= -\lproduct{\fnF(h_{*}), \varphi}_{\Lsp^2}.
\end{align}
Therefore, $h_{*}$ is a weak solution to the equation
\begin{align}
-\laplacian h_{*} = \fnF(h_{*}).
\end{align}
By elliptic regularity, $h_{*}$ is also a strong solution, and
by the maximum principle, $h_{*}$ is constant since $\fnF$ is an
increasing function. Since the only zero of $\fnF$ is at
$t = \mu$, we conclude $h_{*} = \mu$. If $\tilde h_{n_k}$ is any
subsequence of $\tilde h_{n}$, this argument shows it
has a subsequence weakly converging to $\mu$ in $\Hsp^1$ and
strongly in $\Lsp^2$, the
claim of the lemma follows.
$\tilde h_n \to \mu$ strongly in $\Wsp^{2,p}$ for any positive $p$.
We will prove that any subsequence of $\tilde h_n$ has another
subsequence converging to $\mu$ in $\Wsp^{2,p}$, implying the
lemma. To simplify
notation, we denote subsequences of $\tilde h_n$ by the
same symbol.
From the previous lemma, $\tilde h_n \to \mu$ strongly in
$\Lsp^2$. After passing to a subsequence if necessary, we can
assume that $\tilde h_n \to \mu$ pointwise almost
everywhere. We apply the
dominated convergence theorem to deduce the limit,
\begin{align}
\norm{\laplacian \tilde h_n}_{\Lsp^{p}}
= \norm{\fnF(\tilde h_n + \fnv_n)}_{\Lsp^p}
\to \norm{\fnF(\mu)}_{\Lsp^p} = 0.
\end{align}
If $p = 2$, by the standard elliptic estimates, there is a
constant $C$, such that,
\begin{align}
\norm{\tilde h_n - \mu}_{\Hsp^2} \leq C \pbrk{
\norm{\laplacian \tilde h_n}_{\Lsp^2} +
\norm{\tilde h_n - \mu}_{\Lsp^2}} \to 0.
\end{align}
By Sobolev's embedding, $\tilde h_n \to \mu$ uniformly in
$C^0$, hence also in $\Lsp^p$ for any positive $p$. We apply one
more time the elliptic estimate,
\begin{align}
\norm{\tilde h_n - \mu}_{\Wsp^{2,p}} \leq C \pbrk{
\norm{\laplacian \tilde h_n}_{\Lsp^p} +
\norm{\tilde h_n - \mu}_{\Lsp^p}} \to 0.
\end{align}
As a consequence of this lemma and Sobolev's embedding, we have
the convergence,
\begin{align}
\norm{\tilde h_n - \mu}_{C^1} \to 0,
\end{align}
for any arbitrary sequence $\set{\vb x_n} \subset \moduli^{1,1}(\surface)$,
such that $\vb x_n \to \vb x \in \Delta$. This proves the following
The limit,
\begin{align}
\lim_{d(x_1, x_2) \to 0}\norm*{\tilde h(x; x_1, x_2) - \mu
= 0,
\end{align}
holds, where $d(x_1, x_2)$ is the Riemannian distance in $\surface$.
Let $\surface^2_{\diag} = (\surface\times\surface)\setminus\diag$
endowed with the product metric. As differentiable manifolds,
$\moduli^{1,1}$ and
$\surface^2_\diag$ are equivalent. In what follows, we will consider
$\tilde h$ as a
function $\surface \times \surface^2_{\diag} \to \reals$. Let $U \subset
\surface$ be an open and dense subset and let
$\varphi:U \to V \subset \cpx$ be a
holomorphic chart. In what follows we denote points on the surface as $x$
and points on $\cpx$ as $z$, so $z = \varphi(x)$ for $x \in U$. We also
assume vortices
and antivortices are both located in $U$, such that up to a holomorphic
$\tilde h: \surface \times V^2_\diag \to \reals$, where $V^2_\diag
= V^2\setminus \diag_V$ and
$\diag_V \subset \cpx^2$ is the diagonal set. On this chart partial
derivatives $\partial_{z_j}\tilde h$ are well defined functions
\begin{align}
\partial_{z_j}\tilde h : \surface \times V^2_\diag \to \cpx.
\end{align}
We denote the covariant derivative and Laplacian with respect to the first
variable by $\grad$ and $\laplacian$ and emphasize that
the metric on $V^2_\diag$ is the push forward of the metric induced
by the surface. Our aim is
to estimate the rate at which the second
derivatives $\grad \partial_{z_j}\tilde h$ grow as a sequence $\vb z_n \in
V^2_\diag$ diverges to the diagonal set. This will
allow us to prove that the moduli space is incomplete.
Since, $\laplacian$ and $\partial_{z_j}$ commute, $\partial_{z_j}\tilde h$
is the solution to the elliptic problem,
\begin{align}
-\laplacian\partial_{z_j}\tilde h = \fnV(h)
\partial_{z_j} \tilde h + \sign_j\,\fnV(h)\,\partial_{z_j} \fnv_j,
\end{align}
where $v_j(x) = 4\pi \,G(x, \varphi^{-1}(z_j))$.
Let $\dist_j(x) = \dist(x, x_j)$, $x_j = \varphi^{-1}(z_j)$, we know
there is a uniform constant $C$,
such that the derivative of Green's function is
bounded [Aubin, 2013],
\begin{align}
\abs{\grad G(x, x_j)} < \frac{C}{\dist_j}, &&
\abs{\grad_2 G(x, x_j)} < \frac{C}{\dist_j},
\end{align}
where $\grad_2 G$ is the covariant derivative with respect to the
second variable.
Recall in holomorphic coordinates the metric is
$e^{\Lambda(z)}\abs{dz}^2$, hence, if $z_j$ is restricted to a
bounded domain,
\begin{align}
\abs{\del_{z_j}\fnv_j} \leq 4\pi e^{-\Lambda(z_j)}\,\abs{\grad_2
< \frac{C}{d_j}.
\end{align}
For any positive constant
$C_1$, there is another constant $C$, such that, for all
$x, x_1, x_2 \in U$,
\begin{align}
\frac{\dist_{12}^2}{C_1\dist_1^2 + \dist_2^2}
& \leq C,\label{eq:vol-1}
\\
\frac{\dist_j\dist_k^2}{(C_1\dist_1^2 + \dist_2^2)^2}
&\leq \frac{C}{\dist_{12}},\label{eq:vol-2}
\end{align}
where $\set{\dist_j, \dist_k} = \set{\dist_1, \dist_2}$
and $\dist_{12}=\dist(x_1, x_2)$.
By the triangle inequality and Cauchy-Schwarz,
\begin{align}
\dist_{12} \leq \dist_1 + \dist_2 \leq C\,(\dist_1^2 + \dist_2^2)^{1/2},
\end{align}
on the other hand, any two norms in a finite dimensional vector
space are equivalents, hence, there is another constant such
\begin{align}
(\dist_1^2 + \dist_2^2)^{1/2} \leq C\,(C_1\dist_1^2 + \dist_2^2)^{1/2},
\end{align}
from these two inequalities we obtain the first claim of the
lemma. For the second claim, it is enough to prove that the
\begin{align}
\frac{\dist_1\dist_2^2}{(C_1\dist_1^2 + \dist_2^2)^2}
\leq \frac{C}{\dist_{12}},
\end{align}
holds, the remaining case being equivalent to this one after
relabelling $\dist_1$ and $\dist_2$. Let us note that since,
\begin{align}
\dist_1\dist_2 \leq \half (\dist_1^2+\dist_2^2)
\leq C\,(C_1\dist_1^2 + \dist_2^2),
\end{align}
is sufficient to prove that,
\begin{align}
\frac{\dist_2}{C_1\dist_1^2 + \dist_2^2}
\leq \frac{C}{\dist_{12}}.\label{eq:vol-dist2-c1norm-bound}
\end{align}
$\dist_2 \leq \dist_1$, by the triangle inequality we have,
\begin{align}
\dist_2\dist_{12}
&\leq \dist_1\dist_2 + \dist_2^2\nonumber\\
&\leq \dist_1^2 + \dist_2^2\nonumber\\
&\leq C\,(C_1\dist_1^2 + \dist_2^2),
\end{align}
hence (<ref>). On the other hand, if
$\dist_1\leq \dist_2$, repeating the
previous step, we find that
\begin{align}
\dist_1\dist_{12} \leq C\,(C_1\dist_1^2 + \dist_2^2),
\end{align}
this inequality, together with (<ref>) and the triangle
inequality, implies,
\begin{align}
\frac{\dist_2}{C_1\dist_1^2+\dist_2^2}
\leq \frac{\dist_1}{C_1\dist_1^2+\dist_2^2}
+ \frac{\dist_{12}}{C_1\dist_1^2+\dist_2^2}
\leq \frac{C}{\dist_{12}}.
\end{align}
In any case, we conclude that equation (<ref>) holds.
There is a constant $C$ such that for any pair of distinct
points $x_1, x_2 \in \surface$,
\begin{align}
\abs*{G(x_1,x_2) - \frac{1}{2\pi}\log\,\dist(x_1,x_2)} \leq C.
\end{align}
We cover $\surface$ with a finite cover of metric disks
$\disk_{R_j/2}(p_j)$ such that $R_j < \delta$, where $\delta$ is the
injectivity radius of the metric and for each disk there is a
chart $\varphi_j: U_j \to \cpx$, $\disk_{R_j}(p_j) \subset
U_j$. Let $R = \min\set{R_j}$, for any pair of distinct points $x_1,
\in \surface$,
such that $d(x_1, x_2) < R/2$,
there is a disk such that $x_1, x_2 \in \disk_{R_j}(p_j)$.
For any disk in the cover, let $R_j'$ be a positive radius, such that,
\begin{align}
|\varphi_j(x) - \varphi_j(p_j)| < R_j', \qquad
\forall x \in \disk_{R_j}(p_j).
\end{align}
Let $z_j = \varphi_j(p_j)$ and let us denote by $D_{R'_j}(z_j) \subset
\cpx$ the
holomorphic disk of radius $R'_j$ centred at $z_j$. For any small
$\epsilon > 0$ there are continuous functions $\tilde G_j: D_{R'_j +
\epsilon}(z_j) \times D_{R'_j + \epsilon}(z_j) \to \reals$ such that
if $x_1, x_2 \in \disk_{R_j}(p_j)$,
\begin{align}
G(x_1, x_2) = \frac{1}{2\pi} \log\,\abs{\varphi_j(x_1) -
\varphi_j(x_2)}
+ \tilde G_j(\varphi_j(x_1), \varphi_j(x_2)).
\end{align}
If $\exp \Lambda_j(z)$ is the conformal factor of the metric in the
chart $\varphi_j$, let
\begin{equation}
\begin{aligned}
M_j &= \max \set{e^{\Lambda_j(z)/2}\,:\, z \in
\overline{D_{R'_j}(z_j)}},
\\
m_j &= \min \set{e^{\Lambda_j(z)/2}\,:\, z \in
\overline{D_{R'_j}(z_j)}},
\end{aligned}
\end{equation}
and $M = \max_j \set{M_j}$, $m = \min_j \set{m_j}$. Since each
$\disk_{R_j}(p_j)$ is geodesically convex, for any $x_1, x_2 \in
\disk_{R_j}(p_j)$,
\begin{align}
m\,\abs{\varphi_j(x_1) - \varphi_j(x_2)} \leq d(x_1, x_2)
\leq M\,\abs{\varphi_j(x_1) - \varphi_j(x_2)}.
\end{align}
Taking the log of this inequality we find a positive constant such that,
\begin{align}
\abs{\,d(x_1, x_2) - \log\,\abs{\varphi_j(x_1) - \varphi_j(x_2)}\,}
\leq C,
\end{align}
whenever $x_1, x_2 \in \disk_{R_j}(p_j)$. Since each function $\tilde
G_j$ is continuous in the compact set $\overline{D_{R'_j}(z_j)}$, we find
another constant such that,
\begin{multline}
\abs*{G(x_1, x_2) - \frac{1}{2\pi}\log\,d(x_1, x_2)} =\\
\abs*{\frac{1}{2\pi}\pbrk{\log\,\abs{\varphi_j(x_1) - \varphi_j(x_2)} -
\log\,d(x_1,
x_2)} + \tilde G_j(\varphi_j(x_1), \varphi_j(x_2))}
\leq C.
\end{multline}
This proves the inequality whenever $d(x_1, x_2) < R/2$. Since $G$ and
the distance function are continuous on the compact set,
\begin{align}
\set{(x_1, x_2) \in \surface \times \surface \,:\, d(x_1, x_2) \geq
\frac{R}{2}},\
\end{align}
we can find a second constant satisfying the inequality whenever
$d(x_1, x_2) \geq R/2$. Taking the maximum of both constants
concludes the lemma.
Let $D$ be any bounded domain on $\cpx$. For any $p > 0$, there is a
constant $C$, independent of $z_1, z_2 \in D$, $z_1 \neq z_2$ such
that, if $x_j = \varphi^{-1}(z_j)$,
\begin{align}
\norm{\fnV(h)\partial_{z_j} v_j}_{\Lsp^p} \leq \frac{C}{\dist(x_1, x_2)}.
\end{align}
By lemma <ref>, there is
a constant, such that for all $x, y \in \surface$, $x \neq y$,
\begin{align}
\abs*{G(x, y) - \frac{1}{2\pi} \log \dist(x,y)} \leq C.
\end{align}
\begin{align}
\abs{\fnV(h)\partial_{z_j}\fnv_j} = \abs*{
\frac{4e^{\fnv_1}e^{\fnv_2}e^{\tilde h}}{
(e^{\fnv_1}e^{\tilde h} + e^{\fnv_2})^2
} \partial_{z_j}\fnv_j} \leq
C \abs*{\frac{4 \dist_1^2\dist_2^2\,e^{\tilde h}}{
(\dist_1^2e^{\tilde h} + \dist_2^2)^2
} \frac{1}{\dist_j}},
\end{align}
where the constant depends on $D$.
Since $\tilde h$ is uniformly
bounded on $\surface$, there are constants $C$, $C_1$, such that
by lemma <ref>.
\begin{align}
\abs{\fnV(h)\partial_{z_j}\fnv_j} \leq
C \abs*{\frac{\dist_1^2\dist_2^2}{
(\dist_1^2C_1 + \dist_2^2)^2
} \frac{1}{\dist_j}}
\leq \frac{C}{\dist(x_1,x_2)},
\end{align}
inequality implies the claim.
The proof of the lemma depends only on properties of Green's
function, we could repeat the
proof of lemma <ref> using $\grad \fnv_j$
instead of $\partial_{z_j}v_j$ to prove for
any given domain $D\subset \cpx$ the existence of a constant,
independent of $z_1, z_2 \in D$ , such
\begin{align}
\norm{V(h)\grad v_j}_{\Lsp^p} \leq \frac{C}{d(x_1, x_2)}.
\end{align}
In the next lemmas we prove that the bilinear form,
\begin{align}
\Bop:\Hsp^1\times \Hsp^1 \to \reals,
\Bop(\phi,\psi) = \lproduct{\grad\phi, \grad\psi}^2_{\Lsp^2}
+ \lproduct{\fnV(h)\phi, \psi}_{\Lsp^2},
\end{align}
is coercive with a uniform coercivity constant.
If $\fnV_n: \surface \to \reals$ is a sequence of continuous,
uniformly bounded functions converging pointwise to the
continuous function $\fnV_*$, and $\phi_n\to \phi_*$ in $\Lsp^2$,
\begin{align}
\lproduct{\fnV_n, \phi_n^2}_{\Lsp^2} \to \lproduct{\fnV_*,
\phi_*^2}_{\Lsp^2}.
\end{align}
We have,
\begin{align}
\abs{\lproduct{\fnV_n,\phi_n^2}_{\Lsp^2} -
\lproduct{\fnV_*,\phi_*^2}_{\Lsp^2}}
\leq
\abs{\lproduct{\fnV_n,\phi_n^2 - \phi_*^2}_{\Lsp^2}}
\abs{\lproduct{\fnV_n - \fnV_*, \phi_*^2}_{\Lsp^2}}.
\end{align}
Since the functions $\fnV_n$ are uniformly bounded, there is a
constant $C$ such that,
\begin{align}
\abs{\lproduct{\fnV_n,\phi_n^2 - \phi_*^2}_{\Lsp^2}}
&\leq C\,\lproduct{
\abs{\phi_n - \phi_*},
\abs{\phi_n + \phi_*}}_{\Lsp^2}\nonumber\\
&\leq C\,\norm{\phi_n - \phi_*}_{\Lsp^2}\,\norm{\phi_n +
\phi_*}_{\Lsp^2},
\end{align}
by the convergence $\phi_n \to \phi_*$ in $\Lsp^2$, we obtain
the limit
\begin{align}
\abs{\lproduct{\fnV_n,\phi_n^2 - \phi_*^2}_{\Lsp^2}} \to 0.
\end{align}
Since there is a constant $C$ such that the functions
$(\fnV_n - \fnV_*)\phi_*^2$ are bounded by the measurable
function $C\phi_*^2$ and $\fnV_n - \fnV_* \to 0$ pointwise, by
the dominated convergence theorem,
\begin{align}
\abs{\lproduct{\fnV_n - \fnV_*, \phi_*^2}_{\Lsp^2}} \to 0.
\end{align}
\begin{align}
\abs{\lproduct{\fnV_n,\phi_n^2}_{\Lsp^2} -
\lproduct{\fnV_*,\phi_*^2}_{\Lsp^2}} \to 0,
\end{align}
concluding the proof of the lemma.
There is a positive constant $C$, independent of $(x_1, x_2) \in
\surface^2_\diag$, such
that for any $\phi \in \Hsp^1$,
\begin{align}
C\norm{\phi}^2_{\Hsp^1} \leq \Bop(\phi, \phi).
\end{align}
By the bilinearity of $\Bop$, it is sufficient to prove the
lemma assuming $\norm{\phi}_{\Hsp^1} = 1$.
Let us assume towards a contradiction the statement is false, in
this case there is a sequence
$(\phi_n,\vb{x}_n) \subset \Hsp^1\times \surface^2_{\diag}$,
with $\norm{\phi_n}_{\Hsp^1} = 1$, such that,
\begin{align}
\Bop(\phi_n, \phi_n) = \norm{\grad\phi_n}_{\Lsp^2}^2 +
\lproduct{\fnV_n, \phi_n^2}_{\Lsp^2} \to 0,
\end{align}
where $\fnV_n = \fnV(h_n)$ is the potential function determined
by $h_n$, the solution to the Taubes equation with data
$\vb{x}_n$. Since the functions $\fnV_n$ are non-negative,
\begin{align}
\norm{\grad\phi_n}_{\Lsp^2} \to 0,
\lproduct{\fnV_n, \phi_n^2}_{\Lsp^2} \to 0.
\end{align}
Passing to a subsequence if necessary, we can assume
$\phi_n \wto \phi_*$ in $\Hsp^1$ and strongly in $\Lsp^2$ and
$\vb{x}_n \to \vb{x}_{*}$ in $\surface\times \surface$. Since
the functions
\begin{align}
e^{\fnv_j}: \surface\times\surface \to \reals
\end{align}
are continuous and $\tilde h_n$ varies continuously with the
initial data, if $\vb{x}_{*} \not\in \diag$, we have the uniform
convergence $\fnV_n \to \fnV_* = \fnV(h_{*})$, where $h_{*}$ is
the solution to the Taubes equation determined by $\vb{x}_{*}$. On
the other hand, if $\vb{x}_{*} \in \diag$, we know that
$\tilde h_n \to \mu$ in $C^1$, hence, we have pointwise
$\fnV_n \to \fnV_{*} \equiv 4\exp(\mu) (\exp(\mu) + 1)^{-2}$. In
any case, by our previous lemma,
\begin{align}
\lproduct{\fnV_n, \phi_n^2}_{\Lsp^2} \to
\lproduct{\fnV_*, \phi_*^2}_{\Lsp^2},
\end{align}
but this limit is zero, hence $\phi_{*} = 0$ almost everywhere
and $\phi_n \to 0$ in $\Hsp^1$ strongly, a contradiction.
Let $D \subset V$ be any bounded domain. There is a positive constant
$C(D)$, such that
\begin{align}
\norm{\partial_{z_j}\tilde h}_{C^1} \leq \frac{C}{\dist_{12}},
\text{and}
\norm{\grad \tilde h}_{C^1} \leq \frac{C}{\dist_{12}},
\end{align}
for all $z_1, z_2 \in D$
with $z_1 \neq z_2$, where $\tilde h(x) = \tilde h(x; \varphi^{-1}(z_1),
\varphi^{-1}(z_2))$ and
$\dist_{12} = \dist(x_1, x_2)$.
$\partial_{z_j}\tilde h$ is a solution to the equation
\begin{align}
-\laplacian \partial_{z_j}\tilde h
= \fnV(h)\partial_{z_j}\tilde h + \sign_j\fnV(h)\partial_{z_j}\fnv_j.
\end{align}
By lemma <ref>, there is a positive
constant $C_1$
independent of $z_1, z_2$, such that
\begin{align}
C_1\,\norm{\phi}_{\Hsp^1}^2 \leq \norm{\grad\phi}_{\Lsp^2}^2 +
\lproduct{\fnV(h)\,\phi, \phi}_{\Lsp^2},
\end{align}
for all $\phi \in \Hsp^1$. As in the proof of
lemma <ref>, a
second uniform constant, dependent on $D$, can be found such that,
\begin{align}
\norm{\fnV(h)\partial_{z_j}\fnv_j}_{\Lsp^2} \leq \frac{C_2}{\dist_{12}}.
\end{align}
By the Lax-Milgram theorem, we obtain the bound,
\begin{align}
\norm{\partial_{z_j}\tilde h}_{\Hsp^1} \leq \frac{C}{\dist_{12}},
\end{align}
where $C = C_2/C_1$. Now we follow a recursive argument: by
Schauder's estimates,
$\norm{\partial_{z_j}\tilde h}_{\Hsp^2}$ is also bounded
by $C\,d_{12}^{-1}$ for some constant $C$. By Sobolev's
embedding, there is another constant such that
$\norm{\partial_{z_j}\tilde h}_{C^0}$ is also bounded by
$C\,d_{12}^{-1}$. Thence, for any given $p > 2$,
\begin{align}
\norm{\partial_{z_j}\tilde h}_{\Lsp^p} \leq \frac{C}{\dist_{12}}.
\end{align}
By the elliptic estimates,
\begin{align}
\norm{\partial_{z_j}\tilde h}_{\Wsp^{2, p}}
&\leq C\,(\norm{\laplacian \partial_{z_j}\tilde h}_{\Lsp^p}
+ \norm{\tilde h}_{\Lsp^p})\nonumber\\
&\leq C\,(\norm{\fnV(h)\partial_{z_j}\tilde h}_{\Lsp^p}
+ \norm{\fnV(h)\partial_{z_j}\fnv_j}_{\Lsp^p} + \norm{\partial_{z_j}\tilde
&\leq \frac{C}{\dist_{12}},
\end{align}
for the last inequality we have used that the function $\fnV(t)$
is bounded. Sobolev's embedding implies the claimed
\begin{align}
\norm{\partial_{z_j}\tilde h}_{C^1} \leq \frac{C}{\dist_{12}}.
\end{align}
This argument is also valid for $\grad \tilde h$, because it is a
solution to the elliptic problem,
\begin{align}
-\laplacian (\grad \tilde h) = \fnV(h)\grad \tilde h + \fnV(h)(
\grad\fnv_{1} - \grad\fnv_{2}),
\end{align}
and the upper bound
\begin{align}
\norm{\fnV(h)\grad \fnv_{j}} \leq \frac{C}{\dist_{12}}
\end{align}
also holds.
For latter application, we need to translate this estimate to a
holomorphic chart.
Let $\varphi:U \subset \surface \to V \subset \cpx$ be a holomorphic
chart and let $D$ be a geodesically convex neighbourhood such that
$\overline D \subset U$, there is a positive constant
$C$, such that for all $z_1, z_2 \in \varphi(D)$,
\begin{align}
C\,\abs{z_1 - z_2} \leq d(\varphi^{-1}(z_1), \varphi^{-1}(z_2)).
\end{align}
The conformal factor is
continuous positive function on
$V$ and $\varphi(\overline D)$ is compact, hence there is a constant
$C > 0$, such that for all $z \in \varphi(D)$,
\begin{align}
C^2 \leq e^{\Lambda(z)}.
\end{align}
Since $D$ is geodesically convex, for any pair $z_1, z_2 \in
\varphi(D)$, there is a curve $\gamma: [0, 1] \to \varphi(D)$ joining $z_1$
to $z_2$
such that $\varphi^{-1}\circ \gamma$ is a minimizing geodesic joining
$\varphi^{-1}(z_1)$ to $\varphi^{-1}(z_2)$, hence,
\begin{align}
C\int_0^1\abs{\dot \gamma}\,ds \leq \int_0^1 e^{\Lambda/2}\abs{\dot
\gamma}\,ds = \dist(\varphi^{-1}(z_1), \varphi^{-1}(z_2)).
\end{align}
By the triangle inequality,
\begin{align}
\abs{z_1 - z_2} = \abs*{\int_0^1\dot\gamma}\leq \int_0^1 \abs {\dot
\gamma}\,ds,
\end{align}
yielding the result.
The advantage of the holomorphic chart is that it makes computations
possible, on the other hand, the Riemannian distance
is a geometric invariant defined globally on the surface and better suited
to prove analytical properties of the solutions to the Taubes equation.
For the next lemma, notice that if $\surface_1 \times \surface_2$ is a
product of
Riemmann surfaces, for any function $f: \surface_1 \times \surface_2
\to \cpx$ in local coordinates $\varphi_j:U_j \to \cpx$,
$\varphi_j(x_j) = z_j$,
\begin{align}
\del_{x_1}\del_{x_2}f = \partial_{z_1z_2}f\,dz^1\otimes dz^2
\in \Omega^{(2,0)}(\surface_1\times\surface_2).
\end{align}
In the product metric, $dz^1$ and $dz^2$ are orthogonal, hence,
\begin{align}
\abs{\partial_{x_1}\partial_{x_2}f} = \abs{\partial_{z_1, z_2}f}\,
\abs{dz^1}\,\abs{dz^2}.
\end{align}
For any holomorphic chart $\varphi: U\subset \surface \to V \subset \cpx$
and any geodesically convex neighbourhood $D$ such that $\overline D
\subset U$, there is a constant $C > 0$
such that, for all $z_1, z_2 \in \varphi(D)$, $z_1 \neq z_2$,
\begin{align}
\abs{\del_{z_1}b_1(z_1, z_2)} \leq \frac{C}{\abs{z_1 - z_2}},
\end{align}
where the coefficient $b_1$ appearing in the metric
of $\moduli^{1,1}(\surface)$ is defined as in (<ref>).
If $z_1, z_2 \in \varphi(D)$, there is a smooth
function $\tilde v: \varphi(\overline D) \times \varphi(\overline D) \times
\varphi(\overline D) \to \reals$, such that
all triples $z, z_1, z_2$ of points in the domain with $z_1 \neq z_2$,
\begin{align}
v(\varphi^{-1}(z)) = \log\,\abs{z - z_1}^2 - \log\,\abs{z - z_2}^2+ \tilde
v(z, z_1, z_2).
\end{align}
\begin{align}
b_1(z_1, z_2) &= 2\eval{\bar\del}{z=z_1}(h(\varphi^{-1}(z))
- \log\,\abs{z - z_1}^2)\nonumber\\
&= 2\eval{\bar\del}{z = z_1}(\tilde h(\varphi^{-1}(z)) - \log\,\abs{z -
z_2}^2 +
\tilde v(z,z_1,z_2))\nonumber\\
&= 2\,\bar\del_z\tilde h(\varphi^{-1}(z_1); \varphi^{-1}(z_1),
\varphi^{-1}(z_2)) - \frac{2}{\bar z_1 - \bar
z_2} + 2\,\bar\del_z\tilde v(z_1, z_1, z_2),
\end{align}
where $\bar\del_z$ refers to complex derivatives with respect to the first
entry. In the following calculation we denote $\tilde h(\varphi^{-1}(z_1);
\varphi^{-1}(z_1), \varphi^{-1}(z_2))$ by $\tilde h$ and $\tilde v(z_1,
z_1, z_2)$ by $\tilde v$, whence,
\begin{align}
\del_{z_1}b_1 &= 2 \pbrk{
\del_z\bar\del_z\tilde h
+{\del_{z_1}}\bar\del_z\tilde h
+\del_{z}\bar\del_z \tilde v
+\del_{z_1}\bar\del_z \tilde v}\nonumber\\
&= 2 \pbrk{ -\frac{e^{\Lambda(z_1)}}{2}\laplacian_{\surface}\tilde h
+ \bar\del_z \del_{z_1} \tilde h
+\del_{z}\bar\del_z \tilde v
+\del_{z_1}\bar\del_z \tilde v}\nonumber\\
&= 2\pbrk{\frac{e^{\Lambda(z_1)}}{2}\Fstable(h)
+ \bar\del_z \del_{z_1} \tilde h
+\del_{z}\bar\del_z \tilde v
+\del_{z_1}\bar\del_z \tilde v}.
\end{align}
Since $\varphi(\overline D)$ is compact, $\Lambda(z_1)$ and the last two
terms are
bounded functions on $\varphi(D)$ by continuity. Since function $F(t)$ is
bounded, we conclude the same statement for the first term. For the second
term, if $x = \varphi^{-1}(z)$ and $x_j =
\varphi^{-1}(z_j)$, we
have by lemma <ref> and
proposition <ref>,
\begin{align}
\abs*{\bar\del_z \del_{z_1} \tilde h} &=
e^{\Lambda(z_1)/2} \abs*{\bar\del_z \del_{z_1} \tilde
h} \abs*{dz}\nonumber\\
&= \abs{\bar \partial_x\partial_{z_1}\tilde h(x_1, \varphi^{-1}(z_1),
\varphi^{-1}(z_2))}\nonumber\\
&\leq \frac{C}{\dist(x_1,x_2)}\nonumber\\
&\leq \frac{C}{\abs{z_1 - z_2}}.
\end{align}
Therefore the lemma is proved.
The moduli space is incomplete. There is a
Cauchy sequence $\set{\vb x_n} \subset \moduli^{1,1}(\surface)$
such that $\vb x_n \to \vb x \in \diag$ as a sequence in
$\surface\times \surface$.
Let $\varphi: U\subset \surface \to \cpx$ be an holomorphic chart defined on
an open and dense neighbourhood $U$.
Let $z_1 \in \cpx$ be chosen such that $\varphi^{-1}(s z_1)$, $0 \leq s
\leq
1$ is contained in a geodesically convex neighbourhood of
Let us define the curve,
\begin{align}
\gamma: (0, 1] \to \cpx^2_\diag, &&
\gamma(s) = (s\,z_1, 0).
\end{align}
Let $z(s) = s z_1$ and let
$\varphi^{-1}_*\gamma(s) = (\varphi^{-1}(z(s)), \varphi^{-1}(0))$,
be the push forward of the curve $\gamma$ to the moduli space, hence,
\begin{align}
\abs{\varphi^{-1}_*\dot\gamma}_\moduli^2 = (e^{\Lambda(z)}(1-\tau) +
\del_{z_1}b_1)\,\abs{z_1}^2,
\end{align}
where we denote by $\abs{\cdot}_\moduli$
the norm of vectors in $T_{\varphi^{-1}_*\gamma(s)}\moduli^{1,1}$.
By Lemma <ref> there is a constant $C$,
such that,
\begin{align}
\abs{\del_{z_1}b_1} \leq \frac{C}{\abs z} = \frac{C}{s\,\abs {z_1}}.
\end{align}
Since the conformal factor is a continuous positive function
defined on the whole plane, there is another constant, also
denoted $C$, such that,
\begin{align}
\abs {\varphi^{-1}_*\dot \gamma}_\moduli \leq \frac{C}{s^{1/2}}.
\end{align}
Let $\ell[\gamma, a, b]$ be the arc-length of the segment
$\gamma|_{[a, b]}$, $a, b \in (0, 1)$, there is another constant,
also denoted by $C$, such that,
\begin{align}
\ell[\gamma, a, b] = \int_a^b\abs{\varphi^{-1}_*\dot \gamma}_\moduli\,ds
\leq C (b^{1/2} - a^{1/2}),
\end{align}
\begin{align}
\dist(\varphi^{-1}_* \gamma(b), \varphi^{-1}_* \gamma(a)) \leq
C\,(b^{1/2} - a^{1/2}).
\end{align}
This inequality shows if $\set{s_n} \subset (0, 1]$ is any
converging sequence $s_n \to 0$, the new sequence,
\begin{align}
\vb x_n = \varphi^{-1}_*\gamma(s_n) \in \moduli^{1,1}(\surface),
\end{align}
is Cauchy, however $\gamma$ is continuous which
implies $\vb x_n \to (\varphi^{-1}(0), \varphi^{-1}(0)) \in
\diag_\surface$.
Therefore, the moduli space
is incomplete.
§ THE VOLUME OF THE MODULI SPACE
We conclude this chapter computing the volume of the moduli space
$\moduli^{1,1}(\surface)$ for the
round sphere and flat tori. As it will turn out, the existence of a Lie
group of isometries will play an important role in the
calculations. Symmetries were studied for their relation to conservation laws
in a Schrodinger-Chern-Simons model by Manton and Nasir
in [Manton and Nasir, 1999], for the Riemann sphere, symmetries of the coefficients
the $\Lsp^2$ metric for vortices of a non-relativistic Chern-Simons model
were treated by Romão [Romao, 2001]. We follow similar
ideas for asymmetric vortices of the $O(3)$ Sigma model. There is a
general conjecture for the volume of the moduli space by
Romão-Speight [Romão and Speight, 2020], which can be stated as follows,
Given a compact Riemann surface $\surface$ of genus $g$ and total area
$\abs\surface$, let,
\begin{align*}
J_\pm &= 2\pi (1 \mp \tau)\abs\surface - 4\pi^2(k_\pm - k_\mp),\\
K_\pm &= \mp 2\pi^2,
\end{align*}
then the total volume of the moduli space $\moduli^{k_+, k_-}(\surface)$ is,
\begin{equation*}
\vol(\moduli^{k_+, k_-}(\surface)) =
\sum_{l = 0}^g \frac{g!(g - l)!}{(-1)^l l!} \prod_{\sigma = \pm}
\sum_{j_\sigma = l}^g
\frac{
(2\pi)^{2l} J_\sigma^{k_\sigma - j_\sigma} K_\sigma^{j_\sigma - l}
(j_\sigma - l)! (g - j_\sigma)! (k_\sigma - j_\sigma)!
\end{equation*}
For $\surface = \sphere_{round}$, they corroborated
it for a
vortex-antivortex pair and $\tau = 0$. We aim to confirm the
conjecture on the round sphere and flat tori for vortex-antivortex
pairs and general $\tau$.
§.§ The Riemann sphere
On the round sphere, the three dimensional Lie group of orthogonal
transformations, $O(3)$, acts by
isometries. The vortex equations are invariant under
isometric actions on the domain, if $\isometry: \surface \to
\surface$ is an isometry and $u$ is the solution of the Taubes equation
with vortex set $\vset$ and antivortex set $\avset$, then $u \circ
\isometry$ is the solution with data $\isometry^{-1}(\vset)$,
$\isometry^{-1}(\avset)$. We will make use of this symmetry to obtain
conservation laws for the non-trivial coefficients $b_j$ and an explicit
formula in the subspace of vortices and
antivortices located at antipodal positions. This formula will lead us
to the volume formula. We will prove the following theorem,
Recall the conformal factor of the sphere of
radius $R$ in a stereographic projection chart with coordinate $z$ is,
\begin{align}
\cf=\frac{4R^{2}}{\left(1+\abs z^{2}\right)^{2}}.
\end{align}
We can give an explicit description of the coefficients in the metric
in the case of only $k_{+}$ coincident vortices or $k_{-}$ coincident
antivortices. By rotational symmetry,
the function $u$ depends only on the chordal distance to
either the vortex or antivortex [Manton and Nasir, 1999],
the coefficients $b_{\pm}$ in this case are,
\begin{equation}
b_{\pm} = -\frac{ 2 k_{\pm}z_{\pm} }{ 1 +
\abs{z_{\pm}}^{2}}.\label{eq:bpm-antipodal-position}
\end{equation}
The proof relies on the rotational symmetry
of the configuration and is analogous to the proof for $n$ coincident
Ginzburg-Landau vortices on the sphere that can be found in
[Manton, 1993]. With this identity at hand, we prove the
following theorem,
The volume of the moduli space $\moduli^{k_{+},0}\left( \sphere
\right)$ is,
\begin{equation}
\label{eq:vol-moduli-k-0}
\vol \left( \moduli^{k_{+},0}(\sphere) \right)
\end{equation}
and the volume of $\moduli^{0, k_-}(\sphere)$ can be obtained from
equation (<ref>) by changing $\tau$ into
$-\tau$. For a vortex-antivortex pair, the volume of $\moduli^{1,1}\left(
\sphere\right)$ is
\begin{align}
\vol \left( \moduli^{1,1}(\sphere) \right)
\end{align}
For $k_+ = 0$ or $k_- = 0$ we follow ideas of
Manton-Nasir [Manton and Nasir, 1999], as their proof relies on the topology
of the symmetric product $(\sphere)^N/S_N$, $S_N$ being the $N$ symmetric
group, and can be adapted easily to vortices of the $O(3)$ Sigma model of the
same type. For the case $k_+ = k_- = 1$, we extend the proof given
by Romão-Speight <cit.> for the symmetric case. For
general $\tau$ we no longer have the symmetry $(z_1, z_2) \mapsto (z_2, z_1)$,
instead, we complement the symmetries induced by $SO(3)$ in the moduli space
the symmetry $(z_1, z_2) \mapsto (\conj z_1, \conj z_2)$ to deduce a suitable
formula for the volume of a general Kähler metric on $\sphere\!_{\diag}$.
§.§.§ $k_+$ vortices of the same type
If there are $k_+$ vortices on $\sphere$ and no antivortices, the moduli
space is isomorphic to $\pSpace^{n}$, the complex projective space of
dimension $k_+$ [Manton and Nasir, 1999]. The subspace
$\moduli_0^{k_+,0}(\sphere) \subset \moduli^{k_+,0}(\sphere)$ of $k_+$
coincident vortices on the other hand is isomorphic to $\pSpace^1$,
and can be parametrized with the coordinate $z_+$ of the coincident
vortices. By equation (<ref>) we know how
to compute the coefficient $b_{+}$ in $\moduli_0^{k_+,0}(\sphere)$,
\begin{equation}
b_{+}=-\frac{2 k_+ z_+}{1+\abs {z_+}^2}.\label{eq:b-formula-k-0}
\end{equation}
The metric in $\moduli^{k_+,0}_0(\sphere)$ therefore is,
\begin{align}
ds^{2} & =2 k_+\pi\left((1-\tau)\cf + \frac{\del b_{+}}{\del
z_+}\right)\abs{dz_+}^{2}\nonumber \\
& = k_+
\pi\left(2(1-\tau) - \frac{k_+}{R^2}
\right)\Omega\abs{dz_+}^{2},\label{eq:metric-moduli-kp-0}
\end{align}
as can be seen, the metric is a multiple of the round metric,
hence, the volume of $\moduli^{k_+,0}_0(\sphere)$ is,
\begin{equation}
4\pi^{2} R^{2}k_+ \left( 2 (1-\tau)-\frac{k_+}{R^{2}}
\right),\label{eq:vol-moduli-N}
\end{equation}
this volume is $k_+$ times the volume of the generating cycle in
\begin{equation}
4\pi^{2}R^{2}\left( 2 (1-\tau)-\frac{k_+}{R^{2}}
\right).\label{eq:generating-cycle-vol}
\end{equation}
The total volume of the moduli space therefore is,
\begin{equation}
\vol\left(\moduli^{k_+,0}(\sphere)\right) =
\frac{\left(8\pi^{2}R^{2}(1-\tau)-4\pi^{2} k_+
\right)^{k_+}}{k_+!},\label{eq:total-volume-moduli-space}
\end{equation}
the proof of the volume formula in $\moduli^{0,k_-}(\sphere)$ is analogous,
\begin{equation}
\vol\left(\moduli^{0,k_-}(\sphere)\right) =
\frac{\left(8\pi^{2}R^{2}(1+\tau)-4\pi^{2} k_-
\right)^{k_-}}{k_-!}.%\label{eq:total-volume-moduli-space}
\end{equation}
§.§.§ The moduli space of vortex-antivortex pairs
In general, there is no explicit expression for the
coefficients $b_j$ of the metric if the cores are at general position, however,
can deduce from the invariance of
the Taubes equation under the action of $O(3)$ several constraints on the
coefficients due to symmetry. Before doing so, we need a general lemma that
will also be necessary for flat tori in the next section.
Let $\varphi: U \subset \surface \to V\subset \cpx$ be a
holomorphic chart, containing the core set $\mathcal{Z}$ of a
point in the
moduli space $\moduli^{1,1}(\surface)$.
For any bounded domain $D \subset V$, such that $\mathcal{Z}
\subset \varphi^{-1}(D)$, there are continuous functions $\tilde b_j:
D\times D \to
\cpx$,
$j = 1, 2$,
such that:
* If $\varphi(\mathcal{Z}) = \left\{z_1, z_2\right\}$, where
$z_1$ $(z_2)$ is the vortex (antivortex),
\begin{align}
b_j(z_1, z_2) = \frac{-2\,\sign_j}{\bar z_1 - \bar z_2} + \tilde
b_j(z_1, z_2),
\end{align}
where $b_j$, $j = 1, 2$, are the non-trivial coefficients in the
metric, defined in lemma <ref>.
\begin{align}
\lim_{|z_1 - z_2| \to 0} \tilde b_j(z_1, z_2) = 0.
\end{align}
On $\varphi^{-1}(D)$, Green's function
can be written as
\begin{align}
G(x_1, x_2) = \frac{1}{2\pi}\,\log\,\abs{\varphi(x_1) - \varphi(x_2)} +
\tilde G(x_1,
\end{align}
with a smooth regular part $\tilde G: \varphi^{-1}(D)\times
\varphi^{-1}(D) \to \reals$. Therefore, the solution $h$ to the Taubes
equation can be written as
\begin{align}
h(x; x_1, x_2) = \tilde h(x; x_1, x_2) + \log\,\abs{\varphi(x) -
\varphi(x_1)}^2 - \log\,\abs{\varphi(x) - \varphi(x_2)}^2 + \tilde v(x;
x_1, x_2),
\end{align}
\begin{align}\label{eq:tilde-v-vav-sigma}
\tilde v(x; x_1, x_2) = 4\pi\, \tilde G(x, x_1)
- 4\pi\, \tilde G(x, x_2),
\end{align}
and $\tilde h(x; x_1, x_2)$ can be extended in $C^1$ to the coincidence set
$x_1 = x_2$ by corollary <ref>.
Denoting $h(\varphi^{-1}(z); \varphi^{-1}(z_1),
\varphi^{-1}(z_2))$ and $\tilde h(\varphi^{-1}(z); \varphi^{-1}(z_1),
\varphi^{-1}(z_2))$ as $h$, $\tilde h$, etcetera,
\begin{align}
b_j(z_1, z_2) &= 2\,\conj\del|_{z=z_j}
\pbrk{\sign_j\,h - \log\abs{z - z_j}}\nonumber\\
&= 2\,\conj\del_{z=z_j}
\pbrk{\sign_j\tilde h - \log\abs{z - z_k} + \sign_j\tilde v}\nonumber\\
&= \frac{-2}{\conj z_j - \conj z_k}
+ 2\,\sign_j\conj\del|_{z=z_j}(\tilde h + \tilde v)\nonumber\\
&= \frac{-2\,\sign_j}{\conj z_1 - \conj z_2} + \tilde b_j,
\end{align}
where the regular part $\tilde b_j$ is continuous in $D\times D$. This
proves the first statement. The second statement is a consequence of
corollary <ref> and the fact that
by (<ref>),
\begin{align}
\lim_{|z_1 - z_2| \to 0} \conj\del\lvert_{z=z_j}\pbrk{\tilde
\varphi^{-1}(z_1), \varphi^{-1}(z_2))} = 0.
\end{align}
Suppose $\gamma:U_{1} \subset \cpx \to U_{2}
\subset \cpx$ is
a holomorphic change of coordinates in ambient space, such that
$z_{k}\in U_{1}$
for all cores. There are pairs of corresponding
coefficients $b_{s}(z_{1},\ldots,z_{n})$,
in each of the charts. Let $z'=\gamma(z)$,
$z'_{k}=\gamma(z_{k})$, as in [Romao, 2001], we have the
transformation rule
\begin{equation}
b'_j = \frac{1}{\conj{\gamma'_j}}b_j-\frac{\conj{\gamma''_j}} {\left(
\conj{ \gamma'_j}\right)^{2}}.\label{eq:b-coeffs-change-of-coords}
\end{equation}
Manton and Nasir noted in [Manton and Nasir, 1999] that
equation (<ref>) is similar to the transformation
rule for the Levi-Civita connection on $\sphere$ and resembles the
topological nature of the coefficients $b_j$. In the sphere, the group of
isometries is large, in the sense that it is a Lie group, and each of this
isometries induces a holomorphic change of coordinates on the moduli space.
We exploit this remark to prove the following lemmas.
In the projective chart, the coefficients $b_j$ satisfy the identities,
\begin{gather}
\sum_k(2\,\conj z_k + \conj z_k^2\,b_k + \conj
b_k) = C,\label{eq:bs-gen-rel}\\
\sum_k\conj z_kb_k\in\reals,\label{eq:z-axis-symmetry-b}
\end{gather}
for some constant $C$. For a vortex-antivortex pair, $C = 0$.
Romão deduced similar identities for vortices of a modified Chern-Simons
on the sphere in [Romao, 2001], employing the action of $SO(3)$ on
the moduli space.
Let $\mathbb{S}^2_\Delta$ be the diagonal in the product $\sphere
\times \sphere$. The orthogonal group acts diagonally on the moduli space
$\moduli^{1,1}(\sphere) \cong (\sphere \times \sphere)
\setminus \mathbb{S}^2_\Delta$ by isometries. We can always assume
there is a projective chart such that the pair is located with the
vortex at $z_1 = \epsilon$ and the antivortex at $z_2 =
-\epsilon$. From (<ref>) and the fact that
\begin{equation}
b_{j}(\epsilon, -\epsilon) =
\conj{b_j}(\epsilon, -\epsilon),\label{eq:refl-invariance}
\end{equation}
we conclude,
\begin{equation}
b_{1}(\epsilon,-\epsilon) +
\end{equation}
The $\Lsp^2$ metric in $\moduli^{1,1}(\sphere)$ is Kähler and invariant
under the diagonal action of $O(3)$, given any
pair $(z_1, z_2) \in \moduli^{1,1}(\sphere)$, we can always find a
rotation of $\sphere$ such that in south pole stereographic
projection, $z_1 = \epsilon$, $z_2 = -\epsilon$. In this way, we have
a diffeomorphism,
\begin{equation}
(\sphere \times \sphere) \setminus \mathbb{S}^2_\Delta \cong (0, 1]
\times SO(3),
\label{eq:moduli11-so3-param}
\end{equation}
hence, the moduli space can be parametrized as $(0, 1]\times SO(3)$.
Let $g$ be a Kähler metric in $\sphere\times\sphere$ such that if $o
\in O(3)$ and $(z_1, z_2) \in \sphere\times\sphere$, then the action
\begin{equation}
o * (z_1, z_2) = (o * z_1, o * z_2),
\end{equation}
is by isometries. Let $E_0 = \del_\epsilon$ and let $E_j \in
\mathfrak{so}(3)$ be the left invariant vector field corresponding to
rotations with respect to the $j$-th coordinate axis in $\reals^3$.
Then there exists a function
\begin{equation}
A : (0, 1] \to \reals,
\end{equation}
and a real constant $c$ such that in the parametrization
\begin{multline}
\label{eq:so3-diag-inv-metric}
g = A\left(\frac{1-\epsilon^{2}}{1+\epsilon^{2}}\,(\sigma^{1})^{2} +
\frac{1+\epsilon^{2}}{1-\epsilon^{2}}(\sigma^{2})^{2}\right) -
\frac{1}{\epsilon}
\frac{dA}{d\epsilon}
\left((\sigma^{0})^{2}+\epsilon^{2}(\sigma^{3})^{2}\right)\\
+ \frac{c}{1+\epsilon^{2}}\left(\sigma^{0}\sigma^{2}
+ \frac{\epsilon(1-\epsilon^{2})}{1+\epsilon^{2}} \sigma^{1}
\sigma^{3}\right),
\end{multline}
where $\sigma^k\in T^*((0,1]\times SO(3))$ is the co-vector dual to
$E_k$, $k = 0, \ldots, 3$. For this metric, the volume is,
\begin{equation}
\label{eq:vol-formula}
\vol\left(\sphere\times\sphere\right) =
\end{equation}
Applying lemma <ref> to the $\Lsp^2$ metric,
we obtain,
The $\Lsp^2$ metric on $\moduli^{1,1}(\sphere)$ has the structure
provided by Lemma <ref>, with
\begin{gather}
A = 2\pi \left(
\frac{4R^2}{1 + \epsilon^2} - \epsilon\,b_1 - 2R^2 - 1
\right),
\label{eq:A-formula}\\
c = 8\pi R^2\tau.
\label{eq:c-formula}
\end{gather}
To compute the constant $c$, we calculate
$g\left(E_{0},E_{2}\right)$. Tangent vectors $E_{0}$, $E_{2}$ in
projective coordinates $(z_1, z_2) \in \sphere \times \sphere$ with
respect to the south pole are,
\begin{equation}
E_{0} = \frac{\del}{\del x_1} - \frac{\del}{\del x_{2}}, \qquad
E_{2} = \frac{1+\epsilon^2}{2}\,\left( \frac{\del}{\del x_{1}} +
\frac{\del}{\del x_2} \right).\label{eq:E0-E2-proj}
\end{equation}
where $z_k = x_k + i y_k$. Thence,
\begin{align}
& =\frac{1+\epsilon^{2}}{2}g\left(\frac{\del}{\del
x_{1}}-\frac{\del}{\del x_{2}},\frac{\del}{\del
x_{1}}+\frac{\del}{\del x_{2}}\right)\nonumber \\
& =\frac{1+\epsilon^{2}}{2}2\pi\left(\cf(1+\tau)-\cf(1-\tau)+\frac{\del
b_{1}}{\del z_{1}}+\frac{\del b_{2}}{\del z_{1}}-\frac{\del
b_{1}}{\del z_{2}}-\frac{\del b_{2}}{\del
\end{align}
To simplify (<ref>), we use the symmetries of the
coefficients $b_j$, lemma <ref>,
\begin{align}
\sum_j\left(\pdv{z_1}{b_j} - \pdv{z_2}{b_j}\right) &=
\frac{1}{2} \sum_j{\dv\epsilon}b_j(\epsilon,-\epsilon) -
\frac{i}{2}\sum_j
\left(\pdv{y_1}{b_j} - \pdv{y_2}{b_j}\right)\nonumber\\
&=\frac{1}{2}{\dv \epsilon} (b_1 + b_2) + \frac{1}{2\epsilon}
(b_1 + b_2)\nonumber\\
&= 0.
\end{align}
\begin{equation}
\label{eq:gl2-e0-e2}
\moduliMetric(E_0, E_2) = \frac{8\pi R^2\tau}{1 + \epsilon^2}
\end{equation}
and consequently $c = 8\pi R^2\tau$. Let us compute
$\moduliMetric(E_0, E_0)$,
\begin{align}
\nonumber
\moduliMetric(E_0, E_0) &= \moduliMetric \left( \pdv{x_1} -
\pdv{x_2}, \pdv{x_1} -\pdv{x_2} \right)\\
&= 2 \pi \left( \cf (1 + \tau) + \cf (1 - \tau)
+ \pdv{z_1}{b_1} - \pdv{z_1}{b_2}
- \pdv{z_2}{b_1} + \pdv{z_2}{b_2} \right).
\end{align}
Again by symmetry,
\begin{equation}
\label{eq:db1-db2-prop-at-qeps}
\pdv{z_1}{b_j} - \pdv{z_2}{b_j} = \frac{1}{2}
\frac{d b_j}{d\epsilon}
\end{equation}
\begin{equation}
\label{eq:gl2-e0-e0}
\moduliMetric(E_0, E_0) = 2\pi \left( \frac{8R^2}{(1 +
\epsilon^2)^2} +
\frac{d b_1}{d\epsilon}
+ \frac{1}{\epsilon}b_1 \right).
\end{equation}
Comparing (<ref>) and (<ref>),
\begin{equation}
\label{eq:A-diff-eq}
- \frac{1}{\epsilon} \frac{d A}{d\epsilon} = 2\pi \left(
\frac{8R^2}{(1 + \epsilon^2)^2} + \frac{db_1}{d\epsilon}
+ \frac{1}{\epsilon}b_1 \right),
\end{equation}
Solving this equation, we find,
\begin{equation}
\label{eq:A-diff-eq-sol}
A = \frac{8\pi R^2}{1 + \epsilon^2} - 2\pi\epsilon b_1 +
\mathrm{const}.
\end{equation}
From the regularity condition $\lim_{\epsilon\to 1}A(\epsilon) = 0$ used
to compute the formula for the volume of the moduli space and the
explicit formula (<ref>) for $b_1$ in the
antipodal case, the constant is
\begin{equation}
\label{eq:A-sol-const}
\mathrm{const}. = -2\pi (2R^2 + 1).
\end{equation}
\begin{equation}
A = 2\pi \left( \frac{4R^2}{1 + \epsilon^2} - \epsilon b_1 - 2R^2 -
\end{equation}
We claim that
\begin{equation}
\label{eq:lim-eps-b1}
\lim_{\epsilon\to 0} \epsilon b_1 = -1
\end{equation}
as can be seen numerically in figure <ref> for the
symmetric case in the unit sphere.
For a vortex-antivortex pair,
\begin{equation}
b_1(\epsilon, -\epsilon) = 2 \eval{\frac{\del}{\del
x}}{z=\epsilon} h_{\epsilon} - \frac{1}{\epsilon}.
\label{eq:vav-b1}
\end{equation}
Since $h_{\epsilon} \to \mu$ in $C^1$ as $\epsilon\to 0$,
\begin{equation}
\lim_{\epsilon\to 0} \epsilon\,b_1(\epsilon, -\epsilon) = -1.
\label{eq:the-limit-conclusion}
\end{equation}
Applying lemmas <ref>
and <ref>, the volume of the moduli space is
\begin{equation}\label{eq:volume-conjecture-sphere-formula}
\vol \left( \moduli^{1,1}(\sphere) \right) = \left( 8\pi^2
R^2\right)^2 (1 - \tau^2).
\end{equation}
Notice that another way to express the volume is as $4\pi^2(1-\tau^2)\vol
(\sphere)$, which corresponds to the volume of a product of spheres,
each factor weighted by $2\pi (1\pm\tau)$, the effective mass of a core,
hence, it is expected that as $\tau \to \pm 1$, the volume vanishes,
because of the negligible weight of one of the factors.
Three views of the declination data of $\tilde h_{\epsilon}$,
the regular part of the solution to the Taubes equation, for
three different values of the asymmetry parameter $\tau$ on the
unit sphere.
Top. Vortex and
antivortex are symmetric, with the same effective
mass. Middle and bottom. The antivortex becomes more
massive. We solved
from $\epsilon = 1$ down to $0.05$ in steps of
$0.05$, except that for $\tau = 0.5$, the computation
stopped at $\epsilon = .20$ due to algorithm divergence.
As $\epsilon \to 0$ the data shows how $\tilde h_\epsilon$ flattens
as expected.
Top. Real profile of $\epsilon b$ in the symmetric
case. The limit $\lim_{\epsilon\to 0} \epsilon b = -1$ is apparent
in the numerical data. Bottom. Real profile of a
vortex-antivortex pair located at $\pm \epsilon$ on the real axis
of the extended complex plane for several values of
$\epsilon$. In both cases, the domain is the unit sphere, the bottom
plot shows the behaviour of the real profile of
$\tilde h$ as $\epsilon \to 0$ in the south pole of the domain. The dashed
horizontal line is $\log \left( (1 - \tau)(1 +
\tau)^{-1} \right)$. The data shows how the
regular part of the solution to the Taubes equation converges to this
constant value as the pair collides at the north pole.
§.§ Flat tori
In this section we compute the volume of the moduli space for a flat
tori, to
this end, we extend the coefficients $b_q$ in
the $\Lsp^2$ metric to a global object and relate it to the volume of
$\moduli^{1,1}(\mathbb{T}^2)$ in lemma <ref>.
Consider a holomorphic chart $\varphi: U \subset \mathbb{T}^2 \to \cpx$
on an open and dense set $U$, with coordinates $z = \varphi(x)$, $x \in U$. Let
us define,
\begin{align}
b_U = b_j\,d\bar z^j \in \Omega^{(0,1)}((U \times U) \setminus \diag_U).
\end{align}
In general $b_U$ is only well defined on a chart, however, flat tori admit
atlases such that the holomorphic changes of coordinates are
translations. Since
translations have trivial second derivatives, by
(<ref>) $b_U$ extends to a global form $b \in
\Omega^{(0,1)}(\moduli^{1,1}(\mathbb{T}^2))$. By the symmetries of the
coefficients $b_j$, this form is holomorphic, as the following short
calculation shows in coordinates:
\begin{align}
\bar\del b_U &= \sum_{i,j} \bar\del_{z_i}b_j\,d\bar z^i\wedge d\bar
&= -\sum_{i,j} \bar\del_{z_j}b_i\,d\bar z^j\wedge d\bar z^i\nonumber\\
&= -\bar\del b_U,
\end{align}
hence, $\bar \del b_U = 0$.
To compute the volume of flat tori, we will use the $(1,1)$-form $\del b$ to
define another form in the moduli space which is more convenient
for calculations. Let $\proj_j:\mathbb{T}^2\times \mathbb{T}^2\to \mathbb{T}^2$
be the canonical projection map onto the $j$-th factor of the product. Let
us define the form
\begin{align}
\kform_0 = 2\pi\,(1-\tau)\,\proj_1^{*}\,\kform_{\mathbb{T}^2} + 2\pi\,(1 +
\tau)\, \proj_2^{*}\,\kform_{\mathbb{T}^2}.
\end{align}
The Kähler form on the moduli space can be written as,
\begin{align}
\kform &= \kform_0 + \pi i\,\del b \in
\Lambda^{1,1}(\moduli^{1,1}(\mathbb{T}^2)).
\end{align}
Notice that,
\begin{align}
\vform &= \half\,\kform\wedge\kform\nonumber\\
&= \vform_0 + \pi i\,\kform_0\wedge \del b -
\frac{\pi^2}{2} \del b \wedge \del b,
\end{align}
where $\vform_0 = \half \kform_0\wedge\kform_0$ is the restriction of
the volume form in the product $\mathbb{T}^2\times\mathbb{T}^2$ to the moduli
Let $\diag_\epsilon$ be the $\epsilon$-tubular neighbourhood of
the diagonal set of $\mathbb{T}^2\times\mathbb{T}^2$ for small
$\epsilon$. The volume of the moduli space can be computed as,
\begin{multline}
\Volume(\moduli^{1,1}(\mathbb{T}^2)) =
4\pi^2 (1 - \tau^2)\,\Volume(\mathbb{T}^2)^2 \\
+ \lim_{\epsilon\to 0}\int_{\mathbb{T}^2 \times \mathbb{T}^2
\setminus \diag_\epsilon}
\pbrk{\pi i\, \kform_0\wedge \del b
- \frac{\pi^2}{2}
\del b \wedge \del b}.\label{eq:vol-int-lim}
\end{multline}
\begin{align}
\Volume(\moduli^{1,1}(\mathbb{T}^2))
&= \lim_{\epsilon\to 0}
\int_{\mathbb{T}^2 \times \mathbb{T}^2 \setminus \diag_\epsilon}
\vform\nonumber\\
&= \int_{\mathbb{T}^2\times\mathbb{T}^2}\vform_0
+ \lim_{\epsilon\to 0}\int_{\mathbb{T}^2 \times
\mathbb{T}^2 \setminus \diag_\epsilon}
\pbrk{\pi i\,\kform_0\wedge \del b
- \frac{\pi^2}{2} \del b \wedge \del b}.
\end{align}
On the other hand,
\begin{align}
\vform_0 = 4\pi^2(1 - \tau^2)\,\proj_1^{*}\,\kform_{\mathbb{T}^2} \wedge
\proj_2^{*}\,\kform_{\mathbb{T}^2}.
\end{align}
Applying Fubini and the change of variables theorems,
\begin{align}
\int_{\mathbb{T}^2\times\mathbb{T}^2}\vform_0
&= 4\pi^2(1-\tau^2)\pbrk{\int_{\mathbb{T}^2}\kform_{\mathbb{T}^2}}^2
= 4\pi^2(1-\tau^2)\Volume(\mathbb{T}^2)^2.
\end{align}
This concludes the proof of the lemma.
According to lemma <ref>, to compute the volume
of $\moduli^{1,1}(\mathbb{T})$, we must compute the two non-trivial
terms in (<ref>).
Let $\pi: \cpx \to \mathbb{T}^2$ be the canonical covering map and let
$R \subset \cpx$ be an open parallelogram such that $\pi|_R: R \to
\mathbb{T}^2$ is a bi-holomorphism onto its image and $U = \pi|_R(R)$ is
open and dense. On the local coordinates $\pi|_R^{-1}:U \to R$, there is a
$c \in \cpx$ such that for any pair of
different points $z_1, z_2 \in R$,
\begin{align}
b_1(z_1, z_2) + b_2(z_1, z_2) &= c,
\end{align}
If $\mathcal{I} : \mathbb{T}^2 \to \mathbb{T}^2$ is an isometry, the Taubes
equation is invariant under $\mathcal{I}$,
\begin{align}
h(\mathcal{I}(x); \mathcal{I}(x_1), \mathcal{I}(x_2)) = h(x; x_1, x_2),
\end{align}
$x, x_1, x_2 \in \mathbb{T}^2$, $x_1 \neq x_2$.
By construction, there is a $v \in \cpx$ such that
$\mathcal{I}_\varphi = \varphi\circ \mathcal{I} \circ \varphi^{-1}(z) = z +
v$ for $z \in \varphi(\mathcal{I}^{-1}(U)\cap U)$. For small $v$, the
translation $\mathcal{I}_\varphi$ maps a neighbourhood,
not necessarily connected,
$N \subset R$ of $x_1$ and $x_2$ into $R$.
This implies $b_j$ has the symmetries,
\begin{align}
b_j(z_1 + v, z_2 + v) = b_j(z_1, z_2),
\end{align}
$v$ small. Hence,
\begin{align}
\del_{z_1}b_j + \del_{z_2}b_j = \bar\del_{z_1}b_j + \bar\del_{z_2}b_j = 0.
\end{align}
Applying the symmetries of the coefficients $b_j$,
\begin{align}
\del_{z_j}(b_1 + b_2) = \bar\del_{z_1}\bar b_j + \bar\del_{z_2}\bar b_j = 0.
\end{align}
\begin{align}
\bar\del_{z_j}(b_1 + b_2) = 0.
\end{align}
Hence $b_1 + b_2$ is constant on the connected neighbourhood $R$.
In a flat torus $\mathbb{T}^2$, for the $(1,1)$ form $\del b$ we
\begin{align}
\del b \wedge \del b = 0.
\end{align}
We apply the previous lemma to prove the proposition.
By lemma <ref>, there is an open and dense
set $U \subset \mathbb{T}^2$ and a chart $\varphi: U \to R\subset \cpx$,
$R$ an open parallelogram, such that in this local coordinates $b_1 + b_2$ is
a constant. Denoting points in $R$ as $z_j$, a direct calculation shows,
\begin{align}
b_U \wedge \del b_U
(b_2\,\del_{z_1}b_1 - b_1\,\del_{z_1}b_2)\,
dz_1\wedge d\bar z_1 \wedge d\bar z_2\nonumber\\
&\quad + (-b_2\,\del_{z_2}b_1 + b_1\,\del_{z_2} b_2)\,
d\bar z_1 \wedge dz_2 \wedge d\bar z_2\nonumber\\
-c\,\del_{z_1}b_2\, dz_1\wedge d\bar z_1 \wedge d\bar z_2
-c\,\del_{z_2}b_1\,d\bar z_1 \wedge dz_2 \wedge d\bar z_2.
\end{align}
Since $b_1$ and $b_2$ add to a constant,
\begin{align}
\del b_U \wedge \del b_U
-c\,(\del_{z_2}\del_{z_1} b_2 + \del_{z_1}\del_{z_2}b_1)\,
dz_1 \wedge d\bar z_1 \wedge dz_2 \wedge d\bar z_2 = 0.
\end{align}
Since $U$ is dense, we conclude $\del b \wedge \del b \equiv 0$.
By this proposition and lemma <ref>, to compute the volume
of the moduli space, we have to integrate
$\kform_0\wedge \del b$.
For a flat torus $\mathbb{T}^2$, the volume of the moduli space is,
\begin{align}
\Volume(\moduli^{1,1}(\mathbb{T}^2))
4\pi^2(1 - \tau^2)\,\Volume(\mathbb{T}^2)^2 +
\end{align}
Notice that the first term of the formula is
similar to the case of the
sphere (<ref>), however, the second
term is new, bearing in mind
the volume conjecture, <ref>, one can argue the
extra term is related to the genus of the base surface, however, it
is not clear how to relate our computation to this fact and
the relation is open to
future work.
\begin{align}
\mathbb{T}^2({\epsilon})
&= (\mathbb{T}^2 \times \mathbb{T}^2) \setminus \diag_{\epsilon},\\
\kform_j &= \proj^{*}_j\kform_{\mathbb{T}^2}, \qquad j = 1, 2,
\end{align}
and let $k$ be the complementary index of $j$, such that
$\set{j, k} = \set{1, 2}$. By Fubini's theorem,
\begin{align}
\int_{{\mathbb{T}^2}(\epsilon)}\kform_0\wedge \del b
2\pi \sum_j (1 - \sign_j\tau)\int_{\mathbb{T}^2}\pbrk{
\int_{\mathbb{T}^2 \setminus
\disk_{\epsilon}(x_j)} \inc_k^{*}\del b}\, \kform_{\mathbb{T}^2},
\end{align}
where for any given $x_j \in \mathbb{T}^2$, $\inc_k: \mathbb{T}^2
\hookrightarrow
\mathbb{T}^2\times\mathbb{T}^2$ is the
inclusion of the torus as the k-th factor of the product anchored at $x_j$.
$b$ is well defined globally,
\begin{align}
\int_{\mathbb{T}^2\setminus\disk_{\epsilon}(x_j)} \inc_k^{*}\del b
\int_{\del \mathbb{T}^2\setminus\disk_{\epsilon}(x_j)} \inc_k^{*}b
-\int_{\del \disk_{\epsilon}(x_j)} \inc_k^{*}b,
\end{align}
where we always orient a submanifold by the outward pointing
normal. Let $\varphi: U \to \cpx$ be a holomorphic chart defined on an open
and dense set $U$. If $x_j \in U$, for small $\epsilon$, $\disk_\epsilon(x_j)
\subset U$. Assume $j = 1$, $k = 2$, in the chart,
\begin{align}
(\varphi^{-1})^*\inc_2^{*}b = b_2 d\bar z.
\end{align}
If $z_1 = \varphi(x_1)$ and $D(z_1) \subset \cpx$ is a
bounded domain and neighbourhood of $z_1$, by lemma <ref>,
\begin{align}
b_2 = \frac{2}{\conj z_1 - \conj z_2} + \tilde b_2(z_1, z_2),\qquad
z_2 \in D(z_1).
\end{align}
If $D_\epsilon(z_j) = \varphi(\disk_\epsilon(x_j))$, by Cauchy's residue
\begin{align}
\int_{\del \disk_{\epsilon}(x_1)} \inc_2^{*}b =
-2\int_{\del D_\epsilon(z_1)} \frac{d\conj z}{\conj z - \conj z_1}
+ \int_{\del D_\epsilon(z_1)}\tilde b_2(z_1,z)d\conj z
= 4\pi i + \int_{\del D_\epsilon(z_1)}\tilde b_2(z_1,z)d\conj z.
\end{align}
If $j = 2$, $k = 1$, we find a similar result,
\begin{align}
\int_{\del \disk_{\epsilon}(x_2)} \inc_1^{*}b =
4\pi i + \int_{\del D_\epsilon(z_2)} \tilde b_1(z, z_2) d\conj z.
\end{align}
Since $\tilde b_k$ is a continuous function in a neighbourhood of each
$z_j \in \cpx$,
\begin{align}
\lim_{\epsilon \to 0} \int_{\del D_\epsilon(z_j)} \tilde b_k\,d\conj z = 0.
\end{align}
Hence, since $U$ is dense in $\mathbb{T}^2$,
\begin{align}
lim_{\epsilon\to 0}
\int_{\mathbb{T}^2}\pbrk{
\int_{\mathbb{T}^2 \setminus
\disk_{\epsilon}(x_j)} \inc_k^{*}\del b} \kform_{\mathbb{T}^2}
&= -4\pi i \Volume(\mathbb{T}^2)\nonumber\\
&\quad - \frac{i}{2} \int_{\cpx}
lim_{\epsilon\to 0} \pbrk{\int_{\del D_{\epsilon}(z_j)} \tilde b_k}
e^{\Lambda(z_j)} dz_j\wedge d\conj z_j\nonumber\\
&= -4\pi i \Volume(\mathbb{T}^2).
\end{align}
\begin{align}
\int_{\moduli^{1,1}(\mathbb{T}^2)}\kform_0\wedge\del b
&= 2\pi\sum_j(1 - \sign_j\tau)
lim_{\epsilon\to 0}
\int_{\mathbb{T}^2}\pbrk{
\int_{\mathbb{T}^2 \setminus
\disk_{\epsilon}(x_j)} \inc_k^{*}\del b} \kform_{\mathbb{T}^2}.\nonumber\\
&= 2\pi\sum_j(1 - \sign_j\tau)
\left(-4\pi i \Volume(\mathbb{T}^2) \right)\nonumber\\
&= -16\pi^2\,i\,\Volume(\mathbb{T}^2).
\end{align}
By lemma <ref>, we conclude the volume formula.
CHAPTER: CHERN-SIMONS DEFORMATIONS OF VORTICES
§ DYNAMICS OF THE MODULI SPACE OF GINZBURG-LANDAU VORTICES WITH A
CHERN-SIMONS TERM
⟨#2 ⟩
⟨#2 ⟩
CHAPTER: CONCLUSION
In this work we focused on geometric models of vortices and
antivortices of the $O(3)$ Sigma model. We emphasised the geometric nature of
the interaction of a vortex-antivortex pair on the moduli space.
We were able to prove that the $\Lsp^2$ metric in the moduli space is
both on the euclidean plane and on a compact surface. We also analysed the
properties of the interaction on the plane, focusing on
scattering of vortex-antivortex pairs.
We also computed the volume of the moduli space on spheres and flat tori,
corroborating the work of Speight and Rõmao who conjectured a
formula for the volume of the moduli space for a general surface.
The fact that the moduli space is incomplete imposed some technical
difficulties on the proofs, that we overcame by analysing the
behaviour of solutions to the Taubes equations in the collision of a
vortex and an antivortex.
Finally, we added a Chern-Simons interaction term to our model and
applied the geodesic approximation ideas to determine the extra term in the
metric of the moduli space for small perturbations due to the
interaction. Our analysis indicates that the extra term
can be extended to the coincidence set.
Some questions remain opened, representing an opportunity for future
work. The short range
approximation formula for the metric on the space of vortex-antivortex
pairs of the euclidean plane relies on uniform convergence of the
family $\tilde h_{\epsilon}/\epsilon$, as $\epsilon \to 0$, where
$\tilde h_{\epsilon}$ is the regular part of the Taubes equation.
Numerical evidence
suggests this conjecture is true. Should it be the case, we would be
able to prove formally that the Gaussian curvature of
$\moduli^{1,1}_0(\plane)$ diverges as $\epsilon \to 0$
as expected from the numerical evidence and we could
also justify analytically the effective potential of Ricci magnetic
geodesics. The equivalent conjecture for a compact surface would allow
to compute the volume formula for a general surface, where we no
longer have the extra symmetries that we used for the task.
In conclusion, geometric ideas to study field theory originated in the realm of
superconductivity with the Ginzburg-Landau functional at critical
coupling, but they have proved to be fruitful in a broader context. In
particular, for asymmetric vortex-antivortex systems of the $O(3)$ Sigma model,
where with these ideas one can understand dynamics from a geometric point of
[Alqahtani and Speight, 2015]
LS Alqahtani and JM Speight.
Ricci magnetic geodesic motion of vortices and lumps.
Journal of Geometry and Physics, 98:0 556–574, 2015.
[Aubin, 2013]
Thierry Aubin.
Some nonlinear problems in Riemannian geometry.
Springer Science & Business Media, 2013.
[Bradlow, 1990]
Steven B Bradlow.
Vortices in Holomorphic Line Bundles over Closed Kahler Manifolds.
Commun. Math. Phys, 1350 (1):0 1–17, 1990.
URL <https://projecteuclid.org:443/euclid.cmp/1104201917>.
[Carroll, 2004]
Sean Carroll.
Spacetime and geometry : an introduction to general
Addison Wesley, San Francisco, 2004.
ISBN 0805387323.
[Chae and Nam, 2001]
D Chae and Hee Seok Nam.
On the condensate multivortex solutions of the self-dual
maxwell-chern-simmons cp1 model.
Annales de l'Institut Henri Poincare, 0 (2):0
887–906, 2001.
ISSN 03779017.
[Chen and Chern, 2019]
Zhi-You Chen and Jann-Long Chern.
The analysis of solutions for Maxwell–Chern–Simons O(3) sigma
Calculus of Variations and Partial Differential Equations,
580 (4):0 147, jul 2019.
ISSN 1432-0835.
URL <https://doi.org/10.1007/s00526-019-1590-4>.
[Chiacchio and Ricciardi, 2007]
Francesco Chiacchio and Tonia Ricciardi.
Multiple vortices for a self-dual CP(1) Maxwell-Chern-Simons model.
Nonlinear Differential Equations and Applications, 130
(5-6):0 563–584, 2007.
ISSN 10219722.
[Chipot, 2011]
Michel Chipot.
Handbook of differential equations: stationary partial
differential equations.
Elsevier, 2011.
[Christ, 1991]
Michael Christ.
On the $\bar\partial$ equation in weighted $\mathrm{L}^2$ norms in
The Journal of Geometric Analysis, 10 (3):0
193–230, 1991.
[Collie and Tong, 2008]
Benjamin Collie and David Tong.
Dynamics of Chern-Simons vortices.
Physical Review D - Particles, Fields, Gravitation and
Cosmology, 2008.
ISSN 15507998.
[Deimling, 2010]
Klaus Deimling.
Nonlinear functional analysis.
Courier Corporation, 2010.
[Demoulini and Stuart, 2009]
Sophia Demoulini and David Stuart.
Adiabatic Limit and the Slow Motion of Vortices in a
Chern-Simons-Schrödinger System.
Communications in Mathematical Physics, 2900
(2):0 597–632, sep 2009.
ISSN 0010-3616, 1432-0916.
[Evans, 2010]
L.C. Evans.
Partial Differential Equations.
Graduate studies in mathematics. American Mathematical Society, 2010.
ISBN 9780821849743.
[Flood and Speight, 2018]
S. P. Flood and J. M. Speight.
Chern-Simons deformation of vortices on compact domains.
Journal of Geometry and Physics, 133:0 153–167, 2018.
ISSN 03930440.
URL <https://doi.org/10.1016/j.geomphys.2018.07.009>.
[Gilbarg and Trudinger, 2015]
David Gilbarg and Neil S Trudinger.
Elliptic partial differential equations of second order.
springer, 2015.
[Goldstein et al., 2002]
Herbert Goldstein, Charles Poole, and John Safko.
Classical mechanics, 2002.
[Han, 2000]
Jongmin Han.
Existence of topological multivortex solutions in the self-dual gauge
Proceedings of the Royal Society of Edinburgh Section A:
Mathematics, 1300 (6):0 1293–1309, 2000.
[Han and Lin, 2014]
Jongmin Han and Chang-Shou Lin.
Multiplicity for Self-Dual Condensate Solutions in the
Maxwell-Chern-Simons O(3) Sigma Model.
Communications in Partial Differential Equations, 390
(8):0 1424–1450, 2014.
URL <https://doi.org/10.1080/03605302.2014.908909>.
[Han and Nam, 2005]
Jongmin Han and Hee Seok Nam.
On the topological multivortex solutions of the self-dual
Maxwell-Chern-Simons gauged O(3) sigma model.
Letters in Mathematical Physics, 730 (1):0
17–31, 2005.
ISSN 03779017.
[Han and Song, 2011]
Jongmin Han and Kyungwoo Song.
Existence and asymptotics of topological solutions in the self-dual
Maxwell-Chern-Simons O(3) sigma model.
Journal of Differential Equations, 2500 (1):0
204–222, 2011.
ISSN 00220396.
URL <http://dx.doi.org/10.1016/j.jde.2010.08.003>.
[Hwang, 2003]
AD Hwang.
A symplectic look at surfaces of revolution.
Enseignement Mathematique, pages 1–17, 2003.
[Jackiw et al., 1990]
R Jackiw, Kimyeong Lee, and Erick J. Weinberg.
Self-dual Chern-Simons solitons.
420 (10):0 3488–3499, 1990.
[Jaffe and Taubes, 1980]
A Jaffe and CH Taubes.
Monopoles and Vortices.
Birkhäuser, Basel, 1980.
[Kim and Min, 1992]
Sung Ku Kim and Hyunsoo Min.
Statistical interactions between chern-simons vortices.
Physics Letters B, 2810 (1):0 81 – 85, 1992.
ISSN 0370-2693.
[Kim and Lee, 1994]
Yoonbai Kim and Kimyeong Lee.
Vortex dynamics in self-dual Chern-Simons-Higgs systems.
Physical Review D, 490 (4):0 2041–2054, feb
URL <http://0.link.aps.org/doi/10.1103/PhysRevD.49.2041>.
[Kim and Lee, 2002]
Yoonbai Kim and Kimyeong Lee.
First and second order vortex dynamics.
Physical Review D, 660 (4):0 045016, aug 2002.
URL <http://0.link.aps.org/doi/10.1103/PhysRevD.66.045016>.
[Kimm et al., 1996]
Kyoungtae Kimm, Kimyeong Lee, and Taejin Lee.
Anyonic Bogomol'nyi solitons in a gauged O(3) $\sigma$ model.
Phys. Rev. D, 530 (8):0 4436–4440, 1996.
URL <https://link.aps.org/doi/10.1103/PhysRevD.53.4436>.
[Krusch and Speight, 2010]
Steffen Krusch and J Martin Speight.
Exact moduli space metrics for hyperbolic vortex polygons.
Journal of Mathematical Physics, 510 (2):0
022304, 2010.
[Lee et al., 1990]
Choonkyu Lee, Kimyeong Lee, and Hyunsoo Min.
Self-dual Maxwell Chern-Simons solitons.
Physics Letters B, 2520 (1):0 79–83, 1990.
ISSN 0370-2693.
[Lieb, 1997]
Elliott H. Lieb.
Graduate studies in mathematics, v. 14. American Mathematical
Society, Providence, R.I, 1997.
ISBN 0821806327.
[Manton and Sutcliffe, 2004]
N. Manton and P. Sutcliffe.
Topological Solitons.
Cambridge Monographs on Mathematical Physics. Cambridge
University Press, 2004.
ISBN 978-1-139-45469-8.
[Manton, 1982]
N. S. Manton.
A remark on the scattering of BPS monopoles.
Physics Letters B, 1100 (1):0 54 – 56, 1982.
ISSN 0370-2693.
[Manton, 1997]
N. S. Manton.
First Order Vortex Dynamics.
Annals of Physics, 2560 (1):0 114–131, may
ISSN 0003-4916.
[Manton and Nasir, 1999]
N S Manton and S M Nasir.
Conservation laws in a first-order dynamical system of vortices.
Nonlinearity, 120 (4):0 851,
URL <http://stacks.iop.org/0951-7715/12/i=4/a=306>.
[Manton and Nasir, 1999]
N S Manton and S M Nasir.
Volume of vortex moduli spaces.
Communications in mathematical physics, 1990
(3):0 591–604, 1999b.
[Manton and Nasir, 1999]
N S Manton and S M Nasir.
Conservation laws in a first-order dynamical system of vortices.
Nonlinearity, 120 (4):0 851,
URL <http://stacks.iop.org/0951-7715/12/i=4/a=306>.
[Manton, 1993]
N.S. Manton.
Statistical mechanics of vortices.
Nuclear Physics B, 4000 (1):0 624 – 632,
ISSN 0550-3213.
[Manton and Nasir, 1999]
NS Manton and SM Nasir.
Volume of vortex moduli spaces.
Communications in mathematical physics, 1990
(3):0 591–604, 1999d.
[Manton and Speight, 2003]
N.S. Manton and J.M. Speight.
Asymptotic interactions of critically coupled vortices.
Communications in Mathematical Physics, 2360
(3):0 535–555, Jun 2003.
ISSN 1432-0916.
URL <https://doi.org/10.1007/s00220-003-0842-4>.
[Mawhin, 1999]
Jean Mawhin.
Leray-Schauder degree: a half century of extensions and
Topological Methods in Nonlinear Analysis, 140
(2):0 195, 1999.
ISSN 1230-3429.
[Mitrea and Taylor, 1999]
Marius Mitrea and Michael Taylor.
Boundary layer methods for lipschitz domains in riemannian manifolds.
Journal of Functional Analysis, 1630 (2):0
181–251, 1999.
[Mitrea and Taylor, 2000]
Marius Mitrea and Michael Taylor.
Potential theory on lipschitz domains in riemannian manifolds: Holder
continuous metric tensors.
Communications in Partial Differential Equations, 250
(7-8):0 1487–1536, 2000.
[Nagy, 2017]
Ákos Nagy.
The Berry Connection of the Ginzburg-Landau Vortices.
Communications in Mathematical Physics, 3500
(1):0 105–128, 2 2017.
ISSN 0010-3616.
URL <http://link.springer.com/10.1007/s00220-016-2701-0>.
[Payne and Weinberger, 1960]
L. E. Payne and H. F. Weinberger.
An optimal poincaré inequality for convex domains.
Archive for Rational Mechanics and Analysis, 50
(1):0 286–292, 1960.
ISSN 1432-0673.
URL <https://doi.org/10.1007/BF00252910>.
[Romão and Speight, 2020]
N. M. Romão and J. M. Speight.
The Geometry of the Space of BPS Vortex–Antivortex Pairs.
Communications in Mathematical Physics, 3790
(2):0 723–772, 2020.
ISSN 14320916.
URL <https://doi.org/10.1007/s00220-020-03824-y>.
[Romao, 2001]
Nuno M Romao.
Quantum chern–simons vortices on a sphere.
Journal of Mathematical Physics, 420 (8):0
3445–3469, 2001.
[Samols, 1992]
T M Samols.
Mathematical Physics Vortex Scattering.
Commun. Math. Phys, 1450 (1):0 149–179, 1992.
URL <https://projecteuclid.org:443/euclid.cmp/1104249538>.
[Schaefer, 1955]
Helmut Schaefer.
Über die Methode der a priori-Schranken.
Mathematische Annalen, 129:0 415–416, 1955.
[Schroers, 1996]
B. J. Schroers.
The spectrum of Bogomol'nyi solitons in gauged linear sigma models.
Nuclear Physics B, 4750 (1):0 440 – 468,
ISSN 0550-3213.
[Schroers, 1995]
BJ Schroers.
Bogomol'nyi solitons in a gauged O (3) sigma model.
Physics Letters B, 3560 (2-3):0 291–296,
[Sibner et al., 2000]
Lesley Sibner, Robert Sibner, and Yisong Yang.
Abelian gauge theory on Riemann surfaces and new topological
Proceedings of the Royal Society of London A: Mathematical,
Physical and Engineering Sciences, 4560 (1995):0 593–613,
ISSN 1364-5021.
[Speight, 1997]
J. M. Speight.
Static intervortex forces.
Phys. Rev. D, 550 (6):0 3830–3835, mar 1997.
URL <https://link.aps.org/doi/10.1103/PhysRevD.55.3830>.
[Spruck and Yang, 2009]
Joel Spruck and Yisong Yang.
Proof of the Julia–Zee Theorem.
Communications in Mathematical Physics, 2910
(2):0 347–356, oct 2009.
ISSN 0010-3616, 1432-0916.
[Strachan, 1992]
I. A. B. Strachan.
Low-velocity scattering of vortices in a modified Abelian Higgs
Journal of Mathematical Physics, 330 (1):0
102–110, 1992.
URL <https://doi.org/10.1063/1.529949>.
[Stuart, 1994]
D. Stuart.
Dynamics of abelian Higgs vortices in the near Bogomolny regime.
Communications in Mathematical Physics, 1590
(1):0 51–91, 1994.
URL <https://projecteuclid.org:443/euclid.cmp/1104254491>.
[Taubes, 1980]
Clifford Henry Taubes.
Arbitrary $n$-vortex solutions to the first order ginzburg-landau
Comm. Math. Phys., 720 (3):0 277–292, 1980.
URL <https://projecteuclid.org:443/euclid.cmp/1103907703>.
[Tong, 2002]
David Tong.
Ns5-branes, t-duality and worldsheet instantons.
Journal of High Energy Physics, 20020 (07):0
013, 2002.
[Tong and Wong, 2014]
David Tong and Kenny Wong.
Vortices and impurities.
Journal of High Energy Physics, 20140 (1):0
90, jan 2014.
ISSN 1029-8479.
URL <http://link.springer.com/10.1007/JHEP01(2014)090>.
[Yang, 1999]
Yisong Yang.
Strings of opposite magnetic charges in a gauge field theory.
In Proceedings of the Royal Society of London A: Mathematical,
Physical and Engineering Sciences, volume 455, pages 601–629. The Royal
Society, 1999a.
[Yang, 1999]
Yisong Yang.
Strings of opposite magnetic charges in a gauge field theory.
In Proceedings of the Royal Society of London A:
Mathematical, Physical and Engineering Sciences, volume 455, pages
601–629. The Royal Society, 1999b.
[Yang, 2000]
Yisong Yang.
On a System of Nonlinear Elliptic Equations Arising in
Theoretical Physics.
Journal of Functional Analysis, 1700 (1):0 1
– 36, 2000.
ISSN 0022-1236.
[Zhongwei, 1995]
Shen Zhongwei.
$L^p$ estimates for Schrodinger operators with certain potentials.
Annales de l'Institut Henri Poincare, 450
(2):0 513–546, 1995.
[Alqahtani and Speight, 2015]
LS Alqahtani and JM Speight.
Ricci magnetic geodesic motion of vortices and lumps.
Journal of Geometry and Physics, 98:0 556–574, 2015.
[Aubin, 2013]
Thierry Aubin.
Some nonlinear problems in Riemannian geometry.
Springer Science & Business Media, 2013.
[Bradlow, 1990]
Steven B Bradlow.
Vortices in Holomorphic Line Bundles over Closed Kahler Manifolds.
Commun. Math. Phys, 1350 (1):0 1–17, 1990.
URL <https://projecteuclid.org:443/euclid.cmp/1104201917>.
[Carroll, 2004]
Sean Carroll.
Spacetime and geometry : an introduction to general
Addison Wesley, San Francisco, 2004.
ISBN 0805387323.
[Chae and Nam, 2001]
D Chae and Hee Seok Nam.
On the condensate multivortex solutions of the self-dual
maxwell-chern-simmons cp1 model.
Annales de l'Institut Henri Poincare, 0 (2):0
887–906, 2001.
ISSN 03779017.
[Chen and Chern, 2019]
Zhi-You Chen and Jann-Long Chern.
The analysis of solutions for Maxwell–Chern–Simons O(3) sigma
Calculus of Variations and Partial Differential Equations,
580 (4):0 147, jul 2019.
ISSN 1432-0835.
URL <https://doi.org/10.1007/s00526-019-1590-4>.
[Chiacchio and Ricciardi, 2007]
Francesco Chiacchio and Tonia Ricciardi.
Multiple vortices for a self-dual CP(1) Maxwell-Chern-Simons model.
Nonlinear Differential Equations and Applications, 130
(5-6):0 563–584, 2007.
ISSN 10219722.
[Chipot, 2011]
Michel Chipot.
Handbook of differential equations: stationary partial
differential equations.
Elsevier, 2011.
[Christ, 1991]
Michael Christ.
On the $\bar\partial$ equation in weighted $\mathrm{L}^2$ norms in
The Journal of Geometric Analysis, 10 (3):0
193–230, 1991.
[Collie and Tong, 2008]
Benjamin Collie and David Tong.
Dynamics of Chern-Simons vortices.
Physical Review D - Particles, Fields, Gravitation and
Cosmology, 2008.
ISSN 15507998.
[Deimling, 2010]
Klaus Deimling.
Nonlinear functional analysis.
Courier Corporation, 2010.
[Demoulini and Stuart, 2009]
Sophia Demoulini and David Stuart.
Adiabatic Limit and the Slow Motion of Vortices in a
Chern-Simons-Schrödinger System.
Communications in Mathematical Physics, 2900
(2):0 597–632, sep 2009.
ISSN 0010-3616, 1432-0916.
[Evans, 2010]
L.C. Evans.
Partial Differential Equations.
Graduate studies in mathematics. American Mathematical Society, 2010.
ISBN 9780821849743.
[Flood and Speight, 2018]
S. P. Flood and J. M. Speight.
Chern-Simons deformation of vortices on compact domains.
Journal of Geometry and Physics, 133:0 153–167, 2018.
ISSN 03930440.
URL <https://doi.org/10.1016/j.geomphys.2018.07.009>.
[Gilbarg and Trudinger, 2015]
David Gilbarg and Neil S Trudinger.
Elliptic partial differential equations of second order.
springer, 2015.
[Goldstein et al., 2002]
Herbert Goldstein, Charles Poole, and John Safko.
Classical mechanics, 2002.
[Han, 2000]
Jongmin Han.
Existence of topological multivortex solutions in the self-dual gauge
Proceedings of the Royal Society of Edinburgh Section A:
Mathematics, 1300 (6):0 1293–1309, 2000.
[Han and Lin, 2014]
Jongmin Han and Chang-Shou Lin.
Multiplicity for Self-Dual Condensate Solutions in the
Maxwell-Chern-Simons O(3) Sigma Model.
Communications in Partial Differential Equations, 390
(8):0 1424–1450, 2014.
URL <https://doi.org/10.1080/03605302.2014.908909>.
[Han and Nam, 2005]
Jongmin Han and Hee Seok Nam.
On the topological multivortex solutions of the self-dual
Maxwell-Chern-Simons gauged O(3) sigma model.
Letters in Mathematical Physics, 730 (1):0
17–31, 2005.
ISSN 03779017.
[Han and Song, 2011]
Jongmin Han and Kyungwoo Song.
Existence and asymptotics of topological solutions in the self-dual
Maxwell-Chern-Simons O(3) sigma model.
Journal of Differential Equations, 2500 (1):0
204–222, 2011.
ISSN 00220396.
URL <http://dx.doi.org/10.1016/j.jde.2010.08.003>.
[Hwang, 2003]
AD Hwang.
A symplectic look at surfaces of revolution.
Enseignement Mathematique, pages 1–17, 2003.
[Jackiw et al., 1990]
R Jackiw, Kimyeong Lee, and Erick J. Weinberg.
Self-dual Chern-Simons solitons.
420 (10):0 3488–3499, 1990.
[Jaffe and Taubes, 1980]
A Jaffe and CH Taubes.
Monopoles and Vortices.
Birkhäuser, Basel, 1980.
[Kim and Min, 1992]
Sung Ku Kim and Hyunsoo Min.
Statistical interactions between chern-simons vortices.
Physics Letters B, 2810 (1):0 81 – 85, 1992.
ISSN 0370-2693.
[Kim and Lee, 1994]
Yoonbai Kim and Kimyeong Lee.
Vortex dynamics in self-dual Chern-Simons-Higgs systems.
Physical Review D, 490 (4):0 2041–2054, feb
URL <http://0.link.aps.org/doi/10.1103/PhysRevD.49.2041>.
[Kim and Lee, 2002]
Yoonbai Kim and Kimyeong Lee.
First and second order vortex dynamics.
Physical Review D, 660 (4):0 045016, aug 2002.
URL <http://0.link.aps.org/doi/10.1103/PhysRevD.66.045016>.
[Kimm et al., 1996]
Kyoungtae Kimm, Kimyeong Lee, and Taejin Lee.
Anyonic Bogomol'nyi solitons in a gauged O(3) $\sigma$ model.
Phys. Rev. D, 530 (8):0 4436–4440, 1996.
URL <https://link.aps.org/doi/10.1103/PhysRevD.53.4436>.
[Krusch and Speight, 2010]
Steffen Krusch and J Martin Speight.
Exact moduli space metrics for hyperbolic vortex polygons.
Journal of Mathematical Physics, 510 (2):0
022304, 2010.
[Lee et al., 1990]
Choonkyu Lee, Kimyeong Lee, and Hyunsoo Min.
Self-dual Maxwell Chern-Simons solitons.
Physics Letters B, 2520 (1):0 79–83, 1990.
ISSN 0370-2693.
[Lieb, 1997]
Elliott H. Lieb.
Graduate studies in mathematics, v. 14. American Mathematical
Society, Providence, R.I, 1997.
ISBN 0821806327.
[Manton and Sutcliffe, 2004]
N. Manton and P. Sutcliffe.
Topological Solitons.
Cambridge Monographs on Mathematical Physics. Cambridge
University Press, 2004.
ISBN 978-1-139-45469-8.
[Manton, 1982]
N. S. Manton.
A remark on the scattering of BPS monopoles.
Physics Letters B, 1100 (1):0 54 – 56, 1982.
ISSN 0370-2693.
[Manton, 1997]
N. S. Manton.
First Order Vortex Dynamics.
Annals of Physics, 2560 (1):0 114–131, may
ISSN 0003-4916.
[Manton and Nasir, 1999]
N S Manton and S M Nasir.
Conservation laws in a first-order dynamical system of vortices.
Nonlinearity, 120 (4):0 851,
URL <http://stacks.iop.org/0951-7715/12/i=4/a=306>.
[Manton and Nasir, 1999]
N S Manton and S M Nasir.
Volume of vortex moduli spaces.
Communications in mathematical physics, 1990
(3):0 591–604, 1999b.
[Manton and Nasir, 1999]
N S Manton and S M Nasir.
Conservation laws in a first-order dynamical system of vortices.
Nonlinearity, 120 (4):0 851,
URL <http://stacks.iop.org/0951-7715/12/i=4/a=306>.
[Manton, 1993]
N.S. Manton.
Statistical mechanics of vortices.
Nuclear Physics B, 4000 (1):0 624 – 632,
ISSN 0550-3213.
[Manton and Nasir, 1999]
NS Manton and SM Nasir.
Volume of vortex moduli spaces.
Communications in mathematical physics, 1990
(3):0 591–604, 1999d.
[Manton and Speight, 2003]
N.S. Manton and J.M. Speight.
Asymptotic interactions of critically coupled vortices.
Communications in Mathematical Physics, 2360
(3):0 535–555, Jun 2003.
ISSN 1432-0916.
URL <https://doi.org/10.1007/s00220-003-0842-4>.
[Mawhin, 1999]
Jean Mawhin.
Leray-Schauder degree: a half century of extensions and
Topological Methods in Nonlinear Analysis, 140
(2):0 195, 1999.
ISSN 1230-3429.
[Mitrea and Taylor, 1999]
Marius Mitrea and Michael Taylor.
Boundary layer methods for lipschitz domains in riemannian manifolds.
Journal of Functional Analysis, 1630 (2):0
181–251, 1999.
[Mitrea and Taylor, 2000]
Marius Mitrea and Michael Taylor.
Potential theory on lipschitz domains in riemannian manifolds: Holder
continuous metric tensors.
Communications in Partial Differential Equations, 250
(7-8):0 1487–1536, 2000.
[Nagy, 2017]
Ákos Nagy.
The Berry Connection of the Ginzburg-Landau Vortices.
Communications in Mathematical Physics, 3500
(1):0 105–128, 2 2017.
ISSN 0010-3616.
URL <http://link.springer.com/10.1007/s00220-016-2701-0>.
[Payne and Weinberger, 1960]
L. E. Payne and H. F. Weinberger.
An optimal poincaré inequality for convex domains.
Archive for Rational Mechanics and Analysis, 50
(1):0 286–292, 1960.
ISSN 1432-0673.
URL <https://doi.org/10.1007/BF00252910>.
[Romão and Speight, 2020]
N. M. Romão and J. M. Speight.
The Geometry of the Space of BPS Vortex–Antivortex Pairs.
Communications in Mathematical Physics, 3790
(2):0 723–772, 2020.
ISSN 14320916.
URL <https://doi.org/10.1007/s00220-020-03824-y>.
[Romao, 2001]
Nuno M Romao.
Quantum chern–simons vortices on a sphere.
Journal of Mathematical Physics, 420 (8):0
3445–3469, 2001.
[Samols, 1992]
T M Samols.
Mathematical Physics Vortex Scattering.
Commun. Math. Phys, 1450 (1):0 149–179, 1992.
URL <https://projecteuclid.org:443/euclid.cmp/1104249538>.
[Schaefer, 1955]
Helmut Schaefer.
Über die Methode der a priori-Schranken.
Mathematische Annalen, 129:0 415–416, 1955.
[Schroers, 1996]
B. J. Schroers.
The spectrum of Bogomol'nyi solitons in gauged linear sigma models.
Nuclear Physics B, 4750 (1):0 440 – 468,
ISSN 0550-3213.
[Schroers, 1995]
BJ Schroers.
Bogomol'nyi solitons in a gauged O (3) sigma model.
Physics Letters B, 3560 (2-3):0 291–296,
[Sibner et al., 2000]
Lesley Sibner, Robert Sibner, and Yisong Yang.
Abelian gauge theory on Riemann surfaces and new topological
Proceedings of the Royal Society of London A: Mathematical,
Physical and Engineering Sciences, 4560 (1995):0 593–613,
ISSN 1364-5021.
[Speight, 1997]
J. M. Speight.
Static intervortex forces.
Phys. Rev. D, 550 (6):0 3830–3835, mar 1997.
URL <https://link.aps.org/doi/10.1103/PhysRevD.55.3830>.
[Spruck and Yang, 2009]
Joel Spruck and Yisong Yang.
Proof of the Julia–Zee Theorem.
Communications in Mathematical Physics, 2910
(2):0 347–356, oct 2009.
ISSN 0010-3616, 1432-0916.
[Strachan, 1992]
I. A. B. Strachan.
Low-velocity scattering of vortices in a modified Abelian Higgs
Journal of Mathematical Physics, 330 (1):0
102–110, 1992.
URL <https://doi.org/10.1063/1.529949>.
[Stuart, 1994]
D. Stuart.
Dynamics of abelian Higgs vortices in the near Bogomolny regime.
Communications in Mathematical Physics, 1590
(1):0 51–91, 1994.
URL <https://projecteuclid.org:443/euclid.cmp/1104254491>.
[Taubes, 1980]
Clifford Henry Taubes.
Arbitrary $n$-vortex solutions to the first order ginzburg-landau
Comm. Math. Phys., 720 (3):0 277–292, 1980.
URL <https://projecteuclid.org:443/euclid.cmp/1103907703>.
[Tong, 2002]
David Tong.
Ns5-branes, t-duality and worldsheet instantons.
Journal of High Energy Physics, 20020 (07):0
013, 2002.
[Tong and Wong, 2014]
David Tong and Kenny Wong.
Vortices and impurities.
Journal of High Energy Physics, 20140 (1):0
90, jan 2014.
ISSN 1029-8479.
URL <http://link.springer.com/10.1007/JHEP01(2014)090>.
[Yang, 1999]
Yisong Yang.
Strings of opposite magnetic charges in a gauge field theory.
In Proceedings of the Royal Society of London A: Mathematical,
Physical and Engineering Sciences, volume 455, pages 601–629. The Royal
Society, 1999a.
[Yang, 1999]
Yisong Yang.
Strings of opposite magnetic charges in a gauge field theory.
In Proceedings of the Royal Society of London A:
Mathematical, Physical and Engineering Sciences, volume 455, pages
601–629. The Royal Society, 1999b.
[Yang, 2000]
Yisong Yang.
On a System of Nonlinear Elliptic Equations Arising in
Theoretical Physics.
Journal of Functional Analysis, 1700 (1):0 1
– 36, 2000.
ISSN 0022-1236.
[Zhongwei, 1995]
Shen Zhongwei.
$L^p$ estimates for Schrodinger operators with certain potentials.
Annales de l'Institut Henri Poincare, 450
(2):0 513–546, 1995.
[Alqahtani and Speight, 2015]
LS Alqahtani and JM Speight.
Ricci magnetic geodesic motion of vortices and lumps.
Journal of Geometry and Physics, 98:0 556–574, 2015.
[Aubin, 2013]
Thierry Aubin.
Some nonlinear problems in Riemannian geometry.
Springer Science & Business Media, 2013.
[Bradlow, 1990]
Steven B Bradlow.
Vortices in Holomorphic Line Bundles over Closed Kahler Manifolds.
Commun. Math. Phys, 1350 (1):0 1–17, 1990.
URL <https://projecteuclid.org:443/euclid.cmp/1104201917>.
[Carroll, 2004]
Sean Carroll.
Spacetime and geometry : an introduction to general
Addison Wesley, San Francisco, 2004.
ISBN 0805387323.
[Chae and Nam, 2001]
D Chae and Hee Seok Nam.
On the condensate multivortex solutions of the self-dual
maxwell-chern-simmons cp1 model.
Annales de l'Institut Henri Poincare, 0 (2):0
887–906, 2001.
ISSN 03779017.
[Chen and Chern, 2019]
Zhi-You Chen and Jann-Long Chern.
The analysis of solutions for Maxwell–Chern–Simons O(3) sigma
Calculus of Variations and Partial Differential Equations,
580 (4):0 147, jul 2019.
ISSN 1432-0835.
URL <https://doi.org/10.1007/s00526-019-1590-4>.
[Chiacchio and Ricciardi, 2007]
Francesco Chiacchio and Tonia Ricciardi.
Multiple vortices for a self-dual CP(1) Maxwell-Chern-Simons model.
Nonlinear Differential Equations and Applications, 130
(5-6):0 563–584, 2007.
ISSN 10219722.
[Chipot, 2011]
Michel Chipot.
Handbook of differential equations: stationary partial
differential equations.
Elsevier, 2011.
[Christ, 1991]
Michael Christ.
On the $\bar\partial$ equation in weighted $\mathrm{L}^2$ norms in
The Journal of Geometric Analysis, 10 (3):0
193–230, 1991.
[Collie and Tong, 2008]
Benjamin Collie and David Tong.
Dynamics of Chern-Simons vortices.
Physical Review D - Particles, Fields, Gravitation and
Cosmology, 2008.
ISSN 15507998.
[Deimling, 2010]
Klaus Deimling.
Nonlinear functional analysis.
Courier Corporation, 2010.
[Demoulini and Stuart, 2009]
Sophia Demoulini and David Stuart.
Adiabatic Limit and the Slow Motion of Vortices in a
Chern-Simons-Schrödinger System.
Communications in Mathematical Physics, 2900
(2):0 597–632, sep 2009.
ISSN 0010-3616, 1432-0916.
[Evans, 2010]
L.C. Evans.
Partial Differential Equations.
Graduate studies in mathematics. American Mathematical Society, 2010.
ISBN 9780821849743.
[Flood and Speight, 2018]
S. P. Flood and J. M. Speight.
Chern-Simons deformation of vortices on compact domains.
Journal of Geometry and Physics, 133:0 153–167, 2018.
ISSN 03930440.
URL <https://doi.org/10.1016/j.geomphys.2018.07.009>.
[Gilbarg and Trudinger, 2015]
David Gilbarg and Neil S Trudinger.
Elliptic partial differential equations of second order.
springer, 2015.
[Goldstein et al., 2002]
Herbert Goldstein, Charles Poole, and John Safko.
Classical mechanics, 2002.
[Han, 2000]
Jongmin Han.
Existence of topological multivortex solutions in the self-dual gauge
Proceedings of the Royal Society of Edinburgh Section A:
Mathematics, 1300 (6):0 1293–1309, 2000.
[Han and Lin, 2014]
Jongmin Han and Chang-Shou Lin.
Multiplicity for Self-Dual Condensate Solutions in the
Maxwell-Chern-Simons O(3) Sigma Model.
Communications in Partial Differential Equations, 390
(8):0 1424–1450, 2014.
URL <https://doi.org/10.1080/03605302.2014.908909>.
[Han and Nam, 2005]
Jongmin Han and Hee Seok Nam.
On the topological multivortex solutions of the self-dual
Maxwell-Chern-Simons gauged O(3) sigma model.
Letters in Mathematical Physics, 730 (1):0
17–31, 2005.
ISSN 03779017.
[Han and Song, 2011]
Jongmin Han and Kyungwoo Song.
Existence and asymptotics of topological solutions in the self-dual
Maxwell-Chern-Simons O(3) sigma model.
Journal of Differential Equations, 2500 (1):0
204–222, 2011.
ISSN 00220396.
URL <http://dx.doi.org/10.1016/j.jde.2010.08.003>.
[Hwang, 2003]
AD Hwang.
A symplectic look at surfaces of revolution.
Enseignement Mathematique, pages 1–17, 2003.
[Jackiw et al., 1990]
R Jackiw, Kimyeong Lee, and Erick J. Weinberg.
Self-dual Chern-Simons solitons.
420 (10):0 3488–3499, 1990.
[Jaffe and Taubes, 1980]
A Jaffe and CH Taubes.
Monopoles and Vortices.
Birkhäuser, Basel, 1980.
[Kim and Min, 1992]
Sung Ku Kim and Hyunsoo Min.
Statistical interactions between chern-simons vortices.
Physics Letters B, 2810 (1):0 81 – 85, 1992.
ISSN 0370-2693.
[Kim and Lee, 1994]
Yoonbai Kim and Kimyeong Lee.
Vortex dynamics in self-dual Chern-Simons-Higgs systems.
Physical Review D, 490 (4):0 2041–2054, feb
URL <http://0.link.aps.org/doi/10.1103/PhysRevD.49.2041>.
[Kim and Lee, 2002]
Yoonbai Kim and Kimyeong Lee.
First and second order vortex dynamics.
Physical Review D, 660 (4):0 045016, aug 2002.
URL <http://0.link.aps.org/doi/10.1103/PhysRevD.66.045016>.
[Kimm et al., 1996]
Kyoungtae Kimm, Kimyeong Lee, and Taejin Lee.
Anyonic Bogomol'nyi solitons in a gauged O(3) $\sigma$ model.
Phys. Rev. D, 530 (8):0 4436–4440, 1996.
URL <https://link.aps.org/doi/10.1103/PhysRevD.53.4436>.
[Krusch and Speight, 2010]
Steffen Krusch and J Martin Speight.
Exact moduli space metrics for hyperbolic vortex polygons.
Journal of Mathematical Physics, 510 (2):0
022304, 2010.
[Lee et al., 1990]
Choonkyu Lee, Kimyeong Lee, and Hyunsoo Min.
Self-dual Maxwell Chern-Simons solitons.
Physics Letters B, 2520 (1):0 79–83, 1990.
ISSN 0370-2693.
[Lieb, 1997]
Elliott H. Lieb.
Graduate studies in mathematics, v. 14. American Mathematical
Society, Providence, R.I, 1997.
ISBN 0821806327.
[Manton and Sutcliffe, 2004]
N. Manton and P. Sutcliffe.
Topological Solitons.
Cambridge Monographs on Mathematical Physics. Cambridge
University Press, 2004.
ISBN 978-1-139-45469-8.
[Manton, 1982]
N. S. Manton.
A remark on the scattering of BPS monopoles.
Physics Letters B, 1100 (1):0 54 – 56, 1982.
ISSN 0370-2693.
[Manton, 1997]
N. S. Manton.
First Order Vortex Dynamics.
Annals of Physics, 2560 (1):0 114–131, may
ISSN 0003-4916.
[Manton and Nasir, 1999]
N S Manton and S M Nasir.
Conservation laws in a first-order dynamical system of vortices.
Nonlinearity, 120 (4):0 851,
URL <http://stacks.iop.org/0951-7715/12/i=4/a=306>.
[Manton and Nasir, 1999]
N S Manton and S M Nasir.
Volume of vortex moduli spaces.
Communications in mathematical physics, 1990
(3):0 591–604, 1999b.
[Manton and Nasir, 1999]
N S Manton and S M Nasir.
Conservation laws in a first-order dynamical system of vortices.
Nonlinearity, 120 (4):0 851,
URL <http://stacks.iop.org/0951-7715/12/i=4/a=306>.
[Manton, 1993]
N.S. Manton.
Statistical mechanics of vortices.
Nuclear Physics B, 4000 (1):0 624 – 632,
ISSN 0550-3213.
[Manton and Nasir, 1999]
NS Manton and SM Nasir.
Volume of vortex moduli spaces.
Communications in mathematical physics, 1990
(3):0 591–604, 1999d.
[Manton and Speight, 2003]
N.S. Manton and J.M. Speight.
Asymptotic interactions of critically coupled vortices.
Communications in Mathematical Physics, 2360
(3):0 535–555, Jun 2003.
ISSN 1432-0916.
URL <https://doi.org/10.1007/s00220-003-0842-4>.
[Mawhin, 1999]
Jean Mawhin.
Leray-Schauder degree: a half century of extensions and
Topological Methods in Nonlinear Analysis, 140
(2):0 195, 1999.
ISSN 1230-3429.
[Mitrea and Taylor, 1999]
Marius Mitrea and Michael Taylor.
Boundary layer methods for lipschitz domains in riemannian manifolds.
Journal of Functional Analysis, 1630 (2):0
181–251, 1999.
[Mitrea and Taylor, 2000]
Marius Mitrea and Michael Taylor.
Potential theory on lipschitz domains in riemannian manifolds: Holder
continuous metric tensors.
Communications in Partial Differential Equations, 250
(7-8):0 1487–1536, 2000.
[Nagy, 2017]
Ákos Nagy.
The Berry Connection of the Ginzburg-Landau Vortices.
Communications in Mathematical Physics, 3500
(1):0 105–128, 2 2017.
ISSN 0010-3616.
URL <http://link.springer.com/10.1007/s00220-016-2701-0>.
[Payne and Weinberger, 1960]
L. E. Payne and H. F. Weinberger.
An optimal poincaré inequality for convex domains.
Archive for Rational Mechanics and Analysis, 50
(1):0 286–292, 1960.
ISSN 1432-0673.
URL <https://doi.org/10.1007/BF00252910>.
[Romão and Speight, 2020]
N. M. Romão and J. M. Speight.
The Geometry of the Space of BPS Vortex–Antivortex Pairs.
Communications in Mathematical Physics, 3790
(2):0 723–772, 2020.
ISSN 14320916.
URL <https://doi.org/10.1007/s00220-020-03824-y>.
[Romao, 2001]
Nuno M Romao.
Quantum chern–simons vortices on a sphere.
Journal of Mathematical Physics, 420 (8):0
3445–3469, 2001.
[Samols, 1992]
T M Samols.
Mathematical Physics Vortex Scattering.
Commun. Math. Phys, 1450 (1):0 149–179, 1992.
URL <https://projecteuclid.org:443/euclid.cmp/1104249538>.
[Schaefer, 1955]
Helmut Schaefer.
Über die Methode der a priori-Schranken.
Mathematische Annalen, 129:0 415–416, 1955.
[Schroers, 1996]
B. J. Schroers.
The spectrum of Bogomol'nyi solitons in gauged linear sigma models.
Nuclear Physics B, 4750 (1):0 440 – 468,
ISSN 0550-3213.
[Schroers, 1995]
BJ Schroers.
Bogomol'nyi solitons in a gauged O (3) sigma model.
Physics Letters B, 3560 (2-3):0 291–296,
[Sibner et al., 2000]
Lesley Sibner, Robert Sibner, and Yisong Yang.
Abelian gauge theory on Riemann surfaces and new topological
Proceedings of the Royal Society of London A: Mathematical,
Physical and Engineering Sciences, 4560 (1995):0 593–613,
ISSN 1364-5021.
[Speight, 1997]
J. M. Speight.
Static intervortex forces.
Phys. Rev. D, 550 (6):0 3830–3835, mar 1997.
URL <https://link.aps.org/doi/10.1103/PhysRevD.55.3830>.
[Spruck and Yang, 2009]
Joel Spruck and Yisong Yang.
Proof of the Julia–Zee Theorem.
Communications in Mathematical Physics, 2910
(2):0 347–356, oct 2009.
ISSN 0010-3616, 1432-0916.
[Strachan, 1992]
I. A. B. Strachan.
Low-velocity scattering of vortices in a modified Abelian Higgs
Journal of Mathematical Physics, 330 (1):0
102–110, 1992.
URL <https://doi.org/10.1063/1.529949>.
[Stuart, 1994]
D. Stuart.
Dynamics of abelian Higgs vortices in the near Bogomolny regime.
Communications in Mathematical Physics, 1590
(1):0 51–91, 1994.
URL <https://projecteuclid.org:443/euclid.cmp/1104254491>.
[Taubes, 1980]
Clifford Henry Taubes.
Arbitrary $n$-vortex solutions to the first order ginzburg-landau
Comm. Math. Phys., 720 (3):0 277–292, 1980.
URL <https://projecteuclid.org:443/euclid.cmp/1103907703>.
[Tong, 2002]
David Tong.
Ns5-branes, t-duality and worldsheet instantons.
Journal of High Energy Physics, 20020 (07):0
013, 2002.
[Tong and Wong, 2014]
David Tong and Kenny Wong.
Vortices and impurities.
Journal of High Energy Physics, 20140 (1):0
90, jan 2014.
ISSN 1029-8479.
URL <http://link.springer.com/10.1007/JHEP01(2014)090>.
[Yang, 1999]
Yisong Yang.
Strings of opposite magnetic charges in a gauge field theory.
In Proceedings of the Royal Society of London A: Mathematical,
Physical and Engineering Sciences, volume 455, pages 601–629. The Royal
Society, 1999a.
[Yang, 1999]
Yisong Yang.
Strings of opposite magnetic charges in a gauge field theory.
In Proceedings of the Royal Society of London A:
Mathematical, Physical and Engineering Sciences, volume 455, pages
601–629. The Royal Society, 1999b.
[Yang, 2000]
Yisong Yang.
On a System of Nonlinear Elliptic Equations Arising in
Theoretical Physics.
Journal of Functional Analysis, 1700 (1):0 1
– 36, 2000.
ISSN 0022-1236.
[Zhongwei, 1995]
Shen Zhongwei.
$L^p$ estimates for Schrodinger operators with certain potentials.
Annales de l'Institut Henri Poincare, 450
(2):0 513–546, 1995.
|
Semiclassical Trans-Series from the Perturbative Hopf-Algebraic DSEs
Semiclassical Trans-Series from the Perturbative
Hopf-Algebraic Dyson–Schwinger Equations:
$\boldsymbol{\phi^{3}}$ QFT in 6 Dimensions††This paper is a contribution to
the Special Issue on Algebraic Structures in Perturbative Quantum Field Theory
in honor of Dirk Kreimer for his 60th birthday. The full collection is
available at https://www.emis.de/journals/SIGMA/Kreimer.html
Michael BORINSKY a, Gerald V. DUNNE b and Max MEYNIG b M. Borinsky, G.V. Dunne
and M. Meynig
a) Nikhef Theory Group, Amsterdam 1098 XG, The Netherlands
<EMAIL_ADDRESS>
b) Department of Physics, University of Connecticut, Storrs CT 06269-3046, USA
<EMAIL_ADDRESS><EMAIL_ADDRESS>
Received April 07, 2021, in final form September 16, 2021; Published online
September 23, 2021
We analyze the asymptotically free massless scalar $\phi^{3}$ quantum field
theory in 6 dimensions, using resurgent asymptotic analysis to find the trans-
series solutions which yield the non-perturbative completion of the divergent
perturbative solutions to the Kreimer–Connes Hopf-algebraic Dyson–Schwinger
equations for the anomalous dimension. This scalar conformal field theory is
asymptotically free and has a real Lipatov instanton. In the Hopf-algebraic
approach we find a trans-series having an intricate Borel singularity
structure, with three distinct but resonant non-perturbative terms, each
repeated in an infinite series. These expansions are in terms of the
renormalized coupling. The resonant structure leads to powers of logarithmic
terms at higher levels of the trans-series, analogous to logarithmic terms
arising from interactions between instantons and anti-instantons, but arising
from a purely perturbative formalism rather than from a semi-classical
analysis.
renormalons; resurgence; non-perturbative corrections; quantum field theory;
renormalization; Hopf algebra; trans-series
81T15; 81Q15; 34E10
## 1 Introduction
The seminal work of Kreimer and Connes showed that there is an underlying
Hopf-algebraic structure to the renormalization of quantum field theory (QFT)
[28, 29, 61]. This new perspective has led to deep insights into QFT, and also
to novel computational methods that have enabled significant progress in
higher order perturbative computations [14, 16, 17, 20, 22, 23, 27, 60, 62,
63, 64, 81, 83, 84, 85, 88, 89, 90, 91]. The Hopf-algebraic formulation of QFT
is inherently perturbative in nature, so an important open question is to
understand how the non-perturbative features of QFT arise naturally within the
perturbative Hopf algebra structure. In a recent paper [18] we showed how this
works for 4 dimensional massless Yukawa theory, using Écalle’s theory of
resurgent trans-series and alien calculus [3, 33, 44, 48, 80, 82]. Here we
extend this analysis to a conformal field theory: massless scalar $\phi^{3}$
theory in six dimensional space-time. This QFT has been studied extensively
from numerous directions, and has many interesting features, both perturbative
and non-perturbative. The theory is asymptotically free for real coupling $g$
[26, 30, 67, 68], and has a Yang–Lee edge singularity when $g$ is imaginary
[50]. The perturbative beta function and anomalous dimensions have been
computed to 4 loop order [53] (and very recently to 5 loop order [19, 20]).
The perturbative Hopf algebra structure of Dyson–Schwinger equations of this
model was formulated in the pioneering papers [22, 23]. On the non-
perturbative side, this QFT has a real Lipatov instanton when $g$ is real, for
which the conventional one-instanton semi-classical analysis [21, 66, 78, 93]
of the fluctuation determinant has been studied [76, 77]. Further extensions
to multi-dimensional cubic interactions have many interesting applications and
implications for conformal quantum field theories in general [11, 19, 38, 39,
49, 52, 53, 54, 55, 56, 57]. For other analyses of resurgence properties of
renormalization group and Dyson–Schwinger equations see [5, 6, 7, 8, 9, 13].
Our technical analysis is based on the fundamental result [22, 23, 62, 63]
that the Dyson–Schwinger equations have a recursive Hopf-algebraic structure
which, when combined with the renormalization group equations describing the
anomalous scaling under re-scaling of parameters and in the absence of vertex
renormalization, reduces the problem to a non-linear ordinary differential
equation (ODE), where the variable is the renormalized coupling. This Hopf-
algebraic approximation goes well beyond the familiar rainbow [42] and chain
[22, 23] approximations to the Dyson–Schwinger equations. These results cast
the Hopf algebra renormalization problem in a form in which very high orders
of perturbation theory become accessible, and as we show here it also enables
direct access to the associated non-perturbative structure. We employ the
trans-series approach to the resurgence properties of non-linear differential
equations, along the lines of [31, 32, 33]. Our main new result is that the
perturbative Hopf algebra formulation encodes a non-perturbative trans-series
that involves powers of all three trans-monomial elements: $x$, ${\rm
e}^{-1/x}$, and $\log(x)$, all expressed in terms of the renormalized
coupling. Moreover, this trans-series has the form of an all-orders multi-
instanton expansion, and the logarithms appear with the characteristic
structure of logarithmic terms arising from the interaction of instantons and
anti-instantons.111This logarithmic structure does not occur for the 4
dimensional Yukawa model studied in [18]. Logarithmic terms are familiar in
semi-classical computations [1, 40, 41, 45, 46, 47, 65, 71, 79, 92, 94, 95],
and have been studied in differential equations where resonant Borel
singularities $\pm A$ interact [4, 37, 51], but here we find a quite different
resonant Borel structure, with three resonant singularities of the same sign
yet in integer multiples. All this non-perturbative information is encoded in
the original perturbative Hopf-algebraic formulation, which at first sight
makes no explicit mention of instantons, let alone interactions between
instantons and anti-instantons.
## 2 Perturbative Hopf-algebraic analysis
of massless $\boldsymbol{\phi^{3}}$ theory in 6 dimensions
In this paper we analyze the massless scalar $\phi^{3}$ theory in 6
dimensional spacetime. This is the critical dimension in which the theory is
asymptotically free [68] and in which it has a Lipatov instanton [66, 76, 77].
We analyze the non-perturbative features arising in the Hopf-algebraic
approach of [22, 23, 62, 63]. The Lagrangian density is
$\displaystyle{\mathcal{L}}=\frac{1}{2}(\partial_{\mu}\phi)^{2}-\frac{g}{3!}\phi^{3}.$
As in [22, 23] we consider the renormalized scalar self-energy
$\displaystyle\Pi\big{(}q^{2}\big{)}:=\leavevmode\hbox to28.85pt{\vbox
to14.63pt{\pgfpicture\makeatletter\hbox{\hskip 0.2pt\lower-7.31319pt\hbox
to0.0pt{\pgfsys@beginscope\pgfsys@invoke{
}\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{
}\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{
}\pgfsys@setlinewidth{0.4pt}\pgfsys@invoke{ }\nullfont\hbox
to0.0pt{\pgfsys@beginscope\pgfsys@invoke{ }
{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{\pgfsys@beginscope\pgfsys@invoke{
}\pgfsys@beginscope\pgfsys@invoke{
}\color[rgb]{1,1,1}\definecolor[named]{pgfstrokecolor}{rgb}{1,1,1}\pgfsys@color@gray@stroke{1}\pgfsys@invoke{
}\pgfsys@color@gray@fill{1}\pgfsys@invoke{
}\definecolor[named]{pgffillcolor}{rgb}{1,1,1}\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@curveto{21.33957pt}{3.92854pt}{18.15492pt}{7.11319pt}{14.22638pt}{7.11319pt}\pgfsys@curveto{10.29784pt}{7.11319pt}{7.11319pt}{3.92854pt}{7.11319pt}{0.0pt}\pgfsys@curveto{7.11319pt}{-3.92854pt}{10.29784pt}{-7.11319pt}{14.22638pt}{-7.11319pt}\pgfsys@curveto{18.15492pt}{-7.11319pt}{21.33957pt}{-3.92854pt}{21.33957pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@fill\pgfsys@invoke{
} \pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope\pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope}\pgfsys@beginscope\pgfsys@invoke{ }\pgfsys@invoke{
}{0.0}{1.0}{0.0pt}{0.0pt}{ \lxSVG@closescope
}{0}\definecolor[named]{pgf@tempcolor}{rgb}{0,0,0}\pgfsys@invoke{\lxSVG@setpatternuncolored@{3}{0}{0}{0}
}{}\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@curveto{21.33957pt}{3.92854pt}{18.15492pt}{7.11319pt}{14.22638pt}{7.11319pt}\pgfsys@curveto{10.29784pt}{7.11319pt}{7.11319pt}{3.92854pt}{7.11319pt}{0.0pt}\pgfsys@curveto{7.11319pt}{-3.92854pt}{10.29784pt}{-7.11319pt}{14.22638pt}{-7.11319pt}\pgfsys@curveto{18.15492pt}{-7.11319pt}{21.33957pt}{-3.92854pt}{21.33957pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
} \pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@moveto{8.11319pt}{0.0pt}\pgfsys@curveto{8.11319pt}{0.55229pt}{7.66548pt}{1.0pt}{7.11319pt}{1.0pt}\pgfsys@curveto{6.5609pt}{1.0pt}{6.11319pt}{0.55229pt}{6.11319pt}{0.0pt}\pgfsys@curveto{6.11319pt}{-0.55229pt}{6.5609pt}{-1.0pt}{7.11319pt}{-1.0pt}\pgfsys@curveto{7.66548pt}{-1.0pt}{8.11319pt}{-0.55229pt}{8.11319pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@moveto{22.33957pt}{0.0pt}\pgfsys@curveto{22.33957pt}{0.55229pt}{21.89186pt}{1.0pt}{21.33957pt}{1.0pt}\pgfsys@curveto{20.78728pt}{1.0pt}{20.33957pt}{0.55229pt}{20.33957pt}{0.0pt}\pgfsys@curveto{20.33957pt}{-0.55229pt}{20.78728pt}{-1.0pt}{21.33957pt}{-1.0pt}\pgfsys@curveto{21.89186pt}{-1.0pt}{22.33957pt}{-0.55229pt}{22.33957pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{0.0pt}{0.0pt}\pgfsys@lineto{7.11319pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@lineto{28.45276pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope{{
{}{}{}{}{}}{{{}}{{}}}}{}{}\hss}\pgfsys@discardpath\pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope\hss}}\lxSVG@closescope\endpgfpicture}}$
and take all propagator self-insertions into account. This Hopf-algebraic
approach is depicted by the Dyson–Schwinger equation
$\displaystyle\begin{aligned} \leavevmode\hbox to28.85pt{\vbox
to14.63pt{\pgfpicture\makeatletter\hbox{\hskip 0.2pt\lower-7.31319pt\hbox
to0.0pt{\pgfsys@beginscope\pgfsys@invoke{
}\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{
}\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{
}\pgfsys@setlinewidth{0.4pt}\pgfsys@invoke{ }\nullfont\hbox
to0.0pt{\pgfsys@beginscope\pgfsys@invoke{ }
{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{\pgfsys@beginscope\pgfsys@invoke{
}\pgfsys@beginscope\pgfsys@invoke{
}\color[rgb]{1,1,1}\definecolor[named]{pgfstrokecolor}{rgb}{1,1,1}\pgfsys@color@gray@stroke{1}\pgfsys@invoke{
}\pgfsys@color@gray@fill{1}\pgfsys@invoke{
}\definecolor[named]{pgffillcolor}{rgb}{1,1,1}\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@curveto{21.33957pt}{3.92854pt}{18.15492pt}{7.11319pt}{14.22638pt}{7.11319pt}\pgfsys@curveto{10.29784pt}{7.11319pt}{7.11319pt}{3.92854pt}{7.11319pt}{0.0pt}\pgfsys@curveto{7.11319pt}{-3.92854pt}{10.29784pt}{-7.11319pt}{14.22638pt}{-7.11319pt}\pgfsys@curveto{18.15492pt}{-7.11319pt}{21.33957pt}{-3.92854pt}{21.33957pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@fill\pgfsys@invoke{
} \pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope\pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope}\pgfsys@beginscope\pgfsys@invoke{
}\definecolor[named]{pgf@tempcolor}{rgb}{0,0,0}\pgfsys@invoke{\lxSVG@setpatternuncolored@{3}{0}{0}{0}
}{}\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@curveto{21.33957pt}{3.92854pt}{18.15492pt}{7.11319pt}{14.22638pt}{7.11319pt}\pgfsys@curveto{10.29784pt}{7.11319pt}{7.11319pt}{3.92854pt}{7.11319pt}{0.0pt}\pgfsys@curveto{7.11319pt}{-3.92854pt}{10.29784pt}{-7.11319pt}{14.22638pt}{-7.11319pt}\pgfsys@curveto{18.15492pt}{-7.11319pt}{21.33957pt}{-3.92854pt}{21.33957pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
} \pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@moveto{8.11319pt}{0.0pt}\pgfsys@curveto{8.11319pt}{0.55229pt}{7.66548pt}{1.0pt}{7.11319pt}{1.0pt}\pgfsys@curveto{6.5609pt}{1.0pt}{6.11319pt}{0.55229pt}{6.11319pt}{0.0pt}\pgfsys@curveto{6.11319pt}{-0.55229pt}{6.5609pt}{-1.0pt}{7.11319pt}{-1.0pt}\pgfsys@curveto{7.66548pt}{-1.0pt}{8.11319pt}{-0.55229pt}{8.11319pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@moveto{22.33957pt}{0.0pt}\pgfsys@curveto{22.33957pt}{0.55229pt}{21.89186pt}{1.0pt}{21.33957pt}{1.0pt}\pgfsys@curveto{20.78728pt}{1.0pt}{20.33957pt}{0.55229pt}{20.33957pt}{0.0pt}\pgfsys@curveto{20.33957pt}{-0.55229pt}{20.78728pt}{-1.0pt}{21.33957pt}{-1.0pt}\pgfsys@curveto{21.89186pt}{-1.0pt}{22.33957pt}{-0.55229pt}{22.33957pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{0.0pt}{0.0pt}\pgfsys@lineto{7.11319pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@lineto{28.45276pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope{{
{}{}{}{}{}}{{{}}{{}}}}{}{}\hss}\pgfsys@discardpath\pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope\hss}}\lxSVG@closescope\endpgfpicture}}&=\frac{1}{2}\leavevmode\hbox
to28.85pt{\vbox to14.63pt{\pgfpicture\makeatletter\hbox{\hskip
0.2pt\lower-7.31319pt\hbox to0.0pt{\pgfsys@beginscope\pgfsys@invoke{
}\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{
}\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{
}\pgfsys@setlinewidth{0.4pt}\pgfsys@invoke{ }\nullfont\hbox
to0.0pt{\pgfsys@beginscope\pgfsys@invoke{ }
{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@moveto{8.11319pt}{0.0pt}\pgfsys@curveto{8.11319pt}{0.55229pt}{7.66548pt}{1.0pt}{7.11319pt}{1.0pt}\pgfsys@curveto{6.5609pt}{1.0pt}{6.11319pt}{0.55229pt}{6.11319pt}{0.0pt}\pgfsys@curveto{6.11319pt}{-0.55229pt}{6.5609pt}{-1.0pt}{7.11319pt}{-1.0pt}\pgfsys@curveto{7.66548pt}{-1.0pt}{8.11319pt}{-0.55229pt}{8.11319pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@moveto{22.33957pt}{0.0pt}\pgfsys@curveto{22.33957pt}{0.55229pt}{21.89186pt}{1.0pt}{21.33957pt}{1.0pt}\pgfsys@curveto{20.78728pt}{1.0pt}{20.33957pt}{0.55229pt}{20.33957pt}{0.0pt}\pgfsys@curveto{20.33957pt}{-0.55229pt}{20.78728pt}{-1.0pt}{21.33957pt}{-1.0pt}\pgfsys@curveto{21.89186pt}{-1.0pt}{22.33957pt}{-0.55229pt}{22.33957pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{0.0pt}{0.0pt}\pgfsys@lineto{7.11319pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@lineto{28.45276pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@curveto{21.33957pt}{3.92854pt}{18.15492pt}{7.11319pt}{14.22638pt}{7.11319pt}\pgfsys@curveto{10.29784pt}{7.11319pt}{7.11319pt}{3.92854pt}{7.11319pt}{0.0pt}\pgfsys@curveto{7.11319pt}{-3.92854pt}{10.29784pt}{-7.11319pt}{14.22638pt}{-7.11319pt}\pgfsys@curveto{18.15492pt}{-7.11319pt}{21.33957pt}{-3.92854pt}{21.33957pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope{{
{}{}{}{}{}}{{{}}{{}}}}{}{}\hss}\pgfsys@discardpath\pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope\hss}}\lxSVG@closescope\endpgfpicture}}+\leavevmode\hbox
to43.08pt{\vbox to28.85pt{\pgfpicture\makeatletter\hbox{\hskip
0.2pt\lower-14.42638pt\hbox to0.0pt{\pgfsys@beginscope\pgfsys@invoke{
}\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{
}\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{
}\pgfsys@setlinewidth{0.4pt}\pgfsys@invoke{ }\nullfont\hbox
to0.0pt{\pgfsys@beginscope\pgfsys@invoke{ }
{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{{}}{}{{}{}{}{{}}{{{{}{}{}{}}} {{}{}{}{}}}}{} {} {} {}
{}{}\pgfsys@beginscope\pgfsys@invoke{
}\color[rgb]{1,1,1}\definecolor[named]{pgfstrokecolor}{rgb}{1,1,1}\pgfsys@color@gray@stroke{1}\pgfsys@invoke{
}\pgfsys@color@gray@fill{1}\pgfsys@invoke{
}\definecolor[named]{pgffillcolor}{rgb}{1,1,1}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@curveto{7.11319pt}{-7.8571pt}{13.48247pt}{-14.22638pt}{21.33957pt}{-14.22638pt}\pgfsys@curveto{29.19667pt}{-14.22638pt}{35.56595pt}{-7.8571pt}{35.56595pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} \pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{\pgfsys@beginscope\pgfsys@invoke{
}\pgfsys@beginscope\pgfsys@invoke{
}\color[rgb]{1,1,1}\definecolor[named]{pgfstrokecolor}{rgb}{1,1,1}\pgfsys@color@gray@stroke{1}\pgfsys@invoke{
}\pgfsys@color@gray@fill{1}\pgfsys@invoke{
}\definecolor[named]{pgffillcolor}{rgb}{1,1,1}\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@moveto{28.45276pt}{0.0pt}\pgfsys@curveto{28.45276pt}{3.92854pt}{25.26811pt}{7.11319pt}{21.33957pt}{7.11319pt}\pgfsys@curveto{17.41103pt}{7.11319pt}{14.22638pt}{3.92854pt}{14.22638pt}{0.0pt}\pgfsys@curveto{14.22638pt}{-3.92854pt}{17.41103pt}{-7.11319pt}{21.33957pt}{-7.11319pt}\pgfsys@curveto{25.26811pt}{-7.11319pt}{28.45276pt}{-3.92854pt}{28.45276pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@fill\pgfsys@invoke{
} \pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope\pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope}\pgfsys@beginscope\pgfsys@invoke{
}\definecolor[named]{pgf@tempcolor}{rgb}{0,0,0}\pgfsys@invoke{\lxSVG@setpatternuncolored@{3}{0}{0}{0}
}{}\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@moveto{28.45276pt}{0.0pt}\pgfsys@curveto{28.45276pt}{3.92854pt}{25.26811pt}{7.11319pt}{21.33957pt}{7.11319pt}\pgfsys@curveto{17.41103pt}{7.11319pt}{14.22638pt}{3.92854pt}{14.22638pt}{0.0pt}\pgfsys@curveto{14.22638pt}{-3.92854pt}{17.41103pt}{-7.11319pt}{21.33957pt}{-7.11319pt}\pgfsys@curveto{25.26811pt}{-7.11319pt}{28.45276pt}{-3.92854pt}{28.45276pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
} \pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@moveto{8.11319pt}{0.0pt}\pgfsys@curveto{8.11319pt}{0.55229pt}{7.66548pt}{1.0pt}{7.11319pt}{1.0pt}\pgfsys@curveto{6.5609pt}{1.0pt}{6.11319pt}{0.55229pt}{6.11319pt}{0.0pt}\pgfsys@curveto{6.11319pt}{-0.55229pt}{6.5609pt}{-1.0pt}{7.11319pt}{-1.0pt}\pgfsys@curveto{7.66548pt}{-1.0pt}{8.11319pt}{-0.55229pt}{8.11319pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@moveto{15.22638pt}{0.0pt}\pgfsys@curveto{15.22638pt}{0.55229pt}{14.77867pt}{1.0pt}{14.22638pt}{1.0pt}\pgfsys@curveto{13.67409pt}{1.0pt}{13.22638pt}{0.55229pt}{13.22638pt}{0.0pt}\pgfsys@curveto{13.22638pt}{-0.55229pt}{13.67409pt}{-1.0pt}{14.22638pt}{-1.0pt}\pgfsys@curveto{14.77867pt}{-1.0pt}{15.22638pt}{-0.55229pt}{15.22638pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{28.45276pt}{0.0pt}\pgfsys@moveto{29.45276pt}{0.0pt}\pgfsys@curveto{29.45276pt}{0.55229pt}{29.00505pt}{1.0pt}{28.45276pt}{1.0pt}\pgfsys@curveto{27.90047pt}{1.0pt}{27.45276pt}{0.55229pt}{27.45276pt}{0.0pt}\pgfsys@curveto{27.45276pt}{-0.55229pt}{27.90047pt}{-1.0pt}{28.45276pt}{-1.0pt}\pgfsys@curveto{29.00505pt}{-1.0pt}{29.45276pt}{-0.55229pt}{29.45276pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{28.45276pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{35.56595pt}{0.0pt}\pgfsys@moveto{36.56595pt}{0.0pt}\pgfsys@curveto{36.56595pt}{0.55229pt}{36.11824pt}{1.0pt}{35.56595pt}{1.0pt}\pgfsys@curveto{35.01366pt}{1.0pt}{34.56595pt}{0.55229pt}{34.56595pt}{0.0pt}\pgfsys@curveto{34.56595pt}{-0.55229pt}{35.01366pt}{-1.0pt}{35.56595pt}{-1.0pt}\pgfsys@curveto{36.11824pt}{-1.0pt}{36.56595pt}{-0.55229pt}{36.56595pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{35.56595pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{0.0pt}{0.0pt}\pgfsys@lineto{7.11319pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@lineto{14.22638pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{28.45276pt}{0.0pt}\pgfsys@lineto{35.56595pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{35.56595pt}{0.0pt}\pgfsys@lineto{42.67914pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}{}{}{{}}{{{{}{}{}{}}}{{}{}{}{}}}}{} {} {} {}
{}{}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@curveto{7.11319pt}{7.8571pt}{13.48247pt}{14.22638pt}{21.33957pt}{14.22638pt}\pgfsys@curveto{29.19667pt}{14.22638pt}{35.56595pt}{7.8571pt}{35.56595pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope{{
{}{}{}{}{}}{{{}}{{}}}}{}{}\hss}\pgfsys@discardpath\pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope\hss}}\lxSVG@closescope\endpgfpicture}}+\leavevmode\hbox
to64.42pt{\vbox to50.19pt{\pgfpicture\makeatletter\hbox{\hskip
0.2pt\lower-25.09616pt\hbox to0.0pt{\pgfsys@beginscope\pgfsys@invoke{
}\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{
}\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{
}\pgfsys@setlinewidth{0.4pt}\pgfsys@invoke{ }\nullfont\hbox
to0.0pt{\pgfsys@beginscope\pgfsys@invoke{ }
{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{{}}{}{{}{}{}{{}}{{{{}{}{}{}}} {{}{}{}{}}}}{} {} {} {}
{}{}\pgfsys@beginscope\pgfsys@invoke{
}\color[rgb]{1,1,1}\definecolor[named]{pgfstrokecolor}{rgb}{1,1,1}\pgfsys@color@gray@stroke{1}\pgfsys@invoke{
}\pgfsys@color@gray@fill{1}\pgfsys@invoke{
}\definecolor[named]{pgffillcolor}{rgb}{1,1,1}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@curveto{7.11319pt}{-7.8571pt}{13.48247pt}{-14.22638pt}{21.33957pt}{-14.22638pt}\pgfsys@curveto{29.19667pt}{-14.22638pt}{35.56595pt}{-7.8571pt}{35.56595pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} \pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope{{}}{}{{}}{}{{}{}{}{{}}{{{{}{}{}{}}} {{}{}{}{}}}}{} {} {} {}
{}{}\pgfsys@beginscope\pgfsys@invoke{
}\color[rgb]{1,1,1}\definecolor[named]{pgfstrokecolor}{rgb}{1,1,1}\pgfsys@color@gray@stroke{1}\pgfsys@invoke{
}\pgfsys@color@gray@fill{1}\pgfsys@invoke{
}\definecolor[named]{pgffillcolor}{rgb}{1,1,1}{}\pgfsys@moveto{35.56595pt}{0.0pt}\pgfsys@curveto{35.56595pt}{-7.8571pt}{41.93523pt}{-14.22638pt}{49.79233pt}{-14.22638pt}\pgfsys@curveto{57.64943pt}{-14.22638pt}{64.0187pt}{-7.8571pt}{64.0187pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} \pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{\pgfsys@beginscope\pgfsys@invoke{
}\pgfsys@beginscope\pgfsys@invoke{
}\color[rgb]{1,1,1}\definecolor[named]{pgfstrokecolor}{rgb}{1,1,1}\pgfsys@color@gray@stroke{1}\pgfsys@invoke{
}\pgfsys@color@gray@fill{1}\pgfsys@invoke{
}\definecolor[named]{pgffillcolor}{rgb}{1,1,1}\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@moveto{28.45276pt}{0.0pt}\pgfsys@curveto{28.45276pt}{3.92854pt}{25.26811pt}{7.11319pt}{21.33957pt}{7.11319pt}\pgfsys@curveto{17.41103pt}{7.11319pt}{14.22638pt}{3.92854pt}{14.22638pt}{0.0pt}\pgfsys@curveto{14.22638pt}{-3.92854pt}{17.41103pt}{-7.11319pt}{21.33957pt}{-7.11319pt}\pgfsys@curveto{25.26811pt}{-7.11319pt}{28.45276pt}{-3.92854pt}{28.45276pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@fill\pgfsys@invoke{
} \pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope\pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope}\pgfsys@beginscope\pgfsys@invoke{
}\definecolor[named]{pgf@tempcolor}{rgb}{0,0,0}\pgfsys@invoke{\lxSVG@setpatternuncolored@{3}{0}{0}{0}
}{}\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@moveto{28.45276pt}{0.0pt}\pgfsys@curveto{28.45276pt}{3.92854pt}{25.26811pt}{7.11319pt}{21.33957pt}{7.11319pt}\pgfsys@curveto{17.41103pt}{7.11319pt}{14.22638pt}{3.92854pt}{14.22638pt}{0.0pt}\pgfsys@curveto{14.22638pt}{-3.92854pt}{17.41103pt}{-7.11319pt}{21.33957pt}{-7.11319pt}\pgfsys@curveto{25.26811pt}{-7.11319pt}{28.45276pt}{-3.92854pt}{28.45276pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
} \pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{\pgfsys@beginscope\pgfsys@invoke{
}\pgfsys@beginscope\pgfsys@invoke{
}\color[rgb]{1,1,1}\definecolor[named]{pgfstrokecolor}{rgb}{1,1,1}\pgfsys@color@gray@stroke{1}\pgfsys@invoke{
}\pgfsys@color@gray@fill{1}\pgfsys@invoke{
}\definecolor[named]{pgffillcolor}{rgb}{1,1,1}\pgfsys@moveto{42.67914pt}{0.0pt}\pgfsys@moveto{49.79233pt}{0.0pt}\pgfsys@curveto{49.79233pt}{3.92854pt}{46.60768pt}{7.11319pt}{42.67914pt}{7.11319pt}\pgfsys@curveto{38.7506pt}{7.11319pt}{35.56595pt}{3.92854pt}{35.56595pt}{0.0pt}\pgfsys@curveto{35.56595pt}{-3.92854pt}{38.7506pt}{-7.11319pt}{42.67914pt}{-7.11319pt}\pgfsys@curveto{46.60768pt}{-7.11319pt}{49.79233pt}{-3.92854pt}{49.79233pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{42.67914pt}{0.0pt}\pgfsys@fill\pgfsys@invoke{
} \pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope\pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope}\pgfsys@beginscope\pgfsys@invoke{
}\definecolor[named]{pgf@tempcolor}{rgb}{0,0,0}\pgfsys@invoke{\lxSVG@setpatternuncolored@{3}{0}{0}{0}
}{}\pgfsys@moveto{42.67914pt}{0.0pt}\pgfsys@moveto{49.79233pt}{0.0pt}\pgfsys@curveto{49.79233pt}{3.92854pt}{46.60768pt}{7.11319pt}{42.67914pt}{7.11319pt}\pgfsys@curveto{38.7506pt}{7.11319pt}{35.56595pt}{3.92854pt}{35.56595pt}{0.0pt}\pgfsys@curveto{35.56595pt}{-3.92854pt}{38.7506pt}{-7.11319pt}{42.67914pt}{-7.11319pt}\pgfsys@curveto{46.60768pt}{-7.11319pt}{49.79233pt}{-3.92854pt}{49.79233pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{42.67914pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
} \pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope{{}}{}{{}}{}{{}{}{}{{}}{{{{}{}{}{}}} {{}{}{}{}}}}{} {} {} {}
{}{}\pgfsys@beginscope\pgfsys@invoke{
}\color[rgb]{1,1,1}\definecolor[named]{pgfstrokecolor}{rgb}{1,1,1}\pgfsys@color@gray@stroke{1}\pgfsys@invoke{
}\pgfsys@color@gray@fill{1}\pgfsys@invoke{
}\definecolor[named]{pgffillcolor}{rgb}{1,1,1}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@curveto{7.11319pt}{-13.74994pt}{18.25941pt}{-24.89616pt}{32.00935pt}{-24.89616pt}\pgfsys@curveto{45.7593pt}{-24.89616pt}{56.90552pt}{-13.74994pt}{56.90552pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} \pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@moveto{8.11319pt}{0.0pt}\pgfsys@curveto{8.11319pt}{0.55229pt}{7.66548pt}{1.0pt}{7.11319pt}{1.0pt}\pgfsys@curveto{6.5609pt}{1.0pt}{6.11319pt}{0.55229pt}{6.11319pt}{0.0pt}\pgfsys@curveto{6.11319pt}{-0.55229pt}{6.5609pt}{-1.0pt}{7.11319pt}{-1.0pt}\pgfsys@curveto{7.66548pt}{-1.0pt}{8.11319pt}{-0.55229pt}{8.11319pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@moveto{15.22638pt}{0.0pt}\pgfsys@curveto{15.22638pt}{0.55229pt}{14.77867pt}{1.0pt}{14.22638pt}{1.0pt}\pgfsys@curveto{13.67409pt}{1.0pt}{13.22638pt}{0.55229pt}{13.22638pt}{0.0pt}\pgfsys@curveto{13.22638pt}{-0.55229pt}{13.67409pt}{-1.0pt}{14.22638pt}{-1.0pt}\pgfsys@curveto{14.77867pt}{-1.0pt}{15.22638pt}{-0.55229pt}{15.22638pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{28.45276pt}{0.0pt}\pgfsys@moveto{29.45276pt}{0.0pt}\pgfsys@curveto{29.45276pt}{0.55229pt}{29.00505pt}{1.0pt}{28.45276pt}{1.0pt}\pgfsys@curveto{27.90047pt}{1.0pt}{27.45276pt}{0.55229pt}{27.45276pt}{0.0pt}\pgfsys@curveto{27.45276pt}{-0.55229pt}{27.90047pt}{-1.0pt}{28.45276pt}{-1.0pt}\pgfsys@curveto{29.00505pt}{-1.0pt}{29.45276pt}{-0.55229pt}{29.45276pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{28.45276pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{35.56595pt}{0.0pt}\pgfsys@moveto{36.56595pt}{0.0pt}\pgfsys@curveto{36.56595pt}{0.55229pt}{36.11824pt}{1.0pt}{35.56595pt}{1.0pt}\pgfsys@curveto{35.01366pt}{1.0pt}{34.56595pt}{0.55229pt}{34.56595pt}{0.0pt}\pgfsys@curveto{34.56595pt}{-0.55229pt}{35.01366pt}{-1.0pt}{35.56595pt}{-1.0pt}\pgfsys@curveto{36.11824pt}{-1.0pt}{36.56595pt}{-0.55229pt}{36.56595pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{35.56595pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{49.79233pt}{0.0pt}\pgfsys@moveto{50.79233pt}{0.0pt}\pgfsys@curveto{50.79233pt}{0.55229pt}{50.34462pt}{1.0pt}{49.79233pt}{1.0pt}\pgfsys@curveto{49.24004pt}{1.0pt}{48.79233pt}{0.55229pt}{48.79233pt}{0.0pt}\pgfsys@curveto{48.79233pt}{-0.55229pt}{49.24004pt}{-1.0pt}{49.79233pt}{-1.0pt}\pgfsys@curveto{50.34462pt}{-1.0pt}{50.79233pt}{-0.55229pt}{50.79233pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{49.79233pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{56.90552pt}{0.0pt}\pgfsys@moveto{57.90552pt}{0.0pt}\pgfsys@curveto{57.90552pt}{0.55229pt}{57.45781pt}{1.0pt}{56.90552pt}{1.0pt}\pgfsys@curveto{56.35323pt}{1.0pt}{55.90552pt}{0.55229pt}{55.90552pt}{0.0pt}\pgfsys@curveto{55.90552pt}{-0.55229pt}{56.35323pt}{-1.0pt}{56.90552pt}{-1.0pt}\pgfsys@curveto{57.45781pt}{-1.0pt}{57.90552pt}{-0.55229pt}{57.90552pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{56.90552pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{0.0pt}{0.0pt}\pgfsys@lineto{7.11319pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@lineto{14.22638pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{28.45276pt}{0.0pt}\pgfsys@lineto{35.56595pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{49.79233pt}{0.0pt}\pgfsys@lineto{56.90552pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{56.90552pt}{0.0pt}\pgfsys@lineto{64.0187pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}{}{}{{}}{{{{}{}{}{}}}{{}{}{}{}}}}{} {} {} {}
{}{}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@curveto{7.11319pt}{13.74994pt}{18.25941pt}{24.89616pt}{32.00935pt}{24.89616pt}\pgfsys@curveto{45.7593pt}{24.89616pt}{56.90552pt}{13.74994pt}{56.90552pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope{{
{}{}{}{}{}}{{{}}{{}}}}{}{}\hss}\pgfsys@discardpath\pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope\hss}}\lxSVG@closescope\endpgfpicture}}+\cdots-\text{subtractions}\end{aligned}$
(2.1)
with the appropriate BPHZ subtractions indicated. Another way to describe the
relevant set of graphs is to start with the one-loop graph and add all
possible iterated and multiple insertions of this graph into one of the
propagators. Figure 1 shows the resulting low order diagrams and compares this
Hopf expansion with two other common approximations to the Dyson–Schwinger
equations: the rainbow approximation [42] and the chain approximation [22,
23]. The Hopf expansion includes a much larger class of diagrams than either
the rainbow or the chain approximation, and leads to a much richer non-
perturbative structure. The differences between these approximations is
discussed further below, in Sections 3 and 4.222The effect of including also
the vertex corrections will be addressed in future work.
$\leavevmode\hbox to28.85pt{\vbox
to14.63pt{\pgfpicture\makeatletter\hbox{\hskip 0.2pt\lower-7.31319pt\hbox
to0.0pt{\pgfsys@beginscope\pgfsys@invoke{
}\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{
}\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{
}\pgfsys@setlinewidth{0.4pt}\pgfsys@invoke{ }\nullfont\hbox
to0.0pt{\pgfsys@beginscope\pgfsys@invoke{ }
{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{\pgfsys@beginscope\pgfsys@invoke{
}\pgfsys@beginscope\pgfsys@invoke{
}\color[rgb]{1,1,1}\definecolor[named]{pgfstrokecolor}{rgb}{1,1,1}\pgfsys@color@gray@stroke{1}\pgfsys@invoke{
}\pgfsys@color@gray@fill{1}\pgfsys@invoke{
}\definecolor[named]{pgffillcolor}{rgb}{1,1,1}\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@curveto{21.33957pt}{3.92854pt}{18.15492pt}{7.11319pt}{14.22638pt}{7.11319pt}\pgfsys@curveto{10.29784pt}{7.11319pt}{7.11319pt}{3.92854pt}{7.11319pt}{0.0pt}\pgfsys@curveto{7.11319pt}{-3.92854pt}{10.29784pt}{-7.11319pt}{14.22638pt}{-7.11319pt}\pgfsys@curveto{18.15492pt}{-7.11319pt}{21.33957pt}{-3.92854pt}{21.33957pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@fill\pgfsys@invoke{
} \pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope\pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope}\pgfsys@beginscope\pgfsys@invoke{
}\definecolor[named]{pgf@tempcolor}{rgb}{0,0,0}\pgfsys@invoke{\lxSVG@setpatternuncolored@{3}{0}{0}{0}
}{}\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@curveto{21.33957pt}{3.92854pt}{18.15492pt}{7.11319pt}{14.22638pt}{7.11319pt}\pgfsys@curveto{10.29784pt}{7.11319pt}{7.11319pt}{3.92854pt}{7.11319pt}{0.0pt}\pgfsys@curveto{7.11319pt}{-3.92854pt}{10.29784pt}{-7.11319pt}{14.22638pt}{-7.11319pt}\pgfsys@curveto{18.15492pt}{-7.11319pt}{21.33957pt}{-3.92854pt}{21.33957pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
} \pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@moveto{8.11319pt}{0.0pt}\pgfsys@curveto{8.11319pt}{0.55229pt}{7.66548pt}{1.0pt}{7.11319pt}{1.0pt}\pgfsys@curveto{6.5609pt}{1.0pt}{6.11319pt}{0.55229pt}{6.11319pt}{0.0pt}\pgfsys@curveto{6.11319pt}{-0.55229pt}{6.5609pt}{-1.0pt}{7.11319pt}{-1.0pt}\pgfsys@curveto{7.66548pt}{-1.0pt}{8.11319pt}{-0.55229pt}{8.11319pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@moveto{22.33957pt}{0.0pt}\pgfsys@curveto{22.33957pt}{0.55229pt}{21.89186pt}{1.0pt}{21.33957pt}{1.0pt}\pgfsys@curveto{20.78728pt}{1.0pt}{20.33957pt}{0.55229pt}{20.33957pt}{0.0pt}\pgfsys@curveto{20.33957pt}{-0.55229pt}{20.78728pt}{-1.0pt}{21.33957pt}{-1.0pt}\pgfsys@curveto{21.89186pt}{-1.0pt}{22.33957pt}{-0.55229pt}{22.33957pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{0.0pt}{0.0pt}\pgfsys@lineto{7.11319pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@lineto{28.45276pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope{{
{}{}{}{}{}}{{{}}{{}}}}{}{}\hss}\pgfsys@discardpath\pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope\hss}}\lxSVG@closescope\endpgfpicture}}=\frac{1}{2}\leavevmode\hbox
to28.85pt{\vbox to14.63pt{\pgfpicture\makeatletter\hbox{\hskip
0.2pt\lower-7.31319pt\hbox to0.0pt{\pgfsys@beginscope\pgfsys@invoke{
}\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{
}\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{
}\pgfsys@setlinewidth{0.4pt}\pgfsys@invoke{ }\nullfont\hbox
to0.0pt{\pgfsys@beginscope\pgfsys@invoke{ }
{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@moveto{8.11319pt}{0.0pt}\pgfsys@curveto{8.11319pt}{0.55229pt}{7.66548pt}{1.0pt}{7.11319pt}{1.0pt}\pgfsys@curveto{6.5609pt}{1.0pt}{6.11319pt}{0.55229pt}{6.11319pt}{0.0pt}\pgfsys@curveto{6.11319pt}{-0.55229pt}{6.5609pt}{-1.0pt}{7.11319pt}{-1.0pt}\pgfsys@curveto{7.66548pt}{-1.0pt}{8.11319pt}{-0.55229pt}{8.11319pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@moveto{22.33957pt}{0.0pt}\pgfsys@curveto{22.33957pt}{0.55229pt}{21.89186pt}{1.0pt}{21.33957pt}{1.0pt}\pgfsys@curveto{20.78728pt}{1.0pt}{20.33957pt}{0.55229pt}{20.33957pt}{0.0pt}\pgfsys@curveto{20.33957pt}{-0.55229pt}{20.78728pt}{-1.0pt}{21.33957pt}{-1.0pt}\pgfsys@curveto{21.89186pt}{-1.0pt}{22.33957pt}{-0.55229pt}{22.33957pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{0.0pt}{0.0pt}\pgfsys@lineto{7.11319pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@lineto{28.45276pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@curveto{21.33957pt}{3.92854pt}{18.15492pt}{7.11319pt}{14.22638pt}{7.11319pt}\pgfsys@curveto{10.29784pt}{7.11319pt}{7.11319pt}{3.92854pt}{7.11319pt}{0.0pt}\pgfsys@curveto{7.11319pt}{-3.92854pt}{10.29784pt}{-7.11319pt}{14.22638pt}{-7.11319pt}\pgfsys@curveto{18.15492pt}{-7.11319pt}{21.33957pt}{-3.92854pt}{21.33957pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope{{
{}{}{}{}{}}{{{}}{{}}}}{}{}\hss}\pgfsys@discardpath\pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope\hss}}\lxSVG@closescope\endpgfpicture}}+\frac{1}{2}\leavevmode\hbox
to43.08pt{\vbox to28.85pt{\pgfpicture\makeatletter\hbox{\hskip
0.2pt\lower-14.42638pt\hbox to0.0pt{\pgfsys@beginscope\pgfsys@invoke{
}\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{
}\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{
}\pgfsys@setlinewidth{0.4pt}\pgfsys@invoke{ }\nullfont\hbox
to0.0pt{\pgfsys@beginscope\pgfsys@invoke{ }
{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{{}}{}{{}{}{}{{}}{{{{}{}{}{}}} {{}{}{}{}}}}{} {} {} {}
{}{}\pgfsys@beginscope\pgfsys@invoke{
}\color[rgb]{1,1,1}\definecolor[named]{pgfstrokecolor}{rgb}{1,1,1}\pgfsys@color@gray@stroke{1}\pgfsys@invoke{
}\pgfsys@color@gray@fill{1}\pgfsys@invoke{
}\definecolor[named]{pgffillcolor}{rgb}{1,1,1}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@curveto{7.11319pt}{-7.8571pt}{13.48247pt}{-14.22638pt}{21.33957pt}{-14.22638pt}\pgfsys@curveto{29.19667pt}{-14.22638pt}{35.56595pt}{-7.8571pt}{35.56595pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} \pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@moveto{8.11319pt}{0.0pt}\pgfsys@curveto{8.11319pt}{0.55229pt}{7.66548pt}{1.0pt}{7.11319pt}{1.0pt}\pgfsys@curveto{6.5609pt}{1.0pt}{6.11319pt}{0.55229pt}{6.11319pt}{0.0pt}\pgfsys@curveto{6.11319pt}{-0.55229pt}{6.5609pt}{-1.0pt}{7.11319pt}{-1.0pt}\pgfsys@curveto{7.66548pt}{-1.0pt}{8.11319pt}{-0.55229pt}{8.11319pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@moveto{15.22638pt}{0.0pt}\pgfsys@curveto{15.22638pt}{0.55229pt}{14.77867pt}{1.0pt}{14.22638pt}{1.0pt}\pgfsys@curveto{13.67409pt}{1.0pt}{13.22638pt}{0.55229pt}{13.22638pt}{0.0pt}\pgfsys@curveto{13.22638pt}{-0.55229pt}{13.67409pt}{-1.0pt}{14.22638pt}{-1.0pt}\pgfsys@curveto{14.77867pt}{-1.0pt}{15.22638pt}{-0.55229pt}{15.22638pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{28.45276pt}{0.0pt}\pgfsys@moveto{29.45276pt}{0.0pt}\pgfsys@curveto{29.45276pt}{0.55229pt}{29.00505pt}{1.0pt}{28.45276pt}{1.0pt}\pgfsys@curveto{27.90047pt}{1.0pt}{27.45276pt}{0.55229pt}{27.45276pt}{0.0pt}\pgfsys@curveto{27.45276pt}{-0.55229pt}{27.90047pt}{-1.0pt}{28.45276pt}{-1.0pt}\pgfsys@curveto{29.00505pt}{-1.0pt}{29.45276pt}{-0.55229pt}{29.45276pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{28.45276pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{35.56595pt}{0.0pt}\pgfsys@moveto{36.56595pt}{0.0pt}\pgfsys@curveto{36.56595pt}{0.55229pt}{36.11824pt}{1.0pt}{35.56595pt}{1.0pt}\pgfsys@curveto{35.01366pt}{1.0pt}{34.56595pt}{0.55229pt}{34.56595pt}{0.0pt}\pgfsys@curveto{34.56595pt}{-0.55229pt}{35.01366pt}{-1.0pt}{35.56595pt}{-1.0pt}\pgfsys@curveto{36.11824pt}{-1.0pt}{36.56595pt}{-0.55229pt}{36.56595pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{35.56595pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{0.0pt}{0.0pt}\pgfsys@lineto{7.11319pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@lineto{14.22638pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@lineto{28.45276pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{28.45276pt}{0.0pt}\pgfsys@lineto{35.56595pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{35.56595pt}{0.0pt}\pgfsys@lineto{42.67914pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}{}{}{{}}{{{{}{}{}{}}}{{}{}{}{}}}}{} {} {} {}
{}{}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@curveto{7.11319pt}{7.8571pt}{13.48247pt}{14.22638pt}{21.33957pt}{14.22638pt}\pgfsys@curveto{29.19667pt}{14.22638pt}{35.56595pt}{7.8571pt}{35.56595pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}{}{}{{}}{{{{}{}{}{}}}{{}{}{}{}}}}{} {} {} {}
{}{}{}\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@curveto{14.22638pt}{3.92854pt}{17.41103pt}{7.11319pt}{21.33957pt}{7.11319pt}\pgfsys@curveto{25.26811pt}{7.11319pt}{28.45276pt}{3.92854pt}{28.45276pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope{{
{}{}{}{}{}}{{{}}{{}}}}{}{}\hss}\pgfsys@discardpath\pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope\hss}}\lxSVG@closescope\endpgfpicture}}+\frac{1}{2}\leavevmode\hbox
to57.31pt{\vbox to43.08pt{\pgfpicture\makeatletter\hbox{\hskip
0.2pt\lower-21.53957pt\hbox to0.0pt{\pgfsys@beginscope\pgfsys@invoke{
}\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{
}\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{
}\pgfsys@setlinewidth{0.4pt}\pgfsys@invoke{ }\nullfont\hbox
to0.0pt{\pgfsys@beginscope\pgfsys@invoke{ }
{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{{}}{}{{}{}{}{{}}{{{{}{}{}{}}} {{}{}{}{}}}}{} {} {} {}
{}{}\pgfsys@beginscope\pgfsys@invoke{
}\color[rgb]{1,1,1}\definecolor[named]{pgfstrokecolor}{rgb}{1,1,1}\pgfsys@color@gray@stroke{1}\pgfsys@invoke{
}\pgfsys@color@gray@fill{1}\pgfsys@invoke{
}\definecolor[named]{pgffillcolor}{rgb}{1,1,1}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@curveto{7.11319pt}{-11.78566pt}{16.6671pt}{-21.33957pt}{28.45276pt}{-21.33957pt}\pgfsys@curveto{40.23842pt}{-21.33957pt}{49.79233pt}{-11.78566pt}{49.79233pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} \pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@moveto{8.11319pt}{0.0pt}\pgfsys@curveto{8.11319pt}{0.55229pt}{7.66548pt}{1.0pt}{7.11319pt}{1.0pt}\pgfsys@curveto{6.5609pt}{1.0pt}{6.11319pt}{0.55229pt}{6.11319pt}{0.0pt}\pgfsys@curveto{6.11319pt}{-0.55229pt}{6.5609pt}{-1.0pt}{7.11319pt}{-1.0pt}\pgfsys@curveto{7.66548pt}{-1.0pt}{8.11319pt}{-0.55229pt}{8.11319pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@moveto{15.22638pt}{0.0pt}\pgfsys@curveto{15.22638pt}{0.55229pt}{14.77867pt}{1.0pt}{14.22638pt}{1.0pt}\pgfsys@curveto{13.67409pt}{1.0pt}{13.22638pt}{0.55229pt}{13.22638pt}{0.0pt}\pgfsys@curveto{13.22638pt}{-0.55229pt}{13.67409pt}{-1.0pt}{14.22638pt}{-1.0pt}\pgfsys@curveto{14.77867pt}{-1.0pt}{15.22638pt}{-0.55229pt}{15.22638pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@moveto{22.33957pt}{0.0pt}\pgfsys@curveto{22.33957pt}{0.55229pt}{21.89186pt}{1.0pt}{21.33957pt}{1.0pt}\pgfsys@curveto{20.78728pt}{1.0pt}{20.33957pt}{0.55229pt}{20.33957pt}{0.0pt}\pgfsys@curveto{20.33957pt}{-0.55229pt}{20.78728pt}{-1.0pt}{21.33957pt}{-1.0pt}\pgfsys@curveto{21.89186pt}{-1.0pt}{22.33957pt}{-0.55229pt}{22.33957pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{35.56595pt}{0.0pt}\pgfsys@moveto{36.56595pt}{0.0pt}\pgfsys@curveto{36.56595pt}{0.55229pt}{36.11824pt}{1.0pt}{35.56595pt}{1.0pt}\pgfsys@curveto{35.01366pt}{1.0pt}{34.56595pt}{0.55229pt}{34.56595pt}{0.0pt}\pgfsys@curveto{34.56595pt}{-0.55229pt}{35.01366pt}{-1.0pt}{35.56595pt}{-1.0pt}\pgfsys@curveto{36.11824pt}{-1.0pt}{36.56595pt}{-0.55229pt}{36.56595pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{35.56595pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{42.67914pt}{0.0pt}\pgfsys@moveto{43.67914pt}{0.0pt}\pgfsys@curveto{43.67914pt}{0.55229pt}{43.23143pt}{1.0pt}{42.67914pt}{1.0pt}\pgfsys@curveto{42.12685pt}{1.0pt}{41.67914pt}{0.55229pt}{41.67914pt}{0.0pt}\pgfsys@curveto{41.67914pt}{-0.55229pt}{42.12685pt}{-1.0pt}{42.67914pt}{-1.0pt}\pgfsys@curveto{43.23143pt}{-1.0pt}{43.67914pt}{-0.55229pt}{43.67914pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{42.67914pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{49.79233pt}{0.0pt}\pgfsys@moveto{50.79233pt}{0.0pt}\pgfsys@curveto{50.79233pt}{0.55229pt}{50.34462pt}{1.0pt}{49.79233pt}{1.0pt}\pgfsys@curveto{49.24004pt}{1.0pt}{48.79233pt}{0.55229pt}{48.79233pt}{0.0pt}\pgfsys@curveto{48.79233pt}{-0.55229pt}{49.24004pt}{-1.0pt}{49.79233pt}{-1.0pt}\pgfsys@curveto{50.34462pt}{-1.0pt}{50.79233pt}{-0.55229pt}{50.79233pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{49.79233pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{0.0pt}{0.0pt}\pgfsys@lineto{7.11319pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@lineto{14.22638pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@lineto{21.33957pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@lineto{35.56595pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{35.56595pt}{0.0pt}\pgfsys@lineto{42.67914pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{42.67914pt}{0.0pt}\pgfsys@lineto{49.79233pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{49.79233pt}{0.0pt}\pgfsys@lineto{56.90552pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}{}{}{{}}{{{{}{}{}{}}}{{}{}{}{}}}}{} {} {} {}
{}{}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@curveto{7.11319pt}{11.78566pt}{16.6671pt}{21.33957pt}{28.45276pt}{21.33957pt}\pgfsys@curveto{40.23842pt}{21.33957pt}{49.79233pt}{11.78566pt}{49.79233pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}{}{}{{}}{{{{}{}{}{}}}{{}{}{}{}}}}{} {} {} {}
{}{}{}\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@curveto{14.22638pt}{7.8571pt}{20.59566pt}{14.22638pt}{28.45276pt}{14.22638pt}\pgfsys@curveto{36.30986pt}{14.22638pt}{42.67914pt}{7.8571pt}{42.67914pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}{}{}{{}}{{{{}{}{}{}}}{{}{}{}{}}}}{} {} {} {}
{}{}{}\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@curveto{21.33957pt}{3.92854pt}{24.52422pt}{7.11319pt}{28.45276pt}{7.11319pt}\pgfsys@curveto{32.3813pt}{7.11319pt}{35.56595pt}{3.92854pt}{35.56595pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope{{
{}{}{}{}{}}{{{}}{{}}}}{}{}\hss}\pgfsys@discardpath\pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope\hss}}\lxSVG@closescope\endpgfpicture}}+\frac{1}{2}\leavevmode\hbox
to71.53pt{\vbox to57.31pt{\pgfpicture\makeatletter\hbox{\hskip
0.2pt\lower-28.65276pt\hbox to0.0pt{\pgfsys@beginscope\pgfsys@invoke{
}\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{
}\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{
}\pgfsys@setlinewidth{0.4pt}\pgfsys@invoke{ }\nullfont\hbox
to0.0pt{\pgfsys@beginscope\pgfsys@invoke{ }
{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{{}}{}{{}{}{}{{}}{{{{}{}{}{}}} {{}{}{}{}}}}{} {} {} {}
{}{}\pgfsys@beginscope\pgfsys@invoke{
}\color[rgb]{1,1,1}\definecolor[named]{pgfstrokecolor}{rgb}{1,1,1}\pgfsys@color@gray@stroke{1}\pgfsys@invoke{
}\pgfsys@color@gray@fill{1}\pgfsys@invoke{
}\definecolor[named]{pgffillcolor}{rgb}{1,1,1}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@curveto{7.11319pt}{-15.71422pt}{19.85173pt}{-28.45276pt}{35.56595pt}{-28.45276pt}\pgfsys@curveto{51.28017pt}{-28.45276pt}{64.0187pt}{-15.71422pt}{64.0187pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} \pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@moveto{8.11319pt}{0.0pt}\pgfsys@curveto{8.11319pt}{0.55229pt}{7.66548pt}{1.0pt}{7.11319pt}{1.0pt}\pgfsys@curveto{6.5609pt}{1.0pt}{6.11319pt}{0.55229pt}{6.11319pt}{0.0pt}\pgfsys@curveto{6.11319pt}{-0.55229pt}{6.5609pt}{-1.0pt}{7.11319pt}{-1.0pt}\pgfsys@curveto{7.66548pt}{-1.0pt}{8.11319pt}{-0.55229pt}{8.11319pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@moveto{15.22638pt}{0.0pt}\pgfsys@curveto{15.22638pt}{0.55229pt}{14.77867pt}{1.0pt}{14.22638pt}{1.0pt}\pgfsys@curveto{13.67409pt}{1.0pt}{13.22638pt}{0.55229pt}{13.22638pt}{0.0pt}\pgfsys@curveto{13.22638pt}{-0.55229pt}{13.67409pt}{-1.0pt}{14.22638pt}{-1.0pt}\pgfsys@curveto{14.77867pt}{-1.0pt}{15.22638pt}{-0.55229pt}{15.22638pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@moveto{22.33957pt}{0.0pt}\pgfsys@curveto{22.33957pt}{0.55229pt}{21.89186pt}{1.0pt}{21.33957pt}{1.0pt}\pgfsys@curveto{20.78728pt}{1.0pt}{20.33957pt}{0.55229pt}{20.33957pt}{0.0pt}\pgfsys@curveto{20.33957pt}{-0.55229pt}{20.78728pt}{-1.0pt}{21.33957pt}{-1.0pt}\pgfsys@curveto{21.89186pt}{-1.0pt}{22.33957pt}{-0.55229pt}{22.33957pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{28.45276pt}{0.0pt}\pgfsys@moveto{29.45276pt}{0.0pt}\pgfsys@curveto{29.45276pt}{0.55229pt}{29.00505pt}{1.0pt}{28.45276pt}{1.0pt}\pgfsys@curveto{27.90047pt}{1.0pt}{27.45276pt}{0.55229pt}{27.45276pt}{0.0pt}\pgfsys@curveto{27.45276pt}{-0.55229pt}{27.90047pt}{-1.0pt}{28.45276pt}{-1.0pt}\pgfsys@curveto{29.00505pt}{-1.0pt}{29.45276pt}{-0.55229pt}{29.45276pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{28.45276pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{42.67914pt}{0.0pt}\pgfsys@moveto{43.67914pt}{0.0pt}\pgfsys@curveto{43.67914pt}{0.55229pt}{43.23143pt}{1.0pt}{42.67914pt}{1.0pt}\pgfsys@curveto{42.12685pt}{1.0pt}{41.67914pt}{0.55229pt}{41.67914pt}{0.0pt}\pgfsys@curveto{41.67914pt}{-0.55229pt}{42.12685pt}{-1.0pt}{42.67914pt}{-1.0pt}\pgfsys@curveto{43.23143pt}{-1.0pt}{43.67914pt}{-0.55229pt}{43.67914pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{42.67914pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{49.79233pt}{0.0pt}\pgfsys@moveto{50.79233pt}{0.0pt}\pgfsys@curveto{50.79233pt}{0.55229pt}{50.34462pt}{1.0pt}{49.79233pt}{1.0pt}\pgfsys@curveto{49.24004pt}{1.0pt}{48.79233pt}{0.55229pt}{48.79233pt}{0.0pt}\pgfsys@curveto{48.79233pt}{-0.55229pt}{49.24004pt}{-1.0pt}{49.79233pt}{-1.0pt}\pgfsys@curveto{50.34462pt}{-1.0pt}{50.79233pt}{-0.55229pt}{50.79233pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{49.79233pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{56.90552pt}{0.0pt}\pgfsys@moveto{57.90552pt}{0.0pt}\pgfsys@curveto{57.90552pt}{0.55229pt}{57.45781pt}{1.0pt}{56.90552pt}{1.0pt}\pgfsys@curveto{56.35323pt}{1.0pt}{55.90552pt}{0.55229pt}{55.90552pt}{0.0pt}\pgfsys@curveto{55.90552pt}{-0.55229pt}{56.35323pt}{-1.0pt}{56.90552pt}{-1.0pt}\pgfsys@curveto{57.45781pt}{-1.0pt}{57.90552pt}{-0.55229pt}{57.90552pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{56.90552pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{64.0187pt}{0.0pt}\pgfsys@moveto{65.0187pt}{0.0pt}\pgfsys@curveto{65.0187pt}{0.55229pt}{64.571pt}{1.0pt}{64.0187pt}{1.0pt}\pgfsys@curveto{63.46642pt}{1.0pt}{63.0187pt}{0.55229pt}{63.0187pt}{0.0pt}\pgfsys@curveto{63.0187pt}{-0.55229pt}{63.46642pt}{-1.0pt}{64.0187pt}{-1.0pt}\pgfsys@curveto{64.571pt}{-1.0pt}{65.0187pt}{-0.55229pt}{65.0187pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{64.0187pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{0.0pt}{0.0pt}\pgfsys@lineto{7.11319pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@lineto{14.22638pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@lineto{21.33957pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@lineto{28.45276pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{28.45276pt}{0.0pt}\pgfsys@lineto{42.67914pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{42.67914pt}{0.0pt}\pgfsys@lineto{49.79233pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{49.79233pt}{0.0pt}\pgfsys@lineto{56.90552pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{56.90552pt}{0.0pt}\pgfsys@lineto{64.0187pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{64.0187pt}{0.0pt}\pgfsys@lineto{71.1319pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}{}{}{{}}{{{{}{}{}{}}}{{}{}{}{}}}}{} {} {} {}
{}{}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@curveto{7.11319pt}{15.71422pt}{19.85173pt}{28.45276pt}{35.56595pt}{28.45276pt}\pgfsys@curveto{51.28017pt}{28.45276pt}{64.0187pt}{15.71422pt}{64.0187pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}{}{}{{}}{{{{}{}{}{}}}{{}{}{}{}}}}{} {} {} {}
{}{}{}\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@curveto{14.22638pt}{11.78566pt}{23.78029pt}{21.33957pt}{35.56595pt}{21.33957pt}\pgfsys@curveto{47.35161pt}{21.33957pt}{56.90552pt}{11.78566pt}{56.90552pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}{}{}{{}}{{{{}{}{}{}}}{{}{}{}{}}}}{} {} {} {}
{}{}{}\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@curveto{21.33957pt}{7.8571pt}{27.70885pt}{14.22638pt}{35.56595pt}{14.22638pt}\pgfsys@curveto{43.42305pt}{14.22638pt}{49.79233pt}{7.8571pt}{49.79233pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}{}{}{{}}{{{{}{}{}{}}}{{}{}{}{}}}}{} {} {} {}
{}{}{}\pgfsys@moveto{28.45276pt}{0.0pt}\pgfsys@curveto{28.45276pt}{3.92854pt}{31.6374pt}{7.11319pt}{35.56595pt}{7.11319pt}\pgfsys@curveto{39.49449pt}{7.11319pt}{42.67914pt}{3.92854pt}{42.67914pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope{{
{}{}{}{}{}}{{{}}{{}}}}{}{}\hss}\pgfsys@discardpath\pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope\hss}}\lxSVG@closescope\endpgfpicture}}+\cdots\hfill\vspace{-3ex}$
($a$) Rainbow approximation
$\leavevmode\hbox to28.85pt{\vbox
to14.63pt{\pgfpicture\makeatletter\hbox{\hskip 0.2pt\lower-7.31319pt\hbox
to0.0pt{\pgfsys@beginscope\pgfsys@invoke{
}\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{
}\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{
}\pgfsys@setlinewidth{0.4pt}\pgfsys@invoke{ }\nullfont\hbox
to0.0pt{\pgfsys@beginscope\pgfsys@invoke{ }
{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{\pgfsys@beginscope\pgfsys@invoke{
}\pgfsys@beginscope\pgfsys@invoke{
}\color[rgb]{1,1,1}\definecolor[named]{pgfstrokecolor}{rgb}{1,1,1}\pgfsys@color@gray@stroke{1}\pgfsys@invoke{
}\pgfsys@color@gray@fill{1}\pgfsys@invoke{
}\definecolor[named]{pgffillcolor}{rgb}{1,1,1}\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@curveto{21.33957pt}{3.92854pt}{18.15492pt}{7.11319pt}{14.22638pt}{7.11319pt}\pgfsys@curveto{10.29784pt}{7.11319pt}{7.11319pt}{3.92854pt}{7.11319pt}{0.0pt}\pgfsys@curveto{7.11319pt}{-3.92854pt}{10.29784pt}{-7.11319pt}{14.22638pt}{-7.11319pt}\pgfsys@curveto{18.15492pt}{-7.11319pt}{21.33957pt}{-3.92854pt}{21.33957pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@fill\pgfsys@invoke{
} \pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope\pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope}\pgfsys@beginscope\pgfsys@invoke{
}\definecolor[named]{pgf@tempcolor}{rgb}{0,0,0}\pgfsys@invoke{\lxSVG@setpatternuncolored@{3}{0}{0}{0}
}{}\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@curveto{21.33957pt}{3.92854pt}{18.15492pt}{7.11319pt}{14.22638pt}{7.11319pt}\pgfsys@curveto{10.29784pt}{7.11319pt}{7.11319pt}{3.92854pt}{7.11319pt}{0.0pt}\pgfsys@curveto{7.11319pt}{-3.92854pt}{10.29784pt}{-7.11319pt}{14.22638pt}{-7.11319pt}\pgfsys@curveto{18.15492pt}{-7.11319pt}{21.33957pt}{-3.92854pt}{21.33957pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
} \pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@moveto{8.11319pt}{0.0pt}\pgfsys@curveto{8.11319pt}{0.55229pt}{7.66548pt}{1.0pt}{7.11319pt}{1.0pt}\pgfsys@curveto{6.5609pt}{1.0pt}{6.11319pt}{0.55229pt}{6.11319pt}{0.0pt}\pgfsys@curveto{6.11319pt}{-0.55229pt}{6.5609pt}{-1.0pt}{7.11319pt}{-1.0pt}\pgfsys@curveto{7.66548pt}{-1.0pt}{8.11319pt}{-0.55229pt}{8.11319pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@moveto{22.33957pt}{0.0pt}\pgfsys@curveto{22.33957pt}{0.55229pt}{21.89186pt}{1.0pt}{21.33957pt}{1.0pt}\pgfsys@curveto{20.78728pt}{1.0pt}{20.33957pt}{0.55229pt}{20.33957pt}{0.0pt}\pgfsys@curveto{20.33957pt}{-0.55229pt}{20.78728pt}{-1.0pt}{21.33957pt}{-1.0pt}\pgfsys@curveto{21.89186pt}{-1.0pt}{22.33957pt}{-0.55229pt}{22.33957pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{0.0pt}{0.0pt}\pgfsys@lineto{7.11319pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@lineto{28.45276pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope{{
{}{}{}{}{}}{{{}}{{}}}}{}{}\hss}\pgfsys@discardpath\pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope\hss}}\lxSVG@closescope\endpgfpicture}}=\frac{1}{2}\leavevmode\hbox
to28.85pt{\vbox to14.63pt{\pgfpicture\makeatletter\hbox{\hskip
0.2pt\lower-7.31319pt\hbox to0.0pt{\pgfsys@beginscope\pgfsys@invoke{
}\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{
}\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{
}\pgfsys@setlinewidth{0.4pt}\pgfsys@invoke{ }\nullfont\hbox
to0.0pt{\pgfsys@beginscope\pgfsys@invoke{ }
{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@moveto{8.11319pt}{0.0pt}\pgfsys@curveto{8.11319pt}{0.55229pt}{7.66548pt}{1.0pt}{7.11319pt}{1.0pt}\pgfsys@curveto{6.5609pt}{1.0pt}{6.11319pt}{0.55229pt}{6.11319pt}{0.0pt}\pgfsys@curveto{6.11319pt}{-0.55229pt}{6.5609pt}{-1.0pt}{7.11319pt}{-1.0pt}\pgfsys@curveto{7.66548pt}{-1.0pt}{8.11319pt}{-0.55229pt}{8.11319pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@moveto{22.33957pt}{0.0pt}\pgfsys@curveto{22.33957pt}{0.55229pt}{21.89186pt}{1.0pt}{21.33957pt}{1.0pt}\pgfsys@curveto{20.78728pt}{1.0pt}{20.33957pt}{0.55229pt}{20.33957pt}{0.0pt}\pgfsys@curveto{20.33957pt}{-0.55229pt}{20.78728pt}{-1.0pt}{21.33957pt}{-1.0pt}\pgfsys@curveto{21.89186pt}{-1.0pt}{22.33957pt}{-0.55229pt}{22.33957pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{0.0pt}{0.0pt}\pgfsys@lineto{7.11319pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@lineto{28.45276pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@curveto{21.33957pt}{3.92854pt}{18.15492pt}{7.11319pt}{14.22638pt}{7.11319pt}\pgfsys@curveto{10.29784pt}{7.11319pt}{7.11319pt}{3.92854pt}{7.11319pt}{0.0pt}\pgfsys@curveto{7.11319pt}{-3.92854pt}{10.29784pt}{-7.11319pt}{14.22638pt}{-7.11319pt}\pgfsys@curveto{18.15492pt}{-7.11319pt}{21.33957pt}{-3.92854pt}{21.33957pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope{{
{}{}{}{}{}}{{{}}{{}}}}{}{}\hss}\pgfsys@discardpath\pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope\hss}}\lxSVG@closescope\endpgfpicture}}+\frac{1}{2}\leavevmode\hbox
to43.08pt{\vbox to28.85pt{\pgfpicture\makeatletter\hbox{\hskip
0.2pt\lower-14.42638pt\hbox to0.0pt{\pgfsys@beginscope\pgfsys@invoke{
}\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{
}\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{
}\pgfsys@setlinewidth{0.4pt}\pgfsys@invoke{ }\nullfont\hbox
to0.0pt{\pgfsys@beginscope\pgfsys@invoke{ }
{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{{}}{}{{}{}{}{{}}{{{{}{}{}{}}} {{}{}{}{}}}}{} {} {} {}
{}{}\pgfsys@beginscope\pgfsys@invoke{
}\color[rgb]{1,1,1}\definecolor[named]{pgfstrokecolor}{rgb}{1,1,1}\pgfsys@color@gray@stroke{1}\pgfsys@invoke{
}\pgfsys@color@gray@fill{1}\pgfsys@invoke{
}\definecolor[named]{pgffillcolor}{rgb}{1,1,1}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@curveto{7.11319pt}{-7.8571pt}{13.48247pt}{-14.22638pt}{21.33957pt}{-14.22638pt}\pgfsys@curveto{29.19667pt}{-14.22638pt}{35.56595pt}{-7.8571pt}{35.56595pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} \pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@moveto{8.11319pt}{0.0pt}\pgfsys@curveto{8.11319pt}{0.55229pt}{7.66548pt}{1.0pt}{7.11319pt}{1.0pt}\pgfsys@curveto{6.5609pt}{1.0pt}{6.11319pt}{0.55229pt}{6.11319pt}{0.0pt}\pgfsys@curveto{6.11319pt}{-0.55229pt}{6.5609pt}{-1.0pt}{7.11319pt}{-1.0pt}\pgfsys@curveto{7.66548pt}{-1.0pt}{8.11319pt}{-0.55229pt}{8.11319pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@moveto{15.22638pt}{0.0pt}\pgfsys@curveto{15.22638pt}{0.55229pt}{14.77867pt}{1.0pt}{14.22638pt}{1.0pt}\pgfsys@curveto{13.67409pt}{1.0pt}{13.22638pt}{0.55229pt}{13.22638pt}{0.0pt}\pgfsys@curveto{13.22638pt}{-0.55229pt}{13.67409pt}{-1.0pt}{14.22638pt}{-1.0pt}\pgfsys@curveto{14.77867pt}{-1.0pt}{15.22638pt}{-0.55229pt}{15.22638pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{28.45276pt}{0.0pt}\pgfsys@moveto{29.45276pt}{0.0pt}\pgfsys@curveto{29.45276pt}{0.55229pt}{29.00505pt}{1.0pt}{28.45276pt}{1.0pt}\pgfsys@curveto{27.90047pt}{1.0pt}{27.45276pt}{0.55229pt}{27.45276pt}{0.0pt}\pgfsys@curveto{27.45276pt}{-0.55229pt}{27.90047pt}{-1.0pt}{28.45276pt}{-1.0pt}\pgfsys@curveto{29.00505pt}{-1.0pt}{29.45276pt}{-0.55229pt}{29.45276pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{28.45276pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{35.56595pt}{0.0pt}\pgfsys@moveto{36.56595pt}{0.0pt}\pgfsys@curveto{36.56595pt}{0.55229pt}{36.11824pt}{1.0pt}{35.56595pt}{1.0pt}\pgfsys@curveto{35.01366pt}{1.0pt}{34.56595pt}{0.55229pt}{34.56595pt}{0.0pt}\pgfsys@curveto{34.56595pt}{-0.55229pt}{35.01366pt}{-1.0pt}{35.56595pt}{-1.0pt}\pgfsys@curveto{36.11824pt}{-1.0pt}{36.56595pt}{-0.55229pt}{36.56595pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{35.56595pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{0.0pt}{0.0pt}\pgfsys@lineto{7.11319pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@lineto{14.22638pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@lineto{28.45276pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{28.45276pt}{0.0pt}\pgfsys@lineto{35.56595pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{35.56595pt}{0.0pt}\pgfsys@lineto{42.67914pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}{}{}{{}}{{{{}{}{}{}}}{{}{}{}{}}}}{} {} {} {}
{}{}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@curveto{7.11319pt}{7.8571pt}{13.48247pt}{14.22638pt}{21.33957pt}{14.22638pt}\pgfsys@curveto{29.19667pt}{14.22638pt}{35.56595pt}{7.8571pt}{35.56595pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}{}{}{{}}{{{{}{}{}{}}}{{}{}{}{}}}}{} {} {} {}
{}{}{}\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@curveto{14.22638pt}{3.92854pt}{17.41103pt}{7.11319pt}{21.33957pt}{7.11319pt}\pgfsys@curveto{25.26811pt}{7.11319pt}{28.45276pt}{3.92854pt}{28.45276pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope{{
{}{}{}{}{}}{{{}}{{}}}}{}{}\hss}\pgfsys@discardpath\pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope\hss}}\lxSVG@closescope\endpgfpicture}}+\frac{1}{4}\leavevmode\hbox
to64.42pt{\vbox to50.19pt{\pgfpicture\makeatletter\hbox{\hskip
0.2pt\lower-25.09616pt\hbox to0.0pt{\pgfsys@beginscope\pgfsys@invoke{
}\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{
}\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{
}\pgfsys@setlinewidth{0.4pt}\pgfsys@invoke{ }\nullfont\hbox
to0.0pt{\pgfsys@beginscope\pgfsys@invoke{ }
{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{{}}{}{{}{}{}{{}}{{{{}{}{}{}}} {{}{}{}{}}}}{} {} {} {}
{}{}\pgfsys@beginscope\pgfsys@invoke{
}\color[rgb]{1,1,1}\definecolor[named]{pgfstrokecolor}{rgb}{1,1,1}\pgfsys@color@gray@stroke{1}\pgfsys@invoke{
}\pgfsys@color@gray@fill{1}\pgfsys@invoke{
}\definecolor[named]{pgffillcolor}{rgb}{1,1,1}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@curveto{7.11319pt}{-13.74994pt}{18.25941pt}{-24.89616pt}{32.00935pt}{-24.89616pt}\pgfsys@curveto{45.7593pt}{-24.89616pt}{56.90552pt}{-13.74994pt}{56.90552pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} \pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@moveto{8.11319pt}{0.0pt}\pgfsys@curveto{8.11319pt}{0.55229pt}{7.66548pt}{1.0pt}{7.11319pt}{1.0pt}\pgfsys@curveto{6.5609pt}{1.0pt}{6.11319pt}{0.55229pt}{6.11319pt}{0.0pt}\pgfsys@curveto{6.11319pt}{-0.55229pt}{6.5609pt}{-1.0pt}{7.11319pt}{-1.0pt}\pgfsys@curveto{7.66548pt}{-1.0pt}{8.11319pt}{-0.55229pt}{8.11319pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@moveto{15.22638pt}{0.0pt}\pgfsys@curveto{15.22638pt}{0.55229pt}{14.77867pt}{1.0pt}{14.22638pt}{1.0pt}\pgfsys@curveto{13.67409pt}{1.0pt}{13.22638pt}{0.55229pt}{13.22638pt}{0.0pt}\pgfsys@curveto{13.22638pt}{-0.55229pt}{13.67409pt}{-1.0pt}{14.22638pt}{-1.0pt}\pgfsys@curveto{14.77867pt}{-1.0pt}{15.22638pt}{-0.55229pt}{15.22638pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{28.45276pt}{0.0pt}\pgfsys@moveto{29.45276pt}{0.0pt}\pgfsys@curveto{29.45276pt}{0.55229pt}{29.00505pt}{1.0pt}{28.45276pt}{1.0pt}\pgfsys@curveto{27.90047pt}{1.0pt}{27.45276pt}{0.55229pt}{27.45276pt}{0.0pt}\pgfsys@curveto{27.45276pt}{-0.55229pt}{27.90047pt}{-1.0pt}{28.45276pt}{-1.0pt}\pgfsys@curveto{29.00505pt}{-1.0pt}{29.45276pt}{-0.55229pt}{29.45276pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{28.45276pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{35.56595pt}{0.0pt}\pgfsys@moveto{36.56595pt}{0.0pt}\pgfsys@curveto{36.56595pt}{0.55229pt}{36.11824pt}{1.0pt}{35.56595pt}{1.0pt}\pgfsys@curveto{35.01366pt}{1.0pt}{34.56595pt}{0.55229pt}{34.56595pt}{0.0pt}\pgfsys@curveto{34.56595pt}{-0.55229pt}{35.01366pt}{-1.0pt}{35.56595pt}{-1.0pt}\pgfsys@curveto{36.11824pt}{-1.0pt}{36.56595pt}{-0.55229pt}{36.56595pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{35.56595pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{49.79233pt}{0.0pt}\pgfsys@moveto{50.79233pt}{0.0pt}\pgfsys@curveto{50.79233pt}{0.55229pt}{50.34462pt}{1.0pt}{49.79233pt}{1.0pt}\pgfsys@curveto{49.24004pt}{1.0pt}{48.79233pt}{0.55229pt}{48.79233pt}{0.0pt}\pgfsys@curveto{48.79233pt}{-0.55229pt}{49.24004pt}{-1.0pt}{49.79233pt}{-1.0pt}\pgfsys@curveto{50.34462pt}{-1.0pt}{50.79233pt}{-0.55229pt}{50.79233pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{49.79233pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{56.90552pt}{0.0pt}\pgfsys@moveto{57.90552pt}{0.0pt}\pgfsys@curveto{57.90552pt}{0.55229pt}{57.45781pt}{1.0pt}{56.90552pt}{1.0pt}\pgfsys@curveto{56.35323pt}{1.0pt}{55.90552pt}{0.55229pt}{55.90552pt}{0.0pt}\pgfsys@curveto{55.90552pt}{-0.55229pt}{56.35323pt}{-1.0pt}{56.90552pt}{-1.0pt}\pgfsys@curveto{57.45781pt}{-1.0pt}{57.90552pt}{-0.55229pt}{57.90552pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{56.90552pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{0.0pt}{0.0pt}\pgfsys@lineto{7.11319pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@lineto{14.22638pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@lineto{28.45276pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{28.45276pt}{0.0pt}\pgfsys@lineto{35.56595pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{35.56595pt}{0.0pt}\pgfsys@lineto{49.79233pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{49.79233pt}{0.0pt}\pgfsys@lineto{56.90552pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{56.90552pt}{0.0pt}\pgfsys@lineto{64.0187pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}{}{}{{}}{{{{}{}{}{}}}{{}{}{}{}}}}{} {} {} {}
{}{}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@curveto{7.11319pt}{13.74994pt}{18.25941pt}{24.89616pt}{32.00935pt}{24.89616pt}\pgfsys@curveto{45.7593pt}{24.89616pt}{56.90552pt}{13.74994pt}{56.90552pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}{}{}{{}}{{{{}{}{}{}}}{{}{}{}{}}}}{} {} {} {}
{}{}{}\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@curveto{14.22638pt}{3.92854pt}{17.41103pt}{7.11319pt}{21.33957pt}{7.11319pt}\pgfsys@curveto{25.26811pt}{7.11319pt}{28.45276pt}{3.92854pt}{28.45276pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}{}{}{{}}{{{{}{}{}{}}}{{}{}{}{}}}}{} {} {} {}
{}{}{}\pgfsys@moveto{35.56595pt}{0.0pt}\pgfsys@curveto{35.56595pt}{3.92854pt}{38.7506pt}{7.11319pt}{42.67914pt}{7.11319pt}\pgfsys@curveto{46.60768pt}{7.11319pt}{49.79233pt}{3.92854pt}{49.79233pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope{{
{}{}{}{}{}}{{{}}{{}}}}{}{}\hss}\pgfsys@discardpath\pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope\hss}}\lxSVG@closescope\endpgfpicture}}+\frac{1}{8}\leavevmode\hbox
to85.76pt{\vbox to71.53pt{\pgfpicture\makeatletter\hbox{\hskip
0.2pt\lower-35.76595pt\hbox to0.0pt{\pgfsys@beginscope\pgfsys@invoke{
}\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{
}\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{
}\pgfsys@setlinewidth{0.4pt}\pgfsys@invoke{ }\nullfont\hbox
to0.0pt{\pgfsys@beginscope\pgfsys@invoke{ }
{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{{}}{}{{}{}{}{{}}{{{{}{}{}{}}} {{}{}{}{}}}}{} {} {} {}
{}{}\pgfsys@beginscope\pgfsys@invoke{
}\color[rgb]{1,1,1}\definecolor[named]{pgfstrokecolor}{rgb}{1,1,1}\pgfsys@color@gray@stroke{1}\pgfsys@invoke{
}\pgfsys@color@gray@fill{1}\pgfsys@invoke{
}\definecolor[named]{pgffillcolor}{rgb}{1,1,1}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@curveto{7.11319pt}{-19.64278pt}{23.03636pt}{-35.56595pt}{42.67914pt}{-35.56595pt}\pgfsys@curveto{62.32191pt}{-35.56595pt}{78.24509pt}{-19.64278pt}{78.24509pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} \pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@moveto{8.11319pt}{0.0pt}\pgfsys@curveto{8.11319pt}{0.55229pt}{7.66548pt}{1.0pt}{7.11319pt}{1.0pt}\pgfsys@curveto{6.5609pt}{1.0pt}{6.11319pt}{0.55229pt}{6.11319pt}{0.0pt}\pgfsys@curveto{6.11319pt}{-0.55229pt}{6.5609pt}{-1.0pt}{7.11319pt}{-1.0pt}\pgfsys@curveto{7.66548pt}{-1.0pt}{8.11319pt}{-0.55229pt}{8.11319pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@moveto{15.22638pt}{0.0pt}\pgfsys@curveto{15.22638pt}{0.55229pt}{14.77867pt}{1.0pt}{14.22638pt}{1.0pt}\pgfsys@curveto{13.67409pt}{1.0pt}{13.22638pt}{0.55229pt}{13.22638pt}{0.0pt}\pgfsys@curveto{13.22638pt}{-0.55229pt}{13.67409pt}{-1.0pt}{14.22638pt}{-1.0pt}\pgfsys@curveto{14.77867pt}{-1.0pt}{15.22638pt}{-0.55229pt}{15.22638pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{28.45276pt}{0.0pt}\pgfsys@moveto{29.45276pt}{0.0pt}\pgfsys@curveto{29.45276pt}{0.55229pt}{29.00505pt}{1.0pt}{28.45276pt}{1.0pt}\pgfsys@curveto{27.90047pt}{1.0pt}{27.45276pt}{0.55229pt}{27.45276pt}{0.0pt}\pgfsys@curveto{27.45276pt}{-0.55229pt}{27.90047pt}{-1.0pt}{28.45276pt}{-1.0pt}\pgfsys@curveto{29.00505pt}{-1.0pt}{29.45276pt}{-0.55229pt}{29.45276pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{28.45276pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{35.56595pt}{0.0pt}\pgfsys@moveto{36.56595pt}{0.0pt}\pgfsys@curveto{36.56595pt}{0.55229pt}{36.11824pt}{1.0pt}{35.56595pt}{1.0pt}\pgfsys@curveto{35.01366pt}{1.0pt}{34.56595pt}{0.55229pt}{34.56595pt}{0.0pt}\pgfsys@curveto{34.56595pt}{-0.55229pt}{35.01366pt}{-1.0pt}{35.56595pt}{-1.0pt}\pgfsys@curveto{36.11824pt}{-1.0pt}{36.56595pt}{-0.55229pt}{36.56595pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{35.56595pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{49.79233pt}{0.0pt}\pgfsys@moveto{50.79233pt}{0.0pt}\pgfsys@curveto{50.79233pt}{0.55229pt}{50.34462pt}{1.0pt}{49.79233pt}{1.0pt}\pgfsys@curveto{49.24004pt}{1.0pt}{48.79233pt}{0.55229pt}{48.79233pt}{0.0pt}\pgfsys@curveto{48.79233pt}{-0.55229pt}{49.24004pt}{-1.0pt}{49.79233pt}{-1.0pt}\pgfsys@curveto{50.34462pt}{-1.0pt}{50.79233pt}{-0.55229pt}{50.79233pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{49.79233pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{56.90552pt}{0.0pt}\pgfsys@moveto{57.90552pt}{0.0pt}\pgfsys@curveto{57.90552pt}{0.55229pt}{57.45781pt}{1.0pt}{56.90552pt}{1.0pt}\pgfsys@curveto{56.35323pt}{1.0pt}{55.90552pt}{0.55229pt}{55.90552pt}{0.0pt}\pgfsys@curveto{55.90552pt}{-0.55229pt}{56.35323pt}{-1.0pt}{56.90552pt}{-1.0pt}\pgfsys@curveto{57.45781pt}{-1.0pt}{57.90552pt}{-0.55229pt}{57.90552pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{56.90552pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{71.1319pt}{0.0pt}\pgfsys@moveto{72.1319pt}{0.0pt}\pgfsys@curveto{72.1319pt}{0.55229pt}{71.68419pt}{1.0pt}{71.1319pt}{1.0pt}\pgfsys@curveto{70.5796pt}{1.0pt}{70.1319pt}{0.55229pt}{70.1319pt}{0.0pt}\pgfsys@curveto{70.1319pt}{-0.55229pt}{70.5796pt}{-1.0pt}{71.1319pt}{-1.0pt}\pgfsys@curveto{71.68419pt}{-1.0pt}{72.1319pt}{-0.55229pt}{72.1319pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{71.1319pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{78.24509pt}{0.0pt}\pgfsys@moveto{79.24509pt}{0.0pt}\pgfsys@curveto{79.24509pt}{0.55229pt}{78.79738pt}{1.0pt}{78.24509pt}{1.0pt}\pgfsys@curveto{77.6928pt}{1.0pt}{77.24509pt}{0.55229pt}{77.24509pt}{0.0pt}\pgfsys@curveto{77.24509pt}{-0.55229pt}{77.6928pt}{-1.0pt}{78.24509pt}{-1.0pt}\pgfsys@curveto{78.79738pt}{-1.0pt}{79.24509pt}{-0.55229pt}{79.24509pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{78.24509pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{0.0pt}{0.0pt}\pgfsys@lineto{7.11319pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@lineto{14.22638pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@lineto{28.45276pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{28.45276pt}{0.0pt}\pgfsys@lineto{35.56595pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{35.56595pt}{0.0pt}\pgfsys@lineto{49.79233pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{49.79233pt}{0.0pt}\pgfsys@lineto{56.90552pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{56.90552pt}{0.0pt}\pgfsys@lineto{71.1319pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{71.1319pt}{0.0pt}\pgfsys@lineto{78.24509pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{56.90552pt}{0.0pt}\pgfsys@lineto{85.35828pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}{}{}{{}}{{{{}{}{}{}}}{{}{}{}{}}}}{} {} {} {}
{}{}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@curveto{7.11319pt}{19.64278pt}{23.03636pt}{35.56595pt}{42.67914pt}{35.56595pt}\pgfsys@curveto{62.32191pt}{35.56595pt}{78.24509pt}{19.64278pt}{78.24509pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}{}{}{{}}{{{{}{}{}{}}}{{}{}{}{}}}}{} {} {} {}
{}{}{}\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@curveto{14.22638pt}{3.92854pt}{17.41103pt}{7.11319pt}{21.33957pt}{7.11319pt}\pgfsys@curveto{25.26811pt}{7.11319pt}{28.45276pt}{3.92854pt}{28.45276pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}{}{}{{}}{{{{}{}{}{}}}{{}{}{}{}}}}{} {} {} {}
{}{}{}\pgfsys@moveto{35.56595pt}{0.0pt}\pgfsys@curveto{35.56595pt}{3.92854pt}{38.7506pt}{7.11319pt}{42.67914pt}{7.11319pt}\pgfsys@curveto{46.60768pt}{7.11319pt}{49.79233pt}{3.92854pt}{49.79233pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}{}{}{{}}{{{{}{}{}{}}}{{}{}{}{}}}}{} {} {} {}
{}{}{}\pgfsys@moveto{56.90552pt}{0.0pt}\pgfsys@curveto{56.90552pt}{3.92854pt}{60.09016pt}{7.11319pt}{64.0187pt}{7.11319pt}\pgfsys@curveto{67.94725pt}{7.11319pt}{71.1319pt}{3.92854pt}{71.1319pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope{{
{}{}{}{}{}}{{{}}{{}}}}{}{}\hss}\pgfsys@discardpath\pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope\hss}}\lxSVG@closescope\endpgfpicture}}+\cdots\hfill\vspace{-4ex}$
($b$) Chain approximation
$\leavevmode\hbox to28.85pt{\vbox
to14.63pt{\pgfpicture\makeatletter\hbox{\hskip 0.2pt\lower-7.31319pt\hbox
to0.0pt{\pgfsys@beginscope\pgfsys@invoke{
}\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{
}\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{
}\pgfsys@setlinewidth{0.4pt}\pgfsys@invoke{ }\nullfont\hbox
to0.0pt{\pgfsys@beginscope\pgfsys@invoke{ }
{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{\pgfsys@beginscope\pgfsys@invoke{
}\pgfsys@beginscope\pgfsys@invoke{
}\color[rgb]{1,1,1}\definecolor[named]{pgfstrokecolor}{rgb}{1,1,1}\pgfsys@color@gray@stroke{1}\pgfsys@invoke{
}\pgfsys@color@gray@fill{1}\pgfsys@invoke{
}\definecolor[named]{pgffillcolor}{rgb}{1,1,1}\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@curveto{21.33957pt}{3.92854pt}{18.15492pt}{7.11319pt}{14.22638pt}{7.11319pt}\pgfsys@curveto{10.29784pt}{7.11319pt}{7.11319pt}{3.92854pt}{7.11319pt}{0.0pt}\pgfsys@curveto{7.11319pt}{-3.92854pt}{10.29784pt}{-7.11319pt}{14.22638pt}{-7.11319pt}\pgfsys@curveto{18.15492pt}{-7.11319pt}{21.33957pt}{-3.92854pt}{21.33957pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@fill\pgfsys@invoke{
} \pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope\pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope}\pgfsys@beginscope\pgfsys@invoke{
}\definecolor[named]{pgf@tempcolor}{rgb}{0,0,0}\pgfsys@invoke{\lxSVG@setpatternuncolored@{3}{0}{0}{0}
}{}\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@curveto{21.33957pt}{3.92854pt}{18.15492pt}{7.11319pt}{14.22638pt}{7.11319pt}\pgfsys@curveto{10.29784pt}{7.11319pt}{7.11319pt}{3.92854pt}{7.11319pt}{0.0pt}\pgfsys@curveto{7.11319pt}{-3.92854pt}{10.29784pt}{-7.11319pt}{14.22638pt}{-7.11319pt}\pgfsys@curveto{18.15492pt}{-7.11319pt}{21.33957pt}{-3.92854pt}{21.33957pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
} \pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@moveto{8.11319pt}{0.0pt}\pgfsys@curveto{8.11319pt}{0.55229pt}{7.66548pt}{1.0pt}{7.11319pt}{1.0pt}\pgfsys@curveto{6.5609pt}{1.0pt}{6.11319pt}{0.55229pt}{6.11319pt}{0.0pt}\pgfsys@curveto{6.11319pt}{-0.55229pt}{6.5609pt}{-1.0pt}{7.11319pt}{-1.0pt}\pgfsys@curveto{7.66548pt}{-1.0pt}{8.11319pt}{-0.55229pt}{8.11319pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@moveto{22.33957pt}{0.0pt}\pgfsys@curveto{22.33957pt}{0.55229pt}{21.89186pt}{1.0pt}{21.33957pt}{1.0pt}\pgfsys@curveto{20.78728pt}{1.0pt}{20.33957pt}{0.55229pt}{20.33957pt}{0.0pt}\pgfsys@curveto{20.33957pt}{-0.55229pt}{20.78728pt}{-1.0pt}{21.33957pt}{-1.0pt}\pgfsys@curveto{21.89186pt}{-1.0pt}{22.33957pt}{-0.55229pt}{22.33957pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{0.0pt}{0.0pt}\pgfsys@lineto{7.11319pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@lineto{28.45276pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope{{
{}{}{}{}{}}{{{}}{{}}}}{}{}\hss}\pgfsys@discardpath\pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope\hss}}\lxSVG@closescope\endpgfpicture}}=\frac{1}{2}\leavevmode\hbox
to28.85pt{\vbox to14.63pt{\pgfpicture\makeatletter\hbox{\hskip
0.2pt\lower-7.31319pt\hbox to0.0pt{\pgfsys@beginscope\pgfsys@invoke{
}\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{
}\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{
}\pgfsys@setlinewidth{0.4pt}\pgfsys@invoke{ }\nullfont\hbox
to0.0pt{\pgfsys@beginscope\pgfsys@invoke{ }
{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@moveto{8.11319pt}{0.0pt}\pgfsys@curveto{8.11319pt}{0.55229pt}{7.66548pt}{1.0pt}{7.11319pt}{1.0pt}\pgfsys@curveto{6.5609pt}{1.0pt}{6.11319pt}{0.55229pt}{6.11319pt}{0.0pt}\pgfsys@curveto{6.11319pt}{-0.55229pt}{6.5609pt}{-1.0pt}{7.11319pt}{-1.0pt}\pgfsys@curveto{7.66548pt}{-1.0pt}{8.11319pt}{-0.55229pt}{8.11319pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@moveto{22.33957pt}{0.0pt}\pgfsys@curveto{22.33957pt}{0.55229pt}{21.89186pt}{1.0pt}{21.33957pt}{1.0pt}\pgfsys@curveto{20.78728pt}{1.0pt}{20.33957pt}{0.55229pt}{20.33957pt}{0.0pt}\pgfsys@curveto{20.33957pt}{-0.55229pt}{20.78728pt}{-1.0pt}{21.33957pt}{-1.0pt}\pgfsys@curveto{21.89186pt}{-1.0pt}{22.33957pt}{-0.55229pt}{22.33957pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{0.0pt}{0.0pt}\pgfsys@lineto{7.11319pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@lineto{28.45276pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@curveto{21.33957pt}{3.92854pt}{18.15492pt}{7.11319pt}{14.22638pt}{7.11319pt}\pgfsys@curveto{10.29784pt}{7.11319pt}{7.11319pt}{3.92854pt}{7.11319pt}{0.0pt}\pgfsys@curveto{7.11319pt}{-3.92854pt}{10.29784pt}{-7.11319pt}{14.22638pt}{-7.11319pt}\pgfsys@curveto{18.15492pt}{-7.11319pt}{21.33957pt}{-3.92854pt}{21.33957pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope{{
{}{}{}{}{}}{{{}}{{}}}}{}{}\hss}\pgfsys@discardpath\pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope\hss}}\lxSVG@closescope\endpgfpicture}}+\frac{1}{2}\leavevmode\hbox
to43.08pt{\vbox to28.85pt{\pgfpicture\makeatletter\hbox{\hskip
0.2pt\lower-14.42638pt\hbox to0.0pt{\pgfsys@beginscope\pgfsys@invoke{
}\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{
}\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{
}\pgfsys@setlinewidth{0.4pt}\pgfsys@invoke{ }\nullfont\hbox
to0.0pt{\pgfsys@beginscope\pgfsys@invoke{ }
{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{{}}{}{{}{}{}{{}}{{{{}{}{}{}}} {{}{}{}{}}}}{} {} {} {}
{}{}\pgfsys@beginscope\pgfsys@invoke{
}\color[rgb]{1,1,1}\definecolor[named]{pgfstrokecolor}{rgb}{1,1,1}\pgfsys@color@gray@stroke{1}\pgfsys@invoke{
}\pgfsys@color@gray@fill{1}\pgfsys@invoke{
}\definecolor[named]{pgffillcolor}{rgb}{1,1,1}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@curveto{7.11319pt}{-7.8571pt}{13.48247pt}{-14.22638pt}{21.33957pt}{-14.22638pt}\pgfsys@curveto{29.19667pt}{-14.22638pt}{35.56595pt}{-7.8571pt}{35.56595pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} \pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@moveto{8.11319pt}{0.0pt}\pgfsys@curveto{8.11319pt}{0.55229pt}{7.66548pt}{1.0pt}{7.11319pt}{1.0pt}\pgfsys@curveto{6.5609pt}{1.0pt}{6.11319pt}{0.55229pt}{6.11319pt}{0.0pt}\pgfsys@curveto{6.11319pt}{-0.55229pt}{6.5609pt}{-1.0pt}{7.11319pt}{-1.0pt}\pgfsys@curveto{7.66548pt}{-1.0pt}{8.11319pt}{-0.55229pt}{8.11319pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@moveto{15.22638pt}{0.0pt}\pgfsys@curveto{15.22638pt}{0.55229pt}{14.77867pt}{1.0pt}{14.22638pt}{1.0pt}\pgfsys@curveto{13.67409pt}{1.0pt}{13.22638pt}{0.55229pt}{13.22638pt}{0.0pt}\pgfsys@curveto{13.22638pt}{-0.55229pt}{13.67409pt}{-1.0pt}{14.22638pt}{-1.0pt}\pgfsys@curveto{14.77867pt}{-1.0pt}{15.22638pt}{-0.55229pt}{15.22638pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{28.45276pt}{0.0pt}\pgfsys@moveto{29.45276pt}{0.0pt}\pgfsys@curveto{29.45276pt}{0.55229pt}{29.00505pt}{1.0pt}{28.45276pt}{1.0pt}\pgfsys@curveto{27.90047pt}{1.0pt}{27.45276pt}{0.55229pt}{27.45276pt}{0.0pt}\pgfsys@curveto{27.45276pt}{-0.55229pt}{27.90047pt}{-1.0pt}{28.45276pt}{-1.0pt}\pgfsys@curveto{29.00505pt}{-1.0pt}{29.45276pt}{-0.55229pt}{29.45276pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{28.45276pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{35.56595pt}{0.0pt}\pgfsys@moveto{36.56595pt}{0.0pt}\pgfsys@curveto{36.56595pt}{0.55229pt}{36.11824pt}{1.0pt}{35.56595pt}{1.0pt}\pgfsys@curveto{35.01366pt}{1.0pt}{34.56595pt}{0.55229pt}{34.56595pt}{0.0pt}\pgfsys@curveto{34.56595pt}{-0.55229pt}{35.01366pt}{-1.0pt}{35.56595pt}{-1.0pt}\pgfsys@curveto{36.11824pt}{-1.0pt}{36.56595pt}{-0.55229pt}{36.56595pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{35.56595pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{0.0pt}{0.0pt}\pgfsys@lineto{7.11319pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@lineto{14.22638pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@lineto{28.45276pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{28.45276pt}{0.0pt}\pgfsys@lineto{35.56595pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{35.56595pt}{0.0pt}\pgfsys@lineto{42.67914pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}{}{}{{}}{{{{}{}{}{}}}{{}{}{}{}}}}{} {} {} {}
{}{}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@curveto{7.11319pt}{7.8571pt}{13.48247pt}{14.22638pt}{21.33957pt}{14.22638pt}\pgfsys@curveto{29.19667pt}{14.22638pt}{35.56595pt}{7.8571pt}{35.56595pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}{}{}{{}}{{{{}{}{}{}}}{{}{}{}{}}}}{} {} {} {}
{}{}{}\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@curveto{14.22638pt}{3.92854pt}{17.41103pt}{7.11319pt}{21.33957pt}{7.11319pt}\pgfsys@curveto{25.26811pt}{7.11319pt}{28.45276pt}{3.92854pt}{28.45276pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope{{
{}{}{}{}{}}{{{}}{{}}}}{}{}\hss}\pgfsys@discardpath\pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope\hss}}\lxSVG@closescope\endpgfpicture}}+\frac{1}{2}\leavevmode\hbox
to57.31pt{\vbox to43.08pt{\pgfpicture\makeatletter\hbox{\hskip
0.2pt\lower-21.53957pt\hbox to0.0pt{\pgfsys@beginscope\pgfsys@invoke{
}\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{
}\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{
}\pgfsys@setlinewidth{0.4pt}\pgfsys@invoke{ }\nullfont\hbox
to0.0pt{\pgfsys@beginscope\pgfsys@invoke{ }
{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{{}}{}{{}{}{}{{}}{{{{}{}{}{}}} {{}{}{}{}}}}{} {} {} {}
{}{}\pgfsys@beginscope\pgfsys@invoke{
}\color[rgb]{1,1,1}\definecolor[named]{pgfstrokecolor}{rgb}{1,1,1}\pgfsys@color@gray@stroke{1}\pgfsys@invoke{
}\pgfsys@color@gray@fill{1}\pgfsys@invoke{
}\definecolor[named]{pgffillcolor}{rgb}{1,1,1}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@curveto{7.11319pt}{-11.78566pt}{16.6671pt}{-21.33957pt}{28.45276pt}{-21.33957pt}\pgfsys@curveto{40.23842pt}{-21.33957pt}{49.79233pt}{-11.78566pt}{49.79233pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} \pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@moveto{8.11319pt}{0.0pt}\pgfsys@curveto{8.11319pt}{0.55229pt}{7.66548pt}{1.0pt}{7.11319pt}{1.0pt}\pgfsys@curveto{6.5609pt}{1.0pt}{6.11319pt}{0.55229pt}{6.11319pt}{0.0pt}\pgfsys@curveto{6.11319pt}{-0.55229pt}{6.5609pt}{-1.0pt}{7.11319pt}{-1.0pt}\pgfsys@curveto{7.66548pt}{-1.0pt}{8.11319pt}{-0.55229pt}{8.11319pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@moveto{15.22638pt}{0.0pt}\pgfsys@curveto{15.22638pt}{0.55229pt}{14.77867pt}{1.0pt}{14.22638pt}{1.0pt}\pgfsys@curveto{13.67409pt}{1.0pt}{13.22638pt}{0.55229pt}{13.22638pt}{0.0pt}\pgfsys@curveto{13.22638pt}{-0.55229pt}{13.67409pt}{-1.0pt}{14.22638pt}{-1.0pt}\pgfsys@curveto{14.77867pt}{-1.0pt}{15.22638pt}{-0.55229pt}{15.22638pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@moveto{22.33957pt}{0.0pt}\pgfsys@curveto{22.33957pt}{0.55229pt}{21.89186pt}{1.0pt}{21.33957pt}{1.0pt}\pgfsys@curveto{20.78728pt}{1.0pt}{20.33957pt}{0.55229pt}{20.33957pt}{0.0pt}\pgfsys@curveto{20.33957pt}{-0.55229pt}{20.78728pt}{-1.0pt}{21.33957pt}{-1.0pt}\pgfsys@curveto{21.89186pt}{-1.0pt}{22.33957pt}{-0.55229pt}{22.33957pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{35.56595pt}{0.0pt}\pgfsys@moveto{36.56595pt}{0.0pt}\pgfsys@curveto{36.56595pt}{0.55229pt}{36.11824pt}{1.0pt}{35.56595pt}{1.0pt}\pgfsys@curveto{35.01366pt}{1.0pt}{34.56595pt}{0.55229pt}{34.56595pt}{0.0pt}\pgfsys@curveto{34.56595pt}{-0.55229pt}{35.01366pt}{-1.0pt}{35.56595pt}{-1.0pt}\pgfsys@curveto{36.11824pt}{-1.0pt}{36.56595pt}{-0.55229pt}{36.56595pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{35.56595pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{42.67914pt}{0.0pt}\pgfsys@moveto{43.67914pt}{0.0pt}\pgfsys@curveto{43.67914pt}{0.55229pt}{43.23143pt}{1.0pt}{42.67914pt}{1.0pt}\pgfsys@curveto{42.12685pt}{1.0pt}{41.67914pt}{0.55229pt}{41.67914pt}{0.0pt}\pgfsys@curveto{41.67914pt}{-0.55229pt}{42.12685pt}{-1.0pt}{42.67914pt}{-1.0pt}\pgfsys@curveto{43.23143pt}{-1.0pt}{43.67914pt}{-0.55229pt}{43.67914pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{42.67914pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{49.79233pt}{0.0pt}\pgfsys@moveto{50.79233pt}{0.0pt}\pgfsys@curveto{50.79233pt}{0.55229pt}{50.34462pt}{1.0pt}{49.79233pt}{1.0pt}\pgfsys@curveto{49.24004pt}{1.0pt}{48.79233pt}{0.55229pt}{48.79233pt}{0.0pt}\pgfsys@curveto{48.79233pt}{-0.55229pt}{49.24004pt}{-1.0pt}{49.79233pt}{-1.0pt}\pgfsys@curveto{50.34462pt}{-1.0pt}{50.79233pt}{-0.55229pt}{50.79233pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{49.79233pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{0.0pt}{0.0pt}\pgfsys@lineto{7.11319pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@lineto{14.22638pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@lineto{21.33957pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@lineto{35.56595pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{35.56595pt}{0.0pt}\pgfsys@lineto{42.67914pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{42.67914pt}{0.0pt}\pgfsys@lineto{49.79233pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{49.79233pt}{0.0pt}\pgfsys@lineto{56.90552pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}{}{}{{}}{{{{}{}{}{}}}{{}{}{}{}}}}{} {} {} {}
{}{}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@curveto{7.11319pt}{11.78566pt}{16.6671pt}{21.33957pt}{28.45276pt}{21.33957pt}\pgfsys@curveto{40.23842pt}{21.33957pt}{49.79233pt}{11.78566pt}{49.79233pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}{}{}{{}}{{{{}{}{}{}}}{{}{}{}{}}}}{} {} {} {}
{}{}{}\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@curveto{14.22638pt}{7.8571pt}{20.59566pt}{14.22638pt}{28.45276pt}{14.22638pt}\pgfsys@curveto{36.30986pt}{14.22638pt}{42.67914pt}{7.8571pt}{42.67914pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}{}{}{{}}{{{{}{}{}{}}}{{}{}{}{}}}}{} {} {} {}
{}{}{}\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@curveto{21.33957pt}{3.92854pt}{24.52422pt}{7.11319pt}{28.45276pt}{7.11319pt}\pgfsys@curveto{32.3813pt}{7.11319pt}{35.56595pt}{3.92854pt}{35.56595pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope{{
{}{}{}{}{}}{{{}}{{}}}}{}{}\hss}\pgfsys@discardpath\pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope\hss}}\lxSVG@closescope\endpgfpicture}}+\frac{1}{4}\leavevmode\hbox
to64.42pt{\vbox to50.19pt{\pgfpicture\makeatletter\hbox{\hskip
0.2pt\lower-25.09616pt\hbox to0.0pt{\pgfsys@beginscope\pgfsys@invoke{
}\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{
}\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{
}\pgfsys@setlinewidth{0.4pt}\pgfsys@invoke{ }\nullfont\hbox
to0.0pt{\pgfsys@beginscope\pgfsys@invoke{ }
{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{{}}{}{{}{}{}{{}}{{{{}{}{}{}}} {{}{}{}{}}}}{} {} {} {}
{}{}\pgfsys@beginscope\pgfsys@invoke{
}\color[rgb]{1,1,1}\definecolor[named]{pgfstrokecolor}{rgb}{1,1,1}\pgfsys@color@gray@stroke{1}\pgfsys@invoke{
}\pgfsys@color@gray@fill{1}\pgfsys@invoke{
}\definecolor[named]{pgffillcolor}{rgb}{1,1,1}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@curveto{7.11319pt}{-13.74994pt}{18.25941pt}{-24.89616pt}{32.00935pt}{-24.89616pt}\pgfsys@curveto{45.7593pt}{-24.89616pt}{56.90552pt}{-13.74994pt}{56.90552pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} \pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@moveto{8.11319pt}{0.0pt}\pgfsys@curveto{8.11319pt}{0.55229pt}{7.66548pt}{1.0pt}{7.11319pt}{1.0pt}\pgfsys@curveto{6.5609pt}{1.0pt}{6.11319pt}{0.55229pt}{6.11319pt}{0.0pt}\pgfsys@curveto{6.11319pt}{-0.55229pt}{6.5609pt}{-1.0pt}{7.11319pt}{-1.0pt}\pgfsys@curveto{7.66548pt}{-1.0pt}{8.11319pt}{-0.55229pt}{8.11319pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@moveto{15.22638pt}{0.0pt}\pgfsys@curveto{15.22638pt}{0.55229pt}{14.77867pt}{1.0pt}{14.22638pt}{1.0pt}\pgfsys@curveto{13.67409pt}{1.0pt}{13.22638pt}{0.55229pt}{13.22638pt}{0.0pt}\pgfsys@curveto{13.22638pt}{-0.55229pt}{13.67409pt}{-1.0pt}{14.22638pt}{-1.0pt}\pgfsys@curveto{14.77867pt}{-1.0pt}{15.22638pt}{-0.55229pt}{15.22638pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{28.45276pt}{0.0pt}\pgfsys@moveto{29.45276pt}{0.0pt}\pgfsys@curveto{29.45276pt}{0.55229pt}{29.00505pt}{1.0pt}{28.45276pt}{1.0pt}\pgfsys@curveto{27.90047pt}{1.0pt}{27.45276pt}{0.55229pt}{27.45276pt}{0.0pt}\pgfsys@curveto{27.45276pt}{-0.55229pt}{27.90047pt}{-1.0pt}{28.45276pt}{-1.0pt}\pgfsys@curveto{29.00505pt}{-1.0pt}{29.45276pt}{-0.55229pt}{29.45276pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{28.45276pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{35.56595pt}{0.0pt}\pgfsys@moveto{36.56595pt}{0.0pt}\pgfsys@curveto{36.56595pt}{0.55229pt}{36.11824pt}{1.0pt}{35.56595pt}{1.0pt}\pgfsys@curveto{35.01366pt}{1.0pt}{34.56595pt}{0.55229pt}{34.56595pt}{0.0pt}\pgfsys@curveto{34.56595pt}{-0.55229pt}{35.01366pt}{-1.0pt}{35.56595pt}{-1.0pt}\pgfsys@curveto{36.11824pt}{-1.0pt}{36.56595pt}{-0.55229pt}{36.56595pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{35.56595pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{49.79233pt}{0.0pt}\pgfsys@moveto{50.79233pt}{0.0pt}\pgfsys@curveto{50.79233pt}{0.55229pt}{50.34462pt}{1.0pt}{49.79233pt}{1.0pt}\pgfsys@curveto{49.24004pt}{1.0pt}{48.79233pt}{0.55229pt}{48.79233pt}{0.0pt}\pgfsys@curveto{48.79233pt}{-0.55229pt}{49.24004pt}{-1.0pt}{49.79233pt}{-1.0pt}\pgfsys@curveto{50.34462pt}{-1.0pt}{50.79233pt}{-0.55229pt}{50.79233pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{49.79233pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{56.90552pt}{0.0pt}\pgfsys@moveto{57.90552pt}{0.0pt}\pgfsys@curveto{57.90552pt}{0.55229pt}{57.45781pt}{1.0pt}{56.90552pt}{1.0pt}\pgfsys@curveto{56.35323pt}{1.0pt}{55.90552pt}{0.55229pt}{55.90552pt}{0.0pt}\pgfsys@curveto{55.90552pt}{-0.55229pt}{56.35323pt}{-1.0pt}{56.90552pt}{-1.0pt}\pgfsys@curveto{57.45781pt}{-1.0pt}{57.90552pt}{-0.55229pt}{57.90552pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{56.90552pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{0.0pt}{0.0pt}\pgfsys@lineto{7.11319pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@lineto{14.22638pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@lineto{28.45276pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{28.45276pt}{0.0pt}\pgfsys@lineto{35.56595pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{35.56595pt}{0.0pt}\pgfsys@lineto{49.79233pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{49.79233pt}{0.0pt}\pgfsys@lineto{56.90552pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{56.90552pt}{0.0pt}\pgfsys@lineto{64.0187pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}{}{}{{}}{{{{}{}{}{}}}{{}{}{}{}}}}{} {} {} {}
{}{}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@curveto{7.11319pt}{13.74994pt}{18.25941pt}{24.89616pt}{32.00935pt}{24.89616pt}\pgfsys@curveto{45.7593pt}{24.89616pt}{56.90552pt}{13.74994pt}{56.90552pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}{}{}{{}}{{{{}{}{}{}}}{{}{}{}{}}}}{} {} {} {}
{}{}{}\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@curveto{14.22638pt}{3.92854pt}{17.41103pt}{7.11319pt}{21.33957pt}{7.11319pt}\pgfsys@curveto{25.26811pt}{7.11319pt}{28.45276pt}{3.92854pt}{28.45276pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}{}{}{{}}{{{{}{}{}{}}}{{}{}{}{}}}}{} {} {} {}
{}{}{}\pgfsys@moveto{35.56595pt}{0.0pt}\pgfsys@curveto{35.56595pt}{3.92854pt}{38.7506pt}{7.11319pt}{42.67914pt}{7.11319pt}\pgfsys@curveto{46.60768pt}{7.11319pt}{49.79233pt}{3.92854pt}{49.79233pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope{{
{}{}{}{}{}}{{{}}{{}}}}{}{}\hss}\pgfsys@discardpath\pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope\hss}}\lxSVG@closescope\endpgfpicture}}+\frac{1}{2}\leavevmode\hbox
to71.53pt{\vbox to57.31pt{\pgfpicture\makeatletter\hbox{\hskip
0.2pt\lower-28.65276pt\hbox to0.0pt{\pgfsys@beginscope\pgfsys@invoke{
}\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{
}\pgfsys@color@rgb@fill{0}{0}{0}\pgfsys@invoke{
}\pgfsys@setlinewidth{0.4pt}\pgfsys@invoke{ }\nullfont\hbox
to0.0pt{\pgfsys@beginscope\pgfsys@invoke{ }
{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{}{{}}{{}}\hbox{\hbox{{\pgfsys@beginscope\pgfsys@invoke{ }{{}{{}}{}{}}
\pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope}}}
{{}}{}{{}}{}{{}{}{}{{}}{{{{}{}{}{}}} {{}{}{}{}}}}{} {} {} {}
{}{}\pgfsys@beginscope\pgfsys@invoke{
}\color[rgb]{1,1,1}\definecolor[named]{pgfstrokecolor}{rgb}{1,1,1}\pgfsys@color@gray@stroke{1}\pgfsys@invoke{
}\pgfsys@color@gray@fill{1}\pgfsys@invoke{
}\definecolor[named]{pgffillcolor}{rgb}{1,1,1}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@curveto{7.11319pt}{-15.71422pt}{19.85173pt}{-28.45276pt}{35.56595pt}{-28.45276pt}\pgfsys@curveto{51.28017pt}{-28.45276pt}{64.0187pt}{-15.71422pt}{64.0187pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} \pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@moveto{8.11319pt}{0.0pt}\pgfsys@curveto{8.11319pt}{0.55229pt}{7.66548pt}{1.0pt}{7.11319pt}{1.0pt}\pgfsys@curveto{6.5609pt}{1.0pt}{6.11319pt}{0.55229pt}{6.11319pt}{0.0pt}\pgfsys@curveto{6.11319pt}{-0.55229pt}{6.5609pt}{-1.0pt}{7.11319pt}{-1.0pt}\pgfsys@curveto{7.66548pt}{-1.0pt}{8.11319pt}{-0.55229pt}{8.11319pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@moveto{15.22638pt}{0.0pt}\pgfsys@curveto{15.22638pt}{0.55229pt}{14.77867pt}{1.0pt}{14.22638pt}{1.0pt}\pgfsys@curveto{13.67409pt}{1.0pt}{13.22638pt}{0.55229pt}{13.22638pt}{0.0pt}\pgfsys@curveto{13.22638pt}{-0.55229pt}{13.67409pt}{-1.0pt}{14.22638pt}{-1.0pt}\pgfsys@curveto{14.77867pt}{-1.0pt}{15.22638pt}{-0.55229pt}{15.22638pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@moveto{22.33957pt}{0.0pt}\pgfsys@curveto{22.33957pt}{0.55229pt}{21.89186pt}{1.0pt}{21.33957pt}{1.0pt}\pgfsys@curveto{20.78728pt}{1.0pt}{20.33957pt}{0.55229pt}{20.33957pt}{0.0pt}\pgfsys@curveto{20.33957pt}{-0.55229pt}{20.78728pt}{-1.0pt}{21.33957pt}{-1.0pt}\pgfsys@curveto{21.89186pt}{-1.0pt}{22.33957pt}{-0.55229pt}{22.33957pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{28.45276pt}{0.0pt}\pgfsys@moveto{29.45276pt}{0.0pt}\pgfsys@curveto{29.45276pt}{0.55229pt}{29.00505pt}{1.0pt}{28.45276pt}{1.0pt}\pgfsys@curveto{27.90047pt}{1.0pt}{27.45276pt}{0.55229pt}{27.45276pt}{0.0pt}\pgfsys@curveto{27.45276pt}{-0.55229pt}{27.90047pt}{-1.0pt}{28.45276pt}{-1.0pt}\pgfsys@curveto{29.00505pt}{-1.0pt}{29.45276pt}{-0.55229pt}{29.45276pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{28.45276pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{42.67914pt}{0.0pt}\pgfsys@moveto{43.67914pt}{0.0pt}\pgfsys@curveto{43.67914pt}{0.55229pt}{43.23143pt}{1.0pt}{42.67914pt}{1.0pt}\pgfsys@curveto{42.12685pt}{1.0pt}{41.67914pt}{0.55229pt}{41.67914pt}{0.0pt}\pgfsys@curveto{41.67914pt}{-0.55229pt}{42.12685pt}{-1.0pt}{42.67914pt}{-1.0pt}\pgfsys@curveto{43.23143pt}{-1.0pt}{43.67914pt}{-0.55229pt}{43.67914pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{42.67914pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{49.79233pt}{0.0pt}\pgfsys@moveto{50.79233pt}{0.0pt}\pgfsys@curveto{50.79233pt}{0.55229pt}{50.34462pt}{1.0pt}{49.79233pt}{1.0pt}\pgfsys@curveto{49.24004pt}{1.0pt}{48.79233pt}{0.55229pt}{48.79233pt}{0.0pt}\pgfsys@curveto{48.79233pt}{-0.55229pt}{49.24004pt}{-1.0pt}{49.79233pt}{-1.0pt}\pgfsys@curveto{50.34462pt}{-1.0pt}{50.79233pt}{-0.55229pt}{50.79233pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{49.79233pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{56.90552pt}{0.0pt}\pgfsys@moveto{57.90552pt}{0.0pt}\pgfsys@curveto{57.90552pt}{0.55229pt}{57.45781pt}{1.0pt}{56.90552pt}{1.0pt}\pgfsys@curveto{56.35323pt}{1.0pt}{55.90552pt}{0.55229pt}{55.90552pt}{0.0pt}\pgfsys@curveto{55.90552pt}{-0.55229pt}{56.35323pt}{-1.0pt}{56.90552pt}{-1.0pt}\pgfsys@curveto{57.45781pt}{-1.0pt}{57.90552pt}{-0.55229pt}{57.90552pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{56.90552pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
}
{{}}{}{{}}{}{{{}}{}{}{}{}{}{}{}{}}{}\pgfsys@moveto{64.0187pt}{0.0pt}\pgfsys@moveto{65.0187pt}{0.0pt}\pgfsys@curveto{65.0187pt}{0.55229pt}{64.571pt}{1.0pt}{64.0187pt}{1.0pt}\pgfsys@curveto{63.46642pt}{1.0pt}{63.0187pt}{0.55229pt}{63.0187pt}{0.0pt}\pgfsys@curveto{63.0187pt}{-0.55229pt}{63.46642pt}{-1.0pt}{64.0187pt}{-1.0pt}\pgfsys@curveto{64.571pt}{-1.0pt}{65.0187pt}{-0.55229pt}{65.0187pt}{0.0pt}\pgfsys@closepath\pgfsys@moveto{64.0187pt}{0.0pt}\pgfsys@fillstroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{0.0pt}{0.0pt}\pgfsys@lineto{7.11319pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@lineto{14.22638pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@lineto{21.33957pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@lineto{28.45276pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{28.45276pt}{0.0pt}\pgfsys@lineto{42.67914pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{42.67914pt}{0.0pt}\pgfsys@lineto{49.79233pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{49.79233pt}{0.0pt}\pgfsys@lineto{56.90552pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{56.90552pt}{0.0pt}\pgfsys@lineto{64.0187pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}}
{}{}{}\pgfsys@moveto{64.0187pt}{0.0pt}\pgfsys@lineto{71.1319pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}{}{}{{}}{{{{}{}{}{}}}{{}{}{}{}}}}{} {} {} {}
{}{}{}\pgfsys@moveto{7.11319pt}{0.0pt}\pgfsys@curveto{7.11319pt}{15.71422pt}{19.85173pt}{28.45276pt}{35.56595pt}{28.45276pt}\pgfsys@curveto{51.28017pt}{28.45276pt}{64.0187pt}{15.71422pt}{64.0187pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}{}{}{{}}{{{{}{}{}{}}}{{}{}{}{}}}}{} {} {} {}
{}{}{}\pgfsys@moveto{14.22638pt}{0.0pt}\pgfsys@curveto{14.22638pt}{11.78566pt}{23.78029pt}{21.33957pt}{35.56595pt}{21.33957pt}\pgfsys@curveto{47.35161pt}{21.33957pt}{56.90552pt}{11.78566pt}{56.90552pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}{}{}{{}}{{{{}{}{}{}}}{{}{}{}{}}}}{} {} {} {}
{}{}{}\pgfsys@moveto{21.33957pt}{0.0pt}\pgfsys@curveto{21.33957pt}{7.8571pt}{27.70885pt}{14.22638pt}{35.56595pt}{14.22638pt}\pgfsys@curveto{43.42305pt}{14.22638pt}{49.79233pt}{7.8571pt}{49.79233pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} {{}}{}{{}}{}{{}{}{}{{}}{{{{}{}{}{}}}{{}{}{}{}}}}{} {} {} {}
{}{}{}\pgfsys@moveto{28.45276pt}{0.0pt}\pgfsys@curveto{28.45276pt}{3.92854pt}{31.6374pt}{7.11319pt}{35.56595pt}{7.11319pt}\pgfsys@curveto{39.49449pt}{7.11319pt}{42.67914pt}{3.92854pt}{42.67914pt}{0.0pt}\pgfsys@stroke\pgfsys@invoke{
} \pgfsys@invoke{\lxSVG@closescope }\pgfsys@endscope{{
{}{}{}{}{}}{{{}}{{}}}}{}{}\hss}\pgfsys@discardpath\pgfsys@invoke{\lxSVG@closescope
}\pgfsys@endscope\hss}}\lxSVG@closescope\endpgfpicture}}+\vspace{-3ex}$
${}+\frac{1}{8}\leavevmode\hbox to85.76pt{\vbox
to71.53pt{\pgfpicture\makeatletter\hbox{\hskip 0.2pt\lower-35.76595pt\hbox
to0.0pt{\pgfsys@beginscope\pgfsys@invoke{
}\definecolor{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@rgb@stroke{0}{0}{0}\pgfsys@invoke{
|
# Effects of backreaction and exponential nonlinear electrodynamics on the
holographic superconductors
A<EMAIL_ADDRESS>F. Shaker 1 1 Physics Department and
Biruni Observatory, College of Sciences, Shiraz University, Shiraz 71454, Iran
2 Research Institute for Astronomy and Astrophysics of Maragha (RIAAM), P.O.
Box 55134-441, Maragha, Iran
###### Abstract
We analytically study the properties of a $(2+1)$-dimensional $s$-wave
holographic superconductor in the presence of exponential nonlinear
electrodynamics. We consider the case in which the scalar and gauge fields
back react on the background metric. Employing the analytical Sturm-Liouville
method, we find that in the black hole background, the nonlinear
electrodynamics correction will affect the properties of the holographic
superconductors. We find that with increasing both backreaction and nonlinear
parameters, the scalar hair condensation on the boundary will develop more
difficult. We obtain the relation connecting the critical temperature with the
charge density. Our analytical results support that, even in the presence of
the nonlinear electrodynamics and backreaction, the phase transition for the
holographic superconductor still belongs to the second order and the critical
exponent of the system always takes the mean-field value $1/2$.
## I Introduction
The AdS/CFT correspondence is an equivalence between a conformal field theory
(CFT) in $d$ spacetime dimensions, and a theory of gravity in
$(d+1)$-dimensional anti-de Sitter (AdS) spacetime AdS.CFT1 ; AdS.CFT2 ;
AdS.CFT3 . The $d$-dimensional theory does not have a gravitational force, and
is to be viewed as a hologram of the $(d+1)$-dimensional theory. The AdS/CFT
correspondence is a well-known approach to explore strongly coupled field
theories in which certain questions become computationally smooth and
conceptually more explicit. The AdS/CFT correspondence can be applied to
condensed matter phenomena. In condensed matter physics, there are many
strongly coupled systems such as superconductors. In this regards, it was
recently argued that it is quite possible to shed some light on the problem of
understanding the mechanism of the high temperature superconductors in
condensed matter physics, by studying a classical general relativity in one
higher dimensional spacetime Har1 ; Har2 . The holographic superconductivity
is a phenomenon associated with asymptotic AdS black holes. The studies on the
holographic superconductors have received a lot of attentions HS.HM ; HS.H ;
HS.FGR ; Wang1 ; Wang2 .
Most studies on the holographic superconductors are focused on the cases where
the gauge field is in the form of the linear Maxwell field. But nonlinear
electrodynamics is constructed by the desire to find non-singular field
theories. One may consider nonlinear electrodynamics as a possible mechanism
for avoiding the singularity of the point-like charged particle at the origin.
The nonlinear extension of the original Maxwell electrodynamics in the context
of holographic superconductors have arisen intensive investigations b.Nu.GB.BI
; p.BI&LN.4D ; p.BI.4D ; JC.p.Nu.BI.4D ; P.BGRL ; P.JPC ; b.BI.5D ; Shey2 . In
particular, in order to see what difference will appear for holographic
superconductor in the presence of Born-Infeld (BI) nonlinear electrodynamics,
compared with the case of linear Maxwell electrodynamics, the authors of Ref.
LPJW2015.p.BI have studied condensation and critical phenomena of the
holographic superconductors with BI electrodynamics in $d$-dimensional
spacetime. Their analytical results indicate that the nonlinear BI
electrodynamics decreases the critical temperature of the holographic
superconductor. It was observed that the higher BI corrections make it harder
for the condensation to form but do not affect the critical phenomena of the
system LPJW2015.p.BI .
It is also interesting to investigate the effects of gauge and scalar fields
of the holographic superconductor on the background geometry. Although if we
ignore this backreaction, the problem is simplified, but retains most of the
interesting physics since the nonlinear interactions are retained. Indeed,
considering the holographic superconductor model away from the probe limit may
bring rich physics. Therefore, many authors have tried to study the
holographic superconductors away from the probe limit PJWC2012 ; B1 ; B2 ; B3
; B4 ; B5 ; B6 . Employing the analytical Sturm-Liouville method, the effects
of both backreaction and BI nonlinear parameter on the critical temperature as
well as scalar condensation were explored in Ref. BIBR . Furthermore, the
relation between the critical temperature and charge density was established
BIBR . It was shown that it is more difficult to have scalar condensation in
BI electrodynamics when the backreaction is taken into account BIBR .
In the present work we would like to extend our analytical study on the
backreacting holographic superconductors by considering another form of the
higher order corrections to the gauge field, i.e., the exponential form of
nonlinear electrodynamics. It was shown that when the backreaction is taken
into account, even the uncharged scalar field can form a condensation in the
$(2+1)$-dimensional holographic superconductor model Har2 . Numerical studies
on the holographic superconductors with exponential nonlinear (EN)
electrodynamics are carried out in the probe limit ZPCJ.P.N.N . It was shown
that the higher nonlinear electrodynamics corrections makes the condensation
harder to form ZPCJ.P.N.N . As far as we know, analytical study on the
holographic superconductor in the presence of EN electrodynamics and away from
the probe limit has not been done. Considering exponential form of the higher
corrections to the gauge field, we shall analytically investigate the
properties of the holographic superconductors when the gauge and scalar field
do back react on the metric background. We shall use the analytical Sturm-
Liouville eigenvalue problem. We will also compare our results with those for
the holographic superconductors with BI nonlinear electrodynamics with
backreaction given in BIBR .
This paper is outlined as follows. In section II, we introduce the action and
basic field equations of the $(2+1)$-holographic superconductor with EN
electrodynamics with backreaction. In section III, we compute the critical
temperature in terms of the charge density and disclose its dependence on the
both nonlinear and backreaction parameters. Section IV, includes step-by-step
computations for obtaining the critical exponent and the condensation values
of the holographic superconductor and provides explanations about them.
Section V will help us to collect the obtained results briefly.
## II Basic Equations of Holographic Superconductors with Backreactions
The action of Einstein gravity coupled to a charged complex scalar field in
the presence of nonlinear electrodynamics is described by
$\displaystyle S=\int
d^{4}x\sqrt{-g}\left[\frac{1}{2\kappa^{2}}(R-2\Lambda)+\mathcal{L}(\mathcal{F})-|\nabla\psi-
iqA\psi|^{2}-m^{2}|\psi|^{2}\right],$ (1)
where $\kappa$ is the usual four dimensional gravitational constant,
$\kappa^{2}=8\pi G_{4}$, $\Lambda=-{3}/{L^{2}}$ is the cosmological constant,
where $L$ is the AdS radius which will be scaled unity in our calculations.
$R$ and $g$ are, respectively, representing the Ricci scalar and the
determinant of the metric. $A$ is the gauge field and $\psi$ represents a
scalar field with charge $q$ and mass $m$. $\mathcal{L}(\mathcal{F})$ is a
simple generalization of Maxwell Lagrangian in a exponential form Hendi
$\mathcal{L}(\mathcal{F})=\frac{1}{4b}\left(e^{-b\mathcal{F}}-1\right),$ (2)
where $b$ is the nonlinear parameter, $\mathcal{F}=F_{\mu\nu}F^{\mu\nu}$ and
$F^{\mu\nu}$ is the electromagnetic field tensor. Expanding this nonlinear
Lagrangian for small $b$, the leading order term is the linear Maxwell theory,
$\mathcal{L}(\mathcal{F})=-\mathcal{F}/{4}+\mathcal{O}(b)$. The plane-
symmetric black hole solution with an asymptotically AdS behavior including
the backreaction is described by the metric,
$\displaystyle ds^{2}$
$\displaystyle=-f(r)e^{-\chi(r)}dt^{2}+\frac{dr^{2}}{f(r)}+r^{2}(dx^{2}+dy^{2}).$
(3)
We adopt the following gauge choices for the vector field and the scalar
field,
$\displaystyle A_{\mu}=\left(\phi(r),0,0,0\right),\ \ \ \psi=\psi(r),$ (4)
with these functions being real-valued. Then, we need to establish the
Einstein equations by varying action (1) with respect to the metric. We find
$\displaystyle
R^{\mu\nu}-\frac{g^{\mu\nu}}{2}R-\frac{3}{L^{2}}g^{\mu\nu}=\kappa^{2}T^{\mu\nu},$
(5)
where the energy momentum tensor is given by
$\displaystyle T^{\mu\nu}$ $\displaystyle=$
$\displaystyle\frac{1}{4b}g^{\mu\nu}\left(e^{-b\mathcal{F}}-1\right)+e^{-b\mathcal{F}}F_{\sigma}^{\
\mu}F^{\sigma\nu}-m^{2}g^{\mu\nu}|\psi|^{2}-g^{\mu\nu}|\nabla\psi-
iqA\psi|^{2}$ (6)
$\displaystyle+\left[(\nabla^{\nu}+iqA^{\nu})\psi^{*}(\nabla^{\mu}-iqA^{\mu})\psi+\mu\leftrightarrow\nu\right].$
Variation with respect to the scalar field yields
$\displaystyle(\nabla_{\mu}-iqA_{\mu})(\nabla^{\mu}-iqA^{\mu})\psi-m^{2}\psi=0,$
(7)
while the electrodynamic equation,
$\displaystyle\nabla_{\mu}\left(F^{\mu\nu}e^{-b\mathcal{F}}\right)=iq\Bigg{[}\psi^{*}(\nabla^{\mu}-iqA^{\mu})\psi-\psi(\nabla^{\mu}+iqA^{\mu})\psi^{*}\Bigg{]},$
(8)
is obtained by varying action (1) with respect to the gauge field. These
equations can easily reduce to those of the holographic superconductor in
Maxwell theory Har2 , provided $b\rightarrow 0$. Calculations of the Einstein,
scalar and electrodynamic field equations, with respect to the metric (3),
yield the following expressions,
$\displaystyle\chi^{\prime}$ $\displaystyle+2r\kappa^{2}\left(\psi^{\prime
2}+\frac{q^{2}e^{\chi}\phi^{2}\psi^{2}}{f^{2}}\right)=0,$ (9) $\displaystyle
f^{\prime}-\left(\frac{3r}{L^{2}}-\frac{f}{r}\right)-\frac{\chi^{\prime}}{2}f+r\kappa^{2}\left[m^{2}\psi^{2}+\frac{1}{4b}(1-e^{2b\phi^{\prime
2}e^{\chi}})+\phi^{\prime 2}e^{\chi+2b\phi^{\prime 2}e^{\chi}}\right]=0,$ (10)
$\displaystyle\phi^{\prime\prime}(1+4be^{\chi}\phi^{\prime
2})+\frac{2}{r}\phi^{\prime}(1+rbe^{\chi}\chi^{\prime}\phi^{\prime
2})+\frac{\chi^{\prime}\phi^{\prime}}{2}-\frac{2q^{2}\phi\psi^{2}}{f}e^{-2be^{\chi}\phi^{\prime
2}}=0,$ (11)
$\displaystyle\psi^{\prime\prime}+\left(\frac{f^{\prime}}{f}+\frac{2}{r}-\frac{\chi^{\prime}}{2}\right)\psi^{\prime}+\left(\frac{q^{2}e^{\chi}\phi^{2}}{f^{2}}-\frac{m^{2}}{f}\right)\psi=0,$
(12)
where the prime denotes derivative with respect to $r$. We further assume
there exists an event horizon $r_{+}$ for which $f(r_{+})=0$, and thus the
corresponding Hawking temperature of the black hole reads
$\displaystyle T=\frac{f^{\prime}(r_{+})e^{-\chi(r_{+})/2}}{4\pi}.$ (13)
For the case with $b\rightarrow 0$, Eqs. (9)-(12) coincide with their
corresponding equations presented in PJWC2012 . Also in the probe limit where
$\kappa=0$, Eqs. (11) and (12) go back to the $(2+1)$-dimensional holographic
superconductor model studied in ZPCJ.P.N.N . In this case the solution of Eq.
(10) is
$\displaystyle
f(r)=\frac{r^{2}}{L^{2}}\left(1-\frac{r_{+}^{3}}{r^{3}}\right).$ (14)
It should be noted that we can set the charge parameter, $q$, as unity and
keep $\kappa^{2}$ finite when the backreaction is taken into account by
adopting the scaling symmetry ssym . When the Hawking temperature is above the
critical temperature $T>T_{c}$, the system leads to the well-known exact black
holes as $b\rightarrow 0$ with the metric coefficient and the potential
function given by
$\displaystyle
f(r)=\frac{r^{2}}{L^{2}}-\frac{1}{r}\left(\frac{r_{+}^{3}}{L^{2}}+\frac{\kappa^{2}\rho^{2}}{2r_{+}}\right)+\frac{\kappa^{2}\rho^{2}}{2r^{2}},\
\ \ \phi\approx\mu-\frac{\rho}{r}.$ (15)
On the dual side, $\mu$ and $\rho$ are, respectively, the chemical potential
and charge density of the holographic superconductor. When $\kappa=0$, the
metric coefficient $f(r)$ recovers the case of Schwarzschild AdS black holes
(14). For investigating the properties of dual model in superconducting phase,
i.e., $\psi(r)\neq 0$, we need the suitable boundary conditions. Examining the
behavior of the fields near the horizon, we find the suitable boundary
conditions as
$\displaystyle\phi(r_{+})=0,\ \ \
\psi(r_{+})=\frac{f^{\prime}(r_{+})\psi^{\prime}(r_{+})}{m^{2}},$ (16)
and hence the metric functions $\chi$ and $f(r)$ satisfy
$\displaystyle\chi^{\prime}(r_{+})$
$\displaystyle=-2r_{+}\kappa^{2}\left(\psi^{\prime
2}(r_{+})+\frac{q^{2}e^{\chi(r_{+})}\phi^{\prime
2}(r_{+})\psi^{2}(r_{+})}{f^{\prime 2}(r_{+})}\right),$ (17) $\displaystyle
f^{\prime}(r_{+})=\frac{3r_{+}}{L^{2}}-r_{+}\kappa^{2}\left[m^{2}\psi^{2}(r_{+})+\frac{1}{4b}(1-e^{2b\phi^{\prime
2}(r_{+})e^{\chi(r_{+})}})+\phi^{\prime 2}(r_{+})e^{\chi(r_{+})+2b\phi^{\prime
2}(r_{+})e^{\chi(r_{+})}}\right].$ (18)
The asymptotic behavior of the fields, corresponding to the solution of Eqs.
(11) and (12) in the limit $r\rightarrow\infty$, are given by
$\displaystyle\phi\approx\mu-\frac{\rho}{r},$ (19)
$\displaystyle\psi\approx\frac{\psi_{-}}{r^{\Delta_{-}}}+\frac{\psi_{+}}{r^{\Delta_{+}}},$
(20)
where
$\displaystyle\Delta_{\pm}=\frac{3}{2}\pm\frac{\sqrt{9+4m^{2}}}{2},$ (21)
is the conformal dimension of the dual operator $\mathcal{O_{\pm}}$ in the
boundary field theory. Here $\psi_{+}$ and $\psi_{-}$ can be considered as the
source and the vacuum expectation values of the dual operator. Hereafter, we
set $\psi_{+}=0$ and investigate the condensation of
$\psi_{-}=<\mathcal{O_{-}}>$, analytically. In what follows we choose the
scalar to have $m^{2}=-2$, and hence the corresponding dual operator has mass
dimension $\Delta_{-}=1$.
## III Analytical study and critical tempreture
In this section, we investigate the analytical properties of a
$(2+1)$-holographic superconductor in the framework of EN electrodynamics. We
study the problem by taking the backreaction into account. We find the
critical temperature $T_{c}$ via the Sturm-Liouville variational approach.
Further, we obtain a relationship between the critical temperature and the
charge density and investigate the effects of both backreaction and EN
parameter on the critical temperature. In order to get the solutions in
superconducting phase, we can define a new variable $z={r_{+}}/{r}$. Then, the
equations of motion can be rewritten as
$\displaystyle\chi^{\prime}-2\kappa^{2}\left(z{\psi^{\prime}}^{2}+\frac{r_{+}^{2}}{z^{3}f^{2}}e^{\chi}\phi^{2}\psi^{2}\right)=0,$
(22) $\displaystyle
f^{\prime}-\frac{f}{z}+\frac{3r_{+}^{2}}{z^{3}}-\frac{\chi^{\prime}f}{2}-\frac{\kappa^{2}r_{+}^{2}}{z^{3}}\left[m^{2}\psi^{2}+\frac{1}{4b}(1-e^{\frac{2bz^{4}}{r_{+}^{2}}\phi^{\prime
2}e^{\chi}})+\frac{z^{4}}{r_{+}^{2}}\phi^{\prime
2}e^{\chi+\frac{2bz^{4}}{r_{+}^{2}}\phi^{\prime 2}e^{\chi}}\right]=0,$ (23)
$\displaystyle\phi^{\prime\prime}\Bigg{(}1+\frac{4bz^{4}}{r_{+}^{2}}e^{\chi}\phi^{\prime
2}\Bigg{)}+\frac{8bz^{3}}{r^{2}_{+}}e^{\chi}\phi^{\prime
3}+\frac{2bz^{4}}{r_{+}^{2}}e^{\chi}\phi^{\prime
3}\chi^{\prime}+\frac{\phi^{\prime}\chi^{\prime}}{2}-\frac{2r_{+}^{2}\psi^{2}}{fz^{4}}e^{-\frac{2bz^{4}}{r^{2}_{+}}e^{\chi}\phi^{\prime
2}}\phi=0,$ (24)
$\displaystyle\psi^{\prime\prime}-\left(\frac{\chi^{\prime}}{2}-\frac{f^{\prime}}{f}\right)\psi^{\prime}-\frac{r_{+}^{2}}{z^{4}}\left(\frac{m^{2}}{f}-\frac{e^{\chi}\phi^{2}}{f^{2}}\right)\psi=0,$
(25)
where now the prime denotes derivative with respect to $z$. When $b\rightarrow
0$, the above equations restore the corresponding equations in Ref. PJWC2012 ,
while in the absence of the backreaction, Eqs. (24) and (25) reduce to their
corresponding equations in Ref. ZPCJ.P.N.N . Following the perturbation
scheme, since close to the critical point, the value of the scalar operator is
small, it can be introduced as an expansion parameter
$\displaystyle\epsilon\equiv<\mathcal{O}_{i}>,$ (26)
with $i=+$ or $i=-$. Besides, near the critical point the scalar and gauge
fields are small and therefore we can expand the gauge field $\phi$, the
scalar field $\psi$, and the metric functions $f(z)$, $\chi(z)$ as PJWC2012
$\displaystyle\psi=\epsilon\psi_{1}+\epsilon^{3}\psi_{3}+\epsilon^{5}\psi_{5}+...,$
(27)
$\displaystyle\phi=\phi_{0}+\epsilon^{2}\phi_{2}+\epsilon^{4}\phi_{4}+...,$
(28) $\displaystyle f=f_{0}+\epsilon^{2}f_{2}+\epsilon^{4}f_{4}+...,$ (29)
$\displaystyle\chi=\epsilon^{2}\chi_{2}+\epsilon^{4}\chi_{4}+...,$ (30)
where the metric function $f(z)$ and $\chi(z)$ are expanded around the
Reissner-Nordström AdS spacetime. Also, the chemical potential $\mu$ may be
expanded as PJWC2012
$\displaystyle\mu=\mu_{0}+\epsilon^{2}\delta\mu_{2}+...,$ (31)
where $\delta\mu_{2}$ is positive. Thus, near the phase transition, the order
parameter as a function of the chemical potential has the form
$\displaystyle\epsilon\thickapprox\Bigg{(}\frac{\mu-\mu_{0}}{\delta\mu_{2}}\Bigg{)}^{1/2}.$
(32)
whose critical exponent $\beta=1/2$ is the same as in the Ginzburg-Landau mean
field theory. The phase transition can take place when
$\mu\rightarrow\mu_{0}$. In this case the critical value of the chemical
potential is given by $\mu_{c}=\mu_{0}$.
From Eq. (24) the equation for $\phi$ is obtained at zeroth order as
$\displaystyle\phi^{\prime\prime}(z)\Bigg{(}1+4b\frac{z^{4}}{r_{+c}^{2}}\phi^{\prime
2}\Bigg{)}+\frac{8bz^{3}}{r^{2}_{+c}}\phi^{\prime 3}(z)=0,$ (33)
which admits the following solutions for the gauge field
$\displaystyle\phi(z)=\int_{1}^{z}{dz\frac{r_{+c}}{2z^{2}\sqrt{b}}\sqrt{L_{W}\left(\frac{4bz^{4}\beta^{2}}{r_{+c}^{2}}\right)}}.$
(34)
In the above expression $\beta$ is an integration constant and
$L_{W}(x)={LambertW(x)}$ is the Lambert function which satisfies Lambert
$L_{W}(x)e^{L_{W}(x)}=x,$ (35)
and has the following series expansion
$L_{W}(x)=x-x^{2}+\frac{3}{2}x^{3}-\frac{8}{3}x^{4}+....$ (36)
Obviously, the series (36) converges provided $|x|<1$. If we expand the
solution (34) for small $b$ and keep the only linear terms in $b$, we arrive
at
$\displaystyle\phi(z)=-\beta(1-z)+\frac{2\beta^{3}b}{5r_{+c}^{2}}\left(1-z^{5}\right)+\mathcal{O}(b^{2}).$
(37)
Differentiating Eqs. (19) and (37) with respect to $z$ and equating them at
$z=0$, we find $\beta=-{\rho}/{r_{+c}}$. Rearranging Eq. (37) and using the
relation $\beta$, we arrive at
$\displaystyle\phi_{0}(z)=\lambda
r_{+c}(1-z)\Bigg{\\{}1-\frac{2}{5}b\lambda^{2}(1+z+z^{2}+z^{3}+z^{4})\Bigg{\\}},\
\ \ b\lambda^{2}<1,$ (38)
where
$\displaystyle\lambda=\frac{\rho}{r^{2}_{+c}},$ (39)
and we have neglected $\mathcal{O}(b^{2})$. Thus, to zeroth order the equation
for $f$ is solved as
$\displaystyle
f_{0}(z)=r_{+}^{2}g(z)=r_{+}^{2}\left[\frac{1}{z^{2}}-z-\frac{\kappa^{2}\lambda^{2}}{2}z(1-z)+\frac{b}{10}\kappa^{2}\lambda^{4}z\left(1-z^{5}\right)\right].$
(40)
At the first order, the behavior of $\psi$ at the asymptotic AdS boundary is
given by
$\displaystyle\psi_{1}\approx\frac{\psi_{-}}{r_{+}^{\Delta_{-}}}z^{\Delta_{-}}+\frac{\psi_{+}}{r_{+}^{\Delta_{+}}}z^{\Delta_{+}}.$
(41)
Next, we introduce a variational trial function $F(z)$ near the boundary
$\displaystyle\psi_{1}(z)=\frac{<\mathcal{O}_{i}>}{\sqrt{2}r_{+}^{\triangle_{i}}}z^{\triangle_{i}}F(z),$
(42)
with the boundary condition $F(0)=1$ and $F^{\prime}(0)=0$. Then, we can
obtain the equation of motion for $F(z)$ by substituting (42) into Eq. (25).
We find
$\displaystyle
F^{\prime\prime}(z)+\Bigg{[}\frac{2\Delta}{z}+\frac{g^{\prime}}{g}\Bigg{]}F^{\prime}(z)+\Bigg{[}\frac{\Delta}{z}\Bigg{(}\frac{\Delta-1}{z}+\frac{g^{\prime}}{g}\Bigg{)}-\frac{m^{2}}{z^{4}g}\Bigg{]}F(z)$
$\displaystyle+\frac{\lambda^{2}(1-z)^{2}}{z^{4}g^{2}}\Bigg{[}1-\frac{4}{5}b\lambda^{2}\Bigg{(}1+z+z^{2}+z^{3}+z^{4}\Bigg{)}\Bigg{]}F(z)=0.$
(43)
Defining the new functions
$\displaystyle T(z)$ $\displaystyle=$ $\displaystyle
z^{2\Delta_{i}+1}\Bigg{[}2(z^{-3}-1)-\kappa^{2}\lambda^{2}(1-z)+\frac{b}{5}\kappa^{2}\lambda^{4}(1-z^{5})\Bigg{]},$
(44) $\displaystyle P(z)$ $\displaystyle=$
$\displaystyle\frac{\Delta_{i}}{z}\Bigg{(}\frac{\Delta_{i}-1}{z}+\frac{g^{\prime}}{g}\Bigg{)}-\frac{m^{2}}{z^{4}g},$
(45) $\displaystyle Q(z)$ $\displaystyle=$
$\displaystyle\frac{(1-z)^{2}}{z^{4}g^{2}}\Bigg{[}1-\frac{4}{5}b\lambda^{2}\Bigg{(}1+z+z^{2}+z^{3}+z^{4}\Bigg{)}\Bigg{]}.$
(46)
we can rewrite Eq. (III) as
$\displaystyle TF^{\prime\prime}+T^{\prime}F^{\prime}+PF+\lambda^{2}QF=0.$
(47)
According to the Sturm-Liouville eigenvalue problem Gan , the eigenvalue
$\lambda^{2}$ can be obtained by minimizing the expression
$\displaystyle\lambda^{2}=\frac{\int_{0}^{1}T\left(F^{\prime
2}-PF^{2}\right)dz}{\int_{0}^{1}TQF^{2}dz},$ (48)
where we have chosen the trial function in the form $F(z)=1-\alpha z^{2}$. In
order to simplify our calculations, we express the backreaction parameter as
PJWC2012
$\displaystyle\kappa_{n}=n\Delta\kappa,\ \ \ n=0,1,2,...$ (49)
where $\Delta\kappa=\kappa_{n+1}-\kappa_{n}$ is the step size of our iterative
procedure. The main purpose is to work in the small backreaction approximation
so that all the functions can be expanded by $\kappa^{2}$ and the $\kappa^{4}$
term can be neglected. Furthermore, we retain the terms that are linear in
nonlinear parameter $b$ and keep terms upto $\mathcal{O}(b)$. So we use the
following relations
$\displaystyle\kappa^{2}\lambda^{2}=\kappa_{n}^{2}\lambda^{2}=\kappa_{n}^{2}(\lambda^{2}|_{\kappa_{n-1}})+\mathcal{O}\left[(\Delta\kappa)^{4}\right],$
(50) $\displaystyle
b\lambda^{2}=b\left(\lambda^{2}|_{b=0}\right)+\mathcal{O}(b^{2}),$ (51)
and
$\displaystyle
b\kappa^{2}\lambda^{4}=b\kappa_{n}^{2}(\lambda^{4}|_{\kappa_{n-1},b=0})+\mathcal{O}(b^{2})+\mathcal{O}[(\Delta\kappa)^{4}],$
(52)
where we have assumed $\kappa_{-1}=0$, $\lambda^{2}|_{\kappa_{-1}}=0$ and
$\lambda^{2}|_{b=0}$ is the value of $\lambda^{2}$ for $b=0$. Now we are going
to compute the critical temperature $T_{c}$. First of all, we start with the
following equation
$\displaystyle T_{c}=\frac{f^{\prime}(r_{+c})}{4\pi}.$ (53)
From Eq. (18), $f^{\prime}(r_{+c})$ is expressed as
$\displaystyle
f^{\prime}(r_{+c})=3r_{+c}-\kappa^{2}r_{+c}\left[\frac{{{\phi_{0}}^{\prime}}^{2}(r_{+c})}{2}+\frac{3}{2}b{{\phi_{0}}^{\prime}}^{4}(r_{+c})\right].$
(54)
Substituting Eq. (38) in the above equation, and then inserting the result
back into Eq. (53), we arrive at the following expression for the critical
temperature,
$\displaystyle
T_{c}=\frac{1}{4\pi}\sqrt{\frac{\rho}{\lambda}}\Bigg{[}3-\frac{\kappa_{n}^{2}(\lambda^{2}|_{\kappa_{n-1}})}{2}+\frac{1}{2}b\kappa_{n}^{2}(\lambda^{4}|_{\kappa_{n-1},b=0})\Bigg{]}.$
(55)
With these obtained computations out of the analytical approach at hand, we
are in a position to present the results of the critical temperature $T_{c}$
for a $(2+1)$-dimensional holographic superconductors in the presence of both
EN electrodynamics as well as backreaction. To do this, we assume the
nonlinear parameter $b$ is small, by choosing it as $b=0,0.1,0.2,0.3$. We also
get the values $m^{2}=-2$, $\Delta_{i}=\Delta_{-}=1$ and $\Delta\kappa=0.05$.
As an example, we bring the details of our calculations for the case of $n=4$
and summarize all results in table $1$.
For $b=0$, we obtain $\lambda^{2}$ From Eq. (48) as
$\displaystyle\lambda^{2}=\frac{-8.279205\alpha^{2}+4.924220\alpha-4.957900}{-3.579048+0.878281\alpha-0.153258\alpha^{2}}.$
(56)
From it we get the minimum eigenvalues of $\lambda^{2}$ and the corresponding
value of $\alpha$, as $\lambda^{2}_{\rm min}=1.2593$ at $\alpha=0.2361$. And
thus the critical temperature is obtained from Eq. (55) as
$T_{c}=0.2235\sqrt{\rho}$, which is in good agreement with the result of
PJWC2012 .
For $b=0.1$, we find
$\displaystyle\lambda^{2}=\frac{-4.95286+4.91624\alpha-8.27411\alpha^{2}}{-3.035+0.673477\alpha-0.109325\alpha^{2}},$
(57)
which has a minimum value $\lambda^{2}_{\rm min}=1.4757$ at $\alpha=0.2417$,
and we can get the critical temperature $T_{c}=0.2147\sqrt{\rho}$.
For $b=0.2$, we arrive at
$\displaystyle\lambda^{2}=\frac{-4.94387+4.90127\alpha-8.26415\alpha^{2}}{-2.49088+0.468581\alpha-0.0652154\alpha^{2}},$
(58)
whose minimum is $\lambda^{2}_{\rm min}=1.7811$ at $\alpha=0.24955$ and the
critical temperature becomes $T_{c}=0.2046\sqrt{\rho}$.
For $b=0.3$, we have
$\displaystyle\lambda^{2}=\frac{-4.93051+4.87819\alpha-8.24877\alpha^{2}}{-1.94661+0.263502\alpha-0.021046\alpha^{2}},$
(59)
which attains its minimum $\lambda^{2}_{\rm min}=2.2451$ at $\alpha=0.26133$
and the critical temperature reads $T_{c}=0.1927\sqrt{\rho}$. We summarize our
results for the critical temperature in cases of different values of nonlinear
and backreaction parameters In table $1$. From this table, we see that, for
fixed value of the backreaction parameter, with the nonlinear parameter $b$
getting stronger, the critical temperature decreases. Similarly, for a fixed
value of the nonlinear parameter $b$, the critical temperature drops as the
backreaction parameter increases. Thus, we conclude that the critical
temperature becomes smaller and so, make the condensation harder when we
increase the values of both backreaction and nonlinear parameters. These
features were also observed in the study a $(2+1)$-dimensional holographic
superconductors with backreaction when the gauge field is in the form of BI
nonlinear electrodynamics BIBR . Comparing the results obtained here with
those of BIBR , we observe that the effect of the EN corrections on the
condensation with respect to the BI nonlinear one is stronger when the
backreactions is taken into account in both cases. In other words, the
formation of scalar hair in the presence of EN electrodynamics is harder
compared to the case of BI nonlinear electrodynamics. Obviously, our analytic
results back up the findings in other articles. In the case of $b=\kappa=0$,
we observe that the analytic results for the critical temperature are
consistent with both the analytical results of Ref. P.ZGJZ as well as the
numerical result of Ref. Har2 . Also, we confirm the numerical result found in
Ref. ZPCJ.P.N.N when the backreaction parameter $\kappa$ is equal to zero. On
the other hand, for $b=0$ the data obtained for the critical temperature, is
analogous to those reported for the holographic superconductors with
backreaction in Maxwell theory PJWC2012 .
n | b=0 | | b=0.1 | | b=0.2 | | b=0.3 |
---|---|---|---|---|---|---|---|---
| BI | EN | BI | EN | BI | EN | BI | EN
0 | 0.2250 | 0.2250 | 0.2228 | 0.2161 | 0.2206 | 0.2060 | 0.2184 | 0.1942
1 | 0.2249 | 0.2249 | 0.2227 | 0.2160 | 0.2204 | 0.2059 | 0.2181 | 0.1941
2 | 0.2246 | 0.2246 | 0.2225 | 0.2158 | 0.2203 | 0.2057 | 0.2180 | 0.1938
3 | 0.2241 | 0.2241 | 0.2220 | 0.2153 | 0.2199 | 0.2050 | 0.2176 | 0.1934
4 | 0.2235 | 0.2235 | 0.2214 | 0.2147 | 0.2192 | 0.2046 | 0.2170 | 0.1927
5 | 0.2226 | 0.2226 | 0.2208 | 0.2141 | 0.2184 | 0.2038 | 0.2162 | 0.1919
6 | 0.2216 | 0.2216 | 0.2196 | 0.2130 | 0.2174 | 0.2029 | 0.2152 | 0.1909
Table $1$: The critical temperature $T_{c}/\sqrt{\rho}$ for holographic
superconductors in the presence of BI and EN electrodynamics. Here we have
taken $\kappa_{n}=n\Delta\kappa$ where $\Delta\kappa=0.05$. The results for BI
case are invoked from Ref. BIBR .
## IV CRITICAL EXPONENT AND THE CONDENSATION OF THE SCALAR OPERATOR
We use the Sturm-Liouville method to analytically examine the scalar
condensation and the order of the phase transition with backreactions near the
critical temperature. With the help of Eq. (42), when $T$ is close to $T_{c}$,
the equation of motion (24) can be rewritten as
$\displaystyle\phi^{\prime\prime}\left(1+\frac{4bz^{4}}{r_{+}^{2}}\phi^{\prime
2}\right)+\frac{8bz^{3}}{r_{+}^{2}}{\phi^{\prime}}^{3}=\frac{\langle\mathcal{O}\rangle^{2}}{r_{+}^{2}}\mathcal{B}(z)\phi(z),$
(60)
$\displaystyle\mathcal{B}(z)=\frac{F^{2}(z)}{1-z^{3}}\left(1-\frac{2bz^{4}}{r_{+}^{2}}\phi^{\prime
2}(z)\right)\left[1+\frac{\kappa^{2}z^{3}}{1+z+z^{2}}\left(\frac{\lambda^{2}}{2}-\frac{b\lambda^{4}}{10}\xi(z)\right)\right],$
(61)
where $\zeta(z)=1+z+z^{2}+z^{3}+z^{4}$. Since the parameter
${\langle\mathcal{O}\rangle^{2}}/{r_{+}^{2}}$ is very small, we can expand
$\phi(z)$ as
$\displaystyle\frac{\phi(z)}{r_{+}}=\lambda(1-z)\left(1-\frac{2}{5}b\lambda^{2}\xi(z)\right)+\frac{\langle\mathcal{O}\rangle^{2}}{r_{+}^{2}}\chi(z).$
(62)
Substituting Eq. (62) into Eq. (60), we can obtain the equation of motion for
$\chi(z)$ as
$\displaystyle\Bigg{[}K(z)\chi^{\prime}(z)\Bigg{]}^{\prime}$ $\displaystyle=$
$\displaystyle(1+4b\lambda^{2}z^{4})^{1/2}\frac{\lambda F^{2}}{1+z+z^{2}}$
(63)
$\displaystyle\times\Bigg{\\{}1-\frac{2}{5}b\lambda^{2}(\xi(z)+5z^{4})+\frac{z^{3}}{1+z+z^{2}}\left(\frac{\kappa^{2}\lambda^{2}}{2}-\frac{b\kappa^{2}\lambda^{4}}{10}(3\xi(z)+10z^{4})\right)\Bigg{\\}}.$
with $\chi(1)=0=\chi^{\prime}(1)$ and we have defined
$\displaystyle K(z)=\left(1+4b\lambda^{2}z^{4}\right)^{3/2}.$ (64)
Integrating both sides of Eq. (63) between $z=0$ to $z=1$, we reach
$\displaystyle\chi^{\prime}(0)$ $\displaystyle=$
$\displaystyle-\lambda\int_{0}^{1}dz(1+4b\lambda^{2}z^{4})^{1/2}\frac{F^{2}}{1+z+z^{2}}$
(65)
$\displaystyle\times\Bigg{\\{}1-\frac{2}{5}b\lambda^{2}(\xi(z)+5z^{4})+\frac{z^{3}}{1+z+z^{2}}\left(\frac{\kappa^{2}\lambda^{2}}{2}-\frac{b\kappa^{2}\lambda^{4}}{10}(3\xi(z)+10z^{4})\right)\Bigg{\\}}.$
Equating $\phi(z)$ from Eqs. (19) and (62), we arrive at
$\displaystyle\frac{\mu}{r_{+}}-\frac{\rho}{r_{+}^{2}}z$ $\displaystyle=$
$\displaystyle\lambda(1-z)\Bigg{\\{}1-\frac{2}{5}b\lambda^{2}\xi(z)\Bigg{\\}}+\frac{\langle\mathcal{O}\rangle^{2}}{r_{+}^{2}}\chi(z)$
(66) $\displaystyle=$
$\displaystyle\lambda(1-z)\Bigg{\\{}1-\frac{2}{5}b\lambda^{2}\xi(z)\Bigg{\\}}+\frac{\langle\mathcal{O}\rangle^{2}}{r_{+}^{2}}\left(\chi(0)+z\chi^{\prime}(0)+...\right),$
where in the last step we have expanded $\chi(z)$ around $z=0$. Considering
the coefficients of $z$ term in both sides of Eq. (66), we find that
$\displaystyle\frac{\rho}{r_{+}^{2}}=\lambda-\frac{\langle\mathcal{O}\rangle^{2}}{r_{+}^{2}}\chi^{\prime}(0).$
(67)
Substituting $\chi^{\prime}(0)$ from Eq. (65) in the above relation, we get
$\displaystyle\frac{\rho}{r_{+}^{2}}=\lambda\Bigg{\\{}1+\frac{\langle\mathcal{O}\rangle^{2}}{r_{+}^{2}}\mathcal{A}\Bigg{\\}},$
(68)
where
$\displaystyle\mathcal{A}$ $\displaystyle=$
$\displaystyle\int_{0}^{1}dz(1+4b\lambda^{2}z^{4})^{1/2}\frac{F^{2}}{1+z+z^{2}}$
(69)
$\displaystyle\times\Bigg{\\{}1-\frac{2}{5}b\lambda^{2}(\xi(z)+5z^{4})+\frac{z^{3}}{1+z+z^{2}}\left(\frac{\kappa^{2}\lambda^{2}}{2}-\frac{b\kappa^{2}\lambda^{4}}{10}(3\xi(z)+10z^{4})\right)\Bigg{\\}}.$
Using Eqs. (13), (18) and (38), and taking into account the fact that $T$ is
very close to $T_{c}$, we can deduce
$\displaystyle r_{+}=\frac{4\pi
T}{\left[3-\frac{\kappa^{2}\lambda^{2}}{2}+\frac{b}{2}\kappa^{2}\lambda^{4}\right]}.$
(70)
Eqs. (39) and (70) show that Eq. (68) can be rewritten as
$\displaystyle
T_{c}^{2}-T^{2}=\langle\mathcal{O}\rangle^{2}\frac{\mathcal{A}}{(4\pi)^{2}}\left[3-\frac{\kappa^{2}\lambda^{2}}{2}+\frac{b}{2}\kappa^{2}\lambda^{4}\right]^{2}.$
(71)
Thus, we find the expectation value $\langle\mathcal{O}\rangle$ near the
critical point as
$\displaystyle\langle\mathcal{O}\rangle=\gamma T_{c}\sqrt{1-\frac{T}{T_{c}}},$
(72)
where $\gamma$ is the condensation parameter of the system which is given by
$\displaystyle\gamma=\frac{4\pi\sqrt{2}}{\sqrt{\mathcal{A}}}\left[3-\frac{\kappa^{2}\lambda^{2}}{2}+\frac{b}{2}\kappa^{2}\lambda^{4}\right]^{-1}.$
(73)
The relation obtained in Eq. (72) is valid for small nonlinear coupling and
backreaction parameters and satisfies
$\langle\mathcal{O}\rangle\sim\sqrt{1-\frac{T}{T_{c}}}$. Therefore, the
analytical result supports that the phase transition for the superconductor
belongs to the second order and the critical exponent of the system takes the
mean-field value $1/2$. This implies that considering nonlinear coupling and
backreaction parameters the value of the critical exponent will not be
altered. As we see in table $2$, condensation values $\gamma$ increases with
increasing the nonlinear parameter $b$ for the fixed parameter $\kappa$. Also,
we see the same behavior between the condensation values $\gamma$ and the
backreaction parameter with a fixed value of the nonlinear parameter $b$. This
means that the condensation becomes harder to be formed by considering both
the nonlinear corrections to the gauge field and taking the backreactions into
account. It should be noted that, at a temperature slightly below $T_{c}$ for
the $(2+1)$-dimensional holographic superconductors with backreaction,
condensation values for both BI and EN holographic superconductors have the
same behaviour, as we see in table $2$. Also, because of the larger parameter
$\gamma$, effect of the EN electrodynamics on the condensation of the scalar
operators is bigger than that of BI case. This implies that the scalar hair is
more difficult to be developed in the holographic superconductors with EN
electrodynamics.
n | b=0 | | b=0.1 | | b=0.2 | | b=0.3 |
---|---|---|---|---|---|---|---|---
| BI | EN | BI | EN | BI | EN | BI | EN
0 | 8.07 | 8.07 | 8.1801 | 8.5298 | 8.3094 | 9.1579 | 8.4696 | 10.0355
1 | 8.09 | 8.09 | 8.1869 | 8.5331 | 8.3212 | 9.1616 | 8.4890 | 10.0399
2 | 8.11 | 8.11 | 8.1943 | 8.5443 | 8.3237 | 9.1742 | 8.4893 | 10.0565
3 | 8.115 | 8.115 | 8.2121 | 8.5630 | 8.3417 | 9.1951 | 8.5023 | 10.0818
4 | 8.13 | 8.13 | 8.2370 | 8.5889 | 8.3669 | 9.2241 | 8.5277 | 10.1014
5 | 8.16 | 8.16 | 8.2909 | 8.6425 | 8.3994 | 9.2615 | 8.5606 | 10.1639
6 | 8.20 | 8.20 | 8.3079 | 8.6617 | 8.4391 | 9.3070 | 8.6007 | 10.2193
Table 2: The values of the condensation parameter $\gamma$ for holographic
superconductors in the presence of EN electrodynamics. Here we have taken
$\kappa_{n}=n\Delta\kappa$ where $\Delta\kappa=0.05$. We have also provided
the results for BI holographic superconductor from Ref. BIBR , for comparison.
## V Conclusions
We have introduced a different type of gravity dual models, i.e., the charged
AdS black holes in the context of Einstein-nonlinear electrodynamics with a
scalar field. We have assumed the EN electrodynamics as the gauge field, and
analytically investigated the behavior of the $(2+1)$-dimensional holographic
superconductors. We have worked in a limit in which the scalar and gauge
fields backreact on the background metric. We have employed the Sturm-
Liouville analytic method to explore the problem. We have found the influence
of the nonlinear corrections to the gauge filed as well as the backreaction
effects on the critical temperature and the process of the scalar field
condensation. We observed that the formation of the scalar hair condensation
on the boundary becomes harder in the presence of nonlinear electrodynamics.
This is mainly caused by the decreasing of the critical temperature when the
both nonlinear and backreaction parameters become stronger. This phenomenon
was also obtained in the study of the effect of the BI and backreaction
parameters in the $(2+1)$-dimensional holographic superconductors BIBR .
Comparing these different models show that for a specific $b$ the critical
temperature $T_{c}$ becomes larger for a holographic superconductor with BI
nonlinear electrodynamics comparing to the case with EN electrodynamics. This
implies that the scalar hair is more difficult to develop in the latter case
than the former one. We have also given the critical exponent for the EN
holographic superconductor model with backreaction, which still takes the
mean-field value $1/2$. We found out that the condensation parameter $\gamma$
in Eq. (72) increases with (i) increasing the nonlinear parameter $b$ with a
fixed backreaction parameter $\kappa$, (ii) increasing the backreaction
parameter with a fixed value of the nonlinear parameter $b$. This implies that
both the nonlinear corrections to the gauge field as well as backreaction,
cause the formation of condensation harder.
###### Acknowledgements.
We thank Shiraz University Research Council. This work has been supported
financially by Research Institute for Astronomy and Astrophysics of Maragha
(RIAAM), Iran.
## References
* (1) J. M. Maldacena, The large-N limit of superconformal field theories and supergravity, Adv. Theor. Math. Phys. 2, 231 (1998).
* (2) S. S. Gubser, I. R. Klebanov, and A. M. Polyakov, Gauge theory correlators from non-critical string theory, Phys. Lett. B 428, 105 (1998).
* (3) E. Witten, Anti-de Sitter space and holography, Adv. Theor. Math. Phys. 2, 253 (1998).
* (4) S. A. Hartnoll, C. P. Herzog and G. T. Horowitz, Building a Holographic Superconductor, Phys. Rev. Lett. 101, 031601 (2008) [arXiv:0803.3295].
* (5) S. A. Hartnoll, C. P. Herzog and G. T. Horowitz, Holographic superconductors, JHEP 12, 015 (2008) [arXiv:0810.1563].
* (6) G. T. Horowitz and M. M. Roberts, Holographic Superconductors with Various Condensates, Phys. Rev. D 78 , 126008 (2008) [arXiv:0810.1077].
* (7) S. A. Hartnoll, Lectures on holographic methods for condensed matter physics, Class. Quant. Grav. 26, 224002 (2009) [arXiv:0903.3246].
* (8) S. Franco, A. Garcia-Garcia, and D. Rodriguez-Gomez, A general class of holographic superconductors, JHEP 04, 092 (2010) [arXiv:0906.1214].
* (9) Q. Y. Pan and B. Wang, General holographic superconductor models with Gauss-Bonnet corrections, Phys. Lett. B 693, 159 (2010) [arXiv:1005.4743].
* (10) X. H. Ge, B. Wang, S. F. Wu and G. H. Yang, Analytical study on holographic superconductors in external magnetic field, JHEP 08, 108 (2010) [arXiv:1002.4901].
* (11) Y. Liu, Y. Peng, and B. Wang, Gauss-Bonnet holographic superconductors in Born-Infeld electrodynamics with backreactions, [arXiv:1202.3586].
* (12) Y. Liu, Y. Gong, and B. Wang, Non-equilibrium condensation process in holographic superconductor with nonlinear electrodynamics, DOI: 10.1007/JHEP02(2016)116, [arXiv:1505.03603].
* (13) D. Roychowdhury, Effect of external magnetic field on holographic superconductors in presence of nonlinear corrections, Phys Rev D 86, 106009 (2012) [arXiv:1211.0904].
* (14) J. Jing and S. Chen, Holographic superconductors in the Born-Infeld electrodynamics, Phys. Lett. B 686, 68 (2010) [arXiv:1001.4227].
* (15) R. Banerjee, S. Gangopadhyay, D. Roychowdhury and, A. Lala, Holographic s-wave condensate with nonlinear electrodynamics: A nontrivial boundary value problem, Phys. Rev. D 87, 104001 (2013) [arXiv:1208.5902].
* (16) J. Jing, Q. Pan, S. Chen, Holographic Superconductors with Power-Maxwell field, JHEP 11, 045 (2011) [arXiv:1106.5181].
* (17) Y. Liu, and B. Wang, Perturbations around the AdS Born-Infeld black holes, Phys. Rev. D 85, 046011 (2012) [arXiv:1111.6729].
* (18) A. Sheykhi, H. R. Salahi, A. Montakhab, Analytical and Numerical Study of Gauss-Bonnet Holographic Superconductors with Power-Maxwell Field, JHEP 04, 058 (2016) [arXiv:1603.00075].
* (19) C. Lai, Q. Pan, J. Jing, and Y. Wang, On analytical study of holographic superconductors with Born-Infeld electrodynamics, Phys. Lett. B 749, 437 (2015) [arXiv:1508.05926].
* (20) Q. Pan, J. Jing, B. Wang, and S. Chen, Analytical study on holographic superconductors with backreactions, JHEP 06, 087 (2012) [arXiv:1205.3543].
* (21) S. S. Gubser and A. Nellore, Low-temperature behavior of the Abelian Higgs model in anti-de Sitter space, JHEP 04, 008 (2009) [arXiv:0810.4554].
* (22) Y. Brihaye and B. Hartmann, Holographic Superconductors in 3+1 dimensions away from the probe limit, Phys. Rev. D 81, 126008 (2010) [ arXiv:1003.5130].
* (23) G. T. Horowitz and B. Way, Complete Phase Diagrams for a Holographic Superconductor/Insulator System, JHEP 11, 011 (2010) [arXiv:1007.3714].
* (24) A. Akhavan and M. Alishahiha, P-Wave Holographic Insulator/Superconductor Phase Transition, Phys. Rev. D 83, 086003 (2011) [arXiv:1011.6158].
* (25) S. Gangopadhyay, Analytic study of properties of holographic superconductors away from the probe limit, Phys. Lett. B 724, 176 (2013) [arXiv:1302.1288].
* (26) T. Konstandin, G. Nardini, and M. Quiros, Gravitational Backreaction Effects on the Holographic Phase Transition, Phys. Rev. D 82, 083513 (2010) [arXiv:1007.1468].
* (27) A. Sheykhi, F. Shaker, Analytical study of holographic superconductor in Born-Infeld electrodynamics with backreaction, Phys. Lett. B 754, 281 (2016) [arXiv:1601.04035].
* (28) Z. Zhao, Q. Pan, S. Chen, and J. Jing, Notes on holographic superconductor models with the nonlinear electrodynamics, Nucl. Phys. B 871, 98 (2013) [arXiv:1212.6693].
* (29) S. H. Hendi, Asymptotic charged BTZ black hole solutions, JHEP 03, 065 (2012);
S. H. Hendi, A. Sheykhi, Charged rotating black string in gravitating
nonlinear electromagnetic fields, Phys. Rev D 88, 044044 (2013)
[arXiv:1405.6998].
* (30) L. Barclay, R. Gregory, S. Kanno, and P. Sutcliffe, Gauss-Bonnet Holographic Superconductors, JHEP 12, 029 (2010) [arXiv:1009.1991];
R. Gregory, Holographic Superconductivity with Gauss-Bonnet gravity, J. Phys.
Conf. Ser. 283, 012016 (2011) [arXiv:1012.1558];
W. Yao, J. Jing, Holographic entanglement entropy in metal/superconductor
phase transition with Born-Infeld electrodynamics, Nucl. Phys. B 889, 109
(2014) [arXiv:1408.1171].
* (31) M. Abramowitz and I. A. Stegun, _Handbook of Mathematical Functions_ , Dover, New York, (1972).
* (32) S. Gangopadhyay and D. Roychowdhury, Analytic study of properties of holographic superconductors in Born-Infeld electrodynamics, JHEP 05, 002 (2012) [arXiv:1201.6520].
* (33) H. B. Zeng, X. Gao, Y. Jiang, and H. S. Zong, Analytical Computation of Critical Exponents in Several Holographic Superconductors, JHEP 05, 002 (2011) [arXiv:1012.5564].
|
# Transcendence measure of $e^{1/n}$
Marta Dujella Universität Basel, Spiegelgasse 1, 4051 Basel, Switzerland.
<EMAIL_ADDRESS>, Anne-Maria Ernvall-Hytönen University of Helsinki,
PL 68, 00014 Helsingin yliopisto, Finland. anne-maria.ernvall-
<EMAIL_ADDRESS>, Linda Frey Universität Göttingen, Bunsenstraße 3-5,
37073 Göttingen, Germany<EMAIL_ADDRESS>and Bidisha Roy Scuola
Normale di Pisa, Piazza dei Cavlieri,7, 56126, Pisa, Italy.
<EMAIL_ADDRESS>
###### Abstract.
For a given transcendental number $\xi$ and for any polynomial
$P(X)=:\lambda_{0}+\cdots+\lambda_{k}X^{k}\in\mathbb{Z}[X]$, we know that
$P(\xi)\neq 0.$ Let $k\geq 1$ and $\omega(k,H)$ be the infimum of the numbers
$r>0$ satisfying the estimate
$\left|\lambda_{0}+\lambda_{1}\xi+\lambda_{2}\xi^{2}+\ldots+\lambda_{k}\xi^{k}\right|>\frac{1}{H^{r}},$
for all
$(\lambda_{0},\ldots,\lambda_{k})^{T}\in\mathbb{Z}^{k+1}\setminus\\{\overline{0}\\}$
with $\max_{1\leq i\leq k}\\{|\lambda_{i}|\\}\leq H$. Any function greater
than or equal to $\omega(k,H)$ is a transcendence measure of $\xi$. In this
article, we find out a transcendence measure of $e^{1/n}$ which improves a
result proved by Mahler([7]) in 1975.
We thank the organizers of the conference Women in Numbers Europe 4,
especially but not exclusively Valentijn Karemaker and Nirvana Coppola. This
conference set the foundation for this article. Furthermore, we thank the
Universiteit Utrecht for granting unlimited coffee.
## 1\. Introduction
Let $\xi$ be a transcendental number. Then $P(\xi)\neq 0$ for any polynomial
$P$ with integer coefficient.
Let $k,H\geq 1$ and $\omega(k,H)$ is the infimum of the numbers $r>0$
satisfying the estimate
(1)
$\left|\lambda_{0}+\lambda_{1}\xi+\lambda_{2}\xi^{2}+\ldots+\lambda_{k}\xi^{k}\right|>\frac{1}{H^{r}},$
for all
$\overline{\lambda}=(\lambda_{0},\ldots,\lambda_{k})^{T}\in\mathbb{Z}^{k+1}\setminus\\{\overline{0}\\}$
with $\max_{1\leq i\leq k}\\{|\lambda_{i}|\\}\leq H$. Any function greater
than or equal to $\omega(k,H)$ is a transcendence measure of $\xi$.
Bounding transcendence measures of different constants is a classical problem
in number theory. It is widely investigated in particular in the context of
the Napier’s constant $e$. In 1873, Hermite proved it to be transcendental
[5]. This work started with Borel in 1899 [1] when proved that
$\omega(k,H)<c\log\log H$ for some constant $c$ depending on $k$. Popken [9,
10] improved the bound to $k+\frac{c}{\log\log H}$, where $c$ depends on $k$
in 1928–1929. Mahler [8] made the dependance on $k$ explicit in 1931 by
deriving the bound $k+\frac{ck^{2}\log(k+1)}{\log\log H}$, where $c$ is an
absolute constant. This result is already of the shape of modern state of art
results. In 1991, Khassa and Srinivasan [6] showed that $c=98$ is valid for
$\log\log Hd(k+1)^{6k}$ for some constant $d>e^{950}$. This result was
improved by Hata in 1995 [4] who showed that one can choose $c=1$, and also
considerably improved the lower bound for $H$. This was further improved by
Ernvall-Hytönen, Matala-aho and Seppälä ion 2018 [3].
In this paper, we concentrate on the expression
$\left|\lambda_{k}e^{k/n}+\ldots+\lambda_{1}e^{1/n}+\lambda_{0}\right|,$
where $k\geq n\geq 2$. This gives a transcendence measure for roots of $e$.
Since $e$ is transcendental, all its roots and powers are also transcendental.
Ernvall-Hytönen, Matala-aho and Seppälä considered also sparse polynomials in
the context of the transcendence measure of $e$. As a corollary, they also
derived a transcendence measure for integer powers of $e$. However, we were
not able to find any transcendence measures tailored for _roots of $e$_ in the
literature. There are some general results in the literature, for instance, by
Mahler [7] which can be used to derive a bound. Also, the generalized
transcendence measure by Ernvall-Hytönen, Leppälä and Matala-aho can be used
to derive a bound. Our bound will be compared to these bounds in Section 2.
In this article, we prove the following bound:
###### Theorem 1.
Assume $k\geq n\geq 2$. We have
(2)
$\left|\lambda_{0}+\lambda_{1}e^{1/n}+\lambda_{2}e^{2/n}+\ldots+\lambda_{k}e^{k/n}\right|>\frac{1}{H^{r}},$
where $r>\omega(k,H)$ and we can choose
$\omega(k,H)=k+\frac{k^{2}\log k}{\log\log H}\left(1+\frac{0.69}{\log
k-1}\right),$
for $k\geq 5$ and $\omega(k,H)=k+\frac{k^{2}\log k}{\log\log H}d(k)$, where
$d(k)=\begin{cases}&3.319\textrm{ for }k=2\\\ &1.145\textrm{ for }k=3\\\
&1.114\textrm{ for }k=4\end{cases}$
and $\log H\geq s(n,k)e^{s(n,k)}$ with $s(n,k)=(k+n)(\log(k+n))^{2}$.
We follow the approach used in [3].
## 2\. Earlier results and comparisons to our bound
The following result can be obtained as a corollary of a much more general
result presented in by Mahler in 1975.
###### Theorem 2 (Theorem 1 in [7]).
Take $a_{i}=i$ for $i=0,\ldots,k$ (with $k\geq 2$) and $a=n$ a positive
integer. Let $\lambda_{0},\ldots,\lambda_{k}$ be integers not all zero (in
Mahler’s paper $x_{0},\ldots,x_{k}$), $C(r)=(k+1)^{2}r\sqrt{(\log(n+k+1)\log
r}$ and $T$ be the product of the non-zero $\lambda_{i}$. Then for $r$ the
smallest integer for which
$\displaystyle\frac{(r-1)!}{e^{2C(r-1)}}\leq\max\\{|x_{i}|\\}<\frac{r!}{e^{2C(r)}}$
we have
(3)
$\displaystyle|T(\lambda_{0}+\lambda_{1}e^{\frac{1}{n}}+\ldots+\lambda_{k}e^{\frac{k}{n}})|>\frac{\max\\{|\lambda_{i}|\\}}{e^{(2(k+1)-\frac{1}{4})C(r)}}.$
The following result is due to Ernvall-Hytönen, Leppälä and Matala-aho, and
can be obtained as a corollary of a much more general result presented in [2].
###### Theorem 3 (Corollary of [2]).
We have
$\left|\lambda_{0}+\lambda_{1}e^{1/n}+\cdots+\lambda_{k}e^{k/n}\right|>\frac{M^{1-\hat{\delta}(M)}}{h_{0}h_{1}\ldots
h_{k}},$
where $M=\max_{0\leq i\leq k}\\{|\lambda_{i}|\\}$,
$\hat{\delta}(M)\leq\frac{\hat{B}(\overline{\alpha})}{\sqrt{\log\log M}}\leq
c_{k}k^{2}\sqrt{\log(g_{1}(\overline{\alpha})(1+g_{3}(\overline{\alpha})))}/\sqrt{\log\log
M}$ and $h_{i}=\max\\{1,|\lambda_{i}|\\}$, for $i=1,\ldots,k$. Moreover,
$c_{k}=13$ if $k<3$ and $12$ otherwise.
In particular, for $\overline{\alpha}=(0,\frac{1}{n},\ldots,\frac{k}{n})$ we
have $g_{1}(\overline{\alpha})=n$ and $g_{3}(\overline{\alpha})=\frac{k}{n}$.
Therefore,
$\left|\lambda_{0}+\lambda_{1}e^{1/n}+\cdots+\lambda_{k}e^{k/n}\right|>\frac{M}{h_{0}\cdots
h_{k}M^{\frac{c_{k}k^{2}\sqrt{\log(n(1+k/n))}}{\sqrt{\log\log M}}}}.$
Let us now compare these results with our bound.
###### Example 4.
Let us look at the family of polynomials with
$\frac{H}{2}\leq|\lambda_{i}|\leq H$ for all coefficients $\lambda_{i}$ when
$1\leq i\leq k$ and compare our result with results in [2] and [7].
Our result gives the bound
$\left|\lambda_{0}+\lambda_{1}e^{1/n}+\cdots+\lambda_{k}e^{k/n}\right|>H^{-k-\frac{k^{2}\log
k}{\log\log H}\left(1+\frac{0.639}{\log k-1}\right)}$
The bound by Ernvall-Hytönen, Matala-aho and Leppälä:
$\displaystyle\left|\lambda_{0}+\lambda_{1}e^{1/n}+\cdots+\lambda_{k}e^{k/n}\right|$
$\displaystyle>\frac{H}{h_{0}\cdots
h_{k}H^{\frac{c_{k}k^{2}\sqrt{\log(n(1+k/n))}}{\sqrt{\log\log H}}}}$
This bound is certainly not better than
$\left(\frac{H}{2}\right)^{-k-\frac{12k^{2}\sqrt{\log(n(1+k/n))}}{\sqrt{\log\log
H}}}=H^{-k+k\frac{\log 2}{\log
H}-\frac{12k^{2}\sqrt{\log(n+k)}}{\sqrt{\log\log H}}},$
which is weaker than ours for large values of $H$ because for large
$\sqrt{\log\log H}$ grows slower than $\log\log H$.
Mahler 1975:
(4)
$\displaystyle|\lambda_{0}+\lambda_{1}e^{\frac{1}{n}}+\ldots+\lambda_{k}e^{\frac{k}{n}}|>\frac{\max\\{|\lambda_{i}|\\}}{Te^{(2(k+1)-\frac{1}{4})C(r)}},$
where $T$ is the product of all non-zero $\lambda_{i}$’s.
(5)
$\displaystyle\frac{\max\\{|\lambda_{i}|\\}}{Te^{(2(k+1)-\frac{1}{4})C(r)}}$
$\displaystyle\leq\frac{1}{\left(\frac{H}{2}\right)^{k}e^{(2(k+1)-1/4)C(r)}}.$
Mahler gives the bound $\frac{\log x}{\log\log x}<r<\frac{6\log x}{\log\log
x}$, where in his notation $x$ is the maximum of the absolute values of the
coefficients of the polynomial. In our setting, this inequality would
approximately translate to
$\frac{\log H}{\log\log H}<r<\frac{6\log H}{\log\log H}.$
We are losing some accuracy here because we only expected the coefficients of
the polynomial to be on the interval $[\frac{H}{2},H]$. However, for the
current purposes, this is not an issue.
The denominator of (5) can now be written as
$\left(\frac{H}{2}\right)^{k}e^{(2(k+1)-1/4)(2(k+1)-1/4)C(r)}=H^{k-\frac{k\log
2-(2(k+1)-1/4)C(r)}{\log H}}=H^{k-\frac{k\log 2}{\log
H}+\frac{(2(k+1)-1/4)C(r)}{\log H}}.$
Let us now look at the expression $\frac{(2(k+1)-1/4)C(r)}{\log H}$. Let us
use the expression for $C(r)$ and the bound for $r$:
$C(r)=(k+1)^{2}r\sqrt{(\log(n+k+1)\log r}\approx(k+1)^{2}\frac{\log
H}{\log\log H}\sqrt{(\log(n+k+1)\log\frac{\log H}{\log\log H}}\\\
\approx(k+1)^{2}\frac{\log H\sqrt{\log k}}{\sqrt{\log\log H}},$
where we used the bound $n\leq k$ to estimate $\log(n+k+1)\approx\log k$, and
that for large H, $\log\frac{\log H}{\log\log H}\approx\log\log H$. Hence,
$\frac{(2(k+1)-1/4)C(r)}{\log H}\approx\frac{(2(k+1)-1/4)}{\log
H}\cdot(k+1)^{2}\frac{\log H\sqrt{\log k}}{\sqrt{\log\log
H}}\approx\frac{2(k+1)^{2}\sqrt{\log k}}{\sqrt{\log\log H}}.$
Hence, our bound is also better than Mahler’s bound, because $\sqrt{\log\log
H}$ grows slower than $\log\log H$, and the numerator is bigger (dependance on
$k^{3}$ instead of $k^{2}$).
## 3\. Preliminaries and the outline of the method
Ernvall-Hytönen, Seppälä and Matala-aho [3] used the following approach:
Assume that there is a sequence of simultaneous approximations
$L_{m,j}(h)=B_{m,0}(h)\Theta_{j}+B_{m,j}(h)$
with $m=0,1,\dots,k$ and $j=1,2,\dots,k$. Further assume
$B_{m,j}(\ell)\in\mathbb{Z}$ for all $m,j\in\\{0,1,\dots,k\\}$. Assume further
that the coefficients $B_{m,j}$ satisfy the following determinant condition:
$\begin{vmatrix}B_{0,0}&B_{0,1}&\cdots&B_{0,k}\\\
B_{1,0}&B_{1,1}&\cdots&B_{1,k}\\\ \vdots&\vdots&\ddots&\vdots\\\
B_{k,0}&B_{m,1}&\cdots&B_{k,k}\\\ \end{vmatrix}\neq 0.$
Pick the functions $Q(h)$, $q(h)$, $R(h)$ and $r(h)$ to be such that they
satisfy the following inequalities:
$B_{m,0}(h)\leq Q(h)=e^{q(h)}$
and
$\sum_{j=1}^{k}|L_{m,j}(h)|\leq R(h)=e^{-r(h)},$
for all $h\geq h_{0}$, where the functions are of the form
$q(h)=ah\log h+bh$
and
$-r(h)=-ch\log h+dh.$
Assume that $z(y)$ is the inverse function of the function $y(z)=z\log z$.
Further, denote
$B=b+\frac{ad}{c},\quad C=a,\quad D=a+b+e^{-s(m)},\quad F^{-1}=2e^{D},\quad
v=c-\frac{d}{s(m)},\quad h_{1}=\max\\{h_{0},e,e^{s(m)}\\}.$
Our choice will be $s(n,k)=(n+k)(\log(n+k))^{2}$, and we will actually have
$h_{1}=e^{s(n,k)}$. Under the assumptions above, they proved the following
lemma:
###### Lemma 5 ([3]).
Let $m\geq 1$ and $\log(2H)\geq vh_{1}\log h_{1}$. Then under the assumptions
above
(6)
$|\lambda_{0}+\lambda_{1}\Theta_{1}+\cdots+\lambda_{m}\Theta_{m}|>F(2H)^{-\frac{a}{c}-\epsilon(H)},$
where
$\epsilon(H)\log(2H)=Bz\left(\frac{\log(2H)}{v}\right)+C\log\left(z\left(\frac{\log(2H)}{v}\right)\right).$
Furthermore, they gave the following construction for the approximations in
the case of $e^{\alpha_{j}}$: Write
$\overline{\alpha}=(\alpha_{0},\ldots,\alpha_{k})$ and set
(7)
$\displaystyle\Omega(x,\overline{\alpha})=\prod_{j=0}^{m}(\alpha_{j}-x)^{\ell_{j}}=\sum_{i=0}^{L}\sigma_{i}x^{i},$
where $L=\ell_{0}+\ell_{1}+\cdots+\ell_{m}$ and
$\sigma_{i}=\sigma_{i}(\overline{\ell},\overline{\alpha})$. Then choosing
$A_{0}(t)=\sum_{i=\ell_{0}}^{L}t^{L-i}i!\sigma_{i},$
we get
(8) $e^{\alpha_{j}}A_{0}(t)-A_{j}(t)=R_{j}(t),$
where $A_{j}(t)$ is a polynomial with integer coefficients and
$\begin{cases}\deg A_{0}(t)=L-\ell_{0}\\\ \deg A_{j}(t)=L-\ell_{j}\\\
\mathrm{ord}_{t=0}R_{j}(t)\geq L+1.\end{cases}$
Notice that the polynomials depend on the values of
$\ell_{0},\ell_{1},\dots,\ell_{m}$ and on $\overline{\alpha}$. We will
explicitly describe $A_{j}$ and $R_{j}$ in the following chapter.
In the following, we will be choosing $\Theta_{j}=e^{j/n}$ for some $n\geq 2$.
We will then proceed in the same fashion as in [3] to construct the explicit
polynomials used in the simultaneous approximations and to bound them.
Finally, we simplify the estimate given by (6).
## 4\. Explicit polynomial construction
We start by constructing the simultaneous approximations of the powers of the
roots of $e$. For estimating the required term, we set
$\overline{\alpha}=(\alpha_{0},\ldots,\alpha_{k})$ with $\alpha_{s}=s/n$, for
$s=0,1,\ldots k$. Let
$\overline{\ell}=(l_{0},\ldots,l_{k})\in\mathbb{Z}^{k+1}_{\geq 1}$ and
$L=l_{0}+\ldots+\l_{k}$. As explained in the previous chapter, we get the
following approximation formulas for $j=1,\ldots,k$
$e^{\alpha_{j}t}A_{0}(t)-A_{j}(t)=L_{j}(t),$
where
$A_{0}(t)=\sum_{i=\ell_{0}}^{L}t^{L-i}i!\sigma_{i}.$
With a direct computation (similarly as in [3]), we obtain
$\displaystyle\sigma_{i}$ $\displaystyle=(-1)^{i}\sum_{\ell_{0}+i_{1}+\ldots
i_{k}=i}\binom{\ell_{1}}{i_{1}}\ldots\binom{\ell_{k}}{i_{k}}\left(\frac{1}{n}\right)^{\ell_{1}-i_{1}}\ldots\left(\frac{k}{n}\right)^{\ell_{k}-i_{k}}$
$\displaystyle=(-1)^{i}\sum_{\ell_{0}+i_{1}+\ldots
i_{k}=i}\binom{\ell_{1}}{i_{1}}\ldots\binom{\ell_{k}}{i_{k}}n^{-L+i}2^{\ell_{2}-i_{2}}\dots
k^{\ell_{k}-i_{k}}$
Furthermore, $\sigma_{i}=0$ when $0\leq i<\ell_{0}$ and so
$A_{0}(t)=\sum_{i=0}^{L}t^{L-i}i!\sigma_{i}.$
We wish to now bound the polynomials. Laplace transform gives us a tool to
switch from sums to integrals, which is helpful in estimates. Since
$\frac{i!\sigma_{i}(\overline{\ell},\overline{\alpha})}{t^{i+1}}=\mathcal{L}(\sigma_{i}(\overline{\ell},\overline{\alpha})x^{i})(t)$
(where $\mathcal{L}$ denotes the Laplace transform), we have
$A_{0}(t)=\sum_{i=0}^{L}t^{L-i}i!\sigma_{i}=t^{L+1}\sum_{i=0}^{L}\mathcal{L}(\sigma_{i}x^{i})(t)=t^{L+1}\int_{0}^{\infty}e^{-xt}\Omega(x)dx,$
where $\Omega(x):=\Omega(x,\overline{\alpha})$ is given by (7). Now for any
$\alpha_{j}$, we have
$e^{\alpha_{j}t}A_{0}(t)=t^{L+1}\int_{0}^{\infty}e^{(\alpha_{j}-x)t}\Omega(x)dx=t^{L+1}\left(\int_{0}^{\alpha_{j}}+\int_{\alpha_{j}}^{\infty}\right)e^{(\alpha_{j}-x)t}\Omega(x)dx.$
Changing the variable in the second integral: $y=x-\alpha$ gives us:
$e^{\alpha_{j}t}A_{0}(t)=t^{L+1}\int_{0}^{\alpha_{i}}e^{(\alpha_{j}-x)t}\Omega(x)dx+t^{L+1}\int_{0}^{\infty}e^{-yt}\Omega(y+\alpha_{j})dy$
Hence we get
$\displaystyle A_{j}(t)=t^{L+1}\int_{0}^{\infty}e^{-yt}\Omega(y+\alpha_{j})dy$
and
$\displaystyle
L_{j}(t)=t^{L+1}\int_{0}^{\alpha_{j}}e^{(\alpha_{j}-x)t}\Omega(x)dx$
for $j=1,\ldots,k$.
We can make the result stronger if the terms in (8) are as small as possible.
But at the same time, we want to keep the coefficients in $A_{j}(t)$ as
integers. Therefore, we try to find as large common factors in coefficients as
possible.
To do that, we proceed as in [3]. We start by picking very specific values of
$\ell_{0},\ell_{1},\dots,\ell_{k}$ in relation to each other.
For any $u$ with $0\leq u\leq k$, we take
$\ell_{s}^{(u)}=\begin{cases}\ell-1&\mbox{if }s=u\\\
\ell&\mbox{otherwise}\end{cases}$ and
$\overline{\ell}^{(u)}=(\ell_{0}^{(u)},\ldots,\ell_{k}^{(u)})$. For these
values of $\overline{\ell}$, we denote $A_{j}(t)=A_{\overline{\ell},j}(t)$ by
$A_{u,j}(t)$ and $L_{j}(t)=L_{\overline{\ell},j}(t)$ by $L_{u,j}(t)$.
$A_{0}(t)=\sum_{i=\ell_{0}}^{L}t^{L-i}i!(-1)^{i}\sum_{\ell_{0}+i_{1}+\ldots
i_{k}=i}\binom{\ell_{1}}{i_{1}}\ldots\binom{\ell_{k}}{i_{k}}n^{-L+i}2^{\ell_{2}-i_{2}}\dots
k^{\ell_{k}-i_{k}}.$
For our chosen $\overline{\ell}$-s we always have
$\ell_{0}\in\\{\ell,\ell-1\\}$, so we can see that
$\displaystyle\frac{n^{L-\ell+1}}{(\ell-1)!}A_{u,0}(t)\in\mathbb{Z}[t].$
Similarly, we can look at polynomials $A_{u,j}(t)$ for $j=1,\ldots,k$. We have
$A_{u,j}(t)=\int_{0}^{\infty}e^{-yt}\Omega(y+\alpha_{j})dy=t^{L+1}\sum_{i=0}^{L}\mathcal{L}(\sigma_{i}(\overline{\ell},\overline{\beta^{(j)}}))=\sum_{i=0}^{L}t^{L-i}i!\sigma_{i}(\overline{\ell},\overline{\beta^{(j)}}),$
where
$\overline{\beta^{(j)}}=(\alpha_{0}-\alpha_{j},\ldots,\alpha_{k}-\alpha_{j})$
for each $j$. Since the coefficients
$\sigma_{i}(\overline{\ell},\overline{\beta^{(j)}})$ are defined using
polynomial $\Omega(\overline{\ell},\overline{\beta^{(j)}})$, and in the
product representation of $\Omega$, the term
$(t-\beta_{j}^{(j)})^{\ell_{j}}=(t-(\alpha_{j}-\alpha_{j}))^{\ell_{j}}=t^{\alpha_{j}}$
will define the lowest degree of terms occuring in
$\Omega(\overline{\ell},\overline{\beta})$, and thereby in $A_{u,j}$, we have
$\sigma_{i}(\overline{\ell},\overline{\beta})=0$ unless $i\geq\ell_{j}$.
Hence, all the coefficients in the representation of $A_{u,j}(t)$ have the
factor $\ell_{j}!$, which is again either $\ell$ or $\ell-1$. On the other
hand, these terms have the denominator $n^{L-i}$, so again
$\frac{n^{L-\ell+1}}{(\ell-1)!}A_{u,j}(t)\in\mathbb{Z}[t].$
## 5\. Estimation of $A_{u,0}(1)$
Next, we would like to estimate the term $A_{u,0}(t)$, for which its
representation as an integral will be useful. We have fixed
$\overline{\alpha}=(1/n,\ldots,k/n)$ and so
$\Omega(x)=\prod_{j=0}^{k}\left(j/n-x\right)$. Furthermore, for the choices of
$\overline{\ell}$ as in the previous section we have $L-1=(k+1)\ell$.
Therefore, $A_{u,0}(t)$ looks like
$\displaystyle A_{u,0}(t)$
$\displaystyle=t^{(k+1)\ell}\int_{0}^{\infty}e^{-yt}\prod_{j=0}^{k}\left(\frac{j}{n}-x\right)^{\ell_{s}}dx$
$\displaystyle=t^{(k+1)\ell}\int_{0}^{\infty}e^{-xt}(-x)^{\ell}\left(\frac{1}{n}-x\right)^{\ell}\cdots\left(\frac{u}{n}-x\right)^{\ell-1}\cdots\left(\frac{k}{n}-x\right)^{\ell}dx.$
Note that $\left|\frac{x^{\ell}(x-1/n)\cdots(x-k/n)}{(x-u/n)}\right|\leq
x^{(k+1)\ell-1}\leq x^{(k+1)\ell}$, for $x>\frac{k}{n}.$ This gives us an idea
of how the function inside the integral behaves: while $0\leq
x\leq\frac{k}{n}$, the function stays relatively small, and it touches zero at
points $0,\frac{1}{n},\dots,\frac{k}{n}$. However, when $x\geq\frac{k}{n}$, it
starts behaving roughly as $x^{(k+1)\ell-1}e^{-x}$.
Therefore, we split the above integral in the following way
(9)
$\int_{0}^{\infty}e^{-xt}\prod_{j=0}^{k}\left(\frac{j}{n}-x\right)^{\ell_{s}}dx=\left(\int_{0}^{k/n}+\int_{k/n}^{2(k+1)\ell}+\int_{2(k+1)\ell}^{\infty}\right)e^{-xt}\prod_{j=0}^{k}\left(\frac{j}{n}-x\right)^{\ell_{s}}dx.$
Now we treat the above integrals with $t=1$. In that case,
(10)
$\displaystyle\left|\left(\int_{k/n}^{2(k+1)\ell}+\int_{2(k+1)\ell}^{\infty}\right)e^{-x}\frac{\prod_{s=0}^{k}\left(\frac{s}{n}-x\right)^{\ell}}{\left(\frac{u}{n}-x\right)}dx\right|\leq\left(\int_{k/n}^{2(k+1)\ell}+\int_{2(k+1)\ell}^{\infty}\right)e^{-x}x^{(k+1)\ell-1}dx:=I_{1}+I_{2}.$
For estimating the above integrals, we first consider $I_{1}$. In that case,
$\displaystyle|I_{1}|=\left|\int_{k/n}^{2(k+1)\ell}e^{-x}x^{(k+1)\ell-1}dx\right|<2(k+1)\ell
e^{-(k+1)\ell+1}((k+1)\ell-1)^{(k+1)\ell-1}$
because the expression $e^{-x}x^{(k+1)\ell-1}$ is maximal when
$x=(k+1)\ell-1$.
Let us now move to $I_{2}$.
###### Lemma 6.
Let $c>1$ be a constant. We have
$\int_{c\ell(k+1)}^{\infty}e^{-x}\frac{\left(\prod_{j=0}^{k}\left|\frac{j}{n}-x\right|\right)^{\ell}}{\left|\frac{k^{\prime}}{n}-x\right|}\leq\frac{c}{c-1}e^{-c(k+1)l}(c(k+1)\ell)^{\ell(k+1)-1}.$
for any $0\leq k^{\prime}\leq k$.
###### Proof.
We can partially integrate:
$\int e^{-x}x^{t}dx=\left[-e^{-x}x^{t}\right]+\int e^{-x}tx^{t-1},$
which gives us the series expansion for the integral above:
$\int_{c\ell(k+1)}^{\infty}e^{-x}x^{\ell(k+1)-1}dx\\\
=e^{-c(k+1)\ell}(c(k+1)\ell)^{\ell(k+1)-1}+(\ell(k+1)-1)e^{-c(k+1)\ell}(c(k+1)\ell)^{\ell(k+1)-2}\\\
+(\ell(k+1)-1)(\ell(k+1)-2)e^{-c(k+1)\ell}(c(k+1)\ell)^{\ell(k+1)-3}+\ldots\\\
\leq
e^{-c(k+1)l}(c(k+1)\ell)^{\ell(k+1)-1}\left(1+\frac{(k+1)\ell-1}{c(k+1)\ell}+\frac{(k+1)\ell-1}{c(k+1)\ell}\cdot\frac{(k+1)\ell-2}{c(k+1)\ell}\ldots\right)\\\
\leq\frac{c}{c-1}e^{-c(k+1)l}(c(k+1)\ell)^{\ell(k+1)-1}.$
∎
Notice that if we pick $c=2$, we get the following corollary:
###### Corollary 7.
We have the following estimate
$|I_{2}|\leq\int_{2\ell(k+1)}^{\infty}e^{-x}x^{\ell(k+1)-1}dx\leq
2e^{-2(k+1)l}(2(k+1)\ell)^{\ell(k+1)-1}.$
Next, it remains to get a bound for
$\displaystyle\int_{0}^{k/n}e^{-y}\frac{\prod_{s=0}^{k}\left(\frac{s}{n}-y\right)^{\ell}}{\left(\frac{u}{n}-y\right)}dy.$
###### Lemma 8.
Assume $k\geq 5$. Now
$\int_{0}^{k/n}e^{-y}\frac{\prod_{s=0}^{k}\left(\frac{s}{n}-y\right)^{\ell}}{\left(\frac{u}{n}-y\right)}dy\leq\frac{(k!)^{\ell}}{120^{\ell-1}n^{k\ell-5(\ell-1)}}c(n)^{\ell-1},$
where $c(n)=\displaystyle\max_{0\leq y\leq
1}\prod_{s=0}^{5}\left|\frac{s}{n}-y\right|^{\ell-1}.$ Furthermore,
$|c(n)|\leq 1$.
###### Proof.
Observe that $\displaystyle\max_{v\leq y\leq
v+1}\prod_{s=0}^{k}\left|\frac{s}{n}-y\right|^{\ell-1}\leq\max_{0\leq y\leq
1}\prod_{s=0}^{k}\left|\frac{s}{n}-y\right|^{\ell-1},$ for positive integer
$v$ with $0\leq v\leq k-1$. Therefore,
$\left|\frac{\prod_{s=0}^{k}\left(\frac{s}{n}-y\right)^{\ell}}{\left(\frac{u}{n}-y\right)}\right|\leq\frac{k!}{n^{k}}\max_{0\leq
y\leq 1}\prod_{s=0}^{k}\left|\frac{s}{n}-y\right|^{\ell-1}.$
Hence
$\displaystyle\left|\int_{0}^{k/n}e^{-y}\frac{\prod_{s=0}^{k}\left(\frac{s}{n}-y\right)^{\ell}}{\left(\frac{u}{n}-y\right)}dy\right|$
$\displaystyle\leq\int_{0}^{k/n}e^{-y}\frac{k!}{n^{k}}\max_{0\leq y\leq
1}\prod_{s=0}^{k}\left|\frac{s}{n}-y\right|^{\ell-1}dy$
$\displaystyle\leq\frac{k!}{n^{k}}\frac{(k!)^{\ell-1}}{(5!)^{\ell-1}n^{(k-5)(\ell-1)}}\int_{0}^{k/n}e^{-y}\max_{0\leq
y\leq 1}\prod_{s=0}^{5}\left|\frac{s}{n}-y\right|^{\ell-1}dy$
$\displaystyle\leq\frac{(k!)^{\ell}}{120^{\ell-1}n^{k\ell-5(\ell-1)}}c(n)^{\ell-1},\mbox{
writing }c(n)=:\displaystyle\max_{0\leq y\leq
1}\prod_{s=0}^{5}\left|\frac{s}{n}-y\right|.$
By checking small values individually, and bounding
$\prod_{s=0}^{5}\left|\frac{s}{n}-y\right|\leq 1$
for $n\geq 5$, we obtain $|c(n)|\leq 1$. ∎
###### Lemma 9.
Assume $k\leq 5$. Now
$\int_{0}^{k/n}e^{-y}\frac{\prod_{s=0}^{k}\left(\frac{s}{n}-y\right)^{\ell}}{\left(\frac{u}{n}-y\right)}dy\leq\frac{k!}{n^{k}}c(n,k)^{\ell-1},$
where
$c(n,k)=\begin{cases}0.049&\textrm{for }(n,k)=(2,2)\\\
\frac{1}{16}&\textrm{for }(n,k)=(2,3)\\\ \frac{1}{81}&\textrm{for
}(n,k)=(3,3)\\\ 0.114&\textrm{for }(n,k)=(2,4)\\\ 0.015&\textrm{for
}(n,k)=(3,4)\\\ 0.004&\textrm{for }(n,k)=(4,4)\par\end{cases}$
###### Proof.
We simply bound
$\int_{0}^{k/n}e^{-y}\frac{\prod_{s=0}^{k}\left(\frac{s}{n}-y\right)^{\ell}}{\left(\frac{u}{n}-y\right)}dy\leq\frac{k!}{n^{k}}\max_{0\leq
y\leq 1}\prod_{s=0}^{k}\left|y-\frac{s}{n}\right|\int_{0}^{k/n}e^{-y}dy,$
where the integral can be bounded to be at most $1$, and the individual maxima
can be determined using WolframAlpha.
∎
###### Lemma 10.
Assume $k\geq 2$. We have
$\int_{0}^{\infty}e^{-y}\frac{\prod_{s=0}^{k}\left(\frac{s}{n}-y\right)^{\ell}}{\left(\frac{u}{n}-y\right)}dy\leq\exp\left(\log
4+(k+1)\ell\log 2+(\ell(k+1)-1)\log((k+1)\ell)-\ell(k+1)+1\right)$
###### Proof.
Assume first $k\geq 5$. Now taking the above three estimations into account,
we obtain
$\displaystyle\left|A_{u,0}(1)\right|=\left|\int_{0}^{\infty}e^{-y}(-y)^{\ell}(\frac{1}{n}-y)^{\ell}\cdots(\frac{u}{n}-y)^{\ell-1}\cdots(\frac{k}{n}-y)^{\ell}dy\right|$
$\displaystyle\leq\frac{(k!)^{\ell}}{120^{\ell-1}n^{k\ell-5(\ell-1)}}c(n)^{\ell-1}+2(k+1)\ell
e^{-(k+1)\ell+1}((k+1)\ell-1)^{(k+1)\ell-1}+2e^{-2(k+1)l}(2(k+1)\ell)^{\ell(k+1)-1}$
$\displaystyle\leq\frac{(k!)^{\ell}}{n^{k\ell-5(\ell-1)}}\left(\frac{c(n)}{120}\right)^{\ell-1}+\left(2(k+1)\ell+2\cdot
2^{\ell(k+1)-1}\right)e^{-(k+1)\ell+1}((k+1)\ell)^{\ell(k+1)-1}$
$\displaystyle\leq\frac{(k!)^{\ell}}{n^{k\ell-5(\ell-1)}}\left(\frac{c(n)}{120}\right)^{\ell-1}+3\cdot
2^{(k+1)\ell}e^{-(k+1)\ell+1}((k+1)\ell)^{\ell(k+1)-1}$ $\displaystyle\leq
4\cdot 2^{(k+1)\ell}e^{-(k+1)\ell+1}((k+1)\ell)^{\ell(k+1)-1},\qquad\mbox{ for
}n\geq 2$ $\displaystyle\leq\exp\left(\log 4+(k+1)\ell\log
2+(\ell(k+1)-1)\log((k+1)\ell)-\ell(k+1)+1\right)$
For $k\in\\{2,3,4\\}$ the only thing that changes is the first term, but
because $c(n,k)<1$ for all choices of $k$ and $n$ that interest us, we have
that
$\frac{k!}{n^{k}}c(n,k)^{\ell-1}\leq
2^{(k+1)\ell}e^{-(k+1)\ell+1}((k+1)\ell)^{\ell(k+1)-1},$
so the final bound is also valid in for these values. ∎
Finally, we actually need to get a bound for $A_{u,0}^{\star}(1)$. The
following corollary gives us the desired bound.
###### Corollary 11.
For $k\geq 3$ and $\ell\geq\exp(s(n,k))$, we have
$\left|\frac{n^{L-\ell+1}}{(\ell-1)!}A_{u,0}(1)\right|\leq\exp\left(\ell
k\log\ell+\ell\left[k\log k+k\log n+0.72k+0.000003\right]\right)$
For $k=2$, we have
$\displaystyle\left|\frac{n^{L-\ell+1}}{(\ell-1)!}A_{u,0}(1)\right|\leq\exp(2\ell\log\ell+3.377257\ell+2\ell\log
2).$
###### Proof.
Applying the previous lemma and Stirling’s formula we have
$\displaystyle\log\left|A_{u,0}(1)\right|-\log\left((l-1)!\right)$
$\displaystyle\leq\ell
k\log\ell+\ell\left((k+1)\log(k+1)+\log\ell-\log(\ell-1)-k+(k+1)\log 2\right)$
$\displaystyle-\log\ell-\log(k+1)+\frac{1}{2}\log(\ell-1)+\log
4-\log\sqrt{2\pi}$
We first deal with the case when $k\geq 3$. With this assumption, because
$\log\ell\geq s(n,k)\geq s(2,3)$ we have that
$\log(\ell)-\log(\ell-1)=\int_{\ell-1}^{\ell}\frac{dx}{x}\leq\frac{1}{\ell-1}<0.000003.$
Furthermore,
$(k+1)\log(k+1)-k\log(k)=k\log(1+1/k)+\log(k+1)\leq 1+\log(k+1).$
and
$-k+(k+1)\log 2+1+\log(k+1)<0.72k.$
Additionally, for all $k\geq 2$
$\frac{-\log\ell-\log(k+1)+\frac{1}{2}\log(\ell-1)+\log
4-\log\sqrt{2\pi}}{\ell}<0$
and goes to $0$ as $\ell$ grows.
Thus, multiplying by $n^{L-\ell+1}=n^{k\ell}$ we get
$\left|\frac{n^{L-\ell+1}}{(\ell-1)!}A_{u,0}(1)\right|\leq\exp\left(\ell
k\log\ell+\ell\left[k\log k+k\log n+0.72k+0.000003\right]\right)$
Let us now move to the case $k=2$. We have
$\log(\ell)-\log(\ell-1)=\int_{\ell-1}^{\ell}\frac{dx}{x}\leq\frac{1}{\ell-1}<0.00046.$
For $k=2$ see that
$\displaystyle\log\left|A_{u,0}(1)\right|-\log\left((l-1)!\right)$
$\displaystyle\leq(\log 4-3\ell+3\ell
log(6\ell))-(\ell-1)\log(\ell-1)+(\ell-1)-\frac{1}{2}\log(\ell-1)-\log\sqrt{2\pi}$
$\displaystyle\leq 3\ell\log\ell+\ell\left(3\log 6-3+\frac{\log
4-1}{\ell}-\frac{\ell\log(\ell-1)}{\ell}+1+\frac{\log(\ell-1)}{2\ell}-\frac{\log\sqrt{2\pi}}{\ell}\right)$
$\displaystyle\leq 2\ell\log\ell+\ell\left(3\log 6-2+\frac{\log
4/\sqrt{2\pi}-1}{\ell}+\log(\frac{\ell}{\ell-1})+\frac{\log(\ell-1)}{2\ell}\right)$
$\displaystyle\leq 2\ell\log\ell+3.377\ell.$
Therefore, in this case,
$\displaystyle\left|\frac{n^{L-\ell+1}}{(\ell-1)!}A_{u,0}(1)\right|\leq\exp(2\ell\log\ell+3.377257\ell+2\ell\log
2)$.
∎
## 6\. Integrals corresponding to terms $L_{u,j}^{\star}$
Next we need to get a suitable bound on terms $L_{u,j}^{\star}$, or more
precisely their sum $\sum_{j=1}^{k}\left|L_{u,j}^{\star}\right|$. The
following lemma will be useful
###### Lemma 12.
Let $k\geq 3$ and $n\geq 2$. Then
$\displaystyle\max_{0<x<k/n}\left|x\left(\frac{1}{n}-x\right)\left(\frac{2}{n}-x\right)\cdot\ldots\cdot\left(\frac{k}{n}-x\right)\right|\leq\frac{k!}{6n^{k+1}}.$
If $k=2$, then we have
$\max_{0<x<2/n}|x(\frac{1}{n}-x)(\frac{2}{n}-x)|\leq\frac{2}{3\sqrt{3}n^{3}}$.
###### Proof.
By doing a change of variable $y=nx$ the expression on the left becomes
$\displaystyle\max_{0<x<k/n}\left|x\left(\frac{1}{n}-x\right)\cdot\ldots\cdot\left(\frac{k}{n}-x\right)\right|=\frac{1}{n^{k+1}}\max_{0<y<k}\left|y(1-y)\cdot\ldots\cdot(k-y)\right|.$
By analyzing the function $\left|y(1-y)(2-y)\cdot\ldots\cdot(k-y)\right|$ we
see that its maximum on the interval $(0,k)$ is attained for the first time
already on the interval $(0,1)$. If $k\geq 3$ we have the following
$\displaystyle\max_{0<y<k}\left|y(1-y)\ldots(k-y)\right|$
$\displaystyle\leq\max_{0<y<1}|(4-y)\ldots(k-y)|\cdot\max_{0<y<1}|y(1-y)(2-y)(3-y)|$
$\displaystyle\leq\frac{k!}{3!}\max_{0<y<1}y(1-y)(2-y)(3-y).$
By taking the derivative we can see that the function $y(y-1)(y-2)(y-3)$
achieves its maximum $1$ for $y=(3\pm\sqrt{5})/2$, which finally implies that
$\displaystyle\max_{0<x<k/n}\left|x\left(\frac{1}{n}-x\right)\cdot\ldots\cdot\left(\frac{k}{n}-x\right)\right|\leq\frac{k!}{6n^{k+1}}.$
Similarly for $k=2$ we need to analyze the function $y(y-1)(y-2)$, whose
maximum $\frac{2}{3\sqrt{3}}$ is achieved for $y=1\pm\frac{1}{\sqrt{3}}$, from
which the claim follows. ∎
###### Lemma 13.
Let $k\geq 2$ and $n\geq 2$. Then
$\displaystyle|L_{u,j}^{*}(1)|\leq
n^{L-\ell+1}\frac{(e^{\frac{j}{n}}-1)(k!)^{\ell}}{(\ell-1)!(c(k)n^{k+1})^{\ell-1}n^{k}},$
where $c(k)=6$ for $k\geq 3$ and $c(2)=3\sqrt{3}$.
###### Proof.
Let $j\in\\{1,\ldots,k\\}$. By the definition of $|L_{u,j}^{*}(1)|$ we have
$\displaystyle|L_{u,j}^{*}(1)|(\ell-1)!$
$\displaystyle=n^{L-\ell+1}e^{\frac{j}{n}}\int_{0}^{\frac{j}{n}}e^{-x}\frac{\prod_{r=0}^{k}|\frac{r}{n}-x|^{\ell}}{|\frac{u}{n}-x|}dx$
$\displaystyle=n^{L-\ell+1}e^{\frac{j}{n}}\int_{0}^{\frac{j}{n}}e^{-x}\prod_{r=0}^{k}|\frac{r}{n}-x|^{\ell-1}\frac{\prod_{r=0}^{k}|\frac{r}{n}-x|}{|\frac{u}{n}-x|}dx$
$\displaystyle\leq
n^{L-\ell+1}e^{\frac{j}{n}}\int_{0}^{\frac{j}{n}}e^{-x}\prod_{r=0}^{k}|\frac{r}{n}-x|^{\ell-1}\frac{k!}{n^{k}}dx.$
Because $j\geq k$, we have that
$\max_{0<x<j/n}\prod_{r=0}^{k}|\frac{r}{n}-x|\leq\max_{0<x<k/n}\prod_{r=0}^{k}|\frac{r}{n}-x|$
which is at most $\frac{k!}{c(k)n^{k+1}}$ due to the previous lemma. So we
further have
$\displaystyle|L_{u,j}^{*}(1)|(\ell-1)!$ $\displaystyle\leq
n^{L-\ell+1}e^{\frac{j}{n}}\int_{0}^{\frac{j}{n}}e^{-x}\left(\frac{k!}{c(k)n^{k+1}}\right)^{\ell-1}\frac{k!}{n^{k}}dx$
$\displaystyle\leq
n^{L-\ell+1}\frac{(k!)^{l}}{(c(k)n^{k+1})^{\ell-1}n^{k}}e^{\frac{j}{n}}\int_{0}^{\frac{j}{n}}e^{-x}dx$
$\displaystyle\leq
n^{L-\ell+1}\frac{(k!)^{\ell}}{(c(k)n^{k+1})^{\ell-1}n^{k}}(e^{\frac{j}{n}}-1)$
∎
###### Lemma 14.
Let $k\geq 2$ and $c(k)$ as in the previous lemma. We have
$\sum_{j=1}^{k}|L_{u,j}^{*}|\leq\frac{(k!)^{\ell}}{c(k)^{\ell-1}(\ell-1)!}n^{2-\ell}e^{(k+1)/n}.$
###### Proof.
We have the following:
$\sum_{j=1}^{k}(e^{j/n}-1)<\sum_{j=1}^{k}e^{j/n}=\frac{e^{(k+1)/n}-e^{1/n}}{e^{1/n}-1}<\frac{e^{(k+1)/n}}{e^{1/n}-1}.$
Since
$e^{1/n}-1=\int_{0}^{1/n}e^{x}dx>\frac{1}{n},$
this can be further estimated to
$\sum_{j=1}^{k}(e^{j/n}-1)<ne^{(k+1)/n}.$
By summing up the above estimation for $j=1,\dots,k$ we get
$\sum_{j=1}^{k}|L_{u,j}^{*}|\leq
n^{L-\ell+1}\frac{(k!)^{\ell}}{(c(k)n^{k+1})^{\ell-1}n^{k}(\ell-1)!}ne^{(k+1)/n}.$
We can further simplify the above expression, by noticing that
$\displaystyle\frac{n^{L-\ell+1}n}{(n^{k+1})^{\ell-1}n^{k}}=n^{2-\ell},$
which follows from $L=k\ell-1$. This finally gives us the bound
$\displaystyle\sum_{j=1}^{k}|L_{u,j}^{*}|\leq\frac{(k!)^{\ell}}{c(k)^{\ell-1}(\ell-1)!}n^{2-\ell}e^{(k+1)/n}.$
∎
To make this bound suitable for application in $r(\ell)=\exp(R(\ell)$ we need
to simplify it further.
###### Lemma 15.
Let $k\geq 3$ and $\ell\geq e^{s(k,n)}=e^{(k+n)(\log(k+n))^{2}}$. We have
$\sum_{j=1}^{k}|L_{u,j}^{*}|\leq\exp\left(-\ell\log\ell+\ell(k\log
k-0.81k-\log n+0.174)\right).$
For $k=2$ we have
$\sum_{j=1}^{2}|L_{u,j}^{*}|\leq\exp\left(-\ell\log\ell-0.64\ell\right).$
###### Proof.
First let $k\geq 3$. We need to simplify the following expression:
$\log\left(\frac{(k!)^{\ell}}{6^{\ell-1}(\ell-1)!}n^{2-\ell}e^{(k+1)/n}\right)=\ell\log(k!)-(\ell-1)\log
6-\log(\ell-1)!+(2-\ell)\log n+\frac{k+1}{n}.$
We have $\log(\ell-1)!=\log\ell!-\log\ell$. Further, we can use Stirling’s
formula to bound the factorials:
$\sqrt{2\pi\ell}\left(\frac{\ell}{e}\right)^{\ell}e^{1/(12\ell+1)}<\ell!<\sqrt{2\pi\ell}\left(\frac{\ell}{e}\right)^{\ell}e^{1/(12\ell)}.$
Hence
$\log\ell!>\frac{1}{2}\log(2\pi)+\frac{1}{2}\log\ell+\ell\log\ell-\ell+\frac{1}{12\ell+1}$
and similarly for $k!$:
$\log k!<\frac{1}{2}\log(2\pi)+\frac{1}{2}\log k+k\log k-k+\frac{1}{12k}.$
Hence, we have
$\displaystyle\log\left(\frac{(k!)^{\ell}}{6^{\ell-1}(\ell-1)!}n^{2-\ell}e^{(k+1)/n}\right)=$
$\displaystyle\ell\left(\frac{1}{2}\log(2\pi)+\frac{1}{2}\log k+k\log
k-k+\frac{1}{12k}\right)-(\ell-1)\log 6$
$\displaystyle-\log\ell!+\log\ell+(2-\ell)\log n+\frac{k+1}{n}$
$\displaystyle=$ $\displaystyle\ell\left(\frac{1}{2}\log(2\pi)+\frac{1}{2}\log
k+k\log k-k+\frac{1}{12k}\right)-(\ell-1)\log 6+\log\ell$
$\displaystyle-\left(\frac{1}{2}\log(2\pi)+\frac{1}{2}\log\ell+\ell\log\ell-\ell+\frac{1}{12\ell+1}\right)+(2-\ell)\log
n+\frac{k+1}{n}$ $\displaystyle=$ $\displaystyle-\ell\log\ell+\ell(k\log
k-0.81k-\log n+0.16)+\frac{1}{2}\log\ell$ $\displaystyle+2\log
n+\frac{k+1}{n}+0.88,$
because $\log 6-\frac{1}{2}\log(2\pi)-\frac{1}{12\ell+1}<0.88$ and
$\frac{1}{2}\log(2\pi)-\log 6+\frac{1}{12k}+1<0.16$ and $\frac{1}{2}\log
k-k<-0.81k$. We can further simplify by using the inequality
$\left(\frac{1}{2}\log\ell+2\log n+\frac{k+1}{n}+0.88\right)<0.00004\ell$
to obtain
$\log\left(\frac{(k!)^{\ell}}{6^{\ell-1}(\ell-1)!}n^{2-\ell}e^{(k+1)/n}\right)<-\ell\log\ell+\ell(k\log
k-0.81k-\log n+0.17).$
Similar calculation for $k=2$ gives us explicitly:
$\displaystyle\log\left(\sum_{j=1}^{2}|L_{u,j}^{*}|\right)\leq$
$\displaystyle-\ell\log\ell+\ell(2\log 2+\frac{1}{2}\log
2-2+\frac{1}{24}-\log(3\sqrt{3})+\frac{1}{2}\log(2\pi)+1-\log n)$
$\displaystyle+\log(3\sqrt{3})-\frac{1}{2}\log(2\pi)+\frac{1}{2}\log\ell-\frac{1}{12\ell+1}+2\log
n+\frac{k+1}{n}$ $\displaystyle\leq$
$\displaystyle-\ell\log\ell+\ell(0.0456-\log n)+0.0035\ell$
$\displaystyle\leq$
$\displaystyle-\ell\log\ell-0.65\ell+0.0035\ell\leq-\ell\log\ell-0.64\ell.$
∎
## 7\. Transcendence measure for $e^{1/n}$
We are now ready to put together the bounds
$q(\ell)=\ell k\log\ell+\ell\left[k\log k+k\log n+0.72k+0.000003\right]$
which is true for $k\geq 3$ and for $k=2$
$q(\ell)=2\ell\log\ell+\ell(3.377257+2\log n).$
Estimating sum of $L_{u,j}^{*}$, for $k\geq 3$, we obtained
$-r(\ell)=-\ell\log\ell+\ell(k\log k-0.81k-\log n+0.17)$
and for $k=2$, it is
$-r(\ell)=-\ell\log\ell-0.64\ell.$
Using notation from Section 3 in [3], equations $(8)$ and $(9)$, we have
$\displaystyle s(n,k)$ $\displaystyle=(k+n)(\log(k+n))^{2}\quad\textrm{this
function is in place of $s(m)$ there}$ $\displaystyle a$ $\displaystyle=k$
$\displaystyle b$ $\displaystyle=k\log k+k\log n+0.72k+0.000003$
$\displaystyle c$ $\displaystyle=1$ $\displaystyle d$ $\displaystyle=k\log
k-0.81k-\log n+0.17.$
Now with the notation from Section 3 in [3], equation $(10)$, we have
$\displaystyle B$ $\displaystyle=b+\frac{ad}{c}=k\log k+k\log
n+0.72k+0.000003+k(k\log k-0.81k-\log n+0.17)$ $\displaystyle=k\log
k+0.89k+0.000003+k^{2}\log k-0.81k^{2}$ $\displaystyle C$ $\displaystyle=a=k$
$\displaystyle D$ $\displaystyle=a+b+ae^{-s(k,n)}=k+k\log k+k\log
n+0.72k+0.000003+\frac{k}{e^{(k+n)(\log(k+n))^{2}}}$ $\displaystyle F^{-1}$
$\displaystyle=2e^{D}$ $\displaystyle v$ $\displaystyle=1-\frac{k\log
k-0.81k-\log n+0.17}{(k+n)(\log(k+n))^{2}}$ $\displaystyle n_{1}$
$\displaystyle=e^{(n+k)(\log(n+k))^{2}}.$
Now we have
$|\lambda_{0}+\lambda_{1}e^{1/n}+\cdots+\lambda_{k}e^{k/n}|>F(2H)^{-a/c-\epsilon(H)},$
where
$\epsilon(H)=\frac{1}{\log(2H)}\left(Bz\left(\frac{\log(2H)}{v}\right)+C\log\left(z\left(\frac{\log(2H)}{v}\right)\right)\right).$
The term $H^{-a/c}=H^{-k}$ will form the main term, and everything else will
be put together into the second term in the exponent (of $H$). This second
term will be formed of the terms
$F2^{-a/c}(2H)^{-\epsilon(H)}$
For large $k$, we have
$1<|\Lambda|2(2H)^{\frac{a}{c}}e^{\epsilon(H)\log(2H)+D}=|\Lambda|H^{\frac{a}{c}+Y}=|\Lambda|H^{k+Y},$
where
$\displaystyle Y:$ $\displaystyle=\frac{1}{\log
H}\left(Bz\left(\frac{\log(2H)}{v}\right)+C\log\left(z\left(\frac{\log(2H)}{v}\right)\right)+D+(k+1)\log
2\right)$ $\displaystyle\leq\frac{1}{\log
H}\left(\frac{uB}{v}\frac{\log(2H)}{\log\log(2H)}+k\log\left(\frac{u}{v}\frac{\log(2H)}{\log\log(2H)}\right)+D+(k+1)\log
2\right)$
and
$u=1+\frac{\log(s(n,k))}{s(n,k)}.$
We use the fact that $\log H\geq s(n,k)e^{s(n,k)}$. Not, we need to estimate
the terms involving $D$ and $B$.
$\displaystyle D+(k+1)\log 2$ $\displaystyle=k+k\log k+k\log
n+0.72k+0.000003+\frac{k}{e^{(k+n)(\log(k+n))^{2}}}+(k+1)\log 2$
$\displaystyle=k+k\log k+k\log
n+k\left(0.72+\frac{0.000003}{k}+\frac{1}{e^{(k+n)(\log(k+n))^{2}}}+\log
2+\frac{\log 2}{k}\right)$ $\displaystyle\leq k\log k+k\log n+3.4k.$
For the next term, we observe
$k\log\left(\frac{u}{v}\frac{\log(2H)}{\log\log(2H)}\right)\leq
k\log(u\log(2H)).$
Therefore, we get
$\displaystyle Y$ $\displaystyle\leq\frac{1}{\log
H}\left(\frac{uB}{v}\frac{\log(2H)}{\log\log(2H)}+k\log(u\log(2H))+k\log
k+k\log n+3.4k\right)$ $\displaystyle\leq\frac{1}{\log
H}\left(\frac{uB}{v}\frac{\log(2H)}{\log\log(2H)}+k\log(2\log H)+k\log(2\log
2)+k\log k+k\log n+3.4k\right)$ $\displaystyle\leq\frac{1}{\log
H}\left(\frac{uB}{v}\frac{\log(2H)}{\log\log(2H)}+k\log(2\log
H)+\frac{k}{2}+k\log k+k\log n+3.4k\right)$ $\displaystyle\leq\frac{1}{\log
H}\left(\frac{uB}{v}\frac{\log(2H)}{\log\log(2H)}+\frac{69}{10}k\log(2\log
H)\right)$ $\displaystyle\leq\frac{1}{\log\log H}\left(\frac{\log(2H)}{\log
H}\cdot\frac{uB}{v}+\frac{6.9\log\log H\cdot k\log(2\log H)}{\log H}\right)$
$\displaystyle=\frac{u}{v\log\log H}\left(B+\frac{1}{\log
H}\left(\log(2)B+\frac{v\log\log H\cdot 6.9k\log(2\log H)}{u}\right)\right)$
We have
$Y\leq\begin{cases}&\frac{u}{v\log\log H}\left(B+0.744754115\right),\mbox{ for
}k=2\\\ &\frac{u}{v\log\log H}\left(B+0.04386773\right),\mbox{ for }k=3\\\
&\frac{u}{v\log\log H}\left(B+0.00075786\right),\mbox{ for }k=4\\\
&\frac{u}{v\log\log H}\left(B+0.00000412\right),\mbox{ for }k=5\\\
&\frac{u}{v\log\log H}\left(B+7.976\times 10^{-9}\right),\mbox{ for }k=6\\\
\end{cases}$
For $k\geq 6,$ we observe
$Y\leq\frac{u}{v\log\log H}\left(B+10^{-8}\right).$
For calculating the small values of $k$, we do it case by case.
For $k=2$, take $n=2$ and recall
$\displaystyle b=3.377257+2\log 2$ $\displaystyle d=-0.64$ $\displaystyle
a=2\mbox{ and }d=1$
which implies
$3.878864\mbox{ and }D=7.159781\qquad\frac{u}{v}\leq 1.151906$
We also have $\log H\geq s(2,2)e^{s(2,2)}=7.69\times e^{7.69}$
$\displaystyle Y$ $\displaystyle\leq\frac{1}{\log
H}\left(\frac{uB}{v}\frac{\log(2H)}{\log\log(2H)}+2\log\left(\frac{u}{v}\frac{\log(2H)}{\log\log(2H)}\right)+D+3\log
2\right)$ $\displaystyle\leq\frac{9.202255}{\log\log H}$
Let us now look at the case $k\geq 3$. For simplicity’s sake, write
$Y\leq\frac{u}{v\log\log H}(B+\theta)$
Define
$f(k,n)=\frac{u}{vk^{2}\log
k}\left(B+\theta\right)=\frac{\left(1+\frac{\log((k+n)(\log(k+n))^{2})}{(k+n)(\log(k+n))^{2}}\right)\left(1+\frac{1}{k}+\frac{0.89}{k\log
k}+\frac{0.000003+\theta}{k^{2}\log k}-\frac{0.81}{\log
k}\right)}{1-\frac{k\log k-0.81k-\log n+0.17}{(k+n)(\log(k+n))^{2}}}$
The expression can be further simplified to
$f(k,n)=\frac{\left(1+\frac{1}{(k+n)\log(k+n)}+\frac{2\log\log(k+n)}{(k+n)(\log(k+n))^{2}}\right)\left(1+\frac{1}{k}+\frac{0.89}{k\log
k}+\frac{0.000003+\theta}{k^{2}\log k}-\frac{0.81}{\log
k}\right)}{1-\frac{k\log k-0.81k-\log n+0.17}{(k+n)(\log(k+n))^{2}}}$
Let us look at the numerator. We have
$\left(1+\frac{1}{(k+n)\log(k+n)}+\frac{2\log\log(k+n)}{(k+n)(\log(k+n))^{2}}\right)\left(1+\frac{1}{k}+\frac{0.89}{k\log
k}+\frac{0.000003+\theta}{k^{2}\log k}-\frac{0.81}{\log k}\right)\\\
=1+\frac{1}{k}+\frac{0.89}{k\log k}+\frac{0.000003+\theta}{k^{2}\log
k}-\frac{0.81k}{\log
k}+\frac{1}{(k+n)\log(k+n)}+\frac{1}{k(k+n)\log(k+n)}+\frac{0.89}{k(k+n)\log
k\log(k+n)}\\\ +\frac{0.000003+\theta}{k^{2}\log
k(k+n)\log(k+n)}-\frac{0.81}{(k+n)\log(k+n)\log
k}+\frac{2\log\log(k+n)}{(k+n)(\log(k+n))^{2}}+\frac{2\log\log(k+n)}{k(k+n)(\log(k+n))^{2}}\\\
+\frac{2\cdot 0.89\log\log(k+n)}{k(k+n)\log
k(\log(k+n))^{2}}+\frac{2\cdot(0.000003+\theta)\log\log(k+n)}{k^{2}(k+n)\log
k(\log(k+n))^{2}}-\frac{2\cdot 0.81\log\log(k+n)}{(k+n)(\log(k+n))^{2}\log
k}.$
We can now verify using WolframAlpha that
$\frac{1}{k(k+n)\log(k+n)}+\frac{0.89}{k(k+n)\log
k\log(k+n)}+\frac{0.000003+\theta}{k^{2}\log
k(k+n)\log(k+n)}-\frac{0.81}{(k+n)\log(k+n)\log k}<0$
and
$\frac{2\log\log(k+n)}{k(k+n)(\log(k+n))^{2}}\\\ +\frac{2\cdot
0.89\log\log(k+n)}{k(k+n)\log
k(\log(k+n))^{2}}+\frac{2\cdot(0.000003+\theta)\log\log(k+n)}{k^{2}(k+n)\log
k(\log(k+n))^{2}}-\frac{2\cdot 0.81\log\log(k+n)}{(k+n)(\log(k+n))^{2}\log
k}<0.$
for $k\geq 3$. Furthermore, the denominator can be written as
$1-\frac{k\log
k}{(k+n)(\log(k+n))^{2}}+\frac{0.81k}{(k+n)(\log(k+n))^{2}}+\frac{\log
n}{(k+n)(\log(k+n))^{2}}-\frac{0.17}{(k+n)(\log(k+n))^{2}}\\\ >1-\frac{k\log
k}{(k+n)(\log(k+n))^{2}}+\frac{0.81k}{(k+n)(\log(k+n))^{2}},$
since $\frac{\log
n}{(k+n)(\log(k+n))^{2}}-\frac{0.17}{(k+n)(\log(k+n))^{2}}>0$. We can thus
estimate
$f(n,k)<\frac{1+\frac{1}{k}+\frac{0.89}{k\log
k}+\frac{0.000003+\theta}{k^{2}\log k}-\frac{0.81}{\log
k}+\frac{1}{(k+n)\log(k+n)}+\frac{2\log\log(k+n)}{(k+n)(\log(k+n))^{2}}}{1-\frac{k\log
k}{(k+n)(\log(k+n))^{2}}+\frac{0.81k}{(k+n)(\log(k+n))^{2}}}.$
If $k=3$, we have $f(2,3)<1.145$ and $f(3,3)<1.08$, so $f(2,3)$ gives the
larger value.
For $k=4$, we have $f(2,4)<1.114$, $f(3,4)<1.05$ and $f(4,4)<1$, so $f(2,4)$
yields the largest bound. Estimating
$\frac{1}{(k+n)\log(k+n)}+\frac{2\log\log(k+n)}{(k+n)(\log(k+n))^{2}}<\frac{0.81k}{(k+n)(\log(k+n))^{2}}$
(which is based on the fact that the second term on the left side is at most
$1+\frac{2}{e}<1.74$ and the term on the right side is at least $>1.75$ if
$k\geq 5$ and $n\leq k$), and
$\frac{k\log k}{(k+n)(\log(k+n))^{2}}<\frac{1}{\log k},$
we can further simplify the expression:
$f(n,k)<\frac{1+\frac{1}{k}+\frac{0.89}{k\log
k}+\frac{0.000003+\theta}{k^{2}\log k}-\frac{0.81}{\log
k}+\frac{0.81k}{(k+n)(\log(k+n))^{2}}}{1-\frac{1}{\log
k}+\frac{0.81k}{(k+n)(\log(k+n))^{2}}}$
Now
$\frac{1}{k}+\frac{0.89}{k\log k}+\frac{0.000003+\theta}{k^{2}\log
k}=\frac{1}{\log k}\left(\frac{\log
k}{k}+\frac{0.89}{k}+\frac{0.000003+\theta}{k^{2}\log k}\right)\\\
<\frac{1}{\log k}\left(0.299+0.149+0.0000003\right)<\frac{0.5}{\log k},$
where the terms are estimated using $k\geq 5$. Hence
$f(n,k)<\frac{1-\frac{0.31}{\log
k}+\frac{0.81k}{(k+n)(\log(k+n))^{2}}}{1-\frac{1}{\log
k}+\frac{0.81k}{(k+n)(\log(k+n))^{2}}}=1+\frac{0.69}{\log k-1+\frac{0.81k\log
k}{(k+n)(\log(k+n))^{2}}}<1+\frac{0.69}{\log k-1}$
Hence,
$Y\leq\frac{k^{2}\log k}{\log\log H}f(n,k)<\frac{k^{2}\log k}{\log\log
H}\left(1+\frac{0.69}{\log k-1}\right).$
.
## References
* [1] Borel É. Sur la nature arithmétique du nombre $e$, C. R. Acad. Sci. Paris, 128 (1899), 596–599.
* [2] Ernvall-Hytönen, A.-M., Leppälä, K., & Matala-aho, T. (2015). An explicit Baker-type lower bound of exponential values. Proceedings of the Royal Society of Edinburgh: Section A Mathematics, 145(6), 1153-1182. `doi:10.1017/S0308210515000049`
* [3] Ernvall-Hytönen, A.-M., Matala-aho, T. & Seppälä, L. (2019). On Mahler’s Transcendence Measure for e. Constr Approx 49, 405–444. `https://doi.org/10.1007/s00365-018-9429-3`
* [4] Hata M. Remarks on Mahler’s Transcendence Measure for e, J. Number Theory 54 (1995), 81–92.
* [5] Hermite Ch. Sur la fonction exponentielle, C. R. Acad. Sci. 77 (1873), 18–24, 74–79, 226–233, 285–293.
* [6] Khassa D. S. and Srinivasan S. A transcendence measure for e, J. Indian Math. Soc. 56 (1991), 145–152.
* [7] Mahler, K. (1975). On a paper by A. Baker on the approximation of rational powers of e Tom 27 / 1975 Acta Arithmetica 27, 61-87. `doi:10.4064/aa-27-1-61-87`
* [8] Mahler K. Zur Approximation der Exponentialfunktion und des Logarithmus. Teil I, J. Reine Angew. Math. 166 (1931), 118–136.
* [9] Popken J. Sur la nature arithmétique du nombre e, C. R. Acad. Sci. Paris, 186 (1928), 1505–1507.
* [10] Popken J. Zur Transzendenz von e, Math. Z. 29 (1929), 525–541.
|
# Accurate quantum simulation of molecular ground and excited states with a
transcorrelated Hamiltonian
Ashutosh Kumar<EMAIL_ADDRESS>Theoretical Division, Los Alamos National
Laboratory, Los Alamos, NM 87545, USA Ayush Asthana Department of Chemistry,
Virginia Tech, Blacksburg, VA 24061, USA Conner Masteran Department of
Chemistry, Virginia Tech, Blacksburg, VA 24061, USA Edward F. Valeev
Department of Chemistry, Virginia Tech, Blacksburg, VA 24061, USA Yu Zhang
<EMAIL_ADDRESS>Theoretical Division, Los Alamos National Laboratory, Los Alamos,
NM 87545, USA Lukasz Cincio Theoretical Division, Los Alamos National
Laboratory, Los Alamos, NM 87545, USA Sergei Tretiak Theoretical Division,
Los Alamos National Laboratory, Los Alamos, NM 87545, USA Center for
Integrated Nanotechnologies, Los Alamos National Laboratory, Los Alamos, NM
87545, USA Pavel A. Dub<EMAIL_ADDRESS>Chemistry Division, Los Alamos
National Laboratory, Los Alamos, NM 87545, USA
###### Abstract
NISQ era devices suffer from a number of challenges like limited qubit
connectivity, short coherence times and sizable gate error rates. Thus,
quantum algorithms are desired that require shallow circuit depths and low
qubit counts to take advantage of these devices. We attempt to reduce quantum
resource requirements for molecular simulations on a quantum computer, a
promising application on NISQ devices, while maintaining the desired accuracy
with the help of classical quantum chemical theories of canonical
transformation and explicit correlation. In this work, compact ab initio
Hamiltonians are generated classically through an approximate similarity
transformation of the Hamiltonian with a) an explicitly correlated two-body
unitary operator with generalized pair excitations that remove the Coulombic
electron-electron singularities from the Hamiltonian and b) a unitary one-body
operator to efficiently capture the orbital relaxation effects required for
accurate description of the excited states. The resulting transcorelated
Hamiltonians are able to describe both ground and excited states of molecular
systems in a balanced manner. Using the fermionic-ADAPT-VQE method based on
the unitary coupled cluster with singles and doubles (UCCSD) ansatz and only a
minimal basis set (ANO-RCC-MB), we demonstrate that the transcorrelated
Hamiltonians can produce ground state energies comparable to the much larger
cc-pVTZ basis. This leads to a potential reduction in the number of required
CNOT gates by more than three orders of magnitude for the chemical species
studied in this work. Furthermore, using the qEOM formalism in conjunction
with the transcorrelated Hamiltonian, we reduce the errors in excitation
energies by an order of magnitude. The transcorrelated Hamiltonians developed
here are Hermitian and contain only one- and two-body interaction terms and
thus can be easily combined with any quantum algorithm for accurate electronic
structure simulations.
## I Introduction
In 1982, Feynman envisioned the idea of simulating quantum mechanical
processes occurring in nature through devices that operate on the principles
of quantum mechanics themselves [1]. Since then, a lot of important progress
has been made in the development of such quantum mechanical devices, also
referred to as quantum computers as they promise a near-exponential speed-up
(“quantum advantage”) over classical computers for a wide variety of
computational tasks[2]. Solving the many-body electronic Schrödinger equation
is quite naturally one of the most promising applications for quantum
computers. A number of algorithms based on quantum phase estimation (QPE)[3,
4], adiabatic state preparation[5, 6], variational quantum optimization[7, 8]
have been developed and refined to calculate the ground and low-lying excited
states of the many-body systems with the aim of realizing the promised
“quantum advantage” in the field of electronic structure theory. The
contemporary quantum hardware, however, is still in its infancy and suffers
from a variety of challenges, such as limited qubit connectivity, short
coherence times and sizable gate error rates. Furthermore, the mostly one-to-
one correspondence between spin-orbitals and qubits ensures that the quantum
simulations can only utilize a minimal number of qubits which can, at most,
only give a qualitative description of the desired solution. Thus, a lot of
efforts lately have focused on reducing the quantum resource requirements for
electronic structure simulations on noisy intermediate-scale quantum (NISQ)
devices[9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19]. A majority of these
efforts utilize the variational quantum eigensolver (VQE) algorithm in
conjunction with unitary coupled-cluster based ansatzes[8, 7, 20, 21] and are
able to produce highly compact quantum circuits through a variational
minimization of the expectation value of the Hamiltonian with respect to the
circuit parameters, and hence by construction, are more suited for
contemporary quantum hardware. Some specific examples include development of
adaptive ansatzes for simulation of ground[9, 10, 11] and excited-states[12,
13], correlation informed permutation of qubits (PermVQE)[14] or qubits
clustering (ClusterVQE)[19] approaches, construction of highly compact
molecular Hamiltonians through a basis-set free formalism[15] utilizing pair-
natural orbital (PNO) based compression [22, 23] in conjunction with multi-
resolution[24] strategies, low rank factorization techniques for approximating
operators[16]. In a similar work, Bauman et al. employed the QPE algorithm and
double unitary coupled-cluster (DUCC) formalism to downfold or embed many-body
correlation effects into active-spaces of effective Hamiltonians for both
ground[17] and excited states[18]. The explicitly correlated theories[25, 26,
27, 28], which are routinely used in classical electronic structure
calculations to accelerate the convergence of electronic energies and other
molecular properties with respect to the size of basis sets, is another
attractive approach, which has the potential to significantly reduce the
computational resources required for accurate quantum simulations. It is well
known that the traditional many-body wavefunctions generated from a
superposition of single slater determinants which are nothing but an
antisymmetrized product of one-electron orbitals fail to capture short-range
dynamic correlation effects efficiently.[26]. The singularity of the Coulombic
electron-electron interactions near the coalescent point introduces cusps[29]
in the wavefunction, which can only be described accurately by using a large
number of one-electron basis functions. The explicitly correlated methods
alleviate this problem through an explicit parametrization of the wavefunction
in terms of inter-electronic distances and are hence referred to as “R12” or
“F12” methods. Other F12-based formalisms have also been developed which focus
on removing the singularities in the Hamiltonian itself[30, 31, 32]. In the
transcorrelated Hamiltonian approach originally introduced by Boys and
Handy[30] with later improvements by Ten-no[31] and Luo[32], singularity-free
Hamiltonians are generated through a similarity transformation of the
Hamiltonian with a geminal correlation operator $\hat{A}$,
$\hat{H}\to\hat{\bar{H}}=e^{-\hat{A}}\hat{H}e^{\hat{A}}\quad.$ (1)
Yanai and Shiozaki[33] utilized ideas from both the canonical transformation
theory[34, 35] and the transcorrelated approach to construct canonical
transcorrelated (CT-F12) Hamiltonians, where unlike the previous works on the
transcorrelation theory, they used a unitary geminal operator which ensures
that the transformed Hamiltonian is Hermitian, making the formalism quite
robust and easy to use. In an earlier work by some of us[36], the CT-F12
Hamiltonian, in conjunction with the UCCSD ansatz and VQE algorithm was able
to produce near cc-pVTZ quality ground state correlation energies of several
small molecular species, with the much smaller 6-31g basis. McArdle and
Tew[37] and more recently, Sokolov and co-workers[38], also employed the
transcorrelated approach to improve the accuracy of quantum simulations, where
they had to make use of the imaginary-time evolution algorithms due to the
non-Hermiticity of their transformed Hamiltonian. In another application of
explicit correlation strategies in quantum computing, Schleich and co-
workers[39] recently employed the $[2]_{R12}$ formalism developed by Valeev
and Torheyden[40], where one- and two-body reduced density matrices obtained
from the quantum simulation was used to formulate a correction to the energy.
However, these a posteriori corrections have only been developed for ground-
state correlation energies, while a priori strategies like CT-F12 can be
potentially combined with any many-body quantum theory to calculate different
molecular properties corresponding to both ground and excited states.
In this work, we look to extend the CT-F12 formalism to the simulation of
molecular excited states as many interesting chemical phenomena in nature
involve excited states in one way or the other. Preliminary investigations
with the CT-F12 Hamiltonian revealed a very strong bias towards the ground-
state and the basis-set convergence of excited-state properties like
excitation energies was noticeably slower than the regular Hamiltonian itself.
Similar observations were also noticed in the framework of explicitly
correlated coupled-cluster response theory [41]. One reason for this
unbalanced description of the ground and excited states can be attributed to
the absence of virtual orbitals in the definition of the geminal operator.
This is quite obvious in the case of valence excited states where an accurate
simulation would require the inclusion of dynamic correlation effects between
an electron in occupied and virtual orbital in the effective Hamiltonian. For
example, the ${}^{1}P(2p\leftarrow 2s)$ state of the Be atom would be very
poorly described with the above formalism as both the occupied orbitals (1s
and 2s) are of S symmetry and hence the resulting pair won’t contribute at all
to the given excited state[42]. Furthermore, the basis-set convergence of the
energies of Rydberg-like excited states are dominated by non-dynamical
electron correlation effects, and one needs to incorporate orbital relaxations
in the effective Hamiltonian for their accurate description.
In this work, we have added pairs involving virtual orbitals in the definition
of the geminal operator along with the introduction of a singles operator in
the similarity transformation procedure. It should be noted that Watson et
al.[43] had implemented a similar approach for accurate molecular computations
on a classical computer with a minimal basis. In this paper, we develop an
improved recipe for obtaining the singles operator (see sec. II for details)
and apply the method to achieve a massive reduction in quantum resources for
the molecular ground and excited state calculations on a quantum computer.
Within the framework of VQE algorithms, different strategies have been
developed to compute the energies of the low-lying excited states of molecular
species[7, 44]. State-specific methods, that compute one excited state at a
time, include folded spectrum method [7] and orthogonality constrained VQE
method[44]. A more general approach, Quantum Krylov subspace expansion
methods[45, 46, 47, 48] diagonalize the Hamiltonian in a small subspace and
can provide a number of low-lying excited states together. In addition,
subspace expansion based on excited determinants has been proposed[49, 50],
including equation of motion (EOM) operator based qEOM method[51] that
provides a size-intensive approach for excitation energy calculations. In
order to demonstrate the advantage of transcorrelated methods in molecular
simulations on a quantum computer, we use a classical simulator to compute the
ground state energies through fermionic-ADAPT-VQE[9] and excitation energies
using the formalism of qEOM implemented on top of the ground state
calculation.
This manuscript is organized as follows: Section II introduces the formalism
of the canonical transcorrelation procedure. Computational details are
provided in Sec. III. In Sec. IV, we assess the accuracy of the
transcorrelated Hamiltonian in calculations of ground and excited states
energies of a number of small molecular species and estimate the potential
reduction in quantum resources with this approach. We give a summary of our
findings in Sec. V.
## II Theory
The molecular Hamiltonian in the second-quantized formalism can be written as,
$\hat{H}=h_{\nu}^{\mu}\hat{E}_{\mu}^{\nu}+\frac{1}{2}g_{\nu\kappa}^{\mu\lambda}\hat{E}_{\mu\lambda}^{\nu\kappa}$
(2)
where $h_{\nu}^{\mu}$ and $g_{\nu\kappa}^{\mu\lambda}$ refer to the one and
two-electron elements of the Hamiltonian and $\mu$, $\nu$, $\kappa$, $\lambda$
indices refer to the orbitals in the infinite orbital basis. Please refer to
Fig. 1 for a detailed description of the orbital spaces along with their
labels, used in this work. Here, $\hat{E}_{\mu}^{\nu}$ is the spin-free or
spin-summed excitation operator,
$E_{\mu}^{\nu}=a_{\mu\sigma}^{\dagger}a_{\nu\sigma}$, where
$a_{\mu\sigma}^{\dagger}$ and $a_{\mu\sigma}$ ($\sigma\in\\{\alpha\beta\\}$)
are the usual creation and annihilation operators respectively, with $\sigma$
referring to the the spin label. We have followed the Einstein summation
convention throughout this work.
Figure 1: Schematic notation used for different orbital spaces. OBS here
refers to the finite orbital basis while CABS is the complementary auxiliary
basis set, the orthogonal complement to the OBS space.
The canonical transcorrelated (CT) theory[33] aims to incorporate the missing
dynamic electron correlation effects into an effective Hamiltonian through a
similarity transformation,
$\hat{\bar{H}}=e^{\hat{A}^{\dagger}}\hat{H}e^{\hat{A}}$, where $\hat{A}$ is
usually an anti-Hermitian many-body operator which makes
$e^{\hat{A}^{\dagger}}$ and $e^{\hat{A}}$ operations unitary, thus maintaining
the Hermiticity of the Hamiltonian. Utilizing the Baker–Campbell–Hausdorff
(BCH) expansion, the effective Hamiltonian can be expressed in terms of
(nested) commutators,
$\hat{\bar{H}}=\hat{H}+[\hat{H},\hat{A}]+\frac{1}{2!}[[\hat{H},\hat{A}],\hat{A}]+\ldots.$
(3)
The CT theory introduces additional approximations of a) restricting the above
expansion to only double commutators and b) approximating the full Hamiltonian
($\hat{H}$) by its mean-field constituent, the Fock operator($\hat{F}$) in the
double commutator term,
$\hat{\bar{H}}\approx\hat{H}+{[\hat{H},\hat{A}]}_{1,2}+\frac{1}{2}{{[[\hat{F},\hat{A}]}_{1,2},\hat{A}]}_{1,2}\quad.$
(4)
The resulting CT-F12 Hamiltonian is correct at least through second-order in
perturbation[52]. It should be noted that similar approximations have also
been used in CC-F12 theories as well[53]. The notation $[..]_{1,2}$ means only
one- and two-body operators generated from the given commutator are retained
directly while the three-body operators are included through an approximate
decomposition into one- and two-body operators using the extended normal
ordering approach of Mukherjee and Kutzelnigg[54, 55, 56]. In spin orbitals
representation, the full decomposition can be written as[57],
$\displaystyle a_{stu}^{pqr}\hskip 0.72229pt=$
$\displaystyle\tilde{a}_{stu}^{pqr}+9\left(D_{s}^{p}\wedge
a_{tu}^{qr}\right)-36\left(D_{s}^{p}\wedge D_{t}^{q}\wedge D_{u}^{r}\right)$
(5) $\displaystyle+9\left(D_{st}^{pq}\wedge
a_{u}^{r}\right)+24\left(D_{s}^{p}\wedge
D_{t}^{q}D_{u}^{r}\right)-9\left(D_{st}^{pq}\wedge D_{u}^{r}\right)$
$\displaystyle+\lambda^{pqr}_{stu},$
where $\wedge$ denotes antisymmetrization over all upper and lower indices[57]
with the corresponding prefactor of $(\frac{1}{n})^{2}$, where $n$ is the
particle rank of the original undecomposed operator (n = 6 for three-body
decompositions). In this work, the first term $\tilde{a}_{stu}^{pqr}$, which
is the three-body fluctuation operator in normal ordered form with respect to
a reference and the last term $\lambda^{pqr}_{stu}$ which refers to the three-
body density cumulants, have been dropped from the above decomposition. The
final spin-free equations for this approximate decomposition can be found in
Ref.[56]. The mean-field one-body Fock operator in equation 3. is defined as,
$\begin{split}&\hat{F}=f_{\nu}^{\mu}\hat{E}_{\mu}^{\nu},\\\
&f_{\nu}^{\mu}=h_{\nu}^{\mu}+D_{\kappa}^{\lambda}\left(g_{\nu\kappa}^{\mu\lambda}-\frac{1}{2}g_{\kappa\nu}^{\mu\lambda}\right)\end{split}$
(6)
where,
$D_{\nu}^{\mu}=\left\langle\Psi_{0}\left|\hat{E}_{\nu}^{\mu}\right|\Psi_{0}\right\rangle$
is the one-body reduced density matrix associated with the reference
wavefunction $\Psi_{0}$. We have used the following form of the transformation
operator $\hat{A}$ in this work,
$\begin{split}&\hat{A}=\hat{A}^{\mathrm{F}12}+\hat{S^{\prime}},\\\
&\hat{A}^{\mathrm{F}12}=\frac{1}{2}G_{pq}^{\alpha\beta}\left(\hat{E}_{pq}^{\alpha\beta}-\hat{E}_{\alpha\beta}^{pq}\right),\\\
&\hat{S^{\prime}}=G_{p}^{\alpha}\left(\hat{E}_{p}^{\alpha}-\hat{E}_{\alpha}^{p}\right).\end{split}$
(7)
The amplitudes corresponding to the geminal operator $\hat{A}^{\mathrm{F}12}$
is defined as,
$G_{pq}^{\alpha\beta}=\frac{3}{8}\left\langle\alpha\beta\left|\hat{Q}_{12}\hat{F}_{12}\right|pq\right\rangle+\frac{1}{8}\left\langle\alpha\beta\left|\hat{Q}_{12}\hat{F}_{12}\right|qp\right\rangle.$
(8)
Here, we have made use of the SP ansatz[58, 59] of Ten-no where the geminal
amplitudes (1/8, 3/8) are fixed and are obtained by satisfying the first-order
cusp conditions for the singlet and triplet electron pairs respectively. We
also chose a slater-type geminal (STG) as the two-body correlation factor,
$\hat{F}_{12}(r_{12})=-\gamma^{-1}\exp\left(-\gamma r_{12}\right),$ (9)
where $\gamma$ is a scale-length parameter whose values are in practice tuned
to a given orbital basis set[60]. The strong orthogonality projector
$\hat{Q}_{12}$,
$\hat{Q}_{12}=1-\hat{V}_{1}\hat{V}_{2},$ (10)
where $\hat{V}_{i}$ projects the one-electron states into virtual orbitals
(a,b) of the orbital basis set (OBS), ensures that geminal matrix elements
involving products of virtual orbitals like $\langle
ab|\hat{Q}_{12}\hat{F}_{12}|pq\rangle=0$. Thus, all the geminal matrix
elements considered here contain at least one external (CABS) index.
In an earlier work by some of us[36], only occupied-occupied pairs were
included in the definition of the geminal operator. This made the
transcorrelated Hamiltonian very biased towards the ground state wavefunction.
The orbital pairs involving the virtual orbitals, especially the chemically
important ones (usually defined in active spaces) can contribute very
significantly to the excited state wavefunction and thus one needs to satisfy
the cusp conditions for such pairs as well. Thus, we included the occupied-
virtual and virtual-virtual geminal pairs as well in this work, so that our
Hamiltonian can treat multiple states at an equal footing. This approach is
very common in quasi-degenerate perturbation theory[61] where a multi-
configuration reference wavefunction is deployed instead of the regular
Hartree-Fock wavefunction to remove any kind of biasedness from the
Hamiltonian. However, not all virtual orbitals are equally important and one
can always generate an active space instead. One way to do this is to look at
the eigenvalues (also called occupation number) of the virtual-virtual block
of the MP2 one-body reduced density matrix[41]. and choose only those natural
virtual orbitals with occupation numbers greater than a given threshold.
Furthermore, we have also added a singles operator in the similarity
transformation procedure to incorporate orbital relaxation effects in the
Hamiltonian. In the equation of motion based formalisms, the excited state
wavefunction is often characterized by dominant contributions from the singles
excitation operator (for example $R^{a}_{i}$ amplitudes in EOM-CCSD) and thus
addition of quality singles amplitudes is essential. In order to define our
singles operator, we look towards the “CABS singles” approach usually employed
in the explicit correlation theory to accelerate the basis set convergence of
the energy of the reference wavefunction by allowing for orbital rotations
between the occupied and the missing virtual space i.e. the CABS space.
Following the works of Valeev and Kong[26], these amplitudes have been
determined from the Rayleigh-Schrödinger perturbation theory, with the
following partitioning of the Hamiltonian,
$\hat{H}^{(0)}=\left(\begin{array}[]{ccc}F_{j}^{i}&0&0\\\
0&F_{b}^{a}&F_{b}^{x}\\\ 0&F_{y}^{a}&F_{y}^{x}\end{array}\right),\hskip
3.61371pt\hat{H}^{(1)}=\left(\begin{array}[]{ccc}0&F_{i}^{a}&F_{i}^{x}\\\
F_{a}^{j}&0&0\\\ F_{y}^{j}&0&0\end{array}\right)$ (11)
where $F^{i}_{j}$, $F^{a}_{b}$, $F^{x}_{i}$ and $F^{y}_{x}$ refer to the
occupied-occupied, virtual-virtual, occupied-CABS and CABS-CABS block of the
Fock matrix respectively. In this approach, the occupied-CABS ($G^{x}_{i}$)
block of the singles amplitudes can be obtained by solving the following
equation,
$F^{j}_{i}G^{x}_{j}-F^{x}_{y}G^{y}_{i}=F^{x}_{i}.$ (12)
We chose the perturbative formulation of the “CABS singles” approach in this
work as it has been shown to work quite well for small basis sets[26]. One can
easily extend the above equation to solve for the virtual-CABS ($G^{x}_{a}$)
component of the singles amplitudes as well but this naive approach often
leads to very poor results as the resulting equations are generally very
poorly conditioned due to the small differences in orbital energies
($F^{\mu}_{\mu}$) between virtual and CABS orbitals. In order to overcome
this, we replace the virtual-virtual block of the Fock matrix by a constant
parameter $\epsilon$ which we denote as “shift” in this work,
$\epsilon*G^{x}_{a}-F^{x}_{y}G^{y}_{a}=F^{x}_{a}.$ (13)
This shift parameter is in practice very close to the HOMO energy but can be
adjusted for a given molecule and basis set to add optimal orbital relaxation
effects in the transcorrelated Hamiltonian. This concept is very similar to
the regularization strategies employed in the single reference-based
perturbation theories to treat multi-reference problems[62]. In other
approximations, following Shiozaki and Yanai,[33] $G^{cx}_{ab}$ types of
geminal amplitudes have been ignored as incorporating them would require
addition of full three body operators and not the approximated ones used in
this work.
Finally, the transcorrelated Hamiltonian has the following form,
$\hat{\bar{H}}=\bar{h}_{q}^{p}\hat{E}_{p}^{q}+\frac{1}{2}\bar{g}_{rs}^{pq}\hat{E}_{pq}^{rs},$
(14)
where $\bar{h}_{q}^{p}$ and $\bar{g}_{rs}^{pq}$ are the “perturbed” one- and
two-body interaction terms. The transcorrelated Hamiltonian ($\hat{\bar{H}}$)
is Hermitian but its two-body interaction term has a lower symmetry compared
to the original Hamiltonian:
$\overline{g}^{pq}_{rs}\neq\overline{g}^{ps}_{rq}$ while
$g^{pq}_{rs}=g^{ps}_{rq}$. The construction of the transcorrelated Hamiltonian
(with fully factorized equations) scale as $\mathcal{O}$($N^{6}$), where $N$
is the number of orbitals, with a quadratic dependence on the size of the CABS
basis when approach C[63] is used to evaluate the intermediate B.
## III Computational details
The second quantized expressions required to construct the transcorrelated
Hamiltonian were derived using a python module[64] that automates the
evaluation of single and double commutators of the Hamiltonian with excitation
and de-excitation operators of different orders using Wick’s theorem. The
resulting expressions were automatically converted to the einsum tensor
contraction routines from the numpy[65] library to generate the final
transcorrelated Hamiltonian. All the integrals and intermediates (V, X and
B)[26] were calculated using the MPQC[66] software. We used the approach C to
evaluate the matrix elements of the B intermediate.
In this study, we consider four small hydrogen-containing molecules: H2
(r(H-H) = 0.7 Å), LiH (r(Li-H) = 1.5957 Å), H2O (r(O-H) = 0.957 Å, $\theta$
(H-O-H) = 104.5°) and NH3 (r(N-H) = 1.09 Å, $\theta$ (H-N-H) = 109.427°).
6-31g basis set was used as the OBS in the calculations involving the H2
molecule while the ANO-RCC-MIN[67] basis set was employed for all others. In
the canonical transcorrelation procedure, aug-cc-pVTZ-OptRI[68] basis set was
used as the CABS basis for all the molecules studied in this work except the
LiH molecule where cc-pVDZ-F12-OptRI[69] basis set was used instead due to the
non-availability of the former. All the transcorrelated calculations utilized
the CABS+ approach[70]. Since, optimized $\gamma$ values are only available
for larger basis sets, we chose those $\gamma$ values for a given molecule and
basis set which produced ground state energies in the close vicinity of the
cc-pVTZ value. Similarly, those values of the shift parameter were chosen
which minimized the maximum deviation of the excitation energies from the
reference values. For a detailed analysis, we compare the performance of six
types of Hamiltonians in this work. $\hat{H}$ refers to the regular
untransformed Hamiltonian while $\hat{H}_{\text{F12}}^{(ij)}$ and
$\hat{H}_{\text{F12}}^{(pq)}$ refer to the transcorrelated Hamiltonians
generated by the doubles geminal operator defined by occupied-occupied and
all-all orbital pairs respectively. Consequently, addition of the singles
operator in the similarity transformation procedure leads to S’ +
$\hat{H}_{\text{F12}}^{(ij)}$ and S’ + $\hat{H}_{\text{F12}}^{(pq)}$
transcorelated Hamiltonians. Finally, S’ + $\hat{H}$ is obtained from the
similarity transformation of the regular Hamiltonian with the singles operator
only. The ground state energies associated with all these Hamiltonians were
calculated using the fermionic-ADAPT-VQE method (implemented clasically[71])
with Jordan-Wigner mapping in conjunction with the UCCSD ansatz. Excitation
energies were calculated using the qEOM formalism[51] on top of the ground
state calculations. The reference cc-pVTZ ground and excited state energies
were calculated classically using CCSD and EOM-CCSD methods respectively using
the PySCF software[72]
## IV Results and Discussions
We test the performance of the transcorrelated Hamiltonians by doing quantum
simulations of both ground and excited states on a number of small hydrogen
containing molecules: H2, LiH, H2O and NH3 using 6-31g and the minimal basis
set ANO-RCC-MB. We compare the ground state energies and excitation energies
of the few lowest-lying excited states of these molecules with the
corresponding values obtained with a much larger and more accurate basis set
cc-pVTZ.
Table 1: Deviations obtained in the ground state energy (mEh) and the excitation energies (eV) of the four lowest-lying excited states of the H2 molecule using the 6-31g basis set from the reference cc-pVTZ values (second column) for the six different Hamiltonians (see text for details). Parameters used: ($\gamma$, shift) = (0.7, -0.4) States | Reference | $\hat{H}$ | $\hat{H}_{\text{F12}}^{(ij)}$ | $\hat{H}_{\text{F12}}^{(pq)}$ | S’ + $\hat{H}_{\text{F12}}^{(ij)}$ | S’ + $\hat{H}_{\text{F12}}^{(pq)}$ | S’ + $\hat{H}$
---|---|---|---|---|---|---|---
| cc-pVTZ | | | 6-31G | |
S0 | -1.17101 | 20.8 | 6.8 | 3.3 | 2.5 | -1.6 | 16.0
T1 | 11.30 | 0.08 | 0.46 | 0.50 | 0.24 | 0.25 | -0.13
S1 | 13.89 | 1.76 | 2.14 | 1.45 | 0.26 | -0.15 | -0.11
T2 | 15.25 | 7.84 | 8.23 | 8.15 | -0.23 | -0.31 | -0.60
S2 | 17.60 | 10.75 | 11.14 | 10.43 | 1.01 | 0.17 | 0.62
### IV.1 H2
Table 1 lists the deviations obtained in the ground state energy (mEh) and the
excitation energies (eV) of the four lowest-lying excited states (arranged in
the order of increasing energies) of the H2 molecule using 6-31g basis set for
all the six types of Hamiltonians considered in this work, from the reference
cc-pVTZ values (second column). Si and Ti symbols in the first column refer to
the $i^{th}$ singlet and triplet states respectively. The values of $\gamma$
and shift parameters used in these calculations were 0.7 and -0.4
respectively. The ground state energy with the regular Hamiltonian is around
21 mEh away from the reference value. While all the transcorrelated
Hamiltonians bring this difference down, the ones with the singles operator in
the similarity transformation procedure, S’ + $\hat{H}_{\text{F12}}^{(ij)}$
and S’ + $\hat{H}_{\text{F12}}^{(pq)}$ perform the best with deviations of
only 2.5 mEh and -1.6 mEh respectively. Thus, the S’ +
$\hat{H}_{\text{F12}}^{(pq)}$ Hamiltonian produces even better energies than
the reference cc-pVTZ values and is in fact identical to the total energy
obtained using the cc-pVQZ basis, which is quite close to the complete basis
limit (CBS). Looking at the excitation energies, the
$\hat{H}_{\text{F12}}^{(ij)}$ Hamiltonian gives larger deviations than the
regular Hamiltonian itself. The deviations grow sharply as we move towards the
higher energy excited states, with a max deviation of 11.14 eV observed for
the S2 excited state. This clearly shows that an unbalanced description of the
ground and excited states is obtained when geminal operators are defined with
only occupied-occupied orbital pairs. On adding all the orbital pairs in the
geminal operator ($\hat{H}_{\text{F12}}^{(pq)}$), the max deviation is lowered
but only slightly to 10.43 eV. Adding the singles operator in the similarity
transformation on the other hand has quite a dramatic effect with a max
deviation of 1.01 eV and 0.31 eV for the S’ + $\hat{H}_{\text{F12}}^{(ij)}$
and S’ + $\hat{H}_{\text{F12}}^{(pq)}$ Hamiltonians respectively. For the
triplet excited states (T1, T2), the S’ + $\hat{H}_{\text{F12}}^{(ij)}$
Hamiltonian performs slightly better over its $pq$ counterpart but the
magnitude of improvement in excitation energies for these states are only 0.01
eV and 0.08 eV respectively. It should be noted that the correction due to the
S’ + $\hat{H}_{\text{F12}}^{(ij)}$ Hamiltonian can be seen to have a state-
specific nature. It will always be biased towards those excited states with
dominant contributions from configurations with the “regular” occupied pairs.
On the other hand, the S’ + $\hat{H}_{\text{F12}}^{(pq)}$ Hamiltonian is more
reliable as it also takes care of the missing dynamic correlation effects due
to electrons in occupied-virtual and virtual-virtual orbital pairs. Thus, it
was able to reduce the max deviation in excitation energies from 1.01 eV (S’ +
$\hat{H}_{\text{F12}}^{(ij)}$) to 0.31 eV as seen in table 1. From these
results, it’s quite clear that the S’ + $\hat{H}_{\text{F12}}^{(pq)}$
Hamiltonian should be preferred for accurate simulation of both ground and
excited states.
Furthermore, one can easily see that the basis set convergence of the energies
of these excited states are not dominated by dynamical electron correlation
effects and hence can’t be captured by the geminal operators alone. The last
column of the table illustrates this even more clearly where addition of just
the singles operator to the regular Hamiltonian (S’ + $\hat{H}$) lowers the
max deviation from 17.6 eV (regular Hamiltonian) to 0.62 eV. However, due to
the lack of transcorrelation procedure, the deviations in the ground state
energy for this Hamiltonian still remains quite high at around 16 mEh compared
to the 21 mEh obtained using the regular Hamiltonian. Thus, the quality of the
ground state wavefunction remains to be poor for the S’ + $\hat{H}$
Hamiltonian even though the differences in the ground and excited state
energies are quite accurate due to error cancellations.
Table 2: Deviations obtained in the ground state energy (mEh) and the excitation energies (eV) of the six lowest-lying excited states of the LiH molecule using the ANO-RCC-MIN basis set from the reference cc-pVTZ values (second column) for the six different Hamiltonians (see text for details). The superscript in the first column indicates the degeneracy of the given excited state. Parameters used: ($\gamma$, shift) = (0.7, -0.4) States | Reference | $\hat{H}$ | $\hat{H}_{\text{F12}}^{(ij)}$ | $\hat{H}_{\text{F12}}^{(pq)}$ | S’ + $\hat{H}_{\text{F12}}^{(ij)}$ | S’ + $\hat{H}_{\text{F12}}^{(pq)}$ | S’ + $\hat{H}$
---|---|---|---|---|---|---|---
| cc-pVTZ | | | ANO-RCC-MB | |
S0 | -8.02230 | 29.3 | 15.1 | 15.0 | 8.9 | 8.4 | 27.1
T1 | 3.26 | -0.52 | -0.14 | -0.12 | -0.18 | -0.16 | -0.64
S1 | 3.62 | -0.46 | -0.08 | -0.10 | -0.09 | -0.11 | -0.54
T2(2) | 4.24 | -0.43 | -0.04 | -0.08 | 0.00 | -0.04 | -0.48
S2(2) | 4.61 | -0.47 | -0.07 | -0.14 | -0.09 | -0.15 | -0.57
T3 | 5.77 | 2.02 | 2.41 | 2.33 | 0.24 | 0.14 | 0.20
S3 | 6.37 | 7.06 | 7.46 | 6.44 | 1.65 | 0.55 | 1.17
### IV.2 LiH
Table 2 tabulates the deviation in ground state and excitation energies of six
lowest-lying excited states of the LiH molecule in an identical layout as the
H2 molecule using the same values of $\gamma$ and shift parameters. Here, we
have used a minimal basis set ANO-RCC-MB and froze the core electrons in both
ground and excited state simulations in order to lower the number of required
qubits. The parenthesis in the first column indicates the degeneracy of the
excited state. For example, the second singlet and triplet excited states (S2,
T2) are doubly degenerate. For the ground state energy, just like before, the
transcorrelated Hamiltonians with the singles operator yield the lowest
deviations of 8.9 mEh (S’ + $\hat{H}_{\text{F12}}^{(ij)}$) and 8.4 mEh (S’ +
$\hat{H}_{\text{F12}}^{(pq)}$) respectively from the reference value. Adding
all the orbital pairs to the geminal operator and addition of the singles
operator improves the ground state energy the most. However, unlike the H2
molecule, the cc-pVDZ-F12-OptRI basis set was used as the CABS basis in the
generation of the transcorrelated Hamiltonian due to the non-availability of
the aug-cc-pVTZ-OptRI basis set for Li. We observed that the performance of
the cc-pVDZ-F12-OptRI basis is not as optimal compared to the aug-cc-pVTZ-
OptRI basis set when ANO-RCC-MB basis set is used as the OBS. Thus, these
deviations can be reduced even further with the help of an optimized CABS
basis for minimal basis sets. In the case of the excitation energies, the four
lowest-lying excited states of the LiH molecule have a weaker dependence on
the size of the basis set compared to the $H_{2}$ molecule, with a max
deviation of around 0.5 eV for the regular Hamiltonian. Contrary to the
$H_{2}$ molecule, the regular transcorrelated Hamiltonians (without singles
operators) are able to reduce the max deviation to 0.14 eV. The energies of
these excited states are thus dominated by the dynamical electron correlation
effects. Thus, no major improvements are observed by adding the singles
operators for these four states. Also, the performance of both S’ +
$\hat{H}_{\text{F12}}^{(ij)}$ and S’ + $\hat{H}_{\text{F12}}^{(pq)}$
Hamiltonians are very similar for these states with a maximum difference of
0.06 eV between the two. However, the deviation for the fifth (T3) and sixth
(S3) excited states rises sharply to 2.02 eV and 7.06 eV respectively for the
regular Hamiltonian which gets further increased to 2.41 eV and 7.46 eV for
the $\hat{H}_{\text{F12}}^{(ij)}$ Hamiltonian. The
$\hat{H}_{\text{F12}}^{(pq)}$ Hamiltonian decreases these deviations only
slightly to 2.33 eV and 6.44 eV respectively. The orbital relaxation effects
seem to be very important for these excited states, as can be seen from the
last column of table 2. Indeed, the S’ + $\hat{H}_{\text{F12}}^{(ij)}$ and S’
+ $\hat{H}_{\text{F12}}^{(pq)}$ Hamiltonians are able to bring the max
deviation down to 1.65 eV and 0.55 eV respectively. Thus, just like the H2
molecule, the results obtained using the S’ + $\hat{H}_{\text{F12}}^{(pq)}$
are the most accurate for both ground and excited states.
Table 3: Deviations obtained in the ground state energy (mEh) and the excitation energies (eV) of the seven lowest-lying excited states of the H2O molecule using the ANO-RCC-MIN basis set from the reference cc-pVTZ values (second column) for the six different Hamiltonians (see text for details). Parameters used: ($\gamma$, shift) = (1.4, -0.15) States | Reference | $\hat{H}$ | $\hat{H}_{\text{F12}}^{(ij)}$ | $\hat{H}_{\text{F12}}^{(pq)}$ | S’ + $\hat{H}_{\text{F12}}^{(ij)}$ | S’ + $\hat{H}_{\text{F12}}^{(pq)}$ | S’ + $\hat{H}$
---|---|---|---|---|---|---|---
| cc-pVTZ | | | ANO-RCC-MB | |
S0 | -76.32455 | 342.0 | 151.0 | 121.9 | 33.1 | -1.7 | 233.5
T1 | 7.56 | 3.68 | 5.16 | 3.88 | 1.50 | -0.14 | -0.05
S1 | 8.12 | 4.56 | 6.07 | 4.65 | 1.94 | 0.18 | 0.37
T2 | 9.82 | 3.46 | 4.82 | 3.48 | 1.57 | -0.06 | 0.12
T3 | 9.87 | 4.84 | 6.38 | 4.71 | 1.98 | 0.01 | 0.39
S2 | 10.14 | 5.16 | 6.53 | 4.99 | 2.33 | 0.25 | 0.66
S3 | 10.61 | 4.99 | 6.55 | 4.79 | 2.35 | 0.45 | 0.91
T4 | 11.92 | 3.89 | 5.29 | 3.78 | 1.56 | -0.17 | 0.15
### IV.3 H2O
The calculations on the H2O molecule also employed ANO-RCC-MB basis set and
frozen core settings with the values of $\gamma$ and shift parameter set to
1.4 and -0.15 respectively. From table 3, one can see the regular Hamiltonian
is 342 mEh away from the reference value. The $\hat{H}_{\text{F12}}^{(ij)}$
and $\hat{H}_{\text{F12}}^{(pq)}$ Hamiltonians are able to reduce it to 151
mEh and $\sim$ 122 mEh respectively. Thus, adding all the orbital pairs in the
geminal operator improves the ground state energy by around 29 mEh which is
quite significant. After the addition of the singles operator, we obtained
deviations of 33.1 mEh and -1.7 mEh for the S’ + $\hat{H}_{\text{F12}}^{(ij)}$
and S’ + $\hat{H}_{\text{F12}}^{(pq)}$ Hamiltonians respectively. Thus, just
like in the case of the H2 molecule, we were able to obtain better quality
ground state energies than the reference cc-pVTZ values. The excitation
energies of all the seven lowest-lying excited states of the $H_{2}$O
molecule, unlike the LiH molecule, have a stronger dependence on the size of
the basis set with a max deviation of $\sim$ 5.2 eV for the fifth excited
state (S2) using the regular Hamiltonian. The deviations increase even further
with the $\hat{H}_{\text{F12}}^{(ij)}$ Hamiltonian with a max deviation of
$\sim$ 6.6 eV, again illustrating the unbalanced treatment of the ground and
excited states by this Hamiltonian. Even the $\hat{H}_{\text{F12}}^{(pq)}$
Hamiltonian doesn’t offer any improvement over the regular Hamiltonian and
makes the deviations slightly worse for the three lowest-lying excited states
(T1, S1, T2) while only minor improvements were noted for the next four (T3,
S2, S3, T4). The low-lying excited states of the water molecule is
characterized by a mixture of Rydberg and valence correlation effects. In
essence, the $\hat{H}_{\text{F12}}^{(ij)}$ Hamiltonian has the same effect as
adding polarization functions for a more accurate description of the valence
correlation effects. However, one would require diffuse functions as well in
order to accurately describe the Rydberg character of these states. The effect
of the singles operator can be seen as an injection of diffusivity into the
Hamiltonian. Indeed, the S’ + $\hat{H}_{\text{F12}}^{(pq)}$ Hamiltonian is
able to bring down the max deviation from $\sim$ 5.2 eV (regular H) to $\sim$
0.4 eV. The S’ + $\hat{H}_{\text{F12}}^{(ij)}$ Hamiltonian on the other hand
still produces a max deviation of $\sim$ 2.3 eV. Thus, addition of all orbital
pairs in the geminal operator is very important for an accurate description of
both ground and excited states.
Table 4: Deviations obtained in the ground state energy (mEh) and the excitation energies (eV) of the eight lowest-lying excited states of the NH3 molecule using the ANO-RCC-MIN basis set from the reference cc-pVTZ values (second column) for the six different Hamiltonians (see text for details). Parameters used: ($\gamma$, shift) = (1.1, -0.30) States | Reference | $\hat{H}$ | $\hat{H}_{\text{F12}}^{(ij)}$ | $\hat{H}_{\text{F12}}^{(pq)}$ | S’ + $\hat{H}_{\text{F12}}^{(ij)}$ | S’ + $\hat{H}_{\text{F12}}^{(pq)}$ | S’ + $\hat{H}$
---|---|---|---|---|---|---|---
| cc-pVTZ | | | ANO-RCC-MB | |
S0 | -56.45043 | 270.6 | 105.3 | 72.6 | 24.3 | -13.4 | 197.7
T1 | 6.06 | 3.27 | 4.66 | 3.24 | 1.85 | 0.13 | 0.42
S1 | 6.62 | 4.06 | 5.49 | 3.86 | 2.44 | 0.53 | 1.00
T2 | 8.21 | 3.69 | 5.14 | 3.47 | 2.12 | 0.16 | 0.67
T3 | 8.21 | 3.70 | 5.15 | 3.48 | 2.13 | 0.16 | 0.68
S2 | 8.78 | 4.70 | 6.20 | 4.20 | 2.80 | 0.55 | 1.31
S3 | 8.78 | 4.71 | 6.21 | 4.21 | 2.81 | 0.55 | 1.32
T4 | 10.92 | 3.38 | 4.46 | 3.20 | 1.18 | -0.38 | 0.08
T5 | 10.93 | 3.38 | 4.46 | 3.19 | 1.17 | -0.38 | 0.08
### IV.4 NH3
For the NH3 molecule, the $\gamma$ and shift parameters were chosen to be 1.1
and -0.3 respectively, along with frozen-core settings and ANO-RCC-MB basis
set. The trends for the NH3 molecule as seen from the table 4 are quite
consistent with the previous calculations. The S’ +
$\hat{H}_{\text{F12}}^{(pq)}$ Hamiltonian again provides ground state energies
between cc-pVTZ and cc-pVQZ quality while the regular Hamiltonian yields a
deviation of $\sim$ 271 mEh from the cc-pVTZ value. Looking at the table, one
can see that three pairs ((T2,T3), (S2,S3), ((T4,T5))) of excited states are
nearly degenerate resulting in the appearance of identical deviations with all
the Hamiltonians for these states. For the excitation energies, the max
deviation produced by the S’ + $\hat{H}_{\text{F12}}^{(pq)}$ Hamiltonian is
around 0.5 eV compared to the 4.7 eV and 2.8 eV for the regular and S’ +
$\hat{H}_{\text{F12}}^{(ij)}$ Hamiltonians respectively.
Table 5: Estimation of quantum resources required to simulate the ground state of different molecules and basis sets studied in this work using the UCCSD ansatz assuming Jordan-Wigner mapping and frozen-core settings. All excitation operators are considered along with no circuit transpilation or truncation of the Hamiltonian elements Molecules | Basis | Orbitals | Qubits | Parameters | CNOT Gates
---|---|---|---|---|---
H2 | 6-31G | 4 | 8 | 15 | 768
| cc-pVTZ | 28 | 56 | 783 | 341280
LiH | ANO-RCC-MB | 5 | 10 | 24 | 1616
| cc-pVTZ | 43 | 86 | 1848 | 1249080
H2O | ANO-RCC-MB | 6 | 12 | 92 | 8064
| cc-pVTZ | 57 | 114 | 61904 | 47224272
NH3 | ANO-RCC-MB | 7 | 14 | 204 | 21072
| cc-pVTZ | 71 | 142 | 98892 | 93571664
### IV.5 Quantum resource reduction
Table 5 shows an estimate of the quantum resources required to simulate the
ground state of the molecules studied in this work along with the
corresponding basis sets assuming Jordan-Wigner mapping and frozen-core
settings. All our quantum calculations utilized the UCCSD ansatz and the
number of parameters in 5 refer to the total singles and doubles excitation
operators for the given molecule and basis set. Here, we use the number of
CNOT gates as a measure of the quantum circuit complexity. In order to
estimate the number of CNOT gates, we used the second-quantized particle-hole
formalism for describing the excitation operators and utilized the circuit
designs from Ref. [73]. The number of CNOT gates required for the
exponentiation of a given singles and doubles excitation operator was
calculated and summed to obtain the final numbers. These estimates don’t take
into account any circuit optimization or transpilation and doesn’t use any
kind of truncation schemes for the Hamiltonian matrix elements. However, they
are very useful to describe in a qualitative sense, the massive increase in
quantum resource requirements as the number of qubits increases. For example,
in the case of H2O and NH3 molecules, going from the minimal basis set of ANO-
RCC-MB to the cc-pVTZ basis set results in an increase in the number of CNOT
gates by more than 3 orders of magnitude. A smaller number of CNOT gates
corresponds to a shallow circuit with lower gate errors making the
transcorrelated formalism more suitable for quantum simulations on NISQ
devices. Furthermore, we can also reduce the number of measurements in the
qEOM procedure quite significantly. The measurement of each single matrix
element in the qEOM generalized eigenvalue problem scales as
$\mathcal{O}$($N^{4}$), where N is the number of qubits[51]. For the the NH3
molecule, this would translate into a reduction in the number of measurements
for a given matrix element by a factor of $10^{4}$ $((142/14)^{4})$.
## V Conclusions
We used the canonical transcorrelated theory to construct compact ab initio
Hamiltonians that can drastically reduce the quantum resources required for
accurate simulations of both ground and excited states of molecular systems.
In a work by some of the present authors[36], the transcorrelated Hamiltonians
that were obtained through a similarity transformation of the Hamiltonian with
an explicitly correlated two-body unitary operator greatly accelerated the
recovery of the ground state correlation energies with respect to the size of
the basis set. However, the convergence of the excited state properties like
excitation energies with these Hamiltonians show a completely different trend
and was found to be even slower than the regular Hamiltonian itself. This is
not surprising since excited states can have a very different character than
the ground state. For example, the low-lying excited states of the water
molecule are characterized by a mixture of Rydberg and valence correlation
effects but the traditional explicitly correlated methods can only capture the
missing dynamical electron correlation. Also, the previous formalism was not
able to recover the missing dynamical correlation effects between electrons in
occupied and virtual orbitals due to the absence of orbital pairs involving
virtual orbitals in the definition of the two-body geminal operator. In this
work, we have addressed all these points by re-defining the two-body geminal
operator to include all orbital pairs and added a singles operator in the
similarity transformation procedure to account for the orbital relaxation
effects, resulting in a balanced treatment of both ground and excited states.
The new transcorrelated Hamiltonians can produce ground state energies
comparable to the cc-pVTZ basis, even with a minimal basis set. Furthermore,
it can reduce the errors in the excitation energies by more than an order of
magnitude. This can potentially lead to more than one and three orders of
magnitude reduction in the number of qubits and CNOT gates respectively, in
the VQE procedure. Consequently, the quantum simulations with the
transcorrelated Hamiltonian are expected to be more noise resilient on NISQ
devices.
###### Acknowledgements.
Research presented in this article was supported by the Laboratory Directed
Research and Development (LDRD) program of Los Alamos National Laboratory
(LANL) under project number 20200056DR. LANL is operated by Triad National
Security, LLC, for the National Nuclear Security Administration of U.S.
Department of Energy (contract no. 89233218CNA000001). We thank LANL
Institutional Computing (IC) program for access to HPC resources. AK
acknowledges the help of Tanvi Gujarati (IBM, USA) for help in quantum
resource estimations.
## References
* Feynman [1982] R. P. Feynman, Int. J. Theor. Phys. 121, 467 (1982).
* Nielsen and Chuang [2011] M. A. Nielsen and I. Chuang, _Quantum computation and quantum information_ (Cambridge University Press;, 2011).
* Abrams and Lloyd [1999] D. S. Abrams and S. Lloyd, Phys. Rev. Lett. 83, 5162 (1999).
* Aspuru-Guzik _et al._ [2005] A. Aspuru-Guzik, A. D. Dutoi, P. J. Love, and M. Head-Gordon, Science 309, 1704 (2005).
* Farhi _et al._ [2001] E. Farhi, J. Goldstone, S. Gutmann, J. Lapan, A. Lundgren, and D. Preda, Science 292, 472 (2001).
* Babbush _et al._ [2014] R. Babbush, P. J. Love, and A. Aspuru-Guzik, Sci. Rep. 4, 1 (2014).
* Peruzzo _et al._ [2014] A. Peruzzo, J. McClean, P. Shadbolt, M.-H. Yung, X.-Q. Zhou, P. J. Love, A. Aspuru-Guzik, and J. L. O’brien, Nat. Comm. 5, 1 (2014).
* O’Malley _et al._ [2016] P. J. O’Malley, R. Babbush, I. D. Kivlichan, J. Romero, J. R. McClean, R. Barends, J. Kelly, P. Roushan, A. Tranter, N. Ding, _et al._ , Phys. Rev. X 6, 031007 (2016).
* Grimsley _et al._ [2019] H. R. Grimsley, S. E. Economou, E. Barnes, and N. J. Mayhall, Nat. Comm. 10, 1 (2019).
* Tang _et al._ [2021] H. L. Tang, V. Shkolnikov, G. S. Barron, H. R. Grimsley, N. J. Mayhall, E. Barnes, and S. E. Economou, PRX Quantum 2, 020310 (2021).
* Fedorov _et al._ [2021] D. A. Fedorov, Y. Alexeev, S. K. Gray, and M. Otten, arXiv preprint arXiv:2109.12652 (2021).
* Yordanov _et al._ [2021] Y. S. Yordanov, C. H. Barnes, and D. R. Arvidsson-Shukur, arXiv preprint arXiv:2106.06296 (2021).
* Liu _et al._ [2021] J. Liu, Z. Li, and J. Yang, J. Chem. Phys. 154, 244112 (2021).
* Tkachenko _et al._ [2021] N. V. Tkachenko, J. Sud, Y. Zhang, S. Tretiak, P. M. Anisimov, A. T. Arrasmith, P. J. Coles, L. Cincio, and P. A. Dub, PRX Quantum 2, 020337 (2021).
* Kottmann _et al._ [2021] J. S. Kottmann, P. Schleich, T. Tamayo-Mendoza, and A. Aspuru-Guzik, J. of Phys. Chem. Letters 12, 663 (2021).
* Motta _et al._ [2021] M. Motta, E. Ye, J. R. McClean, Z. Li, A. J. Minnich, R. Babbush, and G. K.-L. Chan, Npj Quantum Inf. 7, 1 (2021).
* Bauman _et al._ [2019a] N. P. Bauman, E. J. Bylaska, S. Krishnamoorthy, G. H. Low, N. Wiebe, C. E. Granade, M. Roetteler, M. Troyer, and K. Kowalski, J. Chem. Phys. 151, 014107 (2019a).
* Bauman _et al._ [2019b] N. P. Bauman, G. H. Low, and K. Kowalski, J. Chem. Phys. 151, 234114 (2019b).
* Zhang _et al._ [2021] Y. Zhang, L. Cincio, C. F. Negre, P. Czarnik, P. Coles, P. M. Anisimov, S. M. Mniszewski, S. Tretiak, and P. A. Dub, arXiv:2106.07619 (2021).
* Romero _et al._ [2018] J. Romero, R. Babbush, J. R. McClean, C. Hempel, P. J. Love, and A. Aspuru-Guzik, Quantum Sci. Technol. 4, 014008 (2018).
* Bartlett _et al._ [1989] R. J. Bartlett, S. A. Kucharski, and J. Noga, Chem. Phys. Lett. 155, 133 (1989).
* Riplinger _et al._ [2016] C. Riplinger, P. Pinski, U. Becker, E. F. Valeev, and F. Neese, J. Chem. Phys. 144, 024109 (2016).
* Kumar _et al._ [2020] A. Kumar, F. Neese, and E. F. Valeev, J. Chem. Phys. 153, 094105 (2020).
* Harrison _et al._ [2004] R. J. Harrison, G. I. Fann, T. Yanai, Z. Gan, and G. Beylkin, J. Chem. Phys. 121, 11587 (2004).
* Kutzelnigg [1985] W. Kutzelnigg, Theor. Chim. Acta 68, 445 (1985).
* Kong _et al._ [2012] L. Kong, F. A. Bischoff, and E. F. Valeev, Chem. Rev. 112, 75 (2012).
* Hättig _et al._ [2012] C. Hättig, W. Klopper, A. Köhn, and D. P. Tew, 112, 4 (2012).
* Ten-no and Noga [2012] S. Ten-no and J. Noga, WIREs Comput Mol Sci 2, 114 (2012).
* Kato [1957] T. Kato, Comm. Pure Appl. Math. 10, 151 (1957).
* Boys and Handy [1969] S. F. Boys and N. C. Handy, Proc. R. Soc. Lond. A. Math. Phys. Sci. 310, 43 (1969).
* Ten-no [2000] S. Ten-no, Chem. Phys. Lett. 330, 169 (2000).
* Luo _et al._ [2010] H. Luo, W. Hackbusch, and H.-J. Flad, Mol. Phys. 108, 425 (2010).
* Yanai and Shiozaki [2012] T. Yanai and T. Shiozaki, J. Comp. Phys. 136, 084107 (2012).
* Yanai and Chan [2006] T. Yanai and G. K.-L. Chan, J. Comp. Phys. 124, 194106 (2006).
* Yanai _et al._ [2010] T. Yanai, Y. Kurashige, E. Neuscamman, and G. K.-L. Chan, J. Comp. Phys. 132, 024105 (2010).
* Motta _et al._ [2020a] M. Motta, T. P. Gujarati, J. E. Rice, A. Kumar, C. Masteran, J. A. Latone, E. Lee, E. F. Valeev, and T. Y. Takeshita, Phys. Chem. Chem. Phys. 22, 24270 (2020a).
* McArdle and Tew [2020] S. McArdle and D. P. Tew, arXiv preprint arXiv:2006.11181 (2020).
* Sokolov _et al._ [2022] I. O. Sokolov, W. Dobrautz, H. Luo, A. Alavi, and I. Tavernelli, arXiv preprint arXiv:2201.03049 (2022).
* Schleich _et al._ [2021] P. Schleich, J. S. Kottmann, and A. Aspuru-Guzik, arXiv preprint arXiv:2110.06812 (2021).
* Torheyden and Valeev [2009] M. Torheyden and E. F. Valeev, J. Chem. Phys. 131, 171103 (2009).
* Neiss _et al._ [2006] C. Neiss, C. Hättig, and W. Klopper, J. Chem. Phys. 125, 064111 (2006).
* Fliegl _et al._ [2006] H. Fliegl, C. Hättig, and W. Klopper, J. Chem. Phys. 124, 044112 (2006).
* Watson and Chan [2016] T. J. Watson and G. K.-L. Chan, J. Chem. Theory Comput. 12, 512 (2016), pMID: 26756223.
* Lee _et al._ [2019] J. Lee, W. J. Huggins, M. Head-Gordon, and K. B. Whaley, J. Chem. Theory Comput. 15, 311 (2019).
* Motta _et al._ [2020b] M. Motta, C. Sun, A. T. Tan, M. J. O’Rourke, E. Ye, A. J. Minnich, F. G. Brandão, and G. K.-L. Chan, Nat. Phys. 16, 205 (2020b).
* Stair _et al._ [2020] N. H. Stair, R. Huang, and F. A. Evangelista, J. Chem. Theory Comput. 16, 2236 (2020).
* Cortes and Gray [2021] C. L. Cortes and S. K. Gray, arXiv preprint arXiv:2109.06868 (2021).
* Yeter-Aydeniz _et al._ [2021] K. Yeter-Aydeniz, B. T. Gard, J. Jakowski, S. Majumder, G. S. Barron, G. Siopsis, T. S. Humble, and R. C. Pooser, Adv. Quantum Tech. , 2100012 (2021).
* Colless _et al._ [2018] J. I. Colless, V. V. Ramasesh, D. Dahlen, M. S. Blok, M. E. Kimchi-Schwartz, J. R. McClean, J. Carter, W. A. de Jong, and I. Siddiqi, Phys. Rev. X 8, 011021 (2018).
* Parrish _et al._ [2019] R. M. Parrish, E. G. Hohenstein, P. L. McMahon, and T. J. Martínez, Phys. Rev. Lett. 122, 230401 (2019).
* Ollitrault _et al._ [2020] P. J. Ollitrault, A. Kandala, C.-F. Chen, P. K. Barkoutsos, A. Mezzacapo, M. Pistoia, S. Sheldon, S. Woerner, J. M. Gambetta, and I. Tavernelli, Phys. Rev. Research 2, 043140 (2020).
* Kutzelnigg [1992] W. Kutzelnigg, Theor. Chim. Acta 83, 263 (1992).
* Köhn and Tew [2010] A. Köhn and D. P. Tew, J. Chem. Phys. 133, 174117 (2010).
* Mukherjee [1997] D. Mukherjee, Chem. Phys. Lett. 274, 561 (1997).
* Kutzelnigg and Mukherjee [1997] W. Kutzelnigg and D. Mukherjee, J. Chem. Phys. 107, 432 (1997).
* Kutzelnigg _et al._ [2010] W. Kutzelnigg, K. Shamasundar, and D. Mukherjee, Mol. Phys. 108, 433 (2010).
* Yanai and Chan [2007] T. Yanai and G. K.-L. Chan, J. Chem. Phys. 127, 104107 (2007).
* Ten-no [2004] S. Ten-no, J. Chem. Phys. 121, 117 (2004).
* Zhang and Valeev [2012] J. Zhang and E. F. Valeev, J. Chem. Theory Comput. 8, 3175 (2012).
* Tew and Klopper [2005] D. P. Tew and W. Klopper, J. Chem. Phys. 123, 074101 (2005).
* Nakano _et al._ [2002] H. Nakano, R. Uchiyama, and K. Hirao, J. Comp. Chem. 23, 1166 (2002).
* Lee and Head-Gordon [2018] J. Lee and M. Head-Gordon, J. Chem. Theory Comput. 14, 5203 (2018), pMID: 30130398\.
* Kedžuch _et al._ [2005] S. Kedžuch, M. Milko, and J. Noga, Int. J. Quantum Chem. 105, 929 (2005).
* [64] https://github.com/ashutoshvt/WicksTheorem/tree/vacuum.
* Harris _et al._ [2020] C. R. Harris, K. J. Millman, S. J. van der Walt, R. Gommers, P. Virtanen, D. Cournapeau, E. Wieser, J. Taylor, S. Berg, N. J. Smith, R. Kern, M. Picus, S. Hoyer, M. H. van Kerkwijk, M. Brett, A. Haldane, J. F. del Río, M. Wiebe, P. Peterson, P. Gérard-Marchant, K. Sheppard, T. Reddy, W. Weckesser, H. Abbasi, C. Gohlke, and T. E. Oliphant, Nature 585, 357 (2020).
* MPQ [2020] http://github.com/ValeevGroup/mpqc (2020).
* Roos _et al._ [2005] B. O. Roos, R. Lindh, P.-Å. Malmqvist, V. Veryazov, and P.-O. Widmark, J. Phys. Chem. A 109, 6575 (2005).
* Yousaf and Peterson [2009] K. E. Yousaf and K. A. Peterson, Chemical Physics Letters 476, 303 (2009).
* Yousaf and Peterson [2008] K. E. Yousaf and K. A. Peterson, J. Chem. Phys. 129, 184108 (2008).
* Valeev [2004] E. F. Valeev, Chem. Phys. Lett. 395, 190 (2004).
* [71] https://github.com/mayhallgroup/adapt-vqe.
* Sun _et al._ [2020] Q. Sun, X. Zhang, S. Banerjee, P. Bao, M. Barbry, N. S. Blunt, N. A. Bogdanov, G. H. Booth, J. Chen, Z.-H. Cui, J. J. Eriksen, Y. Gao, S. Guo, J. Hermann, M. R. Hermes, K. Koh, P. Koval, S. Lehtola, Z. Li, J. Liu, N. Mardirossian, J. D. McClain, M. Motta, B. Mussard, H. Q. Pham, A. Pulkin, W. Purwanto, P. J. Robinson, E. Ronca, E. R. Sayfutyarova, M. Scheurer, H. F. Schurkus, J. E. T. Smith, C. Sun, S.-N. Sun, S. Upadhyay, L. K. Wagner, X. Wang, A. White, J. D. Whitfield, M. J. Williamson, S. Wouters, J. Yang, J. M. Yu, T. Zhu, T. C. Berkelbach, S. Sharma, A. Y. Sokolov, and G. K.-L. Chan, J. Chem. Phys. 153, 024109 (2020).
* Barkoutsos _et al._ [2018] P. K. Barkoutsos, J. F. Gonthier, I. Sokolov, N. Moll, G. Salis, A. Fuhrer, M. Ganzhorn, D. J. Egger, M. Troyer, A. Mezzacapo, _et al._ , Phys. Rev. A 98, 022322 (2018).
|
# Search for secluded dark matter towards the Galactic Centre with the ANTARES
neutrino telescope
A. Albert Université de Strasbourg, CNRS, IPHC UMR 7178, F-67000 Strasbourg,
France Université de Haute Alsace, F-68100 Mulhouse, France S. Alves IFIC -
Instituto de Física Corpuscular (CSIC - Universitat de València) c/
Catedrático José Beltrán, 2 E-46980 Paterna, Valencia, Spain M. André
Technical University of Catalonia, Laboratory of Applied Bioacoustics, Rambla
Exposició, 08800 Vilanova i la Geltrú, Barcelona, Spain M. Anghinolfi INFN -
Sezione di Genova, Via Dodecaneso 33, 16146 Genova, Italy G. Anton
Friedrich-Alexander-Universität Erlangen-Nürnberg, Erlangen Centre for
Astroparticle Physics, Erwin-Rommel-Str. 1, 91058 Erlangen, Germany M. Ardid
Institut d’Investigació per a la Gestió Integrada de les Zones Costaneres
(IGIC) - Universitat Politècnica de València. C/ Paranimf 1, 46730 Gandia,
Spain S. Ardid Institut d’Investigació per a la Gestió Integrada de les
Zones Costaneres (IGIC) - Universitat Politècnica de València. C/ Paranimf 1,
46730 Gandia, Spain J.-J. Aubert Aix Marseille Univ, CNRS/IN2P3, CPPM,
Marseille, France J. Aublin Université de Paris, CNRS, Astroparticule et
Cosmologie, F-75013 Paris, France B. Baret Université de Paris, CNRS,
Astroparticule et Cosmologie, F-75013 Paris, France S. Basa Aix Marseille
Univ, CNRS, CNES, LAM, Marseille, France B. Belhorma National Center for
Energy Sciences and Nuclear Techniques, B.P.1382, R. P.10001 12, Morocco M.
Bendahman Université de Paris, CNRS, Astroparticule et Cosmologie, F-75013
Paris, France University Mohammed V in Rabat, Faculty of Sciences, 4 av. Ibn
Battouta, B.P. 1014, R.P. 10000 Rabat, Morocco F. Benfenati INFN - Sezione
di Bologna, Viale Berti-Pichat 6/2, 40127 Bologna, Italy Dipartimento di
Fisica e Astronomia dell’Università, Viale Berti Pichat 6/2, 40127 Bologna,
Italy V. Bertin Aix Marseille Univ, CNRS/IN2P3, CPPM, Marseille, France S.
Biagi INFN - Laboratori Nazionali del Sud (LNS), Via S. Sofia 62, 95123
Catania, Italy M. Bissinger Friedrich-Alexander-Universität Erlangen-
Nürnberg, Erlangen Centre for Astroparticle Physics, Erwin-Rommel-Str. 1,
91058 Erlangen, Germany J. Boumaaza University Mohammed V in Rabat, Faculty
of Sciences, 4 av. Ibn Battouta, B.P. 1014, R.P. 10000 Rabat, Morocco M.
Bouta University Mohammed I, Laboratory of Physics of Matter and Radiations,
B.P.717, Oujda 6000, Morocco M.C. Bouwhuis Nikhef, Science Park, Amsterdam,
The Netherlands H. Brânzaş Institute of Space Science, RO-077125 Bucharest,
Măgurele, Romania R. Bruijn Nikhef, Science Park, Amsterdam, The Netherlands
Universiteit van Amsterdam, Instituut voor Hoge-Energie Fysica, Science Park
105, 1098 XG Amsterdam, The Netherlands J. Brunner Aix Marseille Univ,
CNRS/IN2P3, CPPM, Marseille, France J. Busto Aix Marseille Univ, CNRS/IN2P3,
CPPM, Marseille, France B. Caiffi INFN - Sezione di Genova, Via Dodecaneso
33, 16146 Genova, Italy D. Calvo IFIC - Instituto de Física Corpuscular
(CSIC - Universitat de València) c/ Catedrático José Beltrán, 2 E-46980
Paterna, Valencia, Spain A. Capone INFN - Sezione di Roma, P.le Aldo Moro 2,
00185 Roma, Italy Dipartimento di Fisica dell’Università La Sapienza, P.le
Aldo Moro 2, 00185 Roma, Italy L. Caramete Institute of Space Science,
RO-077125 Bucharest, Măgurele, Romania J. Carr Aix Marseille Univ,
CNRS/IN2P3, CPPM, Marseille, France V. Carretero IFIC - Instituto de Física
Corpuscular (CSIC - Universitat de València) c/ Catedrático José Beltrán, 2
E-46980 Paterna, Valencia, Spain S. Celli INFN - Sezione di Roma, P.le Aldo
Moro 2, 00185 Roma, Italy Dipartimento di Fisica dell’Università La Sapienza,
P.le Aldo Moro 2, 00185 Roma, Italy M. Chabab LPHEA, Faculty of Science -
Semlali, Cadi Ayyad University, P.O.B. 2390, Marrakech, Morocco. T. N. Chau
Université de Paris, CNRS, Astroparticule et Cosmologie, F-75013 Paris, France
R. Cherkaoui El Moursli University Mohammed V in Rabat, Faculty of Sciences,
4 av. Ibn Battouta, B.P. 1014, R.P. 10000 Rabat, Morocco T. Chiarusi INFN -
Sezione di Bologna, Viale Berti-Pichat 6/2, 40127 Bologna, Italy M. Circella
INFN - Sezione di Bari, Via E. Orabona 4, 70126 Bari, Italy A. Coleiro
Université de Paris, CNRS, Astroparticule et Cosmologie, F-75013 Paris, France
R. Coniglione INFN - Laboratori Nazionali del Sud (LNS), Via S. Sofia 62,
95123 Catania, Italy P. Coyle Aix Marseille Univ, CNRS/IN2P3, CPPM,
Marseille, France A. Creusot Université de Paris, CNRS, Astroparticule et
Cosmologie, F-75013 Paris, France A. F. Díaz Department of Computer
Architecture and Technology/CITIC, University of Granada, 18071 Granada, Spain
G. de Wasseige Université de Paris, CNRS, Astroparticule et Cosmologie,
F-75013 Paris, France C. Distefano INFN - Laboratori Nazionali del Sud
(LNS), Via S. Sofia 62, 95123 Catania, Italy I. Di Palma INFN - Sezione di
Roma, P.le Aldo Moro 2, 00185 Roma, Italy Dipartimento di Fisica
dell’Università La Sapienza, P.le Aldo Moro 2, 00185 Roma, Italy A. Domi
Nikhef, Science Park, Amsterdam, The Netherlands Universiteit van Amsterdam,
Instituut voor Hoge-Energie Fysica, Science Park 105, 1098 XG Amsterdam, The
Netherlands C. Donzaud Université de Paris, CNRS, Astroparticule et
Cosmologie, F-75013 Paris, France Université Paris-Sud, 91405 Orsay Cedex,
France D. Dornic Aix Marseille Univ, CNRS/IN2P3, CPPM, Marseille, France D.
Drouhin Université de Strasbourg, CNRS, IPHC UMR 7178, F-67000 Strasbourg,
France Université de Haute Alsace, F-68100 Mulhouse, France T. Eberl
Friedrich-Alexander-Universität Erlangen-Nürnberg, Erlangen Centre for
Astroparticle Physics, Erwin-Rommel-Str. 1, 91058 Erlangen, Germany T. van
Eeden Nikhef, Science Park, Amsterdam, The Netherlands D. van Eijk Nikhef,
Science Park, Amsterdam, The Netherlands N. El Khayati University Mohammed V
in Rabat, Faculty of Sciences, 4 av. Ibn Battouta, B.P. 1014, R.P. 10000
Rabat, Morocco A. Enzenhöfer Aix Marseille Univ, CNRS/IN2P3, CPPM,
Marseille, France P. Fermani INFN - Sezione di Roma, P.le Aldo Moro 2, 00185
Roma, Italy Dipartimento di Fisica dell’Università La Sapienza, P.le Aldo
Moro 2, 00185 Roma, Italy G. Ferrara INFN - Laboratori Nazionali del Sud
(LNS), Via S. Sofia 62, 95123 Catania, Italy F. Filippini INFN - Sezione di
Bologna, Viale Berti-Pichat 6/2, 40127 Bologna, Italy Dipartimento di Fisica
e Astronomia dell’Università, Viale Berti Pichat 6/2, 40127 Bologna, Italy L.
Fusco Aix Marseille Univ, CNRS/IN2P3, CPPM, Marseille, France Y. Gatelet
Université de Paris, CNRS, Astroparticule et Cosmologie, F-75013 Paris, France
P. Gay Université de Paris, CNRS, Astroparticule et Cosmologie, F-75013
Paris, France Laboratoire de Physique Corpusculaire, Clermont Université,
Université Blaise Pascal, CNRS/IN2P3, BP 10448, F-63000 Clermont-Ferrand,
France H. Glotin LIS, UMR Université de Toulon, Aix Marseille Université,
CNRS, 83041 Toulon, France R. Gozzini IFIC - Instituto de Física Corpuscular
(CSIC - Universitat de València) c/ Catedrático José Beltrán, 2 E-46980
Paterna, Valencia, Spain R. Gracia Ruiz Nikhef, Science Park, Amsterdam, The
Netherlands K. Graf Friedrich-Alexander-Universität Erlangen-Nürnberg,
Erlangen Centre for Astroparticle Physics, Erwin-Rommel-Str. 1, 91058
Erlangen, Germany C. Guidi INFN - Sezione di Genova, Via Dodecaneso 33,
16146 Genova, Italy Dipartimento di Fisica dell’Università, Via Dodecaneso
33, 16146 Genova, Italy S. Hallmann Friedrich-Alexander-Universität
Erlangen-Nürnberg, Erlangen Centre for Astroparticle Physics, Erwin-Rommel-
Str. 1, 91058 Erlangen, Germany H. van Haren Royal Netherlands Institute for
Sea Research (NIOZ), Landsdiep 4, 1797 SZ ’t Horntje (Texel), the Netherlands
A.J. Heijboer Nikhef, Science Park, Amsterdam, The Netherlands Y. Hello
Géoazur, UCA, CNRS, IRD, Observatoire de la Côte d’Azur, Sophia Antipolis,
France J.J. Hernández-Rey IFIC - Instituto de Física Corpuscular (CSIC -
Universitat de València) c/ Catedrático José Beltrán, 2 E-46980 Paterna,
Valencia, Spain J. Hößl Friedrich-Alexander-Universität Erlangen-Nürnberg,
Erlangen Centre for Astroparticle Physics, Erwin-Rommel-Str. 1, 91058
Erlangen, Germany J. Hofestädt Friedrich-Alexander-Universität Erlangen-
Nürnberg, Erlangen Centre for Astroparticle Physics, Erwin-Rommel-Str. 1,
91058 Erlangen, Germany F. Huang Aix Marseille Univ, CNRS/IN2P3, CPPM,
Marseille, France G. Illuminati Université de Paris, CNRS, Astroparticule et
Cosmologie, F-75013 Paris, France INFN - Sezione di Bologna, Viale Berti-
Pichat 6/2, 40127 Bologna, Italy Dipartimento di Fisica e Astronomia
dell’Università, Viale Berti Pichat 6/2, 40127 Bologna, Italy C. W. James
International Centre for Radio Astronomy Research - Curtin University,
Bentley, WA 6102, Australia B. Jisse-Jung Nikhef, Science Park, Amsterdam,
The Netherlands M. de Jong Nikhef, Science Park, Amsterdam, The Netherlands
Huygens-Kamerlingh Onnes Laboratorium, Universiteit Leiden, The Netherlands
P. de Jong Nikhef, Science Park, Amsterdam, The Netherlands Universiteit van
Amsterdam, Instituut voor Hoge-Energie Fysica, Science Park 105, 1098 XG
Amsterdam, The Netherlands M. Kadler Institut für Theoretische Physik und
Astrophysik, Universität Würzburg, Emil-Fischer Str. 31, 97074 Würzburg,
Germany O. Kalekin Friedrich-Alexander-Universität Erlangen-Nürnberg,
Erlangen Centre for Astroparticle Physics, Erwin-Rommel-Str. 1, 91058
Erlangen, Germany U. Katz Friedrich-Alexander-Universität Erlangen-Nürnberg,
Erlangen Centre for Astroparticle Physics, Erwin-Rommel-Str. 1, 91058
Erlangen, Germany N.R. Khan-Chowdhury IFIC - Instituto de Física Corpuscular
(CSIC - Universitat de València) c/ Catedrático José Beltrán, 2 E-46980
Paterna, Valencia, Spain A. Kouchner Université de Paris, CNRS,
Astroparticule et Cosmologie, F-75013 Paris, France I. Kreykenbohm Dr.
Remeis-Sternwarte and ECAP, Friedrich-Alexander-Universität Erlangen-Nürnberg,
Sternwartstr. 7, 96049 Bamberg, Germany V. Kulikovskiy INFN - Sezione di
Genova, Via Dodecaneso 33, 16146 Genova, Italy C. Lagunas Gualda Deutsches
Elektronen Synchrotron DESY, Platanenallee 6, 15738 Zeuthen, Germany Institut
für Physik, Humboldt-Universität zu Berlin, D-12489 Berlin, Germany R.
Lahmann Friedrich-Alexander-Universität Erlangen-Nürnberg, Erlangen Centre
for Astroparticle Physics, Erwin-Rommel-Str. 1, 91058 Erlangen, Germany R. Le
Breton Université de Paris, CNRS, Astroparticule et Cosmologie, F-75013
Paris, France S. LeStum Aix Marseille Univ, CNRS/IN2P3, CPPM, Marseille,
France D. Lefèvre Mediterranean Institute of Oceanography (MIO), Aix-
Marseille University, 13288, Marseille, Cedex 9, France; Université du Sud
Toulon-Var, CNRS-INSU/IRD UM 110, 83957, La Garde Cedex, France E. Leonora
INFN - Sezione di Catania, Via S. Sofia 64, 95123 Catania, Italy G. Levi
INFN - Sezione di Bologna, Viale Berti-Pichat 6/2, 40127 Bologna, Italy
Dipartimento di Fisica e Astronomia dell’Università, Viale Berti Pichat 6/2,
40127 Bologna, Italy M. Lincetto Aix Marseille Univ, CNRS/IN2P3, CPPM,
Marseille, France D. Lopez-Coto Dpto. de Física Teórica y del Cosmos &
C.A.F.P.E., University of Granada, 18071 Granada, Spain S. Loucatos
Université de Paris, CNRS, Astroparticule et Cosmologie, F-75013 Paris, France
IRFU, CEA, Université Paris-Saclay, F-91191 Gif-sur-Yvette, France L. Maderer
Université de Paris, CNRS, Astroparticule et Cosmologie, F-75013 Paris, France
J. Manczak IFIC - Instituto de Física Corpuscular (CSIC - Universitat de
València) c/ Catedrático José Beltrán, 2 E-46980 Paterna, Valencia, Spain M.
Marcelin Aix Marseille Univ, CNRS, CNES, LAM, Marseille, France A. Margiotta
INFN - Sezione di Bologna, Viale Berti-Pichat 6/2, 40127 Bologna, Italy
Dipartimento di Fisica e Astronomia dell’Università, Viale Berti Pichat 6/2,
40127 Bologna, Italy A. Marinelli INFN - Sezione di Napoli, Via Cintia 80126
Napoli, Italy J.A. Martínez-Mora Institut d’Investigació per a la Gestió
Integrada de les Zones Costaneres (IGIC) - Universitat Politècnica de
València. C/ Paranimf 1, 46730 Gandia, Spain B. Martino Aix Marseille Univ,
CNRS/IN2P3, CPPM, Marseille, France K. Melis Nikhef, Science Park,
Amsterdam, The Netherlands Universiteit van Amsterdam, Instituut voor Hoge-
Energie Fysica, Science Park 105, 1098 XG Amsterdam, The Netherlands P.
Migliozzi INFN - Sezione di Napoli, Via Cintia 80126 Napoli, Italy A. Moussa
University Mohammed I, Laboratory of Physics of Matter and Radiations,
B.P.717, Oujda 6000, Morocco R. Muller Nikhef, Science Park, Amsterdam, The
Netherlands L. Nauta Nikhef, Science Park, Amsterdam, The Netherlands S.
Navas Dpto. de Física Teórica y del Cosmos & C.A.F.P.E., University of
Granada, 18071 Granada, Spain E. Nezri Aix Marseille Univ, CNRS, CNES, LAM,
Marseille, France B. Ó Fearraigh Nikhef, Science Park, Amsterdam, The
Netherlands A. Păun Institute of Space Science, RO-077125 Bucharest,
Măgurele, Romania G.E. Păvălaş Institute of Space Science, RO-077125
Bucharest, Măgurele, Romania C. Pellegrino INFN - Sezione di Bologna, Viale
Berti-Pichat 6/2, 40127 Bologna, Italy Museo Storico della Fisica e Centro
Studi e Ricerche Enrico Fermi, Piazza del Viminale 1, 00184, Roma INFN -
CNAF, Viale C. Berti Pichat 6/2, 40127, Bologna M. Perrin-Terrin Aix
Marseille Univ, CNRS/IN2P3, CPPM, Marseille, France V. Pestel Nikhef,
Science Park, Amsterdam, The Netherlands P. Piattelli INFN - Laboratori
Nazionali del Sud (LNS), Via S. Sofia 62, 95123 Catania, Italy C. Pieterse
IFIC - Instituto de Física Corpuscular (CSIC - Universitat de València) c/
Catedrático José Beltrán, 2 E-46980 Paterna, Valencia, Spain C. Poirè
Institut d’Investigació per a la Gestió Integrada de les Zones Costaneres
(IGIC) - Universitat Politècnica de València. C/ Paranimf 1, 46730 Gandia,
Spain V. Popa Institute of Space Science, RO-077125 Bucharest, Măgurele,
Romania T. Pradier Université de Strasbourg, CNRS, IPHC UMR 7178, F-67000
Strasbourg, France N. Randazzo INFN - Sezione di Catania, Via S. Sofia 64,
95123 Catania, Italy D. Real IFIC - Instituto de Física Corpuscular (CSIC -
Universitat de València) c/ Catedrático José Beltrán, 2 E-46980 Paterna,
Valencia, Spain S. Reck Friedrich-Alexander-Universität Erlangen-Nürnberg,
Erlangen Centre for Astroparticle Physics, Erwin-Rommel-Str. 1, 91058
Erlangen, Germany G. Riccobene INFN - Laboratori Nazionali del Sud (LNS),
Via S. Sofia 62, 95123 Catania, Italy A. Romanov INFN - Sezione di Genova,
Via Dodecaneso 33, 16146 Genova, Italy Dipartimento di Fisica
dell’Università, Via Dodecaneso 33, 16146 Genova, Italy F. Sala Laboratoire
de Physique Théorique et Hautes Énergies, CNRS, Sorbonne Université, Paris,
France A. Sánchez-Losa IFIC - Instituto de Física Corpuscular (CSIC -
Universitat de València) c/ Catedrático José Beltrán, 2 E-46980 Paterna,
Valencia, Spain INFN - Sezione di Bari, Via E. Orabona 4, 70126 Bari, Italy
F. Salesa Greus IFIC - Instituto de Física Corpuscular (CSIC - Universitat de
València) c/ Catedrático José Beltrán, 2 E-46980 Paterna, Valencia, Spain D.
F. E. Samtleben Nikhef, Science Park, Amsterdam, The Netherlands Huygens-
Kamerlingh Onnes Laboratorium, Universiteit Leiden, The Netherlands M.
Sanguineti INFN - Sezione di Genova, Via Dodecaneso 33, 16146 Genova, Italy
Dipartimento di Fisica dell’Università, Via Dodecaneso 33, 16146 Genova, Italy
P. Sapienza INFN - Laboratori Nazionali del Sud (LNS), Via S. Sofia 62, 95123
Catania, Italy J. Schnabel Friedrich-Alexander-Universität Erlangen-
Nürnberg, Erlangen Centre for Astroparticle Physics, Erwin-Rommel-Str. 1,
91058 Erlangen, Germany J. Schumann Friedrich-Alexander-Universität
Erlangen-Nürnberg, Erlangen Centre for Astroparticle Physics, Erwin-Rommel-
Str. 1, 91058 Erlangen, Germany F. Schüssler IRFU, CEA, Université Paris-
Saclay, F-91191 Gif-sur-Yvette, France J. Seneca Nikhef, Science Park,
Amsterdam, The Netherlands M. Spurio INFN - Sezione di Bologna, Viale Berti-
Pichat 6/2, 40127 Bologna, Italy Dipartimento di Fisica e Astronomia
dell’Università, Viale Berti Pichat 6/2, 40127 Bologna, Italy Th. Stolarczyk
IRFU, CEA, Université Paris-Saclay, F-91191 Gif-sur-Yvette, France M. Taiuti
INFN - Sezione di Genova, Via Dodecaneso 33, 16146 Genova, Italy Dipartimento
di Fisica dell’Università, Via Dodecaneso 33, 16146 Genova, Italy Y. Tayalati
University Mohammed V in Rabat, Faculty of Sciences, 4 av. Ibn Battouta, B.P.
1014, R.P. 10000 Rabat, Morocco S.J. Tingay International Centre for Radio
Astronomy Research - Curtin University, Bentley, WA 6102, Australia B.
Vallage Université de Paris, CNRS, Astroparticule et Cosmologie, F-75013
Paris, France IRFU, CEA, Université Paris-Saclay, F-91191 Gif-sur-Yvette,
France V. Van Elewyck Université de Paris, CNRS, Astroparticule et
Cosmologie, F-75013 Paris, France Institut Universitaire de France, 75005
Paris, France F. Versari Université de Paris, CNRS, Astroparticule et
Cosmologie, F-75013 Paris, France INFN - Sezione di Bologna, Viale Berti-
Pichat 6/2, 40127 Bologna, Italy Dipartimento di Fisica e Astronomia
dell’Università, Viale Berti Pichat 6/2, 40127 Bologna, Italy S. Viola INFN
- Laboratori Nazionali del Sud (LNS), Via S. Sofia 62, 95123 Catania, Italy
D. Vivolo INFN - Sezione di Napoli, Via Cintia 80126 Napoli, Italy
Dipartimento di Fisica dell’Università Federico II di Napoli, Via Cintia
80126, Napoli, Italy J. Wilms Dr. Remeis-Sternwarte and ECAP, Friedrich-
Alexander-Universität Erlangen-Nürnberg, Sternwartstr. 7, 96049 Bamberg,
Germany S. Zavatarelli INFN - Sezione di Genova, Via Dodecaneso 33, 16146
Genova, Italy A. Zegarelli INFN - Sezione di Roma, P.le Aldo Moro 2, 00185
Roma, Italy Dipartimento di Fisica dell’Università La Sapienza, P.le Aldo
Moro 2, 00185 Roma, Italy J.D. Zornoza IFIC - Instituto de Física
Corpuscular (CSIC - Universitat de València) c/ Catedrático José Beltrán, 2
E-46980 Paterna, Valencia, Spain J. Zúñiga IFIC - Instituto de Física
Corpuscular (CSIC - Universitat de València) c/ Catedrático José Beltrán, 2
E-46980 Paterna, Valencia, Spain (The ANTARES Collaboration)
###### Abstract
Searches for dark matter (DM) have not provided any solid evidence for the
existence of weakly interacting massive particles in the GeV-TeV mass range.
Coincidentally, the scale of new physics is being pushed by collider searches
well beyond the TeV domain. This situation strongly motivates the exploration
of DM masses much larger than a TeV. Secluded scenarios contain a natural way
around the unitarity bound on the DM mass, via the early matter domination
induced by the mediator of its interactions with the Standard Model. High-
energy neutrinos constitute one of the very few direct accesses to energy
scales above a few TeV. An indirect search for secluded DM signals has been
performed with the ANTARES neutrino telescope using data from 2007 to 2015.
Upper limits on the DM annihilation cross section for DM masses up to 6 PeV
are presented and discussed.
###### Contents
1. 1 Introduction
2. 2 Neutrinos from Dark Matter annihilations
3. 3 Detector and data set
4. 4 Analysis method
1. 4.1 Data Selection
2. 4.2 Signal Identification
5. 5 Results and discussion
6. 6 Conclusions
## 1 Introduction
Astrophysical and cosmological observations point to the existence of non-
luminous matter beyond that contained in the Standard Model (SM) of particle
physics. Among the many proposed candidates for such dark matter (DM), weakly
interacting massive particles, with a mass at the electroweak scale, have been
long looked for. They annihilate to ordinary particles detectable far from
their source, are scattered by ordinary matter, and can be produced at
colliders. No clear evidence for their existence has emerged so far from data.
This situation is encouraging the exploration of new regions of the DM
parameter space, and indeed recent years have seen a growing theoretical
interest in DM candidates heavier than about 10 TeV. This mass range is of
even more interest in light of the empty-handed searches for physics beyond
the Standard Model at the LHC, which push new physics models at scales larger
than a few TeV, see e.g. [1, 2]. In turn, these models may naturally host dark
matter candidates with a mass in a similar range, as known since a long time,
for example in supersymmetric theories [3].
Considerations of unitarity of DM annihilation processes imply the existence
of a well-known upper limit, of about 100 TeV, on the DM mass [4], see e.g.
[5, 6] for recent appraisals. This limit holds if some conditions about the
cosmological history of the universe and of DM are respected, and can for
example be easily evaded if the universe was matter dominated between the
freeze-out of dark matter interactions and Big Bang nucleosynthesis, see e.g.
[7]. Secluded DM models [8] naturally constitute a very economical framework
that realises the needed early-matter domination [9, 10, 11, 12]. Here the
dark matter particle interacts sizeably with a mediator, in turn feebly
interacting with SM particles. In these scenarios, the unitarity bound on the
mass of thermal dark matter is avoided thanks to the late time entropy
injection from decays of the mediators, which are responsible for the early-
matter domination. Dark matter masses of 100 TeV and above are therefore
allowed. Such models provide large signals in the so-called indirect detection
searches (because controlled by the dark matter-mediator interaction) with
almost no signal in direct detection and collider experiments (because
controlled by the small mediator-SM coupling). From a technical point of view,
a reliable phenomenological computation of the spectra of SM particles arising
from DM annihilations in secluded model is possible, providing an excellent
motivationfor indirect searches [12]. Indeed, the relevant energy scale is not
the heavy dark matter mass (that would demand a resummation of electroweak
radiation for111 Through all the text, units are chosen such that $c=\hbar=1$.
$m_{\text{DM}}>O(10)$ TeV, see [13] for a recent study that addressed this
challenge), but rather the sub-TeV mediator mass, where the first order
treatment of electroweak corrections [14] implemented in the tool PPPC4DMID
[15] is well under control. Therefore, despite the absence of prior bounds on
the mass of the mediator $m_{V}$ from theory, a reliable computation of the
indirect detection signals with [15] is only possible when $m_{V}<O(10)$ TeV,
otherwise electroweak radiation should be resummed also to compute decays of
the mediator. This condition on the masses also implies that the interaction
between DM particles is long-range, giving rise to phenomena like Sommerfeld
enhancement [16, 17] and bound state formation [18, 19, 20], which
significantly enhance the DM signal at present times with respect to the
‘standard’ case of short-range interactions.
Neutrino telescopes have been used for indirect searches of DM (see for
instance [21] for a recent review). The ANTARES detector has been used before
to search for DM accumulated in the Earth [22], the Sun [23] and the Galactic
Centre [24]. Moreover, there has been a specific search for secluded DM with
ANTARES looking at the Sun [25], and in the public data from IceCube [26, 27].
However, the Sun is not the best source to explore heavy DM due to absorption
of resulting particles in this dense medium, even high-energy neutrinos. Thus,
it seems for this case more appropriate to look at the Galactic Centre and
high-energy neutrinos constitute one of the very few direct accesses to energy
scales above a few TeV. This places the ANTARES telescope in a privileged
position to test this relatively unexplored mass range for dark matter, via
the search for neutrinos possibly coming from dark matter annihilations or
decays. This position is reinforced by the favourable geographical location of
the telescope with respect to the position of the Galactic Centre, where most
of the indirect signal from dark matter is expected to originate. It appears
therefore very well motivated to exploit ANTARES data to test models of dark
matter heavier than a few TeV.
This paper is organised as follows. The production spectra for heavy secluded
dark matter are detailed in Section 2. A description of the experimental setup
is presented in Section 3 (as for the detector and data set used) and 4 (as
for the analysis method). The results of this work are exposed and discussed
in Section 5, summarised and placed in further context in Section 6.
## 2 Neutrinos from Dark Matter annihilations
The neutrino signal at the ANTARES site arises from the annihilation of a pair
of dark matter particles into two mediators. They then decay into neutrinos
and/or other SM particles, which in turn will produce neutrinos via showering
and decays. The mediator lifetime is required to be shorter than about 0.1
seconds to respect limits from Big Bang nucleosynthesis [28]. With this
constraint, the mediator decay process is instantaneous from the astrophysical
point of view, and takes place entirely in the source of interest. The
baryonic matter density in the Galactic Centre is not enough to cause
distortions or absorption effects in outcoming neutrino spectra. The formation
of positronium-like bound states of DM can sizeably contribute to the signal
of interest for ANTARES, via the decay of the bound state into two or more
mediators [18]. The dark matter annihilation cross section, for which limits
will be presented here, is then to be intended as an effective cross section
taking into account also the bound state contribution (see e.g. [29, 11] for
more details).
The energy spectra of the neutrinos per single dark matter annihilation are
computed in two steps. First, the energy spectra of neutrinos from the decay
of a mediator at rest are obtained with the PPPC4DMID tool [15]. Second,
spectra are boosted to the centre of mass frame of the dark matter pair that
annihilates (see [30] for more details on this procedure). Flavour
oscillations then occur between the source and the detector site. In this
analysis, the production of three neutrino flavours was considered in the
Galactic Centre, and oscillated in the long-baseline approximation to obtain
spectra at the Earth surface. Figure 1 shows these spectra for two benchmark
values of the mediator mass $m_{V}$, and for a DM mass of 50 TeV. When
$m_{V}=50$ GeV, electroweak corrections to the spectra are not important.
Considering as an example the $V\to\nu_{\mu}\bar{\nu}_{\mu}$ channel, one can
then understand the shape of its spectrum as follows: in the mediator frame,
the spectrum consists of a delta-function, the energy of each neutrino is half
the mass of the mediator. When this delta is boosted to the frame of the DM
pair, it gives rise to neutrinos spread over all energies and up to the DM
mass, as visible in the left-hand panel of Figure 1. Instead, when
$m_{V}=1000$ GeV electroweak corrections are important, and they are for
example responsible for the “bump” visible at low energies in the
$V\to\nu_{\mu}\bar{\nu}_{\mu}$ channel: the neutrinos from the decay of the
mediator can radiate a $W$ or a $Z$ boson, which in turn will give rise to
more neutrinos at smaller energies. Analogous considerations apply to the
other $V$ decay channels. Apart from oscillation effects, the primary energy
spectra above coincide with the spectra at the ANTARES location, as neutrinos
of these energies propagate undisturbed in the Galaxy. In this analysis, the
following decay channels of mediators $V$ into Standard Model particles have
been considered:
$V\rightarrow\mu^{+}\mu^{-},\;\tau^{+}\tau^{-},\;b\bar{b},\;\nu_{\mu}\bar{\nu}_{\mu}\,.$
(1)
Each of these channels is treated independently with a branching ratio of
100%.
Figure 1: Energy distribution of the muon neutrinos plus antineutrinos at
Earth location, per single annihilation into two mediators $V$ of a pair of DM
particles each with mass of 50 TeV. The mediator decays to the SM pair
indicated in the legend, then all (anti)neutrino flavours coming from that
specific pair are included and contribute via long-distance oscillations to
the muon (anti)neutrinos at Earth location. The mediator mass is 50 GeV in the
left-hand plot and 1 TeV in the right-hand one.
## 3 Detector and data set
The ANTARES neutrino detector is situated underwater in the Mediterranean Sea
40 km offshore from Toulon (France). It is composed of 12 lines instrumented
with photomultiplier tubes for the detection of Cherenkov light [31]. ANTARES
records Cherenkov light induced by charged particles originated in the
interaction of a neutrino inside the detector or in the volume around it.
Based on these recorded signals, the neutrino energy and arrival direction are
reconstructed and constitute the main information of processed data. In the
text that follows the term neutrinos stands for $\nu$ and ${\bar{\nu}}$, as
the events generated by their interactions are seen indistinguishably in
neutrino telescopes. Muons produced in cosmic ray interactions in the
atmosphere form a very large background which is suppressed in analyses by
considering only events with arrival directions crossing the Earth. The
Galactic Centre, located at a declination of $-29.01^{\circ}$S, is visible
from the detector latitude about 70$\%$ of the time [32]. In this analysis, 9
years of muon tracks, mostly induced by upward-going $\nu_{\mu}$ charged
current (CC) interactions and collected between May 2007 and December 2015,
were searched. This sample is composed of 7637 reconstructed tracks recorded
over 2101.6 days of effective livetime. Tracks are reconstructed with a good
angular resolution of the order of 1∘ at the energies relevant for this search
[33]; this data set coincides with the one analysed in previous works [34].
Tracks are reconstructed from the calibrated positions [35] and calibrated hit
times [36] of photomultiplier hits recorded in coincidence with the event. A
quality parameter $\Lambda$ is associated to each reconstructed track, based
on a maximum likelihood obtained for the reconstruction fit [37]. In its
geometrical layout, the ANTARES detector is designed for the detection of
astrophysical neutrino fluxes, which ensures a good coverage of the energy
range necessary for neutrinos from heavy dark matter annihilation. The amount
of Cherenkov photons induced in the paths of the propagating charged particles
is proportional to the amount of deposited energy and, consequently, the
number of hit optical modules, $N_{\mbox{\tiny{HIT}}}$, is a proxy of the
neutrino energy $E_{\nu}$. A set of simulated data has been produced in
correspondence with the environmental and trigger conditions of each ANTARES
data run [38]. To reproduce the expected signal from secluded dark matter, the
simulated event energy is weighted with a factor obtained according to the
energy distributions of each annihilation channel computed following [12] and
shown in Figure 1.
## 4 Analysis method
The signature of secluded dark matter annihilation would be, as other dark
matter signals, very difficult to distinctively identify. In this analysis, a
stricter event selection has been applied with respect to previous searches
for weakly interacting massive particles [24], to setup a more assertive test
of the non-standard scenario, as detailed in Section 4.1. With the preliminary
event selection described in this section, the sample is cleaned off the
majority of atmospheric muons mis-reconstructed as upgoing that failed to be
removed by standard analysis cuts. The remaining atmospheric neutrinos plus a
possible component from dark matter annihilations compose our ‘pre-selected’
data sample. An unbinned maximum likelihood method is applied to this pre-
selected sample to search for signals of secluded dark matter over the
underlying background of atmospheric neutrinos. The discrimination between
atmospheric neutrinos and neutrinos from dark matter annihilation is based on
a space and morphology information on the location of the source, and on a
spectral information based on the knowledge of the energy distribution of each
DM annihilation channel. This method has been used in previous analyses such
as [34, 24].
### 4.1 Data Selection
A set of relaxed starting cuts is initially applied to the data sample in
order to reduce a large fraction of background from atmospheric muons, and
perform consistency checks between data and Monte Carlo simulation. Similarly
to other DM analyses by the ANTARES Collaboration [24, 37, 34], these cuts
regard quality indicators of the reconstructed events: the likelihood
$\Lambda$ for the linear fit interpolating the hit pattern, and the angular
uncertainty $\beta$ estimating the angular error on the track arrival
direction. The condition for the reconstructed event to be coming from across
the Earth is required with a cut on $\theta$, zenith angle of the
reconstructed track (with respect to an axis pointing up to the vertical).
Initially, events fulfilling $\Lambda>-5.6$, $\beta<1^{\circ}$ and
$\theta<90^{\circ}$ are selected. With these starting cuts, the suppression of
atmospheric muons, electronic noise and poorly reconstructed tracks is
ensured. A stricter set of variable cuts is then applied as summarised in what
follows. The data selection was optimised by maximizing the sensitivity on the
signal by varying $\Lambda$ (from $-5.4$ to $-4.8$ in steps of 0.2). A cut on
the neutrino energy is introduced, motivated by the shape of the energy
distribution of annihilating secluded dark matter (see Figure 1), which
favours events in the high-energy end of the spectrum. On the contrary, a
power law describes the atmospheric neutrino spectrum. In this analysis the
number of recorded light hits $N_{\mathrm{HIT}}$ is used as a proxy for the
reconstructed neutrino energy. In addition to the cuts on zenith angle,
$\Lambda$ and $\beta$, $N_{\mathrm{HIT}}$ is varied up to 200 in steps of 1.
The cut value leading to the strongest sensitivity in velocity averaged cross
section for DM annihilation $\langle\sigma v\rangle$ is chosen and applied to
the unblinded data set. While the best value for $\Lambda$ remains fixed at
$-5.2$, consistently with previous similar analyses [34], the cut value on
$N_{\mathrm{HIT}}$ varies according to the dark matter mass $m_{\mathrm{DM}}$,
annihilation channel, and mediator mass $m_{V}$; the corresponding values are
reported in Table 1. The flux of signal events is obtained dividing the number
of signal events (or limit) by the integrated acceptance, defined as the
integral of the effective area $A_{\mathrm{EFF}}$ weighted by the dark matter
annihilation spectrum
$\mathcal{A}(m_{\mathrm{DM}})=\int_{0}^{m_{\mathrm{DM}}}A_{\mathrm{EFF}}^{\nu}(E_{\nu})\frac{dN_{\nu}(E_{\nu})}{dE_{\nu}}dE_{\nu}+A_{\mathrm{EFF}}^{\bar{\nu}}(E_{\nu})\frac{dN_{\bar{\nu}}(E_{\bar{\nu}})}{dE_{\bar{\nu}}}dE_{\bar{\nu}}.$
(2)
Tightening any cut improves the purity of the signal sample but reduces the
acceptance. Initially, values for the $N_{\mathrm{HIT}}$ cuts have been chosen
such to reduce the acceptance to 90%, 75%, 50%, 25%, 10% with respect to the
uncut value (where uncut means including all other cuts). However, the values
25% and 10% result in a suppression of too many events for performing the
likelihood analysis. In these cases, it was impossible to successfully scan
the skymap looking for a signal cluster, and therefore these cut values were
not further considered in the analysis. The corresponding acceptances are
shown in Figure 2 for the four annihilation modes $V\rightarrow b\bar{b}$,
$V\rightarrow\mu^{+}\mu^{-}$, $V\rightarrow\nu_{\mu}\bar{\nu}_{\mu}$ and
$V\rightarrow\tau^{+}\tau^{-}$.
${\begin{array}[]{cc|ccccccccc|cccccc|}&&&&&&&&m_{\mathrm{DM}}&&&&&&\\\
&&&&&&\mathrm{TeV}&&&&&&&\mathrm{PeV}&\\\
&{\mathrm{channel}}&3&15&30&50&100&150&200&400&600&1&1.5&2.5&4&6\\\
\hline\cr\hline\cr&\mu&31&33&35&36&38&39&40&77&82&87&92&97&102&106\\\
m_{V}=&\tau&31&33&34&35&37&38&39&74&77&82&86&91&96&99\\\
\mathrm{50\,GeV}&b&29&31&32&32&33&34&35&36&37&38&39&40&75&78\\\
&\nu_{\mu}&31&34&36&38&40&75&78&86&91&97&102&107&111&113\\\
\hline\cr&\mu&31&33&35&36&38&39&40&77&82&87&92&97&102&106\\\
m_{V}=&\tau&31&33&34&35&37&38&39&74&77&82&86&91&96&99\\\
\mathrm{250\,GeV}&b&29&31&32&32&33&34&35&36&37&38&39&71&75&78\\\
&\nu_{\mu}&31&34&36&38&70&75&78&86&91&97&102&107&111&113\\\
\hline\cr&\mu&31&33&35&36&38&39&52&77&82&87&92&97&102&106\\\
m_{V}=&\tau&31&33&34&35&37&38&39&74&77&82&86&91&96&99\\\
\mathrm{1\,TeV}&b&29&31&32&32&33&34&35&36&37&38&39&71&75&78\\\
&\nu_{\mu}&31&34&36&38&40&75&78&86&91&97&102&107&111&113\\\ \hline\cr\\\
\end{array}}$
Table 1: Cut values on the number of hits $N_{\mathrm{HIT}}$, optimised for
best sensitivity for each dark matter mass $m_{\mathrm{DM}}$, annihilation
channel, mediator mass $m_{V}$.
Integrated acceptance [m2]
$b$channel$\mu$channel
Integrated acceptance [m2]
$\nu_{\mu}$channel$m_{\textrm{DM}}$[GeV]$m_{\textrm{DM}}$[GeV]$m_{\textrm{DM}}$[GeV]$m_{\textrm{DM}}$[GeV]$\tau$channel
Integrated acceptance [m2]
Integrated acceptance [m2]
Figure 2: Integrated acceptances, as defined in Equation 2, computed for four
annihilation spectra (indicated over each panel). Acceptances computed without
any $N_{\mathrm{HIT}}$ cut are shown with a red line; different cut values in
$N_{\mathrm{HIT}}$ as reported in Table 1, leading to a reduction of
acceptance from 90% to 10%
, are indicated with blue shades from brighter to darker respectively.
The search for a signal of secluded dark matter is optimised in a $blind$ way,
according to which the events are shuffled in right ascension to ensure the
unbiased optimisation of selection cuts. When blinding, a random number
$\alpha_{\mathrm{blind}}\in[-180^{\circ},180^{\circ}]$ is assigned as right
ascension of the reconstructed arrival direction. After establishing the best
sensitivities, the original right ascension coordinate is set back. The
blinding procedure does not alter the expected sky distribution of atmospheric
neutrinos, in which the declination coordinate is maintained.
### 4.2 Signal Identification
Based on the information about the expected signal, an unbinned maximum
likelihood algorithm has been used as a search method. Unbinned likelihood is
a fitting method based on the prior knowledge of probability distribution
functions (PDFs) of signal and background discriminating variables. The method
used here does not differ from the one already applied in other ANTARES
analyses such as [34, 24], and comprehends the following steps:
1. 1.
Computation of PDFs for the signal, based on the spectra described in [12] and
DM halo morphology described with Navarro–Frenk–White (NFW) profile of spatial
mass distribution of DM profile [39]
$\rho_{DM}(r)=\rho_{s}\;r_{s}/(r(1+r^{2}/r_{s}^{2}))$, with
$\rho_{s}=1.40\cdot 10^{7}M_{\odot}$/kpc3 and $r_{s}=16.1$ kpc [24].
Computation of PDF for the distribution of atmospheric background events is
obtained from blind data.
2. 2.
Generation of $10^{4}$ pseudo-skymaps for each DM parameter choice
($m_{\mathrm{DM}}$, $m_{V}$, annihilation channel). A number of signal events
scanned in steps of 1 between 0 and 50 is injected in addition to the total
number of events taken from the background sample. Each choice of a number of
signal events makes a population of pseudo-skymaps.
3. 3.
Maximisation of likelihood yielding a test statistic (TS) distribution for
each population of pseudo-skymaps. A convolution with a Poisson function is
performed to include fluctuations expected in the distribution of signal
events.
4. 4.
Computation of 90% confidence level (CL) median upper limit in number of
detectable events that will be referred to as sensitivity according to the
Neyman method [40]. The flux sensitivity is obtained dividing the sensitivity
on number of events by the integrated acceptance.
5. 5.
Unblinding: determine the likelihood of the real data distribution and, if no
evidence of excess, computation of limits at 90% CL on flux and velocity
averaged annihilation cross section $\langle\sigma v\rangle$. These results
are presented in Section 5.
In order to quantify the signal component, a TS is defined as the ratio
between the maximum likelihood and the likelihood of the pure background
sample. Sensitivities at 90% CL are obtained comparing the TS distribution for
different numbers of injected signal events with the median of pure background
distribution, and selecting the population which is confused with background
less than 10% of the times. The number of events $n_{s}^{*}$ reconstructed
with maximum likelihood in each set of pseudo-skymaps is subject to
fluctuations following a Poisson distribution. To include fluctuations, a
transformation through a Poisson function, $\mathcal{P}$, is performed,
returning the TS distribution $P(\mathrm{TS})$ as a function of the Poissonian
mean $\mu$:
$P\left(\mathrm{TS}(\mu)\right)=\sum_{n^{*}_{s}=1}^{N}P\left(\mathrm{TS}(n_{s}^{*})\right)\,\mathcal{P}(n_{s}^{*},\mu).$
(3)
As in similar analyses [34], to take into account systematics on the expected
number of $\nu_{\mu}$ CC reconstructed events, a smearing of the test
statistic with a 15% width Gaussian is performed [33].
## 5 Results and discussion
This search for heavy secluded dark matter is performed as a function of three
free parameters: the dark matter candidate mass $m_{\mathrm{DM}}$, the
mediator mass $m_{V}$ (with general condition $m_{V}\ll m_{\mathrm{DM}}$) and
the annihilation channel. As mentioned before, the galactic DM halo profile
has been fixed to the NFW parameterisation [39]. As explained in the previous
section, a set of optimal cuts, identified independently for each parameter
choice (14 dark matter masses, 3 mediator masses and four annihilation
channels) is applied to the data in $14\times 3\times 4$ unblindings. Data is
found to be consistent with the background-only hypothesis. Upper limits at
90% CL on the thermally averaged cross section for self-conjugate DM pair
annihilation are computed for a light ($m_{V}=50$ GeV), a medium ($m_{V}=250$
GeV) and a heavy ($m_{V}=1$ TeV) mediator. The most stringent limits are
obtained in the direct channel $\mathrm{DM}\,\mathrm{DM}\rightarrow
2\,V\rightarrow 4\nu$, which is due to the spectral shape of the 4$\nu$
annihilation mode, which among those considered yields the largest fraction of
neutrinos at high energies. Each annihilation mode is independently considered
with a branching ratio of 100%. Figures 3 and 4 display the results of the
upper limits for each channel separately, alongside with the sensitivities and
corresponding $1\sigma$ and $2\sigma$ containment bands shaded in green and
yellow respectively. 1000 sets of simulated data are generated and used to
determine the sensitivity as the mean expected exclusion, and the bands as the
680 and 950 closest lines to the sensitivity. This procedure allows one to
visualise possible statistical fluctuations of the background: the fact that
our limits stay within the bands means that data are compatible with the
background hypothesis at better than $2\sigma$. Limits have been raised to be
equal to sensitivities in case of underfluctuations, analogously to similar
ANTARES analyses [34].
Figure 3: Upper limits at 90% CL on the thermally averaged DM pair
annihilation cross section $\langle\sigma v\rangle$ for a mediator mass
$m_{V}$= 1 TeV, with 1$\sigma$ and 2$\sigma$ containment bands, for $4\mu$
(top panel, limits as blue boxes) and $4\tau$ (bottom panel, limits as orange
triangles) final states.
Figure 4: Upper limits at 90% CL on the thermally averaged DM pair
annihilation cross section $\langle\sigma v\rangle$ for a mediator mass
$m_{V}$= 1 TeV, with 1$\sigma$ and 2$\sigma$ containment bands, for $4b$ (top
panel, limits as cyan boxes) and $4\nu$ (bottom panel, limits as brown
triangles) final states.
To understand which DM models are tested, Figures 5 and 6 display also the
upper limits together with the two lines indicating the unitarity limit on the
annihilation cross section. They assume, respectively, that annihilation is
dominated by s-wave processes [4, 6]
$\sigma v<\frac{4\pi}{v}\frac{1}{m_{\text{DM}}^{2}},$ (4)
or that DM is a composite state with size $R\simeq(10~{}\text{GeV})^{-1}$
$\sigma
v<\frac{4\pi}{v}\frac{1}{m_{\text{DM}}^{2}}\Big{(}1+m_{\text{DM}}vR\Big{)}^{2}\,.$
(5)
The extra term in Eq. (5) with respect to Eq. (4) is the result of the sum
over all partial waves $\sum_{j=0}^{j_{\text{max}}}(2j+1)$, where
$j_{\text{max}}=m_{\text{DM}}vR$, see e.g. [4, 6] for more details. These two
lines should be regarded as rough indications of which models are tested by
the searches presented in this paper. One for example learns that the DM
models, for which DM masses heavier than 100 TeV can be tested, are those
where more than a single partial wave contributes significantly to the
annihilation cross section. Composite DM is a limiting case where a large
number of partial waves contributes, see e.g. [41, 42, 43]. Other DM models
that evade the unitarity limit feature an indirect detection phenomenology
analogous to the one of secluded models (e.g., supercooled composite DM [44,
45]), so they are also constrained by the searches presented here. The
interest of the limits presented in this paper goes therefore beyond the
cosmological histories with early matter domination, sketched in the
introduction.
## 6 Conclusions
A search in ANTARES data from 2007 to 2015 for a signal, coming from the
Galactic Centre, due to the annihilation of secluded dark matter particles was
presented. Data were found to be consistent with the background-only
hypothesis, so that limits on the velocity averaged cross sections for
annihilation were placed for DM candidate masses between 3 TeV and 6 PeV.
These limits have been compared with theoretical expectations for the maximal
possible annihilation signals in different models.
Previous DM searches with ANTARES have used the information on the energy of
each event (i.e. the number of hits) as an input variable for the likelihood,
and then computed limits integrating between the minimal ANTARES sensitivity
and the largest energy allowed by the signal model [24]. This search instead
constitutes the first case where the information on the energy is used to
preselect events, namely the lowest energy has been varied with the parameters
of the signal model tested, to optimise the sensitivity of ANTARES in testing
it. To the best of our knowledge, this also constitutes the first time that
any telescope tested annihilation signals from DM with masses up to the PeV
range.
## Acknowledgements
The authors acknowledge the financial support of the funding agencies: Centre
National de la Recherche Scientifique (CNRS), Commissariat à l’énergie
atomique et aux énergies alternatives (CEA), Commission Européenne (FEDER fund
and Marie Curie Program), Institut Universitaire de France (IUF), LabEx
UnivEarthS (ANR-10-LABX-0023 and ANR-18-IDEX-0001), Région Île-de-France (DIM-
ACAV), Région Alsace (contrat CPER), Région Provence-Alpes-Côte d’Azur,
Département du Var and Ville de La Seyne-sur-Mer, France; Bundesministerium
für Bildung und Forschung (BMBF), Germany; Istituto Nazionale di Fisica
Nucleare (INFN), Italy; Nederlandse organisatie voor Wetenschappelijk
Onderzoek (NWO), the Netherlands; Executive Unit for Financing Higher
Education, Research, Development and Innovation (UEFISCDI), Romania;
Ministerio de Ciencia, Innovación, Investigación y Universidades (MCIU):
Programa Estatal de Generación de Conocimiento (refs. PGC2018-096663-B-C41,
-A-C42, -B-C43, -B-C44) (MCIU/FEDER), Generalitat Valenciana: Prometeo
(PROMETEO/2020/019), Grisolía (refs. GRISOLIA/2018/119, /2021/192) and GenT
(refs. CIDEGENT/2018/034, /2019/043, /2020/049, /2021/023) programs, Junta de
Andalucía (ref. A-FQM-053-UGR18), La Caixa Foundation (ref. LCF/BQ/IN17/
11620019), EU: MSC program (ref. 101025085), Spain; Ministry of Higher
Education, Scientific Research and Innovation, Morocco, and the Arab Fund for
Economic and Social Development, Kuwait. We also acknowledge the technical
support of Ifremer, AIM and Foselev Marine for the sea operation and the CC-
IN2P3 for the computing facilities. F. Sala acknowledges funding support from
the Initiative Physique des Infinis (IPI), a research training program of the
Idex SUPER at Sorbonne Université.
Figure 5: Upper limits at 90% CL on the thermally averaged cross section for
DM pair annihilation $\langle\sigma v\rangle$, from the analysis of 9 years of
ANTARES data, for mediator masses $m_{V}$= 50 GeV, and for the mediator decay
channels 4$\mu$, 4$\tau$, 4$b$, 4$\nu_{\mu}$. The dashed lines denote the
unitarity limit on the DM annihilation cross section in two limiting cases,
one where only the $s$-wave dominates the scattering, and one where DM is a
composite object with size $R\simeq(10~{}\text{GeV})^{-1}$. In each of these
cases, the parameter space above the related line is theoretically
inaccessible, see text for more details.
Figure 6: Same as Figure 5 for mediator masses $m_{V}$= 250 GeV (top panel)
and 1 TeV (bottom panel).
## References
* [1] ATLAS Collaboration, _https://atlaspo.cern.ch/public/summary $\\_$plots/_, 2022.
* [2] CMS Collaboration, _http://cms-results.web.cern.ch/cms-results/public-results/publications/_ , 2022\.
* [3] S. Dimopoulos, G. F. Giudice and A. Pomarol, _Dark matter in theories of gauge mediated supersymmetry breaking_ , _Phys. Lett._ B389 (1996) 37 [hep-ph/9607225].
* [4] K. Griest and M. Kamionkowski, _Unitarity Limits on the Mass and Radius of Dark Matter Particles_ , _Phys. Rev. Lett._ 64 (1990) 615.
* [5] B. von Harling and K. Petraki, _Bound-state formation for thermal relic dark matter and unitarity_ , _JCAP_ 12 (2014) 033 [1407.7874].
* [6] J. Smirnov and J. F. Beacom, _TeV-Scale Thermal WIMPs: Unitarity and its Consequences_ , _Phys. Rev. D_ 100 (2019) 043029 [1904.11503].
* [7] G. F. Giudice, E. W. Kolb and A. Riotto, _Largest temperature of the radiation era and its cosmological implications_ , _Phys. Rev. D_ 64 (2001) 023508 [hep-ph/0005123].
* [8] M. Pospelov, A. Ritz and M. B. Voloshin, _Secluded WIMP Dark Matter_ , _Phys. Lett._ B662 (2008) 53 [0711.4866].
* [9] A. Berlin, D. Hooper and G. Krnjaic, _PeV-Scale Dark Matter as a Thermal Relic of a Decoupled Sector_ , _Phys. Lett. B_ 760 (2016) 106 [1602.08490].
* [10] A. Berlin, D. Hooper and G. Krnjaic, _Thermal Dark Matter From A Highly Decoupled Sector_ , _Phys. Rev. D_ 94 (2016) 095019 [1609.02555].
* [11] M. Cirelli, P. Panci, K. Petraki, F. Sala and M. Taoso, _Dark Matter’s secret liaisons: phenomenology of a dark U(1) sector with bound states_ , _JCAP_ 1705 (2017) 036 [1612.07295].
* [12] M. Cirelli, Y. Gouttenoire, K. Petraki and F. Sala, _Homeopathic Dark Matter, or how diluted heavy substances produce high energy cosmic rays_ , _JCAP_ 02 (2019) 014 [1811.03608].
* [13] C. W. Bauer, N. L. Rodd and B. R. Webber, _Dark matter spectra from the electroweak to the Planck scale_ , _JHEP_ 06 (2021) 121 [2007.15001].
* [14] P. Ciafaloni, D. Comelli, A. Riotto, F. Sala, A. Strumia and A. Urbano, _Weak Corrections are Relevant for Dark Matter Indirect Detection_ , _JCAP_ 1103 (2011) 019 [1009.0224].
* [15] M. Cirelli, G. Corcella, A. Hektor, G. Hutsi, M. Kadastik, P. Panci et al., _PPPC 4 DM ID: A Poor Particle Physicist Cookbook for Dark Matter Indirect Detection_ , _JCAP_ 1103 (2011) 051 [1012.4515].
* [16] A. Sommerfeld, _Über die Beugung und Bremsung der Elektronen_ , _Ann. Phys._ 403 (1931) 257.
* [17] A. D. Sakharov, _Interaction of an Electron and Positron in Pair Production_ , _Zh. Eksp. Teor. Fiz._ 18 (1948) 631.
* [18] M. Pospelov and A. Ritz, _Astrophysical Signatures of Secluded Dark Matter_ , _Phys. Lett._ B671 (2009) 391 [0810.1502].
* [19] J. D. March-Russell and S. M. West, _WIMPonium and Boost Factors for Indirect Dark Matter Detection_ , _Phys. Lett._ B676 (2009) 133 [0812.0559].
* [20] W. Shepherd, T. M. P. Tait and G. Zaharijas, _Bound states of weakly interacting dark matter_ , _Phys. Rev. D_ 79 (2009) 055022 [0901.2125].
* [21] J. d. D. Zornoza, _Review on Indirect Dark Matter Searches with Neutrino Telescopes_ , _Universe_ 7 (2021) 415.
* [22] A. Albert et al., _Search for Dark Matter Annihilation in the Earth using the ANTARES Neutrino Telescope_ , _Phys. Dark Univ._ 16 (2017) 41 [1612.06792].
* [23] S. Adrian-Martinez et al., _Limits on Dark Matter Annihilation in the Sun using the ANTARES Neutrino Telescope_ , _Phys. Lett. B_ 759 (2016) 69 [1603.02228].
* [24] A. Albert et al., _Search for dark matter towards the Galactic Centre with 11 years of ANTARES data_ , _Phys. Lett. B_ 805 (2020) 135439 [1912.05296].
* [25] S. Adrián-Martínez et al., _A search for Secluded Dark Matter in the Sun with the ANTARES neutrino telescope_ , _JCAP_ 2016 (2016) 016 [1602.07000].
* [26] M. Ardid, I. Felis, A. Herrero and J. A. Martínez-Mora, _Constraining Secluded Dark Matter models with the public data from the 79-string IceCube search for dark matter in the Sun_ , _JCAP_ 2017 (2017) 010 [1701.08863].
* [27] C. Niblaeus, A. Beniwal and J. Edsjö, _Neutrinos and gamma rays from long-lived mediator decays in the Sun_ , _JCAP_ 2019 (2019) 011 [1903.11363].
* [28] K. Jedamzik, _Big bang nucleosynthesis constraints on hadronically and electromagnetically decaying relic neutral particles_ , _Phys. Rev._ D74 (2006) 103509 [hep-ph/0604251].
* [29] K. Petraki, M. Postma and J. de Vries, _Radiative bound-state-formation cross-sections for dark matter interacting via a Yukawa potential_ , _JHEP_ 04 (2017) 077 [1611.01394].
* [30] G. Elor, N. L. Rodd, T. R. Slatyer and W. Xue, _Model-Independent Indirect Detection Constraints on Hidden Sector Dark Matter_ , _JCAP_ 1606 (2016) 024 [1511.08787].
* [31] M. Ageron et al., _ANTARES: the first undersea neutrino telescope_ , _Nucl. Instrum. Meth. A_ 656 (2011) 11 [1104.1607].
* [32] T. Pradier, _Coincidences between Gravitational Wave Interferometers and High Energy Neutrino Telescopes_ , _Nucl. Instrum. Meth. A_ 602 (2009) 268 [0807.2562].
* [33] A. Albert et al., _First all-flavor neutrino pointlike source search with the ANTARES neutrino telescope_ , _Phys. Rev. D_ 96 (2017) 082001 [1706.01857].
* [34] A. Albert et al., _Results from the search for dark matter in the Milky Way with 9 years of data of the ANTARES neutrino telescope_ , _Phys. Lett. B_ 769 (2017) 249 [1612.04595].
* [35] S. Adrian-Martinez et al., _The Positioning System of the ANTARES Neutrino Telescope_ , _JINST_ 7 (2012) T08002 [1202.3894].
* [36] J. A. Aguilar et al., _Time Calibration of the ANTARES Neutrino Telescope_ , _Astropart. Phys._ 34 (2011) 539 [1012.2204].
* [37] S. Adrián-Martínez et al., _Search for Cosmic Neutrino Point Sources with Four Year Data of the ANTARES Telescope_ , _Astrophys. J._ 760 (2012) 53 [1207.3105].
* [38] A. Albert et al., _Monte Carlo simulations for the ANTARES underwater neutrino telescope_ , _JCAP_ 2021 (2021) 064 [2010.06621].
* [39] J. F. Navarro, C. S. Frenk and S. D. M. White, _The Structure of cold dark matter halos_ , _Astrophys. J._ 462 (1996) 563 [astro-ph/9508025].
* [40] J. Neyman, _Lectures and conferences on mathematical statistics_ , _Nature_ 142 (1938) 274.
* [41] A. Mitridate, M. Redi, J. Smirnov and A. Strumia, _Dark Matter as a weakly coupled Dark Baryon_ , _JHEP_ 10 (2017) 210 [1707.05380].
* [42] R. Contino, A. Mitridate, A. Podo and M. Redi, _Gluequark Dark Matter_ , _JHEP_ 02 (2019) 187 [1811.06975].
* [43] M. Geller, S. Iwamoto, G. Lee, Y. Shadmi and O. Telem, _Dark quarkonium formation in the early universe_ , _JHEP_ 06 (2018) 135 [1802.07720].
* [44] I. Baldes, Y. Gouttenoire and F. Sala, _String Fragmentation in Supercooled Confinement and Implications for Dark Matter_ , _JHEP_ 04 (2021) 278 [2007.08440].
* [45] I. Baldes, Y. Gouttenoire, F. Sala and G. Servant, _Supercool Composite Dark Matter beyond 100 TeV_ , 2110.13926.
|
# Stability bounds of a delay visco-elastic rheological model with substrate
friction
Malik A. Dawi 1 Jose J. Muñoz1,2∗
1Laboratori de Càlcul Numèric (LaCàN)
Universitat Politècnica de Catalunya, Barcelona, Spain
2Dept. of Mathematics, Universitat Politècnica de Catalunya, Barcelona, Spain
Centre Internacional de Mètodes Numèrics en Enginyeria (CIMNE), Barcelona,
Spain.
<EMAIL_ADDRESS>
###### Abstract
Cells and tissues exhibit oscillatory deformations during remodelling,
migration or embryogenesis. Although it has been shown that these oscillations
correlate with cell biochemical signalling, it is yet unclear the role of
these oscillations in triggering drastic cell reorganisation events or
instabilities, and the coupling of this oscillatory response with tested
visco-elastic properties.
We here present a rheological model that incorporates elastic, viscous and
frictional components, and that is able to generate oscillatory response
through a delay adaptive process of the rest-length. We analyse its stability
properties as a function of the model parameters and deduce analytical bounds
of the stable domain. While increasing values of the delay and remodelling
rate render the model unstable, we also show that increasing friction with the
substrate destabilise the oscillatory response. Furthermore, we numerically
verify that the extension of the model to non-linear strain measures is able
to generate sustained oscillations that alternate between stable and unstable
regions.
keywords:Oscillations, Delay differential equations, Visco-elasticity,
friction , stability, rheology, cells.
## 1 Introduction
Oscillatory cell deformations are ubiquitous and have been quantified _in
vitro_ [18, 20] and _in vivo_ , for instance in the segmented clock of mice
[27] or during _Drosophila_ fly dorsal closure [23]. These oscillations have
been associated to biochemical dynamics [13], signalling delays [19] or myosin
concentration fluctuations [7]. We here present a rheological model that
explicitly incorporates the delay between the cell length adaptation and the
current stretch.
Time delay has been included in numerous models in biology, with applications
in biochemical negative feedback [15], cell growth and division [1, 11], or
cell maturation [10], but are less common in biomechanics. In our case we
introduce this delay in an evolution law of the cell or tissue rest-length.
Such models with varying rest-length have been applied to stress relaxation
[14], morphogenesis [5], cortical mechanics [8], or endocytosis [4]. They have
the advantage of including a measurable quantity, the rest-length [26], and
also furnishing the observed visco-elastic response. We will here adapt these
models and include the delay response in conjunction with frictional or
adhesive forces from the environment or substrate.
Our visco-elastic model mimics the standard linear solid, but expressed in
terms of delay rest-length changes, which provides the oscillatory character
of the deformation. The stability of such system has been described in [17] or
in [3] for planar frictionless dynamics of monolayers. We here extend such
analysis to a frictional substrate, and deduce the stability conditions as a
function of viscous, stiffness and friction parameters.
The stability analysis is usually carried out through the inspection of the
characteristic equation [2, 22], or semi-discretisation methods [12, 25]. We
resort to the former method, and by analysing the associated Lambert function
[21, 6], we deduce strict and simple bounds of the stability region. We
compare our analysis with some numerical solutions of the Delay Differential
Equations (DDEs).
The article is organised as follows. We describe the visco-elastic model in
Section 2 together with the delay evolution law of the rest-length. In Section
3 the stability of a linear model is analysed, and some bounds as a function
of the model parameters are given. A non-linear extension is presented in
Section 4, which is solved numerically and is analysed with the help of the
results obtained in the linearised model. Our findings are finally discussed
in the Conclusions section.
## 2 Visco-elastic model with delay
We consider a material rheology that mimics the solid standard mode: a purely
elastic stress $\sigma^{e}$ in parallel with a visco-elastic stress
$\sigma^{v}$. Figure 1 shows schematically the two branches. We assume a one-
dimensional domain $D=\left[0,l(t)\right]$, with $l(t)$ a time dependent
apparent (measurable) length of the domain.
The total stress $\sigma$ in $D$ is given by the sum of elastic and
viscoelastic contributions,
$\displaystyle\sigma=\sigma^{e}+\sigma^{v},$
where each stress component is given by
$\sigma^{e}=k_{1}\varepsilon(l(t),l_{0})$ and
$\sigma^{v}=k_{2}\varepsilon^{e}(l(t),L(t))$, with $k_{1}$ and $k_{2}$ the
associated stiffness parameters. The strain measures $\varepsilon(l(t),l_{0})$
and $\varepsilon^{e}(l(t),L(t))$ will be detailed in the next sections for the
linear and non-linear models. As yet we mention that they depend, in addition
to $l(t)$, on the initial length $l_{0}=l(0)$ and the rest-length $L(t)$ of
the visco-elastic branch. This rest-length can be interpreted as an internal
variable, whose evolution mimics the viscous response of Maxwell models [16].
Figure 1: Schematic view of 1-dimensional model, illustrating both elastic and
visco-elastic branches with dissipative friction.
More specifically, $L(t)$ changes according to the following evolution law
$\dot{L}(t)=\gamma(l(t-\tau)-L(t-\tau)),t>0.$ (1)
Henceforth we denote by a superimposed dot the time derivatives, i.e.
$\dot{(\bullet)}=d(\bullet)/dt$. Parameter $\gamma>0$ is the _remodelling
rate_ , which measures the rate at which the cell adapts its length to the
difference $l(t-\tau)-L(t-\tau)$. We have introduced the delay parameter
$\tau\geq 0$ which aims at mimicking the measured time-lag between the
chemical signalling and the internal mechanical remodelling in the cell, as
measured in different systems such as Drosophila dorsal closure [7] or in
wound healing [28], and which in these cases is in the order of a few minutes.
We also include in our model a viscous friction $\sigma_{\eta}$ with the
external substrate or environment, and given by an external force
$\sigma_{\eta}(t)=-\eta\dot{l}(t)$, with $\eta\geq 0$ a viscous coefficient
(see Figure 1). In total, the balance law,
$\sigma_{\eta}=\sigma^{e}+\sigma^{v}$ reads in our case
$-\eta\dot{l}(t)=k_{1}\varepsilon^{e}(l(t),l_{0})+k_{2}\varepsilon^{e}(l(t),L(t)),\
t>0,$ (2)
which should be solved together with the evolution law in (1). Due to the
presence of the delay $\tau$, initial conditions must be specified for
$t\in[-\tau,0]$. For simplicity, we assume constant values
$\displaystyle l(t)=l_{0},\ t\in[-\tau,0],$ (3) $\displaystyle L(t)=L_{0},\
t\in[-\tau,0],$ (4)
with $l_{0}$ and $L_{0}$ given constants. In the next sections we will analyse
the stability and oscillatory regime of the system of Delay Differential
Equations (DDE) for linear and non-linear definitions of the strain measures
$\varepsilon$ and $\varepsilon^{e}$.
## 3 Stability analysis of linear model
### 3.1 Characteristic equations and analytical bounds
In order to ease the stability analysis, we assume here linear definitions of
the strain measures:
$\displaystyle\begin{aligned} \varepsilon(l(t),l_{0})&=l(t)-l_{0},\\\
\varepsilon^{e}(l(t),L(t))&=l(t)-L(t).\end{aligned}$
Inserting these expression into the balance equation (2), the set of DDE turn
into the following form:
$\displaystyle-\eta\dot{l}(t)$
$\displaystyle=k_{1}\left(l(t)-l_{0}\right)+k_{2}\left(l(t)-L(t)\right),$
$\displaystyle t>0$ (5) $\displaystyle\dot{L}(t)$
$\displaystyle=\gamma(l(t-\tau)-L(t-\tau)),$ $\displaystyle t>0$ (6)
with the initial conditions in (3). The coupled system of DDE can be written
in a compact form as
$\dot{\mathcal{L}}(t)+\mathbf{A}\mathcal{L}(t)+\mathbf{B}\mathcal{L}(t-\tau)+\mathbf{c}=\mathbf{0},t>0,$
(7)
with
$\mathcal{L}(t)=\left\\{\begin{array}[]{c}l(t)\\\ L(t)\end{array}\right\\}\ ;\
\mathbf{A}=\left[\begin{array}[]{cc}\frac{k_{1}+k_{2}}{\eta}&-\frac{k_{2}}{\eta}\\\
0&0\end{array}\right]\ ;\ \mathbf{B}=\left[\begin{array}[]{cc}0&0\\\
-\gamma&\gamma\end{array}\right]\ ;\
\mathbf{c}=\left\\{\begin{array}[]{c}\frac{k_{1}l_{0}}{\eta}\\\
0\end{array}\right\\}.$
Generally, the solution of the coupled system of DDE in (7) is characterized
qualitatively (e.g. asymptotic, synchronous, oscillatory) by the exponents or
the roots of the characteristic function [9, 22]. In order to obtain this
characteristic function, one might search for a solution in the form,
$\mathcal{L}(t)=\sum_{i}e^{m_{i}t}\mathcal{L}_{i}+\mathcal{L}_{0},$ (8)
where $\mathcal{L}_{0}$ and $\mathcal{L}_{i}$ are constant vectors that depend
on the chosen initial values, and $m_{i}\in\mathbb{C}$ are the characteristic
exponents. Clearly if all the exponent have negative real parts, i.e.
$Re(m_{i})<0$, the solution is asymptotically stable with time. Substituting
Eq. (8) into Eq. (7) gives for each term in the summation
$\left(m_{i}\mathbf{I}+\mathbf{A}+\mathbf{B}e^{-m_{i}\tau}\right)\mathcal{L}_{i}=\mathbf{0}.$
We remark that the above linear transformation must hold regardless of the
initial conditions, that is to say, the determinant must always vanish. This
allows us to express the characteristic function of the system as the
determinant of the above matrix, which gives
$f(m):=m^{2}+\gamma me^{-m\tau}+\frac{k_{1}+k_{2}}{\eta}m+\frac{\gamma
k_{1}}{\eta}e^{-m\tau}=0.$ (9)
We decompose the characteristic function to real and imaginary parts by
substituting $m=\alpha+i\beta$ and then separating each part, leading to the
following non-linear system of equations,
$\displaystyle\text{Re}\;f(m)$
$\displaystyle=\alpha^{2}-\beta^{2}+\frac{k_{1}+k_{2}}{\eta}\alpha+\gamma
e^{-\alpha\tau}\left(\left(\alpha+\frac{k_{1}}{\eta}\right)\cos(\beta\tau)+\beta\sin(\beta\tau)\right),$
(10) $\displaystyle\text{Im}\;f(m)$
$\displaystyle=2\alpha\beta+\frac{k_{1}+k_{2}}{\eta}\beta+\gamma
e^{-\alpha\tau}\left(\beta\cos(\beta\tau)-\left(\alpha+\frac{k_{1}}{\eta}\right)\sin(\beta\tau)\right).$
The stability regions in the parameters space are defined by the borders where
the number of unstable exponents changes, which means, at least one
characteristic exponents crosses the imaginary axes from left to right. In
such case Eq. (10) will have at least one solution with positive $\alpha$.
Here, we have constructed the phase diagram by solving the system in Eq. (10)
numerically while monitoring the values of $\alpha$ (see Fig. 2). If there is
at least one root with a positive $\alpha$ the solution was considered
unstable.
Figure 2: Phase diagrams for different pairs of material parameters. (a) Plane
$(k_{1},k_{2})$, (b) plane $(k_{1},\eta)$, (c) plane $(k_{2},\eta)$ and (d)
plane $(\tau,\gamma)$. The curves show stability borders for different values
of the off-plane parameters. Continuous lines are obtained with the numerical
solution of Eq. (10). Dashed lines represent the sufficient stability
condition in Eq. (11). The regions which are labeled as stable are those with
negative values for $\alpha$ and those label as unstable indicate the regions
with at least a single positive $\alpha$.
With the aim of furnishing a practical bound for detecting stable solutions,
we also give the following result:
###### Proposition 1.
The solution of the system of delay differential equations in Eq. (7) with
initial conditions in Eq. (3) is stable as long as,
$k_{1}+k_{2}-\gamma\eta-k_{1}\gamma\tau>0.$ (11)
Proof. Condition (11) is derived resorting to the results in [24], and
analysing the so-called D-curves defined as,
$\displaystyle R(\omega):=\text{Re}\;f(i\omega)$
$\displaystyle=-\omega^{2}+\gamma\left(\frac{k_{1}}{\eta}\cos(\omega\tau)+\omega\sin(\omega\tau)\right)$
(12) $\displaystyle S(\omega):=\text{Im}\;f(i\omega)$
$\displaystyle=\frac{k_{1}+k_{2}}{\eta}\omega+\gamma\left(\omega\cos(\omega\tau)-\frac{k_{1}}{\eta}\sin(\omega\tau)\right)$
(13)
with $\omega\in[0,+\infty)$. The functions $R(\omega)$ and $S(\omega)$ provide
infinite parametric curves that mark the region with constant number of
unstable characteristic exponents. In particular, we resort to Theorem 2.19 in
[24], which indicates that the zeros of Eq. (9) have no real positive parts if
and only if,
$S(\rho_{k})\neq 0\quad k=1,..,r,$ (14)
and
$\sum_{k=1}^{r}(-1)^{k}S(\rho_{k})=-1,$ (15)
where $\rho_{1}\geq...\geq\rho_{r}\geq 0$ are the non-negative roots of
$R(\omega)$, with $r$ being an odd number. Moreover, we introduce a polynomial
$S^{-}(\omega)$ which defines a lower bound for the function $S(\omega)$ such
that,
$0<S^{-}(\omega)\leq S(\omega)\quad\quad\text{for}\quad\omega\in(0,+\infty).$
(16)
In case that $S(\omega)$ satisfies the stability conditions in Eq. (14) and
(15), $S^{-}(\omega)$ will also satisfy them by construction. An adequate
choice for the polynomial $S^{-}(\omega)$ can be obtained by exploiting the
following inequalities,
$\cos(\omega\tau)\geq-1,\quad-\sin(\omega\tau)\geq-\omega\tau\quad\quad\text{for}\quad\omega\in(0,+\infty)$
which lead to,
$S^{-}(\omega)=\Big{(}\frac{k_{1}+k_{2}}{\eta}-\gamma-\frac{k_{1}}{\eta}\gamma\tau\Big{)}\omega.$
Since $\omega>0$, the condition in Eq. (16) is satisfied as long as
$k_{1}+k_{2}-\gamma\eta-k_{1}\gamma\tau>0.\quad\quad\qed$
We point out that the main benefit of Proposition 1 is that it counts in the
whole space of system parameters, giving the opportunity to cross check the
stability taking into account the relative variations of system parameters. In
the phase diagrams in the parametric space, condition (11) is indicated by the
dashed lines in Fig. 2. As it can be observed, it indicates stability regions
that are smaller then those obtained by solving numerically Eq. (10). These
plots emphasise the fact that although the bound in Eq. (16) does not provide
a necessary condition, it provides a useful sufficient stability condition.
We remark also two salient conclusion from the expression in the bound, which
are also confirmed in the phase diagrams: increasing values of $\gamma\tau$
have an unstable effect in the lengths $l(t)$ and $L(t)$, as previously
encountered in other models [17], while decreasing values of $\eta$ may render
the oscillations stable. This is an unexpected result, since increasing
viscosity has in general a stabilising or damping effects in mechanics. This
can be explained by highlighting the retardation or delay that viscosity
entails in the stress response, similar to an increase of $\tau$.
### 3.2 Numerical simulations
In order to verify the obtained stability limits, we have preformed some
numerical tests considering the one-dimensional model presented in Fig. 1. The
test mimics a previous compression state that is given by the following
initial conditions,
$\displaystyle l(t)=L(t)=1,$ $\displaystyle\tau<t\leq 0$ (17) $\displaystyle
l(-\tau)=0.9,L(-\tau)=1.$ (18)
In order to compare our results with previous values in the literature and
with more general boundary conditions, we will also test different prescribed
values of $l(t)$ and additional external forces. Indeed, in the presence of a
constant external force $f$, the equilibrium equation in (2) reads,
$\displaystyle-\eta\dot{l}(t)+f$
$\displaystyle=k_{1}\left(l(t)-l_{0}\right)+k_{2}\left(l(t)-L(t)\right)$ (19)
$\displaystyle\dot{L}(t)$
$\displaystyle=\gamma\left(l(t-\tau)-L(t-\tau)\right)$ (20)
#### 3.2.1 Unloaded free conditions
A backward Euler implicit time discretisation of equations in (19) yields the
following set of equations, which are computed sequentially,
$\begin{split}L_{n+1}&=\Delta t\gamma(l_{n-\tau}-L_{n-\tau})\\\
l_{n+1}&=\frac{1}{(\eta/\Delta t+k_{1}+k_{2})}\left(\frac{\eta}{\Delta
t}l_{n}+f_{n+1}+k_{1}L_{0}+k_{2}L_{n+1}\right)\end{split}$ (21)
We here consider the case $f_{n}=0,n=0,1,2\ldots,200/\Delta t$ and $\Delta
t=0.01$, which is found sufficiently accurate when being compared with smaller
values. The resulting evolution of $l_{n}$ and $L_{n}$ is consistent with the
stability analysis of the previous section. The presence of the delay $\tau>0$
produces oscillatory solutions for $l$ and $L$, as it can be seen in Fig. 3.
The stability of these oscillations depends on the model parameters as
indicated in the stability diagrams in Fig. 2. The first case in Fig. 3a
corresponds to stable oscillations, with parameters inside the stability
domain, while the second case in Fig, 3b yields unstable oscillations, with
parameters that exceed the stability limits.
(a) Model parameters: $k_{1}=2$, $k_{2}=3$, $\eta=8$, $\gamma=0.5$, $\tau=6$
(b) Model parameters: $k_{1}=3$, $k_{2}=2$, $\eta=8$, $\gamma=0.5$, $\tau=6$
Figure 3: Time evolution of current length and rest-length for free unloaded
conditions. (a) Parameters belonging to the stable domain. (b) Choice of
parameters that lie outside of the stable domain.
#### 3.2.2 Prescribed deformation
We here choose a constant value of the apparent length $l(t)$, with an initial
discontinuity:
$\displaystyle L(t)=L_{0}=1,$ $\displaystyle\ -\tau\leq t\leq 0,$
$\displaystyle l(-\tau)=0.9,l=l_{0}=1,$ $\displaystyle\ -\tau<t.$
In this case, $\dot{l}(t)=0,t>0$, so the the first differential gives us a
reaction force term equal to $k_{2}(l_{0}-L(t))$, while the DDE reads
$\displaystyle\dot{L}$ $\displaystyle=\gamma(l_{0}-L(t-\tau)).$
This DDE (or equivalent forms) has been extensively studied [22, 17], and is
known to yield oscillatory values of rest-length $L(t)$ whenever
$\gamma\tau>\frac{1}{e}$, and unstable oscillations whenever
$\gamma\tau>\frac{\pi}{2}$. This has been confirmed by the numerical
simulations in Fig. 4.
(a) Model parameters: $\gamma=0.35$, $\tau=4$
(b) Model parameters: $\gamma=0.35$, $\tau=5$
Figure 4: The evolution of the rest-length with fixed values for the apparent
length $l(t)$. The stability is in this case identical to the friction-less
models [17]: (a) Oscillatory solution when $\tau\gamma>\frac{1}{e}$, (b)
unstable solution arise whenever $\tau\gamma>\frac{\pi}{2}$.
#### 3.2.3 Prescribed forces
We now impose and external force $f=0.2$. Since this value only affects the
value of the vector $\mathbf{c}$ in Eq. (5), the stability is consequently
unaffected by the value of $f$. The plots in Fig. 5 confirm this fact. These
plots show the apparent length as a function of time, while the rest-length is
shown as the contourplot on the varying domain $x\in[0,l(t)]$.
(a) Model parameters: $k_{1}=1$, $k_{2}=1$, $\eta=1$, $\gamma=0.5$, $\tau=6$
(b) Model parameters: $k_{1}=1$, $k_{2}=1$, $\eta=3$, $\gamma=0.6$, $\tau=6$
Figure 5: The evolution of the current length and the rest-length (color map)
with prescribed compression forces $f$ ($f(x=0)=0.2,\quad f(x=1)=-0.2$). (a)
the solution inside the stability domain. (b) the time evolution as the
stability limit is exceeded.
## 4 Extension to non-linear: strain–based model
We now use a non-dimensional definition of the strains
$\displaystyle\varepsilon(l(t),l_{0})$
$\displaystyle=\frac{l(t)-l_{0}}{l_{0}},$
$\displaystyle\varepsilon^{e}(l(t),L(t))$
$\displaystyle=\frac{l(t)-L(t)}{L(t)}.$
While this is a more common strain measure, with non-dimensional values, these
expressions, when inserted into the equilibrium equations in (2) yield a set
of non-linear DDE:
$\displaystyle-\eta\dot{l}(t)$
$\displaystyle=k_{1}\left(\frac{l(t)-l_{0}}{l_{0}}\right)+k_{2}\left(\frac{l(t)-L(t)}{L(t)}\right),$
(22) $\displaystyle\dot{L}(t)$
$\displaystyle=\gamma\left(l(t-\tau)-L(t-\tau)\right).$ (23)
We aim at studying the oscillatory character and stability of these equations.
However, due to their non-linearity we cannot directly apply the methodology
previously presented. We aim instead at analysing the linearised form of
equation (22) at time $t_{0}$. By setting $\delta l(t)=l(t)-l(t_{0})$ and
$\delta L(t)=L(t)-L(t_{0})$, the linear terms read,
$\displaystyle-\eta\delta\dot{l}(t)=$ $\displaystyle\frac{k_{1}}{l_{0}}\delta
l+\frac{k_{2}}{L(t_{0})}\delta l(t)-\frac{k_{2}l(t_{0})}{L(t_{0})^{2}}\delta
L(t).$ (24)
It then follows that by defining the modified stiffness parameters,
$\displaystyle\hat{k}_{1}$
$\displaystyle=\frac{k_{1}}{l(t_{0})}+\frac{k_{2}}{L(t_{0})}\left(1-\frac{l(t_{0})}{L(t_{0})}\right),$
(25) $\displaystyle\hat{k}_{2}$
$\displaystyle=\frac{k_{2}l(t_{0})}{L(t_{0})^{2}},$ (26)
equation (24) is equivalent to the linear terms in the equilibrium equation in
(5), but replacing $(k_{1},k_{2})$ by $(\hat{k}_{1},\hat{k}_{2})$ and in terms
of $\delta l(t)$ and $\delta L(t)$ instead of $l(t)$ and $L(t)$. This allows
us to understand some of the numerical solutions obtained for the non-linear
case.
Figure 6a shows the time evolution of $l(t)$ and $L(t)$, which are sustained,
that is, their asymptotic behaviour does not increase nor decrease. We plot in
the parametric space of $k_{1}$ and $k_{2}$ the modified parameters
$\hat{k}_{1}$ and $\hat{k}_{2}$ for each time $t_{0}$, as shown in Fig. 6b. It
can be observed that although the initial values are located in the unstable
region, they in turn oscillate between the unstable and stable region,
reaching a limit cycle that alternates between the two domains.
(a) Time evolution of current length and rest-length
(b) The evolution of $\tilde{k_{1}}$ and $\tilde{k_{2}}$.
Figure 6: Numerical solution with sustained oscillations of the non-linear
model. Parameters: $k_{1}=1$, $k_{2}=1$, $\eta=3$, $\gamma=0.6$, $\tau=5$
We have also tested other parameter settings, with an initial location of
($\hat{k}_{1},\hat{k}_{2}$) in the parametric space farther from the stability
boundary (see Fig. 7). In this case, the system exhibits oscillations that
reach the singular value $L(t)=0$ for some $t>0$, which renders the DDEs in
(22) ill-posed. Instead, when using values that are farther inside the
stability region, as it is the case in Fig. 8, the oscillations stabilise
before reaching this singular value. Although we are not able to furnish
bounds for non-linear stability, we can explain the presence of stable,
sustained, or unstable (or singular) oscillations according to the distance of
the initial value of $(\hat{k}_{1},\hat{k}_{2})$ to the stability boundary of
the linear case.
(a) Time evolution of current length and rest-length
(b) The evolution of $\tilde{k_{1}}$ and $\tilde{k_{2}}$
Figure 7: Numerical solution with unstable oscillations on the non-linear
model. Parameters: $k_{1}=2$, $k_{2}=1$, $\eta=3$, $\gamma=0.6$, $\tau=5$
(a) Time evolution of current length and rest-length
(b) The evolution of $\tilde{k_{1}}$ and $\tilde{k_{2}}$
Figure 8: Numerical solution with stable oscillations of the non-linear model.
Parameter $k_{1}=1$, $k_{2}=2$, $\eta=3$, $\gamma=0.6$, $\tau=5$
## 5 Conclusions
Motivated by the presence of delays and visco-elastic response of tissues, we
have presented a rheological model that includes elastic and viscous
contributions, and also exhibits oscillatory behaviour.
We have analysed the stability of he model when using a linear strain measure
and as a function of the model parameters. We have recovered previous results,
which show that increasing values of the delay $\tau$ and the remodelling rate
$\gamma$ (a quantity that is inversely proportional to tissue viscosity),
render the oscillations unstable. Remarkably, increasing values of the viscous
friction of the domain with respect to external boundary also destabilise the
system.
By studying the characteristic function of the DDE we have provided sufficient
conditions of stability and bounds to the stability region. This analysis have
also allowed us to explain the presence of sustained oscillations in a non-
linear version of the model. This persistent oscillations in the tissue
deformations are frequently observed [18, 20], and in our model are due to the
transition between stable and unstable domains.
We note that despite visco-elastic models based on rest-length changes are
increasingly common [4, 5, 14], their stability in the presence of delayed
response has not been studied. We here provide such an analysis which may also
help to explain the observed sudden deformations in embryo development and
morphogenesis.
## acknowledgements
JJM and MD have been financially supported by the Spanish Ministry of Science,
Innovation and Universities (MICINN) with grant DPI2016-74929-R and by the
local government _Generalitat de Catalunya_ with grant 2017 SGR 1278.
## References
* [1] T. Alarcón, Ph. Getto, and Y. Nakata. Stability analysis of a renewal equation for cell population dynamics with quiescence. SIAM J. Appl. Math., 74(4):1266–1297, 2014.
* [2] F. M. Asl and A. G. Ulsoy. Analysis of a system of linear delay differential equations. J. Dyn. Sys. Meas. Contr., 125:215–223, 2003.
* [3] C. Borja, E. Moral, and J.J. Muñoz. Viscoelasticity and Collective Cell Migration: An interdisciplinary perspective across levels of organization, chapter 5: Effects of time delays and viscoelastic parameters in oscillatory response cell monolayers. Elsevier, 2020. In press.
* [4] K.E. Cavanaugh, M.F. Staddon, E. Munro, S. Banerjee, and M.L. Gardel. RhoA mediates epithelial cell shape changes via mechanosensitive endocytosis. Dev. Cell, 52(2):152–166, 2020.
* [5] R. Clément, C. Collinet, B. Dehapiot, T. Lecuit, and P.F. Lenne. Viscoelastic dissipation stabilizes cell shape changes during tissue morphogenesis. Current Biol., 27(20):3132–3142, 2017.
* [6] R.M. Corless, G.H. Gonnet, D.E.G. Hare, D.J. Jeffrey, and D.E. Knuth. On the Lambert W function. Adv. Comp. Math., 5:329–359, 1996.
* [7] K. Dierkes, A. Sumi, J. Solon, and G. Salbreux. Spontaneous Oscillations of Elastic Contractile Materials with Turnover. Phys. Rev. Letters, 113:148102, 2014.
* [8] K. Doubrovinski, M. Swan, O. Polyakov, and E.F. Wieschaus. Measurement of cortical elasticity in drosophila melanogaster embryos using ferrofluids. Proc. Natl. Acad. Sci. USA, 114(5):1051–1056, 2017.
* [9] T. Erneux. Applied Delay Differential Equations, volume 3 of Surveys and Tutorials in the Applied Mathematical Sciences. Springer, New York, 2009.
* [10] P. Getto, M. Gyllenberg, Y. Nakata, and F. Scarabel. Stability analysis of a state-dependent delay differential equation for cell maturation: analytical and numerical methods. J. Math. Biol., 79:281–328, 2019.
* [11] M. Gyllenberg and H. J. A. M. Heijmans. An abstract delay-differential equation modelling size dependent cell growth and division. SIAM J. Math. Anal., 18(1):74–88, 1987.
* [12] T. Insperger and G. Stépán. Semi-discretization method for delayed systems. Int. J. Num. Meth. Engng., 55(5):503–518, 2002.
* [13] K. Kaouri, P. K. Maini, P. A. Skourides, N. Christodoulou, and S. J. Chapman. A simple mechanochemical model for calcium signalling in embryonic epithelial cells. J. Math. Biol., 78:2059–2092, 2019.
* [14] N. Khalilgharibi, J. Fouchard, N. Asadipour, R. Barrientos, M. Duda, A. Bonfanti, A. Yonis, A. Harris, P. Mosaffa, Y. Fujita, A. Kabla, Y. Mao, B. Baum, J.J. Muñoz, M. Miodownik, and G. Charras. Stress relaxation in epithelial monolayers is controlled by the actomyosin cortex. Nature Phys., 15:839–847, 2019.
* [15] A. Lapytsko and J. Schaber. The role of time delay in adaptive cellular negative feedback systems. J. Theor. Biol., 308:64–73, 2016.
* [16] J.J. Muñoz and S. Albo. Physiology-based model of cell viscoelasticity. Phys. Rev. E, 88(1):012708, 2013.
* [17] J.J. Muñoz, M. Dingle, and M. Wenzel. Mechanical oscillations in biological tissues as a result of delayed rest-length changes. Phys. Rev. E, 98(1):052409, 2018.
* [18] V. Petrolli, M.L. Goff, M. Tadrous, K. Martens, C. Allier, O. Mandula, L. Hervé, S. Henkes, R. Sknepnek, T. Boudou, G. Cappello, and M. Balland. Confinement-induced transition between wave-like collective cell migration modes. Phys. Rev. Letters, 122(16):168101, 2019.
* [19] G. Petrungaro, L. Morelli, and K. Uriu. Information flow in the presence of cell mixing and signalling delays during embryonic development. Sem. Cell Dev. Biol., 93:23–35, 2019.
* [20] G. Peyret, R. Mueller, J. d’Alessandro, S. Begnaud, P. Marcq, R.M. Mège, J.M. Yeomans, A. Doostmohammadi, and B. Ladoux. Sustained oscillations of epithelial cell sheets. Bioph. J., 117(3):454–478, 2019.
* [21] H. Shinozaki and T. Mori. Robust stability analysis of linear time-delay systems by Lambert W function: Some extreme point results. Automat., 42(1):1791–1799, 2006.
* [22] H. Smith. An Introduction to Delay Differential Equations with Applications to the Life Sciences. Texts in Applied Mathematics. Springer, New York, USA, 2011.
* [23] J Solon, A Kaya-Copur, and D Brunner. Pulsed forces timed by a ratchet-like mechanism drive directed tissue movement during dorsal closure. Cell, 58(137):1331–1342, 2009.
* [24] G. Stépán. Retarded dynamical systems: stability and characteristic functions, volume 210 of Pitman Res. Notes Math. Longman Scientific & Technical, Essex, UK, 1989.
* [25] H.T. Sykora, D. Bachrathy, and G. Stépán. Stochastic semi-discretization for linear stochastic delay differential equations. Int. J. Num. Meth. Engng., 119(9):879–898, 2019.
* [26] T.P. J. Wyatt, J. Fouchard, A. Lisica, N. Khalilgharibi, B. Baum, P. Recho, A.J. Kabla, and G.T. Charras. Actomyosin controls planarity and folding of epithelia in response to compression. Num. Math., 19:109–117, 2020. https://doi.org/10.1038/s41563-019-0461-x.
* [27] K. Yoshioka-Kobayashi, M. Matsumiya, Y. Niino, A. Isomura, H. Kori, A. Miyawaki, and R. Kageyama. Coupling delay controls synchronized oscillation in the segmentation clock. Nature, 580(7801):119–123, 2020.
* [28] T. Zulueta-Coarasa and R. Fernandez-Gonzalez. Dynamic force patterns promote collective cell movements during embryonic wound repair. Nature Phys., 14:750–758, 2018.
|
11footnotetext: ∗ Indicates equal contribution.22footnotetext: ${\dagger}$
Indicates corresponding author.
# Seq1F1B: Efficient Sequence-Level Pipeline Parallelism for Large Language
Model Training
Ao Sun1,∗ Weilin Zhao2,∗ Xu Han2,† Cheng Yang1,† Zhiyuan Liu2 Chuan Shi1
Maosong Sun2
1 Beijing University of Posts and Telecommunications, Beijing, China.
2 NLP Group, DCST, IAI, BNRIST, Tsinghua University, Beijing, China.
{maydomine<EMAIL_ADDRESS>
<EMAIL_ADDRESS>
<EMAIL_ADDRESS>
###### Abstract
The emergence of large language models (LLMs) relies heavily on distributed
training strategies, among which pipeline parallelism plays a crucial role. As
LLMs’ training sequence length extends to 32k or even 128k, the current
pipeline parallel methods face severe bottlenecks, including high memory
footprints and substantial pipeline bubbles, greatly hindering model
scalability and training throughput. To enhance memory efficiency and training
throughput, in this work, we introduce an efficient sequence-level one-
forward-one-backward (1F1B) pipeline scheduling method tailored for training
LLMs on long sequences named Seq1F1B. Seq1F1B decomposes batch-level
schedulable units into finer sequence-level units, reducing bubble size and
memory footprint. Considering that Seq1F1B may produce slight extra bubbles if
sequences are split evenly, we design a computation-wise strategy to partition
input sequences and mitigate this side effect. Compared to competitive
pipeline baseline methods such as Megatron 1F1B pipeline parallelism, our
method achieves higher training throughput with less memory footprint.
Notably, Seq1F1B efficiently trains a LLM with 30B parameters on sequences up
to 64k using 64 NVIDIA A100 GPUs without recomputation strategies, a feat
unachievable with existing methods. Our source code is based on Megatron-LM,
and now is avaiable at: https://github.com/MayDomine/Seq1F1B.git.
## 1 Introduction
In recent years, there has been a growing interest in large language models
(LLMs), which have revolutionized various tasks in natural language processing
Touvron et al. [2023], Reid et al. [2024], Jiang et al. [2024], Anil et al.
[2023]. Efficient distributed training strategies Korthikanti et al. [2023],
Rasley et al. [2020], Rajbhandari et al. [2020], Shoeybi et al. [2019],
Narayanan et al. [2021a] play a crucial role in training these large models.
Among these strategies, pipeline parallelism Shoeybi et al. [2019], Huang et
al. [2019], Yang et al. [2021], Qi et al. [2024], Li et al. [2021] stands out
due to its low communication bandwidth requirements and high scalability when
integrated with other distributed training strategies.
Fundamentally, pipeline parallelism involves partitioning the model into
multiple stages, with each computing device responsible for processing a stage
consisting of consecutive layers. This setup inherently leads to “bubbles”—the
idle time caused by the dependencies between the computation of sharded
layers. Several scheduling strategies, such as GPipe Huang et al. [2019], are
proposed to address this problem, significantly reducing pipeline bubbles by
splitting each mini-batch of training examples into several micro-batches.
These strategies come at the expense of increased memory usage, as each stage
must store all hidden states of the micro-batches generated during the forward
passes until the backward passes are completed. Based on GPipe, TeraPipe Li et
al. [2021] further splits each micro-batch along the sequence dimension into
micro-sequences to further reduce pipeline bubbles, but still suffer from the
high memory demand for storing the hidden states of micro-batches.
To address the issue of high memory demand, one-forward-one-backward (1F1B)
scheduling strategies are proposed Harlap et al. [2018], Fan et al. [2021],
Narayanan et al. [2021a] to rewrite GPipe to schedule backward passes in
advance and make backward passes have higher execution priority than forward
passes, without affecting final results. By adopting 1F1B parallel strategies,
the memory demand for storing hidden states can be significantly reduced
without adding extra pipeline bubbles. Other methods such as zero-bubble-
pipeline Qi et al. [2024] and 1F1B-I (1F1B with interleaved stages) Narayanan
et al. [2021a] seek to further reduce the bubbles of 1F1B strategies but at
the cost of more memory overhead and communication cost. Generally, optimizing
pipeline parallelism continues to handle trade-offs between bubble ratio and
memory overhead.
On the other hand, some recent efforts Buckman and Gelada , Reid et al. [2024]
have noticed that long-sequence training benefits LLMs in many aspects,
leading to increasingly longer training contexts for LLMs. However, LLMs can
not simply support training longer sequences due to the quadratic time and
memory complexities of Transformer attention modules in terms of sequence
length Vaswani et al. [2017]. Several efforts Dao et al. [2022], Ding et al.
[2023], Jacobs et al. [2023] to build efficient attention modules have been
proposed to address this issue. Even so, the challenge caused by long
sequences extends beyond attention modules. In distributed training scenarios,
long sequences may cause various parallel methods to fail. For pipeline
parallelism, such as GPipe Huang et al. [2019] and 1F1BHarlap et al. [2018],
Fan et al. [2021], Narayanan et al. [2021a], Qi et al. [2024], these methods
can only use micro-batches as the minimal scheduling units. In extreme cases,
a single micro-batch consisting of long sequences can lead to memory overflow.
Long sequences make the issue of high memory demand in pipeline parallelism
more serious.
A straightforward approach to solving such a problem is to split the micro-
batches along the sequence dimension. However, such simple modification is
challenging for existing 1F1B scheduling strategies such as Qi et al. [2024]
and 1F1B-I Narayanan et al. [2021a] because there are computation dependencies
between forward and backward passes across micro-sequences, making a direct
split along the sequence dimension unfeasible.
To solve such a challenge, we introduce the Seq1F1B, an efficient sequence-
level 1F1B schedule, which has higher efficiency and lower memory demands than
the traditional 1F1B methods. In detail, we introduce a partially ordered
scheduling queue in Seq1F1B to replace the first-in-first-out (FIFO)
scheduling queue in 1F1B, rewriting the scheduling strategy to preserve the
exact forward and backward semantics while providing synchronous pipeline
parallelism. To further improve Seq1F1B, we propose a strategy for balancing
the workload across sub-sequence computations. More specifically, we balance
the sub-sequence computation by designing a solution based on floating-point
operations on each sub-sequence. In this design, we addressed the imbalance of
computational workload caused by the attention mechanism by splitting
sequences based on computational workloads rather than simply dividing them
evenly along the sequence dimension.
Sufficient experiments demonstrate that Seq1F1B significantly outperforms both
the 1F1B and 1F1B-I scheduling strategies in terms of memory efficiency and
training throughput for training LLMs, with the sequence length ranging from
16k to 128k and the model size ranging from 2.7B to 32B. From the experimental
results, the efficiency of Seq1F1B becomes more pronounced as the sequence
length increases and Seq1F1B supports efficiently training a GPT with 30B
parameters on sequences up to 64k tokens using 64 NVIDIA A100 GPUs without any
recomputation strategies, which is unachievable with existing pipeline
parallel methods.
## 2 Related Work
Training large language models requires using a mixture of parallel
strategies, the most important of which are data parallelism, tensor
parallelism, and pipeline parallelism. Data parallelism scales training models
by distributing data across multiple devicesGoyal et al. [2017], Zinkevich et
al. [2010], Li et al. [2020], Xing et al. [2015], each device hosting a model
replica and synchronizing gradients. Zero redundancy optimizer (ZeRO)
Rajbhandari et al. [2020], Rasley et al. [2020], Ren et al. [2021] enhances
data parallelism’s memory efficiency by partitioning model parameters across
devices at the cost of significant communication. Tensor parallelismShoeybi et
al. [2019], Korthikanti et al. [2023] parallelize computation by partitioning
matrix multiplication. In such way, Tensor parallelism effectively enhances
computation efficiency but introduces high communication costs of aggregating
the results of matrix multiplication, making it commonly used within the
multiple workers of a single node. Since this paper focuses on improving
pipeline parallelism, we will show more details for pipeline parallelism next.
For pipeline parallelism, schedules can be broadly categorized into two main
types: synchronous and asynchronous. Asynchronous schedules such as
asynchronous PipeDream Harlap et al. [2018] and PipeMare Yang et al. [2021]
can achieve bubble-free but suffer from the performance degradation of final
trained models because they use outdated parameters to compute gradient
updates. In view of this, our work focuses on synchronous pipeline schedules,
as they ensure consistent semantics across different model parallel
strategies. GPipeHuang et al. [2019], Li et al. [2021] and 1F1BFan et al.
[2021], Narayanan et al. [2021a, b] are the most commonly used pipeline
schedules following synchronous settings. Many other works are built upon
these two foundation schedules.
The original GPipeHuang et al. [2019] simply divides a mini-batch into several
micro-batches. The scheduling process of GPipe has only two phases: the
forward and the backward phases. Only after all forward passes for the micro-
batches within a batch are complete will the backward passes be executed.
During the forward phase, the hidden states of each micro-batch are enqueued
into a first-in-first-out (FIFO) queue $Q$. During the backward phase, these
hidden states are dequeued for their corresponding backward passes. Since the
backward phase happens after all hidden states are queued, GPipe exhibits an
$O(M)$ memory consumption, where $M$ represents the number of micro-batches.
Based on GPipe, other methods such as TeraPipe Li et al. [2021] and Chimera Li
and Hoefler [2021] further optimize the bubble ratio of GPipe through
different techniques. TeraPipe relies on the observation of causal language
modeling — the computation of a given input token only depends on its previous
tokens. Specifically, TeraPipe divides GPipe’s micro-batch into multiple token
spans and replaces the FIFO queue with a last-in-first-out (LIFO) queue to
ensure the correct computation of gradients during attention backward passes.
By leveraging finer scheduling units, TeraPipe effectively reduces the bubble
ratio while being more memory-efficient than GPipe. Chimera adopts
bidirectional pipeline parallelism, where each computing device is responsible
for the workload of multiple stages. While this approach reduces the bubble
ratio, each device has to store redundant parameters (as stages are not evenly
distributed across devices), leading to increased memory usage.
Different from GPipe, which performs backward passes after completing all
forward passes, 1F1B Narayanan et al. [2021a], Fan et al. [2021] alternates
between forward and backward passes (adopting a one-forward-one-backward
pattern) to keep the number of hidden states in the FIFO queue $Q$ constant.
Regardless of the number of micro-batches, 1F1B mitigates excessive memory
usage. Based on 1F1B, 1F1B-INarayanan et al. [2021a] schedule enlarges the
number of pipeline stages, and each device is assigned multiple stages. By
interleaving stages among devices, 1F1B-I reduces the bubble ratio at the cost
of adding more communication operators and slightly increasing memory
consumption. Zero-bubble-pipelineQi et al. [2024] divides the backward passes
into obtaining weight and input gradients separately. Zero-bubble-pipelineQi
et al. [2024] achieves higher pipeline efficiency by delaying weight gradient
computation and using dynamic programming to optimize the schedule. Zero-
bubble-pipeline approach nearly achieves zero-bubble pipeline efficiency but
brings more memory footprint caused by such delayment.
## 3 Methodology
In this section, we detail how our method works, beginning with a preliminary
overview to introduce characteristics of the 1F1B schedule and language
modeling. Then, we can prove why it’s feasible to schedule at the sequence
dimension for micro-batches in 1F1B. Following that, We will then explain how
Seq1F1B works in detail and how to meet the exact semantics of original
language modeling.
Building on this, we will discuss how different sequence-splitting strategies
impact the scheduling order in pipeline parallelism, and we will construct an
optimal solution based on the theoretical, computational load to address the
load-balancing issues associated with sequence-splitting strategies, thereby
enhancing the efficiency of our method.
### 3.1 Preliminary
1F1B includes three phases during one iteration: warm-up, steady, and cooling-
down phase. Assume a 1F1B scheduling scenario where we have $P$ workers, each
responsible for one pipeline stage, such that the size of pipeline parallelism
is $P$ and the number of micro-batches is $M$. Each worker denoted as $i$,
executes a forward pass during the warm-up phase. The number of warm-up micro-
batches for each worker is determined by the Eq. 1.
$\centering\text{w}_{i}=\begin{cases}P-i-1&\text{if }M>P\\\ M&\text{if }M\leq
P\end{cases}\@add@centering$ (1)
When $w_{i}$ equals $M$, 1F1B degrades to the behavior of GPipe. Otherwise,
during the warm-up phase, a worker responsible for an earlier stage performs
one more forward pass than a worker for a subsequent stage. Each forward pass
results in a hidden state that is enqueued in a FIFO queue $Q$ to be used
later for gradient computation during the backward pass. In the steady phase,
each worker performs one forward pass and enqueues the resulting hidden state
into $Q$. Following each forward pass, a hidden state is dequeued from $Q$ and
immediately to perform a backward pass for gradient computation, which is
where the "one-forward-one-backward" (1F1B) name comes from. It is noted that
the bubble ratio is minimal during the steady phase, and the number of one-
forward-one-backward passes in this phase is given by: $M-\text{w}_{i}$. Thus,
as $M$ increases, the proportion of the steady phase increases, which reduces
the bubble ratio. After the steady phase, the 1F1B scheduling enters the
cooling-down phase, which is symmetric to the warm-up phase and involves
performing the same number of backward passes as in the warm-up.
The primary optimization of 1F1B is to ensure that the memory consumption of
the hidden states is independent of $M$. The peak memory consumption for the
hidden states is determined by the number of items in the queue $Q$ at the end
of the warm-up phase, where each worker holds $w_{i}$ hidden states. Assuming
the total memory consumption of all hidden states is $A$, the peak memory
consumption of worker $i$ is $w_{i}\frac{A}{P}$. During the steady and
cooling-down phases, this consumption does not increase.
Figure 1: Execution timeline for the 1F1B and Seq1F1B schedules. Blank spaces
represent idle time, also known as bubbles. The top figure illustrates the
original 1F1B schedule, where each micro-batch is labeled with an ID. The
bottom figure illustrates our Seq1F1B schedule, where the input is split into
two sequences for better illustration. In Seq1F1B’s illustration, light-
colored areas represent the first sequence, while dark-colored areas represent
the second sequence. Notice that the forward pass for the dark-colored
sequence follows the light-colored sequence, whereas, for the backward pass,
the dark-colored sequence precedes the light-colored sequence. Figure 2:
Execution timeline for the 1F1B-I and Seq1F1B-I schedules. The top figure
illustrates the 1F1B-I schedule, where each micro-batch is labeled with an ID,
and different colors distinguish the forward/backward passes of different
stages. The lower part of the figure shows the Seq1F1B-I schedule, where the
input is split into two segments. In Seq1F1B-I, the light-colored areas
represent the first sequence and the dark-colored areas represent the second
sequence.
Language modeling is the most common unsupervised objective in training
language models. In Language modeling’s objective, each token is predicted
sequentially while conditioned on the preceding tokens, embodying the
principles of sequential generation, as formulated in Eq. 2.
$P(\mathbf{x})=\prod_{t=1}^{T}P(x_{t}\mid x_{1},x_{2},\ldots,x_{t-1})$ (2)
In the context of language modeling using Transformers, the unidirectional
attention mechanism ensures that each token in a sequence can only see its
predecessors, including itself.
Given a sequence of tokens $x_{0},x_{1},\ldots,x_{n}$, the output of the
attention mechanism for each token can be computed as follows. Each token
$t_{i}$ is associated with a query vector $q_{i}$, a key vector $k_{i}$, and a
value vector $v_{i}$, which will be used for attention computation. The output
for each token $t_{i}$, denoted as $O_{i}$, is computed by attending over all
previous tokens up to $t_{i}$, as formulated in Eq. 3.
$O_{i}=\text{softmax}\left(\frac{q_{i}\cdot[k_{0},\ldots,k_{i}]^{T}}{\sqrt{d_{k}}}\right)[v_{0},\ldots,v_{i}]$
(3)
Based on these characteristics, it becomes clear that to partition Transformer
computation across the sequence dimension, the attention mechanism must retain
the key and value vectors of all preceding tokens. The forward and backward
passes also need to maintain a specific order. The forward computation of each
token must follow the completion of its predecessor’s computation, while the
backward pass requires the subsequent token’s gradients to complete its
computation.
### 3.2 Seq1F1B
From the illustration 1, we observe that the original 1F1B schedule cannot
accommodate the splitting of micro-batches along the sequence dimension
because the last stage needs to immediately execute a backward pass after
forwarding a micro-batch. A straightforward adaptation method is to divide
each original 1F1B micro-batch into $k$ segments and then execute a $k$FkB
pipeline Li et al. [2021]. Although this schedule can reduce some bubbles in
1F1B, it does not save memory usage.
To achieve a more efficient sequence-level 1F1B pipeline schedule, we propose
Seq1F1B, which is a handcrafted 1F1B schedule for sequence-level input.
Specifically, Seq1F1B partitions the model into consecutive sets of layers and
assigns each worker with the corresponding set (a.k.a pipeline stages). Then,
Seq1F1B initializes the schedule part. Similar to 1F1B, the schedule is
divided into three phases: warm-up, steady, and cooling-down.
$\centering\text{w}_{i}=\begin{cases}P-i-2+k&\text{if }M>P\\\ M&\text{if
}M\leq P\end{cases}\@add@centering$ (4)
During the warm-up phase, the number of warm-up micro-batches of each worker
$i$ is calculated according to Eq. 4, in which $k$ represents the number of
splits in the sequence. This equation ensures that the last stage can perform
a backward pass on the last sequence segment of the first batch when entering
the steady phase, and the worker responsible for each stage performs one more
forward pass than the worker responsible for the subsequent stage. Here, we
construct a partially ordered queue $Q_{s}$, where each pop returns the tail
sequence from the earliest batch that has enqueued. This satisfies the first-
in-first-out principle in the batch dimension and the first-in-last-out
principle in the sequence dimension. In each iteration of the warm-up phase,
workers execute one forward pass and enqueue the corresponding hidden states
onto $Q_{s}$.
In the steady phase, after each worker completes a forward pass, it dequeues
from $Q_{s}$ and performs a backward pass on the dequeued hidden states,
following the standard 1F1B process, except that the units for forward and
backward passes become a sequence segment.
In the cooling-down phase, workers dequeue the remaining warm-up hidden states
from $Q_{s}$ and perform backward passes sequentially.
From the timeline shown in Figure 1, it is evident that the Seq1F1B schedule
offers shorter execution time and significantly fewer bubbles compared to the
original 1F1B schedule. Meanwhile, it can be clearly seen that each worker now
has less memory consumption since the micro-sequence is smaller than the
micro-batch. Another observation is that optimizations similar to the zero-
bubble pipeline can also be applied to Seq1F1B by delaying the gradient
computation associated with weights in the backward pass. We will discuss this
in the following section.
### 3.3 Seq1F1B-I
Figure 3: Execution timeline for the zero-bubble-pipeline’s ZBH1 and Seq1F1B
schedule intergrated with zero-bubble-pipeline’ ZBH1. Where each micro-batch
is labeled with an ID and different colors distinguish the
forward/backward/weight computation of different stages. OPT stands for
optimizer step.
1F1B-I Narayanan et al. [2021a] achieves better efficiency by modifying the
1F1B schedule to support interleaved stages among workers. In 1F1B-I, each
worker is assigned multiple stages. Suppose we have $P$ workers and $V$ stages
$\\{s_{1},s_{2},\ldots,s_{V}\\}$ in our pipeline, where $V$ is a multiple of
$P$. Each worker $i$ will handle $n$ stages
$\\{s_{i},s_{i+P},s_{i+2P},\ldots,s_{i+(n-1)P}\\}$, where $n=\frac{V}{P}$. The
number of warm-up micro-batches of each worker $i$ in 1F1B-I is given in Eq.
5.
$w_{i}=(P-i-1)\times 2+(n-1)\times P$ (5)
After completing $P$ iterations of forward/backward passes, each worker
switches its context to the next stage it is responsible for. From the Figure
2, the above part shows a 1F1B-I pipeline with $P$ as 4 and $V$ as 8, in which
each worker handles 2 stages. 1F1B-I’s schedule reduces the bubble ratio by
interleaving stages among workers. However, this interleaving slightly
increases memory consumption, as the number of warm-up micro-batches $w_{i}$
is greater compared to 1F1B.
Similar to 1F1B-I, Seq1F1B-I further modifies 1F1B-I to achieve sequence-level
scheduling, as shown in bottom part of Figure 2. From the Figure 2, Seq1F1B-I
effectively reduces pipeline bubbles and maintains less memory footprint of
hidden state compared with 1F1B-I. Seq1F1B-I defines the number of warm-up
micro-batches as in Eq. 6.
$w_{i}=(P-i-1)\times 2+(n-1)\times P+k-1$ (6)
in which $k$ represents the number of splits in the sequence. By using the
partially ordered queue, Seq1F1B-I maintains strict order of forward/backward
pass and ensures the consistent semantics of gradient updates. From the
perspective of pipeline bubbles, Seq1F1B-I outperforms both Seq1F1B and
1F1B-I. In terms of memory demands, Seq1F1B-I requires slightly more memory
than Seq1F1B but significantly less than 1F1B-I.
### 3.4 Integration with Zero-bubble-pipeline
From the illustration3, we can see Seq1F1B can integrate with ZB1P method and
further reduce bubbles while reducing memory demands by splitting sequence.
Such integration outperforms simple ZB1P in both memory demands and pipeline
bubbles since sequence-level pipelines naturally have fewer bubbles.
Furthermore, Seq1F1B can integrate with ZB2P and ZBV methods too.
Theoretically, introducing a zero-bubble-pipeline to Seq1F1B should be more
efficient. Even though, such a fine-grained handcraft schedule may have
performance degradation under some settings. We hope our work inspires future
work to solve this problem.
### 3.5 Workload Balance
In this section, we detail the strategy of sequence partition and workload
balance consideration. Previous works, such as Li et al. [2021], have
discussed strategies for sequence partitioning. To achieve efficient pipeline
scheduling, it is crucial that the processing times for each subsequence are
approximately equal to avoid pipeline bubbles. Based on this premise, we
design a computation-wise partition strategy by estimating the FLOPs of
sequences and constructing a theoretical solution aiming to make the FLOPs of
all subsequences as closely as possible.
For a input sequence $S=(x_{1},x_{2},\cdots,x_{n})$, we devide it into $k$
segments $S=[S_{1},\cdots,S_{k}]$. Each segment having a length of $n_{i}$,
where $\sum_{i=1}^{k}n_{i}=n$. We expect the computational amount of each
segment to be roughly the same, that is
$\text{FLOPs}(S_{1})=\text{FLOPs}(S_{2})=\cdots=\text{FLOPs}(S_{k})={\frac{\text{FLOPs}(S)}{k}}.$
(7)
Specifically, we use the method proposed in Hoffmann et al. [2022] to estimate
the FLOPs for each subsequence, as formulated in Eq. 8,
$\begin{aligned}
\text{FLOPs}(S_{i})=2~{}n_{i}~{}P+2~{}L~{}n_{i}\left(\sum_{j=0}^{i}n_{j}\right)d,\forall
i=1\dots k;\ \ \text{FLOPs}(S)=2~{}n~{}P+2~{}L~{}n^{2}d\end{aligned},$ (8)
in which, $L$ is a number of layers, $d$ is dimension of the model, and $P$ is
the total number of parameters in the model. We have $k$ variables in Eq. 8
and $k$ equations in Eq. 7. Therefore, we can set up the equation to get the
optimial segmentation.
## 4 Experiments
### 4.1 Experimental Settings
Table 1: Settings used in experiments for training LLMs. Model | Number of | Attention | Hidden | Sequence | PP | TP | Number of
---|---|---|---|---|---|---|---
Size | Layers | Heads | Size | Length | Size | Size | Micro-batches
2.7B | 32 | 32 | 2560 | 16k / 24k / 32k | 8 | 1 | 32 / 64
7B | 32 | 32 | 4096 | 32k / 64k / 128k | 4 | 8 | 16 / 32
13B | 40 | 40 | 5120 | 32k / 64k / 128k | 4 | 8 | 16 / 32
30B | 64 | 64 | 6144 | 32k / 48k / 64k | 8 | 8 | 32 / 64
Figure 4: Peak Memory consumption of tranining a series of models under
varying sequence lengths and fixed batch settings. “X” means experiments ran
out of memory. We take the maximum memory consumption between all workers for
better clarification.
In experiments, we measure our methods and 1F1B and 1F1B-I under variable
sequence lengths, different numbers of micro-batches, different numbers of
GPUs, different pipeline parallel sizes and tensor parallel sizes. Compared
methods are as follows:
* •
Seq1F1B: Seq1F1B with computation-wise sequence partition strategy.
* •
Seq1F1B-I: Seq1F1B with interleaved stages and computation-wise sequence
partition strategy.
* •
1F1B/1F1B-I: 1F1B and 1F1B with interleaved stages in Megatron implementation.
* •
Seq1F1B w/o cwp: Seq1F1B without computation-wise sequence partition strategy.
* •
Seq1F1B-I w/o cwp: Seq1F1B-I without computation-wise sequence partition
strategy.
All assessments are based on GPT model and model configuration are listed in
Table 1. All experiments focus on long-sequence training since a lot of work
has metioned the importance. For hyperparameter configurations, we set the
number of sequence splits to four and each worker managing two stages in
interleaved settings. Our implementation is based on the open-source Megatron-
LM project Narayanan et al. [2021a] and ensures reproducibility. We adopts
Megatron-V3Korthikanti et al. [2023]’s tensor parallelism in all experiments
since it is necessary for long sequence training.
Our experiments include three cluster settings: 1) 1 node with 8 NVIDIA A100
SXM 80G GPUs interconnected by NvLink. 2) 4 nodes interconnected by a RoCE
RDMA network and each node has 8 NVIDIA A100 SXM 80G GPUs interconnected by
NvLink. 3) 8 nodes interconnected by a RoCE RDMA network and each node has 8
NVIDIA A100 SXM 80G GPUs interconnected by NvLink. Each measurement in the
experiment is repeated 100 times, and the standard deviation is recorded.
### 4.2 Main Results
Table 2: 2.7B GPT training experiments with pipeline parallel size of 8 under
8xA100 setting.
Model Size | 2.7b
---|---
Sequence Length | 16384 | 24576 | 32768
Micro-batch | 16 | 32 | 16 | 32 | 16 | 32
Throughput | 1F1B | 32.0±0.0 | 37.1±0.0 | 27.0±0.0 | 31.4±0.0 | OOM | OOM
(Thousands | 1F1B-I | 36.4±0.0 | 39.7±0.0 | OOM | OOM | OOM | OOM
Tokens/s) | Seq1F1B | 37.3±0.0 | 38.9±0.3 | 32.6±0.0 | 34.2±0.0 | 28.8±0.0 | 30.1±0.2
Seq1F1B-I | 38.0±0.0 | 38.9±0.0 | 33.3±0.0 | 34.3±0.0 | 29.5±0.0 | 30.3±0.0
TFLOPS | 1F1B | 96.9±0.0 | 112.3±0.0 | 95.5±0.1 | 111.1±0.1 | OOM | OOM
per device | 1F1B-I | 110.3±0.1 | 120.2±0.1 | OOM | OOM | OOM | OOM
Seq1F1B | 113.1±0.0 | 117.8±0.8 | 115.2±0.1 | 120.9±0.1 | 116.5±0.1 | 122.0±1.0
| Seq1F1B-I | 115.2±0.0 | 118.0±0.0 | 118.0±0.1 | 121.3±0.1 | 119.4±0.0 | 122.7±0.0
Table 3: 7B GPT training experiments with pipeline parallel size of 4 and
tensor parallel size of 8 under 32xA100 setting.
Model Size | 7b
---|---
Sequence Length | 32768 | 65536 | 131072
Micro-batch | 8 | 16 | 8 | 16 | 8 | 16
Throughput | 1F1B | 48.2±0.1 | 55.3±0.2 | 37.3±0.0 | 43.1±0.0 | OOM | OOM
(Thousands | 1F1B-I | 53.0±0.3 | 56.3±0.4 | 41.7±0.1 | 44.7±0.0 | OOM | OOM
Tokens/s) | Seq1F1B | 53.5±0.3 | 55.8±0.1 | 43.3±0.0 | 45.0±0.1 | 30.4±0.0 | 31.6±0.0
Seq1F1B-I | 47.2±0.9 | 46.2±0.8 | 40.9±0.4 | 41.0±0.3 | 30.0±0.0 | 30.4±0.0
TFLOPS | 1F1B | 99.7±0.2 | 114.5±0.4 | 107.5±0.0 | 124.0±0.1 | OOM | OOM
per device | 1F1B-I | 109.5±0.7 | 116.5±0.8 | 120.0±0.2 | 128.7±0.1 | OOM | OOM
Seq1F1B | 110.6±0.5 | 115.3±0.2 | 124.6±0.1 | 129.7±0.5 | 136.7±0.1 | 142.1±0.0
| Seq1F1B-I | 97.7±1.8 | 95.5±1.6 | 117.8±1.3 | 118.0±0.8 | 135.1±0.2 | 136.6±0.2
Table 4: 13B GPT training experiments with pipeline parallel size of 4 and
tensor parallel size of 8 under 32xA100 setting.
Model Size | 13b
---|---
Sequence Length | 32768 | 49152 | 65536
Micro-batch | 8 | 16 | 8 | 16 | 8 | 16
Throughput | 1F1B | 28.9±0.1 | 33.4±0.1 | 25.3±0.1 | 29.3±0.1 | 22.6±0.1 | 30.0±0.0
(Thousands | 1F1B-I | 32.2±0.2 | 34.4±0.1 | 28.2±0.2 | 30.6±0.1 | OOM | OOM
Tokens/s) | Seq1F1B | 32.9±0.1 | 34.3±0.1 | 29.5±0.1 | 30.8±0.0 | 26.7±0.0 | 27.8±0.0
Seq1F1B-I | 29.7±0.4 | 29.8±0.3 | 28.0±0.2 | 28.3±0.1 | 26.4±0.1 | 26.8±0.1
TFLOPS | 1F1B | 106.7±0.2 | 123.0±0.5 | 109.5±0.5 | 126.2±0.6 | 111.9±0.5 | 135.1±0.2
per device | 1F1B-I | 118.6±0.6 | 126.9±0.4 | 121.9±0.7 | 132.2±0.4 | OOM | OOM
Seq1F1B | 121.2±0.2 | 126.6±0.3 | 127.3±0.4 | 133.1±0.2 | 132.5±0.0 | 137.9±0.0
| Seq1F1B-I | 109.7±1.4 | 110.0±1.1 | 121.0±1.1 | 122.1±0.4 | 130.6±0.3 | 132.8±0.3
Table 5: 30B GPT training experiments with pipeline parallel size of 8 and
tensor parallel size of 8 under 64xA100 setting.
Model Size | 30b
---|---
Sequence Length | 32768 | 49152 | 65536
Micro-batch | 8 | 16 | 8 | 16 | 8 | 16
Throughput | 1F1B | 26.4±0.1 | 31.2±0.2 | OOM | OOM | OOM | OOM
(Thousands | 1F1B-I | OOM | OOM | OOM | OOM | OOM | OOM
Tokens/s) | Seq1F1B | 31.3±0.1 | 33.1±0.2 | 28.2±0.1 | 29.6±0.1 | 25.5±0.0 | 26.8±0.0
Seq1F1B-I | 28.0±0.4 | 28.4±0.2 | 26.5±0.2 | 27.1±0.2 | 24.8±0.1 | 25.2±0.1
TFLOPS | 1F1B | 104.8±0.3 | 123.9±0.7 | OOM | OOM | OOM | OOM
per device | 1F1B-I | OOM | OOM | OOM | OOM | OOM | OOM
Seq1F1B | 124.5±0.2 | 131.5±0.6 | 129.4±0.3 | 135.6±0.3 | 132.6±0.0 | 139.2±0.0
| Seq1F1B-I | 111.1±1.6 | 113.0±1.0 | 121.5±1.1 | 124.2±0.8 | 128.6±0.3 | 130.9±0.6
In Figure 4, we compared the memory consumption of our method with that of
1F1B and 1F1B-I. As can be seen, our method consistently requires less memory
across all settings, and notably, it can support training a 30B model on a
64xA100 cluster, which is impossible for the traditional combination of
pipeline and tensor parallelism. Additionally, we recorded TFLOPS(teraFLOPS)
per GPU in our experiments to measure the hardware utilization of different
methods. From the Table 2, 3, 4 and 5, our method Seq1F1B outperforms 1F1B and
1F1B-I under almost all settings in both training throughput and teraFLOPS.
However, as observed in Table 3,4,5, the Seq1F1B-I may have a performance
degradation under multi-node settings. This could be due to the overly fine-
grained interleaving of stage partitioning and input sequence partitioning,
which also implies more communication calls in Tensor Parallelism (although
the total communication volume remains unchanged), potentially leading to a
decrease in performance. Another observation is that the efficiency of Seq1F1B
becomes more pronounced as the sequence length increases. This is because the
computation time for each micro-sequence extends with longer sequences,
thereby enhancing the benefits derived from sequence partitioning.
### 4.3 Ablation Results
We also conducted all experiments using Seq1F1B without computation-wise
partitioning (Seq1F1B w/o cwp) and Seq1F1B-I without computation-wise
partitioning (Seq1F1B-I w/o cwp) to evaluate the effectiveness of our
computation-wise partition strategy. Under identical settings, employing the
computation-wise partition strategy leads to performance enhancements ranging
from approximately 10-30% for Seq1F1B compared to simply splitting the
sequence.
Across all experimental scales, Seq1F1B consistently surpassed Seq1F1B w/o cwp
in performance. Table 6 highlights the ablation performance for a 2.7B model
with a sequence length of 32k, demonstrating a performance boost of
approximately 28% due to the computation-wise partitioning.
## 5 Conclusion
Table 6: The Ablation experiments based on 2.7B GPT of sequence partitioning
strategies, where “w/o cwp” indicates the absence of a computation-wise
partitioning strategy.
Method | TFLOPS/device | SpeedUp
---|---|---
Seq1F1B w/o cwp | 94.8±0.1 | -
Seq1F1B | 122.0±1.0 | 1.28x
Seq1F1B-I w/o cwp | 103.5±0.1 | -
Seq1F1B-I | 122.7±0.0 | 1.18x
In this paper, we present Seq1F1B, an efficient 1F1B pipeline parallel
scheduling method orienting to training Transformer-based LLMs on long
sequences by decomposing the batch-level schedulable units used by typical
1F1B methods into more fine-grained sequence-level units. To achieve a better
workload balance of the sequence-level pipeline, we design a computation-wise
sequence partition strategy to partition the sequences well. Meanwhile Seq1F1B
can integrate with other pipeline parallel methods such as 1F1B with
interleaved stage or zero-bubble-pipeline. Our evaluations demonstrate that
Seq1F1B outperforms the 1F1B and 1F1B-I scheduling strategies regarding memory
efficiency and training throughput under variable sequence lengths and model
sizes. Moreover, Seq1F1B can support the efficient training of a 30B GPT model
on sequences up to 64k in length using 64xA100 GPUs, without using
recomputation strategies, which is unachievable with existing pipeline
parallel methods. In the future, we will thoroughly combine our method with
other distributed methods to achieve better LLM training acceleration. In
addition, we will systematically release our code to support the community in
training LLMs to process longer sequences more efficiently.
## References
* Touvron et al. [2023] Hugo Touvron, Louis Martin, Kevin Stone, Peter Albert, Amjad Almahairi, Yasmine Babaei, Nikolay Bashlykov, Soumya Batra, Prajjwal Bhargava, Shruti Bhosale, et al. Llama 2: Open foundation and fine-tuned chat models. _arXiv preprint arXiv:2307.09288_ , 2023.
* Reid et al. [2024] Machel Reid, Nikolay Savinov, Denis Teplyashin, Dmitry Lepikhin, Timothy Lillicrap, Jean-baptiste Alayrac, Radu Soricut, Angeliki Lazaridou, Orhan Firat, Julian Schrittwieser, et al. Gemini 1.5: Unlocking multimodal understanding across millions of tokens of context. _arXiv preprint arXiv:2403.05530_ , 2024.
* Jiang et al. [2024] Albert Q Jiang, Alexandre Sablayrolles, Antoine Roux, Arthur Mensch, Blanche Savary, Chris Bamford, Devendra Singh Chaplot, Diego de las Casas, Emma Bou Hanna, Florian Bressand, et al. Mixtral of experts. _arXiv preprint arXiv:2401.04088_ , 2024.
* Anil et al. [2023] Rohan Anil, Andrew M Dai, Orhan Firat, Melvin Johnson, Dmitry Lepikhin, Alexandre Passos, Siamak Shakeri, Emanuel Taropa, Paige Bailey, Zhifeng Chen, et al. Palm 2 technical report. _arXiv preprint arXiv:2305.10403_ , 2023.
* Korthikanti et al. [2023] Vijay Anand Korthikanti, Jared Casper, Sangkug Lym, Lawrence McAfee, Michael Andersch, Mohammad Shoeybi, and Bryan Catanzaro. Reducing activation recomputation in large transformer models. _Proceedings of Machine Learning and Systems_ , 5, 2023.
* Rasley et al. [2020] Jeff Rasley, Samyam Rajbhandari, Olatunji Ruwase, and Yuxiong He. Deepspeed: System optimizations enable training deep learning models with over 100 billion parameters. In _Proceedings of KDD_ , pages 3505–3506, 2020.
* Rajbhandari et al. [2020] Samyam Rajbhandari, Jeff Rasley, Olatunji Ruwase, and Yuxiong He. Zero: Memory optimizations toward training trillion parameter models. In _Proceedings of SC20_ , 2020.
* Shoeybi et al. [2019] Mohammad Shoeybi, Mostofa Patwary, Raul Puri, Patrick LeGresley, Jared Casper, and Bryan Catanzaro. Megatron-lm: Training multi-billion parameter language models using model parallelism. _arXiv preprint arXiv:1909.08053_ , 2019.
* Narayanan et al. [2021a] Deepak Narayanan, Mohammad Shoeybi, Jared Casper, Patrick LeGresley, Mostofa Patwary, Vijay Korthikanti, Dmitri Vainbrand, Prethvi Kashinkunti, Julie Bernauer, Bryan Catanzaro, et al. Efficient large-scale language model training on gpu clusters using megatron-lm. In _Proceedings of the International Conference for High Performance Computing, Networking, Storage and Analysis_ , pages 1–15, 2021a.
* Huang et al. [2019] Yanping Huang, Youlong Cheng, Ankur Bapna, Orhan Firat, Dehao Chen, Mia Chen, HyoukJoong Lee, Jiquan Ngiam, Quoc V Le, Yonghui Wu, et al. Gpipe: Efficient training of giant neural networks using pipeline parallelism. _Advances in neural information processing systems_ , 32, 2019.
* Yang et al. [2021] Bowen Yang, Jian Zhang, Jonathan Li, Christopher Ré, Christopher Aberger, and Christopher De Sa. Pipemare: Asynchronous pipeline parallel dnn training. _Proceedings of Machine Learning and Systems_ , 3:269–296, 2021.
* Qi et al. [2024] Penghui Qi, Xinyi Wan, Guangxing Huang, and Min Lin. Zero bubble (almost) pipeline parallelism. In _The Twelfth International Conference on Learning Representations_ , 2024. URL https://openreview.net/forum?id=tuzTN0eIO5.
* Li et al. [2021] Zhuohan Li, Siyuan Zhuang, Shiyuan Guo, Danyang Zhuo, Hao Zhang, Dawn Song, and Ion Stoica. Terapipe: Token-level pipeline parallelism for training large-scale language models. In _International Conference on Machine Learning_ , pages 6543–6552. PMLR, 2021.
* Harlap et al. [2018] Aaron Harlap, Deepak Narayanan, Amar Phanishayee, Vivek Seshadri, Nikhil Devanur, Greg Ganger, and Phil Gibbons. Pipedream: Fast and efficient pipeline parallel dnn training. _arXiv preprint arXiv:1806.03377_ , 2018.
* Fan et al. [2021] Shiqing Fan, Yi Rong, Chen Meng, Zongyan Cao, Siyu Wang, Zhen Zheng, Chuan Wu, Guoping Long, Jun Yang, Lixue Xia, et al. Dapple: A pipelined data parallel approach for training large models. In _Proceedings of the 26th ACM SIGPLAN Symposium on Principles and Practice of Parallel Programming_ , pages 431–445, 2021.
* [16] Jacob Buckman and Carles Gelada. Compute-optimal Context Size.
* Vaswani et al. [2017] Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Łukasz Kaiser, and Illia Polosukhin. Attention is all you need. In _Proceedings of NeurIPS_ , 2017.
* Dao et al. [2022] Tri Dao, Dan Fu, Stefano Ermon, Atri Rudra, and Christopher Ré. FlashAttention: Fast and memory-efficient exact attention with io-awareness. In _Proceedings of NeurIPS_ , pages 16344–16359, 2022.
* Ding et al. [2023] Jiayu Ding, Shuming Ma, Li Dong, Xingxing Zhang, Shaohan Huang, Wenhui Wang, and Furu Wei. LongNet: Scaling transformers to 1,000,000,000 tokens. _arXiv preprint arXiv:2307.02486_ , 2023.
* Jacobs et al. [2023] Sam Ade Jacobs, Masahiro Tanaka, Chengming Zhang, Minjia Zhang, Leon Song, Samyam Rajbhandari, and Yuxiong He. Deepspeed ulysses: System optimizations for enabling training of extreme long sequence transformer models. _arXiv preprint arXiv:2309.14509_ , 2023.
* Goyal et al. [2017] Priya Goyal, Piotr Dollár, Ross Girshick, Pieter Noordhuis, Lukasz Wesolowski, Aapo Kyrola, Andrew Tulloch, Yangqing Jia, and Kaiming He. Accurate, large minibatch sgd: Training imagenet in 1 hour. _arXiv preprint arXiv:1706.02677_ , 2017.
* Zinkevich et al. [2010] Martin Zinkevich, Markus Weimer, Lihong Li, and Alex Smola. Parallelized stochastic gradient descent. _Advances in neural information processing systems_ , 23, 2010.
* Li et al. [2020] Shen Li, Yanli Zhao, Rohan Varma, Omkar Salpekar, Pieter Noordhuis, Teng Li, Adam Paszke, Jeff Smith, Brian Vaughan, Pritam Damania, et al. Pytorch distributed: experiences on accelerating data parallel training. _Proceedings of the VLDB Endowment_ , 13(12):3005–3018, 2020.
* Xing et al. [2015] Eric P Xing, Qirong Ho, Wei Dai, Jin Kyu Kim, Jinliang Wei, Seunghak Lee, Xun Zheng, Pengtao Xie, Abhimanu Kumar, and Yaoliang Yu. Petuum: A new platform for distributed machine learning on big data. _IEEE Transactions on Big Data_ , 1(02):49–67, 2015.
* Ren et al. [2021] Jie Ren, Samyam Rajbhandari, Reza Yazdani Aminabadi, Olatunji Ruwase, Shuangyan Yang, Minjia Zhang, Dong Li, and Yuxiong He. ZeRO-Offload: Democratizing billion-scale model training. In _Proceedings of ATC_ , pages 551–564, 2021.
* Narayanan et al. [2021b] Deepak Narayanan, Amar Phanishayee, Kaiyu Shi, Xie Chen, and Matei Zaharia. Memory-efficient pipeline-parallel dnn training. In _International Conference on Machine Learning_ , pages 7937–7947. PMLR, 2021b.
* Li and Hoefler [2021] Shigang Li and Torsten Hoefler. Chimera: efficiently training large-scale neural networks with bidirectional pipelines. In _Proceedings of the International Conference for High Performance Computing, Networking, Storage and Analysis_ , pages 1–14, 2021.
* Hoffmann et al. [2022] Jordan Hoffmann, Sebastian Borgeaud, Arthur Mensch, Elena Buchatskaya, Trevor Cai, Eliza Rutherford, Diego de Las Casas, Lisa Anne Hendricks, Johannes Welbl, Aidan Clark, et al. Training compute-optimal large language models. _arXiv preprint arXiv:2203.15556_ , 2022.
|
§ ABSTRACT
Magnetic materials with noncollinear spin textures are promising for spintronic applications. To realize practical devices, control over the length and energy scales of such spin textures is imperative. The chiral helimagnets and exhibit analogous magnetic phase diagrams with different real-space periodicities and field dependence, positioning them as model systems for studying the relative strengths of the microscopic mechanisms giving rise to exotic spin textures. Here, we carry out a comparative study of the electronic structures of and using angle-resolved photoemission spectroscopy and density functional theory. We show that bands in are more dispersive than their counterparts in and connect this result to bonding and orbital overlap in these materials. We also unambiguously distinguish exchange splitting from surface termination effects by studying the dependence of their photoemission spectra on polarization, temperature, and beam size. We find strong evidence that hybridization between intercalant and host lattice electronic states mediates the magnetic exchange interactions in these materials, suggesting that band engineering is a route toward tuning their spin textures. Overall, these results underscore how the modular nature of intercalated transition metal dichalcogenides translates variation in composition and electronic structure to complex magnetism.
§ INTRODUCTION
Next-generation spintronic devices utilize the spin degree of freedom to store information. fert2013,parkin2015 Magnetic materials in which spins order in topologically protected quasiparticles, such as skyrmions or magnetic solitons, are promising platforms for realizing such devices.togawa2016,tokura2021 These chiral spin textures can be manipulated with currents and magnetic fields, which is appealing for various applications in memory, logic, and unconventional computing.tey2022 For practical spintronic devices, optimizing the energy and length scales of the spin textures is important: stability at operationally accessible temperatures and fields as well as high density in thin-film architectures are broadly desirable. Strategies to control the microscopic mechanisms that give rise to complex magnetism are thus needed. In terms of materials design, this can be broadly achieved by tailoring the interactions between spin centers as directed by their spatial arrangements and coordination environments.
The chiral helimagnets and are especially well-suited for device schemes implementing noncollinear spin textures because of their anisotropic layered structures, which are compatible with thin-film architectures.togawa2015,yamasaki2017,wang2017,zhang2022,osorio2022 In these materials, the $S=3/2$ Cr$^{3+}$ centers occupy pseudo-octahedral sites between layers of $2H$-NbS2 or $2H$-TaS2,parkin1980a,parkin1980 forming a $\sqrt{3} \times \sqrt{3}$ superlattice.rouxel1971 They exhibit easy-plane ferromagnetic (FM) behavior with chiral magnetic ordering out-of-plane: the Cr superlattice breaks the inversion symmetry of the transition metal dichalcogenide (TMD) host lattice along the crystallographic $c$-axis, giving rise to a Dzyaloshinskii–Moriya (DM) interaction, also known as antisymmetric exchange.dzyaloshinsky1958,moriya1960 The DM interaction favors out-of-plane spin canting, which competes with FM exchange to produce one-dimensional helical spin textures that propagate along [001]. Importantly, the application of an in-plane magnetic field creates a chiral soliton lattice (CSL) phase with tunable periodicities up to a critical field, $H_\mathrm{c}$, above which a forced ferromagnetic (FFM) state is observed.miyadai_magnetic_1983,togawa_chiral_2012-1,ghimire2013,zhang2021,obeysekera2021,du2021 Both and have Curie temperatures, $T_{\mathrm{C}}$, well above 100 K, and nanoscale soliton wavelengths tunable with fields of 1.5 T or less, thus providing a richly accessible phase space for manipulating chiral spin textures.
Although the magnetic phase diagrams for and are qualitatively analogous, the periodicities and stabilities of their magnetic solitons differ somewhat. Literature reports have established that consistently exhibits a higher $T_{\mathrm{C}}$,togawa_chiral_2012-1,ghimire2013,du2021,kousaka2022,meng2023, higher $H_\mathrm{c}$,ghimire2013,zhang2021,obeysekera2021,meng2023 and a shorter soliton wavelength than the Nb analogue.togawa_chiral_2012-1,kousaka2016,zhang2021,du2021 These observations imply that changing the host lattice from NbS2 to TaS2 alters the relative strengths of magnetic coupling among Cr centers, manifesting in quantitative changes to their magnetic phase diagrams. However, the origin of magnetic exchange interactions in these materials is still a matter of debate,sirica2016,sirica2021,qin2022,hicken2022 and comparative studies have been scant.hicken2022 A detailed investigation of the electronic structures of both and is thus motivated by the fact that these materials are natural platforms for studying how the length and energy scales of chiral spin textures can be tuned through materials chemistry.
Herein, we present a comprehensive investigation of the electronic structures of and using angle-resolved photoemission spectroscopy (ARPES) and density functional theory (DFT) calculations. We show that the Ta analogue has more dispersive bands, consistent with greater orbital overlap in the case of Ta, and discuss implications for their magnetic properties and the rational design of chiral helimagnets. Using polarization-dependent ARPES and orbital-projected DFT calculations, we assign the parity and orbital character of bands, finding excellent agreement between theory and experiment with the exception of additional band splitting near the Fermi level observed in ARPES but not predicted in DFT. ARPES data collected with smaller beam sizes reveal spatial variation in this band splitting, consistent with different surface terminations on the as-cleaved samples. Thus, we distinguish exchange splitting from surface vs. bulk splitting for the first time in these materials. Our findings establish a high degree of similarity in the electronic structures of and and highlight the relevance of their polar layered nature in interpreting surface-sensitive spectroscopic investigations. More generally, these results suggest that careful band structure engineering and Fermi level tuning may prove to be fruitful avenues toward optimizing the spin textures in intercalated TMDs.
First, we briefly outline the electronic and magnetic properties of ($M$ = Nb or Ta) as established in the existing literature. According to a simple electron counting scheme, these compounds can be considered as alternating layers of [Cr_1/3]+ and [$M$S2]-. The intercalant layers consist of Cr^3+ centers occupying 1/3 of the trigonally distorted pseudo-octahedral interstitial sites between layers of $2H$-$M$S2 ($M$ = Nb or Ta). These intercalant layers donate one electron per formula unit to the $M$S2 host lattice layers (Figure <ref>a). The qualitative local $d$-orbital splitting diagrams for the Cr^3+ ($D_{3d}$) and $M$^3+ ($D_{3h}$) centers are shown in Figure <ref>b–c.xie_structure_2022 The electronic structure of the periodic solids are more complex, and in reality, the half-filled TMD bands have both $d_{z^2}$ and $d_{xy}$/$d_{x^2-y^2}$ character.mattheiss1973,yee1991,whangbo1992 Nevertheless, this simplified picture captures (1) charge transfer from the intercalant species to the highest-lying $M$ $d$ bands of the host lattice, and (2) the inherently polar nature of these layered intercalation compounds.
The qualitative magnetic properties of and below $T_{\mathrm{C}}$ are summarized in Figure <ref>d–e. Within each [Cr_1/3]+ layer, the Cr spins exhibit FM coupling through an in-plane exchange constant, $J_{\parallel}$. Between adjacent [Cr_1/3]+ layers, FM coupling through an out-of-plane exchange constant, $J_{\perp}$, competes energetically with spin canting through a DM interaction term, $D$. At zero field, this results in a continuous helical arrangement of spins, or a CHM ground state, with the magnetic soliton wavelength determined by the ratio of $J_{\perp}$ and $D$.chapman2014spin,aczel2018 With increasing $H \perp c$, FM regions aligned with the field grow, effectively unwinding the CHM state to create the CSL phase, in which the distance separating adjacent solitons is a function of the magnitude of $H$. Finally, with fields larger than $H_\mathrm{c}$, an FFM state with saturated magnetization is obtained.miyadai_magnetic_1983,togawa_chiral_2012-1,ghimire2013,han2017,zhang2021,obeysekera2021,meng2023
(a) Crystal structure of , showing formal charges for the $M$S2 and Cr layers from a simple electron-counting picture. (b) and (c) Qualitative $d$-orbital splitting diagrams for isolated Cr and $M$ centers from the local ligand field in . (d) Schematic illustration of the magnetic structure of in the chiral helimagnetic (CHM) state. (e) Schematic representations of spin textures evolving CHM to chiral soliton lattice (CSL) to forced ferromagnetic (FFM) states with increasing applied magnetic field $H \perp c$. (f) and (g) $M(H)$ data for and , respectively, showing transitions between CHM, CSL, and FFM states.
In this study, we investigate the electronic structure of and in a comparative context to tease out differences between the two compounds and connect these to their magnetic phase diagrams. To do so, we grew and characterized single crystals, verified their chiral spin textures with magnetometry, carried out a comprehensive suite of ARPES measurements, and conducted DFT band structure calculations, as detailed below.
§ RESULTS
§.§ Synthesis, Structure, and Magnetism
Single crystals of and were grown via chemical vapor transport using iodine as a transport agent. X-ray diffraction confirmed that both materials crystallize in the noncentrosymmetric space group $P6_322$, with the Cr centers forming a $\sqrt{3} \times \sqrt{3}$ superlattice (Figure <ref> and Tables <ref>–<ref>). exhibits a slightly larger in-plane lattice parameter and smaller out-of-plane lattice parameter ($a$ = 5.7400(7) Å and $c$ = 12.1082(14) Å) compared to ($a$ = 5.7155(5) Å and $c$ = 12.1751(12) Å). Raman spectroscopy revealed sharp vibrational modes associated with the $\sqrt{3} \times \sqrt{3}$ superlatticesfan2021 (Figure <ref>), and energy dispersive X-ray spectroscopy indicated Cr:Nb and Cr:Ta ratios of 0.33(1):1 (Figures <ref> and <ref>).
The metamagnetic transitions across these states with applied magnetic field are observed in the $M(H)$ data for single crystals of and shown in Figure <ref>f–g, confirming the characteristic spin textures in our samples.miyadai_magnetic_1983,zhang2021,obeysekera2021 Both compounds exhibit similar saturation moments (2.7 $\mu_{\mathrm{B}}$/Cr for and 2.8 $\mu_{\mathrm{B}}$/Cr for ), close to the expected spin-only value of 3 $\mu_{\mathrm{B}}$/Cr. The analogous transitions are observed at fields more than an order of magnitude larger for than , with $H_\mathrm{c}$ values of about 0.45 mT for and 16 mT for . This is consistent with shorter soliton wavelengths in the Ta analogue.zhang2021,obeysekera2021,du2021 The $M(T)$ data show pronounced peaks at 110 and 133 K for and , respectively, corresponding to the onset of chiral helimagnetism below these temperatures (Figures <ref>–<ref>).
§.§ Superlattice Effects on Electronic Structure
After obtaining structural and magnetic evidence of highly ordered $\sqrt{3} \times \sqrt{3}$ Cr superlattices in and , we sought to investigate their influence on the electronic structure of these materials. Figure <ref>a illustrates the real-space $1 \times 1$ primitive unit cell for the host lattice TMD and the $\sqrt{3} \times \sqrt{3}$ superlattice unit cell for the intercalated compounds along [001]. The $\sqrt{3} \times \sqrt{3}$ unit cell is rotated by 30° compared to the $1 \times 1$ unit cell. In reciprocal space, the $\sqrt{3} \times \sqrt{3}$ superlattice defines a smaller Brillouin zone that is likewise rotated by 30° relative to the primitive Brillouin zone (Figure <ref>b). To probe the electronic effects of Cr intercalation, we first examined the symmetries of the experimental Fermi surfaces and band dispersions of and using ARPES.
(a) Real-space crystal structure of ($M$ = Nb or Ta) viewed along the crystallographic $c$-axis, with overlaid unit cells for the $1 \times 1$ primitive $M$S2 lattice (dashed green) and the $\sqrt{3} \times \sqrt{3}$ Cr superlattice (solid gray). (b) Surface Brillouin zones for the $1 \times 1$ primitive lattice and $\sqrt{3} \times \sqrt{3}$ superlattice. (c) and (d) ARPES Fermi surfaces of and with the dashed and solid overlaid lines corresponding to the primitive and $\sqrt{3} \times \sqrt{3}$ superlattice Brillouin zones, respectively. (e) and (f) ARPES band dispersions for and along the $\Gamma$–$\mathrm{K}_0$ direction, showing folding of features from $\Gamma$ to $\mathrm{K}_0$ and vice versa (18 K, $h\nu$ = 79 eV).
As shown in Figure <ref>c–d, the Fermi surfaces of both and below $T_{\mathrm{C}}$ ($h\nu$ = 79 eV) display multiple nested barrels around $\Gamma$ and $\mathrm{K}$ of the primitive Brillouin zone, which is indicated by the dashed green hexagons. Notably, six-fold symmetry is clearly observed around the primitive $\mathrm{K}$ (denoted as $\mathrm{K_0}$), in contrast with three-fold symmetry around $\mathrm{K}$ of the host lattice materials $2H$-NbS2 and $2H$-TaS2.elyoubi2021,zhao2017 Additionally, in the intercalated materials, three-fold symmetry is introduced at $\mathrm{K}$ of the $\sqrt{3} \times \sqrt{3}$ superlattice Brillouin zone (denoted as $\mathrm{K_{SL}}$), which is indicated by the solid gray hexagons in Figure <ref>c–d. Hence, the Fermi surfaces of and display the expected symmetries associated with reconstruction and band folding from the $\sqrt{3} \times \sqrt{3}$ Cr superlattice.
The ARPES dispersions show clear evidence of band folding as well (Figure <ref>e–f). Cuts along the $\Gamma$–$\mathrm{K_0}$ direction show the same features at both $\Gamma$ and $\mathrm{K_0}$: both materials display several nested hole pockets and parabolic bands below $E_{\mathrm{F}}$. In contrast, for the host TMDs $2H$-NbS2 and $2H$-TaS2, the bands crossing $E_{\mathrm{F}}$ have different dispersions and energies at $\Gamma$ and $\mathrm{K}$. In the Cr-intercalated materials, the $\sqrt{3} \times \sqrt{3}$ superlattice folds the primitive lattice $\Gamma$ to $\mathrm{K}$ and vice versa, as they both become $\Gamma$ of the superlattice Brillouin zone (denoted as $\mathrm{\Gamma_{SL}}$ in Figure <ref>b). Thus, the presence of the same features at both $\Gamma$ and $\mathrm{K_0}$ in the ARPES of the Cr-intercalated materials is consistent with $\sqrt{3} \times \sqrt{3}$ superlattice band folding.
The band folding in the ARPES data reveals that the $\sqrt{3} \times \sqrt{3}$ Cr superlattice potential is strong in both and . Broadly, this electronic reconstruction is in line with previous literature reports on ,sirica2016,qin2022 as well as other intercalated TMDs with $\sqrt{3} \times \sqrt{3}$ transition metal superlattices.tanaka2022,yang2022,popcevic2022,edwards2023 The features observed in both materials are qualitatively similar; however, at a glance, the hole pockets in appear to be larger than those found in . To contextualize differences in the experimental electronic structures of and , we turned to DFT calculations and quantitative analysis of their band dispersions.
§.§ Relative Band Dispersions
To understand the relative differences between the band structures of and , we started by comparing the host lattice materials, $2H$-NbS2 and $2H$-TaS2. DFT band structure calculations of $2H$-NbS2 and $2H$-TaS2 show that the bands crossing $E_{\mathrm{F}}$ in $2H$-TaS2 are more dispersive compared to the analogous bands in $2H$-NbS2. This can be clearly visualized by comparing the relative spread of the maxima and minima of these respective bands, as illustrated in Figure <ref>a–b: the more dispersive bands in $2H$-TaS2 have a higher-energy maximum and lower-energy minimum compared to $2H$-NbS2. These bands have predominantly Nb or Ta $d_{z^2}$ and $d_{xy}$/$d_{x^2-y^2}$ character, with additional contribution from S $p$ states.zhao2017,elyoubi2021
(a) and (b) DFT band structures of $2H$-NbS2 and $2H$-TaS2, with maxima and minima of the bands crossing $E_{\mathrm{F}}$ indicated by solid navy and mint lines, respectively. (c) and (d) Spin-polarized band structures for and in the FM state, with spin up and spin down bands indicated in red and blue, respectively. (e) and (f) ARPES dispersions of and (18 K, $h\nu$ = 46 eV), with blue circles indicating the peak center positions of the most intense feature from MDC analysis. The Fermi velocities, $v_{\mathrm{F}}$, are obtained from linear fits to the centers between 0 and $-50$ meV.
Next, we calculated the band structures of the Cr-intercalated materials and compared the results to our ARPES data. Due to the surface-sensitive nature of ARPES, we do not expect to experimentally resolve signatures of the CHM state, i.e. out-of-plane spin textures with length scales on the order of tens of nm. Hence, we use spin-polarized band structure calculations of and in their FM states, with the magnetization vector along [100], as proxies for the electronic structure near the surface (Figure <ref>c–d and Figures <ref>–<ref>). Three distinct changes are evident in the DFT band structures of and compared to the host lattices: (1) folding due to the $\sqrt{3} \times \sqrt{3}$ superlattice potential, (2) raising of $E_{\mathrm{F}}$ due to electron transfer from Cr to the host lattice, and (3) introduction of new bands crossing $E_{\mathrm{F}}$ due to Cr–Nb or Cr–Ta hybridization and FM exchange splitting.
Although the Cr-intercalated materials have more complex electronic structures than the host lattices, DFT calculations show that the Ta analogue again has more dispersive bands than the Nb analogue. The amount of charge transfer from Cr to the host lattice appears to be very similar for both materials, as shown by the calculated and experimental magnetic moments (Tables <ref> and <ref>). Thus, the shift of $E_{\mathrm{F}}$ upon intercalation is almost identical. This results in larger hole pockets around $\Gamma$ and $\mathrm{K_0}$ in than and an extra spin-up band crossing $E_{\mathrm{F}}$ at $\Gamma$ and $\mathrm{K}_0$ in . Notably, the ARPES dispersions of and at 18 K ($h\nu$ = 46 eV) show clearly that the most intense hole pocket around $\Gamma$ in the $\Gamma$–$\mathrm{M_{SL}}$ direction is considerably larger at $E = E_{\mathrm{F}}$ in compared to (Figure <ref>e–f), with Fermi wavevectors, $k_{\mathrm{F}}$, of 0.10 Å for and 0.19 Å for . By fitting the momentum distribution curves (MDCs) to Lorentzians between 0 and $-50$ meV, we extracted Fermi velocities, $v_{\mathrm{F}}$, of $1.47(5) \times 10^5$ m/s for and $1.8(2) \times 10^5$ m/s for —thus experimentally quantifying the relative band dispersions between the two systems. The larger experimental $v_{\mathrm{F}}$ for the Ta analogue mirrors the relative trends from the DFT band structures.
§.§ Orbital Character Assignments
To gain insight into the orbital character of the bands, we studied their polarization dependence in ARPES. For the photoemission process, the matrix element term can be described by
\[ | M^{\mathbf{k}}_{f,i}|^2 \propto | \langle \phi^{\mathbf{k}}_f | \hat{\epsilon} \cdot \mathbf{r} | \phi^{\mathbf{k}}_i \rangle |^2 \]
where $\hat{\epsilon}$ is the unit vector along the polarization direction of the light.damascelli2003 The final state wavefunction of the photoelectron, $\phi^{\mathbf{k}}_f$, can be described by a plane wave state, $e^{i\mathbf{kr}}$, with even parity with respect to the mirror plane defined by the analyzer slit and the normal to the sample surface (Figure <ref>a). To obtain a nonvanishing matrix element, $\hat{\epsilon}$ must be even (odd) for an even (odd) initial state wavefunction, $\phi^{\mathbf{k}}_i$. Based on the symmetry operations of space group $P6_322$ (point group $D_6$), and taking $z$ to be parallel to the crystallographic $c$-axis, we expect the even $a_1$ ($d_{z^2}$) states of both Cr and Nb or Ta to be visible with linear horizontal (LH) polarized light (even $\hat{\epsilon}$) but not linear vertical (LV) polarized light (odd $\hat{\epsilon}$).
The $e_1$ ($d_{xz}$, $d_{yz}$) and $e_2$ ($d_{xy}$, $d_{x^2-y^2}$) sets are not symmetric overall with respect to any scattering plane containing the sample surface normal. We illustrate this by considering a horizontal analyzer slit aligned to the $\Gamma$–$\mathrm{K}_0$ direction, and defining $x$ as parallel to the crystallographic $a$-axis of the superlattice unit cell. The resulting scattering plane is the $xz$ plane (Figure <ref>a) and contains $M$–S and Cr–S bonds (Figure <ref>b). As shown in Figure <ref>c, the $d_{xz}$ and $d_{x^2-y^2}$ orbitals are even with respect to the $xz$ plane, but the $d_{yz}$ and $d_{xy}$ orbitals (the other components of the $e_1$ and $e_2$ sets) are odd. Thus, the $e_1$ and $e_2$ sets are not symmetric collectively, and may be visible with both LH and LV polarization.
(a) Geometry of ARPES data collection for a horizontal analyzer slit aligned with the $xz$ scattering plane of the sample. (b) Surface Brillouin zones for the primitive lattice (dotted green) and $\sqrt{3} \times \sqrt{3}$ superlattice (solid gray), along with real space projections of local coordination environments for $M$ = Nb or Ta and Cr. (c) Symmetries of $d$ orbitals for $M$ = Nb or Ta and Cr for the scattering plane defined as the $xz$ plane aligned with the $\Gamma$–$\mathrm{K}_0$ direction.
ARPES data of measured with LV polarization (Figure <ref>a) show stronger intensity from the innermost parabolic bands centered at $\Gamma$ and especially $\mathrm{K_0}$. In contrast, with LH polarization (Figure <ref>b), the sharp outermost dispersive bands around $\Gamma$ and $\mathrm{K_0}$ are more prominent, as well as two sets of more diffuse electron pockets with minima at $\mathrm{M_{SL}}$ and $\mathrm{K_{SL}}$. To compare with the polarization-dependent ARPES data, we plotted the orbital-projected DFT band structure as a function of even ($d_{z^2}$) vs. not symmetric ($d_{xy}$/$d_{x^2-y^2}$ and $d_{xz}$/$d_{yz}$) states in Figure <ref>c. At $\Gamma$/$\mathrm{K_0}$, the innermost parabolic bands have predominantly $d_{xy}$/$d_{x^2-y^2}$ and $d_{xz}$/$d_{yz}$ character, whereas the outermost dispersive bands and electron pockets have more $d_{z^2}$ character. These parities are qualitatively consistent with the experimentally observed polarization dependence. The DFT band structure as a function of Cr vs. Nb character (<ref>d) indicates that all of the parabolic bands at $\Gamma$/$\mathrm{K_0}$ are predominantly Nb-derived, while the electron pockets with minima at $\mathrm{M_{SL}}$ and $\mathrm{K_{SL}}$ are composed of mixed Cr and Nb states. The polarization dependence of the host lattice bands more visible in LV polarization is consistent with $d_{xy}$/$d_{x^2-y^2}$ states folded to $\Gamma$ from $\mathrm{K_0}$ by the $\sqrt{3} \times \sqrt{3}$ superlattice potential.elyoubi2021
(a) and (b) ARPES band dispersions for with linear vertical (LV) and linear horizontal (LH) polarized photons, respectively (18 K, $h\nu$ = 79 eV). (c) and (d) DFT orbital-projected band structures of in the FM state, showing in-plane vs. out-of-plane character, and Cr vs. Nb character, respectively. (e–h) The same as (a–d) for , and Cr vs. Ta character in (h).
The polarization-dependent ARPES data for are similar to those for . The innermost bands at $\Gamma$ and $\mathrm{K_0}$ are more prominent in LV polarization (Figure <ref>e), whereas the outer bands around $\Gamma$ and $\mathrm{K_0}$ and more diffuse electron pockets with minima at $\mathrm{M_{SL}}$ and $\mathrm{K_{SL}}$ are more intense in LH polarization (Figure <ref>f). The orbital-projected DFT band structure reveals analogous parities to the Nb analogue (Figure <ref>g) and similar atomic parentage (Figure <ref>h), albeit with more Cr character in the vicinity of $E_\mathrm{F}$. Less Cr–Ta hybridization may be an effect of the slightly longer $c$ lattice parameter in .
For a more quantitative enumeration of the bands near $E_\mathrm{F}$ observed in ARPES, we fitted the momentum distribution curves (MDCs) of the data collected with LH polarization using multiple Lorentzian peaks along the cuts shown in Figure <ref>a–c. We refer to the dispersive features around $\Gamma$ near $E_{\mathrm{F}}$ as the $\alpha$, $\beta$, and $\gamma$ bands, respectively, and two parabolic bands below $E_{\mathrm{F}}$ as $\delta_1$ and $\delta_2$. Comparison of the full width at half maximum (FWHM) values from fits to the $\Gamma$–$\mathrm{M_{SL}}$ MDCs within 200 meV of $E_{\mathrm{F}}$ indicates that the two middle bands have similar FWHMs, while the outermost band (corresponding to the shallow electron pocket) is considerably broader (Figure <ref>d). We therefore assign the middle two features as split $\beta_1$ and $\beta_2$ bands in the vicinity of $E_{\mathrm{F}}$. At higher binding energies, the MDCs can be fitted well with two copies of the electron pocket bands split by about 250 meV, which we refer to as $\gamma_1$ and $\gamma_2$. Comparing the peak center positions from MDC fitting (Figure <ref>e) with the DFT band structure (Figure <ref>f) shows good qualitative agreement, other than the apparent doubling of the $\beta$ and $\gamma$ bands observed in the ARPES data.
(a) ARPES Fermi surface of . Dashed and solid overlaid lines indicate the primitive and $\sqrt{3} \times \sqrt{3}$ superlattice Brillouin zones, respectively. (b) and (c) MDCs along the $\Gamma$–$\mathrm{M_{SL}}$ and $\mathrm{M_{SL}}$–$\mathrm{K_{SL}}$ directions (cuts indicated by bold blue and teal lines in (a)). Gray lines are multi-Lorentzian fits, and colored circles indicate peak center positions, with band assignments labeled. (d) MDCs along the $\Gamma$–$\mathrm{M_{SL}}$ direction (18 K, $h\nu$ = 46 eV). Shaded regions correspond to the full width at half maximum values determined by multi-Lorentzian fits. (e) Sketch of the proposed band structure of derived from the MDC fits (peak center positions shown by gray circles). (f) DFT band structure of , with corresponding band assignments from MDC analysis indicated by colored overlays.
§.§ Temperature Evolution of Band Structure
Due to the aforementioned band splitting, we sought to probe the effect of magnetic ordering on the electronic structures of and by comparing ARPES data collected below and above $T_{\mathrm{C}}$. The dispersion of at 18 K (Figure <ref>a) vs. 145 K (Figure <ref>b) shows that the hole pockets around $\Gamma$ appear smaller at 18 K compared to 145 K. Nevertheless, multi-Lorentzian fits to the MDCs at $E - E_{\mathrm{F}} = -15$ meV show that the outer dispersive bands around $\Gamma$ crossing $E_{\mathrm{F}}$ display the same splitting at 18 and 145 K, as indicated by the labeled $\beta_1$, $\beta_2$, and $\gamma_2$ peaks in Figure <ref>c–d. A similar change in hole pocket sizes is evident in the dispersions of at 18 K (Figure <ref>e) and 170 K (Figure <ref>f). As with , fitting the MDCs at $E - E_{\mathrm{F}} = -15$ meV indicates that the splitting of the outer bands is observed at both 18 and 170 K (Figure <ref>g–h). For all the MDCs, we modeled the inner features around $\Gamma$ with Lorentzian peaks as well, but we note that ascertaining the effects of temperature on these bands is more challenging due to the lower intensities and a non-negligible background component from inelastic scattering. Nonetheless, the persistence of the $\beta_1$, $\beta_2$, and $\gamma_2$ splitting above $T_{\mathrm{C}}$ and the consistency in its magnitude for both materials prompted us to consider non-magnetic origins.
(a) and (b) ARPES dispersions for taken at 18 and 145 K, respectively. (c) and (d) Momentum distribution curves for for $E - E_{\mathrm{F}} = -15$ meV, and fits to multiple Lorentzian peaks (dotted lines), taken at 18 and 145 K, respectively. (e) and (f) ARPES dispersions for taken at 18 and 170 K, respectively. (g) and (h) Momentum distribution curves for for $E - E_{\mathrm{F}} = -15$ meV, and fits to multiple Lorentzian peaks (dotted lines), taken at 18 and 170 K, respectively. All data were measured with $h\nu = 79$ eV and LH polarization.
§.§ ARPES Measurements with Micron-scale Probes
Motivated by the polar nature of these materials and the observation of unexplained band splitting, we carried out ARPES experiments on with a smaller beam size (2–15 µm) to investigate the possible impact of nonuniform sample surfaces. We identified three types of distinct areas based on their Fermi surfaces (Figure <ref>a–c), core level spectra (Figure <ref>d–e), and band dispersions (Figure <ref>f–h). Spots with the simplest Fermi surfaces and the largest hole pockets around $\Gamma$ and $\mathrm{K_0}$ (Figure <ref>a) have the weakest Cr $2p$ core level spectra (Figure <ref>d). Spots with Fermi surfaces representative of the majority of the samples, with the aforementioned $\beta$ and $\gamma$ band doubling (Figure <ref>b), exhibit Cr $2p$ core level signals of intermediate intensity. Finally, spots with Fermi surfaces missing the broadest outermost pockets around $\Gamma$ and $\mathrm{K_0}$ (Figure <ref>c) show the strongest Cr $2p$ core level spectra. Based on the Cr core level intensities, these areas appear to correspond to low, intermediate, and high relative Cr surface concentrations, respectively. The trend in the S $2p$ core levels from the same spots corroborate this assignment: with decreasing Cr surface coverage, an S peak at lower binding energies grows in (indicated by the green arrow in Figure <ref>e), consistent with more reduced S sites on the surface that are not sharing electron density with Cr.
(a–c) ARPES Fermi surfaces of taken in regions of low, intermediate, and high surface Cr coverage, respectively (20 K, $h\nu$ = 120 eV). (d) and (e) Cr $2p$ and S $2p$ core level spectra, respectively, with the green arrow in (e) indicating the S $2p$ feature at low binding energy. (f–h) ARPES band dispersions for the same regions with low, intermediate, and high surface Cr coverage shown in (a–c). Blue arrows in (g) and (h) indicate non-dispersive features. (i) DFT orbital-projected band structure for showing atomic origin of bands.
The ARPES dispersions from these spots also exhibit notable differences. The “low Cr” spot (Figure <ref>f) exhibits less $\sqrt{3} \times \sqrt{3}$ superlattice reconstruction than the other spots (as seen from the apparent three-fold symmetry around $\mathrm{K_0}$ and different sized hole pockets at $\Gamma$ and $\mathrm{K_0}$) and resembles $2H$-NbS2 with $E_{\mathrm{F}}$ shifted up by approximately 250 meV.elyoubi2021 The intense “X”-shaped feature at $\Gamma$ located at about $-1.6$ eV in “low Cr” is shifted down to about $-2.0$ eV in both “intermediate Cr” (Figure <ref>g) and “high Cr” (Figure <ref>h) consistent with the latter two sampling more electron-doped states on average. In the “intermediate Cr” spot, the $\gamma$ band electron pockets near $E_{\mathrm{F}}$ are split by about 250 meV, as they are in other spectra measured with larger beam sizes. In the “high Cr” spot, the electron pockets near $E_{\mathrm{F}}$ are not noticeably split; instead, only a single set of features resembling the lower $\gamma_1$ band in “intermediate Cr” is observed. Additionally, the presence of flat bands in the “intermediate Cr” and “high Cr” spots (where they are especially prominent), as indicated by the blue arrows in Figure <ref>g–h, coincides with Cr states in the orbital-projected DFT band structures (Figure <ref>i), lending further support to the surface coverage assignments.
§ DISCUSSION
§.§ Band Structure and Magnetic Exchange Interactions
Taking the results from both ARPES and DFT into account, the most pronounced difference in the band structures of and is the more dispersive bands in the Ta analogue. The origin appears to be steeper dispersions in $2H$-TaS2 compared to $2H$-NbS2, i.e. the relative band dispersions of the host lattice materials are retained after Cr intercalation. This trend can be attributed to better overlap facilitated by more extended Ta $5d$ orbitals compared to the Nb $4d$ orbitals. For a more detailed explanation, we briefly discuss the salient bonding interactions in both host lattice materials.
(a) Orbital-projected band structure calculation for $2H$-NbS2, showing Nb $4d$ and S $3p$ character. (b) Orbital-projected band structure calculation for $2H$-TaS2, showing Ta $5d$ and S $3p$ character. (c) Projected density of states (DOS) calculation for $2H$-NbS2. (d) Projected DOS calculation for $2H$-TaS2. The energy spread of the Nb $4d$ and Ta $5d$ bands is indicated on the right.
The bands within about 6 eV of the Fermi level in $2H$-$M$S2 ($M$ = Nb or Ta) are comprised of $M$ $d$ states and S $3p$ states, indicating that $M$–S $d$–$p$ and $M$–$M$ $d$–$d$ interactions are those relevant to determining the strength of the bonding and the resulting dispersivity of the bands. Mixing among the $M$ $d$ orbitals results in the formation of a hybridization gap within the $d$ manifold: the bands crossing $E_\mathrm{F}$ are composed of $d_{z^2}$, $d_{xy}$, and $d_{x^2-y^2}$ orbitals, while the higher-lying $d$ bands have more $d_{xz}$ and $d_{yz}$ character.mattheiss1973 TaS2 has a slightly smaller in-plane lattice constant,fisher1980,meetsma1990 and the $5d$ orbitals are more spatially extended than the $4d$ orbitals in NbS2. This leads to better relative overlap in the Ta analogue, both in terms of $M$–S $d$–$p$ bonds and next-nearest-neighbor $M$–$M$ $d$–$d$ interactions. Hence, overall, the $d$ manifold of TaS2 is more dispersive than that of NbS2, as shown in the band structure and density of states (DOS) calculations in Figure <ref>b and c. In turn, the bands crossing $E_\mathrm{F}$ are also more dispersive in the Ta analogue.
These arguments can also be used to explain why the host lattice-derived bands are more dispersive in than , which we have observed in ARPES and DFT. From the crystal structures, the Ta–S bonds in are slightly shorter (2.488(3) Å on average) than the Nb–S bonds in (2.4931(11) Å on average), and the in-plane lattice constant in is slightly smaller (5.7155(5) Å vs. 5.7400(7) Å), suggestive of stronger Ta–S overlap compared to Nb–S overlap. In addition, the in-plane electrical conductivity of is more than an order of magnitude higher than that of , which is consistent with the more dispersive bands and larger hole pockets in the Ta analogue expected from this analysis.ghimire2013,obeysekera2021
The more dispersive bands in may have implications for the relative strengths of magnetic exchange interactions in these materials. As previously established in the literature, has a higher $T_\mathrm{C}$ than (133 K vs. 110 K for our samples). Given the extensive mixing of intercalant- and host lattice-derived states near $E_{\mathrm{F}}$ shown in DFT and supported by ARPES, our findings suggest that the magnetic exchange interactions involve some degree of itinerancy for Cr-based spins. Nevertheless, due to the considerable distance ($> 5.7$ Å) between Cr sites, magnetic exchange is likely still mediated through TMD-derived states. The modestly higher $T_\mathrm{C}$ of may reflect stronger FM coupling (i.e. larger $J_{\parallel}$ and $J_{\perp}$) in this material, which could arise from itinerant carriers with higher $v_{\mathrm{F}}$. Nevertheless, the shorter soliton wavelength and larger $H_\mathrm{c}$ imply that the ratio $D/J_{\perp}$ is considerably larger than in the Nb analogue, i.e. $D$ (the DM interaction term) increases more than $J_{\perp}$ in going from Nb to Ta. Higher $D$ in the Ta analogue can be ascribed to larger spin–orbit coupling (SOC),zhang2021,obeysekera2021,du2021 and appears to affect the length and energy scales of the chiral spin textures more than changes in $J$.
The fact that and have qualitatively analogous magnetic phase diagrams despite different Fermi wavevectors, $k_\mathrm{F}$, supports the notion that the Ruderman–Kittel–Kasuya–Yosida (RKKY) interaction does not adequately describe exchange coupling in these materials.ghimire2013,sirica2021 According to the RKKY formalism, the sign and magnitude of $J$ would depend closely on the magnitude of $k_\mathrm{F}$, which does not appear to be the case for and . However, the exchange interactions in other magnetic intercalated TMDs, such as Fe_$x$NbS2 and Fe_$x$TaS2, have been proposed to be more RKKY-like.ko2011,zheng2021,wu2022 Based on our results, we posit that TaS2-based materials may have more dispersive bands than NbS2-based materials for intercalants other than Cr. This difference (along with SOC) may be pertinent to the disparate magnetic properties of Nb and Ta analogues in other families of intercalated TMDs (e.g. Fe_1/3NbS2 is an antiferromagnet, whereas Fe_1/3TaS2 is a ferromagnet).
The most obvious route toward band engineering in and is through modulation of the Cr concentration. However, a delicate balance exists between disorder and vacancies in the Cr superlattice and the integrity of the desired spin textures.dyadkin2015,kousaka2022,goodge2023 Co-intercalation of another species into the interstitial spacepan2023 or substitutional doping on the TMD sublattice could constitute other pathways toward tuning the filling level while maintaining a well-ordered Cr superlattice (and hence a globally defined $D$). Future studies in these directions could more definitively probe the effects of band dispersion on the resultant spin textures. We note additionally that the sensitivity of the observed surface states on Cr concentration—as discussed in more detail in the next section—suggests that the surface electronic structure is amenable to tuning through further functionalization.
§.§ Exchange Splitting vs. Surface Termination Effects
Previous ARPES studies on have also reported band splitting near $E_{\mathrm{F}}$ that appears similar to our assignment of $\beta_1$, $\beta_2$, and $\gamma_2$ bands. These works interpreted this phenomenon as exchange splitting.sirica2016,sirica2021,qin2022 However, we observed good agreement between the exchange splitting predicted by our FM spin-polarized DFT band structures and observed in our ARPES results. Specifically, according to the spin-polarized DFT band structure calculations shown in Figure <ref>c–d, the $\alpha$ and $\beta$ bands (as labeled in Figure <ref>f) are an exchange-split pair. Because of different mixing in the spin-majority and spin-minority channels, $\alpha$ has more $d_{xy}$/$d_{x^2-y^2}$ character, while $\beta$ has more $d_{z^2}$ character (Figure <ref>c and g). The corresponding bands observed in ARPES show the expected polarization dependence: the $\alpha$ band, which just touches $E_\mathrm{F}$, is much more visible in LV polarization (Figure <ref>a and e), whereas the $\beta$ band is more prominent in LH polarization (Figure <ref>b and f). This prompted us to consider alternative explanations for the observation of more bands in ARPES than predicted by DFT.
Instead, taking the polar nature of intercalated TMDs into account, we surmised that the observed doubling of $\beta_1$/$\beta_2$ and $\gamma_1$/$\gamma_2$ bands could be attributed to surface termination effects. Previous work suggests that spatially distinct areas of Cr- and $M$S2-termination exist on cleaved surfaces: STM studies have observed islands of intercalants on cleaved crystals of intercalated TMDs,sirica2016,lim_tunable_2022 and recent ARPES studies of V_1/3NbS2edwards2023 and Co_1/3NbS2zhang2023 reported termination-dependent surface states. Thus, to understand the expected effects of surface termination on the ARPES data, we consider the charge distributions for Cr-terminated and $M$S2-terminated surfaces. As summarized in Figure <ref>a, the Cr^3+ centers formally donate one electron per $M$S2 formula unit, leading to alternating layers with $+1$ and $-1$ formal charges. The phenomenon of charge redistribution at polar-to-nonpolar interfaces to prevent a polar catastrophe (i.e. diverging electrostatic potential) is well-documented,nakagawa2006 and we expect analogous redistribution to occur at both the [$M$S2]–vacuum interface and the [Cr_1/3]–vacuum interface. Assuming fully occupied (unoccupied) Cr sites for the Cr- ($M$S2-) terminated regions, the surface formal charges expected from simple electron counting are shown in Figure <ref>. In short, $M$S2-terminated regions should exhibit $M$S2-derived surface states that are hole-doped relative to the bulk, originating from partial electron transfer from the surface TMD layer to compensate for its polarity.
Schematic illustration of local Cr clustering on cleaved surfaces of samples, with two distinct regions corresponding to predominantly $M$S2 and Cr terminations. Formal charges given for the two regions are based on a simple electron-counting picture, assuming completely full (absent) Cr coverage in the Cr- ($M$S2-) terminated regions. The “low Cr,” “intermediate Cr,” and “high Cr” sampling areas (colored boxes) thus have average surface stoichiometries of 0, 1/6, and 1/3.
Building upon this picture, we propose that the three distinct spectral signatures observed with mesoscopic probes (Figure <ref>) can be explained by sampling two different surface terminations, as illustrated in Figure <ref>. The “low Cr” and “high Cr” spots correspond to areas with almost exclusively $M$S2 termination and Cr termination, respectively. The “intermediate Cr” spots contain both terminations, resulting in doubled $\beta$ and $\gamma$ bands that are also observed in data collected with larger probes. The $\beta$ and $\gamma$ bands at lower binding energies are associated with $M$S2-terminated regions. Such effects are broadly consistent with those observed on surfaces of other polar layered materials.hossain2008 Hence, we propose that these additional band doublings do not reflect magnetic ordering, but rather originate from charge redistribution. More generally, the possible contribution of surface states should be considered in other cases where unexpected bands are observed in ARPES studies on magnetic intercalated TMDs.
§ CONCLUSIONS
The electronic structures of the chiral helimagnets and have been investigated using ARPES and DFT. Compared to the host lattice materials $2H$-NbS2 and $2H$-TaS2, the Cr-intercalated materials exhibit band folding from the $\sqrt{3} \times \sqrt{3}$ Cr superlattice, higher $E_{\mathrm{F}}$ from electron transfer from Cr to the host TMD, exchange splitting from the in-plane FM ordering of Cr moments, and new bands from hybridization between Cr and TMD states. The chief difference between the band structures of $2H$-NbS2 and $2H$-TaS2—higher band dispersions in the Ta analogue—are retained after Cr intercalation, resulting in a larger $v_{\mathrm{F}}$ in . This finding may have implications for the higher $T_{\mathrm{C}}$ in via stronger FM coupling mediated by itinerant carriers.
By studying the polarization dependence in ARPES and fitting the MDCs, we find that the experimentally observed band structures agree well with the orbital-projected DFT band structures. The primary features at $E_{\mathrm{F}}$ in both materials consist of dispersive hole pockets at $\Gamma$ (and $\mathrm{K}_0$) and shallow electron pockets centered around $\mathrm{K_{SL}}$. Notably, many bands near $E_{\mathrm{F}}$ have significant Cr character in both materials, indicating that a rigid band model is insufficient for modeling the effects of Cr intercalation. Additional copies of bands crossing $E_{\mathrm{F}}$ that are not predicted by DFT are assigned to surface states originating from TMD-terminated regions. The observation of three distinct regions in ARPES experiments with smaller spot sizes is consistent with Cr, TMD, and mixed surface terminations. These results indicate that the polar nature of the surfaces of intercalated TMDs affects the band splitting observed in ARPES data.
It has been well-established that and have analogous magnetic phase diagrams with different energy scales and different wavelengths of magnetic solitons. Our results show that the electronic structures of these two isostructural materials are broadly analogous, with more dispersive bands in the Ta analogue. This finding suggests that band structure engineering and Fermi level tuning may allow for further modulation of the magnitude of $J$, and hence the length and energy scales of magnetic solitons in these materials. To maintain the $\sqrt{3} \times \sqrt{3}$ Cr superlattice, substitutional doping on the TMD sites or co-intercalation of other charged species may be viable routes toward controlling the properties of these chiral helimagnets, which are promising platforms for studying the interplay between electronic structure and the microscopic mechanisms underlying noncollinear magnetism.
§ METHODS
§.§ Crystal Growth
Single crystals of and were grown using chemical vapor transport using iodine as a transport agent. For , high-purity powders of elemental Cr, Nb, and S in a 0.6:1:2 ratio and 5 mg/cm$^3$ of I_2 were sealed under vacuum in a fused quartz ampoule approximately 48 cm long. The ampoule was placed in a three-zone tube furnace with the hot end zone and middle zone maintained at 1050
°C and the cold (growth) zone maintained at 850 °C for 14 days before cooling to room temperature. For , high-purity powders of elemental Cr, Ta, and S in a 0.47:1:2.1 ratio and 2 mg/cm$^3$ of I_2 were sealed under vacuum in a fused quartz ampoule approximately 25 cm long. The ampoule was placed in a two-zone tube furnace with the hot zone maintained at 1100 °C and the cold (growth) zone maintained at 1000 °C for 14 days before cooling to room temperature. Shiny plate-shaped crystals with a silvery metallic luster and hexagonal habit up to $4 \times 4 \times 0.5$ mm in size were obtained.
§.§ Structural and Compositional Characterization
Single crystal X-ray diffraction was collected on a Rigaku XtaLAB P200 with Mo K$\alpha$ radiation at 295 K. Data reduction and scaling and empirical absorption correction were performed in CrysAlis Pro. Structures were solved by direct methods using SHELXTsheldrick2015 and refined against $F^2$ on all data by full-matrix least squares with SHELXL.sheldrick2015a Raman spectroscopy was collected on a Horiba LabRAM HR Evolution with an ultra-low frequency filter using 633 nm laser excitation and powers between 1 and 8 mW. Energy dispersive X-ray spectroscopy was acquired on a FEI Quanta 3D FEG or a Scios 2 DualBeam scanning electron microscope with an accelerating voltage of 20 kV.
§.§ Magnetometry
DC magnetization measurements were carried out on a Quantum Design Physical Property Measurement System Dynacool equipped with a 12 T magnet using either the Vibrating Sample Magnetometer option or the AC Measurement System II option. Single crystals were affixed to quartz sample holders with GE Varnish such that the magnetic field was applied perpendicular to the crystallographic $c$ axis.
§.§ ARPES Measurements
ARPES data were collected at the Quantum Materials Spectroscopy Centre (QMSC) of the Canadian Light Source (CLS) and at Beamline 7.0.2 of the Advanced Light Source (ALS) on both the microARPES and nanoARPES endstations using Scienta Omicron R4000 hemispherical electron analyzers. The beam spot sizes were approximately $20 \times 100$ µm at QMSC, $15 \times 15$ µm on microARPES, and $2 \times 2$ µm on nanoARPES. Results were reproduced on multiple samples at both beamlines with the exception of spatial variation observed with smaller spot sizes. Samples were cooled down to the base temperature of 20 K or below and cleaved in situ by carefully knocking off alumina posts affixed to the top surface of the sample with silver epoxy. All measurements were conducted at pressures lower than $5 \times 10^{-11}$ Torr. The primary datasets were collected at photon energies of 46, 79, and 120 eV with linear horizontal and linear vertical polarizations. Data analysis was performed using the PyARPES software package.stansbury2020
§.§ DFT Calculations
First-principles calculations were performed by using the open source plane-wave code Quantum Espresso (QE).QE The optimized norm-conserving Vanderbilt (ONCV) pseudopotentials from the PseudoDojo project ONCV1,van2018pseudodojo were applied. The kinetic energy cut-off for wavefunctions were set to 86 Ry for all the self-consistent calculations; for these calculations, the experimental lattice constants obtained from X-ray diffraction were used.fisher1980,meetsma1990 A $\Gamma$-centered $4\times4\times2$ $\it{k}$-mesh was sampled in the Brillouin zone for both and , and a $8\times8\times2$ $\it{k}$-mesh for both $2H$-NbS2 and $2H$-TaS2. The Perdew–Burke–Ernzerhof (PBE) functional PBE1997 of the spin-polarized generalized gradient approximation (GGA) was used to describe the exchange correlation of electrons.
§ ACKNOWLEDGMENTS
We thank Sinéad Griffin and Ryan Day for helpful discussions, and thank Nicholas Settineri for assistance in obtaining the single crystal XRD data. The experimental work is based upon work supported by the Air Force Office of Scientific Research under AFOSR Award No. FA9550-20-1-0007. L.S.X. acknowledges support from the Arnold and Mabel Beckman Foundation (Award No. 51532) and L'Oréal USA (Award No. 52025) for postdoctoral fellowships. O.G. acknowledges support from an NSF Graduate Research Fellowship grant DGE 1752814, and National GEM Consortium Fellowship. S.H.R. was supported by the QSA, supported by the U.S. Department of Energy, Office of Science, National Quantum Information Science Research Centers. S.H. acknowledges support from the Blavatnik Innovation Fellowship. Y.P. acknowledges the financial support from the Air Force Office of Scientific Research under AFOSR Award No. FA9550-21-1-0087. Confocal Raman spectroscopy was supported by a Defense University Research Instrumentation Program grant through the Office of Naval Research under Award No. N00014-20-1-2599 (D.K.B.). Part of the research described in this paper was performed at the Canadian Light Source, a national research facility of the University of Saskatchewan, which is supported by the Canada Foundation for Innovation (CFI), the Natural Sciences and Engineering Research Council (NSERC), the National Research Council (NRC), the Canadian Institutes of Health Research (CIHR), the Government of Saskatchewan, and the University of Saskatchewan. This research used resources of the Advanced Light Source, which is a DOE Office of Science User Facility under contract no. DE-AC02-05CH11231. Other instrumentation used in this work was supported by grants from the Canadian Institute for Advanced Research (CIFAR–Azrieli Global Scholar, Award No. GS21-011), the Gordon and Betty Moore Foundation EPiQS Initiative (Award no. 10637), the W.M. Keck Foundation (Award No. 993922), and the 3M Foundation through the 3M Non-Tenured Faculty Award (No. 67507585). The computational part used resources of the Center for Functional Nanomaterials, which is a U.S. DOE Office of Science Facility, and the Scientific Data and Computing Center, a component of the Computational Science Initiative, at Brookhaven National Laboratory under Contract No. DE-SC0012704, and the lux supercomputer at the University of California, Santa Cruz, funded by NSF MRI, Grant No. AST 1828315, and used Stampede supercomputer at the University of Texas at Austin's Texas Advanced Computing Center (TACC) through allocation DMR160106 from the Advanced Cyberinfrastructure Coordination Ecosystem: Services & Support (ACCESS) program, which is supported by National Science Foundation grants #2138259, #2138286, #2138307, #2137603, and #2138296. This research was undertaken thanks in part to funding from the Max Planck–UBC–UTokyo Centre for Quantum Materials and the Canada First Research Excellence Fund, Quantum Materials and Future Technologies Program. This project is also funded by the Mitacs Accelerate Program; the QuantEmX Program of the Institute for Complex Adaptive Matter (ICAM); the Moore EPiQS Program (A.D.); and the CIFAR Quantum Materials Program (A.D.).
§ COMPETING INTERESTS
The authors declare no competing interests.
Missing 'biblatex' package
The bibliography requires the 'biblatex' package.
abstractMagnetic skyrmions are nanoscale spin configurations that hold promise as information carriers in ultradense memory and logic devices owing to the extremely low spin-polarized currents needed to move them.
journaltitleNat. Nanotechnol.
titleSkyrmions on the track
abstractRacetrack memory stores digital data in the magnetic domain walls of nanowires. This technology promises to yield information storage devices with high reliability, performance and capacity.
journaltitleNat. Nanotechnol.
titleMemory on the racetrack
issn0031-9015, 1347-4073
journaltitleJ. Phys. Soc. Jpn.
titleSymmetry, Structure, and Dynamics of Monoaxial Chiral Magnets
liliaxie/Zotero/storage/BTC5V7DL/Togawa et al. - 2016 - Symmetry, Structure, and Dynamics of Monoaxial Chi.pdf
journaltitleChem. Rev.
titleMagnetic Skyrmion Materials
abstractThe realization of chiral spin textures, comprising myriad distinct, nanoscale arrangements of spins with topological properties, has established pathways for engineering robust, energy-efficient, and scalable elements for non-volatile nanoelectronics. Particularly, current-induced manipulation of spin textures in nanowire racetracks and tunnel junction based devices are actively investigated for applications in memory, logic, and unconventional computing. In this Article, we paint a background on the progress of spin textures, as well as the relevant state-of-the-art techniques used for their development. In particular, we clarify the competing energy landscape of chiral spin textures─skyrmions and chiral domain walls, to tune their size, density, and zero-field stability. Next, we discuss the spin texture phenomenology and their response to extrinsic factors arising from geometric constraints, interwire interactions, and thermal-electrical effects. Finally, we reveal promising chiral spintronic memory and neuromorphic devices and discuss emerging material and device engineering opportunities.
journaltitleACS Appl. Electron. Mater.
titleChiral Spin Textures for Next-Generation Memory and Unconventional Computing
journaltitlePhys. Rev. B
titleMagnetic soliton confinement and discretization effects arising from macroscopic coherence in a chiral spin soliton lattice
journaltitle2D Mater.
titleExfoliation and van Der Waals Heterostructure Assembly of Intercalated Ferromagnet Cr_1/3TaS2
American Physical Society
journaltitlePhys. Rev. Lett.
titleControlling the Topological Sector of Magnetic Solitons in Exfoliated Cr_1/3NbS2
journaltitleRare Met.
titleMagnetic soliton confinement and discretization effects in Cr_1/3TaS2 nanoflakes
American Physical Society
journaltitlePhys. Rev. B
titleResponse of the chiral soliton lattice to spin-polarized currents
issn1364-2812, 1463-6417
journaltitlePhilos. Mag. B
title3d Transition-Metal Intercalates of the Niobium and Tantalum Dichalcogenides. I. Magnetic Properties
issn1364-2812, 1463-6417
journaltitlePhilos. Mag. B
title3d Transition-Metal Intercalates of the Niobium and Tantalum Dichalcogenides. II. Transport Properties
journaltitleBull. Soc. Chim. Fr.
titleEtude générale de systèmes $M_x$NbS2 (M élément de transition de la première période)
abstractA thermodynamic theory of “weak” ferromagnetism of α-Fe2O3, MnCO3 and CoCO3 is developed on the basis of landau's theory of phase transitions of the second kind. It is shown that the “weak” ferromagnetism is due to the relativistic spin-lattice and the magnetic dipole interactions. A strong dependence of the properties of “weak” ferromagnetics on the magnetic crystalline symmetry is noted and the behaviour of these ferromagnetics in a magnetic field is studied.
journaltitleJ. Phys. Chem. Solids
titleA thermodynamic theory of “weak” ferromagnetism of antiferromagnetics
abstractA theory of anisotropic superexchange interaction is developed by extending the Anderson theory of superexchange to include spin-orbit coupling. The antisymmetric spin coupling suggested by Dzialoshinski from purely symmetry grounds and the symmetric pseudodipolar interaction are derived. Their orders of magnitudes are estimated to be (Δgg) and (Δgg)2 times the isotropic superexchange energy, respectively. Higher order spin couplings are also discussed. As an example of antisymmetric spin coupling the case of CuCl2·2H2O is illustrated. In CuCl2·2H2O, a spin arrangement which is different from one accepted so far is proposed. This antisymmetric interaction is shown to be responsible for weak ferromagnetism in α-Fe2O3, MnCO3, and CrF3. The paramagnetic susceptibility perpendicular to the trigonal axis is expected to increase very sharply near the Néel temperature as the temperature is lowered, as was actually observed in CrF3.
journaltitlePhys. Rev.
titleAnisotropic Superexchange Interaction and Weak Ferromagnetism
issn0031-9015, 1347-4073
journaltitleJ. Phys. Soc. Jpn.
titleMagnetic Properties of Cr_1/3NbS2
issn0031-9007, 1079-7114
journaltitlePhys. Rev. Lett.
titleChiral Magnetic Soliton Lattice on a Chiral Helimagnet
issn1098-0121, 1550-235X
journaltitlePhys. Rev. B
titleMagnetic Phase Transition in Single Crystals of the Chiral Helimagnet Cr_1/3NbS2
liliaxie/Zotero/storage/QP5VGB9F/Ghimire et al. - 2013 - Magnetic phase transition in single crystals of th.pdf
issn0935-9648, 1521-4095
journaltitleAdv. Mater.
titleChiral Helimagnetism and One-Dimensional Magnetic Solitons in a Cr-Intercalated Transition Metal Dichalcogenide
issn2199-160X, 2199-160X
journaltitleAdv. Electron. Mater.
titleThe Magneto-Transport Properties of Cr_1/3TaS2 with Chiral Magnetic Solitons
issn0027-8424, 1091-6490
journaltitleProc. Natl. Acad. Sci.
titleTopological Spin/Structure Couplings in Layered Chiral Magnet Cr_1/3TaS2: The Discovery of Spiral Magnetic Superstructure
journaltitleAPL Mater.
titleAn Emergence of Chiral Helimagnetism or Ferromagnetism Governed by Cr Intercalation in a Dichalcogenide CrNb3S6
abstractThe chiral magnetic soliton, a topological kinklike spin texture, has significant applications in spintronic components. In this work, a crossover of critical behavior is found in Cr1/3TaS2, a chiral magnetic soliton host with the highest TC to date. Angular-dependent magnetization reveals that Cr1/3TaS2 exhibits an easy orientation within the isotropic ab plane, but displays anisotropy with the c axis. By using a modified iterative method, two distinct sets of critical exponents, including β−=0.3190(1) and γ−=1.263(8) for T≤TC, and β+=0.3475(2) and γ+=1.385(5) for T≥TC, are acquired on both sides of the transition. Analysis of the exponents indicates a crossover of the magnetic interaction from a three-dimensional Ising type below TC to a three-dimensional Heisenberg type above TC, implying nontrivial magnetism in this system. Based on universality scaling, a detailed H−T phase diagram around TC is constructed for H⊥c. The crossover of the critical behavior in Cr1/3TaS2 is peculiar to chiral magnetic soliton hosts, suggesting that the three-dimensional magnetic coupling is replaced by a one-dimensional one in the chiral magnetic soliton phase via a phase transition.
journaltitlePhys. Rev. B
titleCrossover of critical behavior and nontrivial magnetism in the chiral soliton lattice host Cr_1/3TaS2
abstractWe report long periodic chiral helimagnetic orderings in ferromagnetic inorganic compounds CrM3S6 (M = Nb and Ta) with a chiral space group of P6322. Magnetization in polycrystalline samples and high resolution powder neutron diffraction were measured. Our powder neutron diffraction measurements in CrM3S6 successfully separated nuclear and magnetic satellite peaks, having the period of hundreds of angstroms along the c— axis. Therefore, we propose that the magnetic ordering in ferromagnetic CrM3S6 is not ferromagnetic, but long periodic chiral helimagnetic ordering.
journaltitleJ. Phys. Conf. Ser.
titleLong Periodic Helimagnetic Ordering in Cr$M$3S6 ($M$ = Nb and Ta)
American Physical Society
journaltitlePhys. Rev. B
titleElectronic Structure of the Chiral Helimagnet and $3d$-Intercalated Transition Metal Dichalcogenide Cr_1/3NbS2
American Physical Society
journaltitlePhys. Rev. B
titleDisentangling Electronic, Lattice, and Spin Dynamics in the Chiral Helimagnet Cr_1/3NbS2
American Physical Society
journaltitlePhys. Rev. B
titlePersistent Exchange Splitting in the Chiral Helimagnet Cr_1/3NbS2
journaltitlePhys. Rev. B
titleEnergy-Gap Driven Low-Temperature Magnetic and Transport Properties in Cr_1/3$M$S2 ($M$ = Nb, Ta)
issn0002-7863, 1520-5126
journaltitleJ. Am. Chem. Soc.
titleStructure and Magnetism of Iron- and Chromium-Intercalated Niobium and Tantalum Disulfides
abstractThe nonrelativistic augmented-plane-wave (APW) method is applied to calculate the electronic band structures of several transition-metal-dichalcogenide (TX2) layer compounds, including materials with the C6(1T−HfS2,1T−TaS2), C27(2H−TaS2,2H−NbSe2), and C7(2H−MoS2) structure types. These calculations involve crystal potentials that are derived from neutral-atom charge densities. The results of these calculations confirm that the group-IVB (1T−HfS2) and group-VIB (2H−MoS2) compounds are semiconductors; the calculated indirect band gaps of 2.7 and 1.2 eV are in reasonable agreement with the observed values of 2.0 and 1.4 eV, respectively. Metallic behavior is predicted for the intermediate group-VB compounds 1T−TaS2, 2H−TaS2, and 2H−NbSe2. A novel feature of the metal d bands in the 2H−TX2 compounds is the occurence of a 1-eV hybridization gap within the dz2 and dxy, dx2−y2 manifolds. This splits off a pair of hybridized d bands which are half-filled in 2H−TaS2 and 2H−NbSe2 and completely filled in 2H−MoS2. As a result of this hybridization gap, the valence or conduction bandwidths in each of these 2H−TX2 compounds are reduced to about 1 eV.
journaltitlePhys. Rev. B
titleBand Structures of Transition-Metal-Dichalcogenide Layer Compounds
journaltitleInorg. Chem.
shorttitleUtility of semilocalized bonding schemes in extended systems
titleUtility of semilocalized bonding schemes in extended systems: three-center metal-metal bonding in molybdenum sulfide (MoS2), niobium tantalum sulfide bronze (H_$x$(Nb,Ta)S2), and zirconium sulfide (ZrS)
ZE884CV/ic00010a019.html:text/html;Full Text PDF:/Users/liliaxie/Zotero/storage/RARPPQ32/Yee and Hughbanks - 1991 - Utility of semilocalized bonding schemes in extend.pdf:application/pdf
journaltitleJ.. Am. Chem. Soc.
titleAnalogies between the concepts of molecular chemistry and solid-state physics concerning structural instabilities. Electronic origin of the structural modulations in layered transition metal dichalcogenides
nalogies between the concepts of molecular chemis.pdf:application/pdf
AIP Publishing LLC
journaltitleAppl. Phys. Lett.
titleSpin structure of the anisotropic helimagnet Cr_1/3NbS2 in a magnetic field
abstractThe topologically protected, chiral soliton lattice is a unique state of matter offering intriguing functionality, and it may serve as a robust platform for storing and transporting information in future spintronic devices. While the monoaxial chiral magnet Cr1∕3NbS2 is known to host this exotic state in an applied magnetic field, its detailed microscopic origin has remained a matter of debate. Here, we work towards addressing this open question by measuring the spin wave spectrum of Cr1∕3NbS2 over the entire Brillouin zone with inelastic neutron scattering. The well-defined spin wave modes allow us to determine the values of several microscopic interactions for this system. The experimental data are well-explained by a Heisenberg Hamiltonian with exchange constants up to the third nearest neighbor and an easy plane magnetocrystalline anisotropy term. Our work shows that both the second and third nearest neighbor exchange interactions contribute to the formation of the helimagnetic and chiral soliton lattice states in this robust three-dimensional magnet.
journaltitleAppl. Phys. Lett.
titleExtended exchange interactions stabilize long-period magnetic structures in Cr_1/3NbS2
:/Users/liliaxie/Zotero/storage/QBJM8LEA/Aczel et al. - 2018 - Extended exchange interactions stabilize long-peri.pdf:application/pdf;Snapshot:/Users/liliaxie/Zotero/storage/NP9QSWDF/Extended-exchange-interactions-stabilize-long.html:text/html
journaltitlePhys. Rev. B
titleTricritical point and phase diagram based on critical scaling in the monoaxial chiral helimagnet Cr_1/3NbS2
journaltitleNano Lett.
titleExcitations of Intercalated Metal Monolayers in Transition Metal Dichalcogenides
journaltitlePhys. Rev. B
titleFermiology and Electron-Phonon Coupling in the $2H$ and $3R$ Polytypes of NbS2
family=van Wezel,
American Physical Society
journaltitlePhys. Rev. B
titleOrbital Selectivity Causing Anisotropy and Particle-Hole Asymmetry in the Charge Density Wave Gap of $2H$-TaS2
abstractWe study the mechanism of the exceptionally large anomalous Hall effect (AHE) in the noncentrosymmetric antiferromagnet CoNb3S6 by angle-resolved photoemission spectroscopy (ARPES) and magnetotransport measurements. From ARPES measurements of CoNb3S6 and its family compounds (FeNb3S6 and NiNb3S6), we find a band dispersion unique to the Co intercalation existing near the Fermi level. We further demonstrate that a slight deficiency of sulfur in CoNb3S6 eliminates the ferromagnetism and the AHE simultaneously while hardly changing the band structure, indicating that the weak ferromagnetism is responsible for the emergence of the large AHE. Based on our results, we propose Weyl points near the Fermi level to cause the large AHE.
journaltitlePhys. Rev. B
titleLarge anomalous Hall effect induced by weak ferromagnetism in the noncentrosymmetric antiferromagnet CoNb3S6
American Physical Society
journaltitlePhys. Rev. B
titleVisualizing the Out-of-Plane Electronic Dispersions in an Intercalated Transition Metal Dichalcogenide
journaltitlePhys. Rev. B
titleRole of Intercalated Cobalt in the Electronic Structure of Co_1/3NbS2
journaltitleNat. Mater.
titleGiant valley-Zeeman coupling in the surface layer of an intercalated transition metal dichalcogenide
journaltitleRev. Mod. Phys.
titleAngle-Resolved Photoemission Studies of the Cuprate Superconductors
journaltitleInorg. Chem.
titleStoichiometry, structure, and physical properties of niobium disulfide
journaltitleActa Cryst. C
titleStructure of $2H$-TaS2
issn0031-9007, 1079-7114
journaltitlePhys. Rev. Lett.
titleRKKY Ferromagnetism with Ising-Like Spin States in Intercalated Fe_1/4TaS2
liliaxie/Zotero/storage/XBKIDHGE/Ko et al. - 2011 - RKKY Ferromagnetism with Ising-Like Spin States in.pdf
abstractDzyaloshinskii–Moriya interaction (DMI) is vital to form various chiral spin textures, novel behaviors of magnons and permits their potential applications in energy-efficient spintronic devices. Here, we realize a sizable bulk DMI in a transition metal dichalcogenide (TMD) 2H-TaS2 by intercalating Fe atoms, which form the chiral supercells with broken spatial inversion symmetry and also act as the source of magnetic orderings. Using a newly developed protonic gate technology, gate-controlled protons intercalation could further change the carrier density and intensely tune DMI via the Ruderman–Kittel–Kasuya–Yosida mechanism. The resultant giant topological Hall resistivity $${\rho }_{{xy}}ˆ{T}$$of $$1.41{\mathrm{\mu}} \Omega \cdot {{\mathrm{cm}}}$$at $${V}_{g}=-5.2{\mathrm{V}}$$(about $$424 \%$$larger than the zero-bias value) is larger than most known chiral magnets. Theoretical analysis indicates that such a large topological Hall effect originates from the two-dimensional Bloch-type chiral spin textures stabilized by DMI, while the large anomalous Hall effect comes from the gapped Dirac nodal lines by spin–orbit interaction. Dual-intercalation in 2H-TaS2 provides a model system to reveal the nature of DMI in the large family of TMDs and a promising way of gate tuning of DMI, which further enables an electrical control of the chiral spin textures and related electromagnetic phenomena.
journaltitleNat. Commun.
titleTailoring Dzyaloshinskii–Moriya interaction in a transition metal dichalcogenide by dual-intercalation
American Physical Society
journaltitlePhys. Rev. X
titleHighly Tunable Magnetic Phases in Transition-Metal Dichalcogenide Fe_$1/3+\delta$NbS2
journaltitlePhys. Rev. B
titleStructural Disorder versus Chiral Magnetism in Cr_1/3NbS2
abstractTransition metal-intercalated transition metal dichalcogenides (TMDs) are promising platforms for next-generation spintronic devices based on their wide range of electronic and magnetic phases, which can be tuned by varying the host lattice or the identity of the intercalant, along with its stoichiometry and spatial order. Some of these compounds host a chiral magnetic phase in which the helical winding of magnetic moments propagates along a high-symmetry crystalline axis. Previous studies have demonstrated that variation in intercalant concentrations can have a dramatic impact on the formation of chiral domains and ensemble magnetic properties. However, a systematic and comprehensive study of how atomic-scale order and disorder impacts collective magnetic behavior are so far lacking. Here, we leverage a combination of imaging modes in the (scanning) transmission electron microscope (S/TEM) to directly probe (dis)order across multiple length scales and show how subtle changes in the atomic lattice can be leveraged to tune the mesoscale spin textures and bulk magnetic response, with direct implications for the fundamental understanding and technological implementation of such compounds.
notearXiv:2305.06656 [cond-mat]
titleConsequences and control of multi-scale (dis)order in chiral magnetic textures
journaltitlePhys. Rev. Mater.
titleFe and Cr co-intercalation in $2H$-NbS2 single crystals for realization of perpendicular magnetic anisotropy and large anomalous Hall effect
issn1530-6984, 1530-6992
journaltitleNano Lett.
titleTunable Single-Atomic Charges on a Cleaved Intercalated Transition Metal Dichalcogenide
abstractThe Dirac equation combines the two cornerstones of modern physics-quantum mechanics and relativity. There are several manifestations of the Dirac equation in condensed matter systems, such as the quasiparticle dispersion in graphene, topological insulators, Dirac semimetals (DSMs), Weyl semimetals, and d-wave high-temperature superconductors. In a DSM, the massless Dirac fermion has zero chirality, leading to surface states connected adiabatically to a topologically trivial surface state as well as vanishing anomalous Hall effect (AHE). Recently, it is predicted that in the nonrelativistic limit of certain antiferromagnets, there exists a type of chiral 'Dirac-like' fermion, whose dispersion manifests four-fold degenerate crossing points formed by doubly degenerate linear bands, with topologically protected Fermi arcs. Such unconventional chiral fermion, protected by a hidden SU(2) symmetry in the hierarchy of an enhanced crystallographic group, namely spin space group, is not experimentally verified yet. Here, by combining neutron diffraction, angle-resolved photoemission spectroscopy and first-principles calculations, we reveal the existence of the Fermi-arc surface states induced by chiral Dirac-like fermions in collinear antiferromagnet CoNb3S6, which caught great interest due to its surprisingly large AHE. Our transport measurements and theoretical calculations provide a scenario that large Berry curvature embedded in the chiral fermions and weak symmetry breaking are responsible for the emergent AHE. Our work evidences the existence of chiral Dirac-like fermion in CoNb3S6, paving an avenue for exploring new emergent phenomena in quantum materials with unconventional quasiparticle excitations.
notearXiv:2301.12201 [cond-mat]
titleChiral Dirac fermion in a collinear antiferromagnet
journaltitleNat. Mater.
titleWhy some interfaces cannot be sharp
journaltitleNat. Phys.
titleIn situ doping control of the surface of high-temperature superconductors
International Union of Crystallography
abstractThe new computer program SHELXT employs a novel dual-space algorithm to solve the phase problem for single-crystal reflection data expanded to the space group P1. Missing data are taken into account and the resolution extended if necessary. All space groups in the specified Laue group are tested to find which are consistent with the P1 phases. After applying the resulting origin shifts and space-group symmetry, the solutions are subject to further dual-space recycling followed by a peak search and summation of the electron density around each peak. Elements are assigned to give the best fit to the integrated peak densities and if necessary additional elements are considered. An isotropic refinement is followed for non-centrosymmetric space groups by the calculation of a Flack parameter and, if appropriate, inversion of the structure. The structure is assembled to maximize its connectivity and centred optimally in the unit cell. SHELXT has already solved many thousand structures with a high success rate, and is optimized for multiprocessor computers. It is, however, unsuitable for severely disordered and twinned structures because it is based on the assumption that the structure consists of atoms.
journaltitleActa Cryst. A
titleSHELXT – Integrated Space-Group and Crystal-Structure Determination
International Union of Crystallography
abstractThe improvements in the crystal structure refinement program SHELXL have been closely coupled with the development and increasing importance of the CIF (Crystallographic Information Framework) format for validating and archiving crystal structures. An important simplification is that now only one file in CIF format (for convenience, referred to simply as `a CIF') containing embedded reflection data and SHELXL instructions is needed for a complete structure archive; the program SHREDCIF can be used to extract the .hkl and .ins files required for further refinement with SHELXL. Recent developments in SHELXL facilitate refinement against neutron diffraction data, the treatment of H atoms, the determination of absolute structure, the input of partial structure factors and the refinement of twinned and disordered structures. SHELXL is available free to academics for the Windows, Linux and Mac OS X operating systems, and is particularly suitable for multiple-core processors.
journaltitleActa Cryst. C
titleCrystal Structure Refinement with SHELXL
titlePyARPES: An Analysis Framework for Multimodal Angle-Resolved Photoemission Spectroscopies
family=Dal Corso,
abstractQUANTUM ESPRESSO is an integrated suite of computer codes for electronic-structure calculations and materials modeling, based on density-functional theory, plane waves, and pseudopotentials (norm-conserving, ultrasoft, and projector-augmented wave). The acronym ESPRESSO stands for opEn Source Package for Research in Electronic Structure, Simulation, and Optimization. It is freely available to researchers around the world under the terms of the GNU General Public License. QUANTUM ESPRESSO builds upon newly-restructured electronic-structure codes that have been developed and tested by some of the original authors of novel electronic-structure algorithms and applied in the last twenty years by some of the leading materials modeling groups worldwide. Innovation and efficiency are still its main focus, with special attention paid to massively parallel architectures, and a great effort being devoted to user friendliness. QUANTUM ESPRESSO is evolving towards a distribution of independent and interoperable codes in the spirit of an open-source project, where researchers active in the field of electronic-structure calculations are encouraged to participate in the project by contributing their own codes or by implementing their own ideas into existing codes.
journaltitleJ. Phys.: Condens. Matter
titleQUANTUM ESPRESSO: A Modular and Open-Source Software Project for Quantum Simulations of Materials.
American Physical Society
abstractFully nonlocal two-projector norm-conserving pseudopotentials are shown to be compatible with a systematic approach to the optimization of convergence with the size of the plane-wave basis. A reformulation of the optimization is developed, including the ability to apply it to positive-energy atomic scattering states and to enforce greater continuity in the pseudopotential. The generalization of norm conservation to multiple projectors is reviewed and recast for the present purposes. Comparisons among the results of all-electron and one- and two-projector norm-conserving pseudopotential calculations of lattice constants and bulk moduli are made for a group of solids chosen to represent a variety of types of bonding and a sampling of the periodic table.
journaltitlePhys. Rev. B
titleOptimized Norm-Conserving Vanderbilt Pseudopotentials
journaltitleComput. Phys. Commun.
titleThe PseudoDojo: Training and grading a 85 element optimized norm-conserving pseudopotential table
journaltitlePhys. Rev. Lett.
titleGeneralized Gradient Approximation Made Simple
journaltitleNano Lett.
titleExcitations of Intercalated Metal Monolayers in Transition Metal Dichalcogenides
family=Dal Corso,
abstractQUANTUM ESPRESSO is an integrated suite of computer codes for electronic-structure calculations and materials modeling, based on density-functional theory, plane waves, and pseudopotentials (norm-conserving, ultrasoft, and projector-augmented wave). The acronym ESPRESSO stands for opEn Source Package for Research in Electronic Structure, Simulation, and Optimization. It is freely available to researchers around the world under the terms of the GNU General Public License. QUANTUM ESPRESSO builds upon newly-restructured electronic-structure codes that have been developed and tested by some of the original authors of novel electronic-structure algorithms and applied in the last twenty years by some of the leading materials modeling groups worldwide. Innovation and efficiency are still its main focus, with special attention paid to massively parallel architectures, and a great effort being devoted to user friendliness. QUANTUM ESPRESSO is evolving towards a distribution of independent and interoperable codes in the spirit of an open-source project, where researchers active in the field of electronic-structure calculations are encouraged to participate in the project by contributing their own codes or by implementing their own ideas into existing codes.
journaltitleJ. Phys.: Condens. Matter
titleQUANTUM ESPRESSO: A Modular and Open-Source Software Project for Quantum Simulations of Materials.
American Physical Society
abstractFully nonlocal two-projector norm-conserving pseudopotentials are shown to be compatible with a systematic approach to the optimization of convergence with the size of the plane-wave basis. A reformulation of the optimization is developed, including the ability to apply it to positive-energy atomic scattering states and to enforce greater continuity in the pseudopotential. The generalization of norm conservation to multiple projectors is reviewed and recast for the present purposes. Comparisons among the results of all-electron and one- and two-projector norm-conserving pseudopotential calculations of lattice constants and bulk moduli are made for a group of solids chosen to represent a variety of types of bonding and a sampling of the periodic table.
journaltitlePhys. Rev. B
titleOptimized Norm-Conserving Vanderbilt Pseudopotentials
journaltitleComput. Phys. Commun.
titleThe PseudoDojo: Training and grading a 85 element optimized norm-conserving pseudopotential table
journaltitlePhys. Rev. Lett.
titleGeneralized Gradient Approximation Made Simple
issn1098-0121, 1550-235X
journaltitlePhys. Rev. B
titleMagnetic Phase Transition in Single Crystals of the Chiral Helimagnet Cr_1/3NbS2
liliaxie/Zotero/storage/QP5VGB9F/Ghimire et al. - 2013 - Magnetic phase transition in single crystals of th.pdf
journaltitlePhys. Rev. B
titleOut-of-Plane Spin-Orientation Dependent Magnetotransport Properties in the Anisotropic Helimagnet Cr_1/3NbS2
American Physical Society
journaltitlePhys. Rev. B
titleDisentangling Electronic, Lattice, and Spin Dynamics in the Chiral Helimagnet Cr_1/3NbS2
American Physical Society
journaltitlePhys. Rev. B
titlePersistent Exchange Splitting in the Chiral Helimagnet Cr_1/3NbS2
journaltitlePhys. Rev. B
titleMagnetism of chromia from first-principles calculations
journaltitlePhys. Rev. B
titleEffective Coulomb interaction in transition metals from constrained random-phase approximation
journaltitlePhys. Rev. B
titleCharge density wave phase, Mottness, and ferromagnetism in monolayer $1T$-NbSe2
journaltitlePhys. Rev. Lett.
titlePhonon-Assisted Magnetic Mott-Insulating State in the Charge Density Wave Phase of Single-Layer $1T$-NbSe2
issn0031-9015, 1347-4073
journaltitleJ. Phys. Soc. Jpn.
titleMagnetic Properties of Cr_1/3NbS2
journaltitleJ. Solid State Chem.
titleOn the magnetic behavior of new $2H$-NbS2-type derivatives
issn0935-9648, 1521-4095
journaltitleAdv. Mater.
titleChiral Helimagnetism and One-Dimensional Magnetic Solitons in a Cr-Intercalated Transition Metal Dichalcogenide
issn2199-160X, 2199-160X
journaltitleAdv. Electron. Mater.
titleThe Magneto-Transport Properties of Cr_1/3TaS2 with Chiral Magnetic Solitons
issn0027-8424, 1091-6490
journaltitleProc. Natl. Acad. Sci.
titleTopological Spin/Structure Couplings in Layered Chiral Magnet Cr_1/3TaS2: The Discovery of Spiral Magnetic Superstructure
Supporting Information:
Comparative Electronic Structures of the Chiral Helimagnets and
Lilia S. Xie, Oscar Gonzalez, Kejun Li, Matteo Michiardi, Sergey Gorovikov, Sae Hee Ryu, Shannon S. Fender, Marta Zonno, Na Hyun Jo, Sergey Zhdanovich, Chris Jozwiak, Aaron Bostwick, Samra Husremović, Matthew P. Erodici, Cameron Mollazadeh, Andrea Damascelli, Eli Rotenberg, Yuan Ping, D. Kwabena Bediako
§ SINGLE CRYSTAL X-RAY DIFFRACTION
Crystal structures of and from single crystal X-ray diffraction. Representations are shown for and , respectively, along the $a$ crystallographic axis in (a) and (b), and $c$ crystallographic axis in (c) and (d).
Crystal data and structure refinement for and .
Empirical formula CrNb3S6 CrTa3S6
Formula weight (g/mol) 523.09 787.21
Temperature (K) 293(2) 293(2)
Wavelength (Å) 0.71073 0.71073
Crystal system Hexagonal Hexagonal
Space group $P6_322$ $P6_322$
$a$ (Å) 5.7400(7) 5.7155(5)
$c$ (Å) 12.1082(14) 12.1751(12)
Volume (Å$^{-3}$) 345.49(9) 344.44(7)
$Z$ 2 2
Density (calculated) (g/cm$^3$) 3.402 5.680
Absorption coefficient (mm$^{-1}$) 8.082 50.733
$F(000)$ 486 678
Crystal size (mm$^3$) $0.033 \times 0.017 \times 0.013$ $0.119 \times 0.067 \times 0.025$
$\theta$ ($\degree$) 3.365 to 29.272 4.117 to 29.531
Index ranges $-6 \leq h \leq 7$ $-7 \leq h \leq 7$
$-7 \leq k \leq 6$ $-7 \leq k \leq 7$
$-14 \leq l \leq 16$ $-16 \leq l \leq 16$
Reflections collected 2684 2641
Independent reflections 297 302
Completeness to $\theta_\mathrm{full}$ 1.000 0.994
Absorption correction Semi-empirical from equivalents Semi-empirical from equivalents
Refinement method Full-matrix least-squares on $F^2$ Full-matrix least-squares on $F^2$
Data / restraints / parameters 297 / 0 / 17 302 / 0 / 17
Goodness-of-fit on $F^2$ 1.312 1.156
Final $R$ indices [$I > 2\sigma(I)$] $R_1$ = 0.0331, $wR_2$ = 0.0848 $R_1$ = 0.0381, $wR_2$ = 0.1166
$R$ indices (all data) $R_1$ = 0.0435, $wR_2$ = 0.0880 $R_1$ = 0.0437, $wR_2$ = 0.1213
Largest diff. peak and hole ($e$ Å$^{-3}$) 1.81 and $-0.65$ 4.83 and $-1.89$
Atomic coordinates, Wyckoff positions, and equivalent isotropic displacement parameters for .
Atom Labels $x$ $y$ $z$ Site $U_{\mathrm{iso}}$
Cr01 2/3 1/3 3/4 $2c$ 0.0067(6)
Nb02 0 0 1/2 $2a$ 0.0037(4)
Nb03 1/3 2/3 0.50283(6) $4f$ 0.00236(3)
S04 0.6680(3) 0.6675(3) 0.63086(11) $12i$ 0.0048(4)
Atomic coordinates, Wyckoff positions, and equivalent isotropic displacement parameters for .
Atom Labels $x$ $y$ $z$ Site $U_{\mathrm{iso}}$
Cr01 2/3 1/3 3/4 $2c$ 0.0098(7)
Ta02 0 0 1/2 $2a$ 0.0047(5)
Ta03 1/3 2/3 0.50232(4) $4f$ 0.0048(5)
S04 0.6679(3) 0.6685(3) 0.6304(2) $12i$ 0.0058(7)
§ RAMAN SPECTROSCOPY
Raman spectra of and , with modes associated with the $\sqrt{3} \times \sqrt{3}$ superlattice labeled as “SL” and modes associated with the host lattice materials labeled according to symmetry.fan2021
§ ENERGY DISPERSIVE X-RAY SPECTROSCOPY
Representative dispersive X-ray spectroscopy data for a single crystal of with peaks corresponding to Cr, Nb, and S labeled. The atomic ratio determined by fitting the Cr K$\alpha_1$, Nb L$\alpha_1$, and S K$\alpha_1$ peaks was 1.00:3.00:6.30., corresponding to a formula of Cr_0.33NbS_2.10.
Representative dispersive X-ray spectroscopy data for a single crystal of with peaks corresponding to Cr, Ta, and S labeled. The atomic ratio determined by fitting the Cr K$\alpha_1$, Ta M$\alpha_1$, and S K$\alpha_1$ peaks was 1.00:3.07:5.70, corresponding to a formula of Cr_0.33TaS_1.86.
§ MAGNETOMETRY
Temperature-dependent magnetic susceptibility for , measured with $H \perp c$ = 50 Oe.
Temperature-dependent magnetic susceptibility for , measured with $H \perp c$ = 200 Oe.
§ DFT CALCULATIONS
To understand the band structures obtained from ARPES measurements, first-principles calculations for pristine $2H$-NbS2, $2H$-TaS2, and the Cr-intercalated analogs of these host lattices were performed by using the open source plane-wave code Quantum Espresso (QE).QE The optimized norm-conserving Vanderbilt (ONCV) pseudopotentials from the PseudoDojo project ONCV1,van2018pseudodojo were applied. The kinetic energy cut-off for wavefunctions were set to 86 Ry for all the self-consistent calculations; for these calculations, the experimental lattice constants obtained from X-ray diffraction were used. A $\Gamma$-centered $4\times4\times2$ $\it{k}$-mesh was sampled in the Brillouin zone for both and , and a $8\times8\times2$ $\it{k}$-mesh for both $2H$-NbS2 and $2H$-TaS2. The Perdew–Burke–Ernzerhof (PBE) functional PBE1997 of the spin-polarized generalized gradient approximation (GGA) was used to describe the exchange-correlation of electrons. Previous studies incorporating first-principles calculations at the PBE levelghimire2013,bornstein2015,sirica2021 and at the GGA+$U$ level qin2022 used an on-site Coulomb interaction, $U$, of 4 eV for Cr. In this work, different $U$ parameters for Cr, Nb, and Ta were explored. The results obtained from PBE are shown and compared to the experiments in the main text. Calculations obtained for the spin-polarized, orbital-projected, $k_z$-dependent band structures, and the Fermi surfaces are shown in comparison with the results.
§.§ U Parameter for Band Structure from GGA+U Calculations
Cr, Nb, and Ta are transition metals for which the $U$ parameter can be used to describe the on-site Coulomb interaction between localized $d$ electrons.shi2009magnetism,csacsiouglu2011effective However, whether or not the $U$ parameters for the aforementioned transition metals are important for the band structure calculations of and is not clear. It needs more investigation beyond the previous studies using PBEghimire2013,bornstein2015,sirica2021 and the study using GGA+$U$ with a $U$ value of 4 eV for Cr.qin2022 Here, the on-site Coulomb interaction $U$ parameters for Cr and Nb are tested using the values close to those from Ref. <cit.>. The $U$ parameters are discussed in order to clarify the effect of $U$ parameters on the band structure and the necessity of adopting $U$ parameters.
Looking at the band structures with varying $U$(Cr) in Figure <ref>a, overall the band structures of interest within $-$1 eV to 0 eV do not change much with $U$(Cr). Moreover, the band structures near the Fermi level do not show qualitative changes, except for the minor upshift of the bands at $\Gamma$ with increasing $U$(Cr).
The small change of this part of band structures may be explained by the fact that $d$ orbitals of Cr are minor in the composition as shown in Figure <ref>c of the partial density of states (PDOS), and that the $d$ electrons near the Fermi level are nearly delocalized and thus not affected by the $U$ parameter. Likewise, $U$(Nb) can be found to be not important to the band structures of interest from Figure <ref>b. Thus, the band structures from PBE are shown in the main text and compared with experiments.
(a) and (b) Band structures of with the on-site Coulomb interaction for Cr and Nb represented as $U$(Cr) and $U$(Nb), respectively. (c) density of states of calculated with PBE. (a) $U$(Nb) is set to 0 eV and the effect of $U$(Cr) on the band structure is shown when varied from 0 to 4 eV. (b) $U$(Cr) is set to 0 eV and the effect of $U$(Nb) on the band structure is shown when varied from 0 to 3 eV.
(a) and (b) $k_z$-projected band structures of and , respectively, calculated using the PBE functional.
Magnetic moment of .
Stoichiometry Saturation magnetization ($\mu_\text{B}$) Reference Notes
1/3 2.9 <cit.> $M(H)$ (4.2 K)
1/3 3.89 <cit.> $M(H)$ (2 K)
1/3 3.2 <cit.> $M(H)$ (2 K)
0.33(1) 2.68 This work $M(H)$ (2 K)
1/3 2.66 This work DFT (PBE)
Magnetic moment of .
Stoichiometry Saturation magnetization ($\mu_\text{B}$) Reference Notes
1/3 2.97 <cit.> $M(H)$ (2 K)
1/3 2.73 <cit.> $M(H)$ (2 K)
1/3 2.73 <cit.> $M(H)$ (2 K)
0.33(1) 2.82 This work $M(H)$ (2 K)
1/3 2.71 This work DFT (PBE)
Missing 'biblatex' package
The bibliography requires the 'biblatex' package.
abstractMagnetic skyrmions are nanoscale spin configurations that hold promise as information carriers in ultradense memory and logic devices owing to the extremely low spin-polarized currents needed to move them.
journaltitleNat. Nanotechnol.
titleSkyrmions on the track
abstractRacetrack memory stores digital data in the magnetic domain walls of nanowires. This technology promises to yield information storage devices with high reliability, performance and capacity.
journaltitleNat. Nanotechnol.
titleMemory on the racetrack
issn0031-9015, 1347-4073
journaltitleJ. Phys. Soc. Jpn.
titleSymmetry, Structure, and Dynamics of Monoaxial Chiral Magnets
liliaxie/Zotero/storage/BTC5V7DL/Togawa et al. - 2016 - Symmetry, Structure, and Dynamics of Monoaxial Chi.pdf
journaltitleChem. Rev.
titleMagnetic Skyrmion Materials
abstractThe realization of chiral spin textures, comprising myriad distinct, nanoscale arrangements of spins with topological properties, has established pathways for engineering robust, energy-efficient, and scalable elements for non-volatile nanoelectronics. Particularly, current-induced manipulation of spin textures in nanowire racetracks and tunnel junction based devices are actively investigated for applications in memory, logic, and unconventional computing. In this Article, we paint a background on the progress of spin textures, as well as the relevant state-of-the-art techniques used for their development. In particular, we clarify the competing energy landscape of chiral spin textures─skyrmions and chiral domain walls, to tune their size, density, and zero-field stability. Next, we discuss the spin texture phenomenology and their response to extrinsic factors arising from geometric constraints, interwire interactions, and thermal-electrical effects. Finally, we reveal promising chiral spintronic memory and neuromorphic devices and discuss emerging material and device engineering opportunities.
journaltitleACS Appl. Electron. Mater.
titleChiral Spin Textures for Next-Generation Memory and Unconventional Computing
journaltitlePhys. Rev. B
titleMagnetic soliton confinement and discretization effects arising from macroscopic coherence in a chiral spin soliton lattice
journaltitle2D Mater.
titleExfoliation and van Der Waals Heterostructure Assembly of Intercalated Ferromagnet Cr_1/3TaS2
American Physical Society
journaltitlePhys. Rev. Lett.
titleControlling the Topological Sector of Magnetic Solitons in Exfoliated Cr_1/3NbS2
journaltitleRare Met.
titleMagnetic soliton confinement and discretization effects in Cr_1/3TaS2 nanoflakes
American Physical Society
journaltitlePhys. Rev. B
titleResponse of the chiral soliton lattice to spin-polarized currents
issn1364-2812, 1463-6417
journaltitlePhilos. Mag. B
title3d Transition-Metal Intercalates of the Niobium and Tantalum Dichalcogenides. I. Magnetic Properties
issn1364-2812, 1463-6417
journaltitlePhilos. Mag. B
title3d Transition-Metal Intercalates of the Niobium and Tantalum Dichalcogenides. II. Transport Properties
journaltitleBull. Soc. Chim. Fr.
titleEtude générale de systèmes $M_x$NbS2 (M élément de transition de la première période)
abstractA thermodynamic theory of “weak” ferromagnetism of α-Fe2O3, MnCO3 and CoCO3 is developed on the basis of landau's theory of phase transitions of the second kind. It is shown that the “weak” ferromagnetism is due to the relativistic spin-lattice and the magnetic dipole interactions. A strong dependence of the properties of “weak” ferromagnetics on the magnetic crystalline symmetry is noted and the behaviour of these ferromagnetics in a magnetic field is studied.
journaltitleJ. Phys. Chem. Solids
titleA thermodynamic theory of “weak” ferromagnetism of antiferromagnetics
abstractA theory of anisotropic superexchange interaction is developed by extending the Anderson theory of superexchange to include spin-orbit coupling. The antisymmetric spin coupling suggested by Dzialoshinski from purely symmetry grounds and the symmetric pseudodipolar interaction are derived. Their orders of magnitudes are estimated to be (Δgg) and (Δgg)2 times the isotropic superexchange energy, respectively. Higher order spin couplings are also discussed. As an example of antisymmetric spin coupling the case of CuCl2·2H2O is illustrated. In CuCl2·2H2O, a spin arrangement which is different from one accepted so far is proposed. This antisymmetric interaction is shown to be responsible for weak ferromagnetism in α-Fe2O3, MnCO3, and CrF3. The paramagnetic susceptibility perpendicular to the trigonal axis is expected to increase very sharply near the Néel temperature as the temperature is lowered, as was actually observed in CrF3.
journaltitlePhys. Rev.
titleAnisotropic Superexchange Interaction and Weak Ferromagnetism
issn0031-9015, 1347-4073
journaltitleJ. Phys. Soc. Jpn.
titleMagnetic Properties of Cr_1/3NbS2
issn0031-9007, 1079-7114
journaltitlePhys. Rev. Lett.
titleChiral Magnetic Soliton Lattice on a Chiral Helimagnet
issn1098-0121, 1550-235X
journaltitlePhys. Rev. B
titleMagnetic Phase Transition in Single Crystals of the Chiral Helimagnet Cr_1/3NbS2
liliaxie/Zotero/storage/QP5VGB9F/Ghimire et al. - 2013 - Magnetic phase transition in single crystals of th.pdf
issn0935-9648, 1521-4095
journaltitleAdv. Mater.
titleChiral Helimagnetism and One-Dimensional Magnetic Solitons in a Cr-Intercalated Transition Metal Dichalcogenide
issn2199-160X, 2199-160X
journaltitleAdv. Electron. Mater.
titleThe Magneto-Transport Properties of Cr_1/3TaS2 with Chiral Magnetic Solitons
issn0027-8424, 1091-6490
journaltitleProc. Natl. Acad. Sci.
titleTopological Spin/Structure Couplings in Layered Chiral Magnet Cr_1/3TaS2: The Discovery of Spiral Magnetic Superstructure
journaltitleAPL Mater.
titleAn Emergence of Chiral Helimagnetism or Ferromagnetism Governed by Cr Intercalation in a Dichalcogenide CrNb3S6
abstractThe chiral magnetic soliton, a topological kinklike spin texture, has significant applications in spintronic components. In this work, a crossover of critical behavior is found in Cr1/3TaS2, a chiral magnetic soliton host with the highest TC to date. Angular-dependent magnetization reveals that Cr1/3TaS2 exhibits an easy orientation within the isotropic ab plane, but displays anisotropy with the c axis. By using a modified iterative method, two distinct sets of critical exponents, including β−=0.3190(1) and γ−=1.263(8) for T≤TC, and β+=0.3475(2) and γ+=1.385(5) for T≥TC, are acquired on both sides of the transition. Analysis of the exponents indicates a crossover of the magnetic interaction from a three-dimensional Ising type below TC to a three-dimensional Heisenberg type above TC, implying nontrivial magnetism in this system. Based on universality scaling, a detailed H−T phase diagram around TC is constructed for H⊥c. The crossover of the critical behavior in Cr1/3TaS2 is peculiar to chiral magnetic soliton hosts, suggesting that the three-dimensional magnetic coupling is replaced by a one-dimensional one in the chiral magnetic soliton phase via a phase transition.
journaltitlePhys. Rev. B
titleCrossover of critical behavior and nontrivial magnetism in the chiral soliton lattice host Cr_1/3TaS2
abstractWe report long periodic chiral helimagnetic orderings in ferromagnetic inorganic compounds CrM3S6 (M = Nb and Ta) with a chiral space group of P6322. Magnetization in polycrystalline samples and high resolution powder neutron diffraction were measured. Our powder neutron diffraction measurements in CrM3S6 successfully separated nuclear and magnetic satellite peaks, having the period of hundreds of angstroms along the c— axis. Therefore, we propose that the magnetic ordering in ferromagnetic CrM3S6 is not ferromagnetic, but long periodic chiral helimagnetic ordering.
journaltitleJ. Phys. Conf. Ser.
titleLong Periodic Helimagnetic Ordering in Cr$M$3S6 ($M$ = Nb and Ta)
American Physical Society
journaltitlePhys. Rev. B
titleElectronic Structure of the Chiral Helimagnet and $3d$-Intercalated Transition Metal Dichalcogenide Cr_1/3NbS2
American Physical Society
journaltitlePhys. Rev. B
titleDisentangling Electronic, Lattice, and Spin Dynamics in the Chiral Helimagnet Cr_1/3NbS2
American Physical Society
journaltitlePhys. Rev. B
titlePersistent Exchange Splitting in the Chiral Helimagnet Cr_1/3NbS2
journaltitlePhys. Rev. B
titleEnergy-Gap Driven Low-Temperature Magnetic and Transport Properties in Cr_1/3$M$S2 ($M$ = Nb, Ta)
issn0002-7863, 1520-5126
journaltitleJ. Am. Chem. Soc.
titleStructure and Magnetism of Iron- and Chromium-Intercalated Niobium and Tantalum Disulfides
abstractThe nonrelativistic augmented-plane-wave (APW) method is applied to calculate the electronic band structures of several transition-metal-dichalcogenide (TX2) layer compounds, including materials with the C6(1T−HfS2,1T−TaS2), C27(2H−TaS2,2H−NbSe2), and C7(2H−MoS2) structure types. These calculations involve crystal potentials that are derived from neutral-atom charge densities. The results of these calculations confirm that the group-IVB (1T−HfS2) and group-VIB (2H−MoS2) compounds are semiconductors; the calculated indirect band gaps of 2.7 and 1.2 eV are in reasonable agreement with the observed values of 2.0 and 1.4 eV, respectively. Metallic behavior is predicted for the intermediate group-VB compounds 1T−TaS2, 2H−TaS2, and 2H−NbSe2. A novel feature of the metal d bands in the 2H−TX2 compounds is the occurence of a 1-eV hybridization gap within the dz2 and dxy, dx2−y2 manifolds. This splits off a pair of hybridized d bands which are half-filled in 2H−TaS2 and 2H−NbSe2 and completely filled in 2H−MoS2. As a result of this hybridization gap, the valence or conduction bandwidths in each of these 2H−TX2 compounds are reduced to about 1 eV.
journaltitlePhys. Rev. B
titleBand Structures of Transition-Metal-Dichalcogenide Layer Compounds
journaltitleInorg. Chem.
shorttitleUtility of semilocalized bonding schemes in extended systems
titleUtility of semilocalized bonding schemes in extended systems: three-center metal-metal bonding in molybdenum sulfide (MoS2), niobium tantalum sulfide bronze (H_$x$(Nb,Ta)S2), and zirconium sulfide (ZrS)
ZE884CV/ic00010a019.html:text/html;Full Text PDF:/Users/liliaxie/Zotero/storage/RARPPQ32/Yee and Hughbanks - 1991 - Utility of semilocalized bonding schemes in extend.pdf:application/pdf
journaltitleJ.. Am. Chem. Soc.
titleAnalogies between the concepts of molecular chemistry and solid-state physics concerning structural instabilities. Electronic origin of the structural modulations in layered transition metal dichalcogenides
nalogies between the concepts of molecular chemis.pdf:application/pdf
AIP Publishing LLC
journaltitleAppl. Phys. Lett.
titleSpin structure of the anisotropic helimagnet Cr_1/3NbS2 in a magnetic field
abstractThe topologically protected, chiral soliton lattice is a unique state of matter offering intriguing functionality, and it may serve as a robust platform for storing and transporting information in future spintronic devices. While the monoaxial chiral magnet Cr1∕3NbS2 is known to host this exotic state in an applied magnetic field, its detailed microscopic origin has remained a matter of debate. Here, we work towards addressing this open question by measuring the spin wave spectrum of Cr1∕3NbS2 over the entire Brillouin zone with inelastic neutron scattering. The well-defined spin wave modes allow us to determine the values of several microscopic interactions for this system. The experimental data are well-explained by a Heisenberg Hamiltonian with exchange constants up to the third nearest neighbor and an easy plane magnetocrystalline anisotropy term. Our work shows that both the second and third nearest neighbor exchange interactions contribute to the formation of the helimagnetic and chiral soliton lattice states in this robust three-dimensional magnet.
journaltitleAppl. Phys. Lett.
titleExtended exchange interactions stabilize long-period magnetic structures in Cr_1/3NbS2
:/Users/liliaxie/Zotero/storage/QBJM8LEA/Aczel et al. - 2018 - Extended exchange interactions stabilize long-peri.pdf:application/pdf;Snapshot:/Users/liliaxie/Zotero/storage/NP9QSWDF/Extended-exchange-interactions-stabilize-long.html:text/html
journaltitlePhys. Rev. B
titleTricritical point and phase diagram based on critical scaling in the monoaxial chiral helimagnet Cr_1/3NbS2
journaltitleNano Lett.
titleExcitations of Intercalated Metal Monolayers in Transition Metal Dichalcogenides
journaltitlePhys. Rev. B
titleFermiology and Electron-Phonon Coupling in the $2H$ and $3R$ Polytypes of NbS2
family=van Wezel,
American Physical Society
journaltitlePhys. Rev. B
titleOrbital Selectivity Causing Anisotropy and Particle-Hole Asymmetry in the Charge Density Wave Gap of $2H$-TaS2
abstractWe study the mechanism of the exceptionally large anomalous Hall effect (AHE) in the noncentrosymmetric antiferromagnet CoNb3S6 by angle-resolved photoemission spectroscopy (ARPES) and magnetotransport measurements. From ARPES measurements of CoNb3S6 and its family compounds (FeNb3S6 and NiNb3S6), we find a band dispersion unique to the Co intercalation existing near the Fermi level. We further demonstrate that a slight deficiency of sulfur in CoNb3S6 eliminates the ferromagnetism and the AHE simultaneously while hardly changing the band structure, indicating that the weak ferromagnetism is responsible for the emergence of the large AHE. Based on our results, we propose Weyl points near the Fermi level to cause the large AHE.
journaltitlePhys. Rev. B
titleLarge anomalous Hall effect induced by weak ferromagnetism in the noncentrosymmetric antiferromagnet CoNb3S6
American Physical Society
journaltitlePhys. Rev. B
titleVisualizing the Out-of-Plane Electronic Dispersions in an Intercalated Transition Metal Dichalcogenide
journaltitlePhys. Rev. B
titleRole of Intercalated Cobalt in the Electronic Structure of Co_1/3NbS2
journaltitleNat. Mater.
titleGiant valley-Zeeman coupling in the surface layer of an intercalated transition metal dichalcogenide
journaltitleRev. Mod. Phys.
titleAngle-Resolved Photoemission Studies of the Cuprate Superconductors
journaltitleInorg. Chem.
titleStoichiometry, structure, and physical properties of niobium disulfide
journaltitleActa Cryst. C
titleStructure of $2H$-TaS2
issn0031-9007, 1079-7114
journaltitlePhys. Rev. Lett.
titleRKKY Ferromagnetism with Ising-Like Spin States in Intercalated Fe_1/4TaS2
liliaxie/Zotero/storage/XBKIDHGE/Ko et al. - 2011 - RKKY Ferromagnetism with Ising-Like Spin States in.pdf
abstractDzyaloshinskii–Moriya interaction (DMI) is vital to form various chiral spin textures, novel behaviors of magnons and permits their potential applications in energy-efficient spintronic devices. Here, we realize a sizable bulk DMI in a transition metal dichalcogenide (TMD) 2H-TaS2 by intercalating Fe atoms, which form the chiral supercells with broken spatial inversion symmetry and also act as the source of magnetic orderings. Using a newly developed protonic gate technology, gate-controlled protons intercalation could further change the carrier density and intensely tune DMI via the Ruderman–Kittel–Kasuya–Yosida mechanism. The resultant giant topological Hall resistivity $${\rho }_{{xy}}ˆ{T}$$of $$1.41{\mathrm{\mu}} \Omega \cdot {{\mathrm{cm}}}$$at $${V}_{g}=-5.2{\mathrm{V}}$$(about $$424 \%$$larger than the zero-bias value) is larger than most known chiral magnets. Theoretical analysis indicates that such a large topological Hall effect originates from the two-dimensional Bloch-type chiral spin textures stabilized by DMI, while the large anomalous Hall effect comes from the gapped Dirac nodal lines by spin–orbit interaction. Dual-intercalation in 2H-TaS2 provides a model system to reveal the nature of DMI in the large family of TMDs and a promising way of gate tuning of DMI, which further enables an electrical control of the chiral spin textures and related electromagnetic phenomena.
journaltitleNat. Commun.
titleTailoring Dzyaloshinskii–Moriya interaction in a transition metal dichalcogenide by dual-intercalation
American Physical Society
journaltitlePhys. Rev. X
titleHighly Tunable Magnetic Phases in Transition-Metal Dichalcogenide Fe_$1/3+\delta$NbS2
journaltitlePhys. Rev. B
titleStructural Disorder versus Chiral Magnetism in Cr_1/3NbS2
abstractTransition metal-intercalated transition metal dichalcogenides (TMDs) are promising platforms for next-generation spintronic devices based on their wide range of electronic and magnetic phases, which can be tuned by varying the host lattice or the identity of the intercalant, along with its stoichiometry and spatial order. Some of these compounds host a chiral magnetic phase in which the helical winding of magnetic moments propagates along a high-symmetry crystalline axis. Previous studies have demonstrated that variation in intercalant concentrations can have a dramatic impact on the formation of chiral domains and ensemble magnetic properties. However, a systematic and comprehensive study of how atomic-scale order and disorder impacts collective magnetic behavior are so far lacking. Here, we leverage a combination of imaging modes in the (scanning) transmission electron microscope (S/TEM) to directly probe (dis)order across multiple length scales and show how subtle changes in the atomic lattice can be leveraged to tune the mesoscale spin textures and bulk magnetic response, with direct implications for the fundamental understanding and technological implementation of such compounds.
notearXiv:2305.06656 [cond-mat]
titleConsequences and control of multi-scale (dis)order in chiral magnetic textures
journaltitlePhys. Rev. Mater.
titleFe and Cr co-intercalation in $2H$-NbS2 single crystals for realization of perpendicular magnetic anisotropy and large anomalous Hall effect
issn1530-6984, 1530-6992
journaltitleNano Lett.
titleTunable Single-Atomic Charges on a Cleaved Intercalated Transition Metal Dichalcogenide
abstractThe Dirac equation combines the two cornerstones of modern physics-quantum mechanics and relativity. There are several manifestations of the Dirac equation in condensed matter systems, such as the quasiparticle dispersion in graphene, topological insulators, Dirac semimetals (DSMs), Weyl semimetals, and d-wave high-temperature superconductors. In a DSM, the massless Dirac fermion has zero chirality, leading to surface states connected adiabatically to a topologically trivial surface state as well as vanishing anomalous Hall effect (AHE). Recently, it is predicted that in the nonrelativistic limit of certain antiferromagnets, there exists a type of chiral 'Dirac-like' fermion, whose dispersion manifests four-fold degenerate crossing points formed by doubly degenerate linear bands, with topologically protected Fermi arcs. Such unconventional chiral fermion, protected by a hidden SU(2) symmetry in the hierarchy of an enhanced crystallographic group, namely spin space group, is not experimentally verified yet. Here, by combining neutron diffraction, angle-resolved photoemission spectroscopy and first-principles calculations, we reveal the existence of the Fermi-arc surface states induced by chiral Dirac-like fermions in collinear antiferromagnet CoNb3S6, which caught great interest due to its surprisingly large AHE. Our transport measurements and theoretical calculations provide a scenario that large Berry curvature embedded in the chiral fermions and weak symmetry breaking are responsible for the emergent AHE. Our work evidences the existence of chiral Dirac-like fermion in CoNb3S6, paving an avenue for exploring new emergent phenomena in quantum materials with unconventional quasiparticle excitations.
notearXiv:2301.12201 [cond-mat]
titleChiral Dirac fermion in a collinear antiferromagnet
journaltitleNat. Mater.
titleWhy some interfaces cannot be sharp
journaltitleNat. Phys.
titleIn situ doping control of the surface of high-temperature superconductors
International Union of Crystallography
abstractThe new computer program SHELXT employs a novel dual-space algorithm to solve the phase problem for single-crystal reflection data expanded to the space group P1. Missing data are taken into account and the resolution extended if necessary. All space groups in the specified Laue group are tested to find which are consistent with the P1 phases. After applying the resulting origin shifts and space-group symmetry, the solutions are subject to further dual-space recycling followed by a peak search and summation of the electron density around each peak. Elements are assigned to give the best fit to the integrated peak densities and if necessary additional elements are considered. An isotropic refinement is followed for non-centrosymmetric space groups by the calculation of a Flack parameter and, if appropriate, inversion of the structure. The structure is assembled to maximize its connectivity and centred optimally in the unit cell. SHELXT has already solved many thousand structures with a high success rate, and is optimized for multiprocessor computers. It is, however, unsuitable for severely disordered and twinned structures because it is based on the assumption that the structure consists of atoms.
journaltitleActa Cryst. A
titleSHELXT – Integrated Space-Group and Crystal-Structure Determination
International Union of Crystallography
abstractThe improvements in the crystal structure refinement program SHELXL have been closely coupled with the development and increasing importance of the CIF (Crystallographic Information Framework) format for validating and archiving crystal structures. An important simplification is that now only one file in CIF format (for convenience, referred to simply as `a CIF') containing embedded reflection data and SHELXL instructions is needed for a complete structure archive; the program SHREDCIF can be used to extract the .hkl and .ins files required for further refinement with SHELXL. Recent developments in SHELXL facilitate refinement against neutron diffraction data, the treatment of H atoms, the determination of absolute structure, the input of partial structure factors and the refinement of twinned and disordered structures. SHELXL is available free to academics for the Windows, Linux and Mac OS X operating systems, and is particularly suitable for multiple-core processors.
journaltitleActa Cryst. C
titleCrystal Structure Refinement with SHELXL
titlePyARPES: An Analysis Framework for Multimodal Angle-Resolved Photoemission Spectroscopies
family=Dal Corso,
abstractQUANTUM ESPRESSO is an integrated suite of computer codes for electronic-structure calculations and materials modeling, based on density-functional theory, plane waves, and pseudopotentials (norm-conserving, ultrasoft, and projector-augmented wave). The acronym ESPRESSO stands for opEn Source Package for Research in Electronic Structure, Simulation, and Optimization. It is freely available to researchers around the world under the terms of the GNU General Public License. QUANTUM ESPRESSO builds upon newly-restructured electronic-structure codes that have been developed and tested by some of the original authors of novel electronic-structure algorithms and applied in the last twenty years by some of the leading materials modeling groups worldwide. Innovation and efficiency are still its main focus, with special attention paid to massively parallel architectures, and a great effort being devoted to user friendliness. QUANTUM ESPRESSO is evolving towards a distribution of independent and interoperable codes in the spirit of an open-source project, where researchers active in the field of electronic-structure calculations are encouraged to participate in the project by contributing their own codes or by implementing their own ideas into existing codes.
journaltitleJ. Phys.: Condens. Matter
titleQUANTUM ESPRESSO: A Modular and Open-Source Software Project for Quantum Simulations of Materials.
American Physical Society
abstractFully nonlocal two-projector norm-conserving pseudopotentials are shown to be compatible with a systematic approach to the optimization of convergence with the size of the plane-wave basis. A reformulation of the optimization is developed, including the ability to apply it to positive-energy atomic scattering states and to enforce greater continuity in the pseudopotential. The generalization of norm conservation to multiple projectors is reviewed and recast for the present purposes. Comparisons among the results of all-electron and one- and two-projector norm-conserving pseudopotential calculations of lattice constants and bulk moduli are made for a group of solids chosen to represent a variety of types of bonding and a sampling of the periodic table.
journaltitlePhys. Rev. B
titleOptimized Norm-Conserving Vanderbilt Pseudopotentials
journaltitleComput. Phys. Commun.
titleThe PseudoDojo: Training and grading a 85 element optimized norm-conserving pseudopotential table
journaltitlePhys. Rev. Lett.
titleGeneralized Gradient Approximation Made Simple
journaltitleNano Lett.
titleExcitations of Intercalated Metal Monolayers in Transition Metal Dichalcogenides
family=Dal Corso,
abstractQUANTUM ESPRESSO is an integrated suite of computer codes for electronic-structure calculations and materials modeling, based on density-functional theory, plane waves, and pseudopotentials (norm-conserving, ultrasoft, and projector-augmented wave). The acronym ESPRESSO stands for opEn Source Package for Research in Electronic Structure, Simulation, and Optimization. It is freely available to researchers around the world under the terms of the GNU General Public License. QUANTUM ESPRESSO builds upon newly-restructured electronic-structure codes that have been developed and tested by some of the original authors of novel electronic-structure algorithms and applied in the last twenty years by some of the leading materials modeling groups worldwide. Innovation and efficiency are still its main focus, with special attention paid to massively parallel architectures, and a great effort being devoted to user friendliness. QUANTUM ESPRESSO is evolving towards a distribution of independent and interoperable codes in the spirit of an open-source project, where researchers active in the field of electronic-structure calculations are encouraged to participate in the project by contributing their own codes or by implementing their own ideas into existing codes.
journaltitleJ. Phys.: Condens. Matter
titleQUANTUM ESPRESSO: A Modular and Open-Source Software Project for Quantum Simulations of Materials.
American Physical Society
abstractFully nonlocal two-projector norm-conserving pseudopotentials are shown to be compatible with a systematic approach to the optimization of convergence with the size of the plane-wave basis. A reformulation of the optimization is developed, including the ability to apply it to positive-energy atomic scattering states and to enforce greater continuity in the pseudopotential. The generalization of norm conservation to multiple projectors is reviewed and recast for the present purposes. Comparisons among the results of all-electron and one- and two-projector norm-conserving pseudopotential calculations of lattice constants and bulk moduli are made for a group of solids chosen to represent a variety of types of bonding and a sampling of the periodic table.
journaltitlePhys. Rev. B
titleOptimized Norm-Conserving Vanderbilt Pseudopotentials
journaltitleComput. Phys. Commun.
titleThe PseudoDojo: Training and grading a 85 element optimized norm-conserving pseudopotential table
journaltitlePhys. Rev. Lett.
titleGeneralized Gradient Approximation Made Simple
issn1098-0121, 1550-235X
journaltitlePhys. Rev. B
titleMagnetic Phase Transition in Single Crystals of the Chiral Helimagnet Cr_1/3NbS2
liliaxie/Zotero/storage/QP5VGB9F/Ghimire et al. - 2013 - Magnetic phase transition in single crystals of th.pdf
journaltitlePhys. Rev. B
titleOut-of-Plane Spin-Orientation Dependent Magnetotransport Properties in the Anisotropic Helimagnet Cr_1/3NbS2
American Physical Society
journaltitlePhys. Rev. B
titleDisentangling Electronic, Lattice, and Spin Dynamics in the Chiral Helimagnet Cr_1/3NbS2
American Physical Society
journaltitlePhys. Rev. B
titlePersistent Exchange Splitting in the Chiral Helimagnet Cr_1/3NbS2
journaltitlePhys. Rev. B
titleMagnetism of chromia from first-principles calculations
journaltitlePhys. Rev. B
titleEffective Coulomb interaction in transition metals from constrained random-phase approximation
journaltitlePhys. Rev. B
titleCharge density wave phase, Mottness, and ferromagnetism in monolayer $1T$-NbSe2
journaltitlePhys. Rev. Lett.
titlePhonon-Assisted Magnetic Mott-Insulating State in the Charge Density Wave Phase of Single-Layer $1T$-NbSe2
issn0031-9015, 1347-4073
journaltitleJ. Phys. Soc. Jpn.
titleMagnetic Properties of Cr_1/3NbS2
journaltitleJ. Solid State Chem.
titleOn the magnetic behavior of new $2H$-NbS2-type derivatives
issn0935-9648, 1521-4095
journaltitleAdv. Mater.
titleChiral Helimagnetism and One-Dimensional Magnetic Solitons in a Cr-Intercalated Transition Metal Dichalcogenide
issn2199-160X, 2199-160X
journaltitleAdv. Electron. Mater.
titleThe Magneto-Transport Properties of Cr_1/3TaS2 with Chiral Magnetic Solitons
issn0027-8424, 1091-6490
journaltitleProc. Natl. Acad. Sci.
titleTopological Spin/Structure Couplings in Layered Chiral Magnet Cr_1/3TaS2: The Discovery of Spiral Magnetic Superstructure
|
201071-82Nancy, France 71
László Babai
Anandam Banerjee
Raghav Kulkarni
Vipul Naik
# Evasiveness and the Distribution of Prime Numbers
L. Babai University of Chicago, Chicago, IL, USA. , A. Banerjee
Northeastern University, Boston, MA, USA. , R. Kulkarni and V. Naik
###### Abstract.
A Boolean function on $N$ variables is called _evasive_ if its decision-tree
complexity is $N$. A sequence $B_{n}$ of Boolean functions is _eventually
evasive_ if $B_{n}$ is evasive for all sufficiently large $n$.
We confirm the eventual evasiveness of several classes of monotone graph
properties under widely accepted number theoretic hypotheses. In particular we
show that Chowla’s conjecture on Dirichlet primes implies that (a) for any
graph $H$, “forbidden subgraph $H$” is eventually evasive and (b) all
nontrivial monotone properties of graphs with $\leq n^{3/2-\epsilon}$ edges
are eventually evasive. ($n$ is the number of vertices.)
While Chowla’s conjecture is not known to follow from the Extended Riemann
Hypothesis (ERH, the Riemann Hypothesis for Dirichlet’s $L$ functions), we
show (b) with the bound $O(n^{5/4-\epsilon})$ under ERH.
We also prove unconditional results: (a′) for any graph $H$, the query
complexity of “forbidden subgraph $H$” is $\binom{n}{2}-O(1)$; (b′) for some
constant $c>0$, all nontrivial monotone properties of graphs with $\leq cn\log
n+O(1)$ edges are eventually evasive.
Even these weaker, unconditional results rely on deep results from number
theory such as Vinogradov’s theorem on the Goldbach conjecture.
Our technical contribution consists in connecting the topological framework of
Kahn, Saks, and Sturtevant (1984), as further developed by Chakrabarti, Khot,
and Shi (2002), with a deeper analysis of the orbital structure of permutation
groups and their connection to the distribution of prime numbers. Our
unconditional results include stronger versions and generalizations of some
result of Chakrabarti et al.
###### Key words and phrases:
Decision tree complexity, evasiveness, graph property, group action, Dirichlet
primes, Extended Riemann Hypothesis
###### 1991 Mathematics Subject Classification:
F.2.2, F.1.1, F.1.3
22footnotetext: Partially supported by NSF Grant CCF-0830370.
## 1\. Introduction
### 1.1. The framework
A graph property $P_{n}$ of $n$-vertex graphs is a collection of graphs on the
vertex set $[n]=\\{1,\dots,n\\}$ that is invariant under relabeling of the
vertices. A property $P_{n}$ is called monotone (decreasing) if it is
preserved under the deletion of edges. The trivial graph properties are the
empty set and the set of all graphs. A class of examples are the forbidden
subgraph properties: for a fixed graph $H$, let $Q_{n}^{H}$ denote the class
of $n$-vertex graphs that do not contain a (not necessarily induced) subgraph
isomorphic to $H$.
We view a set of labeled graphs on $n$ vertices as a Boolean function on the
$N=\binom{n}{2}$ variables describing adjacency. A Boolean function on $N$
variables is evasive if its deterministic query (decision-tree) complexity is
$N$.
The long-standing Aanderaa-Rosenberg-Karp conjecture asserts that every
nontrivial monotone graph property is evasive. The problem remains open even
for important special classes of monotone properties, such as the forbidden
subgraph properties.
### 1.2. History
In this note, $n$ always denotes the number of vertices of the graphs under
consideration.
Aanderaa and Rosenberg (1973) [17] conjectured a lower bound of
$\Omega(n^{2})$ on the query complexity of monotone graph properties. Rivest
and Vuillemin (1976) [19] verified this conjecture, proving an $n^{2}/16$
lower bound. Kleitman and Kwiatkowski (1980) [10] improved this to $n^{2}/9.$
Karp conjectured that nontrivial monotone graph properties were in fact
evasive. We refer to this statement as the Aanderaa-Rosenberg-Karp (ARK)
conjecture.
In their seminal paper, Kahn, Saks, and Sturtevant [11] observe that non-
evasiveness of monotone Boolean functions has strong topological consequences
(contracibility of the associated simplicial complex). They then use results
of R. Oliver about fixed points of group actions on such complexes to verify
the ARK conjecture when $n$ is a prime-power. As a by-product, they improve
the lower bound for general $n$ to $n^{2}/4.$
Since then, the topological approach of [11] has been influential in solving
various interesting special cases of the ARK conjecture. Yao (1988) [25]
proves that non-trivial monotone properties of bipartite graphs with a given
partition $(U,V)$ are evasive (require $|U||V|$ queries). Triesch (1996) [22]
shows (in the original model) that any monotone property of bipartite graphs
(all the graphs satisfying the property are bipartite) is evasive.
Chakrabarti, Khot, and Shi (2002) [3] introduce important new techniques which
we use; we improve over several of their results (see Section 1.4).
### 1.3. Prime numbers in arithmetic progressions
Dirichlet’s Theorem (1837) (cf. [5]) asserts that if $\gcd(a,m)=1$ then there
exist infinitely many primes $p\equiv a\pmod{m}$. Let $p(m,a)$ denote the
smallest such prime $p$. Let $p(m)=\max\\{p(m,a)\mid\gcd(a,m)=1\\}$. Linnik’s
celebrated theorem (1947) asserts that $p(m)=O(m^{L})$ for some absolute
constant $L$ (cf. [16, Chap. V.]). Heath-Brown [9] shows that $L\leq 5.5$.
Chowla [4] observes that under the Extended Riemann Hypothesis (ERH) we have
$L\leq 2+\epsilon$ for all $\epsilon>0$ and conjectures that $L\leq
1+\epsilon$ suffices:
###### Conjecture 1.1 (S. Chowla [4]).
For every $\epsilon>0$ and every $m$ we have $p(m)=O(m^{1+\epsilon})$.
This conjecture is widely believed; in fact, number theorists suggest as
plausible the stronger form $p(m)=O(m(\log m)^{2})$ [8]. Turán [23] proves the
tantalizing result that for almost all $a$ we have $p(m,a)=O(m\log m)$ .
Let us call a prime $p$ an $\epsilon$-near Fermat prime if there exists an
$s\geq 0$ such that $2^{s}\mid p-1$ and $\frac{p-1}{2^{s}}\leq p^{\epsilon}$.
We need the following weak form of Chowla’s conjecture:
###### Conjecture 1.2 (Weak Chowla Conjecture).
For every $\epsilon>0$ there exist infinitely many $\epsilon$-near Fermat
primes.
In other words, the weak conjecture says that for every $\epsilon$, for
infinitely many values of $s$ we have $p(2^{s},1)<(2^{s})^{1+\epsilon}$.
### 1.4. Main results
For a graph property $P$ we use $P_{n}$ to denote the set of graphs on vertex
set $[n]$ with property $P$. We say that $P$ is eventually evasive if $P_{n}$
is evasive for all sufficiently large $n$.
Our first set of results states that the “forbidden subgraph” property is
“almost evasive” under three different interpretations of this phrase.
###### Theorem 1.3 (Forbidden subgraphs).
For all graphs $H$, the forbidden subgraph property $Q_{n}^{H}$ (a) is
eventually evasive, assuming the Weak Chowla Conjecture; (b) is evasive for
almost all $n$ (unconditionally); and (c) has query complexity
$\binom{n}{2}-O(1)$ for all $n$ (unconditionally).
Part (b) says the asymptotic density of values of $n$ for which the problem is
not evasive is zero. Part (c) improves the bound $\binom{n}{2}-O(n)$ given in
[3]. Parts (a) and (c) will be proved in Section 3. We defer the proof of part
(b) to the journal version.
The term “monotone property of graphs with $\leq m$ edges” describes a
monotone property that fails for all graphs with more than $m$ edges.
###### Theorem 1.4 (Sparse graphs).
All nontrivial monotone properties of graphs with at most $f(n)$ edges are
eventually evasive, where (a) under Chowla’s Conjecture,
$f(n)=n^{3/2-\epsilon}$ for any $\epsilon>0$; (b) under ERH,
$f(n)=n^{5/4-\epsilon}$; and (c) unconditionally, $f(n)=cn\log n$ for some
constant $c>0$. (d) Unconditionally, all nontrivial monotone properties of
graphs with no cycle of length greater than $(n/4)(1-\epsilon)$ are eventually
evasive (for all $\epsilon>0$).
Part (c) of Theorem 1.4 will be proved in Section 4. Parts (a) and (b) follow
in Section 5. The proof of part (d) follows along the lines of part (c); we
defer the details to the journal version of this paper.
We note that the proofs of the unconditional results (c) and (d) in Theorem
1.4 rely on Haselgrove’s version [7] of Vinogradov’s Theorem on Goldbach’s
Conjecture (cf. Sec. 4.2).
Recall that a _topological subgraph_ of a graph $G$ is obtained by taking a
subgraph and replacing any induced path $u-\dots-v$ in the subgraph by an edge
$\\{u,v\\}$ (repeatedly) and deleting parallel edges. A minor of a graph is
obtained by taking a subgraph and contracting edges (repeatedly). If a class
of graphs is closed under taking minors then it is also closed under taking
topological subgraphs but not conversely; for instance, graphs with maximum
degree $\leq 3$ are closed under taking toopological subgraphs but every graph
is a minor of a regular graph of degree 3.
###### Corollary 1.5 (Excluded topological subgraphs).
Let $P$ be a nontrivial class of graphs closed under taking topological
subgraphs. Then $P$ is eventually evasive.
This unconditional result extends one of the results of Chakrabarti et al.
[3], namely, that nontrival classes of graphs closed under taking minors is
eventually evasive.
Corollary 1.5 follows from part (c) of Theorem 1.4 in the light of Mader’s
Theorem which states that if the average degree of a graph $G$ is greater than
$2^{\binom{k+1}{2}}$ then it contains a topological $K_{k}$ [13, 14].
Theorem 1.4 suggests a new stratification of the ARK Conjecture. For a
monotone (decreasing) graph property $P_{n}$, let
$\dim(P_{n}):=\max\\{|E(G)|-1\ |\ G\in P_{n}\\}.$
We can now restate the ARK Conjecture:
###### Conjecture 1.6.
If $P_{n}$ is a non-evasive, non-empty, monotone decreasing graph property
then $\dim(P_{n})=\binom{n}{2}-1.$
## 2\. Preliminaries
### 2.1. Group action
For the basics of group theory we refer to [18]. All groups in this paper are
finite. For groups $\Gamma_{1},\Gamma_{2}$ we use $\Gamma_{1}\leq\Gamma_{2}$
to denote that $\Gamma_{1}$ is a subgroup; and $\Gamma_{1}\lhd\Gamma_{2}$ to
denote that $\Gamma_{1}$ is a (not necessarily proper) normal subgroup. We say
that $\Gamma$ is a $p$-group if $|\Gamma|$ is a power of the prime $p$.
For a set $\Omega$ called the “permutation domain,” let
$\operatorname{Sym}(\Omega)$ denote the symmetric group on $\Omega$,
consisting of the $|\Omega|!$ permutations of $\Omega$. For
$\Omega=[n]=\\{1,\dots,n\\}$, we set $\Sigma_{n}=\operatorname{Sym}([n])$. For
a group $\Gamma$, a homomorphism
$\varphi\,:\,\Gamma\to\operatorname{Sym}(\Omega)$ is called a $\Gamma$-action
on $\Omega$. The action is faithful if $\ker(\varphi)=\\{1\\}$. For
$x\in\Omega$ and $\gamma\in\Gamma$ we denote by $x^{\gamma}$ the image of $x$
under $\varphi(\gamma)$. For $x\in\Omega$ we write
$x^{\Gamma}=\\{x^{\gamma}\,:\,\gamma\in\Gamma\\}$ and call it the orbit of $x$
under the $\Gamma$-action. The orbits partition $\Omega$.
Let $\binom{\Omega}{t}$ denote the set of $t$-subsets of $\Omega$. There is a
natural induced action
$\operatorname{Sym}(\Omega)\to\operatorname{Sym}(\binom{\Omega}{t})$ which
also defines a natural $\Gamma$-action on $\binom{\Omega}{t}$. We denote this
action by $\Gamma^{(t)}$. Similarly, there is a natural induced
$\Gamma$-action on $\Omega\times\Omega$. The orbits of this action are called
the orbitals of $\Gamma$. We shall need the undirected version of this
concept; we shall call the orbits of the $\Gamma$-action on
$\binom{\Omega}{2}$ the u-orbitals (undirected orbitals) of the
$\Gamma$-action.
By an action of the group $\Gamma$ on a structure $\mathfrak{X}$ such as a
group or a graph or a simplicial complex we mean a homomorphism
$\Gamma\to\operatorname{Aut}({\mathfrak{X}})$ where
$\operatorname{Aut}({\mathfrak{X}})$ denotes the automorphism group of
$\mathfrak{X}$.
Let $\Gamma$ and $\Delta$ be groups and let
$\psi\,:\,\Delta\to\operatorname{Aut}(\Gamma)$ be a $\Delta$-action on
$\Gamma$. These data uniquely define a group $\Theta=\Gamma\rtimes\Delta$, the
_semidirect product_ of $\Gamma$ and $\Delta$ with respect to $\psi$. This
group has order $|\Theta|=|\Gamma||\Delta|$ and has the following properites:
$\Theta$ has two subgroups $\Gamma^{*}\cong\Gamma$ and $\Delta^{*}\cong\Delta$
such that $\Gamma^{*}\lhd\Theta$; $\Gamma^{*}\cap\Delta^{*}=\\{1\\}$; and
$\Theta=\Gamma^{*}\Delta^{*}=\\{\gamma\delta\mid\gamma\in\Gamma^{*},\delta\in\Delta^{*}\\}$.
Moreover, identifying $\Gamma$ with $\Gamma^{*}$ and $\Delta$ with
$\Delta^{*}$, for all $\gamma\in\Gamma$ and $\delta\in\Delta$ we have
$\gamma^{\psi(\delta)}=\delta^{-1}\gamma\delta$.
$\Theta$ can be defined as the set $\Delta\times\Gamma$ under the group
operation
$(\delta_{1},\gamma_{1})(\delta_{2},\gamma_{2})=(\delta_{1}\delta_{2},\gamma_{1}^{\psi(\delta_{2})}\gamma_{2})\quad\quad(\delta_{i}\in\Delta,\gamma_{i}\in\Gamma).$
For more on semidirect products, which we use extensively, see [18, Chap. 7].
The group $\operatorname{AGL}(1,q)$ of affine transformations $x\mapsto ax+b$
of $\mathbb{F}_{q}$ ($a\in\mathbb{F}_{q}^{\times}$, $b\in\mathbb{F}_{q}$) acts
on $\mathbb{F}_{q}$. For each $d\mid q-1$, $\operatorname{AGL}(1,q)$ has a
unique subgroup of order $qd$; we call this subgroup $\Gamma(q,d)$. We note
that $\mathbb{F}_{q}^{+}\lhd\Gamma(q,d)$ and $\Gamma(q,d)/\mathbb{F}_{q}^{+}$
is cyclic of order $d$ and is isomorphic to a subgroup $\Delta$ of
$\operatorname{AGL}(1,q)$; $\Gamma(q,d)$ can be described as a semidirect
product $(\mathbb{F}_{q}^{+})\rtimes\Delta$.
### 2.2. Simplicial complexes and monotone graph properties
An abstract simplicial complex ${\mathcal{K}}$ on the set $\Omega$ is a subset
of the power-set of $\Omega$, closed under subsets: if $B\subset
A\in{{\mathcal{K}}}$ then $B\in{{\mathcal{K}}}$. The elements of
${\mathcal{K}}$ are called its faces. The dimension of $A\in{{\mathcal{K}}}$
is $\dim(A)=|A|-1$; the dimension of ${\mathcal{K}}$ is
$\dim({{\mathcal{K}}})=\max\\{\dim(A)\mid A\in{{\mathcal{K}}}\\}$. The Euler
characteristic of ${{\mathcal{K}}}$ is defined as
$\chi({{\mathcal{K}}}):=\sum_{A\in{{\mathcal{K}}},A\neq\emptyset}{(-1)^{\dim(A)}}.$
Let $[n]:=\\{1,2,\ldots,n\\}$ and $\Omega=\binom{[n]}{2}$. Let $P_{n}$ be a
subset of the power-set of $\Omega$, i. e., a set of graphs on the vertex set
$[n]$. We call $P_{n}$ a graph property if it is invariant under the induced
action $\Sigma_{n}^{(2)}$. We call this graph property monotone decreasing if
it is closed under subgraphs, i. e., it is a simplicial complex. We shall omit
the adjective “decreasing.”
### 2.3. Oliver’s Fixed Point Theorem
Let ${\mathcal{K}}\subseteq 2^{\Omega}$ be an abstract simplicial complex with
a $\Gamma$-action. The fixed point complex ${\mathcal{K}}_{\Gamma}$ action is
defined as follows. Let $\Omega_{1},\dots,\Omega_{k}$ be the $\Gamma$-orbits
on $\Omega$. Set
${\mathcal{K}}_{\Gamma}:=\\{S\subseteq[k]\mid\bigcup_{i\in
S}\Omega_{i}\in{\mathcal{K}}\\}.$
We say that a group $\Gamma$ satisfies Oliver’s condition if there exist (not
necessarily distinct) primes $p,q$ such that $\Gamma$ has a (not necessarily
proper) chain of subgroups $\Gamma_{2}\lhd\Gamma_{1}\lhd\Gamma$ such that
$\Gamma_{2}$ is a $p$-group, $\Gamma_{1}/\Gamma_{2}$ is cyclic, and
$\Gamma/\Gamma_{1}$ is a $q$-group.
###### Theorem 2.1 (Oliver [15]).
Assume the group $\Gamma$ satisfies Oliver’s condition. If $\Gamma$ acts on a
nonempty contractible simplicial complex ${\mathcal{K}}$ then
$\chi({\mathcal{K}}_{\Gamma})\equiv 1\pmod{q}.$ (2.1)
In particular, such an action must always have a nonempty invariant face.
### 2.4. The KSS approach and the general strategy
The topological approach to evasiveness, initiated by Kahn, Saks, and
Sturtevant, is based on the following key observation.
###### Lemma 2.2 (Kahn-Saks-Sturtevant [11]).
If $P_{n}$ is a non-evasive graph property then $P_{n}$ is contractible.
Kahn, Saks, and Sturtevant recognized that Lemma 2.2 brought Oliver’s Theorem
to bear on evasiveness. The combination of Lemma 2.2 and Theorem 2.1 suggests
the following general strategy, used by all authors in the area who have
employed the topological method, including this paper: We find primes $p,q$, a
group $\Gamma$ satisfying Oliver’s condition with these primes, and a
$\Gamma$-action on $P_{n}$, such that $\chi(P_{n})\equiv 0\pmod{q}$. By
Oliver’s Theorem and the KSS Lemma this implies that $P_{n}$ is evasive. The
novelty is in finding the right $\Gamma$.
KSS [11] made the assumption that $n$ is a prime power and used as
$\Gamma=\operatorname{AGL}(1,n)$, the group of affine transformations
$x\mapsto ax+b$ over the field of order $n$. While we use subgroups of such
groups as our building blocks, the attempt to combine these leads to hard
problems on the distribution of prime numbers.
Regarding the “forbidden subgraph” property, Chakrabarti, Khot, and Shi [3]
built considerable machinery which we use. Our conclusions are considerably
stronger than theirs; the additional techniques involved include a study of
the orbitals of certain metacyclic groups, a universality property of
cyclotomic graphs derivable using Weil’s character sum estimates, plus the
number theoretic reductions indicated.
For the “sparse graphs” result (Theorem 1.4) we need $\Gamma$ such that all
u-orbitals of $\Gamma$ are large and therefore
$(P_{n})_{\Gamma}=\\{\emptyset\\}$.
In both cases, we are forced to use rather large building blocks of size $q$,
say, where $q$ is a prime such that $q-1$ has a large divisor which is a prime
for Theorem 1.4 and a power of 2 for Theorem 1.3.
## 3\. Forbidden subgraphs
In this section we prove parts (a) and (c) of Theorem 1.3.
### 3.1. The CKS condition
A homomorphism of a graph $H$ to a graph $H^{\prime}$ is a map $f\,:\,V(H)\to
V(H^{\prime})$ such that $(\forall x,y\in V(H))(\\{x,y\\}\in
E(H)\Rightarrow\\{f(x),f(y)\\}\in E(H^{\prime}))$. (In particular,
$f^{-1}(x^{\prime})$ is an independent set in $H$ for all $x^{\prime}\in
V(H^{\prime})$.) Let $Q_{r}^{[[H]]}$ be the set of those $H^{\prime}$ with
$V(H^{\prime})=[r]$ that do not admit an $H\to H^{\prime}$ homomorphism. Let
further $T_{H}:=\min\\{2^{2^{t}}-1\ \mid\ 2^{2^{t}}\geq h\\}$ where $h$
denotes the number of vertices of $H$. The following is the main lemma of
Chakrabarti, Khot, and Shi [3].
###### Lemma 3.1 (Chakrabarti et al. [3]).
If $r\equiv 1\pmod{T_{H}}$ then $\chi(Q_{r}^{[[H]]})\equiv 0\pmod{2}$.
### 3.2. Cliques in generalized Paley graphs
Let $q$ be an odd prime power and $d$ an even divisor of $q-1.$ Consider the
graph $P(q,d)$ whose vertex set is $\mathbb{F}_{q}$ and the adjacency between
the vertices is defined as follows: $i\sim j\iff(i-j)^{d}=1.$ $P(q,d)$ is
called a generalized Paley graph.
###### Lemma 3.2.
If $(q-1)/d\leq q^{1/(2h)}$ then $P(q,d)$ contains a clique on $h$ vertices.
This follows from the following lemma which in turn can be proved by a routine
application of Weil’s character sum estimates (cf. [1]).
###### Lemma 3.3.
Let $a_{1},\ldots,a_{t}$ be distinct elements of the finite field
$\mathbb{F}_{q}.$ Assume $\ell\mid q-1$. Then the number of solutions
$x\in\mathbb{F}_{q}$ to the system of equations $(a_{i}+x)^{(q-1)/\ell}=1$ is
$\frac{q}{\ell^{t}}\pm t\sqrt{q}.$ ∎
Let $\Gamma(q,d)$ be the subgroup of order $qd$ of $\operatorname{AGL}(1,q)$
defined in Section 2.1. Each u-orbital of $\Gamma(q,d)$ is isomorphic to
$P(q,d)$. ∎
###### Corollary 3.4.
If ${\frac{q-1}{d}\leq q^{1/(2h)}}$ then each u-orbital of $\Gamma(q,d)$
contains a clique of size $h.$
### 3.3. $\epsilon$-near-Fermat primes
The numbers in the title were defined in Section 1.3. In this section we prove
Theorem 1.3, part (a).
###### Theorem 3.5.
Let $H$ be a graph on $h$ vertices. If there are infinitely many
$\frac{1}{2h}$-near-Fermat primes then $Q_{n}^{H}$ is eventually evasive.
Proof. Fix an odd prime $p\equiv 2\pmod{T_{H}}$ such that $p\geq|H|.$ If there
are infinitely many $\frac{1}{2h}$-near-Fermat primes then infinitely many of
them belong to the same residue class mod $p$, say $a+\mathbb{Z}p$. Let
$q_{i}$ be the $i$-th $\frac{1}{2h}$-near-Fermat prime such that $q_{i}\geq p$
and $q_{i}\equiv a\pmod{p}.$ Let $r^{\prime}=na^{-1}\pmod{p}$ and
$k^{\prime}=\sum_{i=1}^{r^{\prime}}q_{i}.$ Then $k^{\prime}\equiv n\pmod{p}$
and therefore $n=pk+k^{\prime}$ for some $k$.
Now in order to use Lemma 3.1, we need to write $n$ as a sum of $r$ terms
where $r\equiv 1\pmod{T_{H}}$. We already have $r^{\prime}$ of these terms; we
shall choose each of the remaining $r-r^{\prime}$ terms to be $p$ or $p^{2}$.
If there are $t$ terms equal to $p^{2}$ then this gives us a total of
$r=t+(k-tp)+r^{\prime}$ terms, so we need $t(p-1)\equiv
k+r^{\prime}\pmod{T_{H}}$. By assumption, $p-1\equiv 1\pmod{T_{H}}$; therefore
such a $t$ exists; for large enough $n$, it will also satisfy the constraints
$0\leq t\leq k/p$,
Let now
$\Lambda_{1}:=\left((\mathbb{F}^{+}_{p^{2}})^{t}\times(\mathbb{F}^{+}_{p})^{k-tp}\right)\rtimes\mathbb{F}_{p^{2}}^{\times}$
acting on $[pk]$ with $t$ orbits of size $p^{2}$ and $k-pt$ orbits of size $p$
as follows: on an orbit of size $p^{i}$ ($i=1,2$) the action is
$\operatorname{AGL}(1,p^{i})$. The additive groups act independently, with a
single multiplicative action on top. $\mathbb{F}_{p^{2}}^{\times}$ acts on
$\mathbb{F}_{p}^{+}$ through the group homomorphism
$\mathbb{F}_{p^{2}}^{\times}\to\mathbb{F}_{p}^{\times}$ defined by the map
$x\mapsto x^{p-1}$. Let $B_{j}$ denote an orbit of $\Lambda_{1}$ on $[kp]$.
Now the orbit of any pair $\\{u,v\\}\in{B_{j}\choose 2}$ is a clique of size
$|B_{j}|\geq p\geq h$, therefore a $\Lambda_{1}$-invariant graph cannot
contain an intra-cluster edge.
Let $d_{i}$ be the largest power of 2 that divides $q_{i}-1.$ Let $C_{i}$ be
the subgroup of $\mathbb{F}_{q_{i}}^{\times}$ of order $d_{i}.$ Let
$\displaystyle{\Lambda_{2}:=\prod_{i=1}^{r^{\prime}}\Gamma(q_{i},d_{i}),}$
acting on $[k^{\prime}]$ with $r^{\prime}$ orbits of sizes
$q_{1},\dots,q_{r^{\prime}}$ in the obvious manner.
From Lemma 3.2 we know that the orbit of any $\\{u,v\\}\in{[q_{i}]\choose 2}$
must contain a clique of size $h.$ Hence, an invariant graph cannot contain
any intra-cluster edge.
Overall, let $\Gamma:=\Lambda_{1}\times\Lambda_{2}$, acting on $[n].$ Since
$q_{i}\geq p,$ we have $\gcd(q_{i},p^{2}-1)=1.$ Thus, $\Gamma$ is a “$2$-group
extension of a cyclic extension of a $p$-group” and therefore satisfies
Oliver’s Condition (stated before Theorem 2.1). Hence, assuming $Q_{n}^{H}$ is
non-evasive, Lemma 2.2 and Theorem 2.1 imply
$\chi((Q_{n}^{H})_{\Gamma})\equiv 1\pmod{2}.$
On the other hand, we claim that the fixed-point complex
$(Q_{n}^{H})_{\Gamma}$ is isomorphic to $Q_{r}^{[[H]]}$. The (simple) proof
goes along the lines of Lemma 4.2 of [3]. Thus, by Lemma 3.1 we have
$\chi(Q_{r}^{[[H]]})\equiv 0\pmod{2},$ a contradiction. ∎
### 3.4. Unconditionally, $Q_{n}^{H}$ is only $O(1)$ away from being evasive
In this section, we prove part (c) of Theorem 1.3.
###### Theorem 3.6.
For every graph $H$ there exists a number $C_{H}$ such that the query
complexity of $Q_{n}^{H}$ is $\geq\binom{n}{2}-C_{H}.$
Proof. Let $h$ be the number of vertices of $H$. Let $p$ be the smallest prime
such that $p\geq h$ and $p\equiv 2\pmod{T_{H}}$. So $p<f(H)$ for some function
$f$ by Dirichlet’s Theorem (we don’t need any specific estimates here). Since
$p-1\equiv 1\pmod{T_{H}},$ we have $\gcd(p-1,T_{H})=1$ and therefore
$\gcd(p-1,pT_{H})=1$. Now, by the Chinese Remainder Theorem, select the
smallest positive integer $k^{\prime}$ satisfying $k^{\prime}\equiv
n\pmod{pT_{H}}$ and $k^{\prime}\equiv 1\pmod{p-1}$. Note that
$k^{\prime}<p^{2}T_{H}$. Let $k=(n-k^{\prime})/(pT_{H})$; so we have
$n=kpT_{H}+k^{\prime}$.
Let $N^{\prime}=\binom{n}{2}-\binom{k^{\prime}}{2}$. Consider the following
Boolean function $B_{n}^{H}$ on $N^{\prime}$ variables. Consider graphs $X$ on
the vertex set $[n]$ with the property that they have no edges among their
last $k^{\prime}$ vertices. These graphs can be viewed as Boolean functions of
the remaining $N^{\prime}$ variables. Now we say that such a graph has
property $B_{n}^{H}$ if it does not contain $H$ as a subgraph.
Claim. The function $B_{n}^{H}$ is evasive.
The Claim immediately implies that the query complexity of $Q_{n}^{H}$ is at
least $N^{\prime}$, proving the Theorem with
$C_{H}=\binom{k^{\prime}}{2}<p^{4}T_{H}^{2}<f(H)^{4}T_{H}^{2}$.
To prove the Claim, consider the groups
$\Lambda:=(\mathbb{F}_{p}^{+})^{kT_{H}}\rtimes\mathbb{F}_{p}^{\times}$ and
$\Gamma:=\Lambda\times\mathbb{Z}_{k^{\prime}}$. Here $\Lambda$ acts on
$[pkT_{H}]$ in the obvious way: we divide $[pkT_{H}]$ into $kT_{H}$ blocks of
size $p$; $\mathbb{F}_{p}^{+}$ acts on each block independently and
$\mathbb{F}_{p}^{\times}$ acts on the blocks simultaneously (diagonal action)
so on each block they combine to an $\operatorname{AGL}(1,p)$-action.
$\mathbb{Z}_{k^{\prime}}$ acts as a $k^{\prime}$-cycle on the remaining
$k^{\prime}$ vertices. So $\Gamma$ is a cyclic extension of a $p$-group
(because $\gcd(p-1,k^{\prime})=1$).
If $B_{n}^{H}$ is not evasive then from Theorem 2.1 and Lemma 2.2, we have
$\chi\left((B_{n}^{H})_{\Gamma}\right)=1$.
On the other hand we claim that, $(B_{n}^{H})_{\Gamma}\cong Q_{r}^{[[H]]},$
where $r=kT_{H}+1.$ The proof of this claim is exactly the same as the proof
of Lemma 4.2 of [3]. Thus, from Lemma 3.1, we conclude that
$\chi(Q_{r}^{[[H]]})$ is even. This contradicts the previous conclusion that
$\chi(Q_{r}^{[[H]]})=1.$ ∎
###### Remark 3.7.
Specific estimates on the smallest Dirichlet prime can be used to estimate
$C_{H}$. Linnik’s theorem implies $C_{H}<h^{O(1)}$, extending Theorem 3.6 to
strong lower bounds for variable $H$ up to $h=n^{c}$ for some positive
constant $c$.
## 4\. Sparse graphs: unconditional results
We prove part (c) of Theorem 1.4.
###### Theorem 4.1.
If the non-empty monotone graph property $P_{n}$ is not evasive then
$\dim(P_{n})=\Omega(n\log n).$
### 4.1. The basic group construction
Assume in this section that $n=p^{\alpha}k$ where $p$ is prime. Let
$\Delta_{k}\leq\Sigma_{k}$. We construct the group
$\Gamma_{0}(p^{\alpha},\Delta_{k})$ acting on $[n].$
Let $\Delta=(\mathbb{F}_{p^{\alpha}}^{\times}\times\Delta_{k})$. Let
$\Gamma_{0}(p^{\alpha},\Delta_{k})$ be the semidirect product
$(\mathbb{F}_{p^{\alpha}}^{+})^{k}\rtimes\Delta$ with respect to the
$\Delta$-action on $(\mathbb{F}_{p^{\alpha}}^{+})^{k}$ defined by
$(a,\sigma):(b_{1},\ldots,b_{k})\mapsto(ab_{\sigma^{-1}(1)},\ldots,ab_{\sigma^{-1}(k)}).$
We describe the action of $\Gamma_{0}(p^{\alpha},\Delta_{k})$ on $[n]$.
Partition $[n]$ into $k$ clusters of size $p^{\alpha}$ each. Identify each
cluster with the field of order $p^{\alpha},$ i.e., as a set,
$[n]=[k]\times\mathbb{F}_{p^{\alpha}}.$ The action of
$\gamma=(b_{1},\ldots,b_{k},a,\sigma)$ is described by
$\gamma:(x,y)\mapsto(\sigma(x),ay+b_{\sigma(x)}).$
An unordered pair $(i,j)\in[n]$ is termed an intra-cluster edge if both $i$
and $j$ are in the same cluster, otherwise it is termed an inter-cluster edge.
Note that every u-orbital under $\Gamma$ has only intra-cluster edges or only
inter-cluster edges. Denote by $m_{\operatorname{intra}}$ and
$m_{\operatorname{inter}}$ the minimum sizes of u-orbitals of intra-cluster
and inter-cluster edges respectively.
We denote by $m^{\prime}_{k}$ the minimum size of an orbit in $[k]$ under
$\Delta_{k}$ and by $m^{\prime\prime}_{k}$ the minimum size of a u-orbital in
$[k].$ We then have:
$m_{\operatorname{intra}}\geq\binom{p^{\alpha}}{2}\times m^{\prime}_{k},\qquad
m_{\operatorname{inter}}\geq(p^{\alpha})^{2}\times m^{\prime\prime}_{k}$
Let $m_{k}`:=\min\\{m^{\prime}_{k},m^{\prime\prime}_{k}\\}$ and define $m^{*}$
as the minimum size of a u-orbital in $[n].$ Then
$m^{*}=\min\\{m_{\operatorname{intra}},m_{\operatorname{inter}}\\}=\Omega(p^{2\alpha}m_{k})$
(4.1)
### 4.2. Vinogradov’s Theorem
The Goldbach Conjecture asserts that every even integer can be written as the
sum of two primes. Vinogradov’s Theorem [24] says that every sufficiently
large odd integer $k$ is the sum of three primes $k=p_{1}+p_{2}+p_{3}$. We use
here Haselgrove’s version [7] of Vinogradov’s theorem which states that we can
require the primes to be roughly equal: $p_{i}\sim k/3$. This can be combined
with the Prime Number Theorem to conclude that every sufficiently large even
integer $k$ is a sum of four roughly equal primes.
### 4.3. Construction of the group
Let $n=p^{\alpha}k$ where $p$ is prime. Assume $k$ is not bounded. Write $k$
as a sum of $t\leq 4$ roughly equal primes $p_{i}$. Let
$\Delta_{k}:=\prod_{i}\mathbb{Z}_{p_{i}}$ where $\mathbb{Z}_{p_{i}}$ denotes
the cyclic group of order $p_{i}$ and the direct product is taken over the
distinct $p_{i}$.
$\Delta_{k}$ acts on $[k]$ as follows: partition $k$ into parts of sizes
$p_{1},\dots,p_{t}$ and call these parts $[p_{i}].$ The group
$\mathbb{Z}_{p_{i}}$ acts as a cyclic group on the part $[p_{i}].$ In case of
repetitions, the same factor $\mathbb{Z}_{p_{i}}$ acts on all the parts of
size $p_{i}.$
We follow the notation of Section 4.1 and consider the group
$\Gamma_{0}(p^{\alpha},\Delta_{k})$ with our specific $\Delta_{k}$. We have
$m_{k}=\Omega(k)$ and hence we get, from equation (4.1):
###### Lemma 4.2.
Let $n=p^{\alpha}k$ where $p$ is a prime. For the group
$\Gamma_{0}(p^{\alpha},\Delta_{k})$, we have
$m^{*}=\Omega(p^{2\alpha}k)=\Omega(p^{\alpha}n),$ where $m^{*}$ denotes the
minimum size of a u-orbital.
### 4.4. Proof for the superlinear bound
Let $n=p^{\alpha}k$ where $p^{\alpha}$ is the largest prime power dividing
$n$; so $p^{\alpha}=\Omega(\log n)$; this will be a lower bound on the size of
u-orbitals. Our group $\Gamma$ will be of the general form discussed in
Section 4.1.
Case 1. $p^{\alpha}=\Omega(n^{2/3}).$
Let $\Gamma=\Gamma_{0}(p^{\alpha},\\{1\\})$. Following the notation of Section
4.1, we get $m_{k}^{\prime}=m_{k}^{\prime\prime}=1,$ and this yields that
$m^{*}=\Omega((p^{\alpha})^{2})=\Omega(n^{4/3})=\Omega(n\log n).$ Oliver’s
condition is easily verified for $\Gamma$.
Case 2. $k=\Omega(n^{1/3}).$
Consider the $\Gamma:=\Gamma_{0}(p^{\alpha},\Delta_{k})$ acting on $[n]$ where
$\Delta_{k}$ is as described in Section 4.3. The minimum possible size $m^{*}$
of a u-orbital is $\Omega(np^{\alpha})$ by Lemma 4.2. Finally, since
$p^{\alpha}=\Omega(\log n)$, we obtain $m^{*}=\Omega(n\log n).$
If all $p_{i}$ are co-prime to $p^{\alpha}-1$ then
$\mathbb{F}_{p^{\alpha}}^{\times}\times\Delta_{k}$ becomes a cyclic group and
$\Gamma$ becomes a cyclic extension of a $p$-group.
Since $p_{i}=\Omega(k)=\Omega(n^{1/3})$ for all $i$ and
$p^{\alpha}=O(n^{2/3})$, size considerations yield that at most one $p_{i}$
divides $p^{\alpha}-1$ and $p_{i}^{2}$ does not. Suppose, without loss of
generality, $p_{1}$ divides $p^{\alpha}-1.$ Let $p^{\alpha}-1=p_{1}d,$ then
$d$ must be co-prime to each $p_{i}.$ Thus,
$\Delta=(\mathbb{Z}_{p_{1}}\times\mathbb{Z}_{d})\times(\mathbb{Z}_{p_{1}}\times\ldots\times\mathbb{Z}_{p_{t}})=(\mathbb{Z}_{d}\times\mathbb{Z}_{p_{2}}\times\ldots\times\mathbb{Z}_{p_{r}})\times(\mathbb{Z}_{p_{1}}\times\mathbb{Z}_{p_{1}}).$
Thus, $\Delta$ is a $p_{1}$-group extension of a cyclic group. Hence, $\Gamma$
satisfies Oliver’s Condition (cf. Theorem 2.1). ∎
###### Remark 4.3.
For almost all $n,$ our proof gives a better dimension lower bound of
$\Omega(n^{1+\frac{1+o(1)}{\ln\ln n}}).$
## 5\. Sparse graphs: conditional improvements
In this section we prove parts (a) and (b) of Theorem 1.4.
### 5.1. General Setup
Let $n=pk+r,$ where $p$ and $r$ are prime numbers. Let $q$ be a prime divisor
of $(r-1).$ We partition $[n]$ into two parts of size $pk$ and $r$, denoted by
$[pk]$ and $[r]$ respectively. We now construct a group $\Gamma(p,q,r)$ acting
on $[n]$ as a direct product of a group acting on $[pk]$ and a group acting on
$[r],$ as follows:
$\Gamma=\Gamma(p,q,r):=\Gamma_{0}(p,\Delta_{k})\times\Gamma(r,q)$
Here, $\Gamma_{0}(p,\Delta_{k})$ acts on $[pk]$ and is as defined in Section
4.3, and involves choosing a partition of $k$ into upto four primes that are
all $\Omega(k).$
$\Gamma(r,q)$ is defined as the semidirect product $\mathbb{F}_{r}^{+}\rtimes
C_{q},$ with $C_{q}$ viewed as a subgroup of the group
$\mathbb{F}_{r}^{\times}.$ It acts on $[r]$ as follows: We identify $[r]$ with
the field of size $r.$ Let $(b,a)$ be a typical element of $\Gamma_{r}$ where
$b\in\mathbb{F}_{r}$ and $a\in C_{q}.$ Then, $(b,a):x\mapsto ax+b.$
Thus, $\Gamma=\Gamma(p,q,r)$ acts on $[n].$ Let $m^{*}$ be the minimum size of
the orbit of any edge $(i,j)\in{[n]\choose 2}$ under the action of $\Gamma.$
One can show that
$m^{*}=\Omega(\min\\{p^{2}k,pkr,qr\\}).$ (5.1)
We shall choose $p,q,r$ carefully such that (a) the value of $m^{*}$ is large,
and (b) Oliver’s condition holds for $\Gamma(p,q,r)$.
### 5.2. ERH and Dirichlet primes
The Extended Riemann Hypothesis (ERH) implies the following strong version of
the Prime Number Theorem for arithmetic progressions. Let $\pi(n,D,a)$ denote
the numer of primes $p\leq n$, $p\equiv a\pmod{D}$. Then for $D<n$ we have
$\pi(n,D,a)=\frac{\operatorname{li}(n)}{\varphi(D)}+O(\sqrt{x}\ln x)$ (5.2)
where $\operatorname{li}(n)=\int_{2}^{n}dt/t$ and the constant implied by the
big-Oh notation is absolute (cf. [16, Ch. 7, eqn. (5.12)] or [2, Thm. 8.4.5]).
This result immediately implies “Bertrand’s Postulate for Dirichlet primes:”
###### Lemma 5.1 (Bertrand’s Postulate for Dirichlet primes).
Assume ERH. Suppose the sequence $D_{n}$ satisfies
$D_{n}=o(\sqrt{n}/\log^{2}n)$. Then for all sufficiently large $n$ and for any
$a_{n}$ relatively prime to $D_{n}$ there exists a prime $p\equiv
a_{n}\pmod{D_{n}}$ such that $\frac{n}{2}\leq p\leq n.$
### 5.3. With ERH but without Chowla
We want to write $n=pk+r,$ where $p$ and $r$ are primes, and with $q$ a prime
divisor of $r-1,$ as described in Section 5.1. Specifically, we try for:
$p=\Theta(n^{1/4}),\quad\frac{n}{4}\leq r\leq\frac{n}{2},\quad
q=\Theta(n^{1/4-\epsilon})$
We claim that under ERH, such a partition of $n$ is possible.
To see this, fix some $p=\Theta(n^{1/4})$ such that $\gcd(p,n)=1.$ Fix some
$q=\Theta(n^{1/4-\epsilon}).$ Now, $r\equiv 1\pmod{q}$ and $r\equiv n\pmod{p}$
solves to $r\equiv a\pmod{pq}$ for some $a$ such that $\gcd(a,pq)=1.$ Since
$pq=\Theta(n^{1/2-\epsilon}),$ we can conclude under ERH (using Lemma 5.1)
that there exists a prime $r\equiv a\pmod{pq}$ such that $\frac{n}{4}\leq
r\leq\frac{n}{2}.$ This gives us the desired partition. One can verify that
our $\Gamma$ satisfies Oliver’s Condition. Equation (5.1) gives
$m^{*}=\Omega(n^{5/4-\epsilon}).$ This completes the proof of part (b) of
Theorem 1.4. ∎
### 5.4. Stronger bound using Chowla’s conjecture
Let $a$ and $D$ be relatively prime. Let $p$ be the first prime such that
$p\equiv a\pmod{D}.$ Chowla’s conjecture tells us that $p=O(D^{1+\epsilon})$
for every $\epsilon>0.$ Using this, we show $m^{*}=\Omega(n^{3/2-\epsilon}).$
We can use Chowla’s conjecture, along with the general setup of Section 5.1,
to obtain a stronger lower bound on $m^{*}.$ The new bounds we hope to achieve
are:
$p=\Theta(\sqrt{n}),\quad n^{1-2.5\delta}\leq r\leq n^{1-0.5\delta},\quad
q=\Theta(n^{1/2-\delta})$
Such a partition is always possible assuming Chowla’s conjecture. To see this,
first fix $p=\Theta(n^{1/2}),$ then fix $q=\Theta(n^{1/2-2\delta})$ and find
the least solution for $r\equiv 1\pmod{q}$ and $r\equiv n\pmod{p},$ which is
equivalent to solving for $r\equiv a\pmod{pq}$ for some $a<pq.$ The least
solution will be greater than $pq$ unless $a$ happens to be a prime. In this
case, we add another constraint, say $r\equiv a+1\pmod{3}$ and resolve to get
the least solution greater than $pq.$ Note that $n^{1-2.5\delta}\leq r\leq
n^{1-0.5\delta}.$ Now, from Equation (5.1), we get the lower bound of
$m^{*}=\Omega(n^{3/2-4\delta}).$ This completes the proof of part (a) of
Theorem 1.4. ∎
### Acknowledgment.
Raghav Kulkarni expresses his gratitude to Sasha Razborov for bringing the
subject to his attention and for helpful initial discussions.
## References
* [1] Babai, L., Gál, A., Wigderson, A.: Superpolynomial lower bounds for monotone span programs. Combinatorica 19 (1999), 301–320.
* [2] Bach, E., Shallit, J.: Algorithmic Number Theory, Vol. 1. The MIT Press 1996.
* [3] Chakrabarti, A., Khot, S., Shi, Y.: Evasiveness of Subgraph Containment and Related Properties. SIAM J. Comput. 31(3) (2001), 866-875.
* [4] Chowla, S. On the least prime in the arithmetical progression. J. Indian Math. Soc. 1(2) (1934), 1–3.
* [5] Davenport, H.: Multiplicative Number Theory. (2nd Edn) Springer Verlag, New York, 1980.
* [6] Granville, A., Pomerance, C.: On the least prime in certain arithmetic progressions. J. London Math. Soc. 41(2) (1990), 193–200.
* [7] Haselgrove, C. B.: Some theorems on the analytic theory of numbers. _J. London Math. Soc._ 36 (1951) 273–277
* [8] Heath-Brown, D. R.: Almost-primes in arithmetic progressions and short intervals. Math. Proc. Cambr. Phil. Soc. 83 (1978) 357–376.
* [9] Heath-Brown, D. R.: Zero-free regions for Dirichlet $L$-functions, and the least prime in an arithmetic progression. Proc. London Math. Soc. 64(3) (1992) 265–338.
* [10] Kleitman, D. J., Kwiatkowski, D. J.: Further results on the Aanderaa-Rosenberg Conjecture J. Comb. Th. B 28 (1980), 85–90.
* [11] Kahn, J., Saks, M., Sturtevant, D.: A topological approach to evasiveness. Combinatorica 4 (1984), 297–306.
* [12] Lutz, F. H.: Examples of $\mathbb{Z}$-acyclic and contractible vertex-homogeneous simplicial complexes.. Discrete Comput. Geom. 27 (2002), No. 1, 137–154.
* [13] Mader, W.: Homomorphieeigenschaften und mittlere Kantendichte von Graphen. Math. Ann. 174 (1967), 265–268.
* [14] Mader, W.: Homomorphiesätze für Graphen. Math. Ann. 175 (1968), 154–168.
* [15] Oliver, R.: Fixed-point sets of group actions on finite acyclic complexes. Comment. Math. Helv. 50 (1975), 155–177.
* [16] Prachar, K.: Primzahlverteilung. Springer, 1957.
* [17] Rosenberg A. L.: On the time required to recognize properties of graphs: A problem. SIGACT News 5 (4) (1973), 15–16.
* [18] Rotman, J.: An Introduction to the Theory of Groups. Springer Verlag, 1994.
* [19] Rivest, R.L., Vuillemin, J.: On recognizing graph properties from adjacency matrices. Theoret. Comp. Sci. 3 (1976), 371–384.
* [20] Smith P. A.: Fixed point theorems for periodic transformations. Amer. J. of Math. 63 (1941), 1–8.
* [21] Titchmarsh, E. C.: A divisor problem. Rend. Circ. Mat. Palermo 54 (1930), 419–429.
* [22] Triesch, E.: On the recognition complexity of some graph properties. Combinatorica 16 (2) (1996) 259–268.
* [23] Turán, P.: Über die Primzahlen der arithmetischen Progression. Acta Sci. Math. (Szeged) 8 (1936/37) 226–235.
* [24] Vinogradov, I. M.: The Method of Trigonometrical Sums in the Theory of Numbers (Russian). Trav. Inst. Math. Stekloff 10, 1937.
* [25] Yao, A. C.: Monotone bipartite properties are evasive. SIAM J. Comput. 17 (1988), 517–520.
|
The evolution of the Mdisc-Mstar and Mdot-Mstar correlations traces protoplanetary disc dispersal
Alice Somigliana
1,2email<EMAIL_ADDRESS>Leonardo Testi3,4,
Giovanni Rosotti5,
Claudia Toci1,
Giuseppe Lodato5,
Rossella Anania5,
Benoît Tabone6,
Marco Tazzari4,
Ralf Klessen7, 8,
Ugo Lebreuilly 9,
Patrick Hennebelle 9,
and Sergo Molinari 10
European Southern Observatory, Karl-Schwarzschild-Strasse 2, D-85748 Garching bei München, Germany
Fakultat für Physik, Ludwig-Maximilians-Universität München, Scheinersts. 1, 81679 München, Germany
Dipartimento di Fisica e Astronomia, Universita' di Bologna, Via Gobetti 93/2, I-40122 Bologna, Italy
INAF-Osservatorio Astrofisico di Arcetri, Largo E. Fermi 5, I-50125 Firenze, Italy
Dipartimento di Fisica, Università degli Studi di Milano, Via Celoria 16, I-20133 Milano, Italy
Université Paris-Saclay, CNRS, Institut d'Astrophysique Spatiale, Orsay, France
Universität Heidelberg, Zentrum für Astronomie, Institut für Theoretische Astrophysik, Albert-Ueberle-Str. 2, D-69120 Heidelberg, Germany
Universität Heidelberg, Interdisziplinäres Zentrum für Wissenschaftliches Rechnen, Im Neuenheimer Feld 205, D-69120 Heidelberg, Germany
Université Paris-Saclay, Université Paris Cité, CEA, CNRS, AIM, 91191, Gif-sur-Yvette, France
Istituto Nazionale di Astrofisica-IAPS, Via Fosso del Cavaliere 100, I-00133 Roma, Italy
Observational surveys of entire star-forming regions have provided evidence of power-law correlations between the disc integrated properties and the stellar mass, especially the disc mass ($M_{\mathrm{d}} \propto {M_{\star}}^{\lambda_{\mathrm{m}}}$) and the accretion rate ($\dot M \propto {M_{\star}}^{\lambda_{\mathrm{acc}}}$). Whether the secular disc evolution affects said correlations is still a matter of debate: while the purely viscous scenario has been investigated, other evolutionary mechanisms could have a different impact. In this paper, we study the time evolution of the slopes $\lambda_{\mathrm{m}}$ and $\lambda_{\mathrm{acc}}$ in the wind-driven and viscous-wind hybrid case and compare it to the purely viscous prediction. We use a combination of analytical calculations, where possible, and numerical simulations performed with the 1D population synthesis code , that we also present and release to the community. Assuming $M_{\mathrm{d}}(0) \propto {M_{\star}}^{\lambda_{\mathrm{m}, 0}}$ and $\dot M(0) \propto {M_{\star}}^{\lambda_{\mathrm{acc}, 0}}$ as initial conditions, we find that viscous and hybrid accretion preserve the power-law shape of the correlations, while evolving their slope; on the other hand, MHD winds change the shape of the correlations, bending them in the higher or lower end of the stellar mass spectrum depending on the scaling of the accretion timescale with the stellar mass. However, we show how a spread in the initial conditions conceals this behaviour, leading to power-law correlations with evolving slopes like in the viscous and hybrid case. We analyse the impact of disc dispersal, intrinsic in the wind model and due to internal photoevaporation in the viscous case: we find that the currently available sample sizes ($\sim 30$ discs at 5 Myr) introduce stochastic oscillations in the slopes evolution, which dominate over the physical signatures. We show that we could mitigate this issue by increasing the sample size: with $\sim 140$ discs at 5 Myr, corresponding to the complete Upper Sco sample, we would obtain small enough error bars to use the evolution of the slopes as a proxy for the driving mechanism of disc evolution. Finally, from our theoretical arguments we discuss how the observational claim of steepening slopes necessarily leads to an initially steeper $M_{\mathrm{d}} - M_{\star}$ correlation with respect to $\dot M - M_{\star}$.
The $M_{\mathrm{d}}-M_{\star}$ and $\dot M - M_{\star}$ correlations trace disc dispersal
Somigliana, A. et al.
§ INTRODUCTION
The secular evolution of protoplanetary discs is deeply intertwined with both the planet formation process [Morbidelli et al., 2012] and the accretion onto the central protostar [Hartmann et al., 1998]. Planetesimals, the building blocks of planets, form and evolve within the disc following the dynamics of either the gaseous or solid component, depending on their relative size and their coupling (or lack thereof) with the gas particles; on the other hand, the protostar is fed by the disc itself, through the accretion of material that loses angular momentum and drifts inwards. The ideal ground to explore the connection between protoplanetary discs and their host stars is provided by large surveys of entire star-forming regions, targeting the properties of both discs and protostars; the last decade has seen a significant observational effort in the direction of these population-level studies, also thanks to the advent of facilities like the Atacama Large Millimeter Array (ALMA) (see the PPVII reviews by Manara et al., 2023, Miotello et al., 2023).
Disc masses and accretion rates are arguably the most studied integrated disc properties. Accretion rates are inferred from the spectra of the central stars, which show an excess emission (especially prominent in the UV) when accretion is taking place; surveys performed across different star-forming regions [Muzerolle et al., 2003, Natta et al., 2004, Mohanty et al., 2005, Dullemond et al., 2006, Herczeg & Hillenbrand, 2008, Rigliaco et al., 2011, Manara et al., 2012, Alcalá et al., 2014, Manara et al., 2016, Alcalá et al., 2017, Manara et al., 2017, Venuti et al., 2019, Manara et al., 2020] agree on the presence of a power-law correlation between the accretion rate and the stellar mass, $\dot M \propto {M_{\star}}^{\lambda_{\mathrm{acc, obs}}}$ (following the notation of Somigliana et al., 2022). On the other hand, disc masses have traditionally been determined from observations of the sub-mm continuum emission of the solid component of discs; due to the large number of assumptions involved in converting sub-mm fluxes into total disc masses (see Miotello et al., 2023), one of the current main goals of the protoplanetary disc community is the accurate determination of total disc masses - both from dynamical constraints [Veronesi et al., 2021, Lodato et al., 2023] and direct measurements of the total gas content (e.g., Bergin et al., 2013, Anderson et al., 2022, Trapman et al., 2022). Despite the systematic uncertainties involved in their determination, dust-based disc masses also seem to show a power-law correlation with the stellar mass, $M_{\mathrm{d}} \propto {M_{\star}}^{\lambda_{\mathrm{m, obs}}}$, across different star-forming regions [Ansdell et al., 2016, Ansdell et al., 2017, Barenfeld et al., 2016, Pascucci et al., 2016, Testi et al., 2016, Testi et al., 2022, Sanchis et al., 2020].
The existence of the disc mass-stellar mass and accretion rate-stellar mass correlations is now generally accepted; however, there is no consensus on the physical reason behind their establishment and their evolution with time. While the $\dot M - M_{\star}$ correlation appears to have a roughly constant slope[Throughout this work, we use 'slope' as a synonym of power-law index, referring to the correlations in the logarithmic plane.] of $\lambda_{\mathrm{acc, obs}} \approx 1.8 \pm 0.2$ (as first suggested by Muzerolle et al., 2003 and supported by many of the following works mentioned above), the $M_{\mathrm{d}} - M_{\star}$ correlations is claimed to be steepening with time [Ansdell et al., 2017], from the lowest $\lambda_{\mathrm{m, obs}} ( t \sim 1$ Myr$ ) = 1.7 \pm 0.2$ (Taurus) to the highest $\lambda_{\mathrm{m, obs}} (t \sim 5$ Myr$) = 2.4 \pm 0.4$ (Upper Sco). Whether these correlations reflect the initial conditions of disc populations, or are rather a product of the secular evolution, is still under debate. Both possibilities have been discussed for the $\dot M - M_{\star}$ correlation: [Alexander & Armitage, 2006] have assumed it to hold as initial condition, favouring the correlation to be present in young populations, whereas [Dullemond et al., 2006] have derived it from a simple model of disc formation from a rotating collapsing core, which provided an explanation for evolved disc populations. At the same time, the claimed increase in the slope of $M_{\mathrm{d}} - M_{\star}$ does suggest an evolutionary trend; [Somigliana et al., 2022] have found that, assuming power-law correlations between both $M_{\mathrm{d}}$ and $\dot M$ and the stellar mass as initial conditions, secular evolution can indeed alter the slopes of the correlations themselves (see Section <ref> for details). However, their analysis was limited to the standard viscous evolution paradigm, whereas the driving mechanism of accretion is far from being constrained (see Manara et al., 2023 for a review).
The traditional viscous accretion model prescribes a macroscopic viscosity as the cause of redistribution of angular momentum within the disc [Lynden-Bell & Pringle, 1974, Pringle, 1981]. In this scenario, while part of the material loses angular momentum and moves radially closer to the star, some other material gains the same amount of angular momentum and moves further away, effectively increasing the disc size. The viscous paradigm can explain many key features of disc evolution, but it cannot account for disc dispersal - as determined from the observational evidence of exponentially decreasing fraction of both disc-bearing [Hernández et al., 2007] and accreting [Fedele et al., 2010] sources in star-forming regions with time; furthermore, the low levels of turbulence detected in discs [Pinte et al., 2016, Flaherty et al., 2018, Rosotti, 2023] appear incompatible with the observed evolution. While the discrepancy in the disc and accretion fraction can be mended considering mechanisms such as internal or external photoevaporation [Alexander et al., 2014, Winter et al., 2018], that effectively clear discs on timescales comparable with the observed decline, the tension between the expected and observed amount of turbulence does not appear to be solved yet. On the other hand, the MHD disc winds scenario offers a promising alternative. Pioneering work [Blandford & Payne, 1982, Ferreira, 1997] supported by recent numerical simulations (e.g., Béthune et al., 2017) demonstrated that MHD winds launched from the disc surface have the net effect of removing angular momentum as a consequence of the extraction of material; SMHD wind-driven accretion can even lead to disc dispersal [Armitage et al., 2013, Tabone et al., 2022]. Following disc evolution at population level in numerical simulations remains out of reach; however, three-dimensional core-collapse simulations have shown how non-ideal magnetohydrodynamics and ambipolar diffusion play a fundamental role in shaping the resulting population of early-type young stellar objects [Lebreuilly et al., 2021, Lebreuilly et al., 2024]. While some 3D studies of isolated disc formation have attempted to bridge the gap between Class 0/I and Class II stages [Machida & Hosokawa, 2013, Hennebelle et al., 2020, Xu & Kunz, 2021, Xu & Kunz, 2021, Machida & Basu, 2024, Mauxion et al., 2024], the high numerical cost of the simulations for 3D population synthesis does not allow to follow the evolution of the discs up to very evolved stages where they can be considered isolated from the surrounding environment. MHD wind-driven disc populations can however be modelled in 1D using simple prescriptions as proposed by [Suzuki et al., 2016] or [Tabone et al., 2022]. Detecting characteristic signatures of either of the two evolutionary prescriptions is a compelling issue [Long et al., 2022, Alexander et al., 2023, Somigliana et al., 2023, Trapman et al., 2023, Coleman et al., 2024].
In the context of the evolution of the correlations between the disc properties and the stellar mass, while the purely viscous scenario has been extensively studied by [Somigliana et al., 2022], the wind-driven paradigm remains unexplored; with this paper, we address this deficiency and investigate the impact of MHD wind-driven evolution on the $M_{\mathrm{d}} - M_{\star}$ and $\dot M - M_{\star}$ correlations, with a particular focus on their time evolution and the comparison with the purely viscous paradigm. We also extend the work of [Somigliana et al., 2022] by including internal photoevaporation to the viscous framework. We employ numerical simulations of populations of protoplanetary discs, performed with the population synthesis code , which we also introduce and release to the community.
The paper is structured as follows: in Section <ref>, we present and describe its main features, set up and solution algorithm; in Section <ref>, we discuss the time evolution of the $M_{\mathrm{d}} - M_{\star}$ and $\dot M - M_{\star}$ correlations in three evolutionary scenarios from the theoretical perspective; in Section <ref>, we show the impact of a spread in the initial conditions and dispersal mechanisms on the evolution of the slopes, and we present the numerical results obtained from realistic disc population synthesis; in Section <ref> we interpret the implications of our findings in the context of the observational determination of the slopes, and finally in Section <ref> we draw the conclusions of this work.
§ NUMERICAL METHODS:
In this Section we present the 1D population synthesis code [ and the output analysis library can be installed via the Python Package Index, and . The full documentation and tutorials are available at https://alicesomigliana.github.io/diskpop-docs/index.html. If you use in your work, please cite this paper (Somigliana et al. 2024).]. We describe the master equation for the secular evolution of discs (Section <ref>), the initial conditions to generate a synthetic population (Section <ref>), the solution algorithm (Section <ref>), and the user interface and output (Section <ref>). For a more detailed description, we refer to the code documentation; for a validation of the code, see Appendix <ref>.
§.§ Master equation
The master equation of protoplanetary disc evolution,
\begin{equation}
\label{eq:master_equation}
\begin{split}
\frac{\partial \Sigma}{\partial t} = \frac{3}{r} \frac{\partial}{\partial r} \left[ \frac{1}{\Omega r} \frac{\partial}{\partial r} \left( r^2 \alpha_{\mathrm{SS}} \Sigma {c_s}^2 \right) \right] + \frac{3}{2r} \frac{\partial}{\partial r} \left[ \frac{\alpha_{\mathrm{DW}} \Sigma {c_s}^2}{\Omega} \right] \\
- \frac{3 \alpha_{\mathrm{DW}} \Sigma {c_s}^2}{4 (\lambda-1)r^2 \Omega} - \dot{\Sigma}_{\mathrm{photo}},
\end{split}
\end{equation}
describes the time evolution of the gas surface density in the most general framework, where $\Sigma$ is the gas surface density, $\Omega$ the Keplerian orbital frequency, $\alpha_{\mathrm{SS}}$ the [Shakura & Sunyaev, 1973] $\alpha$ parameter, $\alpha_{\mathrm{DW}}$ the MHD equivalent of $\alpha_{\mathrm{SS}}$ [Tabone et al., 2022], $c_s$ the sound speed, and $\lambda$ the magnetic lever arm parameter, which quantifies the ratio of extracted to initial specific angular momentum. The four terms on the right hand side (RHS) refer to (i) the viscous torque, whose strength is parameterised by $\alpha_{\mathrm{SS}}$, (ii) the wind-driven accretion, which corresponds to an advection term, parameterised by $\alpha_{\mathrm{DW}}$, (iii) mass loss due to MHD disc winds, parameterised by $\lambda$ and (iv) mass loss due to other physical phenomena (in our case, we consider internal and external photoevaporation). Depending on the values of the specific parameters, Equation (<ref>) can describe a purely viscous ($\alpha_{\mathrm{DW}} = 0$), purely MHD wind-driven ($\alpha_{\mathrm{SS}} = 0$) or hybrid ($\alpha_{\mathrm{SS}}, \alpha_{\mathrm{DW}} \neq 0$) evolution, with ($\dot \Sigma_{\mathrm{photo}} \neq 0$) or without ($\dot \Sigma_{\mathrm{photo}} = 0$) the influence of photoevaporation. In the following, we briefly describe the various evolutionary scenarios and the available analytical solutions.
Viscously evolving discs. In the case of purely viscous evolution, the MHD winds parameter $\alpha_{\mathrm{DW}}$ is set to zero. If we also neglect the influence of photoevaporation, Equation (<ref>) reduces to the first term on the RHS and its solution depends on the functional form of the effective viscosity, parameterised as $\nu = \alpha_{\mathrm{SS}} c_s H$ (where $H$ is the vertical height of the disc). A popular analytical solution for viscous discs is the [Lynden-Bell & Pringle, 1974] self-similar solution, which assumes viscosity to scale as a power-law of the radius ($\nu \propto R^{\gamma}$).
MHD winds-driven evolution. There are two classes of analytical solutions to Equation (<ref>) in the MHD wind-driven scenario, associated with a specific prescription of $\alpha_{\mathrm{DW}}$ [Tabone et al., 2022]. We briefly describe their key features, and refer to the original paper for their derivation and an in-depth discussion.
* The simplest class of solutions (so-called hybrid solutions), which highlight the main features of wind-driven accretion in comparison to the viscous model, assume a constant $\alpha_{\mathrm{DW}}$ with time; these solutions depend on the value of $\psi \equiv \alpha_{\mathrm{DW}}/\alpha_{\mathrm{SS}}$, which quantifies the relative strength of the radial and vertical torque.
* Another class of solutions, which describe the unknown evolution of the magnetic field strength, assume a varying $\alpha_{\mathrm{DW}}$ with time. To obtain these, [Tabone et al., 2022] parameterised $\alpha_{\mathrm{DW}}(t) \propto \Sigma_{\mathrm{c}} (t)^{-\omega}$, with $\Sigma_{\mathrm{c}} = M_{\mathrm{d}}(t)/2 \pi {R_c}^2 (t)$ (where $R_c$ is a characteristic radius) and $\omega$ as a free parameter, and neglect the radial transport of angular momentum ($\alpha_{\mathrm{SS}} = 0$).
Photoevaporation. The generic $\dot \Sigma_{\mathrm{photo}}$ term in Equation (<ref>) allows to account for photoevaporative processes, both internal and external. The exact form of $\dot \Sigma_{\mathrm{photo}}$ depends on the specific model considered; therefore, the availability (or lack thereof) of analytical solutions needs to be considered case by case.
allows to evolve populations of discs analytically. In particular, as of this release, it includes implementations of the [Lynden-Bell & Pringle, 1974] self-similar solution and all the analytical solution proposed by [Tabone et al., 2022]. In the cases where Equation (<ref>) cannot be solved analytically, the code relies on the solution algorithm described in Section <ref>.
§.§ Initial conditions and parameters
Every simulation begins with the generation of a synthetic population of Young Stellar Objects (YSOs). Each YSO constitutes of a star and a disc, whose key initial parameters (stellar mass, disc mass, accretion rate, disc radius, evolutionary parameters $\alpha_{\mathrm{SS}}$, $\alpha_{\mathrm{DW}}$, $\lambda$, $\omega$...) can be set by the user. In the following, we describe the standard case where we consider the stellar masses to be distributed according to an Initial Mass Function (IMF) and correlating with the disc mass and radius, and briefly mention the other possible choices; for a deeper discussion, we refer to the documentation.
assembles YSOs by determining their parameters as follows:
* Stellar mass $M_{\star}$: determined following the [Kroupa, 2001] IMF. Other possible choices are a constant mass for all the stars in the population, or a set of custom stellar masses.
* Initial disc mass $M_{\mathrm{d}}$, accretion rate $\dot M$: determined from log-normal distributions of given width and mean value. In the standard case, considers an initial power-law correlation between the initial $M_{\mathrm{d}}$ and $\dot M$ and the stellar mass (see Section <ref> for a detailed discussion), where the normalisation at 1 M$_{\odot}$, the slope and the scatter around the power-laws are free parameters. If the correlations with the stellar mass are neglected, the user sets the mean value and spread of the distributions.
* Accretion parameters ($\alpha_{\mathrm{SS}}$, $\alpha_{\mathrm{DW}}$, $\lambda$, $\omega$): global properties of the whole population, given as input from the user. By setting the parameters controlling accretion, determines the disc radius $R_{\mathrm{d}}$ and accretion timescale $t_{\mathrm{acc}}$ - which are instead disc-specific and linked to the disc mass and accretion rate.
* Internal photoevaporation parameters ($\dot M_{\mathrm{wind}}$, $L_X$): the total photoevaporative mass-loss rate, $\dot M_{\mathrm{wind}}$, can either be set by the user or computed from the stellar X luminosity $L_X$ as [Owen et al., 2012]
\begin{equation*}
\dot M_{\mathrm{wind}} = 6.25 \times 10^{-9} \times \left( \frac{M_{\star}}{M_{\odot}} \right)^{-0.068} \left( \frac{L_X}{10^{30} \mathrm{erg}\mathrm{s}^{-1}} \right)^{1.14} M_{\odot} \mathrm{yr}^{-1}.
\end{equation*}
The surface mass-loss profile $\dot \Sigma_{\mathrm{photo}}$ (Equation B2 in Owen et al., 2012) is then scaled so that $\int 2 \pi R \dot \Sigma_{\mathrm{photo}} \rm{d}R$ is equal to $\dot M_{\mathrm{wind}}$. Like for $M_{\mathrm{d}}$ and $\dot M$, $\dot M_{\rm{wind}}$ (or equivalently $L_X$) is extracted from a log-normal distribution whose mean is determined assuming power-law correlations with the stellar mass, while the normalisation at 1 M$_{\odot}$, the slope and the width of the distribution are free parameters.
* External photoevaporation parameters ($FUV$): FUV flux experienced by each disc, in units of G$_{0}$[G$_{0}$ stands for the Habing unit [Habing], 1968], the flux integral over the range of wavelengths [912 - 2400] $\rm{\mathring{A}}$ weighted by the average value in the solar neighbourhood (1.6$\times 10^{-3}$ erg s$^{-1}$ cm$^{-2}$).. This parameter can be set to any value accessible in the FRIEDv2 grid of mass loss rate [Haworth et al., 2023], spanning from 1 to $10^{5}$ G$_{0}$.
§.§ Solution algorithm
After generating the initial population of YSOs as described above, proceeds to evolve it by integrating the master equation (<ref>). Our solution algorithm employs an operator splitting method: the original equation is separated into different parts over a time step, and the solution to each part is computed separately. Then, all the solutions are combined together to form a solution to the original equation. We split Equation (<ref>) into five different pieces, related to viscosity, wind-driven accretion onto the central star, wind-driven mass loss, internal and external photoevaporation respectively. Furthermore, includes the possibility to trace the dust evolution in the disc, which is split in radial drift and dust diffusion. In the following, we describe the solution algorithm for each process.
* Viscous accretion: the standard viscous solver is based on the freely available code by [Booth et al., 2017]. We assume a radial temperature profile $T \propto R^{-1/2}$, which results in $c_{s} \propto R^{-1/4}$ and $H/R \propto R^{1/4}$. Note that this implies $\nu \propto R$ (i.e., $\gamma = 1$), which will be the case from now on. We assume $H/R = 1/30$ at 1 AU and a mean molecular weight of $2.4$. We refer to the original paper for details on the algorithm.
* Wind-driven accretion: the second term in Equation (<ref>) is effectively an advection term. The general form of the advection equation for a quantity $q$ with velocity $v$ is $\partial_t q(x, t) + v \partial_x q(x, t) = 0$; in the case of wind-driven accretion, the advected quantity is $R \Sigma$, while the advection (inwards) velocity is given by $v_{\mathrm{DW}} = (3 \alpha_{\mathrm{DW}} H c_s)/2R$. We solve the advection equation with an explicit upwind algorithm (used also for dust radial drift).
* Wind-driven mass loss: the mass loss term (third in Equation <ref>) does not involve any partial derivative, and therefore is simply integrated in time multiplying by the time step.
* Internal photoevaporation: effectively, internal photoevaporation (implemented through the model of Owen et al., 2012) is another mass loss term - therefore, as above, its contribution is computed with a simple multiplication by the time step. Once the accretion rate of the disc drops below the photoevaporative mass loss rate, a gap opens in the disc at the radius of influence of photoevaporation: in the model of [Owen et al., 2012], the prescription changes depending on the radial location in the disc, with respect to the gap itself. Later, the gap continues to widen; when it eventually becomes larger than the disc, we stop the evolution and consider the disc as dispersed.
* External photoevaporation: for a given stellar mass and FUV flux experienced by the disc, the mass loss rate arising from external photoevaporation is obtained, at each radial position, from a bi-linear interpolation of the FRIEDv2 grid [Haworth et al., 2023] using the disc surface density at each radial cell. The outside-in depletion of material is implemented following the numerical approach of Sellek et al., 2020: we define the truncation radius, $R_{\mathrm{t}}$, as the position in the disc corresponding to the maximum photoevaporation rate (which is related to the optically thin/thick transition of the wind), and we remove material from each grid cell at $R>R_{\mathrm{t}}$ weighting on the total mass outside this radius. The mass loss attributed to the cell $i$ can be written as:
\begin{equation}
\dot{M}_{\mathrm{ext},i} = \dot{M}_{\mathrm{tot}} \frac{M_{i}}{M(R>R_{\mathrm{t}})},
\end{equation}
where $M_{i}$ is the mass contained in the cell $i$, and $\dot{M}_{\mathrm{tot}}$ is the total mass loss rate outside the truncation radius.
* Dust evolution[As the dust evolution module was forked from Richard Booth's repository, users of who wish to use dust in their work ought to cite [Booth] et al., 2017] together with this paper.: based on the two populations model by [Birnstiel et al., 2012] and the implementation of [Booth et al., 2017]. We consider the dust grain distribution to be described by two representative sizes, a constant monomer size and a time-dependent larger size, which can grow up to the limit imposed by the fragmentation and radial drift barriers. We evolve the dust fraction of both sizes following [Laibe & Price, 2014], and also include a diffusive term: the diffusion comes from the coupling with the turbulent gas, which has the effect of mixing the dust grains, counteracting gradients in concentration [Birnstiel et al., 2010]. The dust-gas relative velocities are computed following [Tanaka et al., 2005] and include feedback on the gas component. We refer to [Booth et al., 2017] for details on the numerical implementation. Dust evolution is included in the release of , however the scientific results presented in this work are based on gas simulations only.
The separate pieces of Equation (<ref>) must be solved over the same time step to be joined in a coherent solution. We calculate the time step for each process imposing the Courant-Friedrichs-Lewy (CFL) condition. The CFL condition reads $\Delta t = C \mathrm{ min}(\Delta x / v)$ and ensures that, within one time step $\Delta t$, the material moving at velocity $v$ does not flow further than one grid spacing $\Delta x$. The Courant number $C$ must be positive and smaller than 1, with $C = 1$ corresponding to the maximum allowed timestep to keep the algorithm stable. In our implementation, we pick $C = 0.5$. We use zero gradients boundary conditions, setting the value of the first and last cell in our grid to that of the second and second to last. We solve the equation on a radial grid of $10^3$ points with power-law spacing and exponent $1/2$, extending from $3 \times 10^{-3}$ au to $10^4$ au. From the physical point of view, this choice corresponds to assuming boundary layer accretion (see., e.g., Popham et al., 1993, Kley & Lin, 1996) - however the difference from magnetic truncation accretion is negligible beyond $\sim 10^{-3}$ au.
After each process has been solved separately, all the pieces are put back together to compute the new surface density, from which the integrated disc quantities are then calculated. As each disc evolves independently of the others in the population, the solver can easily be run in parallel.
§.§ User interface and output
The user interface of is a .json parameters file which includes all the user-dependent parameters. Aside from the number of objects in the population and the evolutionary mechanism, the user can set the chosen IMF (either Kroupa, 2001, single stellar mass, or custom input file), the distributions to draw the disc parameters from (single value, flat, normal, log-normal), as well as the normalisation, slope and spread of the correlations, the times at which snapshots are generated, and the initial dust-to-gas ratio. Furthermore, the user can determine a limit disc mass: this is to be intended as a threshold below which the disc would not be detectable anymore, and is therefore considered dispersed in the simulation as well. When a disc is dispersed, the corresponding YSO turns into a Class III object consisting of the central star only.
The output of is a .hdf5 file containing the properties of both the disc and the star at all chosen time steps for each YSO in the population: this includes the stellar mass, luminosity, temperature, disc mass, accretion rate, accretion timescale, gas and dust surface density, disc radius, dust grain sizes. The output can be easily read and analysed with the dedicated library , released with the code. For a more in-depth description of the parameters, the user interface and the output, we refer to the documentation.
§ THE TIME EVOLUTION OF THE CORRELATIONS BETWEEN DISC PROPERTIES AND STELLAR MASS UNDER DIFFERENT ACCRETION DRIVERS: ANALYTICAL CONSIDERATIONS
H]P1.5cm P5.5cm
0pt2.5ex Parameter Description
0pt2.5ex $\lambda$ Magnetic lever arm parameter
$\psi$ Wind-to-turbulent $\alpha$ ratio
$\omega$ Power-law index of $\alpha_{\mathrm{DW}}$ with $\Sigma_{\mathrm{c}}$
$\xi$ Mass ejection index
$f_{\mathrm{M}, 0}$ Initial mass ejection-to-accretion ratio
0pt2.5ex $\lambda_{\mathrm{m}}$ Power-law index of $M_{\mathrm{d}}$ with $M_{\star}$ (Eq. <ref>)
$\lambda_{\mathrm{acc}}$ Power-law index of $\dot M$ with $M_{\star}$ (Eq. <ref>)
$\zeta$ Power-law index of $R_{\mathrm{d}}$ with $M_{\star}$ (Eq. <ref>)
$\beta$ Power-law index of $H/R$ with $M_{\star}$
$\mu$ Power-law index of $t_{\mathrm{acc}}$ with $M_{\star}$
$\delta$ $\lambda_{\mathrm{m}} - \lambda_{\mathrm{acc}}$
Summary and description of the parameters used throughout the paper: the top block refers to the MHD parameters defined in [Tabone et al., 2022], while the bottom block shows the slopes of the correlations between the disc properties and the stellar mass.
The existence of power-law correlations between the main integrated disc properties - namely the disc mass and stellar accretion rate - and the stellar mass is supported by various surveys across a number of different star-forming regions (e.g., on $M_{\mathrm{d}}-M_{\star}$: Ansdell et al., 2016, Barenfeld et al., 2016, Pascucci et al., 2016, Testi et al., 2016; on $\dot M - M_{\star}$: Muzerolle et al., 2003, Natta et al., 2004, Mohanty et al., 2005, Alcalá et al., 2014, Manara et al., 2016, Alcalá et al., 2017, Manara et al., 2017, Venuti et al., 2019, Manara et al., 2020, Testi et al., 2022). However, whether the establishment and subsequent evolution of said correlations is a product of the secular evolution of discs, or rather an imprint of the initial conditions, remains unclear. [Somigliana et al., 2022] explored a combination of both possibilities, assuming the correlations to hold as initial conditions and investigating the impact of purely viscous evolution; we briefly recall their main theoretical results (Section <ref>) and extend their analysis to the hybrid (Section <ref>) and purely wind-driven (Section <ref>) models from the theoretical perspective. We note that the results presented in this work are based on gas simulations.
Following [Somigliana et al., 2022], we assume power-law correlations between the disc properties and the stellar mass to hold as initial conditions. We focus on the disc mass $M_{\mathrm{d}}$, the stellar accretion rate $\dot M$ and the disc radius $R_{\mathrm{d}}$, and label the slopes of their correlations with the stellar mass $\lambda_{\mathrm{m}}$, $\lambda_{\mathrm{acc}}$, and $\zeta$ respectively. The initial correlations are set as follows:
\begin{equation}
\begin{cases}
M_\mathrm{d} (0) \propto {M_{\star}}^{\lambda_{\mathrm{m}, 0}}, \\
\dot{M} (0) \propto {M_{\star}}^{\lambda_{\mathrm{acc}, 0}}, \\
R_{\mathrm{d}} (0) \propto {M_{\star}}^{\zeta_0}.
\end{cases}
\label{eq:assumptions}
\end{equation}
To analyse the impact of secular evolution on this set of initial conditions, we analytically determine the evolved expressions for $M_\mathrm{d} (t)$, $\dot{M} (t)$ and $R_{\mathrm{d}} (t)$ in the three different scenarios. Table <ref> summarises the parameters introduced in this Section.
§.§ Purely viscous model
The full calculations for the purely viscous case can be found in [Somigliana et al., 2022]. Here, we briefly remind the main assumptions and results, and we refer to the original paper for a detailed discussion.
As mentioned in Section <ref>, assuming a power-law scaling of viscosity with the disc radius ($\nu \propto R^{\gamma}$) allows to solve the viscous evolution equation analytically, recovering the so-called self-similar solution [Lynden-Bell & Pringle, 1974]. In this case, the disc mass and accretion rate read
\begin{equation}
M_\mathrm{d} (t) = M_{\mathrm{d}, 0} \left( 1 + \frac{t}{t_{\nu}} \right)^{1 - \eta},
\label{eq:discmass_ss}
\end{equation}
\begin{equation}
\dot M (t) = (\eta - 1) \frac{M_{\mathrm{d}, 0}}{t_{\nu}} \left( 1 + \frac{t}{t_{\nu}} \right)^{- \eta},
\label{eq:accrate_ss}
\end{equation}
where $\eta = (5/2 - \gamma)/(2 - \gamma)$ and the viscous timescale $t_{\nu} = {R_c}^2/[3(2-\gamma)^2 \nu(R = R_c)]$ at the characteristic radius $R_c$. Because $\dot M_0 \propto M_{\mathrm{d}, 0}/t_{\nu, 0}$, a power-law scaling of $M_{\mathrm{d}, 0}$ and $\dot M_0$ with the stellar mass implies the viscous timescale $t_{\nu, 0}$ to scale as a power-law with the stellar mass as well, which we define as $t_{\nu, 0} \propto {M_{\star}}^{\mu_0}$; furthermore, this scaling corresponds to the difference between the scaling of the disc mass with the stellar mass and of the accretion rate with the stellar mass. Defining $\delta_0 = \lambda_{\mathrm{m}, 0} - \lambda_{\mathrm{acc}, 0}$, in this case $\mu_0 = \delta_0$[The definition of $\delta_0$ might seem redundant at this stage, but it will become important in the following discussion.]; therefore, the scaling of $t_{\nu, 0}$ with the stellar mass is determined by the relative values of $\lambda_{\mathrm{m}, 0}$ and $\lambda_{\mathrm{acc}, 0}$. The main results of [Somigliana et al., 2022] are that (i) viscous evolution maintains the power-law shape of the correlations between the stellar mass and the disc parameters, however (ii) the slope of said correlations may evolve with time, depending on the initial conditions. This is because in a purely viscous framework, the $M_{\mathrm{d}}-\dot M$ correlation is bound to reach a linear correlation with slope unity [Lodato et al., 2017, Rosotti et al., 2017], which implies the two quantities to have the same dependence on the stellar mass, as $M_{\mathrm{d}} / \dot M \propto {M_{\star}}^{\lambda_{\mathrm{m}} - \lambda_{\mathrm{acc}}}$. Therefore, $\lambda_{\mathrm{m}}$ and $\lambda_{\mathrm{acc}}$ must eventually reach the same value, determined by the initial conditions as
\begin{equation}
\lambda_{\mathrm{m, evo}} = \lambda_{\mathrm{acc, evo}} = \frac{3 \lambda_{\mathrm{m}, 0} - \lambda_{\mathrm{acc}, 0}}{2}.
\label{eq:lambdam=lambdaac_evolved}
\end{equation}
Depending on the sign of $\delta_0 = \lambda_{\mathrm{m}, 0} - \lambda_{\mathrm{acc}, 0}$, the initial slopes can either
* steepen, i.e. $\lambda_{\mathrm{m, evo}} > \lambda_{\mathrm{m}, 0}$, if $\delta_0 > 0$ (implying also
$\lambda_{\mathrm{acc, evo}} > \lambda_{\mathrm{acc}, 0}$);
* flatten, i.e. $\lambda_{\mathrm{m, evo}} < \lambda_{\mathrm{m}, 0}$, if $\delta_0 < 0$ (implying also
$\lambda_{\mathrm{acc, evo}} < \lambda_{\mathrm{acc}, 0}$);
* remain constant, i.e. $\lambda_{\mathrm{m, evo}} = \lambda_{\mathrm{m}, 0}$, if $\delta_0 = 0$ (implying
also $\lambda_{\mathrm{acc, evo}} = \lambda_{\mathrm{acc}, 0}$).
Because in the viscous case $\delta_0 = \mu_0$, where we note again that $\mu_0$ is the slope of the correlation between the viscous timescale and the stellar mass ($t_{\nu, 0} \propto {M_{\star}}^{\mu_0}$), we can also interpret these scenarios from the viscous timescale perspective. If $\mu_0 > 0$, meaning that the viscous timescale increases with the stellar mass, discs around less massive stars will have shorter viscous timescales, which leads to a faster evolution, compared to discs around more massive stars, which will in turn have longer viscous timescales. This uneven evolution across the stellar mass spectrum leads to a steepening of the linear correlation, as is visualised by [Somigliana et al., 2022] in Figure 1. The same reasoning, but with opposite/constant trend, applies to the other two scenarios.
§.§ Hybrid model - omega = 0
In the hybrid viscous and MHD winds model, the general analytical solution by [Tabone et al., 2022] gives
\begin{equation}
M_\mathrm{d} (t) = M_0 \left( 1 + \frac{t}{(1 + \psi) t_{\mathrm{acc}, 0}} \right)^{-(\psi + 2 \xi + 1)/2},
\label{eq:mdisc_mhd_omegazero}
\end{equation}
\begin{equation}
\dot{M} (t) = \dot{M}_0 \left( 1 + \frac{t}{(1 + \psi) t_{\mathrm{acc}, 0}} \right)^{-(\psi + 4 \xi + 3)/2},
\label{eq:mdot_mhd_omegazero}
\end{equation}
where $\dot{M}_0$ is defined as
\begin{equation}
\dot{M}_0 = \frac{\psi + 1 + 2 \xi}{\psi + 1} \frac{M_0}{2 t_{\mathrm{acc}, 0}} \frac{1}{(1+f_{\mathrm{M}_0})};
\label{eq:mdot0_mhd_omegazero}
\end{equation}
Time evolution of the slopes of the $M_{\mathrm{d}}-M_{\star}$ and $\dot M-M_{\star}$ correlations, $\lambda_{\mathrm{m}}$ and $\lambda_{\mathrm{acc}}$ (blue and orange solid line respectively) in the hybrid scenario with $\alpha_{\mathrm{SS}} = \alpha_{\mathrm{DW}} = 10^{-3}$ ($\psi = 1$), $\lambda = 3$, $\beta = 0.5$, resulting in $\xi = 0.11$. The dashed lines represent the expected evolved value of both slopes, as in Equation (<ref>). For comparison, we include the viscous evolution as well, represented by grey solid (actual evolution) and dashed (expected evolved value) lines (see Somigliana et al., 2022 for a detailed discussion). The three panels show different values of $\mu_0$, slope of the $t_{\mathrm{acc}, 0} - M_{\star}$ correlation, which is directly linked to the difference between $\lambda_{\mathrm{m}, 0}$ and $\lambda_{\mathrm{acc}, 0}$ (see text for details): we expect the slopes to (i) decrease if $\mu_0 < 0$ (left panel), (ii) remain constant if $\mu_0 = 0$ (central panel), and (iii) increase if $\mu_0 > 0$ (right panel). Contrary to the viscous case, the slopes are not expected to reach the same value anymore, but rather settle to a constant difference of $- \xi/2$. This difference is always negative, meaning that the evolved $\dot M - M_{\star}$ correlation is always steeper than that of $M_{\mathrm{d}} - M_{\star}$ (explaining the lines crossing in the right panel).
in this notation, $\psi = \alpha_{\mathrm{DW}}/\alpha_{\mathrm{SS}}$ represents the relative strength of MHD winds and viscosity,
\begin{equation*}
\xi = \frac{1}{4} (\psi + 1) \left[ \sqrt{1 + \frac{4\psi}{(\lambda-1)(\psi + 1)^2}} - 1 \right]
\end{equation*}
is the mass ejection index quantifying the local mass loss rate to the local accretion rate, and $f_{\mathrm{M}, 0} = (R_{c, 0}/R_{in})^{\xi} - 1$ the dimensionless mass ejection-to-accretion ratio (with $R_{in}$ initial disc radius). If we neglect the MHD-driven mass loss ($\psi \ll 1$ and $\xi \ll 1$, which correspond to $f_{\mathrm{M}, 0} \ll 1$ as well), Equations (<ref>) and (<ref>) reduce to the viscous case; on the other hand, if mass loss is included, it depends on the radial extent of the disc through $f_{\mathrm{M}, 0} + 1$ - which has an impact on the initial accretion rate (Equation <ref>). Because the accretion timescale $t_{\mathrm{acc}}$ is a generalisation of $t_{\nu}$ in the MHD winds framework, the dependence of the two timescales on the stellar mass is exactly equivalent, and we will keep the same notation as above: $t_{\mathrm{acc}} \propto {M_{\star}}^{\mu}$. However, as mentioned above $\dot M_0$ depends on the stellar mass not only through $M_0$ and $t_{\mathrm{acc}, 0}$ as in the viscous case, but also through $f_{\mathrm{M},0} + 1$. As $f_{\mathrm{M},0} + 1 \propto {R_{c, 0}}^{\xi}$, and $R_{c, 0} \propto {M_{\star}}^{\zeta_0}$, the additional dependence will have a slope of $\zeta_0 \xi$. Therefore, in the MHD winds scenario we can link $\delta_0$ with $\mu_0$ as $\delta_0 = \mu_0 + \zeta_0 \xi$. The practical meaning of this difference is that, while in the viscous scenario the difference in slope between the two correlations depends only on the scaling of the viscous timescale with the stellar mass, in the hybrid scenario it depends also on the scaling between the disc radius and the stellar mass. It is important to note that $\xi$ is a small number, typically of the order of $\sim 0.1$, therefore the difference between the viscous and hybrid case is not particularly prominent. For evolved populations, the disc mass and accretion rate read [Tabone et al., 2022]
\begin{equation}
M_\mathrm{d} (t \gg t_{\mathrm{acc}}) \sim M_0 \left(\frac{t}{t_{\mathrm{acc}, 0}} \right)^{-(\psi + 2 \xi + 1)/2},
\label{eq:mdisc_mhd_omegazero_evo}
\end{equation}
\begin{equation}
\dot{M} (t \gg t_{\mathrm{acc}}) \sim \dot{M}_0 \left(\frac{t}{t_{\mathrm{acc}, 0}} \right)^{-(\psi + 4 \xi + 3)/2};
\label{eq:mdot_mhd_omegazero_evo}
\end{equation}
this brings the evolved slopes $\lambda_{\mathrm{m, evo}}$ and $\lambda_{\mathrm{acc, evo}}$ to
\begin{equation}
\begin{split}
\lambda_{\mathrm{m, evo}} = \lambda_{\mathrm{m}, 0} + \frac{1}{2} \mu_0 (\psi + 2\xi + 1), \\
\lambda_{\mathrm{acc, evo}} = \lambda_{\mathrm{acc}, 0} + \frac{1}{2} \mu_0 (\psi + 4\xi + 3),
\end{split}
\label{eq:lambda_evo_mhd}
\end{equation}
Time evolution of $M_{\mathrm{d}} - M_{\star}$ (left) and $\dot M - M_{\star}$ (right) in the pure wind model ($\omega \neq 0$). This plot is obtained with disc population synthesis modelling, without any spread in the initial conditions. Each dot represents a disc in the population at different ages as shown in the colour bar. The initial power-law correlation, shown in light blue, is lost as early as $\sim 1$ Myr (corresponding to $\sim 2 <t_{\mathrm{acc}, 0}>$ with these parameters) due to a downward bending corresponding to lower stellar masses. In this simulation, we have used $N = 100$ discs, $\alpha_{\mathrm{DW}} = 10^{-3}$, $\lambda = 3$, $\omega = 0.25$, $\lambda_{\mathrm{m}, 0} = 2.1$, $\lambda_{\mathrm{acc}, 0} = 1.5$. The set of MHD parameters is based on Tabone et al., 2022.
which reduces to Equation (<ref>) in the viscous case ($\psi \ll 1$, $\xi \ll 1$). Like viscosity, a hybrid secular evolution maintains the power-law shape of the correlation; moreover, Equation (<ref>) provides a theoretical prediction for the evolved slopes: they can steepen, flatten or remain the same as the initial conditions, depending on the involved parameters. As the terms in parentheses in Equation (<ref>) are sums of positive values, the sign of the evolved slopes depends on the sign of $\mu_0$ like in the viscous case. However, there is a difference from the viscous case: as in the hybrid scenario $\mu_0$ and $\delta_0$ do not coincide anymore, a constraint on the value of $\mu_0$ is translated into a constraint on $\delta_0 - \zeta_0 \xi$. In particular, the slopes will increase if $\delta_0 > \zeta_0 \xi$ (corresponding to $\mu_0 > 0$), whereas if $\delta_0 < \zeta_0 \xi$ (corresponding to $\mu_0 < 0$), the slopes will decrease; and finally, if $\delta_0 = \zeta_0 \xi$ (corresponding to $\mu_0 = 0$) the slopes will remain constant in time. Another difference from the viscous scenario is that $\lambda_{\mathrm{m, evo}}$ and $\lambda_{\mathrm{acc, evo}}$ are not expected to reach the same value anymore. The limit difference is given by $\delta_{\mathrm{evo}} = \lambda_{\mathrm{m, evo}} - \lambda_{\mathrm{acc, evo}}$: substituting the values from Equation (<ref>) one finds $\delta_{\mathrm{evo}} = \delta_0 - \mu_0 (\xi + 1)$, which can be further reduced to $\delta_{\mathrm{evo}} = \xi (2 \beta + \frac{1}{2})$ by using the definition of $\mu_0$ as the slope of the $t_{\mathrm{acc}, 0} - M_{\star}$ correlation. In this expression, $\beta$ is the slope of the correlation between the disc aspect ratio and the stellar mass, which comes from definition of the accretion timescale; with a standard $\beta = -1/2$ (a reasonable approximation of the value derived by radiative transfer simulation, see e.g., Sinclair et al., 2020), we obtain $\delta_{\mathrm{evo}} = -\xi/2$. As $\xi$ is positive by definition, in the hybrid scenario $\delta_{\mathrm{evo}}$ is always negative, meaning that in an evolved population, the correlation between the accretion rate and the stellar mass will necessarily be steeper than that between the disc mass and the stellar mass. However, we stress once more that $\xi$ is a small number and therefore the predicted difference is also small. Figure <ref> shows the evolution of the slopes from simulations (with no spread in the initial conditions) for the hybrid model (coloured) compared with the viscous case (grey), which matches the theoretical expectations discussed above. The list of parameters used in the simulations is available in Table <ref>.
Summarising, both the hybrid and viscous secular evolution preserve the power-law shape of the correlations between the disc properties and the stellar mass. The main difference is that the hybrid model does not predict the slopes of the $M_{\mathrm{d}} - M_{\star}$ and $\dot M - M_{\star}$ correlations to reach the same limit value (unlike the viscous case). The predicted difference in the evolved slopes is given by $\delta_{\mathrm{evo}} = -\xi/2$ (of the order of 0.1). The current observational uncertainties on $\delta_{\mathrm{evo}}$ range from 0.4 to 0.8 (Testi et al., 2022, see Table <ref>), 4 to 8 times larger than the predicted difference, making it not observable at this stage.
§.§ Pure wind - $\omega \neq 0$
The solution to the pure wind model (i.e., time-dependent $\alpha_{\mathrm{DW}}$ through $\alpha_{\mathrm{DW}}(t) \propto \Sigma_{\mathrm{c}}(t)^{- \omega}$) by [Tabone et al., 2022] gives
\begin{equation}
M_\mathrm{d} (t) = M_0 \left( 1 - \frac{\omega}{2 t_{\mathrm{acc}, 0}} t \right)^{1/\omega},
\label{eq:mdisc_mhd_omeganotzero}
\end{equation}
\begin{equation}
\dot{M} (t) = \frac{M_0}{2 t_{\mathrm{acc}, 0} (1+f_{\mathrm{M}, 0})} \left( 1 - \frac{\omega}{2 t_{\mathrm{acc}, 0}} t \right)^{-1+1/\omega};
\label{eq:mdot_mhd_omeganotzero}
\end{equation}
in this case, the functional form of Equation (<ref>) and (<ref>) does not allow us to derive a simple analytical expression for $M_\mathrm{d} (t \gg t_{\mathrm{acc}})$ and $\dot M (t \gg t_{\mathrm{acc}})$. Therefore, to explore the evolution of the correlations in the pure wind case, we fully rely on simulations. In order to account for the impact of secular evolution only, we input perfect correlations between the disc properties and the stellar mass - that is, we do not include any spread in the initial conditions. Figure <ref> shows the time evolution of $M_{\mathrm{d}}$ (left panel) and $\dot M$ (right panel) as a function of the stellar mass, from younger (darker) to older (lighter) populations. The input power-law correlation (light blue) corresponds to a line in the logarithmic plane; however, as early as $\sim 1$ Myr (corresponding to $\sim 2<t_{\mathrm{acc}, 0}>$ for this simulation), the input correlation starts to bend downwards at lower stellar masses. This behaviour reveals a significantly different trend from the viscous and hybrid model: in the pure wind scenario, the initial power-law shape of the correlations is not preserved by the secular evolution, but rather broken. In Figure <ref>, the faster evolution of discs around lower mass stars is the consequence of a positive $\mu_0$, implying a positive correlation between the stellar mass and the accretion timescale; a negative correlation between $M_{\star}$ and $t_{\mathrm{acc}, 0}$ (i.e., a negative $\mu_0$) would lead to faster evolution of discs around higher mass stars, causing the correlation to bend towards the other end of the stellar mass spectrum (see Figure <ref>). Another parameter that might impact the evolution of the correlation is the dispersal timescale, $t_{\mathrm{disp}}$; in the wind-driven case, $t_{\mathrm{disp}} \propto t_{\mathrm{acc}}$, therefore $t_{\mathrm{disp}}$ does not introduce any further dependence on the stellar mass.
§.§ Summary
In this Section we have discussed the time evolution of the correlations between the disc properties and the stellar mass from the theoretical point of view. We have performed analytical calculations for the viscous and hybrid scenario, finding that the power-law shape of the initial correlations is preserved during disc evolution. While the slopes are expected to evolve, increasing or decreasing depending on the initial conditions, the limit slopes for evolved populations differ in the two models. While viscosity requires the slope of the $M_{\mathrm{d}} - M_{\star}$ and $\dot M - M_{\star}$ correlations to reach the same value, in the hybrid scenario the two slopes always have different values; moreover, the evolved slope of $\dot M - M_{\star}$ is bound to be higher than that of $M_{\mathrm{d}} - M_{\star}$. Unfortunately however, the difference from the viscous model is a factor four too small to be observed within the current error bars. We have then confirmed our predictions running simulations without any spread in the initial conditions (Figure <ref>). In the pure wind case instead, the functional form of the solution to Equation (<ref>) does not allow for an analytical determination of the evolved slopes; therefore, we entirely relied on simulations. Our results show that, contrary to the viscous and hybrid case, wind-driven accretion does not maintain the power-law shape of the initial correlations: depending on the scaling of the accretion timescale with the stellar mass, the power-laws are broken because of a bending towards higher ($t_{\mathrm{acc}, 0}$ decreasing with $M_{\star}$) or lower ($t_{\mathrm{acc}, 0}$ increasing with $M_{\star}$) stellar masses.
§ POPULATION SYNTHESIS
In Section <ref> we have discussed analytical trends, and presented simulations with no spread to analyse the effect of secular disc evolution alone on the evolution of the slopes. In order to test our theoretical predictions against observational data, we need to account for both a spread in the initial conditions and disc dispersal mechanisms. In this Section, we discuss the impact of both factors on the $M_{\mathrm{d}} - M_{\star}$ and $\dot M - M_{\star}$ slopes and run realistic population synthesis simulations, to determine whether the model-dependent evolutionary features described in Section <ref> would be observable with the currently available data.
§.§ Effects of a spread in the initial conditions
Same as Figure <ref> with the addition of a spread in the initial correlations between the disc properties and the stellar mass ($\sigma_{M_{\mathrm{d}}}(0) = 0.65$ dex, $\sigma_{R_{\mathrm{d}}}(0) = 0.52$ dex). Despite the linear correlation being readily broken in theory (see Section <ref>), the scatter introduced by the spread in the initial conditions simulates the correlation also at more evolved ages. As an example, we show the fitted line at 5 Myr (the age of the oldest observed population) in both panels ($\log_{10}(M_{\mathrm{d}}/M_{\odot}) = 3.1 \log_{10}(M_{\star}/M_{\odot}) - 3.3$, $\log_{10}(\dot M/M_{\odot} \mathrm{yr}^{-1}) = 2.5 \log_{10}(M_{\star}/M_{\odot}) - 10.4$).
The introduction of an observationally-motivated spread in the initial conditions is crucial to produce realistic population synthesis model. In the purely viscous case, [Somigliana et al., 2022] have shown how the spread does not significantly impact the evolution of either the $M_{\mathrm{d}} - M_{\star}$ or $\dot M - M_{\star}$ correlations; the shape of the curves (grey in Figure <ref>) is unaffected, except for their starting point, and the statistical fluctuation - determined as the interval between the 25th and 75th percentile out of 100 realisations of numerical simulations with the same initial conditions - is of the order $\sim 0.1$ for both slopes. This is a factor two less than the smallest observational uncertainty, and therefore does not produce a detectable difference in the predicted results.
Following [Somigliana et al., 2022], we set $\sigma_{M_{\mathrm{d}}}(0) = 0.65$ dex and $\sigma_{R}(0) = 0.52$ dex (determined from Ansdell et al., 2017 and Testi et al., 2022) for the log-normal distributions of $M_{\mathrm{d}}(0)$ and $R_{\mathrm{d}}(0)$ in the hybrid scenario (with $\alpha_{\mathrm{SS}} = \alpha_{\mathrm{DW}} = 10^{-3}$ hence $\psi =1$, $\lambda = 3$, $\omega = 0$). We find that, just like in the purely viscous case, a spread in the initial conditions only shifts the starting point of the curves (coloured in Figure <ref>) and does not have any significant effect on the shape of the evolution of the slopes (see Figure 5-6-7 in Somigliana et al., 2022). The statistical fluctuation for both slopes is again of the order of 0.1, and therefore below the observational error and not impacting our predictions.
On the other hand, wind-driven models with increasing $\alpha_{\mathrm{DW}}$ in time and a spread in the initial conditions (Figure <ref>) behave quite differently from the theoretical expectation discussed in Section <ref>. As the spread introduces a stochastic component, the discs will have higher or lower masses and accretion rates with equal probability; the practical result for the initial correlations is that the bending towards lower stellar masses (approximately $\log_{10}(M_{\star}/M_{\odot}) < 0.5$ in Figure <ref>) is lost to the stochastic displacement of the discs in the $M_{\mathrm{d}} - M_{\star}$ and $\dot M - M_{\star}$ planes. Actually, the resulting distribution of discs both in the $M_{\mathrm{d}} - M_{\star}$ and $\dot M - M_{\star}$ log-log plane does simulate a linear correlation; this implies that, while the stellar mass and the disc properties should not exhibit a linear correlation already after a few $t_{\mathrm{acc}}$, the presence of a spread mimics such correlation, making the wind-driven scenario indistinguishable from the viscous and hybrid ones. The one feature that remains observable, despite the presence of a spread, is the removal of discs around more or less massive stars - depending on the value of $\mu_0$, as discussed in Section <ref>; in the simulation shown in Figure <ref> we have set $\mu_0 > 0$, implying that discs around less massive stars evolve more rapidly and are therefore more readily dispersed, as is visualised by the lack of sources around lower stellar masses at evolved ages.
Summarising, an initial power-law correlation between the disc properties and the stellar mass would keep its power-law shape under wind-driven evolution, like in the viscous or hybrid case; however, the interpretation of the observed correlations is different depending on the theoretical framework. While the viscous and hybrid models preserve an initially established correlation, making the characteristic evolution of the slopes $\lambda_{\mathrm{m}}$ and $\lambda_{\mathrm{acc}}$ a tracer of disc evolution itself, the apparent correlation observed in wind-driven populations is merely a signature of the initial spread, rather than the evolutionary mechanism at play.
Time evolution of the slope of the $M_{\mathrm{d}} - M_{\star}$ correlation for 15 statistical realisations in the purely viscous (left, $\alpha_{\mathrm{SS}} = 10^{-3}$), photoevaporative (centre, $\dot M_w = 4 \times 10^{-10} \mathrm{M}_{\odot}/\mathrm{yr}$, $\alpha_{\mathrm{SS}} = 10^{-3}$) and wind-driven (right, $\omega = 0.25$, $\lambda = 3$, $\alpha_{\mathrm{DW}} = 10^{-3}$) model. The dashed lines show the median evolution, while the shaded area represents the interval between the 25th and the 75th percentiles. The three rows show different sample sizes, increasing from top to bottom. The initial size of each population was chosen to obtain a certain number of discs at 5 Myr (the age of the oldest observed population, Upper Sco) with the different disc fractions. In the top row, we match the current size of the Upper Sco sample($\sim 30$ objects, Testi et al., 2022) with both accretion rate and disc mass measurements; the middle row shows double the current sample size ($\sim 60$ objects), while the bottom row assumes a complete sample ($\sim 140$ objects). While the viscous model produces a remarkably similar evolution for all simulations, the latter two show stochastic oscillations from one realisation to another, suggesting that disc dispersal impacts the observed slope more than the evolutionary model does - at least with the currently available sample sizes; increasing the number of sources significantly mitigates the oscillations. The slope of the $\dot M - M_{\star}$ correlation behaves the same way.
§.§ Accounting for disc dispersal: internal photoevaporation
Out of the three theoretical scenarios discussed so far, wind-driven evolution is the only one that manages to reproduce the disc and accretion fraction (as measured by Hernández et al., 2007 and Fedele et al., 2010 respectively) - and therefore, the only one whose predictions can reasonably be compared with observations. Traditionally, the problem of disc dispersal in viscous populations is addressed by including internal photoevaporation (see e.g., Hollenbach et al., 1994, Clarke et al., 2001, Alexander et al., 2006, Alexander et al., 2006): in this Section, we discuss the impact of internal photoevaporation on the previously described expectations for the evolution of the slopes in the purely viscous scenario. As the lack of analytical solutions to the general equation (<ref>) does not allow for analytical arguments, we base the following discussion on physical considerations.
Internal photoevaporation is a threshold process, that kicks in after the accretion rate drops below the photoevaporative mass-loss rate [Clarke et al., 2001]. The moment where the effect of photoevaporation becomes non-negligible depends therefore on the initial accretion rate: assuming for simplicity a fixed photoevaporation rate for the whole population, as the accretion rate scales positively with the stellar mass we can expect discs around lower mass stars to show the effects of photoevaporation earlier. Moreover, given that the disc mass also scales positively with the stellar mass, said sources correspond to the less massive ones in the population. From these considerations, we can expect discs with lower initial mass to be the first ones to be affected by photoevaporation, causing a steepening of the $M_{\mathrm{d}} - M_{\star}$ and $\dot M - M_{\star}$ correlations. However, photoevaporation also disperses discs: removing sources from the population may alter the expected behaviour, hence the need to perform numerical simulations to understand the evolution of the correlations for a population of discs undergoing internal photoevaporation. Our simulations remove discs either because the photoevaporative gap becomes too large, or because the disc mass or accretion rate fall below a certain detectability threshold. As we mentioned above, the dispersal timescale $t_{\mathrm{disp}}$ might also play a role, if it has a different scaling with the stellar mass with respect to the accretion timescale ($t_{\nu}$ in the viscous case). Our numerical implementation of internal photoevaporation follows [Owen et al., 2010], and we assume a mass-loss rate of $4 \times 10^{-10}$ M$_{\odot}$ yr$^{-1}$ - which allows us to reproduce the observed disc fraction for the set of parameters of our simulation - for all discs in the population; therefore, no further $M_{\star}$ dependence is introduced, and the $t_{\nu} - M_{\star}$ scaling is the only one that matters. We stress that a stellar mass-dependent photoevaporative rate is expected [Picogna et al., 2021]: we explore the influence of such dependence on disc observables in an upcoming work (Malanga et al. in prep.). We discuss the results of our simulations in the following Section.
§.§ What are the slopes tracing?
Comparison of the time evolution of $\lambda_{\mathrm{m}}$ (top) and $\lambda_{\mathrm{acc}}$ (bottom) between the viscous+photoevaporative and wind-driven case including measured slopes from four star-forming regions. To account for statistical fluctuations, each simulation combines 100 realisations of the same initial conditions: the lines show the median evolution, while the shaded area represent the interval between the 25th and 75th percentiles. The simulations in the three columns differ for the initial number of discs, determined to obtain a specific sample size at 5 Myr - currently available sample (left), double the currently available sample (centre) and the complete sample (right). Observed slopes from [Testi et al., 2022].
Figure <ref> shows the time evolution of the $M_{\mathrm{d}} - M_{\star}$ slope for 15 realisations with the same initial conditions for the viscous plus photoevaporative (central panel) and wind-driven (right panel) models, both reproducing the observed disc and accretion fractions, compared to the purely viscous scenario (left panel). The initial number of discs in the populations was determined to recover a certain number of objects at 5 Myr (increasing from top to bottom), and varies in the different simulations, as the decline (or lack thereof) of the disc fraction is model-dependent (see Figure 6 of Somigliana et al., 2023).
The number of objects in the simulation displayed in the first row was set to obtain $\sim$ 30 discs at 5 Myr, corresponding to the currently available sample size in Upper Sco, the oldest observed star-forming region. In the left panel, we see how the evolution of the slope in the purely viscous model is not significantly affected by a spread in the initial conditions: the single realisations resemble each other remarkably well, the only difference being the starting point of the curve (as found by Somigliana et al., 2022). On the other hand, the photoevaporative and wind-driven models have a dissimilar behaviour: each realisation can deviate substantially from the others, as we can particularly notice by the location, amplitude and direction of the bumps. The key difference between these models and the purely viscous case is disc dispersal: the stochastic nature of the slope evolution suggests that it does not trace the underlying secular disc evolution, like in the viscous scenario, but rather carry the signatures of disc dispersal itself - making it impossible to use the evolution of the slopes as a proxy for accretion mechanisms. There are two main factor that play a role in this context:
* Initial conditions and spread in the correlations. The exact evolution of the slopes will depend on the initial conditions, both of the disc mass-stellar mass correlations themselves and of the population-wide parameters. Furthermore, the removal of discs from the population would not impact the results of the fitting procedure only if there was perfect correlations between the disc parameters and the stellar mass; with a spread in the initial correlations, on the other hand, the results may differ depending on which discs in the population are dispersed;
* Small number statistics. Depending on the initial number of objects, disc removal can lead to small samples - so small that it might lead to low number statistics issues. This is the case for the top row of Figure <ref>, where the number of objects at 10 Myr is of the order of 10 or lower.
In this work, we have used one specific set of parameters (summarised in Table <ref>), determined following [Somigliana et al., 2022] (viscous model) and [Tabone et al., 2022] (hybrid and wind-driven), and we leave a deeper exploration of the parameters space to a future work. While the exact shape of the slope evolution, and therefore the accretion model signature, might depend on the initial conditions, the top panel of Figure <ref> shows that with the current sample sizes the noise dominates over the physical evolution. However, the currently available sample of the oldest star-forming region (Upper Sco) with both disc masses (derived from the millimetre flux, Barenfeld et al., 2016) and accretion rate [Manara et al., 2020] estimates is highly incomplete; it is therefore worth investigating whether a higher level of completeness would help reducing the entity of the oscillations, allowing to disentangle between the different evolutionary models.
The central and bottom rows of Figure <ref> show how a larger sample would impact the oscillations of the slope evolution. The simulations in the middle row are performed imposing a double sample size at 5 Myr with respect to the current one ($\sim 60$ discs), while in the bottom row we assume to have the complete Upper Sco sample, totalling $\sim 140$ discs (Carpenter et al. in prep.). We remind that we focus on Upper Sco as the oldest observed star-forming region, which makes it the most affected by disc dispersal.
As expected, statistical significance increases with a larger sample, leading to a decreased impact of the oscillations on the global slope evolution; with the complete sample, in particular, we can reduce the spread in the evolution by a factor of $\sim 2$ compared to the current available data. This argument confirms the importance of larger sample sizes in discriminating between the viscous and wind-driven models, as already suggested by [Alexander et al., 2023] in the context of the accretion rates distribution.
As we mentioned in Section <ref>, our simulations consider discs as dispersed if their masses or accretion rates fall below the imposed detectability threshold of $10^{-12}$ M$_{\odot}$ yr$^{-1}$. We have also included a threshold in accretion rates in post-processing to account for non-accreting objects. From the observational point of view, this latter selection depends on both the instrumental sensitivity and the definition of disc itself: how Class II objects are defined, and in turn how Class III sources are removed from the observed samples, impacts the resulting slope. Summarising, with the current sample sizes, the evolution of the slopes is significantly more affected by disc dispersal than it is by secular evolution; therefore, at the state of the art, it cannot be used as a proxy to disentangle between the different evolutionary models. Increasing the sample size would allow to reduce the effect of low number statistics, potentially allowing to observe the different evolution of the slopes under the two evolutionary mechanisms; we further discuss this possibility in the following Section.
H]|P1cm | P1cm P2cm | P1cm P1.9cm | P1.6cm P5.2cm | P0.2cm|
8c $\mu_0 = \delta_0 - \zeta_0 \xi$
1-7 red8-8black
$\xi$ $\mu_0$ $t_{\mathrm{acc}, 0} - M_{\star}$ $\zeta_0$ $R_{\mathrm{d}} - M_{\star}$ $\delta_0$ 1c!$M_{\mathrm{d}}-M{\star}$ (a) and $\dot M-M{\star}$ (b) 1c!$\Rightarrow$
3*0 bg_visc_light $ < 0 $ bg_visc_light $t_{\mathrm{acc}, 0} \downarrow M_{\star} \uparrow$ 2c|bg_visc_light any bg_visc_light $ < 0 $ 1c!(a) shallower than (b) bg_visc_light 1c!$=$ bg_visc_light
bg_visc $0$ bg_visc $t_{\mathrm{acc}, 0} \leftrightarrow M_{\star} \uparrow$ 2c|bg_visc any bg_visc 0 1c!(a) as steep as (b) bg_visc 1c!$=$ bg_visc
bg_visc_light $ > 0 $ bg_visc_light $t_{\mathrm{acc}, 0} \uparrow M_{\star} \uparrow$ 2c|bg_visc_light any bg_visc_light $ > 0 $ 1c!(a) steeper than (b)bg_visc_light 1c!$=$ bg_visc_light
bg_mhd_light bg_mhd_light bg_mhd_light $ \leq 0 $ bg_mhd_light $R_{\mathrm{d}} \downarrow \leftrightarrow M_{\star} \uparrow$ bg_mhd_light $ \leq 0 $ 1c! (a) shallower or as steep as (b) bg_mhd_light 1c! $=$ bg_mhd_light
bg_mhd_light -2*$ < 0 $ bg_mhd_light -2*$t_{\mathrm{acc}, 0} \downarrow M_{\star} \uparrow$ bg_mhd_light $ > 0 $ bg_mhd_light $R_{\mathrm{d}} \uparrow M_{\star} \uparrow$ bg_mhd_light $ (- \infty, \zeta_0 \xi) $ 1c!(a) shallower or steeper than (b)[ $\zeta_0 \xi $ is positive, therefore $\delta_0$ can either be negative (implying (a) shallower than (b)) or positive (implying (a) steeper than (b)).] bg_mhd_light 1c!$\neq$ bg_mhd_light
bg_mhd bg_mhd bg_mhd $ \leq 0 $ bg_mhd $R_{\mathrm{d}} \downarrow \leftrightarrow M_{\star} \uparrow$ bg_mhd $ \leq 0 $ 1c! (a) shallower or as steep as (b) bg_mhd 1c!$\circ$ bg_mhd
bg_mhd -2*$ 0 $ bg_mhd -2*$t_{\mathrm{acc}, 0} \leftrightarrow M_{\star} \uparrow$ bg_mhd $ > 0 $ bg_mhd $R_{\mathrm{d}} \uparrow M_{\star} \uparrow$ bg_mhd $ \geq 0 $ 1c!(a) steeper or as steep as (b) bg_mhd 1c!$=$ bg_mhd
bg_mhd_light bg_mhd_light bg_mhd_light $ \leq 0 $ bg_mhd_light $R_{\mathrm{d}} \downarrow \leftrightarrow M_{\star} \uparrow$ bg_mhd_light $\left(\zeta_0 \xi, +\infty \right)$ 1c!(a) shallower or steeper than (b)[ $\zeta_0 \xi $ is negative, therefore the same argument as in <ref> holds.] bg_mhd_light 1c! $\circ$ bg_mhd_light
-6*(0, 1) bg_mhd_light -2*$ > 0 $ bg_mhd_light -2*$t_{\mathrm{acc}, 0} \uparrow M_{\star} \uparrow$ bg_mhd_light $ > 0 $ bg_mhd_light $R_{\mathrm{d}} \uparrow M_{\star} \uparrow$ bg_mhd_light $\left(\zeta_0 \xi, +\infty \right)$ 1c!(a) steeper than (b) bg_mhd_light 1c!$=$ bg_mhd_light
1-7 red8-8black
H]|P1.6cm P5cm|
2c $\delta_{\mathrm{evo}} = \xi ( \zeta_0 - \mu_0), \quad \xi \neq 0$
$\delta_{\mathrm{evo}}$ $M_{\mathrm{d}}-M{\star}$ (a) and $\dot M-M{\star}$ (b)
< 0 (a) shallower than (b)
bg_2 0 bg_2 (a) as steep as (b)
> 0 (a) steeper than (b)
Summary of the different possible theoretical scenarios described in Section <ref> to visualise the relative signs of the parameters at play. From left to right in the top table, the columns show (i) $\xi$, a proxy for the evolutionary model (viscous if $\xi = 0$, hybrid or wind-driven otherwise); (ii) $\mu_0$, the slope of the $t_{\mathrm{acc}, 0} - M_{\star}$ correlation and its implication on the correlation itself; (iii) $\zeta_0$, the slope of the $R_{\mathrm{d}}-M_{\star}$ correlation, and its implication on the correlation itself; (iv) $\delta_0 = \lambda_{\mathrm{m}, 0} - \lambda_{\mathrm{acc}, 0}$, the difference between the initial slopes of the $M_{\mathrm{d}}-M_{\star}$ and $\dot M - M_{\star}$ correlations and its implication on their relative steepness. The final column summarises whether the signs of $\mu_0$ and $\delta_0$ are necessarily the same ($=$), necessarily opposite ($\neq$) or can be either ($\circ$). When discussing the implications on correlations, up(down)wards arrows represent an in(de)crease of the parameters, while horizontal arrows describes the lack of correlation. The different cell colours are purely meant to guide the eye. The top table links the initial conditions, while the bottom table summarises the implications of the evolved difference in the slopes.
§ THE OBSERVATIONAL RELEVANCE OF THE SLOPES
Observed star-forming regions have both a spread in the initial conditions in addition to some disc dispersal mechanism (be it photoevaporation or MHD winds); as we discussed in Section <ref>, with the current sample sizes, the statistical significance of the observationally-determined slopes is undermined and their evolution traces disc dispersal, rather than the accretion mechanism. In this Section, we perform a statistical analysis of our simulated slopes and compare them with the currently available measurements; furthermore, we show the relevance of measuring the slopes despite these limitations and discuss the conditions under which they allow us to put constraints on disc evolution.
§.§ Comparison of different evolutionary models
Figure <ref> shows the comparison of the evolution of the slopes between the viscous + photoevaporative (solid line) and wind-driven (dashed line) models for both the $M_{\mathrm{d}}-M_{\star}$ and $\dot M - M_{\star}$ correlations (top and bottom row, respectively), including the measured slopes in four star-forming regions from [Testi et al., 2022] as grey dots. As we mentioned above, both models lead to disc dispersal consistently with the observed disc and accretion fraction (shown in Figure 6 of Somigliana et al., 2023); the three columns show simulations performed with a different initial number of discs, increasing from left to right, to obtain a different sample size at 5 Myr according to the predicted decline of observed discs. Like in Figure <ref>, the number of objects at 5 Myr is $\sim 30$, $\sim 60$ and $\sim 140$ from left to right, increasing from the currently available measurements in Upper Sco to the virtually complete sample. To estimate the effect of statistical fluctuations, given by the spread in the initial conditions, we ran 100 simulations for each set up: the solid and dashed lines represent the median evolution, while the intervals between the 25th and 75th percentile are visualised by the shaded areas. The growing shaded area, particularly visible with smaller sample sizes, is representative of the decreasing amount of sources on which the fit is performed: with the current sample size (left column), which leads to $\sim 30$ discs at 5 Myr, we end up with a 1$\sigma$ deviation from the median value of $\sim 0.5 - 0.6$. Larger sample sizes significantly reduce the scatter, leading to $\sigma \sim 0.4$ with a double sample and $\sigma \sim 0.2$ for the complete sample, reducing the current one by a factor 3. As mentioned in Section <ref>, with the currently available number of objects the dominant role in the evolution of the slopes is played by disc dispersal, which makes it difficult to trace the imprint of the secular evolution. The expected Upper Sco complete sample (right column) allows for a better separation between the two models - particularly for the $M_{\mathrm{d}}-M_{\star}$ correlation: the expected slope in the two scenarios differs by $\sim 0.5$, while the typical uncertainty of the currently measured slopes is between 0.2 and 0.3. Larger sample sizes would further decrease this uncertainty, allowing us to discriminate between the two models based on the slope evolution.
The observed slopes (from Testi et al., 2022) are only included in the left column of Figure <ref> as they refer to the current sample size. The main source of uncertainty in the current measurements is given by Upper Sco, mainly due to the incomplete sample; moreover, it is worth pointing out that external photoevaporation is likely to play a significant role in this region (Anania et al. in prep.). This comparison is meant as a first glance of the parameters space of the observed slopes, and we anticipate a proper exploration of the initial conditions once the full sample will be available.
In the following Section, we discuss the other constraints that we can put on disc evolution, besides identifying the driving accretion mechanism.
§.§ What are the slopes telling us, then?
Despite not allowing to conclusively discriminate between different evolutionary scenarios with the current sample sizes, the slopes of the $M_{\mathrm{d}} - M_{\star}$ and $\dot M - M_{\star}$ correlations can still help with constraining other properties from the theoretical considerations presented in Section <ref>, which we summarise in Table <ref>. If we assume an evolutionary model to begin with, and we can estimate (directly or indirectly) either $\mu_0$ or $\delta_0$, we can constrain the other parameter. When discussing the observational determination of what we have so far referred to as initial conditions, it is important to clarify the meaning of "initial". deals with and evolves Class II, potentially Class III, objects; hence, the initial conditions we input refer to the beginning of the Class II phase, where the protostellar collapse is over and the disc is already formed. From the observational point of view, this means that we expect $\delta_0$ and $\mu_0$ to refer to young Class II objects - around, or younger than, approximately 1 Myr. Earlier phases like the Class 0 and I need a dedicated study, as the accretion of the protostellar envelope is expected to play a prominent role in those stages.
In the following, we discuss the constraints we can put in both directions and comment on their feasibility based on the currently available estimates of $\mu_0$ and $\delta_0$.
§.§.§ Constraining delta0 from mu0
[Ansdell et al., 2017] claimed $\lambda_{\mathrm{m}}$ to be increasing with time. As shown by [Somigliana et al., 2022] and discussed in Section <ref>, increasing slopes imply that discs around less massive stars evolve faster than those around more massive stars; this can be interpreted in terms of increasing accretion timescale with stellar mass, which corresponds to $\mu_0 > 0$ (with $t_{\mathrm{acc}, 0} \propto {M_{\star}}^{\mu_0}$). In this section we discuss the implications of the increasing slope scenario on the initial conditions, $\lambda_{\mathrm{m}, 0}$ and $\lambda_{\mathrm{acc}, 0}$.
The top panel of Table <ref> shows the relation between $\mu_0$ and $\delta_0$ in the different evolutionary models. As $\mu_0 = \delta_0 - \zeta_0 \xi$, in the viscous case (corresponding to $\xi = 0$) we have $\mu_0 = \delta_0$ as mentioned in Section <ref>. This means that, to recover the suggested increasing slopes scenario, $\delta_0$ necessarily needs to be positive - regardless of the value of any other parameters: this translates to the initial $M_{\mathrm{d}} - M_{\star}$ correlation being steeper than $\dot M - M_{\star}$. In the hybrid and wind-driven case, instead, the implication is less straightforward as a positive $\mu_0$ can lead to opposite signs of $\delta_0$: this is determined by the scaling of the disc radius with the stellar mass, which is suggested to be (weakly) positive from observational evidences (e.g., Andrews et al., 2018). In principle, as $\delta_0 \in (\zeta_0 \xi, + \infty)$, the sign of $\zeta_0$ determines whether negative values of $\delta_0$ are possible; however, as $\xi$ is a small number (0.1 in this work), only a limited area of the parameters space would lead to a negative $\delta_0$. Summarising, if we assume increasing slopes ($\mu_0 > 0$) we can constrain the sign of $\delta_0$ regardless of the evolutionary model assumed: in both cases $\delta_0$ needs to be positive, which leads to an initially steeper $M_{\mathrm{d}} - M_{\star}$ than $\dot M - M_{\star}$ correlation.
§.§.§ Constraining mu0 from deltaevo
Instead of assuming increasing slopes, we can start from the currently measured values of $\delta_{\mathrm{evo}}$ and estimate $\delta_0$ in the different evolutionary models. In the viscous case, because $\delta_{\mathrm{evo}} = 0$, we focus on the value of the single slopes instead: as $\lambda_{\mathrm{m, evo}} = \lambda_{\mathrm{acc, evo}} = \delta_0/2 + \lambda_{\mathrm{m}, 0}$, the measured final value of the slopes does not help in constraining $\delta_0$ as it also depends on $\lambda_{\mathrm{m}, 0}$. In the hybrid case, instead, we have $\delta_{\mathrm{evo}} = \xi(\zeta_0 - \mu_0)$, meaning that if we can determine the sign of $\delta_{\mathrm{evo}}$ we can constrain that of $\mu_0$ as well. While in principle the sign of $\zeta_0$ influences that of $\mu_0$, as we mentioned above $\zeta_0$ is likely a small number: therefore, effectively, $\delta_{\mathrm{evo}}$ and $\mu_0$ have opposite signs for the vast majority of the parameters space.
Assuming that the observed disc populations can be considered evolved enough for the above arguments to hold, we can estimate $\delta_{\mathrm{evo}}$ from the most recent and homogeneous measurements available of $\lambda_{\mathrm{m}}$ and $\lambda_{\mathrm{acc}}$ [Testi et al., 2022]. The resulting values of $\delta_{\mathrm{obs}}$ (which we label 'observed' as opposed to the theoretical expectation, 'evolved'), summarised in Table <ref>, are oscillating: out of the four regions L1668, Lupus, Chameleon I and Upper Sco, we find two positive and two negative median values. Moreover, in three cases out of four the uncertainties are so large that $\delta_{\mathrm{obs}}$ would be compatible with both a positive and a negative value. The difficulty in assessing the sign of $\delta_{\mathrm{obs}}$ from the current measurements of the slopes make constraining $\mu_0$ from $\delta_{\mathrm{evo}}$ not trivial. Larger sample sizes would give a better measurement of the slopes and reduce the uncertainty, leading to a more solid determination of the sign of $\delta_{\mathrm{obs}}$ - which would possibly allow to constrain $\mu_0$.
Summarising, the (admittedly not robust) observational evidence pointing towards increasing accretion timescale with stellar mass allows us to constrain the initial correlations between the stellar mass and disc parameters; regardless of the evolutionary model considered, the initial slope of the $M_{\mathrm{d}}-M_{\star}$ correlation needs to be larger than that of $\dot M - M_{\star}$. The other way around, constraining the slope of the accretion timescale - stellar mass correlation from the difference between $\lambda_{\mathrm{m}}$ and $\lambda_{\mathrm{acc}}$ at the present time, requires sample sizes larger by at least a factor two.
H]|P1.5cm | P3.5cm | P1.5cm |
Region Median age [Myr] $\delta_{\mathrm{obs}}$
L1668 1 $-0.3 \pm 0.5$
Lupus 2 $0.1 \pm 0.4$
Cha I 2.8 $-0.7 \pm 0.4$
Upper Sco 4.3 $0.7 \pm 0.8$
Values of $\delta$ derived from the currently available measurements of $\lambda_{\mathrm{m}}$ and $\lambda_{\mathrm{acc}}$ [Testi et al., 2022].
§ CONCLUSIONS
In this paper, we have investigated the impact of disc evolution models on the correlations between the stellar mass and the disc properties - especially the disc mass and the accretion rate. We have explored the purely viscous, wind-driven, viscous and wind hybrid, and photoevaporative models. Assuming power-law correlations to hold as initial conditions, $M_{\mathrm{d}}(0) \propto {M_{\star}}^{\lambda_{\mathrm{m}, 0}}$, $\dot M(0) \propto {M_{\star}}^{\lambda_{\mathrm{acc}, 0}}$, we performed analytical calculations (where possible) and population synthesis simulations for both evolutionary scenarios, and compared them with the purely viscous case discussed in [Somigliana et al., 2022]. Our main results are the following:
* The viscous and hybrid models change the slope of the initial correlations as function of the evolutionary time, but preserve their shape. In the wind-driven model, instead, the correlations deviate from the original power-law shape: this is visualised in the logarithmic plane as a bending of the linear correlation (see Figure <ref>). The bending direction is towards the less or more massive stars depending on the scaling of the accretion timescale with the stellar mass (positive and negative correlation respectively).
* The characteristic behaviour of the slopes in the wind-driven model is concealed by the presence of a spread in the initial conditions, which introduces a scatter in the correlations and makes it no longer possible to detect the bending (Figure <ref>). This leads to a considerably similar evolution of the correlations in the different accretion models.
* Performing our simulations with evolutionary models that match the disc dispersal timescales (intrinsic in the wind-driven model and including internal photoevaporation in the viscous case), we find that the evolution of the slopes is significantly impacted by the removal of discs from the population (Figure <ref>). Different realisations of the same simulation dramatically differ from one another, and show a stochastic behaviour with large variations (Figure <ref>). This has both a physical (presence of a spread in the initial conditions) and a statistical (low number of objects left after a few Myr of evolution) reason.
* While a proper exploration of the parameters space, outside of the scope of this work, would be needed to assess the impact of the initial conditions, with the currently available sample sizes the noise dominates over the physical evolution.
* Increasing the sample size can mitigate the effects of disc dispersal on the evolution of the slope by removing the stochastic effects. We find that, for our parameters choice, the complete sample of Upper Sco ($\sim 140$ sources) at 5 Myr would reduce the oscillations enough to make the slopes a proxy for the evolutionary model (Figure <ref>).
* While the currently available sample sizes do not yet allow to distinguish between the different evolutionary models, we can use them to put some constraints on the initial conditions. We find that in all evolutionary scenarios, the observational claim of increasing slopes leads to an initially steeper correlation between the disc mass and the stellar mass than between the accretion rate and the stellar mass. The other possible way, measuring the current slopes and inferring the correlation between the accretion timescale and the stellar mass from them, provides weaker constraints because of the high uncertainties in the current measurements.
* We have presented and released the 1D Python disc population synthesis code and its output analysis library .
In this work, we have shown how large enough samples of protoplanetary discs can provide a way of distinguishing between the evolutionary models (with a standard set of parameters) through the observation of the time evolution of the correlations between the disc properties and the stellar mass. We have shown how the stochastic fluctuations seen with the currently available observations could be significantly reduced if we had access to the complete Upper Sco sample, consisting of approximately 140 sources at 5 Myr. We strongly support the observational effort in the direction of obtaining larger amounts of data for evolved star-forming regions, and encourage the exploration of the parameters space beyond the standard case.
We thank an anonymous referee for their useful comments that helped us improve the clarity of the manuscript. This work was partly supported by the Italian Ministero dell’Istruzione, Università e Ricerca through the grant Progetti Premiali 2012-iALMA (CUP C52I13000140001), by the Deutsche Forschungsgemeinschaft (DFG, German Research Foundation) - Ref no. 325594231 FOR 2634/2 TE 1024/2-1, by the DFG Cluster of Excellence Origins (www.origins-cluster.de). This project has received funding from the European Union’s Horizon 2020 research and innovation program under the Marie Sklodowska- Curie grant agreement No 823823 (DUSTBUSTERS) and from the European Research Council (ERC) via the ERC Synergy Grant ECOGAL (grant 855130) and ERC Starting Grant DiscEvol (grant 101039651). GR acknowledges support from Fondazione Cariplo, grant No. 2022-1217. Views and opinions expressed are however those of the authors only and do not necessarily reflect those of the European Union or the European Research Council Executive Agency. Neither the European Union nor the granting authority can be held responsible for them.
[Alcalá et al., 2017]
Alcalá, J. M., Manara, C. F., Natta, A., et al. 2017, , 600,
[Alcalá et al., 2014]
Alcalá, J. M., Natta, A., Manara, C. F., et al. 2014, , 561, A2
[Alexander et al., 2014]
Alexander, R., Pascucci, I., Andrews, S., Armitage, P., & Cieza, L.
2014, in Protostars and Planets VI, ed. H. Beuther, R. S. Klessen, C. P.
Dullemond, & T. Henning, 475–496
[Alexander et al., 2023]
Alexander, R., Rosotti, G., Armitage, P. J., et al. 2023, , 524,
[Alexander & Armitage, 2006]
Alexander, R. D. & Armitage, P. J. 2006, , 639, L83
[Alexander et al., 2006]
Alexander, R. D., Clarke, C. J., & Pringle, J. E. 2006a,
, 369, 216
[Alexander et al., 2006]
Alexander, R. D., Clarke, C. J., & Pringle, J. E. 2006b,
, 369, 229
[Anderson et al., 2022]
Anderson, D. E., Cleeves, L. I., Blake, G. A., et al. 2022, , 927,
[Andrews et al., 2018]
Andrews, S. M., Terrell, M., Tripathi, A., et al. 2018, , 865, 157
[Ansdell et al., 2017]
Ansdell, M., Williams, J. P., Manara, C. F., et al. 2017, , 153, 240
[Ansdell et al., 2016]
Ansdell, M., Williams, J. P., van der Marel, N., et al. 2016, ,
828, 46
[Armitage et al., 2013]
Armitage, P. J., Simon, J. B., & Martin, R. G. 2013, , 778, L14
[Barenfeld et al., 2016]
Barenfeld, S. A., Carpenter, J. M., Ricci, L., & Isella, A. 2016,
, 827, 142
[Bergin et al., 2013]
Bergin, E. A., Cleeves, L. I., Gorti, U., et al. 2013, , 493, 644
[Béthune et al., 2017]
Béthune, W., Lesur, G., & Ferreira, J. 2017, , 600, A75
[Birnstiel et al., 2010]
Birnstiel, T., Dullemond, C. P., & Brauer, F. 2010, , 513, A79
[Birnstiel et al., 2012]
Birnstiel, T., Klahr, H., & Ercolano, B. 2012, , 539, A148
[Blandford & Payne, 1982]
Blandford, R. D. & Payne, D. G. 1982, , 199, 883
[Booth et al., 2017]
Booth, R. A., Clarke, C. J., Madhusudhan, N., & Ilee, J. D. 2017,
, 469, 3994
[Clarke et al., 2001]
Clarke, C. J., Gendrin, A., & Sotomayor, M. 2001, , 328, 485
[Coleman et al., 2024]
Coleman, G. A. L., Mroueh, J. K., & Haworth, T. J. 2024, , 527,
[Dullemond et al., 2006]
Dullemond, C. P., Natta, A., & Testi, L. 2006, , 645, L69
[Fedele et al., 2010]
Fedele, D., van den Ancker, M. E., Henning, T., Jayawardhana, R., &
Oliveira, J. M. 2010, , 510, A72
[Ferreira, 1997]
Ferreira, J. 1997, , 319, 340
[Flaherty et al., 2018]
Flaherty, K. M., Hughes, A. M., Teague, R., et al. 2018, , 856, 117
[Habing, 1968]
Habing, H. J. 1968, , 19, 421
[Hartmann et al., 1998]
Hartmann, L., Calvet, N., Gullbring, E., & D'Alessio, P. 1998, ,
495, 385
[Haworth et al., 2023]
Haworth, T. J., Coleman, G. A. L., Qiao, L., Sellek, A. D., &
Askari, K. 2023, , 526, 4315
[Hennebelle et al., 2020]
Hennebelle, P., Commerçon, B., Lee, Y.-N., & Charnoz, S. 2020,
, 635, A67
[Herczeg & Hillenbrand, 2008]
Herczeg, G. J. & Hillenbrand, L. A. 2008, , 681, 594
[Hernández et al., 2007]
Hernández, J., Hartmann, L., Megeath, T., et al. 2007, , 662,
[Hollenbach et al., 1994]
Hollenbach, D., Johnstone, D., Lizano, S., & Shu, F. 1994, , 428,
[Kley & Lin, 1996]
Kley, W. & Lin, D. N. C. 1996, , 461, 933
[Kroupa, 2001]
Kroupa, P. 2001, , 322, 231
[Laibe & Price, 2014]
Laibe, G. & Price, D. J. 2014, , 444, 1940
[Lebreuilly et al., 2021]
Lebreuilly, U., Hennebelle, P., Colman, T., et al. 2021, , 917,
[Lebreuilly et al., 2024]
Lebreuilly, U., Hennebelle, P., Colman, T., et al. 2024, , 682, A30
[Lodato et al., 2023]
Lodato, G., Rampinelli, L., Viscardi, E., et al. 2023, , 518,
[Lodato et al., 2017]
Lodato, G., Scardoni, C. E., Manara, C. F., & Testi, L. 2017, ,
472, 4700
[Long et al., 2022]
Long, F., Andrews, S. M., Rosotti, G., et al. 2022, , 931, 6
[Lynden-Bell & Pringle, 1974]
Lynden-Bell, D. & Pringle, J. E. 1974, , 168, 603
[Machida & Basu, 2024]
Machida, M. N. & Basu, S. 2024, arXiv e-prints, arXiv:2405.08271
[Machida & Hosokawa, 2013]
Machida, M. N. & Hosokawa, T. 2013, , 431, 1719
[Manara et al., 2023]
Manara, C. F., Ansdell, M., Rosotti, G. P., et al. 2023, in
Astronomical Society of the Pacific Conference Series, Vol. 534, Protostars
and Planets VII, ed. S. Inutsuka, Y. Aikawa, T. Muto, K. Tomida, &
M. Tamura, 539
[Manara et al., 2016]
Manara, C. F., Fedele, D., Herczeg, G. J., & Teixeira, P. S. 2016,
, 585, A136
[Manara et al., 2020]
Manara, C. F., Natta, A., Rosotti, G. P., et al. 2020, , 639, A58
[Manara et al., 2012]
Manara, C. F., Robberto, M., Da Rio, N., et al. 2012, , 755, 154
[Manara et al., 2017]
Manara, C. F., Testi, L., Herczeg, G. J., et al. 2017, , 604, A127
[Mauxion et al., 2024]
Mauxion, J., Lesur, G., & Maret, S. 2024, arXiv e-prints,
[Miotello et al., 2023]
Miotello, A., Kamp, I., Birnstiel, T., Cleeves, L. C., & Kataoka, A.
2023, in Astronomical Society of the Pacific Conference Series, Vol. 534,
Protostars and Planets VII, ed. S. Inutsuka, Y. Aikawa, T. Muto,
K. Tomida, & M. Tamura, 501
[Mohanty et al., 2005]
Mohanty, S., Jayawardhana, R., & Basri, G. 2005, , 626, 498
[Morbidelli et al., 2012]
Morbidelli, A., Lunine, J. I., O'Brien, D. P., Raymond, S. N., &
Walsh, K. J. 2012, Annual Review of Earth and Planetary Sciences, 40, 251
[Muzerolle et al., 2003]
Muzerolle, J., Hillenbrand, L., Briceño, C., Calvet, N., &
Hartmann, L. 2003, in Brown Dwarfs, Vol. 211, 141
[Natta et al., 2004]
Natta, A., Testi, L., Muzerolle, J., et al. 2004, , 424, 603
[Owen et al., 2012]
Owen, J. E., Clarke, C. J., & Ercolano, B. 2012, , 422, 1880
[Owen et al., 2010]
Owen, J. E., Ercolano, B., Clarke, C. J., & Alexand er, R. D. 2010,
, 401, 1415
[Pascucci et al., 2016]
Pascucci, I., Testi, L., Herczeg, G. J., et al. 2016, , 831, 125
[Picogna et al., 2021]
Picogna, G., Ercolano, B., & Espaillat, C. C. 2021, , 508, 3611
[Pinte et al., 2016]
Pinte, C., Dent, W. R. F., Ménard, F., et al. 2016, , 816, 25
[Popham et al., 1993]
Popham, R., Narayan, R., Hartmann, L., & Kenyon, S. 1993, , 415,
[Pringle, 1981]
Pringle, J. E. 1981, , 19, 137
[Rigliaco et al., 2011]
Rigliaco, E., Natta, A., Randich, S., Testi, L., & Biazzo, K. 2011,
, 525, A47
[Rosotti, 2023]
Rosotti, G. P. 2023, , 96, 101674
[Rosotti et al., 2017]
Rosotti, G. P., Clarke, C. J., Manara, C. F., & Facchini, S. 2017,
, 468, 1631
[Sanchis et al., 2020]
Sanchis, E., Testi, L., Natta, A., et al. 2020, , 633, A114
[Sellek et al., 2020]
Sellek, A. D., Booth, R. A., & Clarke, C. J. 2020, , 492, 1279
[Shakura & Sunyaev, 1973]
Shakura, N. I. & Sunyaev, R. A. 1973, in IAU Symposium, Vol. 55, X- and
Gamma-Ray Astronomy, ed. H. Bradt & R. Giacconi, 155
[Sinclair et al., 2020]
Sinclair, C. A., Rosotti, G. P., Juhasz, A., & Clarke, C. J. 2020,
, 493, 3535
[Somigliana et al., 2023]
Somigliana, A., Testi, L., Rosotti, G., et al. 2023, , 954, L13
[Somigliana et al., 2020]
Somigliana, A., Toci, C., Lodato, G., Rosotti, G., & Manara, C. F.
2020, , 492, 1120
[Somigliana et al., 2022]
Somigliana, A., Toci, C., Rosotti, G., et al. 2022, , 514, 5927
[Suzuki et al., 2016]
Suzuki, T. K., Ogihara, M., Morbidelli, A., Crida, A., & Guillot, T.
2016, , 596, A74
[Tabone et al., 2022]
Tabone, B., Rosotti, G. P., Cridland, A. J., Armitage, P. J., &
Lodato, G. 2022a, , 512, 2290
[Tabone et al., 2022]
Tabone, B., Rosotti, G. P., Lodato, G., et al. 2022b,
, 512, L74
[Tanaka et al., 2005]
Tanaka, H., Himeno, Y., & Ida, S. 2005, , 625, 414
[Testi et al., 2022]
Testi, L., Natta, A., Manara, C. F., et al. 2022, arXiv e-prints,
[Testi et al., 2016]
Testi, L., Natta, A., Scholz, A., et al. 2016, , 593, A111
[Trapman et al., 2023]
Trapman, L., Rosotti, G., Zhang, K., & Tabone, B. 2023, , 954, 41
[Trapman et al., 2022]
Trapman, L., Zhang, K., van't Hoff, M. L. R., Hogerheijde, M. R., &
Bergin, E. A. 2022, , 926, L2
[Venuti et al., 2019]
Venuti, L., Stelzer, B., Alcalá, J. M., et al. 2019, , 632, A46
[Veronesi et al., 2021]
Veronesi, B., Paneque-Carreño, T., Lodato, G., et al. 2021, ,
914, L27
[Winter et al., 2018]
Winter, A. J., Clarke, C. J., Rosotti, G., et al. 2018, , 478,
[Xu & Kunz, 2021]
Xu, W. & Kunz, M. W. 2021a, , 502, 4911
[Xu & Kunz, 2021]
Xu, W. & Kunz, M. W. 2021b, , 508, 2142
[Alcalá et al., 2017]
Alcalá, J. M., Manara, C. F., Natta, A., et al. 2017, , 600,
[Alcalá et al., 2014]
Alcalá, J. M., Natta, A., Manara, C. F., et al. 2014, , 561, A2
[Alexander et al., 2014]
Alexander, R., Pascucci, I., Andrews, S., Armitage, P., & Cieza, L.
2014, in Protostars and Planets VI, ed. H. Beuther, R. S. Klessen, C. P.
Dullemond, & T. Henning, 475–496
[Alexander et al., 2023]
Alexander, R., Rosotti, G., Armitage, P. J., et al. 2023, , 524,
[Alexander & Armitage, 2006]
Alexander, R. D. & Armitage, P. J. 2006, , 639, L83
[Alexander et al., 2006]
Alexander, R. D., Clarke, C. J., & Pringle, J. E. 2006a,
, 369, 216
[Alexander et al., 2006]
Alexander, R. D., Clarke, C. J., & Pringle, J. E. 2006b,
, 369, 229
[Anderson et al., 2022]
Anderson, D. E., Cleeves, L. I., Blake, G. A., et al. 2022, , 927,
[Andrews et al., 2018]
Andrews, S. M., Terrell, M., Tripathi, A., et al. 2018, , 865, 157
[Ansdell et al., 2017]
Ansdell, M., Williams, J. P., Manara, C. F., et al. 2017, , 153, 240
[Ansdell et al., 2016]
Ansdell, M., Williams, J. P., van der Marel, N., et al. 2016, ,
828, 46
[Armitage et al., 2013]
Armitage, P. J., Simon, J. B., & Martin, R. G. 2013, , 778, L14
[Barenfeld et al., 2016]
Barenfeld, S. A., Carpenter, J. M., Ricci, L., & Isella, A. 2016,
, 827, 142
[Bergin et al., 2013]
Bergin, E. A., Cleeves, L. I., Gorti, U., et al. 2013, , 493, 644
[Béthune et al., 2017]
Béthune, W., Lesur, G., & Ferreira, J. 2017, , 600, A75
[Birnstiel et al., 2010]
Birnstiel, T., Dullemond, C. P., & Brauer, F. 2010, , 513, A79
[Birnstiel et al., 2012]
Birnstiel, T., Klahr, H., & Ercolano, B. 2012, , 539, A148
[Blandford & Payne, 1982]
Blandford, R. D. & Payne, D. G. 1982, , 199, 883
[Booth et al., 2017]
Booth, R. A., Clarke, C. J., Madhusudhan, N., & Ilee, J. D. 2017,
, 469, 3994
[Clarke et al., 2001]
Clarke, C. J., Gendrin, A., & Sotomayor, M. 2001, , 328, 485
[Coleman et al., 2024]
Coleman, G. A. L., Mroueh, J. K., & Haworth, T. J. 2024, , 527,
[Dullemond et al., 2006]
Dullemond, C. P., Natta, A., & Testi, L. 2006, , 645, L69
[Fedele et al., 2010]
Fedele, D., van den Ancker, M. E., Henning, T., Jayawardhana, R., &
Oliveira, J. M. 2010, , 510, A72
[Ferreira, 1997]
Ferreira, J. 1997, , 319, 340
[Flaherty et al., 2018]
Flaherty, K. M., Hughes, A. M., Teague, R., et al. 2018, , 856, 117
[Habing, 1968]
Habing, H. J. 1968, , 19, 421
[Hartmann et al., 1998]
Hartmann, L., Calvet, N., Gullbring, E., & D'Alessio, P. 1998, ,
495, 385
[Haworth et al., 2023]
Haworth, T. J., Coleman, G. A. L., Qiao, L., Sellek, A. D., &
Askari, K. 2023, , 526, 4315
[Hennebelle et al., 2020]
Hennebelle, P., Commerçon, B., Lee, Y.-N., & Charnoz, S. 2020,
, 635, A67
[Herczeg & Hillenbrand, 2008]
Herczeg, G. J. & Hillenbrand, L. A. 2008, , 681, 594
[Hernández et al., 2007]
Hernández, J., Hartmann, L., Megeath, T., et al. 2007, , 662,
[Hollenbach et al., 1994]
Hollenbach, D., Johnstone, D., Lizano, S., & Shu, F. 1994, , 428,
[Kley & Lin, 1996]
Kley, W. & Lin, D. N. C. 1996, , 461, 933
[Kroupa, 2001]
Kroupa, P. 2001, , 322, 231
[Laibe & Price, 2014]
Laibe, G. & Price, D. J. 2014, , 444, 1940
[Lebreuilly et al., 2021]
Lebreuilly, U., Hennebelle, P., Colman, T., et al. 2021, , 917,
[Lebreuilly et al., 2024]
Lebreuilly, U., Hennebelle, P., Colman, T., et al. 2024, , 682, A30
[Lodato et al., 2023]
Lodato, G., Rampinelli, L., Viscardi, E., et al. 2023, , 518,
[Lodato et al., 2017]
Lodato, G., Scardoni, C. E., Manara, C. F., & Testi, L. 2017, ,
472, 4700
[Long et al., 2022]
Long, F., Andrews, S. M., Rosotti, G., et al. 2022, , 931, 6
[Lynden-Bell & Pringle, 1974]
Lynden-Bell, D. & Pringle, J. E. 1974, , 168, 603
[Machida & Basu, 2024]
Machida, M. N. & Basu, S. 2024, arXiv e-prints, arXiv:2405.08271
[Machida & Hosokawa, 2013]
Machida, M. N. & Hosokawa, T. 2013, , 431, 1719
[Manara et al., 2023]
Manara, C. F., Ansdell, M., Rosotti, G. P., et al. 2023, in
Astronomical Society of the Pacific Conference Series, Vol. 534, Protostars
and Planets VII, ed. S. Inutsuka, Y. Aikawa, T. Muto, K. Tomida, &
M. Tamura, 539
[Manara et al., 2016]
Manara, C. F., Fedele, D., Herczeg, G. J., & Teixeira, P. S. 2016,
, 585, A136
[Manara et al., 2020]
Manara, C. F., Natta, A., Rosotti, G. P., et al. 2020, , 639, A58
[Manara et al., 2012]
Manara, C. F., Robberto, M., Da Rio, N., et al. 2012, , 755, 154
[Manara et al., 2017]
Manara, C. F., Testi, L., Herczeg, G. J., et al. 2017, , 604, A127
[Mauxion et al., 2024]
Mauxion, J., Lesur, G., & Maret, S. 2024, arXiv e-prints,
[Miotello et al., 2023]
Miotello, A., Kamp, I., Birnstiel, T., Cleeves, L. C., & Kataoka, A.
2023, in Astronomical Society of the Pacific Conference Series, Vol. 534,
Protostars and Planets VII, ed. S. Inutsuka, Y. Aikawa, T. Muto,
K. Tomida, & M. Tamura, 501
[Mohanty et al., 2005]
Mohanty, S., Jayawardhana, R., & Basri, G. 2005, , 626, 498
[Morbidelli et al., 2012]
Morbidelli, A., Lunine, J. I., O'Brien, D. P., Raymond, S. N., &
Walsh, K. J. 2012, Annual Review of Earth and Planetary Sciences, 40, 251
[Muzerolle et al., 2003]
Muzerolle, J., Hillenbrand, L., Briceño, C., Calvet, N., &
Hartmann, L. 2003, in Brown Dwarfs, Vol. 211, 141
[Natta et al., 2004]
Natta, A., Testi, L., Muzerolle, J., et al. 2004, , 424, 603
[Owen et al., 2012]
Owen, J. E., Clarke, C. J., & Ercolano, B. 2012, , 422, 1880
[Owen et al., 2010]
Owen, J. E., Ercolano, B., Clarke, C. J., & Alexand er, R. D. 2010,
, 401, 1415
[Pascucci et al., 2016]
Pascucci, I., Testi, L., Herczeg, G. J., et al. 2016, , 831, 125
[Picogna et al., 2021]
Picogna, G., Ercolano, B., & Espaillat, C. C. 2021, , 508, 3611
[Pinte et al., 2016]
Pinte, C., Dent, W. R. F., Ménard, F., et al. 2016, , 816, 25
[Popham et al., 1993]
Popham, R., Narayan, R., Hartmann, L., & Kenyon, S. 1993, , 415,
[Pringle, 1981]
Pringle, J. E. 1981, , 19, 137
[Rigliaco et al., 2011]
Rigliaco, E., Natta, A., Randich, S., Testi, L., & Biazzo, K. 2011,
, 525, A47
[Rosotti, 2023]
Rosotti, G. P. 2023, , 96, 101674
[Rosotti et al., 2017]
Rosotti, G. P., Clarke, C. J., Manara, C. F., & Facchini, S. 2017,
, 468, 1631
[Sanchis et al., 2020]
Sanchis, E., Testi, L., Natta, A., et al. 2020, , 633, A114
[Sellek et al., 2020]
Sellek, A. D., Booth, R. A., & Clarke, C. J. 2020, , 492, 1279
[Shakura & Sunyaev, 1973]
Shakura, N. I. & Sunyaev, R. A. 1973, in IAU Symposium, Vol. 55, X- and
Gamma-Ray Astronomy, ed. H. Bradt & R. Giacconi, 155
[Sinclair et al., 2020]
Sinclair, C. A., Rosotti, G. P., Juhasz, A., & Clarke, C. J. 2020,
, 493, 3535
[Somigliana et al., 2023]
Somigliana, A., Testi, L., Rosotti, G., et al. 2023, , 954, L13
[Somigliana et al., 2020]
Somigliana, A., Toci, C., Lodato, G., Rosotti, G., & Manara, C. F.
2020, , 492, 1120
[Somigliana et al., 2022]
Somigliana, A., Toci, C., Rosotti, G., et al. 2022, , 514, 5927
[Suzuki et al., 2016]
Suzuki, T. K., Ogihara, M., Morbidelli, A., Crida, A., & Guillot, T.
2016, , 596, A74
[Tabone et al., 2022]
Tabone, B., Rosotti, G. P., Cridland, A. J., Armitage, P. J., &
Lodato, G. 2022a, , 512, 2290
[Tabone et al., 2022]
Tabone, B., Rosotti, G. P., Lodato, G., et al. 2022b,
, 512, L74
[Tanaka et al., 2005]
Tanaka, H., Himeno, Y., & Ida, S. 2005, , 625, 414
[Testi et al., 2022]
Testi, L., Natta, A., Manara, C. F., et al. 2022, arXiv e-prints,
[Testi et al., 2016]
Testi, L., Natta, A., Scholz, A., et al. 2016, , 593, A111
[Trapman et al., 2023]
Trapman, L., Rosotti, G., Zhang, K., & Tabone, B. 2023, , 954, 41
[Trapman et al., 2022]
Trapman, L., Zhang, K., van't Hoff, M. L. R., Hogerheijde, M. R., &
Bergin, E. A. 2022, , 926, L2
[Venuti et al., 2019]
Venuti, L., Stelzer, B., Alcalá, J. M., et al. 2019, , 632, A46
[Veronesi et al., 2021]
Veronesi, B., Paneque-Carreño, T., Lodato, G., et al. 2021, ,
914, L27
[Winter et al., 2018]
Winter, A. J., Clarke, C. J., Rosotti, G., et al. 2018, , 478,
[Xu & Kunz, 2021]
Xu, W. & Kunz, M. W. 2021a, , 502, 4911
[Xu & Kunz, 2021]
Xu, W. & Kunz, M. W. 2021b, , 508, 2142
§ PARAMETERS USED THROUGHOUT THE PAPER
Table <ref> shows the parameters used throughout the paper. The general simulation parameters refer to the initial correlations (distribution, slopes and spreads), while the following three tables are divided by evolutionary model and display the main parameters for each of them.
General simulation parameters
H]P3.5cm | P1.5cm
0pt1.5ex $\lambda_{\mathrm{m}, 0}$ 2.1
0pt1.5ex$\lambda_{\mathrm{acc}, 0}$ 1.5
0pt1.5ex$\sigma_{\mathrm{M_d}}$ $0.65$ dex
0pt1.5ex$\sigma_{\mathrm{R_d}}$ $0.52$ dex
0pt1.5ex$M_{\mathrm{d}}$, $R_{\mathrm{d}}$ distribution lognormal
Viscous model
H]P1.5cm | P1.5cm
0pt2.2ex $\alpha_{\mathrm{SS}}$ $10^{-3}$
Photoevaporative model
H]P1.5cm | P3cm
0pt1.5ex $\alpha_{\mathrm{SS}}$ $10^{-3}$
0pt1.5ex $\dot M_{w}$ $4 \times 10^{-10}$ M$_{\odot}$ yr$^{-1}$
Wind-driven model
H]P1.5cm | P1.5cm
0pt1.5ex $\alpha_{\mathrm{SS}}$ $10^{-3}$
0pt1.5ex $\alpha_{\mathrm{DW}}$ $10^{-3}$
0pt1.5ex $\lambda$ $3$
0pt1.5ex $\omega$ $0.25$
Values of the parameters used throughout the paper, unless explicitly stated otherwise.
§ MHD MODEL WITH MU<0
As mentioned in Section <ref>, the breaking of the linear correlation between the disc properties and the stellar mass happens towards higher or lower stellar masses depending on the value of $\mu_0$. Figure <ref> shows the a simulation $\mu_0 > 0$, while in Figure <ref> we show the opposite case. As $\mu_0$ is linked to $\delta_0$ through $\mu_0 = \delta_0 - \zeta_0 \xi$, if $\delta_0 < \zeta_0 \xi$ the resulting $\mu_0$ will be negative, leading to a specular bending of the correlation. Given that $\zeta_0 \xi$ is a small number ($\sim 0.1$ in our simulation), this generally requires a negative $\delta_0$. The simulation in Figure <ref> has $\lambda_{\mathrm{m}, 0} = 1.3$ and $\lambda_{\mathrm{acc}, 0} = 1.7$, resulting in $\delta_0 = -0.4$.
Same as Figure <ref>, but with a choice of initial slopes resulting in a negative $\mu_0$ ($\lambda_{\mathrm{m}, 0} = 1.3, \lambda_{\mathrm{acc}, 0} = 1.7$). The bending of the linear correlation happens towards larger disc masses, in agreement with the prediction.
§ VALIDATION OF
Figure <ref> shows the evolution of the gas surface density as a function of the disc radius in the cases of evolution driven by (i) viscosity, (ii) viscosity and internal photoevaporation, (iii) MHD winds, and (iv) viscosity and external photoevaporation for a single disc simulated with . The top left panel, corresponding to viscous evolution (i), shows the key feature of viscous spreading: while the global surface density decreases as a consequence of the accretion onto the central star, the radial extent of the disc increases. This is a consequence of the redistribution of angular momentum, that causes part of the disc material to move towards larger radii. The top right panel, where the disc evolves under the combined effect of viscosity and internal photoevaporation (ii), shows the characteristic two-timescales behaviour [Clarke et al., 2001]: the evolution is effectively viscous in the earliest stages, as long as the accretion rate is larger than the photoevaporative mass-loss rate; then, photoevaporation opens a cavity within the disc, which gets divided into an inner and an outer disc. The inner disc is less extended and therefore has a shorter viscous timescale, which means it evolves much faster and is quickly completely accreted onto the protostar; the outer disc instead keeps on evolving on timescales comparable to the original one, making photoevaporation a two-timescales process. The bottom left panel shows a disc evolving due to MHD winds (iii): the absolute value of the surface density drops faster than in the viscous model, because of the increase of $\alpha_{\mathrm{DW}}$ in time. Furthermore, as angular momentum is removed from the wind (together with material), the disc does not spread but rather shrinks in time, as expected from the theoretical prediction [Tabone et al., 2022]. Finally, the bottom right panel shows the evolution of a disc undergoing external photoevaporation combined with viscosity (iv): the latter dominates at the earliest stages, producing the characteristic features like viscous spreading, while the effect of external photoevaporation is visible at later ages as a truncation of the disc that also halts its spreading. In this case, the disc truncation and the outside-in depletion of disc material is the consequence of the photo-dissociation of gas molecules due to the FUV radiation emitted by massive stars and experienced by the disc. The efficiency of this process depends primarily on the stellar mass and the FUV flux experienced: given a fixed FUV flux, a disc around a lower mass star will lose its material to external winds more easily compared to a disc around a higher mass star, because of the higher gravitational bond of the system. For the same reason, more extended and less massive discs are more prone to external truncation.
Gas surface density as a function of radius for a disc generated with Diskpop at different times as the colour bar shows. The four models are purely viscous (top left, $\alpha_{\mathrm{SS}} = 10^{-3}$), viscous including internal photoevaporation (top right, $\alpha_{\mathrm{SS}} = 10^{-3}$, $\dot M_{\mathrm{w}} = 4 \times 10^{-10}$ M$_{\odot} \mathrm{yr^{-1}}$), wind-driven (bottom left, $\alpha_{\mathrm{DW}} = 10^{-3}$, $\lambda = 3$, $\omega = 0.25$) and viscous including external photoevaporation (bottom right, $\alpha_{\mathrm{SS}} = 10^{-3}$, FUV = 100 $G_0$) respectively.
Figure <ref> shows the isochrones in the $\dot M - M_{\mathrm{d}}$ plane at 0.1, 1 and 10 Myr for the three evolutionary models of viscosity, viscosity and photoevaporation, and MHD winds. Each dot represents a disc in the population, while the solid lines show the analytical prediction (when applicable): in the viscous case, the isochrones read [Lodato et al., 2017]
\begin{equation}
\dot M = \frac{M_{\rm{d}}}{2(2 - \gamma)t} \left[ 1 - \left( \frac{M_{\rm{d}}}{M_0} \right)^{2(2-\gamma)} \right],
\label{eq:theoretical_isochrone_viscous}
\end{equation}
while in the MHD wind-driven scenario [Tabone et al., 2022]
\begin{equation}
\dot M = \frac{1}{\omega (1 + f_{\rm{M}, 0})t} M_{\rm{d}} \left[ \left( \frac{M_{\rm{d}}}{M_0} \right)^{- \omega} - 1 \right].
\label{eq:theoretical_isochrone_MHD}
\end{equation}
The left panel shows the viscous model, where the discs tend more and more towards the theoretical isochrone during their evolution [Lodato et al., 2017]; the central panel includes internal photoevaporation, which has the effect of bending the isochrones once the accretion rate becomes comparable to the photoevaporative mass-loss rate [Somigliana et al., 2020]; finally, the right panel shows an MHD wind-driven population, where the scatter in the $\dot M - M_{\mathrm{d}}$ plane remains significant during the evolution - contrary to the viscous prediction [Somigliana et al., 2023].
Isochrones at 0.1, 1, and 10 Myr for disc populations undergoing viscous, viscous+internal photoevaporation and wind-driven evolution (left, centre, and right panel respectively), with the same parameters as Figure <ref>. Each dot represents a disc, while the solid lines show the theoretical isochrones at the corresponding age, where available.
|
# CALI: Coarse-to-Fine ALIgnments Based Unsupervised Domain Adaptation of
Traversability Prediction for Deployable Autonomous Navigation
Zheng Chen Luddy School of Informatics,
Computing, and Engineering
Indiana University - Bloomington
Indiana 47408
Email<EMAIL_ADDRESS>Durgakant Pushp Luddy School of Informatics,
Computing, and Engineering
Indiana University - Bloomington
Indiana 47408
Email<EMAIL_ADDRESS>Lantao Liu Luddy School of Informatics,
Computing, and Engineering
Indiana University - Bloomington
Indiana 47408
Email<EMAIL_ADDRESS>
###### Abstract
Traversability prediction is a fundamental perception capability for
autonomous navigation. The diversity of data in different domains imposes
significant gaps to the prediction performance of the perception model. In
this work, we make efforts to reduce the gaps by proposing a novel coarse-to-
fine unsupervised domain adaptation (UDA) model – CALI. Our aim is to transfer
the perception model with high data efficiency, eliminate the prohibitively
expensive data labeling, and improve the generalization capability during the
adaptation from easy-to-obtain source domains to various challenging target
domains. We prove that a combination of a coarse alignment and a fine
alignment can be beneficial to each other and further design a first-coarse-
then-fine alignment process. This proposed work bridges theoretical analyses
and algorithm designs, leading to an efficient UDA model with easy and stable
training. We show the advantages of our proposed model over multiple baselines
in several challenging domain adaptation setups. To further validate the
effectiveness of our model, we then combine our perception model with a visual
planner to build a navigation system and show the high reliability of our
model in complex natural environments where no labeled data is available.
111The paper is published in Robotics: Science and Systems (RSS) 2022.
The robot navigation demonstration can be seen in this video:
https://www.youtube.com/watch?v=Nqsegaq_x-o.
## I Introduction
We consider the deployment of autonomous robots in the real-world unstructured
field environments, where the environments can be extremely complex involving
random obstacles (e.g., big rocks, tree stumps, man-made objects), cross-
domain terrains (e.g., combinations of gravel, sand, wet, uneven surfaces), as
well as dense vegetation (tall and low grasses, shrubs, trees). Whenever a
robot is deployed in such an environment, it needs to understand which area of
the captured scene is navigable. A typical solution to this problem is the
visual traversability prediction that can be achieved by learning the scene
semantic segmentation.
Visual traversability prediction has been tackled by using deep neural
networks where the models are typically trained offline with well-labeled
datasets in limited scenarios. However, there is a gap between the data used
to train the model and the real world. It is usually challenging for existing
datasets to well approximate the true distributions of the unseen target
environments where the robot is deployed. Even incrementally collecting and
adding new training data on the fly cannot guarantee the target environments
to be well in-distribution included. In addition, annotating labels for dense
predictions, e.g., semantic segmentation, is prohibitively expensive.
Therefore, developing a generalization-aware deep model is crucial for robotic
systems considering the demands of the practical deployment of deep perception
models and the costs/limits of collecting new data in many robotic
applications, e.g., autonomous driving, search and rescue, and environmental
monitoring.
Figure 1: Transferring models from the available domain to the target domain.
The existing available data might be from either a simulator or collecting
data in certain environments, at a certain time, and with certain sensors. In
contrast, the target deployment might have significantly varying environments,
time, and sensors.
To tackle this challenge, a broadly studied framework is transfer learning
[24] which aims to transfer models between two domains – source domain and
target domain – that have related but different data distributions. The
prediction on target domain can be considered as a strong generalization since
testing data (in target domain) might fall out of the independently and
identically distributed (i.i.d.) assumption and follow a very different
distribution than the training data (in source domain). The “transfer” process
has significant meaning to our model development since we can view the
available public datasets [29, 8, 35, 17] as the source domain and treat the
data in the to-be-deployed environments as the target domain. In this case, we
have access to images and corresponding labels in source domain and images in
target domain, but no access to labels in target domain. Transferring models,
in this set-up, is called Unsupervised Domain Adaptation (UDA) [36, 40].
Domain Alignment (DA) [10, 12, 13, 32, 33] and Class Alignment (CA) [31] are
two conventional ways to tackle the UDA problem. DA treats the deep features
as a whole. It works well for image-level tasks such as image classification,
but has issues with pixel-level tasks such as semantic segmentation, as the
alignment of whole distributions ignores the class features and might misalign
class distributions, even the whole features from the source domain and target
domain are already well-aligned. CA is proposed to solve this issue for dense
predictions with multiple classes.
It is natural and necessary to use CA to tackle the UDA of semantic
segmentation as we need to consider aligning class features. However, CA can
be problematic and might fail to outperform the DA for segmentation, and in a
worse case, might have unacceptable negative transfer, which means the
performance with adaptation is even degraded than that without adaptation.
Intuitively, we need to consider more alignments in CA than in DA. Thus the
searching space might be more complicated, and training might be more unstable
and harder to converge to an expected minima, leading to larger prediction
errors.
To solve the issue of CA, we investigate the relationship of the upper bounds
of the prediction error on target domain between DA and CA and provide a
theoretical analysis of the upper bounds of target prediction error in UDA
setup, and bridge the theoretical analysis and algorithm design for UDA of
traversability prediction.
In summary, our contributions include
* •
We prove that with proper assumptions, the upper bound of CA is upper bounded
by the upper bound of DA. This indicates that constraining the training of CA
using DA can be beneficial. We then propose a novel concept of pseudo-
trilateral game structure (PTGS) for integrating DA and CA.
* •
We propose an efficient coarse-to-fine alignments based UDA model, named CALI,
for traversability prediction. The new proposal includes a trilateral network
structure, novel training losses, and an alternative training process. Our
model design is well supported by theoretical analysis. It is also easy and
stable to train and converge.
* •
We show significant advantages of our proposed model compared to several
baselines in multiple challenging public datasets and one self-collected
dataset. We combine the proposed segmentation model and a visual planner to
build a visual navigation system. The results show high safety and
effectiveness of our model.
## II Related Work
Semantic Segmentation: Semantic segmentation aims to predict a unique human-
defined semantic class for each pixel in the given images. With the prosperity
of deep neural networks, the performance of semantic segmentation has been
boosted significantly, especially by the advent of FCN [20] that first
proposes to use deep convolutional neural nets to predict segmentation.
Following works try to improve the FCN performance by multiple proposals,
e.g., using different sizes of kernels or dilation rates to aggregate multi-
scale features [6, 7, 38]; building image pyramids to create multi-resolution
inputs [41]; applying probabilistic graph to smooth the prediction [19];
compensating features in deeper level by an encoder-decoder structure [30],
and employing attention mechanism to capture the long-range dependencies among
pixels in a more compact and efficient way [28]. We can also see how excellent
the current semantic segmentation SOTA performance is from very recently
released work [42, 37]. However, all of those methods belong to fully-
supervised learning and the performance might catastrophically be degraded
when a domain shift exists between the training data and the data when
deploying. Considering the possible domain shift and developing adaptation-
aware models is extremely practical and urgent.
Unsupervised Domain Adaptation: The main approaches to tackle UDA include
adversarial training (a.k.a., distribution alignment) [10, 12, 13, 32, 31, 33,
21, 34] and self-training [43, 39, 23, 16]. Although self-training is becoming
another dominant method for segmentation UDA in terms of the empirical
results, it still lacks a sound theoretical foundation. In this paper, we only
focus on the alignment-based methods that not only keep close to the UDA
state-of-the-art (SOTA) performance but also are well supported by sound
theoretical analyses [1, 3, 2].
The alignment-based methods adapt models via aligning the distributions from
the source domain and target domain in an adversarial training process, i.e.,
making the deep features of source images and target images indistinguishable
to a discriminator net. Typical approaches to UDA include Domain Alignment
(DA) [10, 12, 13, 32, 33], which aligns the two domains using global features
(aligning the feature tensor from source or target as a whole) and Class
Alignment (CA) [31, 21, 34], which only considers aligning features of each
class from source and target, no matter whether the domain distributions are
aligned or not. In [31], the authors are inspired by the theoretical analysis
of [2] and propose a discrepancy-based model for aligning class features.
There is a clear relation between the theory guidance [2] and the design of
network, loss, and training methods. There are some recent works [21, 34]
similar to the proposed work in spirit and show improved results compared to
[31], but it is still unclear to relate the proposed algorithms with theory
and to understand why the structure/loss/training is designed as the presented
way.
## III Background and Preliminary Materials
We consider segmentation tasks where the input space is
$\mathcal{X}\subset\mathbb{R}^{H\times W\times 3}$, representing the input RGB
images, and the label space is
$\mathcal{Y}\subset\left\\{0,1\right\\}^{H\times W\times K}$, representing the
ground-truth $K$-class segmentation images, where the label for a single pixel
at $(h,w)$ is denoted by a one-hot vector $y^{(h,w)}\in\mathbb{R}^{K}$ whose
elements are by-default 0-valued except at location $(h,w)$ which is labeled
as 1. Domain adaptation has two domain distributions over
$\mathcal{X}\times\mathcal{Y}$, named source domain $\mathcal{D}_{S}$ and
target domain $\mathcal{D}_{T}$. In the setting of UDA for segmentation, we
have access to $m_{s}$ i.i.d. samples with labels
$\mathcal{U}_{S}=\left\\{\mathbf{x}_{si},\mathbf{y}_{si}\right\\}_{i=1}^{m_{s}}$
from $\mathcal{D}_{S}$ and $m_{t}$ i.i.d. samples without labels
$\mathcal{U}_{T}=\left\\{\mathbf{x}_{tj}\right\\}_{j=1}^{m_{t}}$ from
$\mathcal{D}_{T}$.
In the UDA problem, we need to reduce the prediction error on the target
domain. A hypothesis is a function $h:\mathcal{X}\rightarrow\mathcal{Y}$. We
denote the space of $h$ as $\mathcal{H}$. With the loss function
$l(\cdot,\cdot)$, the expected error of $h$ on $\mathcal{D}_{S}$ is defined as
$\epsilon_{S}(h):=\mathbb{E}_{(x,y)\sim\mathcal{D}_{S}}l(h(x),y).$ (1)
Similarly, we can define the expected error of $h$ on $\mathcal{D}_{T}$ as
$\epsilon_{T}(h):=\mathbb{E}_{(x,y)\sim\mathcal{D}_{T}}l(h(x),y).$ (2)
Two important upper bounds related to the source and target error are given in
[2]. Basically,
Theorem 1 For a hypothesis $h$,
$\epsilon_{T}(h)\leq\epsilon_{S}(h)+d_{1}(\mathcal{D}_{S},\mathcal{D}_{T})+\lambda,$
(3)
where $d_{1}(\cdot,\cdot)$ is the $L^{1}$ divergence for two distributions,
and the constant term $\lambda$ does not depend on any $h$. However, it is
claimed in [2] that the bound with $L^{1}$ divergence cannot be accurately
estimated from finite samples, and using $L^{1}$ divergence can unnecessarily
inflate the bound. Another divergence measure is thus introduced to replace
the $L^{1}$ divergence with a new bound derived.
Definition 1 Given two domain distributions $\mathcal{D}_{S}$ and
$\mathcal{D}_{T}$ over $\mathcal{X}$, and a hypothesis space $\mathcal{H}$
that has finite VC dimension, the $\mathcal{H}$-divergence between
$\mathcal{D}_{S}$ and $\mathcal{D}_{T}$ is defined as
$\displaystyle
d_{\mathcal{H}}(\mathcal{D}_{S},\mathcal{D}_{T})=2\sup_{h\in\mathcal{H}}|$
$\displaystyle\text{P}_{x\sim\mathcal{D}_{S}}\left[h(x)=1\right]-$ (4)
$\displaystyle\text{P}_{x\sim\mathcal{D}_{T}}\left[h(x)=1\right]|,$
where $\text{P}_{x\sim\mathcal{D}_{S}}[h(x)=1]$ represents the probability of
$x$ belonging to $\mathcal{D}_{S}$. Same to
$\text{P}_{x\sim\mathcal{D}_{T}}\left[h(x)=1\right]$.
The $\mathcal{H}$-divergence resolves the issues in the $L^{1}$ divergence. If
we replace $d_{1}(\mathcal{D}_{S},\mathcal{D}_{T})$ in Eq. (3) with
$d_{\mathcal{H}}(\mathcal{D}_{S},\mathcal{D}_{T})$, then a new upper bound for
$\epsilon_{T}(h)$, named as $\mathbb{UB}_{1}$, can be written as
$\displaystyle\epsilon_{T}(h)\leq\mathbb{UB}_{1},$ (5)
$\displaystyle\mathbb{UB}_{1}=\epsilon_{S}(h)+d_{\mathcal{H}}(\mathcal{D}_{S},\mathcal{D}_{T})+\lambda.$
An approach to compute the empirical $\mathcal{H}$-divergence is also proposed
in [2].
Lemma 1 For a symmetric hypothesis class $\mathcal{H}$ (one where for every
$h\in\mathcal{H}$, the inverse hypothesis $1-h$ is also in $\mathcal{H}$) and
two sample sets
$\mathcal{U}_{S}=\left\\{x_{i},i=1,\cdots,m_{s},x_{i}\sim\mathcal{D}_{S}\right\\}$
and
$\mathcal{U}_{T}=\left\\{x_{j},j=1,\cdots,m_{t},x_{j}\sim\mathcal{D}_{T}\right\\}$.
$\ \begin{aligned}
\hat{d}_{\mathcal{H}}(\mathcal{D}_{S},\mathcal{D}_{T})=2\Bigg{(}\Bigg{.}1-\min_{\eta\in\mathcal{H}}\Bigg{[}\Bigg{.}&\frac{1}{m_{s}}\sum_{i=1}^{m_{s}}\mathbb{I}[\eta(x_{i})=0]+\\\
&\frac{1}{m_{t}}\sum_{j=1}^{m_{t}}\mathbb{I}[\eta(x_{j})=1]\Bigg{.}\Bigg{]}\Bigg{.}\Bigg{)},\end{aligned}$
(6)
where $\mathbb{I}[a]$ is an indicator function which is 1 if $a$ is true, and
$0$ otherwise.
The second upper bound is based on a new hypothesis called symmetric
difference hypothesis.
Definition 2 For a hypothesis space $\mathcal{H}$, the symmetric difference
hypothesis space $\mathcal{H}\Delta\mathcal{H}$ is the set of hypotheses
$g\in\mathcal{H}\Delta\mathcal{H}\Leftrightarrow g(x)=h(x)\oplus
h^{\prime}(x)~{}~{}~{}\text{for some }h,h^{\prime}\in\mathcal{H},$ (7)
where $\oplus$ denotes an XOR operation. Then we can define the
$\mathcal{H}\Delta\mathcal{H}$-distance as
$\displaystyle
d_{\mathcal{H}\Delta\mathcal{H}}(\mathcal{D}_{S},\mathcal{D}_{T})=2\sup_{h,h^{\prime}\in\mathcal{H}}|$
$\displaystyle\text{P}_{x\sim\mathcal{D}_{S}}\left[h(x)\neq
h^{\prime}(x)\right]-$ (8)
$\displaystyle\text{P}_{x\sim\mathcal{D}_{T}}\left[h(x)\neq
h^{\prime}(x)\right]|.$
Similar to Eq. (5), if we replace $d_{1}(\mathcal{D}_{S},\mathcal{D}_{T})$
with $d_{\mathcal{H}\Delta\mathcal{H}}(\mathcal{D}_{S},\mathcal{D}_{T})$, the
second upper bound for $\epsilon_{T}(h)$, named as $\mathbb{UB}_{2}$, can be
expressed as
$\displaystyle\epsilon_{T}(h)\leq\mathbb{UB}_{2},$ (9)
$\displaystyle\mathbb{UB}_{2}=\epsilon_{S}(h)+d_{\mathcal{H}\Delta\mathcal{H}}(\mathcal{D}_{S},\mathcal{D}_{T})+\lambda,$
where $\lambda$ is the same term as in Eq. (3).
A standard way to achieve the alignment for deep models is to use the
adversarial training method, which is also used in Generative Adversarial
Networks (GANs). Therefore we explain the key concepts of adversarial training
using the example of GANs.
GAN is proposed to learn the distribution $p_{r}$ of a set of given data
$\left\\{\mathbf{x}\right\\}$ in an adversarial manner. The architecture
consists of two networks - a generator $G$, and a discriminator $D$. The $G$
is responsible for generating fake data (with distribution $p_{g}$) from
random noises $\mathbf{z}\sim p_{\mathbf{z}}$ to fool the discriminator $D$
that is instead to accurately distinguish between the fake data and the given
data. Optimization of a GAN involves a mini-maximization over a joint loss for
$G$ and $D$.
$\displaystyle\min_{G}\max_{D}V(G,D)$ (10) $\displaystyle
V(G,D)=\mathbb{E}_{\mathbf{x}\sim
p_{r}}\log\left[D(\mathbf{x})\right]+\mathbb{E}_{\mathbf{z}\sim
p_{\mathbf{z}}}\log\left[1-D(G(z))\right].$
where we use $1$ as the real label and $0$ as the fake label. Training with
Eq. (10) is a bilateral game where the distribution $p_{g}$ is aligned with
the distribution $p_{r}$.
The two bounds (Eq. (5) and Eq. (9)) for the target domain error are
separately given in [2]. It has been independently demonstrated that domain
alignment corresponds to optimizing over $\mathbb{UB}_{1}$ [10], where
optimization over the upper bound $\mathbb{UB}_{1}$ (Eq. (5) with the
divergence Eq. (6)) is proved as equivalent to Eq. (10) with a supervised
learning in the source domain, and that class alignment corresponds to
optimizing over $\mathbb{UB}_{2}$ [31], where the
$d_{\mathcal{H}\Delta\mathcal{H}}$ is approximated by the discrepancy between
two different classifiers.
Training DA is straightforward since we can easily define binary labels for
each domain, e.g., we can use 1 as the source domain label and 0 as the target
domain label. Adversarial training over the domain labels can achieve domain
alignment. For CA, however, it is difficult to implement as we do not have
target labels, hence the target class features are completely unknown to us,
thus leading naively using adversarial training over each class impossible.
The existing way well supported by theory to perform CA [31] is to indirectly
align class features by devising two different classifier hypotheses. The two
classifiers have to be well trained on the source domain and are able to
classify different classes in the source domain with different decision
boundaries. Then considering the shift between source and target domain, the
trained two classifiers might have disagreements on target domain classes.
Note since the two classifiers are already well trained on the source domain,
the agreements of the two classifiers represent those features in the target
domain that are close to the source domain, while in contrast, the features
where disagreements happen indicate that there is a large shift between source
and target. We use the disagreements to approximate the distance between
source and target. If we are able to minimize the disagreements of the two
classifiers, then features of each class between source and target will be
enforced to be well aligned.
Figure 2: An ideal iterative training process by integrating DA and CA.
## IV Methodology
In this work we further investigate the relation between the $\mathbb{UB}_{1}$
and $\mathbb{UB}_{2}$ and prove that $\mathbb{UB}_{1}$ turns out to be an
upper bound of $\mathbb{UB}_{2}$, meaning DA can be a necessary constraint to
CA. This is also consistent to our intuition: DA aligns features globally in a
coarse way while CA aligns features locally in a finer way. Constraining CA
with DA is actually a coarse-to-fine process, which makes the alignment
process efficient and stable. By carefully studying the internal structure of
existing DA and CA work, we propose a novel concept, pseudo-trilateral game
structure, for efficiently integrating DA and CA. We follow our theoretical
analysis and proposed PTGS to guide the development of CALI, including designs
of model structure, losses and training process.
Figure 3: Pseudo-trilateral game structure (PTGS). Three players are in the
game, a feature extractor $G$, a domain discriminator $D$ and a family of
classifiers $C_{s}$. The game between $G$ and $Cs$ is the CA while the game
between $G$ and $D$ is the DA. The DA and CA are connected by sharing the same
feature extractor $G$. Both $D$ and $C_{s}$ are trying to adjust the $G$ such
that the features between source and target generated from $G$ could be well
aligned globally and locally.
Notations used in this paper is explained as follows. we denote the
segmentation model $h$ as $h^{\theta,\phi}(x)=C^{\theta}(G^{\phi}(x))$ which
consists of a feature extractor $G^{\phi}$ parameterized by $\phi$ and a
classifier $C^{\theta}$ parameterized by $\theta$, and $x$ is a sample from
$\mathcal{U}_{S}$ or $\mathcal{U}_{T}$. If multiple classifiers are used, we
will denote the $j^{th}$ classifier as $C_{j}$. We denote the discriminator as
$D^{\psi}$ parameterized by $\psi$.
### IV-A Bounds Relation
We start by examining the relationship between the DA and the CA from the
perspective of target error bound. We propose to use this relation to improve
the segmentation performance of class alignment, which is desired for dense
prediction tasks. We provide the following theorem:
Theorem 2 If we assume there is a hypothesis space $\mathcal{H}$ for
segmentation model $h^{\theta,\phi}$ and a hypothesis space $\mathcal{H}_{D}$
for domain classifiers $D^{\psi}$, and
$\mathcal{H}\Delta\mathcal{H}\subset\mathcal{H}_{D}$, then we have
$\displaystyle\epsilon_{T}(h)$
$\displaystyle\leq\hat{\mathbb{UB}}_{2}\leq\hat{\mathbb{UB}}_{1},$ (11)
$\displaystyle\hat{\mathbb{UB}}_{1}$
$\displaystyle=\epsilon_{S}(h)+\frac{1}{2}d_{\mathcal{H}_{D}}(\mathcal{D}_{S},\mathcal{D}_{T})+\lambda,$
$\displaystyle\hat{\mathbb{UB}}_{2}$
$\displaystyle=\epsilon_{S}(h)+\frac{1}{2}d_{\mathcal{H}\Delta\mathcal{H}}(\mathcal{D}_{S},\mathcal{D}_{T})+\lambda.$
The proof of this theorem is provided in Appendix. VI-A.
Essentially, we limit the hypothesis space $\mathcal{H}$ and $\mathcal{H}_{D}$
in Eq. (11) into the space of deep neural networks. Directly optimizing over
$\hat{\mathbb{UB}_{2}}$ might be hard to converge since
$\hat{\mathbb{UB}_{2}}$ is a tighter upper bound for the prediction error on
target domain. The bounds relation in Eq. (11) shows that the
$\hat{\mathbb{UB}}_{1}$ is an upper bound of $\hat{\mathbb{UB}}_{2}$. This
provides us a clue to improve the training process of class alignment, i.e.,
the domain alignment can be a global constraint and narrow down the searching
space for the class alignment. This also implies that integrating the domain
alignment and class alignment might boost the training efficiency as well as
the prediction performance of UDA. An ideal training process is illustrated in
Fig. 2 where the searching space of $\hat{\mathbb{UB}}_{2}$ (CA) is constantly
bounded by that of $\hat{\mathbb{UB}}_{1}$ (DA), ensuring the whole training
process converge stably. This inspires us to design a new model, and we are
explaining next in details about our model structures, losses and training
process.
### IV-B CALI Structure
The existing DA or CA works usually involve a bilateral game. In CA, the game
is between a feature extractor and a family of classifiers. The two players
are optimized over the discrepancy of the two classifiers (note here the two
players are the two classifiers vs. the feature extractor) in an opposite
manner. In DA, the game happens between a segmentation net and a domain
discriminator. The two players are optimized over the domain discrimination in
an opposite way. It has been empirically showed [33, 32] that DA performs well
if the domain alignment happens to the prediction probability (after
Softmax()). However, according to the identified relation in Eq. (11), the two
upper bounds $\hat{\mathbb{UB}}_{1}$ and $\hat{\mathbb{UB}}_{2}$ need to use
the same feature, hence we connect the domain alignment and class alignment
using a shared feature extractor and propose a novel concept called PTGS (see
Fig. 3) to illustrate an interesting structure to integrate DA and CA. Both
$C_{s}$ and $D$ have game with $G$, but there is no game between $C_{s}$ and
$D$, hence we call this game as pseudo-trilateral game. Furthermore, as
defined in Eq. (8), $h$ and $h^{{}^{\prime}}$ are two different hypotheses,
thus we have to ensure the classifiers in $C_{s}$ are different during the
training.
Following the concept of PTGS, we design the structure of our CALI model as
shown in Fig. 4. Four networks are involved, a shared feature extractor $G$, a
domain discriminator $D$ and two classifiers $C_{1}$ and $C_{2}$. $f$
represents the shared features; $P_{1}/O_{1}$ and $P_{2}/O_{2}$ are the
probability/class predictions for $C_{1}$ and $C_{2}$, respectively; $S/T$
represent the source domain label (1) and target domain label (0); and $L_{1}$
represents the $L_{1}$ distance measure between two probability distributions.
The one-way solid arrows indicate the forward propagation of the data flow
while the two-way dashed arrows indicate losses are generated. The red arrows
represent the source-related data while the blue ones represent the target-
related data. The orange two-way dashed line indicates the structural
regularization loss between the $C_{1}$ and $C_{2}$.
Figure 4: CALI network structure. See Section. IV-B for more details.
### IV-C CALI Losses
We denote raw images from source or target domain as $\mathbf{x}$, and the
label from source domain as $\mathbf{y}$. We use semantic labels in source
domain to train all of the nets, but the domain discriminator, in a supervised
way, see the solid red one-way arrow in Fig. 4. We need to minimize the
supervised segmentation loss since Eq. (11) and other related Eqs suggest that
the source prediction error is also part of the upper bound of target error.
The supervised segmentation loss for training CALI is defined as
$\displaystyle\mathcal{L}_{seg}(G,C_{1},C_{2})=\frac{1}{2}\Bigg{(}\Bigg{.}\mathbb{E}_{(\mathbf{x},\mathbf{y})\sim\mathcal{D}_{S}}\left[-\mathbf{y}\log(C_{1}(G(\mathbf{x})))\right]+$
(12)
$\displaystyle~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}\mathbb{E}_{(\mathbf{x},\mathbf{y})\sim\mathcal{D}_{S}}\left[-\mathbf{y}\log(C_{2}(G(\mathbf{x})))\right]\Bigg{.}\Bigg{)}$
$\displaystyle=-\frac{1}{2}\mathbb{E}_{(\mathbf{x},\mathbf{y})\sim\mathcal{D}_{S}}[\mathbf{y}_{S}\log\left((C_{1}(G(\mathbf{x}))\odot(C_{2}(G(\mathbf{x}))))\right)],$
where $\odot$ represents the element-wise multiplication between two tensors.
To perform domain alignment, we need to define the joint loss function for $G$
and $D$
$\mathcal{V}_{1}(G,D)=-\left(\mathcal{CE}_{S}(\mathbf{x})+\mathcal{CE}_{T}(\mathbf{x})\right),$
(13)
where no segmentation labels but domain labels are used, and we use the
standard cross-entropy to compute the domain classification loss for both
source ($\mathcal{CE}_{S}(\mathbf{x})$) and target data
($\mathcal{CE}_{T}(\mathbf{x})$). We have
$\displaystyle\mathcal{CE}_{S}(\mathbf{x})$
$\displaystyle=\mathbb{E}_{\mathbf{x}\sim\mathcal{D}_{S}}[\mathcal{CE}([1,0]^{T},[D(G(\mathbf{x})),1-D(G(\mathbf{x}))]^{T})]$
(14)
$\displaystyle=\mathbb{E}_{\mathbf{x}\sim\mathcal{D}_{S}}[-\log(D(G(\mathbf{x})))].$
and
$\displaystyle\mathcal{CE}_{T}(\mathbf{x})$
$\displaystyle=\mathbb{E}_{\mathbf{x}\sim\mathcal{D}_{T}}[\mathcal{CE}([0,1]^{T},[D(G(\mathbf{x})),1-D(G(\mathbf{x}))]^{T})]$
(15)
$\displaystyle=\mathbb{E}_{\mathbf{x}\sim\mathcal{D}_{T}}[-\log(1-D(G(\mathbf{x})))],$
Note we include $G$ in Eq. (14) since both the source data and target data are
passed through the feature extractor. This is different than standard GAN,
where the real data is directly fed to $D$, without passing through the
generator.
To perform class alignment, we need to define the joint loss function for $G$,
$C_{1}$, and $C_{2}$
$\displaystyle\mathcal{V}_{2}(G,C_{1},C_{2})=\mathbb{E}_{\mathbf{x}\sim\mathcal{D}_{T}}\left[d(C_{1}(G(\mathbf{x})),C_{2}(G(\mathbf{x})))\right],$
(16)
where $d(\cdot,\cdot)$ is the distance measure between two distributions from
the two classifiers. In this paper, we use the same $L_{1}$ distance in [31]
as the measure, thus $d(p,q)=\frac{1}{K}|p-q|_{1}$, where $p$ and $q$ are two
distributions and $K$ is the number of label classes.
To prevent $C_{1}$ and $C_{2}$ from converging to the same network throughout
the training, we use the cosine similarity as a weight regularization to
maximize the difference of the weights from $C_{1}$ and $C_{2}$, i.e.,
$\mathcal{WR}(C_{1},C_{2})=\frac{\mathbf{w}_{1}\cdot\mathbf{w}_{2}}{\left\|\mathbf{w}_{1}\right\|\left\|\mathbf{w}_{2}\right\|},$
(17)
where $\mathbf{w}_{1}$ and $\mathbf{w}_{2}$ are the weight vectors of $C_{1}$
and $C_{2}$, respectively.
### IV-D CALI Training
We integrate the training processes of domain alignment and class alignment to
systematically train our CALI model. To be consistent with Eq. (11), we adopt
an iterative mechanism that alternates between domain alignment and class
alignment. Our training process is pseudo-coded in Algorithm 1.
1 Input: Source dataset $\mathcal{U}_{s}$; Target dataset $\mathcal{U}_{t}$;
Initial model $G,C_{1},C_{2}$ and $D$; Maximum iterations $M$; Iteration
interval $I$.
2 Output: Updated model parameters $\phi_{G},\theta_{C_{1}},\theta_{C_{2}}$
and $\psi_{D}$
3 Initialization: is_domain=True; is_class=False;
4 for _m $\leftarrow$ 1 to M_ do
5 if _$m\%I==0$ and $m\neq 0$_ then
6 is_domain = not is_domain;
7 is_class = not is_class;
8
// Eq. (12)
9
$\min_{\phi_{G},\theta_{C_{1}},\theta_{C_{2}}}\mathcal{L}_{seg}(G,C_{1},C_{2});$
// Eq. (17)
10 $\min_{\theta_{C_{1}},\theta_{C_{2}}}\mathcal{WR}(C_{1},C_{2});$
11 if _is_domain_ then
// Eq. (13)
12 $\max_{\psi_{D}}\min_{\theta_{G}}\mathcal{V}_{1}(G,D);$
13
14 if _is_class_ then
// Eq. 16
15
$\max_{\theta_{C_{1}},\theta_{C_{2}}}\min_{\phi_{G}}\mathcal{V}_{2}(G,C_{1},C_{2});$
16
17
Return $\phi_{G}$, $\theta_{C_{1}}$, $\theta_{C_{2}}$ and $\psi_{D}$;
Algorithm 1 CALI Training Process
Note the adversarial training order of $\mathcal{V}_{1}$ in Algorithm 1 is
$\max_{\psi_{D}}\min_{\phi_{G}}$, instead of the
$\min_{\phi_{G}}\max_{\psi_{D}}$, meaning in each training iteration we first
train the feature extractor and then the discriminator. The reason for this
order is because we empirically find that the feature from $G$ is relatively
easy for $D$ to discriminate, hence if we train $D$ first, then the $D$ might
become an accurate discriminator in the early stage of training and there will
be no adversarial signals for training $G$, thus making the whole training
fail. The same order applies to training of the pair of $G$ and $Cs$ with
$\mathcal{V}_{2}$.
Figure 5: Qualitative results on adaptation GTA5$\rightarrow$Cityscapes.
Results of our proposed model is listed in the last second column. GT
represents the ground-truth labels.
### IV-E Visual Planner
We design a visual receding horizon planner to achieve feasible visual
navigation by combining the learned image segmentation. Specifically, first we
compute a library of motion primitives [14, 15]
$\mathcal{M}=\left\\{\mathbf{p}_{1},\mathbf{p}_{2},\cdots,\mathbf{p}_{n}\right\\}$
where each
$\mathbf{p}_{*}=\left\\{\mathbf{x}_{1},\mathbf{x}_{2},\cdots,\mathbf{x}_{m}\right\\}$
is a single primitive. We use
$\mathbf{x}_{*}=\begin{bmatrix}x&y&\psi\end{bmatrix}^{T}$ to denote a robot
pose. Then we project the motion primitives to the image plane and compute the
navigation cost function for each primitive based on the evaluation of
collision risk in image space and target progress. Finally, we select the
primitive with minimal cost to execute. The trajectory selection problem can
be defined as:
$\mathbf{p}_{optimal}=\underset{\mathbf{p}}{\text{argmin}}~{}w_{1}\cdot
C_{c}(\mathbf{p})+w_{2}\cdot C_{t}(\mathbf{p}),$ (18)
where $C_{c}(\mathbf{p})=\sum_{j}^{m}c_{c}^{j}$ and
$C_{t}(\mathbf{p})=\sum_{j}^{m}c_{t}^{j}$ are the collision cost and target
cost of one primitive $\mathbf{p}$, and $w_{1}$, $w_{2}$ are corresponding
weights, respectively.
To efficiently evaluate the collision risk in the learned segmentation images,
we first classify the classes in terms of their navigability, e.g., in off-
road environments, grass and mulch are classified as navigable while tree and
bush are classified as non-navigable. In this case, we are able to extract the
boundary between the navigable space and the non-navigable space. We treat the
boundary part close to the bottom line of the image as the obstacle boundary.
We further use the obstacle boundary to generate a Scaled Euclidean Distance
Field (SEDF), where the values fall in $[0,1]$, representing the risk level at
the pixel position. Examples of different SEDF with different scale factors
can be seen in Fig. 6.
Assume $\mathbf{x}^{j}$ is the $j^{th}$ pose in one primitive and its image
coordinates are $\left(u^{j},v^{j}\right)$, then the collision risk for
$\mathbf{x}^{j}$ is
$c_{c}^{j}=E[u^{j},v^{j}],$ (19)
where $E$ represents the SEDF image.
Figure 6: Different SEDFs with varying scale factors of (a) $\alpha=0.25$, (b)
$\alpha=0.55$ and (c) $\alpha=1.00$. Values range from 0 to 1 by the color
from blue to yellow. Figure 7: Qualitative results on adaptation
RUGD$\rightarrow$RELLIS. Results of our proposed model is listed in the last
second column. GT represents the ground-truth labels.
To evaluate target progress during the navigation progress, we propose to use
the distance on $SE(3)$ as the metric. We define three types of frames: world
frame $F_{w}$, primitive pose frame $F_{pj}$, and goal frame $F_{g}$. The
transformation of $F_{pj}$ in $F_{w}$ is denoted as $\mathbf{T}_{wpj}$ while
that of $F_{g}$ in $F_{w}$ is $\mathbf{T}_{wg}$. A typical approach to
represent the distance is to split a pose into a position and an orientation
and define two distances on $\mathbb{R}^{3}$ and $SO(3)$. Then the two
distances can be fused in a weighted manner with two strictly positive scaling
factors $a$ and $b$ and with an exponent parameter $p\in[1,\infty]$ [5]:
$\displaystyle d(\mathbf{T}_{wpj},\mathbf{T}_{wg})=\Bigg{[}\Bigg{.}$
$\displaystyle a\cdot d_{rot}(\mathbf{R}_{wpj},\mathbf{R}_{wg})^{p}+$ (20)
$\displaystyle b\cdot
d_{trans}(\mathbf{t}_{wpj},\mathbf{t}_{wg})^{p}\Bigg{.}\Bigg{]}^{1/p}.$
We use the Euclidean distance as
$d_{trans}(\mathbf{t}_{wpj},\mathbf{t}_{wg})$, the Riemannian distance over
$SO(3)$ as $d_{rot}(\mathbf{R}_{wpj},\mathbf{R}_{wg})$ and set $p$ as $2$.
Then the distance (target cost) between two transformation matrices can be
defined [25] as:
$\displaystyle c_{t}^{j}$ $\displaystyle=d(\mathbf{T}_{wpj},\mathbf{T}_{wg})$
(21)
$\displaystyle=\left[a\cdot\left\|\log(\mathbf{R}_{wpj}^{-1}\mathbf{R}_{wg})\right\|^{2}+b\cdot\left\|\mathbf{t}_{wpj}-\mathbf{t}_{wg}\right\|^{2}\right]^{1/2}.$
## V Experiments
Figure 8: Qualitative results on adaptation RUGD$\rightarrow$MESH. Results of
our proposed model is listed in the last column.
### V-A Datasets
We evaluate CALI together with several baseline methods on a few challenging
domain adaptation scenarios, where several public datasets, e.g., GTA5 [29],
Cityscapes [8], RUGD [35], RELLIS [17], as well as a small self-collected
dataset, named MESH (see the first column of Fig. 8), are investigated. The
GTA5 dataset contains $24966$ synthesized high-resolution images in the urban
environments from a video game and pixel-wise semantic annotations of 33
classes. The Cityscapes dataset consists of $5000$ finely annotated images
whose label is given for 19 commonly seen categories in urban environments,
e.g., road, sidewalk, tree, person, car, etc. The RUGD and RELLIS are two
recently released datasets that aim to evaluate segmentation performance in
off-road environments. The RUGD and the RELLIS contain 24 and 20 classes with
$8000$ and $6000$ images, respectively. RUGD and RELLIS cover various scenes
like trails, creeks, parks, villages, and puddle terrains. Our dataset, MESH,
includes features like grass, trees (particularly challenging in winter due to
foliage loss and monochromatic colors), mulch, etc. It helps us to further
validate the performance of our proposed model for traversability prediction
in challenging scenes, particularly the off-road environments.
TABLE I: Quantitative comparison of different methods in UDA of GTA5$\rightarrow$Cityscapes. mIoU* represents the average mIoU over all of classes. Class | SO | DA | CA | CALI
---|---|---|---|---
Road | 38.86 | 52.80 | 78.56 | 75.36
Sidewalk | 17.47 | 18.95 | 2.79 | 27.12
Building | 63.60 | 61.73 | 43.51 | 67.00
Sky | 58.08 | 54.35 | 46.59 | 60.49
Vegetation | 67.21 | 64.69 | 41.48 | 67.50
Terrain | 7.63 | 7.04 | 8.37 | 9.56
Person | 16.89 | 15.45 | 13.48 | 15.03
Car | 30.32 | 43.41 | 31.64 | 52.25
Pole | 11.61 | 12.38 | 9.68 | 11.91
mIoU* | 34.63 | 36.76 | 30.68 | 42.91
### V-B Implementation Details
To be consistent with our theoretical analysis, the implementation of CALI
only adopts the necessary indications by Eq. (11). First, Eq. (11) requires
that the input of the two upper bounds (one for DA and the other one for CA)
should be the same. Second, nothing else but only domain classification and
hypotheses discrepancy are involved in Eq. (11) and other related analyses
(Eq. (3) - Eq. (9)). Accordingly, we strictly follow the guidance of our
theoretical analyses. First, CALI performs DA in the intermediate-feature
level ($f$ in Fig. 4), instead of the output-feature level used in [33].
Second, we exclude the multiple additional tricks, e.g., entropy-based and
multi-level features based alignment, and class-ratio priors in [33] and
multi-steps training for feature extractor in [31]. We also implement baseline
methods without those techniques for a fair comparison. To avoid possible
degraded performance bought by a class imbalance in the used datasets, we
regroup those rare classes into classes with a higher pixel ratio. For
example, we treat the building, wall, and fence as the same class; the person
and rider as the same class in the adaptation of GTA5$\rightarrow$Cityscapes.
In the adaptation of RUGD$\rightarrow$RELLIS, we treat the tree, bush, and log
as the same class, and the rock and rockbed as the same class. Details about
remapping can be seen in Fig. 14 and Fig. 15 in Appendix. VI-B.
TABLE II: Quantitative comparison of different methods in UDA of RUGD$\rightarrow$RELLIS. mIoU* is the average mIoU over all of classes. Class | SO | DA | CA | CALI
---|---|---|---|---
Dirt | 0.00 | 0.53 | 3.23 | 0.01
Grass | 64.78 | 61.63 | 65.35 | 67.08
Tree | 40.79 | 45.93 | 41.51 | 55.80
Sky | 45.07 | 67.00 | 2.31 | 72.99
Building | 10.90 | 12.29 | 10.91 | 10.28
mIoU* | 32.31 | 37.48 | 24.66 | 41.23
Figure 9: Target discrepancy changes during training process of (a)
GTA5$\rightarrow$Cityscapes; (b) RUGD$\rightarrow$RELLIS; and (c)
RUGD$\rightarrow$MESH.
Figure 10: Using minmax can cause the collapse of training. Figure 11: An
example of collapsed trained model using minmax.
We use the PyTorch [26] framework for implementation. Training images from
source and target domains are cropped to be half of their original image
dimensions. The batch size is set to 1 and the weights of all batch
normalization layers are fixed. We use the ResNet-101 [11] pretrained on
ImageNet [9] as the model $G$ for extracting features. We use the ASPP module
in DeepLab-V2 [6] as the structure for $C_{1}$ and $C_{2}$. We use the similar
structure in [27] as the discriminator $D$, which consists of 5 convolution
layers with kernel $4\times 4$ and with channel size
$\left\\{64,128,256,512,1\right\\}$ and stride of 2. Each convolution layer is
followed by a Leaky-ReLU [22] parameterized by 0.2, but only the last
convolution layer is follwed by a Sigmoid function. During the training, we
use SGD [4] as the optimizer for $G,C_{1}$ and $C_{2}$ with a momentum of 0.9,
and use Adam [18] to optimize $D$ with $\beta_{1}=0.9,\beta_{2}=0.99$. We set
all SGD optimizers a weight decay of $5\text{e-}4$. The initial learning rates
of all SGDs for performing domain alignment are set to $2.5\text{e-}4$ and the
one of Adam is set as $1\text{e-}4$. For class alignment, the initial learning
rate of SGDs is set to $1\text{e-}3$. All of the learning rates are decayed by
a poly learning rate policy, where the initial learning rate is multiplied by
$(1-\frac{iter}{max\\_iters})^{power}$ with $power=0.9$. All experiments are
conducted on a single Nvidia Geforce RTX 2080 Super GPU.
### V-C Comparative Studies
We present comparative experimental results of our proposed model, CALI,
compared to different baseline methods – Source-Only (SO) method, Domain-
Alignment (DA) [33] method, and Class-Alignment [31] method. Specifically, we
first perform evaluations on a sim2real UDA in city-like environments, where
the source domain is represented by GTA5 while the target domain is the
Cityscapes. Then we consider a transfer of real2real in forest environments,
where the source domain and target domain are set as RUGD and RELLIS,
respectively. All models are trained with full access to the images and labels
in the source domain and with only access to the images in the target domain.
The labels in target datasets are only used for evaluation purposes. Finally,
we further validate our model performance for adapting from RUGD to our self-
collected dataset MESH.
To ensure a fair comparison, all the methods use the same feature extractor
$G$; both DA and CALI have the same domain discriminator $D$; both CA and CALI
have the same two classifiers $C_{1}$ and $C_{2}$. We also use the same
optimizers and optimization-related hyperparameters if any is used for models
under comparison.
We use the mean of Intersection over Union (mIoU) as the metric to evaluate
each class and overall segmentation performance on testing images. IoU is
computed as $\frac{n_{tp}}{n_{tp}+n_{fp}+n_{fn}}$, where
$n_{tp},n_{tn},n_{fp}$ and $n_{fn}$ are true positive, true negative, false
positive and false negative, respectively.
#### V-C1 GTA5$\rightarrow$Cityscapes
Quantitative comparison results of GTA5$\rightarrow$Cityscapes are shown in
Table. I, where segmentations are evaluated on 9 classes (as regrouped in Fig.
14). Our proposed method has significant advantages over multiple baseline
methods for most categories and overall performance (mIoU*).
Figure 12: Navigation behaviors in MESH$\\#1$ environment. The left-most
column: top-down view of the environment; Purple triangle: the starting point;
Blue star: the target point; We also show the segmentation (top row) and
planning results (bottom row) at four different moments during the navigation,
as shown from the second column to the last one. Figure 13: Navigation
behaviors in MESH$\\#2$ environment. Same legends with Fig. 12.
In our testing case, SO achieves the highest score for the class person even
without any domain adaptation. One possible reason for this is the deep
features of the source person and the target person from the model solely
trained on source domain, are already well-aligned. If we try to interfere
this well-aligned relation using unnecessary additional efforts, the target
prediction error might be increased (see the mIoU values of the person from
the other three methods). We call this phenomenon as negative transfer, which
also happens to other classes if we compare SO and DA/CA, e.g., sidewalk,
building, sky, vegetation, and so on. In contrast, CALI maintains an improved
performance compared to either SO or DA/CA. We validate our analytical method
for DA and CA (Section. IV-A) by a comparison between CALI and baselines. This
indicates either single DA or CA is problematic for semantic segmentation,
particularly when we strictly follow what the theory supports and do not
include any other training tricks (that might increase the training complexity
and make the training unstable). This implies that integration of DA and CA is
beneficial to each other with significant improvements, and more importantly,
CALI is well theoretically supported, and the training process is easy and
stable.
Fig. 5 shows the examples of qualitative comparison for UDA of
GTA5$\rightarrow$Cityscapes. We find that CALI prediction is less noisy
compared to the baselines methods as shown in the second and third columns
(sidewalk or car on-road), and shows better completeness (part of the car is
missing, see the fourth column).
#### V-C2 RUGD$\rightarrow$RELLIS
We show quantitative results of RUGD$\rightarrow$RELLIS in Table. II, where
only 5 classes222This is because other classes (in Fig. 15) frequently
appearing in source domain (RUGD) are extremely rare in target domain
(RELLIS), hence no prediction for those classes occurs especially considering
the domain shift. are evaluated. It shows the same trend as Table. I. Both
tables show that CA has the negative transfer (compared with SO) issue for
either sim2real or real2real UDA. However, if we constrain the training of CA
with DA, as in our proposed model, then the performance will be remarkably
improved. Some qualitative results are shown in Fig. 7.
#### V-C3 RUGD$\rightarrow$MESH
Our MESH dataset contains only unlabeled images that restrict us to show only
a qualitative comparison for the UDA of RUGD$\rightarrow$MESH, as shown in
Fig. 8. We have collected data in winter forest environments, which are
significantly different than the images in the source domain (RUGD) -
collected in a different season, e.g., summer or spring. These cross-season
scenarios make the prediction more challenging. However, it is more practical
to evaluate the UDA performance of cross-season scenarios, as we might have to
deploy our robot at any time, even with extreme weather conditions, but our
available datasets might be far from covering every season and every weather
condition. From Fig. 8, we can still see the obvious advantages of our
proposed CALI model over other baselines.
### V-D Discussions
In this section, we aim to discuss our model behaviors in more details.
Specifically, first we will explain the advantages of CALI over CA from the
perspective of training process. Second, we will show the vital influence of
mistakenly using wrong order of adversarial training.
The most important part in CA is the discrepancy between the two classifiers,
which is the only training force for the functionality of CA. It has been
empirically studied in [31] that the target prediction accuracy will increase
as the target discrepancy is decreasing, hence the discrepancy is also an
indicator showing if the training is on the right track. We compare the target
discrepancy changes of CALI and our baseline CA in Fig. 9, where the curves
for the three UDA scenarios are presented from (a) to (c) and we only show the
data before iteration 30k. It can be seen that before around iteration 2k, the
target discrepancy of both CALI and CA are drastically decreasing, but
thereafter, the discrepancy of CA starts to increase. On the other hand, if we
impose a DA constraint over the same CA (iteratively), leading to our proposed
CALI, then the target discrepancy will be decreasing as expected. This
validates that integrating DA and CA will make the training process of CA more
stable, thus improving the target prediction accuracy.
As mentioned in Algorithm 1, we have to use adversarial training order of
$\max_{\psi_{D}}\min_{\phi_{G}}$, instead of $\min_{\phi_{G}}\max_{\psi_{D}}$.
The reason for this is related to our designed net structure. Following the
guidance of Eq. (11), we use the same input to the two classifiers and the
domain discriminator, hence the discriminator has to receive the intermediate-
level feature as the input. If we use the order of
$\min_{\phi_{G}}\max_{\psi_{D}}$ in CALI, then the outputs of the
discriminator will be like Fig. 10, where the domain discriminator of CALI
will quickly converge to the optimal state and it can accurately discriminate
if the feature is from source or target domain. In this case, the adversarial
loss for updating the feature extractor will be near 0, hence the whole
training fails, which is validated by changes of the target discrepancy curve,
as shown in Fig. 10, where the discrepancy value is decreasing in a small
amount in the first few iterations and then quickly increase to a high level
that shows the training is divergent and the model is collapsed. This is also
verified by the prediction results at (and after) around iteration 1k, as
shown in Fig. 11, where the first row is the source images while the second
row is the target images.
### V-E Navigation Missions
To further show the effectiveness of our proposed model for real deployments,
we build a navigation system by combining the proposed CALI (trained with
RUGD$\rightarrow$MESH set-up) segmentation model with our visual planner. We
test behaviors of our navigation system in two different forest environments
(named MESH$\\#1$ in Fig. 12 and MESH$\\#2$ in Fig. 13), where our navigation
system shows high reliability. In navigation tasks, the image resolution is
$[400,300]$, and the inference time for pure segmentation inference is around
$33$ frame per second (FPS). However, since a complete perception system
requires several post-processing steps, such as navigability definition, noise
filtering, Scaled Euclidean Distance Field computation, motion primitive
evaluation and so on, the response time for the whole perception pipeline (in
python) is around $8$ FPS without any engineering optimization. The inference
of segmentation for navigation is performed on an Nvidia Tesla T4 GPU. We set
the linear velocity as $0.3m/s$ and control the angular velocity to track the
selected motion primitive. The path length is $32.26m$ in Fig. 12 and $28.63m$
in Fig. 13. Although the motion speed is slow in navigation tasks, as a proof
of concept and with a very basic motion planner, the system behavior is as
expected, and we have validated that the proposed CALI model is able to well
accomplish the navigation tasks in unstructured environments.
## VI Conclusion
We present CALI, a novel unsupervised domain adaptation model specifically
designed for semantic segmentation, which requires fine-grained alignments in
the level of class features. We carefully investigate the relationship between
a coarse alignment and a fine alignment in theory. The theoretical analysis
guides the design of the model structure, losses, and training process. We
have validated that the coarse alignment can serve as a constraint to the fine
alignment and integrating the two alignments can boost the UDA performance for
segmentation. The resultant model shows significant advantages over baselines
in various challenging UDA scenarios, e.g., sim2real and real2real. We also
demonstrate the proposed segmentation model can be well integrated with our
proposed visual planner to enable highly efficient navigation in off-road
environments.
## References
* Ben-David et al. [2007] Shai Ben-David, John Blitzer, Koby Crammer, Fernando Pereira, et al. Analysis of representations for domain adaptation. _Advances in neural information processing systems_ , 19:137, 2007.
* Ben-David et al. [2010] Shai Ben-David, John Blitzer, Koby Crammer, Alex Kulesza, Fernando Pereira, and Jennifer Wortman Vaughan. A theory of learning from different domains. _Machine learning_ , 79(1):151–175, 2010.
* Blitzer et al. [2008] John Blitzer, Koby Crammer, Alex Kulesza, Fernando Pereira, and Jennifer Wortman. Learning bounds for domain adaptation. 2008\.
* Bottou [2010] Léon Bottou. Large-scale machine learning with stochastic gradient descent. In _Proceedings of COMPSTAT’2010_ , pages 177–186. Springer, 2010\.
* Brégier et al. [2018] Romain Brégier, Frédéric Devernay, Laetitia Leyrit, and James L Crowley. Defining the pose of any 3d rigid object and an associated distance. _International Journal of Computer Vision_ , 126(6):571–596, 2018.
* Chen et al. [2017a] Liang-Chieh Chen, George Papandreou, Iasonas Kokkinos, Kevin Murphy, and Alan L Yuille. Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs. _IEEE transactions on pattern analysis and machine intelligence_ , 40(4):834–848, 2017a.
* Chen et al. [2017b] Liang-Chieh Chen, George Papandreou, Florian Schroff, and Hartwig Adam. Rethinking atrous convolution for semantic image segmentation. _arXiv preprint arXiv:1706.05587_ , 2017b.
* Cordts et al. [2016] Marius Cordts, Mohamed Omran, Sebastian Ramos, Timo Rehfeld, Markus Enzweiler, Rodrigo Benenson, Uwe Franke, Stefan Roth, and Bernt Schiele. The cityscapes dataset for semantic urban scene understanding. In _Proc. of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR)_ , 2016.
* Deng et al. [2009] Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. Imagenet: A large-scale hierarchical image database. In _2009 IEEE conference on computer vision and pattern recognition_ , pages 248–255. Ieee, 2009.
* Ganin et al. [2016] Yaroslav Ganin, Evgeniya Ustinova, Hana Ajakan, Pascal Germain, Hugo Larochelle, François Laviolette, Mario Marchand, and Victor Lempitsky. Domain-adversarial training of neural networks. _The journal of machine learning research_ , 17(1):2096–2030, 2016.
* He et al. [2016] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In _Proceedings of the IEEE conference on computer vision and pattern recognition_ , pages 770–778, 2016.
* Hoffman et al. [2016] Judy Hoffman, Dequan Wang, Fisher Yu, and Trevor Darrell. Fcns in the wild: Pixel-level adversarial and constraint-based adaptation. _arXiv preprint arXiv:1612.02649_ , 2016.
* Hoffman et al. [2018] Judy Hoffman, Eric Tzeng, Taesung Park, Jun-Yan Zhu, Phillip Isola, Kate Saenko, Alexei Efros, and Trevor Darrell. Cycada: Cycle-consistent adversarial domain adaptation. In _International conference on machine learning_ , pages 1989–1998. PMLR, 2018.
* Howard and Kelly [2007] Thomas M Howard and Alonzo Kelly. Optimal rough terrain trajectory generation for wheeled mobile robots. _The International Journal of Robotics Research_ , 26(2):141–166, 2007.
* Howard et al. [2008] Thomas M Howard, Colin J Green, Alonzo Kelly, and Dave Ferguson. State space sampling of feasible motions for high-performance mobile robot navigation in complex environments. _Journal of Field Robotics_ , 25(6-7):325–345, 2008.
* Hoyer et al. [2021] Lukas Hoyer, Dengxin Dai, and Luc Van Gool. Daformer: Improving network architectures and training strategies for domain-adaptive semantic segmentation. _arXiv preprint arXiv:2111.14887_ , 2021.
* Jiang et al. [2020] Peng Jiang, Philip Osteen, Maggie Wigness, and Srikanth Saripalli. Rellis-3d dataset: Data, benchmarks and analysis, 2020.
* Kingma and Ba [2014] Diederik P Kingma and Jimmy Ba. Adam: A method for stochastic optimization. _arXiv preprint arXiv:1412.6980_ , 2014.
* Liu et al. [2017] Ziwei Liu, Xiaoxiao Li, Ping Luo, Chen Change Loy, and Xiaoou Tang. Deep learning markov random field for semantic segmentation. _IEEE transactions on pattern analysis and machine intelligence_ , 40(8):1814–1828, 2017.
* Long et al. [2015] Jonathan Long, Evan Shelhamer, and Trevor Darrell. Fully convolutional networks for semantic segmentation. In _Proceedings of the IEEE conference on computer vision and pattern recognition_ , pages 3431–3440, 2015.
* Luo et al. [2019] Yawei Luo, Liang Zheng, Tao Guan, Junqing Yu, and Yi Yang. Taking a closer look at domain shift: Category-level adversaries for semantics consistent domain adaptation. In _Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition_ , pages 2507–2516, 2019.
* Maas et al. [2013] Andrew L Maas, Awni Y Hannun, Andrew Y Ng, et al. Rectifier nonlinearities improve neural network acoustic models. In _Proc. icml_ , volume 30, page 3. Citeseer, 2013.
* Mei et al. [2020] Ke Mei, Chuang Zhu, Jiaqi Zou, and Shanghang Zhang. Instance adaptive self-training for unsupervised domain adaptation. In _Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part XXVI 16_ , pages 415–430. Springer, 2020.
* Pan and Yang [2009] Sinno Jialin Pan and Qiang Yang. A survey on transfer learning. _IEEE Transactions on knowledge and data engineering_ , 22(10):1345–1359, 2009.
* Park [1995] Frank C Park. Distance metrics on the rigid-body motions with applications to mechanism design. 1995\.
* Paszke et al. [2019] Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, et al. Pytorch: An imperative style, high-performance deep learning library. _Advances in neural information processing systems_ , 32:8026–8037, 2019.
* Radford et al. [2015] Alec Radford, Luke Metz, and Soumith Chintala. Unsupervised representation learning with deep convolutional generative adversarial networks. _arXiv preprint arXiv:1511.06434_ , 2015.
* Ranftl et al. [2021] René Ranftl, Alexey Bochkovskiy, and Vladlen Koltun. Vision transformers for dense prediction. In _Proceedings of the IEEE/CVF International Conference on Computer Vision_ , pages 12179–12188, 2021.
* Richter et al. [2016] Stephan R Richter, Vibhav Vineet, Stefan Roth, and Vladlen Koltun. Playing for data: Ground truth from computer games. In _European conference on computer vision_ , pages 102–118. Springer, 2016.
* Ronneberger et al. [2015] Olaf Ronneberger, Philipp Fischer, and Thomas Brox. U-net: Convolutional networks for biomedical image segmentation. In _International Conference on Medical image computing and computer-assisted intervention_ , pages 234–241. Springer, 2015.
* Saito et al. [2018] Kuniaki Saito, Kohei Watanabe, Yoshitaka Ushiku, and Tatsuya Harada. Maximum classifier discrepancy for unsupervised domain adaptation. In _Proceedings of the IEEE conference on computer vision and pattern recognition_ , pages 3723–3732, 2018.
* Tsai et al. [2018] Yi-Hsuan Tsai, Wei-Chih Hung, Samuel Schulter, Kihyuk Sohn, Ming-Hsuan Yang, and Manmohan Chandraker. Learning to adapt structured output space for semantic segmentation. In _Proceedings of the IEEE conference on computer vision and pattern recognition_ , pages 7472–7481, 2018.
* Vu et al. [2019] Tuan-Hung Vu, Himalaya Jain, Maxime Bucher, Matthieu Cord, and Patrick Pérez. Advent: Adversarial entropy minimization for domain adaptation in semantic segmentation. In _Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition_ , pages 2517–2526, 2019.
* Wang et al. [2020] Haoran Wang, Tong Shen, Wei Zhang, Ling-Yu Duan, and Tao Mei. Classes matter: A fine-grained adversarial approach to cross-domain semantic segmentation. In _European Conference on Computer Vision_ , pages 642–659. Springer, 2020.
* Wigness et al. [2019] Maggie Wigness, Sungmin Eum, John G Rogers, David Han, and Heesung Kwon. A rugd dataset for autonomous navigation and visual perception in unstructured outdoor environments. In _International Conference on Intelligent Robots and Systems (IROS)_ , 2019.
* Wilson and Cook [2020] Garrett Wilson and Diane J Cook. A survey of unsupervised deep domain adaptation. _ACM Transactions on Intelligent Systems and Technology (TIST)_ , 11(5):1–46, 2020.
* Xie et al. [2021] Enze Xie, Wenhai Wang, Zhiding Yu, Anima Anandkumar, Jose M Alvarez, and Ping Luo. Segformer: Simple and efficient design for semantic segmentation with transformers. _arXiv preprint arXiv:2105.15203_ , 2021.
* Yu and Koltun [2015] Fisher Yu and Vladlen Koltun. Multi-scale context aggregation by dilated convolutions. _arXiv preprint arXiv:1511.07122_ , 2015.
* Zhang et al. [2017] Yang Zhang, Philip David, and Boqing Gong. Curriculum domain adaptation for semantic segmentation of urban scenes. In _Proceedings of the IEEE international conference on computer vision_ , pages 2020–2030, 2017.
* Zhang [2021] Youshan Zhang. A survey of unsupervised domain adaptation for visual recognition. _arXiv preprint arXiv:2112.06745_ , 2021.
* Zhao et al. [2017] Hengshuang Zhao, Jianping Shi, Xiaojuan Qi, Xiaogang Wang, and Jiaya Jia. Pyramid scene parsing network. In _Proceedings of the IEEE conference on computer vision and pattern recognition_ , pages 2881–2890, 2017.
* Zheng et al. [2021] Sixiao Zheng, Jiachen Lu, Hengshuang Zhao, Xiatian Zhu, Zekun Luo, Yabiao Wang, Yanwei Fu, Jianfeng Feng, Tao Xiang, Philip HS Torr, et al. Rethinking semantic segmentation from a sequence-to-sequence perspective with transformers. In _Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition_ , pages 6881–6890, 2021.
* Zou et al. [2018] Yang Zou, Zhiding Yu, BVK Kumar, and Jinsong Wang. Unsupervised domain adaptation for semantic segmentation via class-balanced self-training. In _Proceedings of the European conference on computer vision (ECCV)_ , pages 289–305, 2018.
## Appendix
### VI-A Proof of Theorem 2
For a hypothesis $h$,
$\displaystyle\epsilon_{T}(h)$
$\displaystyle\leq\epsilon_{T}(h^{*})+\epsilon_{T}(h,h^{*})$ (22)
$\displaystyle=\epsilon_{T}(h^{*})+\epsilon_{S}(h,h^{*})-\epsilon_{S}(h,h^{*})+\epsilon_{T}(h,h^{*})$
$\displaystyle\leq\epsilon_{T}(h^{*})+\epsilon_{S}(h,h^{*})+|\epsilon_{T}(h,h^{*})-\epsilon_{S}(h,h^{*})|$
$\displaystyle\leq\epsilon_{T}(h^{*})+\epsilon_{S}(h,h^{*})+\frac{1}{2}d_{\mathcal{H}\Delta\mathcal{H}}(\mathcal{D}_{S},\mathcal{D}_{T})$
$\displaystyle\leq\epsilon_{T}(h^{*})+\epsilon_{S}(h)+\epsilon_{S}(h^{*})+\frac{1}{2}d_{\mathcal{H}\Delta\mathcal{H}}(\mathcal{D}_{S},\mathcal{D}_{T})$
$\displaystyle=\epsilon_{S}(h)+\frac{1}{2}d_{\mathcal{H}\Delta\mathcal{H}}(\mathcal{D}_{S},\mathcal{D}_{T})+\epsilon_{S}(h^{*})+\epsilon_{T}(h^{*})$
$\displaystyle=\epsilon_{S}(h)+\frac{1}{2}d_{\mathcal{H}\Delta\mathcal{H}}(\mathcal{D}_{S},\mathcal{D}_{T})+\lambda$
$\displaystyle=\epsilon_{S}(h)+\sup_{h,h^{{}^{\prime}}\in\mathcal{H}}|\text{P}_{\mathbf{x}\sim\mathcal{D}_{S}}\left[h(\mathbf{x})\neq
h^{{}^{\prime}}(\mathbf{x})\right]-$
$\displaystyle~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}\text{P}_{\mathbf{x}\sim\mathcal{D}_{T}}\left[h(\mathbf{x})\neq
h^{{}^{\prime}}(\mathbf{x})\right]|+\lambda$
$\displaystyle=\epsilon_{S}(h)+\sup_{g\in\mathcal{H}\Delta\mathcal{H}}|\text{P}_{\mathbf{x}\sim\mathcal{D}_{S}}\left[g(\mathbf{x})=1)\right]-$
$\displaystyle~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}\text{P}_{\mathbf{x}\sim\mathcal{D}_{T}}\left[g(\mathbf{x})=1\right]|+\lambda$
$\displaystyle=\epsilon_{S}(h)+\sup_{g\in\mathcal{H}\Delta\mathcal{H}}|\text{P}_{\mathbf{x}\sim\mathcal{D}_{S}}\left[g(\mathbf{x})=1)\right]+$
$\displaystyle~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}\text{P}_{\mathbf{x}\sim\mathcal{D}_{T}}\left[g(\mathbf{x})=0\right]-1|+\lambda$
$\displaystyle\leq\epsilon_{S}(h)+\sup_{g\in\mathcal{H}\Delta\mathcal{H}}|\text{P}_{\mathbf{x}\sim\mathcal{D}_{S}}\left[g(\mathbf{x})=1)\right]+$
$\displaystyle~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}\text{P}_{\mathbf{x}\sim\mathcal{D}_{T}}\left[g(\mathbf{x})=0\right]|-\inf_{g\in\mathcal{H}\Delta\mathcal{H}}1+\lambda$
$\displaystyle=\epsilon_{S}(h)+\sup_{g\in\mathcal{H}\Delta\mathcal{H}}|\text{P}_{\mathbf{x}\sim\mathcal{D}_{S}}\left[g(\mathbf{x})=1)\right]+$
$\displaystyle~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}\text{P}_{\mathbf{x}\sim\mathcal{D}_{T}}\left[g(\mathbf{x})=0\right]|+\lambda-1$
$\displaystyle=\epsilon_{S}(h)+\frac{1}{2}d_{\mathcal{H}}(\mathcal{D}_{S},\mathcal{D}_{T})+1+\lambda-1$
$\displaystyle=\epsilon_{S}(h)+\frac{1}{2}d_{\mathcal{H}}(\mathcal{D}_{S},\mathcal{D}_{T})+\lambda,$
where $\lambda=\epsilon_{S}(h^{*})+\epsilon_{T}(h^{*})$ and $h^{*}$ is the
ideal joint hypothesis (see the Definition 2 in Section 4.2 of [2]).
We have the $4^{th}$, and the $8^{th}$ line because of the Lemma 3 [2]; the
$5^{th}$ line because of the Theorem 2 [2]; the last second line because of
the Lemma 2 [2]. We have the $11^{th}$ line because $\sup|f_{1}-f_{2}|=\sup
f_{1}-\inf f_{2}\leq\sup|f_{1}|-\inf
f_{2}.~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}\blacksquare$
### VI-B Remapping of Label Space
We regroup the original label classes according to the semantic similarities
among classes. In GTA5 and Cityscapes, we cluster the building, wall and fence
as the same category; traffic light, traffic sign and pole as the same group;
car, train. bicycle, motorcycle, bus and truck as the same class; and treat
the person and rider as the same one. See Fig. 14. Similarly, we also have
regroupings for classes in RUGD and RELLIS, as can be seen in Fig. 15.
Figure 14: Lable remapping for GTA5$\rightarrow$Cityscapes. Name of each new
group is marked as bold. Figure 15: Lable remapping for
RUGD$\rightarrow$RELLIS and RUGD$\rightarrow$MESH. Name of each new group is
marked as bold.
|
HU-Mathematik-2021-06
HU-EP-21/51
SAGEX-21-37-E
Combinatorial Solution of the
Eclectic Spin Chain
Changrim Ahna, Luke Corcoranb, Matthias Staudachera,b
a Department of Physics, Ewha Womans University,
52 Ewhayeodae-gil, Seodaemun-gu, Seoul 03760, S. Korea
b Institut für Physik, Humboldt-Universität zu Berlin,
Zum Großen Windkanal 2, 12489 Berlin, Germany
ahn<EMAIL_ADDRESS>
Abstract
The one-loop dilatation operator in the holomorphic 3-scalar sector of the
dynamical fishnet theory is studied. Due to the non-unitary nature of the
underlying field theory this operator, dubbed in [1] the eclectic spin chain
Hamiltonian, is non-diagonalisable. The corresponding spectrum of Jordan
blocks leads to logarithms in the two-point functions, which is characteristic
of logarithmic conformal field theories. It was conjectured in [2] that for
certain filling conditions and generic couplings the spectrum of the eclectic
model is equivalent to the spectrum of a simpler model, the hypereclectic spin
chain. We provide further evidence for this conjecture, and introduce a
generating function which fully characterises the Jordan block spectrum of the
simplified model. This function is found by purely combinatorial means and is
simply related to the $q$-binomial coefficient.
###### Contents
1. 1 Introduction and Overview
2. 2 The (Hyper)eclectic Spin Chain
1. 2.1 Hamiltonian
2. 2.2 Translation Operator and Cyclicity Classes
3. 2.3 Spectral Problem
3. 3 Hypereclectic with One Wall
1. 3.1 Warmup Examples
2. 3.2 Generating Function
4. 4 General Hypereclectic
1. 4.1 Warmup Examples
2. 4.2 General $L,M,K$
5. 5 Eclectic Spin Chain and Universality
1. 5.1 Eclectic Spin Chain and Level $S$
2. 5.2 Warmup Example
3. 5.3 General Argument for $K=1$
6. 6 Conclusions and Outlook
7. A Unexpected Shortening
8. B Universality Details for $K=1$
9. C Fine Tuning and Cyclicity Classes
## 1 Introduction and Overview
Integrability of gauge and string theories continues to generate exciting new
types of exactly solvable models, ranging from new spin chains to novel
quantum field theories and hitherto unstudied string theories. Frequently this
stems from deformations and/or subtle limit taking of previously investigated
systems. A case at hand is a certain double-scaling limit of the three-
parameter $\gamma$-deformation of $\mathcal{N}=4$ super Yang-Mills theory that
had originally been proposed in [3, 4]. It partially or fully breaks
R-symmetry and thus also supersymmetry, while apparently retaining
conformality and integrability in the planar limit. Its double-scaling limit
was then proposed in [5], combining a strong imaginary $\gamma$-twist with a
vanishing coupling constant. In this limit all gauge field interactions
decouple and one is left with an un-gauged quantum field theory of scalars
$\phi_{i}$ and fermions $\psi_{i}$, $i=1,2,3$. In light of [6] we refer to the
resulting theory as the dynamical fishnet theory, with interaction Lagrangian
$\displaystyle\mathcal{L}^{\text{int}}_{\text{DFN}}$
$\displaystyle=N_{c}\text{tr}\left(\xi_{1}^{2}\phi^{\dagger}_{2}\phi^{\dagger}_{3}\phi_{2}\phi_{3}+\xi_{2}^{2}\phi^{\dagger}_{3}\phi^{\dagger}_{1}\phi_{3}\phi_{1}+\xi_{3}^{2}\phi^{\dagger}_{1}\phi^{\dagger}_{2}\phi_{1}\phi_{2}\right)$
(1.1)
$\displaystyle+N_{c}\text{tr}\left(i\sqrt{\xi_{2}\xi_{3}}(\psi^{3}\phi_{1}\psi^{2}+\bar{\psi}_{3}\phi_{1}^{\dagger}\bar{\psi}_{2})+\text{cyclic}\right).$
This model can be further simplified by taking
$\xi_{1}=\xi_{2}=0,\xi_{3}\equiv\xi$. In this case we recover the bi-scalar
fishnet theory
$\mathcal{L}_{\text{FN}}^{\text{int}}=\xi^{2}N_{c}\text{tr}\left(\phi^{\dagger}_{1}\phi^{\dagger}_{2}\phi_{1}\phi_{2}\right).$
(1.2)
Notably, the theories (1.1) and (1.2) are non-unitary. However, the chiral
nature of these interactions leads to a vast simplification in the Feynman-
diagrammatic structure of many physical quantities. This raises the hope that
the integrability of these models might be more easily understood from first
principles. Recall that the origin of integrability of undeformed or
$\gamma$-deformed $\mathcal{N}=4$ SYM remains shrouded in mystery. On the
contrary, in the chiral models one often observes recursive structures in the
Feynman graphs, and the associated graph-building operator may sometimes be
shown to possess integrable properties [6, 7, 8]. In some cases, correlation
functions are represented by a single Feynman diagram. An example of this is
the fishnet Feynman integrals. These have been shown to enjoy a Yangian
symmetry [9], which in some cases has been sufficient to bootstrap the
integral [10, 11]. In a four-point limit these fishnet graphs reduce to the
celebrated Basso-Dixon correlators, for which integrability has been studied
from various perspectives [12, 13, 14]. The fishnet theory has also been
argued to possess at strong coupling a holographic dual [15, 16, 17].
As written, the theories (1.1) and (1.2) are not strictly conformal, even in
their planar limit [18, 19]. Double trace couplings are generated upon
renormalisation. However, it has been argued in [6, 20] that these coupling
may be fine-tuned as a function of $\xi^{2}$ such that the overall beta-
function becomes identically zero, while preserving integrability. As a
result, one expects to get an integrable logarithmic conformal field theory.
This is a consequence of the models’ non-unitarity: While the state space is
still reducible, it is no longer decomposable. The logarithmic nature of the
underlying CFT poses curious new challenges for the spectral problem of the
theory [1, 2]. In particular, in certain operator sectors the dilatation
operator is no longer diagonalisable. It is known that this leads to the
appearance of logarithms in the two-point functions [21]. For example, in the
simplest case where the dilatation operator acts on an operator pair
$\mathcal{O}_{1},\mathcal{O}_{2}$ as a $2\times 2$ Jordan cell
$\mathfrak{D}\begin{pmatrix}\mathcal{O}_{1}\\\
\mathcal{O}_{2}\end{pmatrix}=\begin{pmatrix}\Delta&1\\\
0&\Delta\end{pmatrix}\begin{pmatrix}\mathcal{O}_{1}\\\
\mathcal{O}_{2}\end{pmatrix},$ (1.3)
the two-point function can be brought into the form
$\langle\mathcal{O}_{i}(x)\mathcal{O}_{j}(0)\rangle=\frac{c}{|x|^{2\Delta}}\begin{pmatrix}\log
x^{2}&1\\\ 1&0\end{pmatrix}.$ (1.4)
An explicit example of this in the fishnet theory for length $5$ operators is
given in [7].
Logarithmic conformal field theories play an important role in two dimensions
[22]. There, due to their direct connection with two-dimensional statistical
mechanics models, they are of great physical interest. Important examples
include models of self-avoiding walks, polymers, and percolation; for recent
progress see [23] and references therein. Often the logarithmic scaling
violations occurring in these models are of both experimental and theoretical
interest. In fact, their mathematical analysis often shows intricate and novel
features as compared to the non-logarithmic case. In higher dimensions,
logarithmic CFTs have been much less studied. Still, given their success story
in two dimensions, it is natural to suspect that they will also be of
considerable value.
A systematic study of the dilatation operator in strongly twisted
$\mathcal{N}=4$ SYM was initiated in [1]. It was found that the mentioned non-
diagonalisability is ubiquitous in these models, leading to a rich structure
of Jordan cells. It was also pointed out that the standard methods of
integrability largely fail when applied to the model’s non-diagonalisable
sectors. This was then studied in more detail in a particularly simple
setting, namely at one-loop and with three scalars of equal chirality, in [2].
The resulting spin chain was dubbed the eclectic spin chain in [1, 2], and an
even simpler model, the hypereclectic spin chain was proposed, but not solved.
Interestingly, the latter appeared to possess an even richer spectrum of
Jordan decompositions as compared to the one in the generic eclectic model.
This phenomenon was called universality in [2].
The current work seamlessly continues [2], and proceeds to find the exact
solution of the hypereclectic model. Curiously, for the moment this does not
use at all the model’s integrability, but instead combines methods of linear
algebra and combinatorics. As a result we obtained an elegant generating
function for the spectrum of Jordan blocks. It is reminiscent of a partition
function, since it can be obtained by computing a trace over the state space
$\mathcal{Z}(q)=\text{tr}\hskip 1.42271ptq^{\hat{S}^{\prime}}\,,$ (1.5)
where $\hat{S}^{\prime}$ is a certain counting operator, which is diagonal in
the canonical basis of tensor product states of the spin chain, see end of
section 4.2. It uniquely encodes in full generality the sizes and
multiplicities of the Hamiltonian’s Jordan block decomposition:
$\displaystyle\mathcal{Z}(q)=\sum_{j}N_{j}[j]_{q}=N_{1}\,q^{0}+N_{2}\,\left(q^{-\frac{1}{2}}+q^{\frac{1}{2}}\right)+N_{3}\,\left(q^{-1}+q^{0}+q^{1}\right)+\ldots\,,$
(1.6)
where $N_{j}$ is the number of Jordan blocks of length $j$, and $[j]_{q}$ is a
$q$-analog of $j$, cf. (3.38). It is easy to see that the $\\{N_{j}\\}$ are
indeed uniquely fixed once one knows $\mathcal{Z}(q)$. We also derive formulas
expressing $\mathcal{Z}(q)$ more explicitly than (1.5) in terms of
$q$-binomial coefficients. For example, for the case corresponding to the
fishnet interaction Lagrangian (1.2), with $L-M$ fields $\phi_{1}$, $M-1$
fields $\phi_{2}$, and a single, non-interacting third field $\phi_{3}$, we
find for the one-loop spectrum of Jordan blocks in the cyclic sector the
(shifted) $q$-binomial coefficients
$Z_{L,M}(q)=\genfrac{[}{]}{0.0pt}{0}{L-1}{M-1}_{q}=\prod_{k=1}^{M-1}\frac{q^{\frac{L-k}{2}}-q^{-\frac{L-k}{2}}}{q^{\frac{k}{2}}-q^{-\frac{k}{2}}}\,.$
(1.7)
Interestingly, this result is also valid for the dynamical fishnet theory
interaction Lagrangian (1.1) (for generic couplings) due to the phenomenon of
universality already pointed out in [2].
The paper is organised as follows. Section 2 recalls the definitions of the
non-hermitian chiral spin chain models at hand. The hypereclectic spin chain
has a particularly simple Hamiltonian, essentially describing chiral right-
movers on a chain along with a number of impenetrable non-movers, which we
call walls. Section 3 derives the exact solution of this model in the case of
a single wall. The partition function method is introduced, and the solution
is expressed in terms of $q$-binomial coefficients. Section 4 generalises
these findings to an arbitrary number of walls. Section 5 analyses the generic
three-parameter eclectic model, and, for the case of a single wall, sketches a
proof of the universality hypothesis. Some remarks on the general case are
made. We end with the short, concluding section 6, where it is also pointed
out that the most important open issue seems to be our current inability to
use integrability to analyse these integrable models. A few appendices A,B,C
give further technical details.
## 2 The (Hyper)eclectic Spin Chain
In this section we collect basic facts about the models under consideration.
### 2.1 Hamiltonian
We consider local single-trace operators in the holomorphic 3-scalar sector of
the theory (1.1)
$\mathcal{O}_{j_{1},j_{2},\dots,j_{L}}(x)=\text{tr}\left(\phi_{j_{1}}\phi_{j_{2}}\cdots\phi_{j_{L}}(x)\right),\hskip
28.45274ptj_{i}\in\\{1,2,3\\}.$ (2.1)
In $\mathcal{N}=4$ SYM the one-loop dilatation operator in the analogous
sector can be written as a sum over permutation operators and enjoys an
$\mathfrak{su}(3)$ symmetry [24]. In the strongly twisted theory (1.1) this
symmetry is broken and the one-loop dilatation operator
$H_{\text{ec}}:\left(\mathbb{C}^{3}\right)^{\otimes
L}\rightarrow\left(\mathbb{C}^{3}\right)^{\otimes L}$ is a sum over chiral
permutation operators [2]
$H_{\text{ec}}=H_{1}+H_{2}+H_{3}=\sum_{i=1}^{L}\left(\xi_{1}\mathcal{H}_{1}^{i,i+1}+\xi_{2}\mathcal{H}_{2}^{i,i+1}+\xi_{3}\mathcal{H}_{3}^{i,i+1}\right).$
(2.2)
The chiral permutation operators
$\mathcal{H}_{i}:\mathbb{C}^{3}\otimes\mathbb{C}^{3}\rightarrow\mathbb{C}^{3}\otimes\mathbb{C}^{3}$
act as follows:
$\mathcal{H}_{1}\ket{32}=\ket{23},\qquad\mathcal{H}_{2}\ket{13}=\ket{31},\qquad\mathcal{H}_{3}\ket{21}=\ket{12},$
(2.3)
and annihilate all other states. Periodic boundary conditions are implemented
$\mathcal{H}_{i}^{L,L+1}\equiv\mathcal{H}_{i}^{L,1}$. We have simplified the
notation for the states of the spin chain by
$\ket{\phi_{j_{1}}\phi_{j_{2}}\cdots\phi_{j_{L}}}\rightarrow\ket{j_{1}j_{2}\cdots
j_{L}}.$ (2.4)
Therefore the Hamiltonian (2.2) scans a state for neighboring fields in chiral
order $\ket{32},\ket{13},$ or $\ket{21},$ and swaps them to anti-chiral order
$\ket{23},\ket{31},$ and $\ket{12}$ respectively. E.g. we have
$\displaystyle
H_{\text{ec}}\ket{321}=\xi_{1}\ket{231}+\xi_{3}\ket{312}+\xi_{2}\ket{123},$
(2.5) $\displaystyle H_{\text{ec}}\ket{123}=0.$
Setting $\xi_{1}=\xi_{2}=0,\xi_{3}\equiv\xi$ we recover the hypereclectic
model
$H_{3}=\xi\sum_{i=1}^{L}\mathcal{H}_{3}^{i,i+1}.$ (2.6)
The Hamiltonians (2.2) and (2.6) are block diagonal with respect to sectors of
fixed numbers $K$ of $\phi_{3}$ fields, $M-K$ of $\phi_{2}$ fields, and $L-M$
of $\phi_{1}$ fields. We define $V^{L,M,K}$ to be the vector subspace of
$\left(\mathbb{C}^{3}\right)^{\otimes L}$ corresponding to these numbers of
fields. Clearly we have
$\text{dim}\hskip 2.84544ptV^{L,M,K}=\frac{L!}{(L-M)!(M-K)!K!}.$ (2.7)
$H_{3}$ corresponds to the one-loop dilatation operator in the fishnet theory,
where we consider $K$ non-dynamical insertions $\phi_{3}$, which act as walls.
For $K=0$ this operator, although non-Hermitian, is diagonalisable via a
coordinate Bethe ansatz [1]. It corresponds essentially to a chiral version of
the XY-model [25].
### 2.2 Translation Operator and Cyclicity Classes
We can further reduce the state space by considering the translation
invariance of these Hamiltonians. Each $H_{i}$ commutes with the translation
operator $U$
$[H_{i},U]=0,\qquad i=1,2,3,$ (2.8)
where $U$ generates a shift along the chain
$U\ket{j_{1}j_{2}\cdots j_{L-1}j_{L}}=\ket{j_{L}j_{1}j_{2}\cdots j_{L-1}}.$
(2.9)
This further implies $[H_{\text{ec}},U]=0$. Therefore we can choose to work in
a basis where $U$ is diagonal. $U$ has $L$ distinct eigenvalues given by the
$L^{th}$ roots of unity
$\omega_{L}^{k}=e^{2\pi ik/L},\hskip 28.45274ptk=0,1,\dots,L-1.$ (2.10)
The $U$-eigenstates in $V^{L,M,K}$ with eigenvalue $\omega_{L}^{k}$ are said
to be in the $k^{th}$ cyclicity class $V^{L,M,K}_{k}$. The $k=0$ cyclicity
class $V^{L,M,K}_{k=0}$ is known as the cyclic sector. The states in the
$k^{th}$ cyclicity class are easily generated by acting repeatedly on a
reference elementary state111We call single ket states $\ket{j_{1}j_{2}\dots
j_{L}}$ elementary. In general states are linear combinations of these. with
$\omega_{L}^{-k}U$. For example, given $\ket{123}\in V^{3,2,1}$ we can form
the cyclic state
$\ket{123}+U\ket{123}+U^{2}\ket{123}=\ket{123}+\ket{312}+\ket{231},$ (2.11)
and states with $k=1$ or $k=2$
$\ket{123}+\omega_{3}^{-1}U\ket{123}+\omega_{3}^{-2}U^{2}\ket{123}=\ket{123}+e^{-2\pi
i/3}\ket{312}+e^{-4\pi i/3}\ket{231},$ (2.12)
$\ket{123}+\omega_{3}^{-2}U\ket{123}+\omega_{3}^{-4}U^{2}\ket{123}=\ket{123}+e^{-4\pi
i/3}\ket{312}+e^{-8\pi i/3}\ket{231}.$ (2.13)
For a given $L,M,K$ counting the number of states in $V^{L,M,K}$ with a given
cyclicity $k$ requires Pólya counting, see for example [26]. We denote the
states in the $k^{th}$ cyclicity class by
$\ket{j_{1}j_{2}\cdots
j_{L}}_{k}\equiv\sum_{l=0}^{L-1}(w^{-k}U)^{l}\ket{j_{1}j_{2}\cdots
j_{L}}\equiv\mathcal{C}_{k}\ket{j_{1}j_{2}\cdots j_{L}},$ (2.14)
where $\mathcal{C}_{k}$ is an (unnormalised) projector222Note that this
projection may also result in a zero vector.
$\mathcal{C}_{k}^{2}\propto\mathcal{C}_{k}$ onto the $k^{th}$ cyclicity class
$V^{L,M,K}_{k}$. For the hypereclectic spin chain we find it more natural to
consider a so-called static basis, which we describe at the beginning of
section 3.
### 2.3 Spectral Problem
Given a Hermitian Hamiltonian $H$ on an $n$-dimensional Hilbert space, it is
well-known that one can construct an orthonormal $H$-eigenbasis $\psi_{j}$,
$j=1,2,\dots,n$, such that
$H\psi_{j}=E_{j}\psi_{j}\qquad j=1,2,\dots,n,$ (2.15)
where $E_{j}\in\mathbb{C}$ are the (possibly degenerate) eigenvalues of $H$.
For non-Hermitian Hamiltonians diagonalisability is not guaranteed, and indeed
the (hyper)eclectic Hamiltonian is nilpotent and therefore non-diagonalisable
in sectors with $K>0$. In this case there is still an essentially unique333Up
to the ordering of the Jordan blocks. form to which the matrix can be brought,
namely its Jordan normal form. Furthermore, it is exactly the structure of the
Jordan normal form which determines how the logarithms appear in the two-point
functions [21].
Let $H^{L,M,K}_{\text{ec}}$ be the eclectic Hamiltonian (2.2) restricted to
$V^{L,M,K}$. Then there exists a set of generalised eigenstates
$\psi^{m_{j}}_{j}$, $j=1,\dots,N$, $m_{j}=1,\dots,l_{j}$, which satisfy
$H^{L,M,K}_{\text{ec}}\psi_{j}^{k}=\psi_{j}^{k-1},\qquad
H^{L,M,K}_{\text{ec}}\psi_{j}^{1}=0.$ (2.16)
We then say there are $N$ Jordan blocks labelled by $j$, each of length
$l_{j}$. We call $\psi_{j}^{l_{j}}$ the top state of the $j^{\text{th}}$
block. Each block has a true eigenstate $\psi^{1}_{j}$ of $H$ with eigenvalue
$0$. In more general situations each block has a generalised eigenvalue
$\mathcal{E}_{j}$ associated to it. However, in our case we have
$\mathcal{E}_{j}=0$ for each $j$ since $H^{L,M,K}_{\text{ec}}$ is nilpotent.
On a Jordan block of length $l$, $H^{L,M,K}_{\text{ec}}$ acts as the $l\times
l$ matrix
$J_{l}=\left(\begin{array}[]{ccccc}0&1&&&0\\\ &0&1&&\\\ &&0&\ddots&\\\
&&&\ddots&1\\\ 0&&&&0\end{array}\right).$ (2.17)
For the rest of this paper our goal will be to determine the Jordan block
spectrum of $H_{\text{ec}}$ as a function of the sector labels $L,M,K$. This
means finding the length and multiplicities of each of the blocks. For
example, consider the sector $L=5,M=3,K=1$, which contains $30$ total states.
For generic values444Interestingly, the couplings can be tuned to give a finer
Jordan block decomposition, see appendix C. of the couplings $\xi_{i}$ there
are $5$ Jordan blocks of length 5, and $5$ Jordan blocks of length 1. We
denote this as
$\text{JNF}_{531}=(5^{5},1^{5}).$ (2.18)
## 3 Hypereclectic with One Wall
In this section we describe a method to determine the full Jordan block
spectrum for the hypereclectic spin chain in sectors where $K=1$, i.e. there
is a single, non-moving $\phi_{3}$ field, which acts as a fixed wall. In these
sectors the model is equivalent to a chiral XY spin chain with open boundary
conditions. The sizes and multiplicities of the Jordan blocks can be read off
very simply from a generating function $Z_{L,M}(q)$, and the states of the
Jordan blocks are determined by algorithmic methods. Throughout this section
we denote the hypereclectic Hamiltonian $H_{3}\equiv H$ and set $\xi=1$.
Since the $\phi_{3}$ field does not move under the action of $H$, we can
further restrict to sectors with a fixed position of $\phi_{3}$. We will
restrict to static states of the form $\ket{j_{1}j_{2}\cdots j_{L-1}3}$, where
$j_{1},j_{2},\dots,j_{L-1}\in\\{1,2\\}$. We will refer to the subspace of
$V^{L,M,1}$ spanned by states of this form as $W^{L,M}$. We can access states
where $\phi_{3}$ is in a different position by acting with the translation
operator $U$, so that the Hilbert space decomposes
$V^{L,M,1}=\bigoplus_{j=0}^{L-1}U^{j}W^{L,M}.$ (3.1)
### 3.1 Warmup Examples
#### General $L$, $M=2$, $K=1$.
The simplest situation is when $M=2$ and $K=1$. This means there are a single
$\phi_{3}$ field, a single $\phi_{2}$ field, and $L-2$ $\phi_{1}$ fields. A
natural basis for $W^{L,2}$ is given by $L-1$ states
$\ket{211\cdots 113},\ket{121\cdots 113},\dots,\ket{111\cdots 123}.$ (3.2)
In this sector the states clearly form a single Jordan block of length $L-1$,
as can be seen by acting with $H$ repeatedly on $\ket{211\cdots 113}$
$\ket{211\cdots 113}\xrightarrow{H}\ket{121\cdots
113}\xrightarrow{H}\cdots\xrightarrow{H}\ket{111\cdots 123}\xrightarrow{H}0.$
(3.3)
We will refer to any state of the form $\ket{2^{M-K}1^{L-M}3^{K}}$ as anti-
locked, and $\ket{1^{L-M}2^{M-K}3^{K}}$ as locked. Similarly for the spaces
$U^{j}W^{L,M}$, $j=1,\dots,L-1$ there is a single Jordan block of length
$L-1$. Therefore for $M=2$ and $K=1$ we have
$\text{JNF}_{L,2,1}=(L-1)^{L},$ (3.4)
meaning there are $L$ blocks of length $L-1$.
#### $L=7,M=3,K=1$.
The situation becomes more intricate with increasing $M$, which we illustrate
with the example $L=7,M=3,K=1$. In this sector there are 4 $\phi_{1}$ fields,
$2$ $\phi_{2}$ fields, and a single $\phi_{3}$ field. In $W^{7,3}$ there are
15 states. We use the important observation that the anti-locked state is
always a top state for the longest Jordan block
$\displaystyle\ket{2211113}$ $\displaystyle H^{0}$ (3.5)
$\displaystyle\rightarrow$ $\displaystyle\ket{2121113}$ $\displaystyle H^{1}$
$\displaystyle\rightarrow$ $\displaystyle\ket{2112113}+\ket{1221113}$
$\displaystyle H^{2}$ $\displaystyle\rightarrow$
$\displaystyle\ket{2111213}+2\ket{1212113}$ $\displaystyle H^{3}$
$\displaystyle\rightarrow$
$\displaystyle\ket{2111123}+3\ket{1211213}+2\ket{1122113}$ $\displaystyle
H^{4}$ $\displaystyle\rightarrow$ $\displaystyle
4\ket{1211123}+5\ket{1121213}$ $\displaystyle H^{5}$
$\displaystyle\rightarrow$ $\displaystyle 5\ket{1112213}+9\ket{1121123}$
$\displaystyle H^{6}$ $\displaystyle\rightarrow$ $\displaystyle
14\ket{1112123}$ $\displaystyle H^{7}$ $\displaystyle\rightarrow$
$\displaystyle 14\ket{1111223}$ $\displaystyle H^{8}$
$\displaystyle\rightarrow$ $\displaystyle 0,$ $\displaystyle H^{9}$
so we have identified a Jordan block of length 9, whose eigenstate is
proportional to the locked state $\ket{1111223}$. However since there are 15
states in the sector there must be additional Jordan blocks.
We note that each of the 15 elementary states appear in the tower of states
(3.5). We classify these 15 states by where in this state tower they appear,
by defining the level $S$ of a state. We give the anti-locked state
$\ket{2211113}$ $S=8$ and the locked state $\ket{1111223}$ $S=0$. In general,
if an elementary state appears in the row $H^{k}$ of (3.5), we give it
$S=8-k$. One notices that the $S$-value for a state is the total number of 1’s
to the right of each of the 2’s. Defining $W_{S}^{7,3}$ to be the vector
subspace of $W^{7,3}$ spanned by states with level $S$, we get
$W^{7,3}=\bigoplus_{S=0}^{8}W_{S}^{7,3},$ (3.6)
and it is clear that
$H:W_{S}^{7,3}\rightarrow W_{S-1}^{7,3},\qquad HW_{0}^{7,3}=0.$ (3.7)
In light of this, the next natural place to look for a top state of a Jordan
block is in $W^{7,3}_{6}$. This is because a single state from each
$W^{7,3}_{S}$ is already contained in the largest Jordan block, and
$W^{7,3}_{6}$ is the space with largest $S$ with dimension larger than 1. We
thus deduce that the top state for the next Jordan block must be of the form
$\alpha\ket{2112113}+\beta\ket{1221113}\in W^{7,3}_{6},$ (3.8)
where $\alpha\neq\beta$ as we want the state to be linearly independent from
the corresponding state in the length 9 block. We act repeatedly on this state
with $H$ until there is a possible choice for $\alpha$ and $\beta$ which makes
the state vanish
$\displaystyle\alpha\ket{1221113}+\beta\ket{2112113}$
$\displaystyle\rightarrow\beta\ket{2111213}+(\alpha+\beta)\ket{1212113}$
$\displaystyle\rightarrow\beta\ket{2111123}+(\alpha+2\beta)\ket{1211213}+(\alpha+\beta)\ket{1122113}$
$\displaystyle\rightarrow(\alpha+3\beta)\ket{1211123}+(2\alpha+3\beta)\ket{1121213}$
$\displaystyle\rightarrow(2\alpha+3\beta)\ket{1112213}+(3\alpha+6\beta)\ket{1121123}$
$\displaystyle\rightarrow(5\alpha+9\beta)\ket{1112123}.$
We see that this yields a zero vector if $5\alpha+9\beta=0$, for example
$\alpha=-9,\beta=5$. Therefore this chain of states determines a Jordan block
of length 5, with top state $5\ket{2112113}-9\ket{1221113}\in W^{7,3}_{6}$ and
eigenstate $-3\ket{1112213}+3\ket{1121123}\in W^{7,3}_{8-6}=W^{7,3}_{2}$.
There must be a single Jordan block of length 1 remaining, and by state
counting this must be contained in $W_{4}^{7,3}$, since this is the only space
with dimension greater than 2. We make the ansatz for the top state
$\alpha^{\prime}\ket{2111123}+\beta^{\prime}\ket{1211213}+\gamma^{\prime}\ket{1122113}\in
W_{4}^{7,3}.$ (3.9)
This is easily checked to be an eigenstate of $H$ for
$\alpha^{\prime}=-\beta^{\prime}=\gamma^{\prime}=1$ and thus determines a
Jordan block of length 1. The story is identical for the remaining spaces
$U^{j}W^{7,3}$, $j=1,\dots,6$, so the overall Jordan normal form for
$L=7,M=3,K=1$ is
$\text{JNF}_{7,3,1}=(9^{7},5^{7},1^{7}).$ (3.10)
Let us step back and look at the state tower (3.5), from which we can see the
dimensions
$\text{dim}\hskip 1.42271ptW_{S}^{7,3},\hskip 28.45274ptS=0,1,\dots,8$ (3.11)
by counting the number of elementary states in each row. We note that these
dimensions form a diamond, in that they start from 1 at $S=8$, increase to a
maximum of 3 at $S=4$, and decrease symmetrically to 1 at $S=0$. We encode
these dimensions in a generating function
$\bar{Z}_{7,3}(q)=\sum_{S=0}^{8}\text{dim}\hskip
1.42271ptW^{7,3}_{S}q^{S}=1+q+2q^{2}+2q^{3}+3q^{4}+2q^{5}+2q^{6}+q^{7}+q^{8}.$
(3.12)
Because of this diamond structure it is actually possible to deduce the Jordan
block structure in $W^{7,3}$ from the generating function, a purely
combinatorial object, up to some possible subtleties described in the next
section. Given the generating function (3.12) we identify the Jordan block of
length 9 by the degree of the polynomial plus 1. We then subtract
$1+q+q^{2}+\dots+q^{8}$ to represent the fact that there is one state at each
level in this largest block. We then normalise the resulting polynomial to
have lowest power $q^{0}$, and repeat the procedure:
$\displaystyle 1+q+2q^{2}+2q^{3}+3q^{4}+2q^{5}+2q^{6}+q^{7}+q^{8}$ (3.13)
$\displaystyle\rightarrow$ $\displaystyle 1+q+2q^{2}+q^{3}+q^{4}$
$\displaystyle\rightarrow$ $\displaystyle 1,$
from which we deduce the Jordan block spectrum $(9,5,1)$. Therefore in the
next section it will be our goal to generalise the arguments of this section
and compute the generating function $\bar{Z}_{L,M}(q)$ for arbitrary $L,M$.
### 3.2 Generating Function
For general $L,M$ we similarly grade the vector space in the static sector by
the action of $H$
$W^{L,M}=\bigoplus_{S=0}^{S_{\text{max}}}W_{S}^{L,M},$ (3.14)
$H:W_{S}^{L,M}\rightarrow W_{S-1}^{L,M},\qquad HW_{0}^{L,M}=0.$ (3.15)
We have in general $S_{\text{max}}=L_{1}M_{1}$, where $L_{1}\equiv L-M$ is the
number of 1’s in the sector and $M_{1}\equiv M-1$ is the number of 2’s. The
anti-locked state is $\ket{2^{M_{1}}1^{L_{1}}3}\in W_{S_{\text{max}}}^{L,M}$
and the locked state is $\ket{1^{L_{1}}2^{M_{1}}3}\in W_{0}^{L,M}$. In general
an elementary state takes the form
$\ket{n_{1},n_{2},\dots,n_{M_{1}}}\equiv|\underbrace{1\cdots
1}_{n_{0}}\mathbf{2}\underbrace{1\cdots
1}_{n_{1}}\mathbf{2}\underbrace{1\cdots
1}_{n_{2}}\cdots\mathbf{2}\underbrace{1\cdots
1}_{n_{M_{1}}}\mathbf{3}\rangle,$ (3.16)
where $n_{j}$ is the number of $1$’s between the $j^{th}$ and $(j+1)^{th}$ 2\.
Clearly they should satisfy
$\sum_{j=0}^{M_{1}}n_{j}=L-M=L_{1}.$ (3.17)
In this notation we can define the level $S$ for such a state which counts the
number of 1’s on the right hand side of each of the 2’s. Explicitly the state
$\ket{n_{1},n_{2},\dots,n_{M_{1}}}$ defined in (3.16) has
$S=\sum_{j=1}^{M_{1}}jn_{j}.$ (3.18)
As before we define $W_{S}^{L,M}$ to be spanned by elementary states with this
level $S$. The Hamiltonian acts on (3.16) as
$H:\ket{n_{1},n_{2},\dots,n_{M_{1}}}\quad\to\quad\sum_{j=1}^{M_{1}}\ket{n_{1},n_{2},\dots,n_{j-1}+1,n_{j}-1,\dots,n_{M_{1}}}.$
(3.19)
(3.18) and (3.19) make it clear that $H$ decreases $S$ to $S-1$.
We now consider the problem of determining the dimensions of the spaces
$W_{S}^{L,M}$. We would like to determine a generating function
$\bar{Z}_{L,M}(q)=\sum_{S=0}^{S_{\text{max}}}\text{dim}\hskip
1.42271ptW^{L,M}_{S}q^{S}.$ (3.20)
These dimensions $\text{dim}\hskip 1.42271ptW^{L,M}_{S}$ are given by the
number of partitions of the integer $S$ into at most $M_{1}$ parts, each less
than or equal to $L_{1}$. Expressing (3.18) as
$S=(n_{1}+n_{2}+\dots+n_{M_{1}})+(n_{2}+\dots+n_{M_{1}})+\dots+n_{M_{1}}$
(3.21)
one can notice that there is one-to-one correpondence between an elementary
vector in (3.16) and such a restricted partition of $S$ in (3.21). For
example, consider the case of the previous section, $L=7,M=3,K=1$. There were
3 elementary states in $W^{7,3}_{4}$:
$\displaystyle\ket{2111123},\qquad(n_{1}+n_{2},n_{2})=(4,0),$ (3.22)
$\displaystyle\ket{1211213},\qquad(n_{1}+n_{2},n_{2})=(3,1),$
$\displaystyle\ket{1122113},\qquad(n_{1}+n_{2},n_{2})=(2,2).$
These correspond to the partitions of the integer $4$ into at most $M_{1}=2$
parts, where each part is less than or equal to $L_{1}=4$. There are 3 such
partitions $4=4=3+1=2+2$.
Such restricted partitions described above can be generated by Gaussian (or
$q$-) binomial coefficients [27]
$\bar{Z}_{L,M}(q)=\sum_{S=0}^{S_{\text{max}}}\text{dim}\hskip
1.42271ptW^{L,M}_{S}q^{S}=\binom{L-1}{M-1}_{q}=\prod_{k=1}^{M-1}\frac{1-q^{L-k}}{1-q^{k}},$
(3.23)
which is always a polynomial in $q$. Note that if we send $q\rightarrow 1$,
the $q$-binomial reduces to the ordinary binomial coefficient and we have
$\sum_{S=0}^{S_{\text{max}}}\text{dim}\hskip
1.42271ptW^{L,M}_{S}=\binom{L-1}{M-1}=\text{dim}\hskip 1.42271ptW^{L,M},$
(3.24)
as expected because of (3.14). (3.23) generates a list of
dimensions555$\mathbf{d}_{S}=\mathbf{d}_{S}(L_{1},M_{1})$, we suppress the
$L_{1},M_{1}$ dependence for now. $\mathbf{d}_{S}\equiv\text{dim}\hskip
1.42271ptW^{L,M}_{S}$
$(\mathbf{d}_{S_{\rm max}},\mathbf{d}_{S_{\rm
max}-1},\dots,\mathbf{d}_{1},\mathbf{d}_{0})\qquad{\rm
with}\quad\mathbf{d}_{0}=\mathbf{d}_{S_{\rm max}}=1.$ (3.25)
Furthermore, from a property of the $q$-binomial coefficient, the dimensions
are increasing from the left to the right until the midpoint, and decreasing
after that, because of the symmetry
$\mathbf{d}_{S}=\mathbf{d}_{\tilde{S}},\qquad{\tilde{S}}\equiv{S_{\rm
max}-S}.$ (3.26)
For the space $W^{L,M}_{S_{\text{max}}}$, there is only one elementary state
$\psi_{0}\equiv\ket{2^{M_{1}}1^{L_{1}}3}$, the anti-locked state. By
successive action of $H$, a Jordan string of states is generated
$\psi_{0}\stackrel{{\scriptstyle
H}}{{\longrightarrow}}H\psi_{0}\stackrel{{\scriptstyle
H}}{{\longrightarrow}}H^{2}\psi_{0}\stackrel{{\scriptstyle
H}}{{\longrightarrow}}\cdots\stackrel{{\scriptstyle
H}}{{\longrightarrow}}H^{S_{\rm max}}\psi_{0}\stackrel{{\scriptstyle
H}}{{\longrightarrow}}0.$ (3.27)
Therefore, this generates a Jordan block of size $S_{\rm max}+1$, the largest
one.
It turns out that the next dimension $\mathbf{d}_{S_{\rm max}-1}$ in (3.25) is
also one, as can be computed from (3.23). This means $W^{L,M}_{S_{\rm max}-1}$
is spanned by $H\psi_{0}$, the first descendant of the anti-locked state in
(3.27). Therefore there is no other independent vector in $W^{L,M}_{S_{\rm
max}-1}$ which can generate a new Jordan string.
The top state of the second Jordan block arises at the first level $S=S_{1}$
below $S_{\text{max}}$ whose dimension is bigger than $1$. We can form
$\mathbf{d}_{S_{1}}-1$ linearly independent potential top states in
$W_{S_{1}}^{L,M}$, which are linearly independent from the $H$-descendant of
the anti-locked state. We denote these states by $\psi^{(S_{1})}_{j}$
($j=1,\dots,\mathbf{d}_{S_{1}}-1$), and make the ansatz
$\psi^{(S_{1})}_{j}=\sum_{i=1}^{\mathbf{d}_{S_{1}}}\alpha_{j}^{(i)}e_{i}^{(S_{1})},$
(3.28)
where $e_{i}^{(S_{1})}$ are the elementary states in $W^{L,M}_{S_{1}}$.
$\alpha_{j}^{(i)}$ are constants which are determined by the condition that
each $\psi^{(S_{1})}_{j}$ constitutes a top state for a new Jordan block. Each
of these states generates a Jordan string
$\psi^{(S_{1})}_{j}\stackrel{{\scriptstyle
H}}{{\longrightarrow}}H\psi^{(S_{1})}_{j}\stackrel{{\scriptstyle
H}}{{\longrightarrow}}H^{2}\psi^{(S_{1})}_{j}\cdots\stackrel{{\scriptstyle
H}}{{\longrightarrow}}H^{S_{1}-{\tilde{S}}_{1}}\psi^{(S_{1})}_{j}\stackrel{{\scriptstyle
H}}{{\longrightarrow}}0,\quad j=1,\dots,\mathbf{d}_{S_{1}}-1.$ (3.29)
The condition
$H^{S_{1}-{\tilde{S}}_{1}}\psi^{(S_{1})}_{j}\stackrel{{\scriptstyle
H}}{{\longrightarrow}}0$ leads to a linear system of equations for the
$\alpha_{j}^{(i)}$ which can be solved to determine the $\mathbf{d}_{S_{1}}-1$
new top states. These new Jordan blocks each have size
$S_{1}-{\tilde{S}}_{1}+1$. The only possible subtlety is the potential for an
‘unexpected shortening’ of the Jordan block, that is the possibility for the
equation $H^{k}\psi^{(S_{1})}_{j}=0$ to admit a solution in the
$\alpha_{j}^{(i)}$ for some $k<S_{1}-\tilde{S}_{1}+1$. While we have not yet
been able to rigorously disprove shortening in full generality, we have
verified for a large number values of $L$ and $M$ that it does not happen. We
were able to perform these extensive tests thanks to a mathematically more
succinct reformulation of the problem, see appendix A for details. We will
assume that shortening cannot occur for the remainder of this paper.
The third set of Jordan blocks occurs at a level $S_{2}$, which is the largest
integer satisfying $\mathbf{d}_{S_{2}}>\mathbf{d}_{S_{1}}$. Then, as before,
we can form $\mathbf{d}_{S_{2}}-\mathbf{d}_{S_{1}}$ linearly independent
potential top states which are linearly independent from $H$-descendants of
the previous vectors, $\psi_{0}$ and $\psi^{(S_{1})}_{j}$. We make a similar
ansatz for these potential top states
$\psi^{(S_{2})}_{j}=\sum_{i=1}^{\mathbf{d}_{S_{2}}}\beta_{j}^{(i)}e_{i}^{(S_{2})},$
(3.30)
where $\beta_{j}^{(i)}$ are constants. These states create new Jordan strings
$\psi^{(S_{2})}_{j}\stackrel{{\scriptstyle
H}}{{\longrightarrow}}H\psi^{(S_{2})}_{j}\stackrel{{\scriptstyle
H}}{{\longrightarrow}}H^{2}\psi^{(S_{2})}_{j}\cdots\stackrel{{\scriptstyle
H}}{{\longrightarrow}}H^{S_{2}-{\tilde{S}}_{2}}\psi^{(S_{2})}_{j}\stackrel{{\scriptstyle
H}}{{\longrightarrow}}0,\quad
j=1,\dots,\mathbf{d}_{S_{2}}-\mathbf{d}_{S_{1}},$ (3.31)
and the final condition
$H^{S_{2}-{\tilde{S}}_{2}}\psi^{(S_{2})}_{j}\stackrel{{\scriptstyle
H}}{{\longrightarrow}}0$ is solved to determine the constants
$\beta_{j}^{(i)}$. This leads to $\mathbf{d}_{S_{2}}-\mathbf{d}_{S_{1}}$
Jordan blocks of size $S_{2}-{\tilde{S}}_{2}+1$. This procedure can be
continued until it reaches the maximum value of the dimension $\mathbf{d}_{S}$
which occurs at $S=[S_{\rm max}/2]$.
We note that for a given $L,M,$ the dimensions $\mathbf{d}_{S}$ are sufficient
to determine the sizes and multiplicities of the Jordan blocks. For example,
for $L=9,M=5$ we compute using (3.23)
$\displaystyle\bar{Z}_{9,5}(q)=$ $\displaystyle
1+q+2q^{2}+3q^{3}+5q^{4}+5q^{5}+7q^{6}+7q^{7}+8q^{8}$ (3.32)
$\displaystyle+7q^{9}+7q^{10}+5q^{11}+5q^{12}+3q^{13}+2q^{14}+q^{15}+q^{16},$
from which we can identify the Jordan normal form of $H$ in $W^{9,5}$ to be
$(17,13,11,9^{2},5^{2},1)$ using the same procedure666With some practice one
can easily and quickly ‘read off’ the Jordan normal form from the generating
function by visual inspection, i.e. this does not involve any calculations,
just a bit of bookkeeping. as (3.13). We can exhaust the Hilbert space by
application of $U^{j},j=1,\dots,8$, so that overall we have
$\text{JNF}_{9,5,1}=(17^{9},13^{9},11^{9},9^{18},5^{18},1^{9}).$ (3.33)
For higher $K$, see the next section 4, we find it necessary to work with a
slightly modified generating function for the dimensions of $W_{S}^{L,M}$ that
is symmetric under $q\rightarrow q^{-1}$:
$Z_{L,M}(q)=q^{-S_{\text{max}/2}}\bar{Z}_{L,M}(q)\equiv\genfrac{[}{]}{0.0pt}{0}{L-1}{M-1}_{q}.$
(3.34)
For example, (3.12) and (3.32) are modified to
$Z_{7,3}(q)=q^{-4}+q^{-3}+2q^{-2}+2q^{-1}+3+2q+2q^{2}+q^{3}+q^{4},$ (3.35)
$\displaystyle Z_{9,5}(q)=q^{-8}$
$\displaystyle+q^{-7}+2q^{-6}+3q^{-5}+5q^{-4}+5q^{-3}+7q^{-2}+7q^{-1}+8$
(3.36) $\displaystyle+7q+7q^{2}+5q^{3}+5q^{4}+3q^{5}+2q^{6}+q^{7}+q^{8}.$
The modified function also provides an elegant way to determine the sizes and
multiplicities of the Jordan blocks in a sector uniquely. We have
$Z_{L,M}(q)=\sum_{j}N_{j}[j]_{q}=N_{1}\,q^{0}+N_{2}\,\left(q^{-\frac{1}{2}}+q^{\frac{1}{2}}\right)+N_{3}\,\left(q^{-1}+q^{0}+q^{1}\right)+\ldots\,,$
(3.37)
where $N_{j}$ is the number of Jordan blocks of length $j$. $[j]_{q}$ is a
modified $q$-number
$[j]_{q}=\frac{q^{j/2}-q^{-j/2}}{q^{1/2}-q^{-1/2}}=\sum_{k=\frac{-j+1}{2}}^{\frac{j-1}{2}}q^{k}.$
(3.38)
For example $Z_{7,3}(q)$ and $Z_{9,5}(q)$ can also be written
$Z_{7,3}(q)=[1]_{q}+[5]_{q}+[9]_{q},$ (3.39)
$Z_{9,5}(q)=[1]_{q}+2[5]_{q}+2[9]_{q}+[11]_{q}+[13]_{q}+[17]_{q},$ (3.40)
reflecting the Jordan block structures $(9,5,1)$ and
$(17,13,11,9^{2},5^{2},1)$ respectively. A generating function which generates
the Jordan block spectrum of all of $V^{L,M,1}$ can be obtained as a trace
over the Hilbert space
$\mathcal{Z}_{L,M}(q)=\text{tr}\hskip
1.42271ptq^{\hat{S}-S_{\text{max}}/2}=L\genfrac{[}{]}{0.0pt}{0}{L-1}{M-1}_{q},$
(3.41)
where $\hat{S}$ acts on elementary states with well-defined values of $S$
$\hat{S}\ket{S}=S\ket{S},$ (3.42)
and is extended by linearity.
#### Cyclicity classes.
We note that instead of considering states in $U^{j}W^{L,M}$ where the
$\phi_{3}$ field is in a fixed position, we could have considered states in
any cyclicity class $k$. If we replaced the states $\ket{j_{1}j_{2}\cdots
j_{L-1}3}\rightarrow\mathcal{C}_{k}\ket{j_{1}j_{2}\cdots j_{L-1}3}$ for any
$k=0,1,\dots,L-1$ the arguments of this section are unchanged because
$[H,\mathcal{C}_{k}]=0$, where $\mathcal{C}_{k}$ is the unnormalised projector
defined in (2.14). Therefore the Jordan normal form of $H$ is the same in
$W^{L,M}$ and $V^{L,M,1}_{k}$ for any $k$.
## 4 General Hypereclectic
Here we discuss the extension of the previous section to sectors with many
walls, i.e. $K>1$. The main observation is that $K>1$ states behave
essentially like a tensor product of $K$ states with $K=1$. Any elementary
state $v\in V^{L,M,K}$ ending in a 3 can be written
$v=v_{1}\otimes v_{2}\otimes\cdots\otimes v_{K},$ (4.1)
where $v_{i}\in W^{\ell_{i}+m_{i}+1,m_{i}+1}$ are elementary states
themselves. We defined $W^{L,M}$ above (3.1). $\ell_{i}$ denotes the number of
1’s in $v_{i}$ and $m_{i}$ denotes the number of 2’s. The hypereclectic
Hamiltonian $H$ acts on states of the form (4.1) as
$Hv=Hv_{1}\otimes v_{2}\otimes\cdots\otimes v_{K}+v_{1}\otimes
Hv_{2}\otimes\cdots\otimes v_{K}+\cdots+v_{1}\otimes v_{2}\otimes\cdots\otimes
Hv_{K}.$ (4.2)
We define $\boldsymbol{\ell}\equiv(\ell_{1},\dots,\ell_{K})$ and
$\boldsymbol{m}\equiv(m_{1},\dots,m_{K})$, which should satisfy
$\sum_{i=1}^{K}\ell_{i}=L-M=L_{1},\qquad\sum_{i=1}^{K}m_{i}=M-K=M_{1}.$ (4.3)
We will denote the spaces $\bigotimes_{i=1}^{K}W^{\ell_{i}+m_{i}+1,m_{i}+1}$
as subsectors, and picking the vectors $\boldsymbol{\ell},\boldsymbol{m}$
corresponds to a choice of subsector. We consider subsectors
$(\boldsymbol{\ell},\boldsymbol{m})$ satisfying (4.3) which are unique up to
application of the translation operator $U^{j}$. In practise this means we
identify
$(\boldsymbol{\ell},\boldsymbol{m})\sim(\boldsymbol{\ell}^{\prime},\boldsymbol{m}^{\prime})$
if $\boldsymbol{\ell},\boldsymbol{\ell}^{\prime}$ and
$\boldsymbol{m},\boldsymbol{m}^{\prime}$ are related by the same cyclic
permutation $\sigma^{n}$
$(\boldsymbol{\ell},\boldsymbol{m})\sim(\boldsymbol{\ell}^{\prime},\boldsymbol{m}^{\prime})\quad\longleftrightarrow\quad(\boldsymbol{\ell}^{\prime},\boldsymbol{m}^{\prime})=(\sigma^{n}\boldsymbol{\ell},\sigma^{n}\boldsymbol{m}),$
(4.4)
$\sigma(\ell_{1},\ell_{2},\dots,\ell_{K})\equiv(\ell_{2},\dots,\ell_{K},\ell_{1}).$
(4.5)
In this way we can describe all the states in $V^{L,M,K}$ using the
translation operator $U$. Overall we have
$V^{L,M,K}=\bigoplus_{(\boldsymbol{\ell},\boldsymbol{m})/\sim}\bigoplus_{j=1}^{L/S_{\boldsymbol{l},\boldsymbol{m}}}U^{j}\bigotimes_{i=1}^{K}W^{\ell_{i}+m_{i}+1,m_{i}+1},$
(4.6)
where we introduced the symmetry factor for a subsector
$S_{\boldsymbol{l},\boldsymbol{m}}$. The symmetry factor reflects the fact
that some subsectors are especially symmetric with respect to cyclicity. This
occurs when there is an $n<K$ such that
$(\sigma^{n}\boldsymbol{\ell},\sigma^{n}\boldsymbol{m})=(\boldsymbol{\ell},\boldsymbol{m}),$
(4.7)
where $\sigma$ is the cyclic permutation defined in (4.5). In this case we
give the subsector a symmetry factor
$S_{\boldsymbol{\ell},\boldsymbol{m}}=K/n$. For example, let $L=14,M=8,K=4$
and take the subsector $\boldsymbol{\ell}=(2,1,2,1),\boldsymbol{m}=(1,1,1,1)$.
We have $\sigma^{2}\boldsymbol{\ell}=\boldsymbol{\ell}$ and
$\sigma^{2}\boldsymbol{m}=\boldsymbol{m}$ and so
$S_{\boldsymbol{\ell},\boldsymbol{m}}=4/2=2$ in this case.
### 4.1 Warmup Examples
#### $L=7,M=4,K=2$.
We begin with the simple example $L=7,M=4,K=2$. In this sector there are three
$\phi_{1}$ fields, two $\phi_{2}$ fields, two $\phi_{3}$ fields and
$\frac{7!}{3!2!2!}=210$ total states. In table 1 we show the 6 inequivalent
choices of $(\boldsymbol{\ell},\boldsymbol{m})$, which corresponds to the 6
ways to decompose the states into $K=1$ states, on which $H$ acts block
diagonally:
| Form of state | Number of states | $\quad\hskip 2.84544pt\boldsymbol{\ell},\boldsymbol{m}$ | JNF
---|---|---|---|---
$\boldsymbol{1}$ | $\ket{111223}\otimes\ket{3}$ | $10\times 1=10$ | $(3,0),(2,0)$ | $7\oplus 3$
$\boldsymbol{2}$ | $\ket{11123}\otimes\ket{23}$ | $4\times 1=4$ | $(3,0),(1,1)$ | 4
$\boldsymbol{3}$ | $\ket{11223}\otimes\ket{13}$ | $6\times 1=6$ | $(2,1),(2,0)$ | $5\oplus 1$
$\boldsymbol{4}$ | $\ket{1123}\otimes\ket{123}$ | $3\times 2=6$ | $(2,1),(1,1)$ | $3\otimes 2$
$\boldsymbol{5}$ | $\ket{1223}\otimes\ket{113}$ | $3\times 1=3$ | $(1,2),(2,0)$ | 3
$\boldsymbol{6}$ | $\ket{1113}\otimes\ket{223}$ | $1\times 1=1$ | $(3,0),(0,2)$ | 1
Table 1: Decomposition of $L=7,M=4,K=2$ states into $K=1$ states. The 3’s
should be regarded as fixed, whereas the 1’s and 2’s can be permuted within
their ket.
All subsectors except for $\boldsymbol{4}$ behave trivially as a single $K=1$
sector under the action of $H$. Their Jordan normal forms were determined in
the previous section and are listed in the table. We look at states of the
form $\boldsymbol{4}$ in a bit more detail. These states have the form of an
$L=4,M=2,K=1$ state and an $L=3,M=2,K=1$ state glued together, which have
Jordan blocks of size 3 and 2 respectively. The natural ‘anti-locked’ state
comes from gluing together the anti-locked states of the respective $K=1$
parts $\ket{2113213}$. We act successively on this state with $H$
$\displaystyle\ket{2113213}\rightarrow\ket{1213213}+\ket{2113123}$ (4.8)
$\displaystyle\rightarrow\ket{1123213}+2\ket{1213123}\rightarrow
3\ket{1123123}\rightarrow 0,$
which is a Jordan block of length 4. There is a further Jordan block of length
2 obtained by making the ansatz for a new top state
$\gamma_{1}\ket{1213213}+\gamma_{2}\ket{2113123},$ (4.9)
and similarly to the last section this gives a Jordan block of length 2 for
$\gamma_{1}=-1,\gamma_{2}=2$. Thus the Jordan decomposition of the subsector
$\boldsymbol{4}$ is $(4,2)$. Since the Jordan decompositions of the $K=1$
sectors are $(3)$ and $(2)$ respectively, we denote this as $3\otimes
2=4\oplus 2$.
At the level of generating functions, we can deduce the Jordan normal form of
the ‘tensor product’ sectors by multiplying the generating functions of the
corresponding $K=1$ sectors. For example, for the subsector $\boldsymbol{4}$
we have
$\displaystyle
Z_{7,4,2}^{\boldsymbol{4}}(q)=Z_{4,2}(q)Z_{3,2}(q)=(q^{-1}+1+q)(q^{-1/2}+q^{1/2})$
(4.10) $\displaystyle=q^{-3/2}+2q^{-1/2}+2q^{1/2}+q^{3/2},$
from which the Jordan normal form $(4,2)$ can be easily deduced using (3.37).
To obtain the full generating function for each of the subsectors in
$L=7,M=4,K=2$ we can simply add the generating functions for each of the
subsectors $\boldsymbol{1},\boldsymbol{2},\dots,\boldsymbol{6}$
$\displaystyle
Z_{7,4,2}(q)=\sum_{\boldsymbol{i}=\boldsymbol{1}}^{\boldsymbol{6}}Z_{7,4,2}^{\boldsymbol{i}}(q)$
$\displaystyle=q^{-3}+2q^{-2}+2q^{-3/2}+4q^{-1}+3q^{-1/2}+6+3q^{1/2}+4q+2q^{3/2}+2q^{2}+q^{3}.$
(4.11)
Using (3.37) leads to the following Jordan normal form:
$\text{JNF}_{7,4,2}=(7,5,4^{2},3^{2},2,1^{2}).$ (4.12)
In this sector there are no subtleties with cyclicity and the rest of the
Hilbert space can be exhausted by application of the translation operator
$U^{j}$, $j=1,\dots,6$. For each $j$ we have the same argument as before, so
the full Jordan block structure can be obtained as seven copies of (4.12)
$\text{JNF}^{\text{tot}}_{7,4,2}=(7^{7},5^{7},4^{14},3^{14},2^{7},1^{14}).$
(4.13)
At the level of the generating function this can be obtained by multiplying
(4.11) by $L=7$. However, there are cases where cyclic symmetry leads to some
subtleties, as we discuss next.
#### $L=8,M=4,K=2$.
Let us consider the case of $L=8,M=4,K=2$. There are $\frac{8!}{4!4!2!}=420$
states in this sector. Therein one finds an
($\boldsymbol{\ell},\boldsymbol{m}$) subsector that is symmetric with respect
to cyclicity. In table 2 we break the states into $K=1$ states as in the
previous section,
| Form of state | Number of states | $\quad\hskip 2.84544pt\boldsymbol{\ell},\boldsymbol{m}$ | Jordan decomposition
---|---|---|---|---
$\boldsymbol{1}$ | $\ket{1111223}\otimes\ket{3}$ | $15\times 1=15$ | $(4,0),(2,0)$ | $9\oplus 5\oplus 1$
$\boldsymbol{2}$ | $\ket{111223}\otimes\ket{13}$ | $10\times 1=10$ | $(3,1),(2,0)$ | $7\oplus 3$
$\boldsymbol{3}$ | $\ket{111123}\otimes\ket{23}$ | $5\times 1=5$ | $(4,0),(1,1)$ | $5$
$\boldsymbol{4}$ | $\ket{11123}\otimes\ket{123}$ | $4\times 2=8$ | $(3,1),(1,1)$ | $4\otimes 2=5\oplus 3$
$\boldsymbol{5}$ | $\ket{11223}\otimes\ket{113}$ | $6\times 1=6$ | $(2,2),(2,0)$ | $7\oplus 3$
$\boldsymbol{6}$ | $\ket{11113}\otimes\ket{223}$ | $1\times 1=1$ | $(4,0),(0,2)$ | 5
$\boldsymbol{7}$ | $\ket{1123}\otimes\ket{1123}$ | $3\times 3=9$ | $(2,2),(1,1)$ | $3\otimes 3=5\oplus 3\oplus 1$
$\boldsymbol{8}$ | $\ket{1223}\otimes\ket{1113}$ | $3\times 1=3$ | $(1,3),(2,0)$ | $3$
Table 2: Decomposition of $L=8,M=4,K=2$ states into $K=1$ states.
where we replaced $4\otimes 2=5\oplus 3$ and $3\otimes 3=5\oplus 3\oplus 1$ by
multiplying the appropriate $K=1$ generating functions and naively extracting
the resulting Jordan block structures using (3.37). We see that
$\boldsymbol{7}$ is the subsector where the issues with cyclicity emerge. For
the other subsectors we can exhaust the rest of the state space by acting with
$U^{j},j=1,\dots,7$. However for subsector $\boldsymbol{7}$ applying the
translation $U^{4}$ maps the states to a state in the same subsector, which
reflects the fact this subsector has a symmetry factor
$S_{\boldsymbol{\ell},\boldsymbol{m}}=2$. Therefore acting with
$U^{j},j=0,1,\dots,7$ leads to a double counting by a factor of 2. We can
realise this at the level of an overall generating function for the
$L=8,M=4,K=2$ sector by multiplying by
$1/S_{\boldsymbol{\ell},\boldsymbol{m}}=1/2$ for the subsector
$\boldsymbol{7}$
$\mathcal{Z}_{8,4,2}(q)=8(Z_{8,4,2}^{\boldsymbol{1}}+Z_{8,4,2}^{\boldsymbol{2}}+Z_{8,4,2}^{\boldsymbol{3}}+Z_{8,4,2}^{\boldsymbol{4}}+Z_{8,4,2}^{\boldsymbol{5}}+Z_{8,4,2}^{\boldsymbol{6}}+\frac{1}{2}Z_{8,4,2}^{\boldsymbol{7}}+Z_{8,4,2}^{\boldsymbol{8}}).$
(4.14)
We compute (4.14) to be
$\mathcal{Z}_{8,4,2}(q)=8q^{-4}+16q^{-3}+52q^{-2}+80q^{-1}+80q+52q^{2}+16q^{3}+8q^{4}.$
(4.15)
Using (3.37) we identify the Jordan normal form to be
$\text{JNF}^{\text{tot}}_{8,4,2}=(9^{8},7^{8},5^{36},3^{28},1^{28}).$ (4.16)
### 4.2 General $L,M,K$
Here we generalise the observations of the previous subsections to arbitrary
$L,M,K$ sectors. Given an $L,M,K$ sector we consider a subsector
$\bigotimes_{i=1}^{K}W^{\ell_{i}+m_{i}+1,m_{i}+1}$ defined by the vectors
$\boldsymbol{\ell},\boldsymbol{m}$. The anti-locked state takes the form
$\Omega=\ket{(2\cdots 21\cdots 1)_{1}\mathbf{3}(2\cdots 21\cdots
1)_{2}\mathbf{3}\cdots(2\cdots 21\cdots 1)_{K}\mathbf{3}},$ (4.17)
where $(\ell_{j},m_{j})$ are the numbers of $1$’s and $2$’s in the
$j^{\text{th}}$ bracket. Recall that we have
$\quad\sum_{j=1}^{K}\ell_{j}=L_{1}=L-M,\quad\sum_{j=1}^{K}m_{j}=M_{1}=M-K.$
(4.18)
As for $K=1$, we can grade the vector space by the action of $H$
$\bigotimes_{i=1}^{K}W^{\ell_{i}+m_{i}+1,m_{i}+1}=\bigoplus_{S=0}^{S_{\text{max}}}W^{\boldsymbol{\ell},\boldsymbol{m}}_{S},$
(4.19)
where $W^{\boldsymbol{\ell},\boldsymbol{m}}_{S_{\text{max}}}$ is spanned by
the anti-locked state and $H$ lowers the level $S\rightarrow S-1$. By acting
successively with $H$ on $\Omega$, we will arrive at the locked state
$|(1\cdots 12\cdots 2)_{1}\mathbf{3}(1\cdots 12\cdots
2)_{2}\mathbf{3}\cdots(1\cdots 12\cdots 2)_{K}\mathbf{3}\rangle.$ (4.20)
There will be many different configurations in the middle with lower values of
$S$. For the anti-locked state we have
$S=S_{\rm
max}=\boldsymbol{\ell}\cdot\boldsymbol{m}=\sum_{j=1}^{K}\ell_{j}m_{j},$ (4.21)
and so the size of the largest Jordan block in each subsector is $S_{\rm
max}+1$. If we define the number of actions of $H$ on the $j^{\text{th}}$
bracket as $n_{j}$, a general state has a level
$S=\sum_{j=1}^{K}s_{j}=S_{\rm max}-N,\qquad s_{j}=\ell_{j}m_{j}-n_{j},\quad
N=\sum_{j=1}^{K}n_{j},\quad{\rm with}\quad 0\leq n_{j}\leq\ell_{j}m_{j}.$
(4.22)
The anti-locked state has $S=S_{\rm max}$ (or $N=0$) and the locked state has
$S=0$ (or $N=S_{\rm max}$).
Now consider states obtained by acting with $H$ $N$-times on the anti-locked
state,
$H^{N}\Omega=\sum_{n_{1}=0}^{\ell_{1}m_{1}}\cdots\sum_{n_{K}=0}^{\ell_{K}m_{K}}\ket{H^{n_{1}}(2\cdots
21\cdots 1)\mathbf{3}H^{n_{2}}(2\cdots 21\cdots 1)\mathbf{3}\cdots
H^{n_{K}}(2\cdots 21\cdots 1)\mathbf{3}}.$ (4.23)
The number of elementary states generated by each $H^{n_{j}}(2\cdots 21\cdots
1)$ was found in section 3.2 to be
$\mathbf{d}_{l_{j}m_{j}-n_{j}}(\ell_{j},m_{j})$, which appeared as a
coefficient of the $q$-binomial $\binom{\ell_{j}+m_{j}}{m_{j}}_{q}$ as defined
in (3.23). Therefore we can compute the number of elementary states at each
level $S$ to be
$\mathbf{D}^{\boldsymbol{\ell},\boldsymbol{m}}_{S_{\text{max}}-N}\equiv\text{dim}\hskip
1.42271ptW^{\boldsymbol{\ell},\boldsymbol{m}}_{S_{\text{max}}-N}=\sum_{n_{1}=0}^{\ell_{1}m_{1}}\cdots\sum_{n_{K}=0}^{\ell_{K}m_{K}}\prod_{j=1}^{K}\mathbf{d}_{l_{j}m_{j}-n_{j}}(\ell_{j},m_{j}),\qquad{\rm
with}\quad\sum_{j=1}^{K}n_{j}=N.$ (4.24)
This can be recast into a generating function
$\displaystyle\bar{Z}^{\boldsymbol{\ell},\boldsymbol{m}}(q)$ $\displaystyle=$
$\displaystyle\sum_{N=0}^{S_{\text{max}}}\mathbf{D}^{\boldsymbol{\ell},\boldsymbol{m}}_{S_{\text{max}}-N}\,q^{N}=\sum_{N=0}^{S_{\text{max}}}\left[\sum_{n_{1}=0}^{\ell_{1}m_{1}}\cdots\sum_{n_{K}=0}^{\ell_{K}m_{K}}\delta_{N,\sum_{i=1}^{K}n_{i}}\prod_{j=1}^{K}\mathbf{d}_{l_{j}m_{j}-n_{j}}(\ell_{j},m_{j})\right]\,q^{N}$
(4.25) $\displaystyle=$
$\displaystyle\sum_{n_{1}=0}^{\ell_{1}m_{1}}\cdots\sum_{n_{K}=0}^{\ell_{K}m_{K}}\prod_{j=1}^{K}\,\left[\mathbf{d}_{l_{j}m_{j}-n_{j}}(\ell_{j},m_{j})q^{n_{j}}\right]=\prod_{j=1}^{K}\,\left[\prod_{k=1}^{m_{j}}\frac{1-q^{\ell_{j}+m_{j}+1-k}}{1-q^{k}}\right],$
using the expression for $K=1$ in (3.23). This may be expressed through
$q$-binomials as
$\bar{Z}^{\boldsymbol{\ell},\boldsymbol{m}}(q)=\prod_{j=1}^{K}\,\binom{l_{j}+m_{j}}{m_{j}}_{q}.$
(4.26)
It proves that the generating function for an
${\boldsymbol{\ell}},\boldsymbol{m}$ is simply a product of the corresponding
$K=1$ generating functions. For example, if we take $L=13,M=7,K=3$ and
consider the subsector $\boldsymbol{\ell}=(3,2,1),\boldsymbol{m}=(2,1,1)$ we
find
$\displaystyle\bar{Z}^{\boldsymbol{\ell},\boldsymbol{m}}(q)=\binom{5}{2}_{q}\binom{3}{1}_{q}\binom{2}{1}_{q}=\sum_{N=0}^{9}\mathbf{D}^{\boldsymbol{l},\boldsymbol{m}}_{9-N}q^{N}$
(4.27)
$\displaystyle=1+3q+6q^{2}+9q^{3}+11q^{4}+11q^{5}+9q^{6}+6q^{7}+3q^{8}+q^{9}.$
Analagously to the $K=1$ case, we can use (3.13) to determine the Jordan block
spectrum in this subsector
$\text{JNF}^{\boldsymbol{\ell},\boldsymbol{m}}_{13,7,3}=(2^{2},4^{3},6^{3},8^{2},10).$
(4.28)
Since the states belonging to a given partition
${\boldsymbol{\ell}},\boldsymbol{m}$ of $(L_{1},M_{1})$ are not mixed with
those in a different partition, the total Jordan block spectrum is just direct
sum of all the spectrum sets. One can sum over all inequivalent partitions
formally. For this purpose, it is necessary to use the modified $q$-binomial
coefficients defined in (3.34)
$Z^{\boldsymbol{\ell},\boldsymbol{m}}(q)=\prod_{j=1}^{K}\genfrac{[}{]}{0.0pt}{0}{\ell_{j}+m_{j}}{m_{j}}_{q}=\prod_{j=1}^{K}\,q^{-\ell_{j}m_{j}/2}\binom{l_{j}+m_{j}}{m_{j}}_{q}.$
(4.29)
For each $\boldsymbol{\ell},\boldsymbol{m}$ subsector we can exhaust the rest
of the state space by acting with the translation operator
$U^{j},j=1,\dots,L-1$. The arguments of this section do not change in these
cases, and so the overall generating function for a subsector can be obtained
by simply multiplying it by $L$. The only exception is
$\boldsymbol{\ell},\boldsymbol{m}$ subsectors which have a symmetry factor
$S_{\boldsymbol{\ell},\boldsymbol{m}}\neq 1$. Adjusting for this possibility,
we can define the generating function for a whole $L,M,K$ sector as a sum over
inequivalent partitions
$\mathcal{Z}_{L,M,K}(q)=\sum_{(\boldsymbol{\ell},\boldsymbol{m})/\sim}\frac{L}{S_{\boldsymbol{\ell},\boldsymbol{m}}}Z^{\boldsymbol{\ell},\boldsymbol{m}}(q).$
(4.30)
This total generating function gives the complete Jordan block spectrum, as in
(3.37):
$\mathcal{Z}_{L,M,K}(q)=\sum_{j}N_{j}[j]_{q}=N_{1}\,q^{0}+N_{2}\,\left(q^{-\frac{1}{2}}+q^{\frac{1}{2}}\right)+N_{3}\,\left(q^{-1}+q^{0}+q^{1}\right)+\ldots\,.$
(4.31)
As for the $K=1$ case, $\mathcal{Z}_{L,M,K}(q)$ can alternatively be computed
as a trace over the entire Hilbert space
$\mathcal{Z}_{L,M,K}(q)=\text{tr}\hskip
1.42271ptq^{\hat{S}-\hat{S}_{\text{max}}/2},$ (4.32)
where $\hat{S}$ measures the level $S$ of an elementary state and
$\hat{S}_{\text{max}}$ measures
$S_{\text{max}}=\boldsymbol{\ell}\cdot\boldsymbol{m}$ of a state in an
$\boldsymbol{\ell},\boldsymbol{m}$ subsector. Both operators are extended to
the full Hilbert space by linearity. We can define
$\hat{S}^{\prime}\equiv\hat{S}-\hat{S}_{\text{max}}/2$ for brevity.
#### Cyclicity classes.
The expression (4.32), which can also be expressed as (4.30), gives a
generating function that describes the Jordan block spectrum of the
hypereclectic model in an arbitrary sector of operators defined by $L,M,K$.
However, in certain circumstances it might be useful to compute the Jordan
block spectrum in a specific cyclicity class $k$, for example the cyclic
sector $k=0$ relevant to quantum field theory. In this case, remarkably the
formula (4.32) still applies
$\mathcal{Z}^{k}_{L,M,K}(q)=\text{tr}_{k}\hskip
1.42271ptq^{\hat{S}^{\prime}},$ (4.33)
where we take care to trace only over states of a fixed cyclicity $k$.
## 5 Eclectic Spin Chain and Universality
In the previous sections we described a method to find the full Jordan block
spectrum of the hypereclectic model, as opposed to the eclectic model (2.2)
which is our main interest. However, we claim a universality hypothesis: The
Jordan block spectrum of the eclectic model for generic couplings
$\xi_{1},\xi_{2},\xi_{3}$ is identical to that of the hypereclectic model,
provided $L,M,K$ satisfy
$L_{1}=L-M\geq K,\qquad M_{1}=M-K\geq K.$ (5.1)
(5.1) implies that the number of $\phi_{3}$ fields in the sector does not
exceed the number of $\phi_{1}$’s or $\phi_{2}$’s. Without loss of generality
we can further take
$L-M\geq M-K\geq K.$ (5.2)
Throughout this section we will consider (5.2) to be satisfied, otherwise we
can simply relabel the fields so that it is. It is possible to fine-tune the
couplings to break down the Jordan block structure in certain cyclicity
classes, as discussed in appendix C. Since the $\phi_{3}$ fields no longer act
as walls it is useful to work with states of a fixed cyclicity $k$, see
section 2.2. For definiteness in the following examples we will restrict to
the cyclic sector $k=0$, which in addition happens to be the case relevant to
quantum field theory.
### 5.1 Eclectic Spin Chain and Level $S$
Recall that for elementary states in $K=1$ sectors we defined a level $S$,
which corresponds to the total number of 1’s to the right of each of the 2’s
in a state. Here we work with cyclic states
$\ket{j_{1}j_{2}\cdots j_{L-1}3}_{0}\equiv\mathcal{C}_{0}\ket{j_{1}j_{2}\cdots
j_{L-1}3}=\sum_{j=0}^{L-1}U^{j}\ket{j_{1}j_{2}\cdots j_{L-1}3}.$ (5.3)
We define $S$ in an analogous manner for states of the form (5.3). For example
the state $\ket{1211213}_{0}$ has $S=4$. Let us define $V_{S}$ to be the
vector subspace of $V^{L,M,1}$ spanned by cyclic states with level $S$.777We
suppress the $L,M$ dependence of $V_{S}$. We saw previously that the
hypereclectic Hamiltonian maps states in $V_{S}$ to $V_{S-1}$
$H_{3}:V_{S}\rightarrow V_{S-1},\qquad H_{3}V_{0}=0.$ (5.4)
Let us investigate the action of the full eclectic Hamiltonian
$H_{\text{ec}}=H_{1}+H_{2}+H_{3}$ on the vector spaces $V_{S}$. We find that
$H_{1}:V_{S}\rightarrow V_{S-L_{1}},\qquad H_{2}:V_{S}\rightarrow
V_{S-M_{1}}.$ (5.5)
Since $L_{1}\geq M_{1}\geq 1$ (5.5) implies that $H_{1}$ and $H_{2}$ decrease
$S$ for a state by a greater than or equal amount to $H_{3}$. This already
makes plausible that they will not affect the Jordan normal form of $H_{3}$,
since $H_{2}$ and $H_{1}$ will annihilate states faster than $H_{3}$. For
example, consider the anti-locked state $\ket{221113}_{0}\in V_{6}$ for
$L=6,M=3,K=1$, so that $L_{1}=3,M_{1}=2$. Then
$\displaystyle H_{1}\ket{221113}_{0}=\ket{211123}_{0}\in V_{3},$ (5.6)
$\displaystyle H_{2}\ket{221113}_{0}=\ket{122113}_{0}\in V_{4},$
$\displaystyle H_{3}\ket{221113}_{0}=\ket{212113}_{0}\in V_{5}.$
### 5.2 Warmup Example
Let us consider the eclectic model for $L=7,M=3,K=1$. In the hypereclectic
model this sector has the Jordan block spectrum $(9,5,1)$ in $W^{9,5}$. Here
we show that the eclectic model has the same Jordan block spectrum in the
cyclic sector.
#### Top block.
The anti-locked state in the cyclic sector $\ket{2211113}_{0}\in V_{8}$ again
determines a Jordan block of length 9. The first descendant of the anti-locked
state is
$\displaystyle
H_{\text{ec}}\ket{2211113}_{0}=\xi_{1}\ket{2111123}_{0}+\xi_{2}\ket{1221113}_{0}+\xi_{3}\ket{2121113}_{0}.$
(5.7)
Note that the coefficients of $\xi_{1},\xi_{2},$ and $\xi_{3}$ are states with
$S=4,6,$ and $7$ respectively, which reflects equations (5.4) and (5.5). In
general acting with a power of $H_{\text{ec}}$ on $\ket{2211113}_{0}$ gives
$H_{\text{ec}}^{n}\ket{2211113}_{0}=H_{3}^{n}\ket{2211113}_{0}\hskip
5.69046pt+\hskip 5.69046pt\text{lower $S$ states}.$ (5.8)
It is then easy to see that $H^{9}\ket{2211113}_{0}=0$ and thus
$\ket{2211113}_{0}$ is the top state for a Jordan block of length 9, as
before.
#### Middle block.
In the hypereclectic case the top state of the next Jordan block is
$\psi^{(6)}=-9\ket{1221113}_{0}+5\ket{2112113}_{0}\in V_{6},$ (5.9)
which satisfies $H_{3}^{5}\psi^{(6)}=0$. Thus $\psi^{(6)}$ determines a Jordan
block of length 5 for $H_{3}$. However, in this case things are a bit trickier
in the eclectic model. We have
$H_{\text{ec}}^{5}\psi^{(6)}=15\xi_{2}\xi_{3}^{4}\ket{1111223}_{0}\neq 0.$
(5.10)
It is however possible to modify the top state (5.9) by adding states of lower
$S$, such that the residual term (5.10) vanishes. In this case it is
sufficient to add states with $S=5$ to $\psi^{(6)}$. Since $\text{dim}\hskip
1.42271ptV_{5}=2$ we can add 2 states, to arrive at a new top state
$\chi^{(6)}=\psi^{(6)}+\gamma_{1}\ket{1212113}_{0}+\gamma_{2}\ket{2111213}.$
(5.11)
This state satisfies
$H_{\text{ec}}^{5}\psi^{(6)}=(-15\xi_{2}+(5\gamma_{1}+4\gamma_{2})\xi_{3})\xi_{3}^{4}\ket{1111223}_{0},$
(5.12)
which is 0 for $5\gamma_{1}+4\gamma_{2}=15\xi_{2}/\xi_{3}$. Note that this
defines a one-parameter family of top states. Therefore the eclectic model
also has a Jordan block of length 5 in this sector, with a slightly modified
top state (5.11) which contains lower $S=5$ states.
#### Bottom block.
In the hypereclectic model the top state for the final Jordan block is
$\psi^{(4)}=\ket{2111123}_{0}-\ket{1211213}_{0}+\ket{1122113}_{0}\in V_{4},$
(5.13)
which satisfies $H_{3}\psi^{(4)}=0$ and thus determines a Jordan block of
length 1. The action of the eclectic Hamiltonian on this state gives a
residual
$H_{\text{ec}}\psi^{(4)}=-\xi_{1}\ket{1111223}_{0}-\xi_{2}\ket{1112213}_{0}-\xi_{2}\ket{1121123}_{0}\neq
0,$ (5.14)
which consists of states with $S=0$ and $S=2$. As before we can eliminate this
residual by adding states of lower $S$ to the top state (5.13). We first try
to add states with $S=3$, and since $\text{dim}\hskip 1.42271ptV_{3}=2$ we add
2 states
$\chi^{(4)}=\psi^{(4)}+\alpha_{1}\ket{1121213}_{0}+\alpha_{2}\ket{1211123}_{0}.$
(5.15)
We check that for $\alpha_{1}=\xi_{2}/\xi_{3},\alpha_{2}=-2\xi_{2}/\xi_{3}$
the $S=2$ states in the residual (5.14) vanish
$H_{\text{ec}}\chi^{(4)}=-\xi_{1}\ket{1111223}_{0}+\frac{\xi_{2}^{2}}{\xi_{3}}\ket{1112123}_{0},$
(5.16)
which is a new residual consisting of an $S=1$ and an $S=0$ state. These can
be removed by adding $S=2$ states into the top state ansatz
$\bar{\chi}^{(4)}=\chi^{(4)}+\beta_{1}\ket{1112213}_{0}+\beta_{2}\ket{1121123}_{0},$
(5.17)
and setting
$\beta_{1}=\xi_{1}/\xi_{2},\beta_{2}=-\xi_{1}/\xi_{2}-\xi_{2}^{2}/\xi_{3}^{2}$.
With these choices for $\alpha_{i}$ and $\beta_{i}$ we have
$H_{\text{ec}}\bar{\chi}^{(4)}=0,$ (5.18)
and so we have identified the Jordan block of length one in this sector of the
eclectic model. In summary, by taking a top state for the hypereclectic model
at a level $S$, we can manufacture a top state (of a Jordan block of the same
length) for the eclectic model by adding appropriate combinations of states
with lower values of $S$. We will argue that it is always possible to add
these states of lower $S$, thus rendering the Jordan block spectra of the
hypereclectic and eclectic models equivalent.
### 5.3 General Argument for $K=1$
Here we sketch a proof of universality for $K=1$, where the filling condition
(5.1) is trivially satisfied, if all three particles are present in the spin
chain state. We find it useful to first introduce the notion of supereclectic
models. These are intermediate models between the eclectic model
$H_{\text{ec}}$ and the hypereclectic model $H_{3}$, defined by setting only a
single coupling $\xi_{1}$ or $\xi_{2}$ equal to zero
$H_{\text{super},i}=H_{i}+H_{3},\qquad i=1,2.$ (5.19)
For both of these cases it is possible to prove rigorously that
$H_{\text{super},i}$ has the same Jordan normal form as $H_{3}$ for generic
couplings. The general strategy of the proof is reminiscent of the example
given in section 5.2. For the hypereclectic model, at a level $S$ satisfying
$\mathbf{d}_{S}>\mathbf{d}_{S+1}$ we can construct
$\mathbf{d}_{S}-\mathbf{d}_{S+1}$ top states
$\psi^{(S)}=\sum_{j=1}^{\mathbf{d}_{S}}\alpha_{j}^{(S)}e_{j}^{(S)},$ (5.20)
where $\alpha_{j}^{(S)}$ are known coefficients and $e_{j}^{(S)}$ are the
elementary states at level $S$. $\psi^{(S)}$ is the top state for a Jordan
block of length $S-\tilde{S}+1$
$H_{3}^{S-\tilde{S}+1}\psi^{(S)}=0,$ (5.21)
where we recall $\tilde{S}=S_{\text{max}}-S=(L-M)(M-1)-S$. We show that it is
always possible to modify the state by adding a linear combination of states
with lower $S$
$\psi^{(S)}\to\psi_{i}^{(S)}=\psi^{(S)}+\sum_{n=0}^{S-1}\varphi^{(n)}$ (5.22)
where $\varphi^{(n)}\in V_{n}$. The modified state is a top state for a Jordan
block of the same length in the supereclectic model $H_{\text{super},i}$
$H_{\text{super},i}^{S-\tilde{S}+1}\psi_{i}^{(S)}=0,$ (5.23)
which renders the Jordan normal forms of $H_{\text{super},i}$ and $H_{3}$
equivalent for generic couplings. More technical details of this proof are
given in appendix B. This argument can then be slightly modified to motivate
that the Jordan normal forms of $H_{\text{ec}}$ and $H_{3}$ are also
equivalent, see again B, even if we have not yet worked out all details of the
proof.
#### Universality for $K>1$.
It is even more complicated to show the universality for $K>1$. One main
difference from the $K=1$ case of the supereclectic models, as explained in
appendix B, is that the action of $h_{j}$ on $\varphi^{(S)}$ in general
generates several states with differing $S$-values. If we interpret
${\hat{S}}(h_{j}\varphi^{(S)})$ in (B.4) as the largest among these and
replace $L_{1}$ with the associated $\ell_{j}$, the same logic should be
valid, so that one can construct for the supereclectic models all subleading
states in (4.24).
For the eclectic model, however, a critical simplification used in (B.22) is
not valid. While we have extensive numerical evidence for general
universality, we are currently unable to provide a proof. We leave this for
future work.
## 6 Conclusions and Outlook
We introduced a generating function $\mathcal{Z}_{L,M,K}(q)$ that we
conjectured (and partially proved) to fully enumerate the Jordan block
spectrum of the hypereclectic model introduced in [2], for any sector of
particles labelled by $L,M,K$. Interestingly, it takes a form reminiscent of a
partition function, where one traces a certain kind of number operator over
the state space. It may also be expressed as a sum over products of (shifted)
$q$-binomial coefficients, which elegantly reduces to a single $q$-binomial
for the case of one wall, i.e. $K=1$. Furthermore, our approach for
determining this generating function yields an algorithmic method for
generating the states of the Jordan blocks. We also provided further strong
evidence and partial proofs for the validity of the universality hypothesis of
[2], i.e. the claim that the spectrum of the hypereclectic and eclectic models
agree for special filling conditions. This is important, as the hypereclectic
model is much easier to handle combinatorially in comparison with the eclectic
one. Apart from its intrinsic value as a new type of solution for a new type
of spin chain, our results appear to be an important starting point for an in-
depth analysis of the indecomposability properties of the dynamical fishnet
theory, cf. (1.1), an integrable logarithmic conformal field theory in four
dimensions. In this context, note that $q$-binomials are ubiquitous in the
analysis of two-dimensional logarithmic conformal field theories, see for
example [28].
There are a number of gaps in our derivations that call for further research.
Firstly, our combinatorial arguments do not rigorously exclude the possibility
of ‘unexpected shortening’ of Jordan blocks, as explained in appendix A.
Secondly, while we made some progress towards a proof of the universality
conjecture, a full proof is still missing. It is possible that the filling of
these two gaps will require entirely new methods.
In this context, note that our results for these integrable models have not
been derived by directly using integrability. Instead, they have been obtained
by linear algebra arguments combined with combinatorics. Still, note that we
were able to provide rather elegant formulas that clean up and organise to a
large degree the (at first sight) incredibly intricate Jordan block structure
of the models. One therefore wonders whether this, at least to us, rather
astonishing fact is not an indirect manifestation of the integrability of
these non-hermitian spin chain models. Understanding our findings from
integrability is not only an interesting intellectual challenge, but might
eventually allow to fill in the above mentioned gaps and incomplete proofs of
this paper. Also, using integrability might lead to more explicit formulas
than (4.30) for $\mathcal{Z}_{L,M,K}(q)$ for $K>1$.
There are numerous further directions for investigations to consider. An
interesting conceptual question is whether the Jordan block spectrum of other
non-diagonalisable spin chains, integrable or not, can also be described by
similar generating functions. Or else, is this something particular to the
(hyper)eclectic spin chain? There would be a few natural ways to test this.
For example, one could study the dilatation operator in other non-
diagonalisable sectors of (dynamical) fishnet theory. These sectors could
contain derivative fields/fermions, and would be more intricate to analyse.
There are also different strong twist limits of $\mathcal{N}=4$ SYM available,
which should contain new diagonalisable models, see [2]. One could also
consider the dilatation operator in the strong twist limit of ABJM theory
[29]. In this case the first quantum correction to the dilatation operator
appears at two loops, and we expect this would be a chiral version of the
alternating spin chain given in [30].
The results of this paper concern the dilatation operator at one-loop order.
It is natural to ask what might happen at higher loops. The dilatation
operator certainly continues to be nilpotent, and therefore is non-
diagonalisable. It would be interesting to see in detailed generality if and
how the dilatation operator at different loop orders refines the Jordan block
spectrum. And clearly if would be exciting to understand the structure of
Jordan blocks on the non-perturbative level. Note that the QSC approach does
not, in its current form, allow to even address the question [31].
### Acknowledgements
We are very thankful to Moritz Kade for helpful discussions, comments on the
draft, and for writing a very useful Mathematica program for numerically
obtaining the Jordan normal form of the hypereclectic spin chain. We would
like to express our sincere gratitude to the Brain Pool Program of the Korean
National Research Foundation (NRF) under grant 2-2019-1283-001-1 for generous
support of this research. MS thanks Ewha Womans University for hospitality in
this difficult period. This project has received funding from NRF grant (NRF-
2016R1D1A1B02007258) (CA) and from the European Union’s Horizon 2020 research
and innovation programme under the Marie Sklodowska-Curie grant agreement No.
764850 ‘SAGEX’ (LC, MS).
## Appendix A Unexpected Shortening
Here we succinctly reformulate the conditions for the unwanted ‘unexpected
shortening’ described in section 3.2. This might be helpful for eventually
finding a rigorous proof. In any case, it was very useful for the extensive
numerical checking of our conjecture: unexpected shortening cannot happen.
In a sector with general $L,M$, $K=1$, we argued for the existence of a top
state in $W^{L,M}_{S}$, where $S$ was such that
$\mathbf{d}_{S}>\mathbf{d}_{S+1}:$
$\psi^{(S)}=\sum_{i=1}^{\mathbf{d}_{S}}\alpha_{i}e_{i}^{(S)},$ (A.1)
where $\alpha_{i}$ are constants and $e_{i}^{(S)}$ are the elementary states
in $W^{L,M}_{S}$. Acting with a power of $H$ on this state gives
$H^{k}\psi^{(S)}=\sum_{i=1}^{\mathbf{d}_{S-k}}\sum_{j=1}^{\mathbf{d}_{S}}A^{(k)}_{ij}\alpha_{j}e_{i}^{(S-k)}=\sum_{i=1}^{\mathbf{d}_{S-k}}(A^{(k)}\alpha)_{i}e_{i}^{(S-k)},$
(A.2)
where $A^{(k)}$ is a $\mathbf{d}_{S-k}\times\mathbf{d}_{S}$ matrix, and
$\alpha$ is a vector of length $\mathbf{d}_{S}$ with entries $\alpha_{i}$. The
top state $\psi^{(S)}$ defines a Jordan block of length $k$ if
$H^{k}\psi^{(S)}=0$, or equivalently the homogeneous linear system
$A^{(k)}\alpha=0$ (A.3)
admits at least one nontrivial solution in $\alpha$. We claim that the rank of
$A^{(k)}$ is always maximal:
$\text{rank}(A^{(k)})=\text{min}(\mathbf{d}_{S-k},\mathbf{d}_{S}).$ (A.4)
In this case, it is well-known that (A.3) can only admit a nontrivial solution
in $\alpha$ if and only if rank$(A^{(k)})<\mathbf{d}_{S}$. Moreover, the
number of independent nontrivial solutions is
$\mathbf{d}_{S}-\text{rank}(A^{(k)})$. Therefore a nontrivial solution only
exists when $\mathbf{d}_{S-k}<\mathbf{d}_{S}$. This occurs precisely when
$k=S-\tilde{S}+1$, as can be deduced from (3.23). Therefore, if the rank of
$A^{(k)}$ is always maximal, the top state $\psi^{(S)}$ determines
$\mathbf{d}_{S}-\mathbf{d}_{S+1}$ Jordan blocks, each of length
$S-\tilde{S}+1$. We checked the rank of $A^{(k)}$ for all top states and for
all values of $k,$ up to $L=30,M=6,$ and always found it to be maximal, in
line with our conjecture.
## Appendix B Universality Details for $K=1$
In this section we prove that $H_{\text{super},i}$, defined in (5.19), has the
same Jordan block structure as the hypereclectic model $H_{3}$ for $K=1$,
under the assumption discussed in appendix A. Then we describe how to modify
these arguments to include the full eclectic Hamiltonian, and sketch a
possible universality proof for $K=1$.
#### Universality for $H_{\text{super},1}$.
We start with the first supereclectic model defined in (5.19),
$H_{\text{super},1}$. Consider a top vector $\psi^{(S)}$ for the hypereclectic
model at a level $S$. This vector determines a Jordan block of length
$n_{S}\equiv S-\tilde{S}+1$
$H_{3}^{n_{S}}\psi^{(S)}=0.$ (B.1)
We can expand $H_{{\rm super},1}^{n_{S}}$ as
$H_{{\rm
super},1}^{n_{S}}=\sum_{k=0}^{n_{S}}\binom{n_{S}}{k}H_{1}^{k}H_{3}^{n_{S}-k}=H_{3}^{n_{S}}+n_{S}H_{1}H_{3}^{n_{S}-1}+\frac{n_{S}(n_{S}-1)}{2}H_{1}^{2}H_{3}^{n_{S}-2}+\cdots,$
(B.2)
where we have used $[H_{1},H_{3}]=0$. We introduce a shorthand notation
$H_{{\rm super},1}^{n_{S}}=\sum_{j=0}^{n_{S}}h_{j},\quad
h_{0}=H_{3}^{n_{S}},\quad
h_{j}\equiv\binom{n_{S}}{k}H_{1}^{j}H_{3}^{n_{S}-j},\quad j=1,\dots,n_{S}.$
(B.3)
B͡ecause of (5.4) and (5.5) each $h_{j}$ lowers the $S$-value of a state by
$j(L_{1}-1)+n_{S}$. In other words, given a vector $\varphi^{(S)}\in V_{S}$ we
have
$\hat{S}(h_{j}\varphi^{(S)})=S-j(L_{1}-1)-n_{S}=(\tilde{S}-1)-j(L_{1}-1).$
(B.4)
In particular, $h_{j}\varphi^{(S)}=0$ if this value is negative. Now let us
consider the $\tilde{S}$ value of a top vector to be in an interval
$\ell(L_{1}-1)\leq\tilde{S}-1<(\ell+1)(L_{1}-1).$ (B.5)
In this case, all operators $h_{j}$ with $j>\ell$ will annihilate the top
vector and its descendants. Therefore, we may consider only operators
$h_{0},h_{1},\dots,h_{\ell}$ and disregard others in (B.3).
For this $S$ value of the top vector of the hypereclectic model $\psi^{(S)}$,
we claim that we can construct a corresponding top vector $\psi^{(S)}_{1}$ of
the supereclectic model $H_{\text{super},1}$, defined by
$H_{{\rm super},1}^{n_{S}}\psi^{(S)}_{1}=0,$ (B.6)
via the ansatz
$\psi^{(S)}_{1}=\varphi_{0}+\varphi_{1}+\cdots+\varphi_{\ell},\qquad\varphi_{0}=\psi^{(S)},$
(B.7)
if the top vector has ${\tilde{S}}$ which satisfies (B.5). The condition (B.6)
can be written as
$\displaystyle(h_{0}+h_{1}+h_{2}+\cdots+h_{\ell})(\varphi_{0}+\varphi_{1}+\cdots+\varphi_{\ell})$
$\displaystyle=$
$\displaystyle(h_{0}\varphi_{0})+(h_{0}\varphi_{1}+h_{1}\varphi_{0})+\cdots+(h_{0}\varphi_{\ell}+h_{1}\varphi_{\ell-1}+\cdots+h_{\ell}\varphi_{0})+\cdots=0,$
where we have grouped terms in a very particular way. The first term
$h_{0}\varphi_{0}$ in (B) vanishes due to (B.1). Now we want to find
$\varphi_{1}$ in the second bracket from the restriction that it vanishes
$h_{0}\varphi_{1}+h_{1}\varphi_{0}=0.$ (B.9)
Since $\hat{S}(h_{1}\varphi_{0})=(\tilde{S}-1)-(L_{1}-1)$ from (B.4), this
equation should be expressed by elementary vectors with this $S$ value. There
are $\mathbf{d}_{(\tilde{S}-1)-(L_{1}-1)}$ of them, which becomes the number
of constraints.888In fact, this is the maximum number of constraints since
some of the elementary vectors may not appear. This equation also determines
$\hat{S}(\varphi_{1})=\hat{S}(h_{1}\varphi_{0})+n_{S}=S-(L_{1}-1)$. Therefore,
$\varphi_{1}$ can be expressed as a linear combination of
$\mathbf{d}_{S-(L_{1}-1)}$ elementary states. Since
$\mathbf{d}_{S-(L_{1}-1)}=\mathbf{d}_{\tilde{S}+(L_{1}-1)}>\mathbf{d}_{(\tilde{S}-1)-(L_{1}-1)}$,
one can solve coefficients of the linear combination from (B.9) (not always
unique). This proves that we can always find the solution $\varphi_{1}$.
We require the next bracket in (B) to vanish:
$h_{0}\varphi_{2}+h_{1}\varphi_{1}+h_{2}\varphi_{0}=0.$ (B.10)
Again, one can find that
$\hat{S}(h_{1}\varphi_{1})=\hat{S}(h_{2}\varphi_{0})=(\tilde{S}-1)-2(L_{1}-1)$,
from which we determine $\hat{S}(\varphi_{2})=S-2(L_{1}-1)$. Since the maximum
number of constraints is smaller than that of the coefficients due to
$\mathbf{d}_{S-2(L_{1}-1)}>\mathbf{d}_{(\tilde{S}-1)-2(L_{1}-1)}$, one can
find $\varphi_{2}$ from the known vectors $\varphi_{1}$ and $\varphi_{0}$
using (B.10).
One can easily generalise this argument up to the $\ell$-th bracket in (B):
$h_{0}\varphi_{\ell}+h_{1}\varphi_{\ell-1}+\cdots+h_{\ell}\varphi_{0}=0,$
(B.11)
where the vectors $\varphi_{j},\ j=0,\dots,\ell-1$ have already been found in
previous steps. Since $\hat{S}(\varphi_{j})=S-j(L_{1}-1)$ we have
$\hat{S}(h_{j}\varphi_{\ell-j})=(\tilde{S}-1)-\ell(L_{1}-1)$ for
$j=1,\dots,\ell$. This determines $S$-value of the unknown vector
$\varphi_{\ell}$ to be $\hat{S}(\varphi_{\ell})=S-\ell(L_{1}-1)$. Again, the
maximum number of constraints in (B.11) is smaller than the number of
coefficients in the expansion of $\varphi_{\ell}$ in terms of elementary
states, which guarantees that we can always find its solution.
There are more terms which we did not include in the second line of (B), but
it is easy to show they all vanish. For example, the $(\ell+1)$-th bracket
would be
$h_{1}\varphi_{\ell}+\cdots+h_{\ell}\varphi_{1}.$ (B.12)
Their $S$-values should be $(\tilde{S}-1)-(\ell+1)(L_{1}-1)$, which is
negative due to (B.5). This means that all these vectors vanish.
This proves our universality conjecture for the supereclectic model
$H_{\text{super},1}$ by constructing the top vector explicitly as
$\psi^{(S)}_{1}=\psi^{(S)}+\varphi_{1}+\cdots+\varphi_{\ell},$ (B.13)
for $\tilde{S}$ in (B.5).
Because $\tilde{S}\leq S_{\rm max}/2$ ($\tilde{S}\leq S$ by definition), the
interval (B.5) is limited by the maximum value of $\ell$ which is
$\ell_{\rm max}=\left[\frac{L_{1}M_{1}}{2(L_{1}-1)}\right],$ (B.14)
where $[x]$ is the largest integer not exceeding $x$.
#### Universality for $H_{\text{super},2}$.
The second supereclectic model $H_{\text{super},2}$ defined in (5.19) can be
analysed in exactly the same way. Again, one can express
$H_{\text{super},2}^{n_{S}}=\sum_{m=0}^{n_{S}}g_{m},\quad g_{m}\equiv\
\binom{n_{S}}{m}H_{2}^{m}H_{3}^{n_{S}-m},\quad m=0,\dots,n_{S},\quad
g_{0}=h_{0}=H_{3}^{n_{S}}.$ (B.15)
Each $g_{m}$ lowers $S$-values as follows:
$\hat{S}(g_{m}\phi^{(S)})=S-m(M_{1}-1)-n_{S}.$ (B.16)
In the same way as before, a top vector with level $S$ (and corresponding
$\tilde{S}$) with
$m(M_{1}-1)\leq\tilde{S}-1<(m+1)(M_{1}-1),$ (B.17)
we only need to consider terms in (B.15) up to $g_{m}$.
The remaining procedure is identical to the previous case. One can always find
${\tilde{\varphi}}_{k}$ from
${\tilde{\varphi}}_{0},\dots,{\tilde{\varphi}}_{k-1}$ using
$g_{0}{\tilde{\varphi}}_{k}+g_{1}{\tilde{\varphi}}_{k-1}+\dots+g_{k}{\tilde{\varphi}}_{0}=0,\quad
k=1,\dots,m.$ (B.18)
This proves the universality conjecture for $H_{\text{super},2}$ by
constructing the top vector explicitly as
$\psi^{(S)}_{2}=\psi^{(S)}+{\tilde{\varphi}}_{1}+\cdots+{\tilde{\varphi}}_{m},$
(B.19)
for $\tilde{S}$ in (B.17), where $m$ should be limited by the maximum value
$m_{\rm max}=\left[\frac{L_{1}M_{1}}{2(M_{1}-1)}\right].$ (B.20)
#### Universality for General Eclectic Model.
Powers of $H_{\rm ec}$ can be written as
$H_{\rm ec}^{n_{S}}=\sum_{k=0}^{n_{S}}\
\binom{n_{S}}{k}\,(H_{1}+H_{2})^{k}H_{3}^{n_{S}-k}.$ (B.21)
This expression can be simplified greatly by observing that
$H_{1}H_{2}=H_{2}H_{1}=0$ in sectors where $K=1$. This can be seen by acting
with $H_{1}$ on any state
$\displaystyle H_{1}|\mathbf{2}1\cdots 1\mathbf{2}1\cdots
1\cdots\mathbf{2}1\cdots 1\mathbf{3}\rangle=|1\cdots 1\mathbf{2}1\cdots
1\cdots\mathbf{2}1\cdots 1\mathbf{2}\mathbf{3}\rangle.$ (B.22)
Then, $H_{2}$ will annihilate the resulting state since it cannot contain
$1\mathbf{3}$. Therefore we can remove any terms with both $H_{1}$ and $H_{2}$
in the expansion (B.21), which leads to
$H_{\rm
ec}^{n_{S}}=h_{0}+(g_{1}+g_{2}+\cdots+g_{n_{S}})+(h_{1}+h_{2}+\cdots+h_{n_{S}}).$
(B.23)
We can restrict the interval for $\tilde{S}$ by the two relations (B.5) and
(B.17). Since $L_{1}\geq M_{1}$, for a given $\ell$ we can find $m$ such that
$m(M_{1}-1)\leq\ell(L_{1}-1)<(m+1)(M_{1}-1).$ (B.24)
In this case, the intersection of the two intervals is given by
$m(M_{1}-1)\leq\ell(L_{1}-1)\leq\tilde{S}-1<(m+1)(M_{1}-1).$ (B.25)
For these values of $S$, the expansion of power of the eclectic Hamiltonian is
truncated to
$H_{\rm
ec}^{n_{S}}=h_{0}+(g_{1}+g_{2}+\cdots+g_{m})+(h_{1}+h_{2}+\cdots+h_{\ell}).$
(B.26)
We now claim that the top vector of the eclectic model can be always
constructed from the hypereclectic top state $\psi^{(S)}=\varphi_{0}$ as
follows:
$\psi^{(S)}_{\rm
ec}=\varphi_{0}+\sum_{i=1}^{m}{\tilde{\varphi}}_{i}+\sum_{j=1}^{\ell}\varphi_{j}.$
(B.27)
Let us provide the detailed proof for the simplest case $m=2,\ell=1$, with
$2(M_{1}-1)\leq(L_{1}-1)\leq\tilde{S}-1<3(M_{1}-1).$ (B.28)
We will show that the top vector for the eclectic model can be constructed as
$\psi^{(S)}_{\rm
ec}=\varphi_{0}+{\tilde{\varphi}}_{1}+{\tilde{\varphi}}_{2}+\varphi_{1}.$
(B.29)
One can expand $H_{\rm ec}^{n_{S}}\psi_{\text{ec}}^{(S)}=0$ as
$\displaystyle(h_{0}+g_{1}+g_{2}+h_{1})(\varphi_{0}+{\tilde{\varphi}}_{1}+{\tilde{\varphi}}_{2}+\varphi_{1})=(h_{0}\varphi_{0})+(g_{0}{\tilde{\varphi}}_{1}+g_{1}\varphi_{0})+$
(B.30) $\displaystyle+$
$\displaystyle(g_{0}{\tilde{\varphi}}_{2}+g_{1}{\tilde{\varphi}}_{1}+g_{2}\varphi_{0})+(g_{0}\varphi_{1}+g_{1}{\tilde{\varphi}}_{2}+g_{2}{\tilde{\varphi}}_{1}+h_{1}\varphi_{0})+\cdots=0.$
The first three brackets in (B.30) have already been solved for
$H_{\text{super},2}$, therefore we only need to consider the fourth term and
ellipsis. The $S$-values of each term have already been computed as
$\hat{S}(g_{1}{\tilde{\varphi}}_{2})=\hat{S}(g_{2}{\tilde{\varphi}}_{1})=({\tilde{S}}-1)-3(M_{1}-1)<\hat{S}(h_{1}\varphi_{0})=({\tilde{S}}-1)-(L_{1}-1).$
(B.31)
Therefore, $\varphi_{1}$ can be determined from $\varphi_{0}$ in the same way
as for $H_{\text{super},1}$ with additional subleading terms in $S$ from the
known ${\tilde{\varphi}}_{1},{\tilde{\varphi}}_{2}$. The terms in the ellipsis
in (B.30) are
$\cdots=g_{1}\varphi_{1}+g_{2}{\tilde{\varphi}}_{2}+h_{1}{\tilde{\varphi}}_{1}+g_{2}{\varphi}_{1}+h_{1}{\tilde{\varphi}}_{2}+h_{1}{\varphi}_{1}.$
(B.32)
The $S$-values for these vectors are given by
$\displaystyle\hat{S}(g_{i}\varphi_{j})=\hat{S}(h_{j}{\tilde{\varphi}}_{i})=({\tilde{S}}-1)-j(L_{1}-1)-i(M_{1}-1),$
$\displaystyle\hat{S}(h_{i}\varphi_{j})=({\tilde{S}}-1)-(i+j)(L_{1}-1),\quad\hat{S}(g_{i}{\tilde{\varphi}}_{j})=({\tilde{S}}-1)-(i+j)(M_{1}-1).$
(B.33)
It is not difficult to see from (B.28) that all these vectors should vanish
since their $S$-values are all negative.
This procedure can now be generalised in principle to any value of $(\ell,m)$,
although it is hard to give general, explicit expressions, since the mixed
interval depends closely on explicit vaues of $L_{1},M_{1}$. It would be
interesting to complete the details of this sketch of a proof of $K=1$
universality.
## Appendix C Fine Tuning and Cyclicity Classes
Although we have proven the universality hypothesis for generic values of the
couplings $\xi_{i}$ for $K=1$, it is possible to fine-tune the couplings to
destroy the Jordan block structures in a particular cyclicity class. We give a
simple example of this occurring, for the sector $L=5,M=3,K=1$. There are 30
states in this sector:
$\displaystyle\mathcal{C}_{k}\ket{22113},\quad\mathcal{C}_{k}\ket{21213},\quad\mathcal{C}_{k}\ket{12213},$
(C.1)
$\displaystyle\mathcal{C}_{k}\ket{21123},\quad\mathcal{C}_{k}\ket{12123},\quad\mathcal{C}_{k}\ket{11223},$
where $\mathcal{C}_{k}$ is the unnormalised projector defined in (2.14) and
$k=0,1,2,3,4$ labels the cyclicity class. In each cyclicity class $k$ the
hypereclectic model $H_{3}$ has Jordan decomposition $(5,1)$, so that the
overall Jordan decomposition is $(5^{5},1^{5})$. The other models related to
$H_{3}$ by permutations of the fields $H_{1}$ and $H_{2}$ have Jordan
decomposition (3,2,1) in each cyclicity class. For generic $\xi_{i}$ we have
argued that the eclectic Hamiltonian $H_{\text{ec}}=H_{1}+H_{2}+H_{3}$ also
has the Jordan decomposition $(5^{5},1^{5})$, since this sector satisfies the
filling conditions (5.2). Setting $\xi_{3}=0$ leads to a Jordan decomposition
$(3^{5},2^{5},1^{5})$. Interestingly, this decomposition can be further
refined by tuning $\xi_{1}$ and $\xi_{2}$. Let us act with
$H_{\text{ec}}|_{\xi_{3}=0}$ on the top state $\mathcal{C}_{k}\ket{22113}$:
$\displaystyle\mathcal{C}_{k}\ket{22113}$
$\displaystyle\rightarrow\omega^{k}\xi_{1}\mathcal{C}_{k}\ket{21123}+\omega^{-k}\xi_{2}\mathcal{C}_{k}\ket{12213}$
(C.2)
$\displaystyle\rightarrow(\omega^{2k}\xi_{1}^{2}+\omega^{-2k}\xi_{2}^{2})\mathcal{C}_{k}\ket{11223}\rightarrow
0,$
where $\omega=e^{2\pi i/5}$ and we used $\mathcal{C}_{k}U^{\pm
1}\psi=\omega^{\pm k}\mathcal{C}_{k}\psi,[H_{i},\mathcal{C}_{k}]=0$. For
generic couplings this gives a length 3 block in each cyclicity class.
However, if we tune the couplings such that
$\xi_{2}^{2}=-\omega^{4k}\xi_{1}^{2}$ the block splits into a 2-block and a
1-block in this cyclicity class $k$. There are two further top states in this
sector:
$\displaystyle\mathcal{C}_{k}\ket{21213}\rightarrow(\xi_{1}\omega^{k}+\xi_{2}\omega^{-k})\mathcal{C}_{k}\ket{12123}\rightarrow
0,$ (C.3)
$\displaystyle\xi_{2}\omega^{-k}\mathcal{C}_{k}\ket{21123}-\xi_{1}\omega^{k}\mathcal{C}_{k}\ket{12213}\rightarrow
0.$
The first of these is a 2-block, which can be broken into two 1-blocks in a
single cyclicity class if $\xi_{2}=-\omega^{2k}\xi_{1}$. The next of these is
always a 1-block. From this example we see explicitly that finer Jordan block
decompositions can be obtained in specific cyclicity classes by tuning the
couplings appropriately.
## References
* [1] A. C. Ipsen, M. Staudacher and L. Zippelius, “The one-loop spectral problem of strongly twisted $\mathcal{N}$ = 4 Super Yang-Mills theory”, Journal of High Energy Physics 2019, L. Zippelius (2019), arxiv:1812.08794, http://dx.doi.org/10.1007/JHEP04(2019)044.
* [2] C. Ahn and M. Staudacher, “The integrable (hyper)eclectic spin chain”, Journal of High Energy Physics 2021, M. Staudacher (2021), arxiv:2010.14515, http://dx.doi.org/10.1007/JHEP02(2021)019.
* [3] S. Frolov, “Lax pair for strings in Lunin-Maldacena background”, Journal of High Energy Physics 2005, 069 (2005), hep-th/0503201, http://dx.doi.org/10.1088/1126-6708/2005/05/069.
* [4] S. Frolov, R. Roiban and A. Tseytlin, “Gauge-string duality for (non)supersymmetric deformations of super-Yang-Mills theory”, Nuclear Physics B 731, 1 (2005), hep-th/0507021, http://dx.doi.org/10.1016/j.nuclphysb.2005.10.004.
* [5] O. Gurdogan and V. Kazakov, “New Integrable 4D Quantum Field Theories from Strongly Deformed Planar N=4 Supersymmetric Yang-Mills Theory”, Physical Review Letters 117, V. Kazakov (2016), arxiv:1512.06704, http://dx.doi.org/10.1103/PhysRevLett.117.201602.
* [6] V. Kazakov, E. Olivucci and M. Preti, “Generalized fishnets and exact four-point correlators in chiral CFT4”, Journal of High Energy Physics 2019, M. Preti (2019), arxiv:1901.00011, http://dx.doi.org/10.1007/JHEP06(2019)078.
* [7] N. Gromov, V. Kazakov, G. Korchemsky, S. Negro and G. Sizov, “Integrability of conformal fishnet theory”, Journal of High Energy Physics 2018, G. Sizov (2018), arxiv:1706.04167, http://dx.doi.org/10.1007/JHEP01(2018)095.
* [8] N. Gromov, V. Kazakov and G. Korchemsky, “Exact correlation functions in conformal fishnet theory”, Journal of High Energy Physics 2019, G. Korchemsky (2019), arxiv:1808.02688, http://dx.doi.org/10.1007/JHEP08(2019)123.
* [9] D. Chicherin, V. Kazakov, F. Loebbert, D. Müller and D.-l. Zhong, “Yangian symmetry for fishnet Feynman graphs”, Physical Review D 96, D. (2017), arxiv:1708.00007, http://dx.doi.org/10.1103/PhysRevD.96.121901.
* [10] F. Loebbert, D. Müller and H. Münkler, “Yangian bootstrap for conformal Feynman integrals”, Physical Review D 101, H. Münkler (2020), arxiv:1912.05561, http://dx.doi.org/10.1103/PhysRevD.101.066006.
* [11] L. Corcoran, F. Loebbert, J. Miczajka and M. Staudacher, “Minkowski box from Yangian bootstrap”, Journal of High Energy Physics 2021, M. Staudacher (2021), arxiv:2012.07852, http://dx.doi.org/10.1007/JHEP04(2021)160.
* [12] B. Basso and L. J. Dixon, “Gluing Ladder Feynman Diagrams into Fishnets”, Physical Review Letters 119, L. J. Dixon (2017), arxiv:1705.03545, http://dx.doi.org/10.1103/PhysRevLett.119.071601.
* [13] S. Derkachov and E. Olivucci, “Exactly Solvable Magnet of Conformal Spins in Four Dimensions”, Physical Review Letters 125, E. Olivucci (2020), arxiv:1912.07588, http://dx.doi.org/10.1103/PhysRevLett.125.031603.
* [14] B. Basso, L. J. Dixon, D. A. Kosower, A. Krajenbrink and D.-l. Zhong, “Fishnet four-point integrals: integrable representations and thermodynamic limits”, Journal of High Energy Physics 2021, D. (2021), arxiv:2105.10514, http://dx.doi.org/10.1007/JHEP07(2021)168.
* [15] N. Gromov and A. Sever, “Derivation of the Holographic Dual of a Planar Conformal Field Theory in 4D”, Physical Review Letters 123, A. Sever (2019), arxiv:1903.10508, http://dx.doi.org/10.1103/PhysRevLett.123.081602.
* [16] N. Gromov and A. Sever, “Quantum fishchain in AdS5”, Journal of High Energy Physics 2019, A. Sever (2019), arxiv:1907.01001, http://dx.doi.org/10.1007/JHEP10(2019)085.
* [17] N. Gromov and A. Sever, “The holographic dual of strongly $\gamma$-deformed $\mathcal{N}$ = 4 SYM theory - derivation, generalization, integrability and discrete reparametrization symmetry”, Journal of High Energy Physics 2020, A. Sever (2020), arxiv:1908.10379, http://dx.doi.org/10.1007/JHEP02(2020)035.
* [18] J. Fokken, C. Sieg and M. Wilhelm, “Non-conformality of ${{\gamma}_{i}}$-deformed $\mathcal{N}$ = 4 SYM theory”, Journal of Physics A: Mathematical and Theoretical 47, 455401 (2014), arxiv:1308.4420, http://dx.doi.org/10.1088/1751-8113/47/45/455401.
* [19] C. Sieg and M. Wilhelm, “On a CFT limit of planar $\gamma_{i}$-deformed $\mathcal{N}$=4 SYM theory”, Physics Letters B 756, 118 (2016), arxiv:1602.05817, http://dx.doi.org/10.1016/j.physletb.2016.03.004.
* [20] D. Grabner, N. Gromov, V. Kazakov and G. Korchemsky, “Strongly $\gamma$-deformed $\mathcal{N}$ Supersymmetric Yang-Mills Theory as an Integrable Conformal Field Theory”, Physical Review Letters 120, G. Korchemsky (2018), arxiv:1711.04786, http://dx.doi.org/10.1103/PhysRevLett.120.111601.
* [21] M. Hogervorst, M. Paulos and A. Vichi, “The ABC (in any D) of logarithmic CFT”, Journal of High Energy Physics 2017, A. Vichi (2017), arxiv:1605.03959, http://dx.doi.org/10.1007/JHEP10(2017)201.
* [22] V. Gurarie, “Logarithmic operators in conformal field theory”, Nuclear Physics B 410, 535 (1993), hep-th/9303160, http://dx.doi.org/10.1016/0550-3213(93)90528-W.
* [23] Y. He and H. Saleur, “A note on the identity module in $c=0$ CFTs”, arxiv:2109.05050.
* [24] J. A. Minahan and K. Zarembo, “The Bethe-ansatz for Script $\mathcal{N}$ = 4 super Yang-Mills”, Journal of High Energy Physics 2003, 013 (2003), hep-th/0212208, http://dx.doi.org/10.1088/1126-6708/2003/03/013.
* [25] E. H. Lieb, T. Schultz and D. Mattis, “Two soluble models of an antiferromagnetic chain”, Annals Phys. 16, 407 (1961).
* [26] M. Spradlin and A. Volovich, “The one-loop partition function of super-Yang-Mills theory on”, Nuclear Physics B 711, 199 (2005), hep-th/0408178, http://dx.doi.org/10.1016/j.nuclphysb.2005.01.007.
* [27] R. P. Stanley, “Enumerative Combinatorics”, 2nd edition, Cambridge University Press (2011).
* [28] M. Flohr and M. Koehn, “What the characters of irreducible subrepresentations of Jordan cells can tell us about LCFT”, J. Phys. A 46, 494007 (2013), arxiv:1307.5844.
* [29] J. Caetano, O. Gurdogan and V. Kazakov, “Chiral limit of N = 4 SYM and ABJM and integrable Feynman graphs”, arxiv:1612.05895.
* [30] J. Minahan and K. Zarembo, “The Bethe ansatz for superconformal Chern-Simons”, Journal of High Energy Physics 2008, 040 (2008), arxiv:0806.3951, http://dx.doi.org/10.1088/1126-6708/2008/09/040.
* [31] V. Kazakov, “Quantum Spectral Curve of $\gamma$-Twisted $\mathcal{N}$= 4 SYM Theory and Fishnet CFT”, Reviews in Mathematical Physics 30, 1840010 (2018), arxiv:1802.02160, http://dx.doi.org/10.1142/S0129055X1840010X.
|
-titleHadron Collider Physics symposium (HCP 2011)
11institutetext: Royal Holloway, University of London.
# A Search for Heavy Resonances in the Dilepton Channel
Daniel Hayden On behalf of the ATLAS Collaboration<EMAIL_ADDRESS>
###### Abstract
There are many extensions to the Standard Model of particle physics which
predict the addition of a U(1) symmetry, and/or extra spatial dimensions,
which give rise to new high mass resonances such as the Z′ and Randall-Sundrum
graviton. The LHC provides a unique opportunity to explore the TeV scale where
these phenomena may become apparent, and can be searched for using the
precision tracking and high energy resolution calorimetry of the ATLAS
detector. This poster presents the search for high mass resonances in the
dilepton channel, and was conducted with an integrated luminosity of 1.08/1.21
fb-1 in the dielectron/dimuon channel respectively, at a centre of mass energy
$\sqrt{s}$ = 7 TeV.
## 1 Introduction
There are many possible extensions to the Standard Model (SM) predicted at the
TeV energy scale which may be visible at the LHC. Many of these extensions
predict extra U(1) symmetry with an associated spin-1 particle Theory:ZP1
Theory:ZP2 . In its simplest form this U(1) symmetry can be arbitrarily added
to the existing SM gauge group, resulting in SU(3) $\times$ SU(2) $\times$
U(1) from the SM, and additionally U(1)′ for the Sequential Standard Model
Z${}^{\prime}_{SSM}$. More rigorously motivated models proceed via the
decomposition of Grand Unified Theories such as E6 $\rightarrow$
SO(10)$\times$U(1)ψ $\rightarrow$ SU(5) $\times$ U(1)χ $\times$ U(1)ψ leading
to Z′($\theta$) = Z${}^{\prime}_{\chi}$cos$\theta$ \+
Z${}^{\prime}_{\psi}$sin$\theta$, where the mixing angle $\theta$ determines
the coupling to fermions and results in various possible models with specific
Z′ states. Other extensions of the SM seek to answer questions such as the
hierarchy problem where the relative weakness of gravity compared to the other
forces of nature can be explained with the use of warped extra dimensions in
theories such as the Randall-Sundrum model Theory:G . A feature of this theory
would be a massive spin-2 particle called the graviton (G∗) which should be
observable at the LHC and have a mass/width that depends on the curvature of
the warped dimension, k, and the reduced Planck scale, $\overline{M}_{Pl}$,
leading to another parameter of interest, the coupling k/$\overline{M}_{Pl}$.
Both of the new particles mentioned would appear as resonances in the dilepton
invariant mass spectrum measured by the ATLAS detector ATLAS , and these
results comprise a search using the detector in this endeavor.
## 2 Dilepton Resonance Search
The search for dilepton resonances was conducted in both the electron and muon
channels separately, which were then combined to give the final result. To
identify candidate events from data, each analysis selected high energy
electron/muon pairs. The main background to a Z′/G∗ search in these channels
is from Drell-Yan, with smaller contributions from $t\bar{t}$, W+jets,
diboson, and QCD events 111QCD events here are defined as semi-leptonic decays
of b and c quarks in the dimuon sample, or at least one electron coming from
photon conversions, semi-leptonic heavy quark decays or a hadronic fake, in
the dielectron sample.. These SM background contributions were estimated using
Monte Carlo (MC) simulation, except for QCD which was estimated from data
using a reverse identification selection sample for electrons, and a non-
isolated sample for muons.
For both the dielectron and dimuon channel analyses, a data quality
requirement is made to ensure parts of the ATLAS detector important for
e/$\gamma$ or $\mu$ analysis respectively are working optimally. The events
are also required to have at least one primary vertex with greater than two
tracks, and pass a single electron trigger with a transverse energy (ET)
greater than 20 GeV or equivalently for the dimuon analysis, a muon trigger
with transverse momentum (pT) greater than 22 GeV.
For an event to be accepted by the analysis in the dielectron channel, an
event must contain at least two electron candidates with ET $>$ 25 GeV and
$|$$\eta$$|$ $<$ 2.47, also excluding the region between the barrel and endcap
calorimeters 1.37 $\leq$ $|$$\eta$$|$ $\leq$ 1.52. The electron candidates
that pass these criteria must have been reconstructed from electromagnetic
cells clusters with an associated charged particle track from the inner
detector. Shower shape variables and hadronic calorimeter leakage, along with
information from the inner detector is then used to strengthen the
identification of the electron candidates. A hit in the first layer of the
pixel detector is required to suppress background from photon conversions.
From the remaining electron candidates the highest ET pair is selected and the
higher ET electron required to pass an isolation threshold of less than 7 GeV
in a cone of 0.2 around the cluster ($\Delta$R =
$\sqrt{(\Delta\eta)^{2}+(\Delta\phi)^{2}}$) to reduce the QCD background.
Finally, the invariant mass of the selected pair must be greater than 70 GeV
to be accepted by the analysis, and no opposite charge requirement is made to
minimise the impact of possible charge mis-identification.
In the dimuon channel, two oppositely charged muons are required. Each muon
must have pT $>$ 25 GeV, and pass quality criteria from the inner detector as
well as having at least three hits in each of the inner, middle, and outer
layers of the muon spectrometer to improve momentum resolution. Muons are
discarded if they have hits in both the barrel and endcap regions because of
residual misalignment. To suppress the cosmic ray background the $z$ position
of the primary vertex is required to be less than 200 mm, and muon tracks must
have a transverse impact parameter $|$$d_{0}$$|$ $<$ 0.2 mm, also being within
1 mm of the primary vertex along the beam-line. To reduce the QCD background
in the muon channel, each muon is required to be isolated such that
$\sum{p_{T}}$($\Delta$R $<$ 0.3)/pT $<$ 0.05. The two highest $p_{T}$ muons
passing this selection form a pair and are required to have an invariant mass
greater than 70 GeV to be accepted by the dimuon analysis.
The dilepton analysis was performed with an integrated luminosity of 1.08 fb-1
in the electron channel, and 1.21 fb-1 in the muon channel. The results of
this analysis can be found in Exotics:EPS , and the main kinematic plots of
interest, namely the invariant mass spectrum for both the electron and muon
channel, are presented in Figure 1.
Figure 1: Invariant mass spectrum for the electron (top) and muon (bottom)
channel dilepton resonance search. Various possible Z${}^{\prime}_{SSM}$
signals are overlayed to show how an expected signal would manifest itself.
## 3 Statistical Analysis
Any excess in the observed data over the SM prediction can be quantified using
a Log Likelihood Ratio (LLR) test:
$LLR=-2ln\frac{{\cal L}(data|N_{sig}+N_{bkg})}{{\cal L}(data|N_{bkg})}$ (1)
In this dataset the greatest excesses give $p$-values of 54% and 24% for the
dielectron and dimuon channels respectively. Therefore as no significant
excess is observed, limits are set on the cross section times branching ratio
($\sigma$B) for the Z${}^{\prime}_{SSM}$ and G∗ decaying to leptons, at 95%
confidence level using the Bayesian Analysis Toolkit (BAT) BAT . BAT
constructs a binned likelihood, combining the electron and muon channel
searches and accounting for observed ($n$) and expected ($\mu$) events with
associated nuisance parameters ($\theta$) on a bin by bin basis:
${\cal L}(data|\sigma
B,\theta_{i})=\prod_{l=1}^{N_{channel}}\prod_{k=1}^{N_{bin}}\frac{\mu_{lk}^{n_{lk}}e^{-\mu_{lk}}}{n_{lk!}}\prod_{i=1}^{N_{sys}}G(\theta_{i},0,1)$
(2)
Employing Bayesian statistics (assuming a flat positive prior so that
$\pi$($\sigma$B) = 1) and treating the nuisance parameters as Gaussian priors,
Markov Chain Monte Carlo is used to reduce the likelihood (${\cal
L}^{\prime}$) and obtain the marginalised posterior probability, which is then
solved for ($\sigma$B)95:
$0.95=\frac{\int_{0}^{{\sigma B}_{95}}{\cal L}^{\prime}(\sigma B)\pi(\sigma
B)d(\sigma B)}{\int_{0}^{\infty}{\cal L}^{\prime}(\sigma B)\pi(\sigma
B)d(\sigma B)}$ (3)
The resulting limits on the Z′/G∗ $\sigma$B are converted into mass exclusion
limits using the theoretical dependence of $\sigma$B as a function of
resonance mass. The $\sigma$B limits are presented in Figure 2. Table 1
summarises the excluded mass values for the models considered. The results
presented here represent a large step forward in the search for heavy dilepton
resonances, exceeding previous experiments’ mass exclusion limits for Z′/G∗
resonances in the dilepton channel. With a total integrated luminosity of
$\sim$5 fb-1 recorded by the ATLAS detector in 2011, this search will soon be
updated probing even further into the TeV scale regime in search of new
physics beyond the current SM.
Figure 2: 95% confidence level $\sigma$B limits for various Z′ models (top),
and RS graviton k/Mpl couplings (bottom).
| E6 Z′ Models
---|---
Model | Z${}^{\prime}_{\psi}$ | Z${}^{\prime}_{N}$ | Z${}^{\prime}_{\eta}$ | Z${}^{\prime}_{I}$
Mass limit [TeV] | 1.49 | 1.52 | 1.54 | 1.56
| E6 Z′ Models |
---|---|---
Model | Z${}^{\prime}_{S}$ | Z${}^{\prime}_{\chi}$ | Z${}^{\prime}_{SSM}$
Mass limit [TeV] | 1.60 | 1.64 | 1.83
G∗ Coupling k/MPl | 0.01 | 0.03 | 0.05 | 0.10
---|---|---|---|---
Mass limit [TeV] | 0.71 | 1.03 | 1.33 | 1.63
Table 1: 95% confidence level lower mass exclusion limits for various Z′
models and RS graviton k/Mpl couplings, decaying to two leptons (dielectron or
dimuon).
## References
* (1) D. London and J. L. Rosner, Extra Gauge Bosons in E(6), Phys. Rev D34 (1986) 1530.
* (2) P. Langacker, The Physics of Heavy Z′ Gauge Bosons, Rev. Mod. Phys 81 (2009) 1199.
* (3) L. Randall and R. Sundrum, A large mass hierarchy from a small extra dimension, Phys. Rev. Lett 83 (1999) 3370.
* (4) ATLAS Collaboration, The ATLAS Experiment at the CERN Large Hadron Collider, JINST 3 S08003 (2008) .
* (5) The ATLAS Collaboration, Search for high-mass dilepton resonances in pp collisions at $\sqrt{s}$ = 7 TeV with the ATLAS detector, Phys. Rev. Lett 107, 272002 (2011) .
* (6) A Caldwell, D. Kollar, K. Kröninger, BAT - The Bayesian Analysis Toolkit, Computer Physics Communication 180 (2009) 2197.
|
# Classical results for alternating virtual links
Hans U. Boden Mathematics & Statistics, McMaster University, Hamilton,
Ontario<EMAIL_ADDRESS>and Homayun Karimi Mathematics & Statistics,
McMaster University, Hamilton, Ontario<EMAIL_ADDRESS>
###### Abstract.
We extend some classical results of Bankwitz, Crowell, and Murasugi to the
setting of virtual links. For instance, we show that an alternating virtual
link is split if and only if it is visibly split, and that the Alexander
polynomial of any almost classical alternating virtual link is alternating.
The first result is a consequence of an inequality relating the link
determinant and crossing number for any non-split alternating virtual link.
The second is a consequence of the matrix-tree theorem of Bott and Mayberry.
We extend the first result to semi-alternating virtual links. We discuss the
Tait conjectures for virtual and welded links and note that Tait’s second
conjecture is not true for alternating welded links.
###### Key words and phrases:
Alternating link, virtual link, split link, checkerboard coloring,
determinant, almost classical link, Alexander polynomial, welded link,
branched double cover, Tait conjectures
###### 2020 Mathematics Subject Classification:
Primary: 57K12
## §1. Introduction
In this paper, we establish conditions that are satisfied by invariants of
alternating virtual links, such as the link determinant and Alexander
polynomial. As an application, we deduce that a reduced alternating virtual
link diagram is split if and only if it is visibly split.
A link is said to be alternating if it admits an alternating diagram, and a
diagram is alternating if the crossings alternate between over and under-
crossing as one travels around any component. This applies to classical and
virtual links, with the proviso that virtual crossings are ignored.
In [Ban30], Bankwitz proved that $\det(L)\geq c(L)$ for any non-split
alternating link $L$, where $\det(L)$ denotes the link determinant and $c(L)$
the crossing number of $L$. In [Cro59a, Mur58], Crowell and Murasugi
independently proved that the Alexander polynomial of an alternating link is
alternating. Here, a Laurent polynomial $\Delta_{L}(t)=\sum c_{i}t^{i}$ is
said to be alternating if its coefficients satisfy $(-1)^{i+j}c_{i}c_{j}\geq
0$.
We extend the results of Bankwitz, Crowell, and Murasugi to alternating
virtual links. Virtual knots were introduced by Kauffman in [Kau99], and they
represent a natural generalization to knots in thickened surfaces up to
stabilization. Classical knots embed faithfully into virtual knot theory
[GPV00], and many invariants from classical knot theory extend in a natural
way to the virtual setting.
For example, the link determinant $\det(L)$ is defined in terms of the
coloring matrix and extends to checkerboard colorable virtual links (defined
below). The link $L$ admits a $p$-coloring if and only if $p$ divides
$\det(L)$. One of our main results is that $\det(L)\geq c(L)$ for any non-
split alternating virtual link $L,$ where $c(L)$ is the classical crossing
number of $L$. This result applies to show that a reduced alternating virtual
link diagram is split if and only if it is visibly split.
The Alexander polynomial $\Delta_{L}(t)$ is defined in terms of the Alexander
module of $L$, and it extends to almost classical virtual links (defined
below). Another one of our main results is that, for any reduced alternating
link $L$ that is almost classical, its Alexander polynomial $\Delta_{L}(t)$ is
alternating. To prove this result, we appeal to the Matrix-Tree Theorem. It
applies to show that many virtual knots cannot be represented by alternating
diagrams.
The link determinant and Alexander polynomial are both invariant under welded
equivalence. Therefore, our main results can be seen as providing restrictions
on a virtual link diagram for it to be welded equivalent to an alternating
virtual link. This is discussed at the end of the paper, where we state open
problems related to the Tait conjectures for welded links.
We provide a brief synopsis of the contents of the rest of this paper. In §2,
we review background material on links in thickened surfaces and virtual and
welded links, together with Cheng coloring and Alexander numbering for virtual
links. In §3, we review the link group and determinant. In §4, we recall the
Matrix-Tree Theorem, which is used to prove one of the main results. In §5, we
prove that split alternating virtual links are visibly split. In §6, we prove
analogous results for semi-alternating links, and in §7, we present a
discussion on the Tait conjectures for welded links and state some interesting
open problems.
## §2. Virtual links
In this section we review the basic properties of virtual links, including
Gauss diagrams, links in thickened surfaces, welded links, ribbon torus links,
alternating virtual links, virtual linking numbers, Cheng colorings, and
Alexander numberings.
Virtual link diagrams. Virtual links are defined as equivalence classes of
virtual link diagrams. Here, a virtual link diagram is an immersion of one or
more circles in the plane with only finitely many regular singularities, each
of which is a double point. Each double point is either classical (indicated
by over- and under-crossings) or virtual (indicated by a circle). Two diagrams
are said to be virtually equivalent if they can be related by planar isotopies
and a series of generalized Reidemeister moves ($r1$)–($r3$) and ($v1$)–($v4$)
depicted in Figure 1.
An orientation for a virtual link is obtained by choosing an orientation for
each component. For a diagram $D$, the orientation is usually indicated by
placing one arrow on each component of $D$.
$r1$$r2$$r3$$v1$$v2$$v3$$v4$$f1$ Figure 1. The generalized Reidemeister moves
($r1$)–($r3$), ($v1$)–($v4$) and the forbidden move ($f1$).
Gauss diagrams. Virtual links can also be defined as equivalence classes of
Gauss diagrams, which consist of one or more circles traversed
counterclockwise, together with directed chords on the circles representing
the classical crossings. The chords point from over-crossings to under-
crossings and are decorated with a sign ($+$ or $-$) to indicate whether the
crossing is positive or negative. Each virtual link diagram determines a Gauss
diagram, and vice versa, and this correspondence is well-defined up to moves
($v1$)–($v4$). The Reidemeister moves can be translated into moves between
Gauss diagrams, and in this way a virtual link can be regarded as an
equivalence class of Gauss diagrams. By convention, the core circles of a
Gauss diagram are oriented counterclockwise.
Notice that the Gauss diagram does not keep track of the virtual crossings. In
effect, the virtual crossings are not really there, rather they are an
inevitable consequence of trying to represent a non-planar virtual link
diagram by a diagram in the plane.
A virtual link diagram is said to be split if its associated Gauss diagram is
disconnected, and a virtual link is split if it can be represented by a split
diagram. For classical links, this agrees with the usual definition. For
virtual links, a diagram can be split and connected. However, any diagram that
is split can be transformed into a disconnected diagram using moves
($v1$)–($v4$).
Links in thickened surfaces. A third approach is to define virtual links as
stable equivalence classes of links in thickened surfaces, and we take a
moment to explain this.
Let $\Sigma$ be a closed, oriented surface and $I=[0,1]$. Consider a link
$\mathcal{L}\subset\Sigma\times I$ in the thickened surface, up to isotopy.
Let $p\colon\Sigma\times I\to\Sigma$ be the projection map.
Stabilization is the operation of adding a 1-handle to $\Sigma$, disjoint from
$p(\mathcal{L})$, and destabilization is the opposite procedure. Two links
$\mathcal{L}\subset\Sigma\times I$ and
$\mathcal{L}^{\prime}\subset\Sigma^{\prime}\times I$ are said to be stably
equivalent if one is obtained from the other by a finite sequence of
stabilizations, destablizations, and orientation-preserving diffeomorphisms of
the pairs $(\Sigma\times I,\Sigma\times\\{0\\})$ and $(\Sigma^{\prime}\times
I,\Sigma^{\prime}\times\\{0\\})$. In [CKS02], Carter, Kamada, and Saito show
there is a one-to-one correspondence between virtual links and stable
equivalence classes of links in thickened surfaces.
Thus, every virtual link can be represented as a link in a thickened surface.
Further, any such link itself can be represented as a link diagram on
$\Sigma.$ A link diagram $\mathcal{D}$ on $\Sigma$ is a tetravalent graph with
over= and under-crossing information drawn at each vertex in the usual way.
A link diagram $\mathcal{D}$ on $\Sigma$ is said to be a split diagram if it
is disconnected, and a link in $\Sigma\times I$ is said to be split if it can
be represented by a split diagram.
Welded links. Two virtual links are said to be welded equivalent if one can be
obtained from the other by a sequence of generalized Reidemeister moves and
the first forbidden move ($f1$) as depicted in Figure 1. In terms of Gauss
diagrams, the first forbidden move corresponds to exchanging two adjacent
arrow feet without changing their signs or arrowheads, see Figure 2.
Therefore, a welded link can also be viewed as an equivalence class of Gauss
diagrams.
Figure 2. The forbidden overpass $(f1)$ for Gauss diagrams.
Ribbon torus links. Every welded link determines a ribbon knotted surface in
$S^{4}$. This is based on a beautiful construction by Satoh [Sat00], which
associates to a welded link $L$ a ribbon torus link $\operatorname{Tube}(L)$
in $S^{4}$. In [Sat00], Satoh shows that every ribbon torus link occurs as
$\operatorname{Tube}(L)$ for some welded link, and that
$\pi_{1}(S^{4}\smallsetminus\operatorname{Tube}(L))$ is isomorphic to the link
group $G_{L}$, defined below.
The correspondence between welded links and ribbon torus links is not one-to-
one, see [Win09]. It is an open problem to determine necessary and sufficient
conditions for two welded knots to represent the same ribbon torus knot (cf.
[Aud16, Question 3.6]).
Alternating virtual links. A virtual link diagram $D$ is called alternating if
the classical crossings alternate between over-crossing and under-crossing as
we go around each component. A Gauss diagram is alternating if it alternates
between arrow heads and tails when going around each of the core circles. A
virtual or welded link is called alternating if it can be represented by an
alternating virtual link diagram. For example, consider the virtual links in
Figure 3. The virtual Borromean rings is alternating, but the virtual Hopf
link is not.
Virtual linking numbers. If $J$ and $K$ are oriented virtual knots, then the
virtual linking number $\operatorname{{\it v}\ell{\it k}}(J,K)$ is defined as
the sum of the writhe of the classical crossings where $J$ goes over $K$.
Using the same definition, we can define $\operatorname{{\it v}\ell{\it
k}}(J,K)$ more generally when $J$ and $K$ are oriented virtual links. Note
that $\operatorname{{\it v}\ell{\it k}}(J,K)$ is additive, namely if
$J=J^{\prime}\cup J^{\prime\prime}$, then $\operatorname{{\it v}\ell{\it
k}}(J,K)=\operatorname{{\it v}\ell{\it k}}(J^{\prime},K)+\operatorname{{\it
v}\ell{\it k}}(J^{\prime\prime},K)$, and likewise if $K=K^{\prime}\cup
K^{\prime\prime}.$ The virtual linking numbers are not symmetric, i.e., it is
not generally true that $\operatorname{{\it v}\ell{\it
k}}(J,K)=\operatorname{{\it v}\ell{\it k}}(K,J)$. For example, consider the
oriented virtual links in Figure 3. For the virtual Hopf link, we have
$\operatorname{{\it v}\ell{\it k}}(J,K)=1$ and $\operatorname{{\it v}\ell{\it
k}}(K,J)=0$, and for the virtual Borromean rings, we have
$\operatorname{{\it v}\ell{\it k}}(I,J)=\operatorname{{\it v}\ell{\it
k}}(J,K)=\operatorname{{\it v}\ell{\it k}}(K,I)=0,\;\operatorname{{\it
v}\ell{\it k}}(J,I)=\operatorname{{\it v}\ell{\it k}}(K,J)=1,\text{ and
}\operatorname{{\it v}\ell{\it k}}(I,K)=-1.$
Figure 3. The virtual Hopf link and the virtual Borromean rings.
Cheng colorings. Given a virtual link diagram $D$, a Cheng coloring of $D$ is
an assignment of integer labels to each arc of $D$ that satisfies the local
rules in Figure 4. An elementary exercise shows if $D$ and $D^{\prime}$ are
two virtual link diagrams that are related by virtual Reidemeister moves, then
$D$ admits a Cheng coloring if and only if $D^{\prime}$ does. A virtual link
$L$ is said to be Cheng colorable if it can be represented by a virtual link
diagram with a Cheng coloring.
Not all virtual links are Cheng colorable. For example, the virtual Hopf link
in Figure 3 is not Cheng colorable. More generally, given a virtual link
$L=K_{1}\cup\cdots\cup K_{m}$ with $m$ components, then an elementary argument
shows that $L$ admits a Cheng coloring if and only if it satisfies
$\operatorname{{\it v}\ell{\it k}}(K_{i},L\smallsetminus
K_{i})=\operatorname{{\it v}\ell{\it k}}(L\smallsetminus K_{i},K_{i})=0$
for each $i=1,\ldots,m.$
Figure 4. Local rules for a Cheng coloring at classical and virtual crossings.
Alexander numberings and almost classical links. Given a virtual link diagram
$D$, an Alexander numbering on $D$ is an assignment of integer labels to each
arc of $D$ that satisfies the local rules in Figure 5. If $D$ admits an
integer labeling that satisfies the local rules mod $p$, then $D$ is said to
be mod $p$ Alexander numberable.
Notice that if $D$ is Alexander numberable, then it is Cheng colorable.
Conversely, a virtual link diagram $D$ is Alexander numberable if and only if
it admits a Cheng coloring such that the arc labels satisfy $b=a-1$ at each
classical crossing. Likewise, a virtual link diagram $D$ is mod $p$ Alexander
numberable if and only if it admits a Cheng coloring such that the arc labels
satisfy $b\equiv a-1$ (mod $p$) at each classical crossing.
Figure 5. Local rules for an Alexander numbering at classical and virtual
crossings.
A virtual link is said to be almost classical if it admits an Alexander
numberable diagram, and it is said to be checkerboard colorable if it admits a
mod 2 Alexander numberable diagram.
Recall from [BGH+17, Theorem 6.1] that a virtual link is almost classical if
and only if it can be represented by a null-homologous link
$\mathcal{L}\subset\Sigma\times I$, or equivalently if $\mathcal{L}$ admits a
Seifert surface.
Recall also from [BCK21, Proposition 1.1] that a virtual link is checkerboard
colorable if and only if it can be represented by a ${\mathbb{Z}}/2$ null-
homologous link $\mathcal{L}\subset\Sigma\times I$, or equivalently if
$\mathcal{L}$ admits an unoriented spanning surface. This is the case if and
only if $\mathcal{L}$ can be represented by a checkerboard colorable diagram
on $\Sigma$.
If a virtual link admits a diagram which is Cheng colorable, then any diagram
for the same link is also Cheng colorable. The reason is that Cheng colorings
extend along generalized Reidemeister moves. The same thing is not true for
Alexander numberings. Indeed, one can easily find two virtual link diagrams
for the same link such that one of them is Alexander numberable and the other
is not. Thus, Alexander numberings of virtual links do not always extend along
generalized Reidemeister moves.
However, if two virtual knot diagrams are Alexander numberable and are related
by generalized Reidemeister moves, then one can arrange that they are related
through Alexander numberable diagrams. More precisely, suppose $D$ and
$D^{\prime}$ are two virtual knot diagrams and
(1) $D=D_{1}\sim D_{2}\sim\cdots\sim D_{r}=D^{\prime}\ $
is a chain of diagrams, where $D_{i+1}$ is obtained from $D_{i}$ by a single
generalized Reidemeister move. If $D$ and $D^{\prime}$ are Alexander
numberable, then there is a chain (1) such that each $D_{i}$ is Alexander
numberable. A similar result holds if $D$ and $D^{\prime}$ are assumed to be
mod $p$ Alexander numberable. These statements can be proved using parity
projection, see [Man10, Nik13]. Any minimal crossing diagram of an almost
classical link is Alexander numberable, and this can also be proved using
parity projection, see [Rus21, BR21].
The corresponding statements for welded knots and links are either not true,
or not known to be true.
## §3. Link group, Alexander module, and determinant
In this section, we introduce the link group and the Alexander module
associated to a virtual link. We also recall the definition of the link
determinant associated to a checkerboard colorable virtual link and show that
$\det(L)=0$ when $L$ is split. Finally, we discuss mod $p$ labelings of
virtual knots and show that $K$ admits a mod $p$ labeling if and only if
$\det(K)=0$ (mod $p$).
Link Group. For classical links, the link group is just the fundamental group
of the complement of the link. For a link $L$, this group is denoted $G_{L}$.
Thus, $G_{L}=\pi_{1}(X_{L})$ where $X_{L}$ is the result of removing an open
tubular neighborhood of $L$ from $S^{3}$.
As an invariant of classical knots, the knot group is an unknot detector,
indeed the only classical knot $K$ whose knot group is infinite cyclic is the
trivial knot. In fact, Waldhausen’s theorem implies that the knot group
together with its peripheral structure is a complete invariant of classical
knots, which is to say that two classical knots are equivalent if and only if
they have isomorphic knot groups with equivalent peripheral structures.
The link group generalizes in a natural way to give an invariant of virtual
links by means of Wirtinger presentations. In fact, the abstract group
together with its peripheral structure are invariant under the first forbidden
move and thus define invariants of the underlying welded link.
Given a virtual link diagram for $L$, we will describe the Wirtinger
presentation of $G_{L}$. Let $D$ be a regular projection of $L$, and suppose
it has $n$ classical crossings. We define a long arc of the diagram $D$ to be
one that goes from one classical under-crossing to the next, passing through
virtual crossings as it goes. Enumerate the long arcs of $D$ by
$x_{1},\ldots,x_{m}$ and the classical crossings by $c_{1},\ldots,c_{n}$.
Figure 6. Arc labels at a crossing.
For each crossing, labeled as in Figure 6, we have the relation
$r_{i}=x_{\ell}x_{j}^{-1}x_{k}^{-1}x_{j}$. The Wirtinger presentation for
$G_{L}$ is then:
(2) $G_{L}=\langle x_{1},\ldots,x_{m}\mid r_{1},\ldots,r_{n}\rangle.$
Alexander module. In order to define the Alexander module, we briefly recall
Fox differentiation. Let $F_{m}$ be the free group on $m$ generators, so
elements of $F_{m}$ are words in $x_{1},\ldots,x_{m}$. For $j=1,\ldots,m$, the
Fox derivative $\partial/\partial x_{j}$ is an endomorphism of
${\mathbb{Z}}[F_{m}]$, the group ring, defined so that $\partial/\partial
x_{j}(1)=0$ and
$\frac{\partial}{\partial x_{j}}(x_{i})=\begin{cases}1&\text{if $i=j$,}\\\
0&\text{otherwise.}\end{cases}$
Further, given words $w,z\in F_{m}$, the Fox derivative satisfies the Leibnitz
rule:
$\frac{\partial}{\partial x_{j}}(wz)=\frac{\partial}{\partial
x_{j}}(w)+w\frac{\partial}{\partial x_{j}}(z).$
These relations completely determine $\partial/\partial x_{j}$ on every word
$w\in F_{m},$ and it is extended linearly to the group ring
${\mathbb{Z}}[F_{m}]$.
We use this to describe the construction of the Alexander module associated to
a link $L$. Let $G_{L}^{\prime}=[G_{L},G_{L}]$ and
$G_{L}^{\prime\prime}=[G_{L}^{\prime},G_{L}^{\prime}]$ be the first and second
commutator subgroups, then the Alexander module is the quotient
$G_{L}^{\prime}/G_{L}^{\prime\prime}$. It is a finitely generated module over
${\mathbb{Z}}[t,t^{-1}]$, the ring of Laurent polynomials, and it is
determined by the Fox Jacobian matrix $A$ as follows. Here, $A$ is the
$n\times m$ matrix with $ij$ entry equal to $\left.\frac{\partial
r_{i}}{\partial x_{j}}\right|_{x_{1},\ldots,x_{m}=t}$. In particular, the Fox
Jacobian is obtained by Fox differentiating the relations $r_{i}$ with respect
to the generators $x_{j}$ and applying the abelianization map $x_{j}\mapsto t$
for $j=1,\ldots,m$. We define the $k$-th elementary ideal $\mathscr{E}_{k}$ as
the ideal of ${\mathbb{Z}}[t,t^{-1}]$ generated by all $(n-k)\times(n-k)$
minors of $A$.
The matrix $A$ depends on the choice of a presentation for $G_{L}$, but the
associated sequence of elementary ideals
$\\{0\\}=\mathscr{E}_{0}\subset\mathscr{E}_{1}\subset\ldots\subset\mathscr{E}_{n}={\mathbb{Z}}[t,t^{-1}]$
does not.
For any classical link $L$, the first elementary ideal $\mathscr{E}_{1}$ is a
principal ideal, and the Alexander polynomial $\Delta_{L}(t)$ is defined as
the generator of $\mathscr{E}_{1}$. The Alexander polynomial is well-defined
up to multiplication by $\pm t^{k}$ for $k\in{\mathbb{Z}}$. It is obtained by
taking the determinant of the Alexander matrix, which is the
$(n-1)\times(n-1)$ matrix obtained by removing a row and column from $A$.
For a virtual link $L$, one can mimic the construction of the Alexander module
by regarding the quotient $G_{L}^{\prime}/G_{L}^{\prime\prime}$ as a module
over ${\mathbb{Z}}[t,t^{-1}],$ This can be used to define elementary ideals
and the Alexander polynomial for virtual links. However, in contrast to the
case of classical links, the first elementary ideal may not be principal. One
way to remedy the situation is to replace the elementary ideals
$\mathscr{E}_{k}$ with the smallest principal ideal containing them. For
instance, this would suggest a way to define an Alexander polynomial for a
virtual link $L$ to be a generator of the principal ideal containing
$\mathscr{E}_{1}$. However, since the link group itself is only an invariant
of the associated welded link, the invariants one obtains in this way will not
be very refined. Indeed, for many virtual knots, the Alexander polynomial is
trivial.
Link determinant for checkerboard colorable virtual links. We review the
definition of the link determinant in terms of the coloring matrix and show
that it extends to an invariant of checkerboard colorable virtual links. We
also prove that the determinant of the coloring matrix is odd for any
checkerboard colorable virtual knot, and that a checkerboard colorable virtual
knot $K$ admits a mod $p$ labeling if and only if $p$ divides $\det(K)$.
Let $L$ be a virtual link that is represented by a checkerboard colorable
diagram $D$ with $n$ classical crossings $\\{c_{1},\ldots,c_{n}\\}$ and $m$
long arcs $\\{a_{1},\ldots,a_{m}\\}.$ If $D$ has $k$ connected components,
then $m=n+k-1$.
Define the $n\times m$ coloring matrix $B(D)$ so that its $ij$ entry is given
by
$\displaystyle b_{ij}(D)$ $\displaystyle=$
$\displaystyle\begin{cases}2,&\text{if $a_{j}$ is the over-crossing arc at
$c_{i}$},\\\ -1,&\text{if $a_{j}$ is one of the under-crossing arcs at
$c_{i}$},\\\ 0,&\text{otherwise}.\end{cases}$
In case $a_{j}$ is coincidentally the over-crossing arc and one of the under-
crossing arcs at $c_{i}$, then we set $b_{ij}(D)=1$. In that case, if $a_{k}$
is the other under-crossing arc at $c_{i}$, then we set $b_{ik}(D)=-1$.
Note that the matrix $B(D)$ is the one obtained by specializing the Fox
Jacobian matrix $A(D)$ at $t=-1$.111This is only true up to sign for any given
row. Here, $A(D)$ is defined in terms of taking Fox derivatives of the
Wirtinger presentation of the link group $G_{D}$ whose generators are given by
the arcs $a_{1},\ldots,a_{m}$ and relations are given by classical crossings
$c_{1},\ldots,c_{n}$ and applying the abelianization homomorphism
$G_{L}\to\left\langle t\right\rangle,\ a_{i}\mapsto t$. For details, see
[BGH+17, Section 5].
Notice that the entries in each row of $B(D)$ sum to zero, therefore, it has
rank at most $n-1$. The proof of the next result is similar to that of [BNW18,
Proposition 2.6].
###### Proposition 3.1.
Any two $(n-1)\times(n-1)$ minors of $B(D)$ are equal up to sign. The absolute
value of the minor is independent of the choice of checkerboard colorable
diagram $D$. It defines an invariant of checkerboard colorable links $L$
called the determinant of $L$ and denoted $\det(L)$.
###### Proof.
As previously noted, the columns of $B(D)$ always sum to zero, and we will use
checkerboard colorability to derive a linear relation among the rows. Recall
that the diagram $D$ is checkerboard colorable if and only if it admits a mod
2 Alexander numbering. For each crossing $c_{i}$ of $D$, let
$\gamma_{i}=(-1)^{\lambda_{i}},$ where $\lambda_{i}\in\\{0,1\\}$ is the
Alexander number on the incoming under-crossing at $c_{i}$. Then we claim that
one obtains a linear relation on the rows by multiplying the $i$-th row of
$B(D)$ by $\gamma_{i}$.
To see why this is true, notice that the columns of $B(D)$ correspond to arcs
of the diagram, and in any given column, there are nonzero entries for each
crossing the arc is involved in. The arc starts and ends with under-crossings,
and the associated column entries are both $-1$. Every time the arc crosses
over another arc, there is an associated column entry equal to 2. Since the
diagram is mod 2 Alexander numberable, the numbers on the transverse arcs
alternate between $0$ and $1$ as one travels along the arc. Consequently, the
coefficients $\gamma_{i}$ alternate in sign as one travels along the arc.
Therefore, after multiplying the $i$-th row by $\gamma_{i}$, this shows that
the entries in each column sum to zero. Furthermore, since each coefficient
$\gamma_{i}$ is a unit, every row of $B(D)$ is a linear combination of the
other rows. This shows that the $(n-1)\times(n-1)$ minors of $B(D)$ are all
equal up to sign. ∎
###### Proposition 3.2.
Suppose $L$ is a checkerboard colorable virtual link. If $L$ is split, then
$\det(L)=0$.
###### Proof.
Suppose $D=D_{1}\cup D_{2}$ is a split checkerboard colorable diagram for $L$.
In each row of the coloring matrix, the nonzero elements are either $2,-1,-1$
or $1,-1$. It follows the rows add up to zero. We consider a simple closed
curve in the plane which separates $D$ into two parts. It follows that the
coloring matrix $B=B(D)$ admits a $2\times 2$ block decomposition of the form
$B=\begin{bmatrix}B_{1}&0\\\ 0&B_{2}\end{bmatrix},$
where $B_{1}$ and $B_{2}$ are the coloring matrices for $D_{1}$ and $D_{2},$
respectively. Since $\det(B_{1})=0=\det(B_{2})$, it follows that the matrix
obtained by removing a row and column from $B$ also has determinant zero. ∎
Next, we define a mod $p$ labeling for a virtual knot diagram.
###### Definition 3.3.
Let $p$ be a prime number. A link diagram can be labeled mod $p$ if each long
arc can be labeled with an integer from $0$ to $p-1$ such that
* (i)
at each crossing the relation $2x-y-z=0\;\text{(mod $p$)}$ holds, where $x$ is
the label on the over-crossing and $y$ and $z$ the other two labels, and
* (ii)
at least two labels are distinct.
If a diagram has a mod $p$ labeling, then multiplying each label by a number
$m$ gives a mod $pm$ labeling, so we assume $p$ is always a prime number.
###### Remark 3.4.
If $p=2$, then the equation $2x-y-z=0\;\text{(mod $p$)}$ indicates at each
crossing the two under-crossings have the same label, hence all the labels are
equal. Therefore, a mod $2$ label for a knot diagram does not exist.
Given a knot diagram, label each long arc with a variable $x_{i}$. At each
crossing we define a relation $2x_{i}-x_{j}-x_{k}=0\;\text{(mod $p$)}$, if the
arc $x_{i}$ crosses over the arcs $x_{j}$ and $x_{k}$. Therefore, a knot can
be labeled mod $p$, if this system has a solution mod $p$ such that not all
$x_{i}$’s are equal to each other.
Fix a variable $x_{j}$. Since $x_{i}=1$ for all $i$, is a solution and adding
two solutions together forms a new solution, if there was a solution such that
not all $x_{i}$’s are equal, then there is a solution with $x_{j}=0$.
Conversely, a nontrivial solution with $x_{j}=0$ results in a labeling of the
knot. So we can delete the $j$-th column and look for the nontrivial solutions
of the resulting system.
Since we assume the knot diagram is checkerboard colorable, and we know for
such a diagram, a linear combination of rows of $B(D)$ is zero, so we can
delete the $j$-th row as well. The result is a square matrix, and a nontrivial
solution means the determinant should be zero mod $p$. The absolute value of
this determinant is $\det(K)$. So, we have the following.
###### Proposition 3.5.
We can mod $p$ label the knot $K$ if and only if $\det(K)=0\;\text{(mod
$p$)}$.
###### Corollary 3.6.
For a checkerboard colorable knot $K$, $\det(K)$ is an odd integer.
###### Proof.
Combining the Proposition 3.5 and Remark 3.4, the result follows. ∎
It would be interesting to compare the link determinant defined here with the
link determinants defined for checkerboard colorable virtual links in terms of
Goeritz matrices [ILL10].
## §4. The Matrix-Tree theorem and an application
In this section, we recall the matrix-tree theorem from [BM54] (cf. [BZH14,
Theorem 13.22]). Using it, we adapt Crowell’s proof [Cro59a] to show that the
Alexander polynomial of any almost classical alternating link is alternating.
Here, we say a Laurent polynomial is alternating if its coefficients alternate
in sign. Specifically, a polynomial $f(t)=\sum
c_{i}t^{i}\in{\mathbb{Z}}[t,t^{-1}]$ is alternating if its coefficients
satisfy $(-1)^{i+j}c_{i}c_{j}\geq 0.$
The spectacular results concerning the Jones polynomial of classical
alternating links are generally not true in the virtual case. For instance,
the span of the Jones polynomial is not equal to the crossing number. For
example, the knot $K=6.90101$ is alternating and has Jones polynomial
$V_{K}(t)=1$.
In [Thi87], Thistlethwaite proved that the Jones polynomial $V_{L}(t)$ of any
non-split, alternating classical link $L$ is alternating. This result does not
extend to virtual links. For example, the virtual knot $K=5.2426$ in Figure 7
is alternating and has Jones polynomial $V_{K}(t)=1/t^{2}+1/t^{3}-1/t^{5}.$
Since $V_{K}(t)$ is not alternating, Thistlethwaite’s result is not true for
virtual links.
Figure 7. A virtual knot diagram for $5.2426$.
Let $L$ be a virtual link. We define the link group $G_{L}$ as in §2. We use
Fox derivatives to define the Jacobian matrix $A$. For virtual knots, the
first elementary ideal $\mathscr{E}_{1}$ is not necessarily principal. We
define the Alexander polynomial $\Delta_{K}(t)$ to be the generator of the
smallest principal ideal containing $\mathscr{E}_{1}$. Since
${\mathbb{Z}}[t,t^{-1}]$ is a gcd domain, it is given by taking the gcd of all
the $(n-1)\times(n-1)$ minors of $A$ . If we remove the $i$-th row and $j$-th
column of $A$ we denote the corresponding minor by $A_{ij}.$
In [NNST12] and [BNW18], the authors showed for almost classical links,
$\mathscr{E}_{1}$ is principal, and the Alexander polynomial $\Delta_{L}(t)$
is given by taking the determinant of the $(n-1)\times(n-1)$ matrix obtained
by removing any row and any column from $A$.
###### Proposition 4.1.
For an almost classical link $L$, the determinant $\det(L)$ is equal to
$|\Delta_{L}(-1)|$.
###### Proof.
If $D$ is a diagram for $L$, the coloring matrix $B(D)$ is exactly the matrix
obtained from the Fox Jacobian matrix by replacing $t$ with $-1$. 222These
matrices are equal up to multiplication by $\pm 1$ in the rows. ∎
###### Remark 4.2.
Since any almost classical knot $K$ is checkerboard colorable, Corollary 3.6
shows that $\Delta_{K}(-1)$ is an odd number (see [BGH+17]).
Next, we state the Matrix-Tree theorem, as proved by Bott and Mayberry in
[BM54]. Tutte had given an earlier proof in [Tut48]. The result goes back to
even earlier work of Kirchhoff, to whom this theorem is usually
attributed.333See [Cro59b] for references to the early papers on the Matrix-
Tree theorem.
Let $\Gamma$ be a finite oriented graph with vertices $\\{c_{i}\mid 1\leq
i\leq n\\}$ and oriented edges $\\{u_{ij}^{\delta}\\}$, such that $c_{i}$ is
the initial point and $c_{j}$ the terminal point of $u_{ij}^{\delta}$. Notice
that $\delta$ enumerates the different edges from $c_{i}$ to $c_{j}$. By a
rooted tree (with root $c_{i}$) we mean a subgraph of $n-1$ edges such that
every point $c_{k}$ is terminal point of a path with initial point $c_{i}$.
Let $a_{ij}$ denote the number of edges with initial point $c_{i}$ and
terminal point $c_{j}$.
###### Theorem 4.3 (Matrix-Tree Theorem).
Let $\Gamma$ be a finite oriented graph without loops ($a_{ii}=0$). The
principal minor $H_{ii}$ of the graph matrix
$H(\Gamma)=\left[\begin{matrix}(\sum_{k\neq
1}a_{k1})&-a_{12}&-a_{13}&\cdots&-a_{1n}\\\ -a_{21}&(\sum_{k\neq
2}a_{k2})&-a_{23}&\cdots&-a_{2n}\\\ \vdots&\vdots&\vdots&&\vdots\\\
-a_{n1}&-a_{n2}&-a_{n3}&\cdots&(\sum_{k\neq n}a_{kn})\end{matrix}\right],$
is equal to the number of rooted trees with root $c_{i}$.
###### Corollary 4.4.
Let $\Gamma$ be a finite oriented loopless graph with a valuation
$f\colon\\{u_{ij}^{\delta}\\}\rightarrow\\{-1,1\\}$ on edges. Then the
principal minor $H_{ii}$ of the matrix $H=[b_{ij}]$, where
$b_{ij}=\begin{cases}\sum_{\delta}f(u_{ij}^{\delta}),\ \ \ i\neq j,\\\
-\sum_{k\neq i}b_{ki},\ \ i=j,\end{cases}$
satisfies the following equation:
$H_{ii}=\sum f(\operatorname{Tr}(i)),$
where the sum is to be taken over all $c_{i}$-rooted trees
$\operatorname{Tr}(i)$, and where
$f(\operatorname{Tr}(i))=\prod_{u_{kj}^{\delta}\in\operatorname{Tr}(i)}f(u_{kj}^{\delta}).$
For a virtual link diagram, there are (at least) two ways one can associate a
$4$-valent graph. One way is to consider the diagram $D$ itself. It has
vertices for the classical and virtual crossings and edges running from one
classical or virtual crossing to the next. This graph is planar. The other way
to associate a graph is to consider vertices only for classical crossings. The
key difference is that in general, this graph is not planar. For an
alternating diagram $D$, we describe this graph and an orientation on it as
follows:
Let $D$ have classical crossings $c_{1},\ldots,c_{n}$. The vertices of
$\Gamma$ are $c_{1},\ldots,c_{n}$. At each vertex consider two out-going edges
corresponding to the over-crossing arc, and two in-coming edges for the under-
crossing arcs (see Figure 8). This is called the source-sink orientation or
the alternate orientation. This orientation is possible because $D$ is
alternating, and an out-going edge at the vertex $c_{i}$, should be an in-
coming edge for the adjacent vertex.
###### Remark 4.5.
In general, any checkerboard colorable diagram $D$ admits a source-sink
orientation. In fact, a diagram is checkerboard colorable if and only if it
admits a source-sink orientation (see [KNS02, Proposition 6]).
Figure 8. The source-sink orientation.
###### Theorem 4.6.
If $L$ is a non-split, almost classical alternating link, then its Alexander
polynomial $\Delta_{L}(t)$ is alternating.
###### Proof.
For the unknot the result is obvious. Assume $D$ has $n\geq 1$ classical
crossings. Orient $D$ and enumerate the crossings by $c_{1},\ldots,c_{n}$.
Label the long arcs by $g_{1},\ldots,g_{n}$. At the crossing $c_{i}$, label
the over-crossing arc $g_{\nu(i)}$ and the under-crossing arcs
$g_{\lambda(i)}$ and $g_{\rho(i)}$ as in Figure 9. Define the relation
$r_{i}=g_{\lambda(i)}g_{\nu(i)}^{-1}g_{\rho(i)}^{-1}g_{\nu(i)}$.
Figure 9. Arc labels at the crossing $c_{i}$.
Now consider the graph $\Gamma$ associated with $D$, with the source-sink
orientation on it. Label the edges by $u_{ij}^{\delta}$. Define the valuation
$f$ as follows. At the crossing $c_{j}$, if $u_{ij}^{\delta}$ corresponds to
$g_{\lambda(j)}$, then $f(u_{ij}^{\delta})=1$, and if it corresponds to
$g_{\rho(j)}$, then $f(u_{ij}^{\delta})=-t$.
Define the matrix $H$ as in the Corollary 4.4. Notice that $D$ is alternating
and there is a one-to-one correspondence between the classical crossings of
$D$ and the set of over-crossing arcs. Therefore, we can choose to label over-
crossing arcs, such that $\nu(i)=i$. The matrix $H$ is the transpose of the
Jacobian matrix $A$. The Alexander polynomial $\Delta_{L}(t)=A_{ii}=H_{ii}$.
By Corollary 4.4,
$H_{ii}=\sum\prod_{u_{kj}^{\delta}\in\operatorname{Tr}(i)}f(u_{kj}^{\delta}).$
Since $f(u_{kj}^{\delta})=1,\text{or}\ -t$, the product
$\prod_{u_{kj}^{\delta}\in\operatorname{Tr}(i)}f(u_{kj}^{\delta})$ is of the
form $(-1)^{l}t^{l}$ and $H_{ii}$ is an alternating polynomial. Therefore,
$\Delta_{L}(t)$ is alternating. ∎
###### Example 4.7.
Up to $6$ classical crossings, the almost classical knots in Table 1 do not
have alternating Alexander polynomials. Therefore, by Theorem 4.6 they do not
admit alternating virtual knot diagrams. $\Diamond$
$K$ | $\Delta_{K}(t)$
---|---
5.2331 | $t^{2}-1+t^{-1}$
6.85091 | $1+t^{-1}-t^{-2}$
6.85774 | $t-1+t^{-2}$
6.87548 | $-t^{2}+2t+1-t^{-1}$
6.87875 | $t+1-2t^{-1}+t^{-2}$
6.89156 | $2t-1-t^{-1}+t^{-2}$
6.89812 | $t^{2}-2+2t^{-1}$
6.90099 | $t-t^{-1}+t^{-3}$
Table 1. Almost classical knots with non-alternating Alexander polynomials.
###### Remark 4.8.
Any integral polynomial $\Delta(t)$ of degree $2n$ satisfying $\Delta(1)=1$
and $\Delta(t)=t^{2n}\Delta(t^{-1})$ is the Alexander polynomial for some
classical knot, see [Sei35]. In [Fox62], Fox asked for a characterization of
Alexander polynomials of alternating knots. If $K$ is an alternating knot,
then $\Delta_{K}(t)=\sum_{j=0}^{2n}(-1)^{j}a_{j}t^{j}$. Fox conjectured that
the Alexander polynomial of any alternating knot satisfies the trapezoidal
inequalities:
$a_{0}<a_{1}<\cdots<a_{n-k}=\cdots=a_{n+k}>\cdots>a_{2n}.$
Fox verified this conjecture for alternating knots up to 11 crossings, and it
has been verified in many other cases [Jon09, HM13, AC21, Che21]. Despite this
progress, Fox’s trapezoidal conjecture remains an intriguing open problem.
Any integral polynomial $\Delta(t)$ satisfying $\Delta(1)=1$ is the Alexander
polynomial for some almost classical knot $K$, see [BCG20]. Is there a way to
characterize the Alexander polynomials of alternating almost classical knots?
Do they satisfy Fox’s trapezoidal inequalities? This has been verified for
almost classical knots up to six classical crossings.
## §5. Split alternating virtual links are visibly split
A classical result of Bankwitz [Ban30] implies that $\det(L)$ is nontrivial
for non-split alternating links. We extend this result to virtual alternating
links and apply it to show that an alternating virtual link $L$ is split if
and only if it is visibly split.
The weak form of the first Tait Conjecture, namely that every knot having a
reduced alternating diagram with at least one crossing is nontrivial, was
first proved by Bankwitz [Ban30] in 1930; and since then, Menasco and
Thistlethwaite [MT91] and Andersson [And95] published simpler proofs. Here we
outline the proof by Balister et al. [BBRS01] and generalize it to alternating
virtual links. This result was first proved for alternating virtual knots by
Cheng [Che15, Proposition 3.3].
Consider the graph $\Gamma$ with vertices $\\{c_{1},\ldots,c_{n}\\}$ as
before.
###### Definition 5.1.
The outdegree of the vertex $c_{i}$, denoted $d^{+}(c_{i})$, is the number of
edges of $\Gamma$ with initial point $c_{i}$. The indegree of the vertex
$c_{i}$, denoted $d^{-}(c_{i})$, is the number of edges of $\Gamma$ with
terminal point $c_{i}$. Therefore,
$d^{+}(c_{i})=\sum_{j=1}^{n}a_{ij}\ \ ,\ \ d^{-}(c_{i})=\sum_{j=1}^{n}a_{ji}.$
###### Definition 5.2.
A walk in a graph is an alternating sequence of vertices and edges, starting
with a vertex $c_{i}$ and ending with a vertex $c_{j}$. A walk is called a
trail if all the edges in that walk are distinct. A circuit is a trail which
starts and ends at a vertex $c_{i}$. An Eulerian circuit is a circuit which
contains all the edges of $\Gamma$. A graph $\Gamma$ is called Eulerian if it
has an Eulerian circuit.
An Eulerian graph is necessarily connected and has $d^{+}(c_{i})=d^{-}(c_{i})$
for every vertex. Let $t_{i}(\Gamma)$ be the number of rooted trees with root
$c_{i}$, then the BEST Theorem is as follows (see [vAEdB51] and [Bol98,
Theorem 13]).
###### Theorem 5.3.
Let $s(\Gamma)$ be the number of Eulerian circuits of $\Gamma$, then
$s(\Gamma)=t_{i}(\Gamma)\prod_{j=1}^{n}(d^{+}(c_{j})-1)!$
In particular, if $\Gamma$ is a two-in two-out oriented graph, i.e.,
$d^{+}(c_{i})=d^{-}(c_{i})=2$ for every $i$, then by Theorems 4.3 and 5.3,
$s(\Gamma)=t_{i}(\Gamma)=H_{ii},\ \ \text{for every}\ i.$
A vertex $c$ of a graph $\Gamma$ is an articulation vertex if $\Gamma$ is the
union of two nontrivial graphs with only the vertex $c$ in common. In [BBRS01]
Balister et al. proved the following result:
Figure 10. The oriented smoothing at $c$.
###### Theorem 5.4.
Let $\Gamma$ be a connected two-in two-out oriented graph with $n\geq 2$
vertices and with no articulation vertex. Then $s(\Gamma)\geq n$.
Given an oriented virtual link diagram $D$, recall that the oriented smoothing
at a crossing $c$ is the diagram with the crossing $c$ removed, see Figure 10.
Recall also that a self-crossing of $D$ is a crossing where one of the
components of the link crosses over itself.
###### Definition 5.5.
Let $D$ be an oriented non-split virtual link diagram. Then a self-crossing
$c$ is said to be nugatory if the oriented smoothing of $D$ at $c$ is a split
link diagram.
The diagram $D$ is said to be reduced if it does not contain any nugatory
crossings.
There is an equivalent definition of nugatory crossing for links in surfaces.
Let $c$ be a crossing of a connected link diagram $\mathcal{D}$ on a surface
$\Sigma$. Then $c$ is said to be nugatory if there is a simple closed curve on
$\Sigma$ that separates $\Sigma$ and intersects $\mathcal{D}$ exactly once at
$c$.
For classical links, nugatory crossings are always removable. For virtual
links, this is no longer true. Indeed, there are examples of virtual knots
that contain essential nugatory crossings, see [BKS20, Example 20]. For welded
links, nugatory crossings are once again always removable, see Remark 6.5
below.
Recall that associated with an alternating virtual link diagram $D$, there is
an oriented two-in two-out graph $\Gamma$. If $D$ has no nugatory crossings,
then $\Gamma$ has no articulation vertex.
###### Corollary 5.6.
Let $K$ be an almost classical knot and $D$ a reduced alternating diagram for
$K$. If $D$ has $n$ classical crossings, then
$|\Delta_{K}(-1)|\geq n.$
###### Proof.
By the proof of Theorem 4.6, $|\Delta_{K}(-1)|$ counts $t_{i}(\Gamma)$ the
number of rooted trees with root $c_{i}$ in the oriented graph $\Gamma$,
associated with the knot diagram $D$. By Theorem 5.3,
$t_{i}(\Gamma)=s(\Gamma)$, and the result follows from Theorem 5.4. ∎
###### Theorem 5.7.
Let $L$ be a non-split virtual link and $D$ a reduced alternating diagram for
$L$. If $D$ has $n$ classical crossings, then the determinant of $L$ satisfies
$\det(L)\geq n.$
###### Proof.
Since $D$ is alternating, we can repeat the proof of Theorem 4.6. By Corollary
4.4, the determinant of $L$ counts the number of spanning trees which is equal
to $s(\Gamma)$. The result follows from Theorem 5.4. ∎
###### Corollary 5.8.
Suppose $L$ is a virtual link which admits an alternating diagram $D$ without
nugatory crossings. Then $L$ is a split link if and only if $D$ is a split
diagram.
###### Proof.
Clearly, if $D$ is a split diagram, then $L$ is split. Suppose then that $D$
is a non-split alternating diagram with $n=n(D)>0$ classical crossings. (If
$n=0$, then $D$ has one component and is an unknot diagram.) Theorem 5.7
implies that $\det(L)\geq n$. Hence $\det(L)\neq 0$, and Proposition 3.2 shows
that $L$ is not split. ∎
## §6. Weakly alternating virtual links
In this section, we extend the results from the previous section to semi-
alternating virtual links, defined below. We also give a formula for the link
determinant of a connected sum, and we use it to show that a semi-alternating
virtual link is split if and only if it is visibly split.
We begin by reviewing the connected sum of virtual links. Suppose $D_{1}$ and
$D_{2}$ are virtual link diagrams and $p_{1},p_{2}$ are points on
$D_{1},D_{2}$ respectively, distinct from the crossings. The connected sum is
denoted $D_{1}\\#D_{2}$ and is formed by removing small arcs from $D_{1}$ and
$D_{2}$ near the basepoints and joining them with trivial unknotted arcs. If
$D_{1}$ and $D_{2}$ are oriented, then the arcs are required to preserve
orientations. The connected sum depends on the choice of diagrams and
basepoints. It does not lead to a well-defined operation on virtual links.
###### Definition 6.1.
A virtual link diagram $D$ is said to be semi-alternating if it can be written
$D=D_{1}\\#\cdots\\#D_{n}$, a connected sum of alternating virtual link
diagrams $D_{1},\ldots,D_{n}$.
The set of semi-alternating virtual links includes, as a proper subset, those
that can be represented as weakly alternating links in thickened surfaces, see
[BKS20, §5].
Every semi-alternating virtual link diagram is checkerboard colorable. This
follows from the fact that every alternating virtual link diagram is
checkerboard colorable (see [Kam02, Lemma 7]), and the observation that the
connected sum of two or more checkerboard colorable diagrams is checkerboard
colorable.
###### Definition 6.2.
A virtual link is said to be $w$-split if it is welded equivalent to a split
virtual link.
Clearly, any virtual link that is split is necessarily $w$-split, but there
are virtual links that are $w$-split but not split. For example, consider the
virtual link $L$ whose Gauss diagram appears on the left of Figure 11. Using
forbidden moves, it is seen to be welded equivalent to the split classical
link $8_{20}\cup\bigcirc$ shown on the right. Thus $L$ is $w$-split.
Let $L^{\prime}=8_{20}\cup\bigcirc$ be the split classical link shown on the
right of Figure 11. Using [Lic97, Definition 3.1 & Theorem 3.5], one can see
that its Jones polynomial satisfies
$\begin{split}(t^{-1/2}-t^{1/2})V(L^{\prime})&=(t-t^{-1})V(8_{20})\\\
&=t^{-6}-t^{-5}-t^{-3}+2t-t^{2}.\end{split}$
In particular, $(t^{-1/2}-t^{1/2})V(L^{\prime})$ lies in
${\mathbb{Z}}[t,t^{-1}].$ (This is true for any classical link with two
components.) On the other hand, direct computation shows that the Jones
polynomial of $L$ satisfies
$\begin{split}(t^{-1/2}-t^{1/2})V(L)=&-t^{-1/2}+3t^{-3/2}+2t^{-2}-3t^{-5/2}-3t^{-3}+2t^{-7/2}+\\\
&\quad 3t^{-4}-t^{-9/2}-3t^{-5}-t^{-11/2}+t^{-6}+t^{-13/2}.\end{split}$
Since $(t^{-1/2}-t^{1/2})V(L)$ does not lie in ${\mathbb{Z}}[t,t^{-1}]$, $L$
cannot be virtually equivalent to a classical link. On the other hand, if $L$
were split, then it would be virtually equivalent to $L^{\prime}$. Since that
is not the case, we see that $L$ is non-split.
Figure 11. The Gauss diagram for a non-split virtual link $L$ (left). Notice
that $L$ is $w$-split; in fact it is welded equivalent to the link
$8_{20}\cup\bigcirc$ (right).
###### Proposition 6.3.
If $L$ is $w$-split, then $\det(L)=0.$
###### Proof.
This follows directly from Proposition 3.2 and the fact that $\det(L)$ is an
invariant of welded links. ∎
There is a nice geometric interpretation of $\det(L)$ in terms of two-fold
branched covers. Let $T=\operatorname{Tube}(L)$ be the ribbon torus link
associated to $L$, and let $X$ be the two-fold cover of $S^{4}$ branched along
$T$. Using the isomorphism $\pi_{1}(S^{4}\smallsetminus T)\cong G_{L}$, one
can identify $\pi_{1}(X)$ with the quotient of $G_{L}$ under the relation
$x_{i}^{2}=1$ for each generator in (2). The coloring matrix is then a
presentation matrix for $H_{1}(X)$. Therefore, $\det(L)=|H_{1}(X)|$ if it is
finite, and $\det(L)=0$ if $H_{1}(X)$ is infinite. Here homology groups are
taken with ${\mathbb{Z}}$ coefficients. Note that, if $L$ is split, then
$H_{1}(X)$ is infinite. This gives an alternative explanation for Propositions
3.2 and 6.3.
###### Theorem 6.4.
If $D=D_{1}\\#D_{2}$ is a connected sum of two checkerboard colorable virtual
link diagrams, then $\det(D)=\det(D_{1})\det(D_{2}).$
###### Proof.
If $D_{1}$ or $D_{2}$ is split, then $D$ is split and
$\det(D)=0=\det(D_{1})\det(D_{2}).$ Therefore, we can assume that $D_{1}$ and
$D_{2}$ are non-split.
There is a proof which is direct and elementary but long. We present an
alternative proof that is shorter and makes use of the interpretation of
$\det(D)$ as the order of the first homology of the two-fold cover of $S^{4}$
branched along $\operatorname{Tube}(D)$. In the following, all homology groups
are taken with ${\mathbb{Z}}$ coefficients.
Let $D_{1}$ and $D_{2}$ be checkerboard colorable virtual link diagrams, and
let $X_{1},X_{2},$ and $X$ be the two-fold covers of $S^{4}$ branched along
$\operatorname{Tube}(D_{1}),\operatorname{Tube}(D_{2})$ and
$\operatorname{Tube}(D)$, respectively. We can then write $X_{1}=A_{1}\cup
B_{1},X_{2}=A_{2}\cup B_{2}$, and $X=A_{1}\cup A_{2}$. Here $A_{i}$ is the
double cover of $D^{4}$ branched along the knotted annulus which is part of
$\operatorname{Tube}(D_{i})$ for $i=1,2$, and $B_{i}$ is the double cover of
$D^{4}$ branched along trivial annulus. In particular,
$A_{i}=X_{i}\smallsetminus\operatorname{Int}(B_{i})$ for $i=1,2$. By [AK80,
Corollary 4.3], $B_{i}$ is diffeomorphic to $S^{2}\times D^{2}$, and
$H_{1}(B_{i})=0$ for $i=1,2.$
Let $M=A_{1}\cap B_{1}=A_{2}\cap B_{2}=A_{1}\cap A_{2}$. Then $M$ is the
3-manifold obtained as the double cover of $S^{3}$ branched along the two
component unlink. Thus $M$ can be identified with the boundary of $B_{1}$ (or
$B_{2}$) and is diffeomorphic to $S^{2}\times S^{1}$. Thus
$H_{1}(M)\cong{\mathbb{Z}}.$
Now consider the decompositions $X_{1}=A_{1}\cup B_{1}$, $X_{2}=A_{2}\cup
B_{2}$, and $X=A_{1}\cup A_{2}$, along with their Mayer-Vietoris sequences in
reduced homology:
(3) $\begin{array}[]{c}\lx@xy@svg{\hbox{\raise 0.0pt\hbox{\kern
6.75pt\hbox{\ignorespaces\ignorespaces\ignorespaces\hbox{\vtop{\kern
0.0pt\offinterlineskip\halign{\entry@#!@&&\entry<EMAIL_ADDRESS>0.0pt\hbox{\hbox{\kern 0.0pt\raise 0.0pt\hbox{\hbox{\kern 3.0pt\raise
0.0pt\hbox{$\textstyle{\cdots\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$}}}}}}}\ignorespaces\ignorespaces\ignorespaces\ignorespaces{}{\hbox{\lx@xy@droprule}}\ignorespaces\ignorespaces{\hbox{\kern
26.31961pt\raise 0.0pt\hbox{{}\hbox{\kern 0.0pt\raise 0.0pt\hbox{\hbox{\kern
3.0pt\hbox{\hbox{\kern 0.0pt\raise 0.0pt\hbox{$\scriptstyle{}$}}}\kern
3.0pt}}}}}}\ignorespaces{\hbox{\kern 30.75pt\raise 0.0pt\hbox{\hbox{\kern
0.0pt\raise
0.0pt\hbox{\lx@xy@tip{1}\lx@xy@tip{-1}}}}}}{\hbox{\lx@xy@droprule}}{\hbox{\lx@xy@droprule}}{\hbox{\kern
30.75pt\raise 0.0pt\hbox{\hbox{\kern 0.0pt\raise 0.0pt\hbox{\hbox{\kern
3.0pt\raise 0.0pt\hbox{$\textstyle{H_{1}(A_{1}\cap
B_{1})\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$}}}}}}}\ignorespaces\ignorespaces\ignorespaces\ignorespaces{}{\hbox{\lx@xy@droprule}}\ignorespaces\ignorespaces\ignorespaces{\hbox{\kern
94.49788pt\raise 5.1875pt\hbox{{}\hbox{\kern 0.0pt\raise
0.0pt\hbox{\hbox{\kern 3.0pt\hbox{\hbox{\kern
0.0pt\raise-0.8264pt\hbox{$\scriptstyle{\varphi_{1}\quad}$}}}\kern
3.0pt}}}}}}\ignorespaces{\hbox{\kern 110.52844pt\raise 0.0pt\hbox{\hbox{\kern
0.0pt\raise
0.0pt\hbox{\lx@xy@tip{1}\lx@xy@tip{-1}}}}}}{\hbox{\lx@xy@droprule}}{\hbox{\lx@xy@droprule}}{\hbox{\kern
110.52844pt\raise 0.0pt\hbox{\hbox{\kern 0.0pt\raise 0.0pt\hbox{\hbox{\kern
3.0pt\raise 0.0pt\hbox{$\textstyle{H_{1}(A_{1})\oplus
H_{1}(B_{1})\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$}}}}}}}\ignorespaces\ignorespaces\ignorespaces\ignorespaces{}{\hbox{\lx@xy@droprule}}\ignorespaces\ignorespaces\ignorespaces{\hbox{\kern
177.7525pt\raise 6.1111pt\hbox{{}\hbox{\kern 0.0pt\raise
0.0pt\hbox{\hbox{\kern 3.0pt\hbox{\hbox{\kern
0.0pt\raise-1.75pt\hbox{$\scriptstyle{\qquad\psi_{1}}$}}}\kern
3.0pt}}}}}}\ignorespaces{\hbox{\kern 213.34296pt\raise 0.0pt\hbox{\hbox{\kern
0.0pt\raise
0.0pt\hbox{\lx@xy@tip{1}\lx@xy@tip{-1}}}}}}{\hbox{\lx@xy@droprule}}{\hbox{\lx@xy@droprule}}{\hbox{\kern
213.34296pt\raise 0.0pt\hbox{\hbox{\kern 0.0pt\raise 0.0pt\hbox{\hbox{\kern
3.0pt\raise
0.0pt\hbox{$\textstyle{H_{1}(X_{1})\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$}}}}}}}\ignorespaces\ignorespaces\ignorespaces\ignorespaces{}{\hbox{\lx@xy@droprule}}\ignorespaces\ignorespaces{\hbox{\kern
253.96655pt\raise 0.0pt\hbox{{}\hbox{\kern 0.0pt\raise 0.0pt\hbox{\hbox{\kern
3.0pt\hbox{\hbox{\kern 0.0pt\raise 0.0pt\hbox{$\scriptstyle{}$}}}\kern
3.0pt}}}}}}\ignorespaces{\hbox{\kern 274.91516pt\raise 0.0pt\hbox{\hbox{\kern
0.0pt\raise
0.0pt\hbox{\lx@xy@tip{1}\lx@xy@tip{-1}}}}}}{\hbox{\lx@xy@droprule}}{\hbox{\lx@xy@droprule}}{\hbox{\kern
274.91516pt\raise 0.0pt\hbox{\hbox{\kern 0.0pt\raise 0.0pt\hbox{\hbox{\kern
3.0pt\raise
0.0pt\hbox{$\textstyle{0.}$}}}}}}}\ignorespaces}}}}\ignorespaces\\\
\lx@xy@svg{\hbox{\raise 0.0pt\hbox{\kern
6.75pt\hbox{\ignorespaces\ignorespaces\ignorespaces\hbox{\vtop{\kern
0.0pt\offinterlineskip\halign{\entry@#!@&&\entry<EMAIL_ADDRESS>0.0pt\hbox{\hbox{\kern 0.0pt\raise 0.0pt\hbox{\hbox{\kern 3.0pt\raise
0.0pt\hbox{$\textstyle{\cdots\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$}}}}}}}\ignorespaces\ignorespaces\ignorespaces\ignorespaces{}{\hbox{\lx@xy@droprule}}\ignorespaces\ignorespaces{\hbox{\kern
26.31961pt\raise 0.0pt\hbox{{}\hbox{\kern 0.0pt\raise 0.0pt\hbox{\hbox{\kern
3.0pt\hbox{\hbox{\kern 0.0pt\raise 0.0pt\hbox{$\scriptstyle{}$}}}\kern
3.0pt}}}}}}\ignorespaces{\hbox{\kern 30.75pt\raise 0.0pt\hbox{\hbox{\kern
0.0pt\raise
0.0pt\hbox{\lx@xy@tip{1}\lx@xy@tip{-1}}}}}}{\hbox{\lx@xy@droprule}}{\hbox{\lx@xy@droprule}}{\hbox{\kern
30.75pt\raise 0.0pt\hbox{\hbox{\kern 0.0pt\raise 0.0pt\hbox{\hbox{\kern
3.0pt\raise 0.0pt\hbox{$\textstyle{H_{1}(A_{2}\cap
B_{2})\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$}}}}}}}\ignorespaces\ignorespaces\ignorespaces\ignorespaces{}{\hbox{\lx@xy@droprule}}\ignorespaces\ignorespaces\ignorespaces{\hbox{\kern
94.49788pt\raise 5.1875pt\hbox{{}\hbox{\kern 0.0pt\raise
0.0pt\hbox{\hbox{\kern 3.0pt\hbox{\hbox{\kern
0.0pt\raise-0.8264pt\hbox{$\scriptstyle{\varphi_{2}\quad}$}}}\kern
3.0pt}}}}}}\ignorespaces{\hbox{\kern 110.52844pt\raise 0.0pt\hbox{\hbox{\kern
0.0pt\raise
0.0pt\hbox{\lx@xy@tip{1}\lx@xy@tip{-1}}}}}}{\hbox{\lx@xy@droprule}}{\hbox{\lx@xy@droprule}}{\hbox{\kern
110.52844pt\raise 0.0pt\hbox{\hbox{\kern 0.0pt\raise 0.0pt\hbox{\hbox{\kern
3.0pt\raise 0.0pt\hbox{$\textstyle{H_{1}(A_{2})\oplus
H_{1}(B_{2})\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$}}}}}}}\ignorespaces\ignorespaces\ignorespaces\ignorespaces{}{\hbox{\lx@xy@droprule}}\ignorespaces\ignorespaces\ignorespaces{\hbox{\kern
177.7525pt\raise 6.1111pt\hbox{{}\hbox{\kern 0.0pt\raise
0.0pt\hbox{\hbox{\kern 3.0pt\hbox{\hbox{\kern
0.0pt\raise-1.75pt\hbox{$\scriptstyle{\qquad\psi_{2}}$}}}\kern
3.0pt}}}}}}\ignorespaces{\hbox{\kern 213.34296pt\raise 0.0pt\hbox{\hbox{\kern
0.0pt\raise
0.0pt\hbox{\lx@xy@tip{1}\lx@xy@tip{-1}}}}}}{\hbox{\lx@xy@droprule}}{\hbox{\lx@xy@droprule}}{\hbox{\kern
213.34296pt\raise 0.0pt\hbox{\hbox{\kern 0.0pt\raise 0.0pt\hbox{\hbox{\kern
3.0pt\raise
0.0pt\hbox{$\textstyle{H_{1}(X_{2})\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$}}}}}}}\ignorespaces\ignorespaces\ignorespaces\ignorespaces{}{\hbox{\lx@xy@droprule}}\ignorespaces\ignorespaces{\hbox{\kern
253.96655pt\raise 0.0pt\hbox{{}\hbox{\kern 0.0pt\raise 0.0pt\hbox{\hbox{\kern
3.0pt\hbox{\hbox{\kern 0.0pt\raise 0.0pt\hbox{$\scriptstyle{}$}}}\kern
3.0pt}}}}}}\ignorespaces{\hbox{\kern 274.91516pt\raise 0.0pt\hbox{\hbox{\kern
0.0pt\raise
0.0pt\hbox{\lx@xy@tip{1}\lx@xy@tip{-1}}}}}}{\hbox{\lx@xy@droprule}}{\hbox{\lx@xy@droprule}}{\hbox{\kern
274.91516pt\raise 0.0pt\hbox{\hbox{\kern 0.0pt\raise 0.0pt\hbox{\hbox{\kern
3.0pt\raise
0.0pt\hbox{$\textstyle{0.}$}}}}}}}\ignorespaces}}}}\ignorespaces\\\
\lx@xy@svg{\hbox{\raise 0.0pt\hbox{\kern
6.75pt\hbox{\ignorespaces\ignorespaces\ignorespaces\hbox{\vtop{\kern
0.0pt\offinterlineskip\halign{\entry@#!@&&\entry<EMAIL_ADDRESS>0.0pt\hbox{\hbox{\kern 0.0pt\raise 0.0pt\hbox{\hbox{\kern 3.0pt\raise
0.0pt\hbox{$\textstyle{\cdots\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$}}}}}}}\ignorespaces\ignorespaces\ignorespaces\ignorespaces{}{\hbox{\lx@xy@droprule}}\ignorespaces\ignorespaces{\hbox{\kern
26.17291pt\raise 0.0pt\hbox{{}\hbox{\kern 0.0pt\raise 0.0pt\hbox{\hbox{\kern
3.0pt\hbox{\hbox{\kern 0.0pt\raise 0.0pt\hbox{$\scriptstyle{}$}}}\kern
3.0pt}}}}}}\ignorespaces{\hbox{\kern 30.75pt\raise 0.0pt\hbox{\hbox{\kern
0.0pt\raise
0.0pt\hbox{\lx@xy@tip{1}\lx@xy@tip{-1}}}}}}{\hbox{\lx@xy@droprule}}{\hbox{\lx@xy@droprule}}{\hbox{\kern
30.75pt\raise 0.0pt\hbox{\hbox{\kern 0.0pt\raise 0.0pt\hbox{\hbox{\kern
3.0pt\raise 0.0pt\hbox{$\textstyle{H_{1}(A_{1}\cap
A_{2})\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$}}}}}}}\ignorespaces\ignorespaces\ignorespaces\ignorespaces{}{\hbox{\lx@xy@droprule}}\ignorespaces\ignorespaces\ignorespaces{\hbox{\kern
94.91109pt\raise 5.1875pt\hbox{{}\hbox{\kern 0.0pt\raise
0.0pt\hbox{\hbox{\kern 3.0pt\hbox{\hbox{\kern
0.0pt\raise-0.8264pt\hbox{$\scriptstyle{\varphi\quad}$}}}\kern
3.0pt}}}}}}\ignorespaces{\hbox{\kern 109.94165pt\raise 0.0pt\hbox{\hbox{\kern
0.0pt\raise
0.0pt\hbox{\lx@xy@tip{1}\lx@xy@tip{-1}}}}}}{\hbox{\lx@xy@droprule}}{\hbox{\lx@xy@droprule}}{\hbox{\kern
109.94165pt\raise 0.0pt\hbox{\hbox{\kern 0.0pt\raise 0.0pt\hbox{\hbox{\kern
3.0pt\raise 0.0pt\hbox{$\textstyle{H_{1}(A_{1})\oplus
H_{1}(A_{2})\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$}}}}}}}\ignorespaces\ignorespaces\ignorespaces\ignorespaces{}{\hbox{\lx@xy@droprule}}\ignorespaces\ignorespaces\ignorespaces{\hbox{\kern
177.0256pt\raise 6.1111pt\hbox{{}\hbox{\kern 0.0pt\raise
0.0pt\hbox{\hbox{\kern 3.0pt\hbox{\hbox{\kern
0.0pt\raise-1.75pt\hbox{$\scriptstyle{\qquad\psi}$}}}\kern
3.0pt}}}}}}\ignorespaces{\hbox{\kern 212.16937pt\raise 0.0pt\hbox{\hbox{\kern
0.0pt\raise
0.0pt\hbox{\lx@xy@tip{1}\lx@xy@tip{-1}}}}}}{\hbox{\lx@xy@droprule}}{\hbox{\lx@xy@droprule}}{\hbox{\kern
212.16937pt\raise 0.0pt\hbox{\hbox{\kern 0.0pt\raise 0.0pt\hbox{\hbox{\kern
3.0pt\raise
0.0pt\hbox{$\textstyle{H_{1}(X)\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$}}}}}}}\ignorespaces\ignorespaces\ignorespaces\ignorespaces{}{\hbox{\lx@xy@droprule}}\ignorespaces\ignorespaces{\hbox{\kern
250.69296pt\raise 0.0pt\hbox{{}\hbox{\kern 0.0pt\raise 0.0pt\hbox{\hbox{\kern
3.0pt\hbox{\hbox{\kern 0.0pt\raise 0.0pt\hbox{$\scriptstyle{}$}}}\kern
3.0pt}}}}}}\ignorespaces{\hbox{\kern 270.94156pt\raise 0.0pt\hbox{\hbox{\kern
0.0pt\raise
0.0pt\hbox{\lx@xy@tip{1}\lx@xy@tip{-1}}}}}}{\hbox{\lx@xy@droprule}}{\hbox{\lx@xy@droprule}}{\hbox{\kern
270.94156pt\raise 0.0pt\hbox{\hbox{\kern 0.0pt\raise 0.0pt\hbox{\hbox{\kern
3.0pt\raise
0.0pt\hbox{$\textstyle{0.}$}}}}}}}\ignorespaces}}}}\ignorespaces\end{array}$
Note that $H_{1}(A_{1}\cap B_{1})\cong{\mathbb{Z}}\cong H_{1}(A_{2}\cap
B_{2})=H_{1}(A_{1}\cap A_{2})$.
We claim that the maps $\varphi_{1},\varphi_{2}$ and $\varphi$ are zero. We
prove this for $\varphi$; the argument for the other cases is similar. It
suffices to show that the maps $H_{1}(A_{1}\cap A_{2})\to H_{1}(A_{i})$
induced by inclusion are zero for $i=1,2$.
Take two points in $S^{3}$, one on each component of the unlink, and join them
by an arc in $S^{3}$ that is otherwise disjoint from the link. The arc lifts
to a loop in the double branched cover, and the loop is a generator of
$H_{1}(A_{1}\cap A_{2})$. However, when pushed into $A_{1}$, the loop does not
link the annulus in $D^{4}$, so it is trivial in $H_{1}(A_{1})$. A similar
argument shows it is also trivial in $H_{1}(A_{2})$. Therefore, the maps
$H_{1}(A_{1}\cap A_{2})\to H_{1}(A_{j})$ are zero for $j=1,2$, and it follows
that $\varphi=0.$
From the claim, it follows that $\psi_{1},\psi_{2},$ and $\psi$ are
isomorphisms. Using (3) and the fact that $H_{1}(B_{1})=0=H_{1}(B_{2})$, we
deduce that
$H_{1}(X_{1})\cong H_{1}(A_{1}),\;H_{1}(X_{2})\cong H_{1}(A_{2}),\text{ and
}H_{1}(X)\cong H_{1}(A_{1})\oplus H_{1}(A_{2}).$
Therefore,
$\begin{split}\det(D)&=|H_{1}(X)|,\\\ &=|H_{1}(A_{1})|\cdot|H_{1}(A_{2})|,\\\
&=|H_{1}(X_{1})|\cdot|H_{1}(X_{2})|,\\\ &=\det(D_{1})\det(D_{2}),\end{split}$
and this completes the proof. ∎
###### Remark 6.5.
We claim that, for welded links, nugatory crossings are always removable. Let
$D$ be a diagram with a nugatory crossing $c$ as in Figure 12 (left). Using
forbidden moves, we can transform $D$ by pulling the over-crossing arc off
$c$, as in Figure 12 (middle). Thus, $D$ is welded equivalent to the diagram
with $c$ removed. Alternatively, one can remove $c$ by making it virtual, as
in Figure 12 (right). At the level of Gauss diagrams, this is equivalent to
deleting the chord associated to $c$.
Figure 12. For welded links, nugatory crossing are removable.
###### Proposition 6.6.
Let $D$ be a virtual link diagram. If $D$ is non-split and alternating, then
$\det(D)\neq 0$.
###### Proof.
In general, if $D$ is a non-reduced virtual link diagram, then successively
removing all the nugatory crossings will produce a reduced diagram
$D^{\prime}$ welded equivalent to $D$. If $D$ is non-split, then $D^{\prime}$
will be too. If $D$ is alternating, then $D^{\prime}$ will be semi-
alternating.
Assume to the contrary that $\det(D)=0.$ Then $\det(D^{\prime})=0$ since
$\det(\cdot)$ is an invariant of welded type. Since $D$ is non-split and
alternating, it follows that $D^{\prime}$ is reduced, non-split, and semi-
alternating. Therefore, we can write
$D^{\prime}=D^{\prime}_{1}\\#\cdots\\#D^{\prime}_{n}$, where
$D^{\prime}_{1},\ldots,D^{\prime}_{n}$ are all reduced alternating diagrams.
By Theorem 6.4,
$0=\det(D^{\prime})=\det(D^{\prime}_{1})\cdots\det(D^{\prime}_{n}),$
thus $\det(D^{\prime}_{i})=0$ for some $1\leq i\leq n.$ Since $D^{\prime}_{i}$
is reduced and alternating, $\det(D^{\prime}_{i})=0$ implies that
$D^{\prime}_{i}$ is split. It follows that $D^{\prime}$ is split, which
implies that $D$ is split, giving the desired contradiction. ∎
###### Corollary 6.7.
Suppose $L$ is a virtual link which admits a semi-alternating diagram $D$,
possibly with nugatory crossings. Then $L$ is $w$-split if and only if $D$ is
a split diagram.
###### Proof.
Clearly if $D$ is split, then $L$ is split and also $w$-split.
On the other hand, suppose $D$ is non-split. Since $D$ is semi-alternating, we
can write $D=D_{1}\\#\cdots\\#D_{n}$, where $D_{1},\ldots,D_{n}$ are all non-
split, alternating diagrams. Proposition 6.6 implies that $\det(D_{i})\neq 0$
for $i=1,\ldots,n.$ Theorem 6.4 implies that
$\det(D)=\prod_{i=1}^{n}\det(D_{i})\neq 0.$ Therefore, $\det(L)\neq 0$, and by
Proposition 6.3, it follows that $L$ is not $w$-split. ∎
## §7. The Tait conjectures for welded links
In his early work on knot tabulation, Tait formulated three far-reaching
conjectures on reduced alternating classical link diagrams [Tai98]. (A link
diagram is reduced if it does not contain a nugatory crossing.) They assert
that, for a non-split link, any two reduced alternating diagrams have the same
crossing number, the same writhe, and are related by a sequence of flype
moves. The first two statements were famously solved by Kauffman, Murasugi,
and Thistlethwaite using the recently discovered Jones polynomial [Kau87,
Mur87, Thi87], and the third statement was subsequently proved by Menasco and
Thistlethwaite [MT93]. The three Tait conjectures lead to a simple and
effective algorithm for tabulating alternating knots and links that has been
implemented [RF04, RF06].
It is an interesting question whether similar results hold for virtual and/or
welded links. For example, analogues of the first and second Tait conjectures
have been established for virtual links using the Jones-Krushkal polynomial
and skein bracket, see [BK19, BKS20].
###### Problem 7.1.
Is the Tait flype conjecture true for alternating virtual links?
Figure 13. The flype move.
The flype move is depicted in Figure 13. By assumption, the tangle $T$ does
not contain any virtual crossings. Allowing the tangle to contain virtual
crossings results in a more general move called a virtual flype move. The
virtual flype move does not, in general, preserve the virtual link type, for
example, see [Kam17, ZJZ04].
It is unknown whether the Tait conjectures hold for welded links. More
generally, what conditions must the invariants of welded link satisfy in order
for it to be alternating?
Since $\det(L)$ is an invariant of welded links, any checkerboard colorable
virtual $L$ with $\det(L)\neq 1$ is nontrivial as a welded link. In
particular, Theorem 5.7 applies to show that any non-split virtual link
represented by a reduced, alternating diagram has $\det(L)\neq 1$ and
therefore, is nontrivial as a welded link.
The Alexander polynomial $\Delta_{L}(t)$ is also an invariant of the welded
type. Therefore, if $L$ is almost classical and $\Delta_{L}(t)$ is not
alternating, then $L$ is not welded equivalent to an alternating link.
Figure 14. Alternating welded knots with 3 and 4 classical crossings.
Figure 14 shows the five alternating welded knots with up to four classical
crossings. All the others can be ruled out using the consideration that
$\det(K)\geq n$, the crossing number.
###### Problem 7.2.
Is the first Tait conjecture true for alternating welded links?
One can find examples of virtual knots which are non-alternating but which
become alternating after adding one crossing. For example, consider Examples
19 and 20, [BKS20]. The first is non-alternating and has six crossings; the
second is alternating and is obtained by adding a nugatory crossing. The two
virtual knots are welded equivalent (see Figure 12). We conjecture that there
exist welded knots which are alternating, but every minimal crossing diagram
for them is non-alternating.
###### Problem 7.3.
Is it possible for an alternating welded knot to represent a non-alternating
classical knot?
Figure 15. Reduced alternating diagrams for the virtual knots $4.106$ and
$4.107$.
Interestingly, there are pairs of reduced alternating virtual knot diagrams
which are equivalent as welded knots but distinct as virtual knots. In
particular, this implies that Tait’s second conjecture is not true for welded
knots.
Figure 16. A sequence of moves on the Gauss diagrams, starting from a diagram
of $4.106$ ending in a diagram for $4.107$.
For example, consider the virtual knots $4.106$ and $4.107$ in Figure 15. Both
are reduced alternating diagrams, but the diagram for $4.106$ has writhe
$w=-2$ whereas the diagram for $4.107$ has writhe $w=0.$ Tait’s second
conjecture holds for reduced alternating virtual knot diagrams [BK19], and
thus comparing the writhes tells us these two are distinct as virtual knots.
However, these diagrams are equivalent as welded knots (see Figure 16 and
Figure 17). Since both diagrams are reduced and alternating, this shows that
the writhe of a reduced alternating diagram is not invariant under welded
equivalence.
This implies that the second Tait conjecture is not true in the welded
category. Since the Tait flype move preserves the writhe, this also shows that
the Tait’s third conjecture, if true, must take a different form in the
virtual and welded settings.
Figure 17. A sequence of moves on the virtual diagrams, starting from a
diagram of $4.106$ ending in a diagram for $4.107$. The fourth and fifth
diagrams are related by an $f1$ move, and this is seen by comparing Gauss
diagrams.
## Acknowledgements
This paper is based on several ideas in the Ph.D. thesis of the second author
[Kar18]. The authors would like to thank Robin Gaudreau, Andy Nicas, Will
Rushworth, and Adam Sikora for their helpful comments and feedback. They would
also like to thank the referee for their input. The first author gratefully
acknowledges grant funding from the Natural Sciences and Engineering Research
Council of Canada.
## References
* [AC21] Marwa E. Alrefai and Nafaa Chbili. On Fox’s trapezoidal conjecture for closed 3-braids. Internat. J. Math., 32(1):Paper No. 2150002, 11, 2021.
* [AK80] Selman Akbulut and Robion Kirby. Branched covers of surfaces in $4$-manifolds. Math. Ann., 252(2):111–131, 1979/80.
* [And95] Peter Andersson. The color invariant for knots and links. Amer. Math. Monthly, 102(5):442–448, 1995.
* [Aud16] Benjamin Audoux. On the welded tube map. In Knot theory and its applications, volume 670 of Contemp. Math., pages 261–284. Amer. Math. Soc., Providence, RI, 2016.
* [Ban30] Carl Bankwitz. über die Torsionszahlen der alternierenden Knoten. Math. Ann., 103(1):145–161, 1930.
* [BBRS01] Paul N. Balister, Béla Bollobás, Oliver M. Riordan, and Alexander D. Scott. Alternating knot diagrams, Euler circuits and the interlace polynomial. European J. Combin., 22(1):1–4, 2001.
* [BCG20] Hans U. Boden, Micah Chrisman, and Robin Gaudreau. Signature and concordance of virtual knots. Indiana Univ. Math. J., 69(7):2395–2459, 2020.
* [BCK21] Hans U. Boden, Micah Chrisman, and Homayun Karimi. The Gordon-Litherland pairing for links in thickened surfaces, 2021\. ArXiv/2107.00426.
* [BGH+17] Hans U. Boden, Robin I. Gaudreau, Eric Harper, Andrew J. Nicas, and Lindsay White. Virtual knot groups and almost classical knots. Fundamenta Mathematicae, 138:101–142, 2017.
* [BK19] Hans U. Boden and Homayun Karimi. The Jones-Krushkal polynomial and minimal diagrams of surface links, 2019. ArXiv/1908.06453, to appear in Ann. Inst. Fourier (Grenoble).
* [BKS20] Hans U. Boden, Homayun Karimi, and Adam S. Sikora. Adequate links in thickened surfaces and the generalized Tait conjectures, 2020. ArXiv/2008.09895, to appear Alg. Geom. Topol.
* [BM54] Raoul H. Bott and John P. Mayberry. Matrices and trees. In Economic activity analysis, pages 391–400. John Wiley and Sons, Inc., New York; Chapman and Hall, Ltd., London, 1954. Edited by Oskar Morgenstern.
* [BNW18] Hans U. Boden, Andrew J. Nicas, and Lindsay White. Alexander invariants of periodic virtual knots. Dissertationes Math., 530:59, 2018.
* [Bol98] Béla Bollobás. Modern graph theory, volume 184 of Graduate Texts in Mathematics. Springer-Verlag, New York, 1998.
* [BR21] Hans U. Boden and William Rushworth. Minimal crossing number implies minimal supporting genus. Bull. Lond. Math. Soc., 53(4):1174–1184, 2021.
* [BZH14] Gerhard Burde, Heiner Zieschang, and Michael Heusener. Knots, volume 5 of De Gruyter Studies in Mathematics. De Gruyter, Berlin, extended edition, 2014.
* [Che15] Zhiyun Cheng. Kauffman-Harary conjecture for alternating virtual knots. J. Knot Theory Ramifications, 24(8):1550046, 13, 2015.
* [Che21] Wenzhao Chen. On two-bridge knots and a conjecture of Hirasawa-Murasugi. J. Knot Theory Ramifications, 30(2):Paper No. 2150007, 16, 2021\.
* [CKS02] J. Scott Carter, Seiichi Kamada, and Masahico Saito. Stable equivalence of knots on surfaces and virtual knot cobordisms. J. Knot Theory Ramifications, 11(3):311–322, 2002. Knots 2000 Korea, Vol. 1 (Yongpyong).
* [Cro59a] Richard Crowell. Genus of alternating link types. Ann. of Math. (2), 69:258–275, 1959.
* [Cro59b] Richard H. Crowell. Nonalternating links. Illinois J. Math., 3:101–120, 1959.
* [Fox62] R. H. Fox. Some problems in knot theory. In Topology of 3-manifolds and related topics (Proc. The Univ. of Georgia Institute, 1961), Ser. Knots Everything, pages 168–176. Prentice-Hall, Englewood Cliffs, N.J., 1962.
* [GPV00] Mikhael Goussarov, Michael Polyak, and Oleg Viro. Finite-type invariants of classical and virtual knots. Topology, 39(5):1045–1068, 2000.
* [HM13] Mikami Hirasawa and Kunio Murasugi. Various stabilities of the Alexander polynomials of knots and links, 2013. ArXiv/1307.1578.
* [ILL10] Young Ho Im, Kyeonghui Lee, and Sang Youl Lee. Signature, nullity and determinant of checkerboard colorable virtual links. J. Knot Theory Ramifications, 19(8):1093–1114, 2010.
* [Jon09] In Dae Jong. Alexander polynomials of alternating knots of genus two. Osaka J. Math., 46(2):353–371, 2009.
* [Kam02] Naoko Kamada. On the Jones polynomials of checkerboard colorable virtual links. Osaka J. Math., 39(2):325–333, 2002.
* [Kam17] Naoko Kamada. Converting virtual link diagrams to normal ones. Topology Appl., 230:161–171, 2017.
* [Kar18] Homayun Karimi. Alternating Virtual Knots. PhD thesis, McMaster University, September 2018. McMaster University.
* [Kau87] Louis H. Kauffman. State models and the Jones polynomial. Topology, 26(3):395–407, 1987.
* [Kau99] Louis H. Kauffman. Virtual knot theory. European J. Combin., 20(7):663–690, 1999.
* [KNS02] Naoko Kamada, Shigekazu Nakabo, and Shin Satoh. A virtualized skein relation for Jones polynomials. Illinois J. Math., 46(2):467–475, 2002.
* [Lic97] W. B. Raymond Lickorish. An introduction to knot theory, volume 175 of Graduate Texts in Mathematics. Springer-Verlag, New York, 1997.
* [Man10] Vassily O. Manturov. Parity in knot theory. Mat. Sb., 201(5):65–110, 2010.
* [MT91] William Menasco and Morwen Thistlethwaite. A geometric proof that alternating knots are nontrivial. Math. Proc. Cambridge Philos. Soc., 109(3):425–431, 1991.
* [MT93] William Menasco and Morwen Thistlethwaite. The classification of alternating links. Ann. of Math. (2), 138(1):113–171, 1993.
* [Mur58] Kunio Murasugi. On the genus of the alternating knot. I, II. J. Math. Soc. Japan, 10:94–105, 235–248, 1958.
* [Mur87] Kunio Murasugi. Jones polynomials and classical conjectures in knot theory. Topology, 26(2):187–194, 1987.
* [Nik13] Igor M. Nikonov. Weak parities and functional maps. Sovrem. Mat. Fundam. Napravl., 51:123–141, 2013. J. Math. Sci. (N.Y.) 214 (2016), no. 5 699–717.
* [NNST12] Takuji Nakamura, Yasutaka Nakanishi, Shin Satoh, and Yumi Tomiyama. Twin groups of virtual 2-bridge knots and almost classical knots. J. Knot Theory Ramifications, 21(10):1250095, 18, 2012.
* [RF04] Stuart Rankin and Ortho Flint. Enumerating the prime alternating links. J. Knot Theory Ramifications, 13(1):151–173, 2004.
* [RF06] Stuart Rankin and Ortho Flint. _Knotilus_ , a computer package for enumerating alternating knots and links, 2006. www.knotilus.math.uwo.ca.
* [Rus21] William Rushworth. A parity for 2-colourable links. Osaka J. Math., 58(4):767–801, 2021.
* [Sat00] Shin Satoh. Virtual knot presentation of ribbon torus-knots. J. Knot Theory Ramifications, 9(4):531–542, 2000.
* [Sei35] Herbert Seifert. Über das Geschlecht von Knoten. Math. Ann., 110(1):571–592, 1935.
* [Tai98] Peter Guthrie Tait. On Knots I, II, III. Scientific papers, I:273–347, 1898.
* [Thi87] Morwen B. Thistlethwaite. A spanning tree expansion of the Jones polynomial. Topology, 26(3):297–309, 1987.
* [Tut48] William T. Tutte. The dissection of equilateral triangles into equilateral triangles. Proc. Cambridge Philos. Soc., 44:463–482, 1948.
* [vAEdB51] Tatyana P. van Aardenne-Ehrenfest and Nicolaas G. de Bruijn. Circuits and trees in oriented linear graphs. Simon Stevin, 28:203–217, 1951.
* [Win09] Blake Winter. The classification of spun torus knots. J. Knot Theory Ramifications, 18(9):1287–1298, 2009.
* [ZJZ04] Paul Zinn-Justin and Jean-Bernard Zuber. Matrix integrals and the generation and counting of virtual tangles and links. J. Knot Theory Ramifications, 13(3):325–355, 2004.
|
# Inelastic x-ray scattering reveals the ergodic to nonergodic transition of
salol, a liquid with local order
L. Comez${}^{\textsf{\footnotesize{\mbox{?}}}}$, D.
Fioretto${}^{\textsf{\footnotesize{\mbox{?}}}}$, J.
Gapinski${}^{\textsf{\footnotesize{\mbox{?}}}}$, G.
Monaco${}^{\textsf{\footnotesize{\mbox{?}}}}$, A.
Patkowski${}^{\textsf{\footnotesize{\mbox{?}}}}$, W.
Steffen${}^{\textsf{\footnotesize{\mbox{?}}}}$
(Received June 25, 2019)
###### Abstract
We have studied the high-frequency dynamics of salol by inelastic x-ray
scattering over a wide temperature range between 50 and 450 K, across the
glass transition. We find that salol efficiently realizes the mechanism of
dynamical arrest described by the mode-coupling theory, as manifested by a
cusp singularity in the behaviour of the non-ergodicity parameter and a $Q$
dependence of the critical non-ergodicity parameter that is in phase with the
static structure factor. These results confront positively the mode-coupling
theory with liquids with local order.
Key words: glass transition, x-ray scattering, mode coupling theory
PACS: 64.70.Pf, 78.70.Ck, 61.20.Lc
###### Abstract
Ìè äîñëäèëè âèñîêîàñòîòíó äèíàìêó ñàëîëà íåïðóæíèì ðîçñþâàííÿì ðåíòãåíâñüêèõ
ïðîìåíâ â øèðîêé îáëàñò òåìïåðàòóð ìæ 50 450 K, ùî ïîêðèâà ïåðåõä â ñòàí ñêëà.
Ìè çíàéøëè, ùî ñàëîë åôåêòèâíî ðåàëçó ìåõàíçì äèíàìíîãî àðåøòó, ÿêèé îïèñóòüñÿ
òåîðþ âçàìîäþèõ ìîä, ùî ïðîÿâëÿòüñÿ ñèíãóëÿðíñòþ â ïîâåäíö ïàðàìåòðó
íååðãîäèíîñò òà â $Q$-çàëåæíîñò êðèòèíîãî ïàðàìåòðó íååðãîäèíîñò, ùî
óçãîäæóòüñÿ ç ñòàòèíèì ñòðóêòóðíèì ôàêòîðîì. Ö ðåçóëüòàòè ïîçèòèâíî
ñïâñòàâëÿþòü òåîðþ âçàìîäþèõ ìîä ç ðäèíàìè, ÿê ìàþòü ëîêàëüíå âïîðÿäêóâàííÿ.
Ключов слова: ïåðåõä â ñòàí ñêëà, ðîçñþâàííÿ ðåíòãåíâñüêèõ ïðîìåíâ, òåîðÿ
âçàìîäþèõ ìîä
## 1 Introduction
At the turn of the second and third millennium, a great deal of work, both
theoretical and experimental, has been devoted to the study of the physics of
disordered systems with particular emphasis on the processes connected with
the structural arrest and glass transition.
The mode coupling theory (MCT) was introduced to provide a self-consistent
treatment of the structural arrest in simple liquids [1]. The idealized
version of MCT describes the glass transition as an ergodic to non-ergodic
transition occurring at a critical temperature $T_{\text{c}}$, associated with
a singular behaviour of the long-time limit of the normalized density
correlator, the so-called non-ergodicity factor $f_{Q}$. The peculiar
temperature ($T$) and wave vector ($Q$) dependences expected for $f_{Q}$ are:
i) a square-root temperature behaviour below $T_{\text{c}}$,
$f_{Q}(T)=f_{Q}^{\text{c}}+h_{Q}\sqrt{(T-T_{\text{c}})/T_{\text{c}}}$, where
$f_{Q}^{\text{c}}$ is the critical non-ergodicity parameter and $h_{Q}$ is the
critical amplitude at a given wavevector $Q$; ii) a $Q$ dependence of
$f_{Q}^{\text{c}}$ and of $h_{Q}$, which are in phase and in antiphase with
the static structure factor $S(Q)$, respectively.
In the same years, the inelastic x-ray scattering (IXS) technique was
developed, capable of detecting the dynamic structure factor $S(Q,\omega)$ of
glasses [2] and glass forming liquids in the nm-1 $Q$-range [3]. IXS, together
with neutron scattering and MD simulations, became the method of choice for
the quantitative test of MCT predictions. Most of the tests were performed in
simple liquids, such as van der Waals molecular liquids, with relatively
simple relaxation patterns [4]. More complex systems were also analyzed, such
as polymers [5, 6], and a good coherence with MCT was found when the
contribution of the structural relaxation was singled out with respect to
those of secondary processes [7], complementing IXS with Brillouin light
scattering measurements [8]. An interesting class of liquids is that of
associated liquids, where local order extends over several neighboring
molecules, giving rise to a pre-peak in the low-$Q$ region of $S(Q)$. A
pioneering work performed by some of us on the associated liquid m-toluidine
provided experimental evidence that hydrogen bond clustering can coexist with
the signature of the ergodic to non-ergodic transition predicted by the MCT
[9, 10].
In the present work, we extend this investigation to salol, a widely studied
glass-forming system [11, 12, 13, 14].
## 2 Experiment
Salol (Phenyl salicylate, Fluka, purity $>98$%) was dried under vacuum for
three days at $95^{\circ}$C. To remove dust as potential sites of
heterogeneous nucleation, the sample was then filtered through a 0.22 µm
Durapore (Millipore company) filter into dust free vials which were sealed
until used in the actual experiment to fill the sample cells.
IXS experiments were performed at the very high energy resolution beamlines
ID16 and ID28 of the European Synchrotron Radiation Facility (ESRF), Grenoble.
The monochromator and analyzer crystals were operated at backscattering
configuration corresponding to an incident photon energy of 21.747 keV and a
total energy resolution of 1.5 meV. Spectra were taken at $Q$-values between 1
and 15 nm-1 in steps of 1 nm-1 at temperatures between 50 and 450 K. All
spectra were corrected for scattering of the empty cell and normalized to the
monitored incoming intensity.
Figure 1: Left-hand panel: IXS spectra of salol taken at $Q=2$ nm-1, at the
indicated temperatures. Right-hand panel: IXS spectra of salol taken at
$T=241$ K, at the indicated $Q$ values. The fitting curves (solid lines), the
quasi-elastic contributions (dash-dotted lines) and the inelastic
contributions (dashed lines) to the total fit are reported together with the
data points (open circles).
In figure 1 we report some IXS spectra for selected temperatures at the fixed
exchanged wave vector $Q=2$ nm-1 (left-hand panel), and some spectra obtained
for different values of $Q$ at $T=241$ K (right-hand panel).
## 3 Results and discussion
All measured spectra show quasielastic and inelastic contributions, whose
characteristic parameters were obtained by fitting the convolution of the
instrumental resolution function $R(\omega)$ with a model for $S(Q,\omega)$
including a delta function for the quasielastic line and a damped harmonic
oscillator for the two inelastic side peaks [15], which are due to the
Brillouin scattering of photons by longitudinal acoustic (LA) modes
propagating in salol. From this fitting procedure, the characteristic
frequency $(\Omega)$ and linewidth $(\Gamma)$ (FWHM) of the LA modes were
obtained, together with the intensities of the quasielastic and inelastic
contributions, $I_{\text{el}}(Q)$ and $I(Q)$.
Figure 2 shows the values of $\Omega$ and $\Gamma$ obtained by the fitting
procedure, as a function of $Q$ and for three selected temperatures. The
almost linear behaviour of $\Omega$ vs. $Q$ allows us to estimate the average
value for the velocity $(v)$ of the LA modes in the low $Q$ regime by fitting
to the first four data points the expression: $\Omega=v\cdot Q$. In the same
region, the almost quadratic behaviour of $\Gamma$ vs. $Q$ gives an estimate
of the unrelaxed value of the longitudinal kinematic viscosity $D_{\text{L}}$
and thermal diffusion $D_{\text{T}}$ through
$D_{\text{L}}+(\gamma-1)D_{\text{T}}=\Gamma/Q^{2}$ with $\gamma$ being the
ratio of specific heats.
Figure 2: Left-hand panels: Linear dependence on $Q$ of the longitudinal
acoustic frequency in the low-$Q$ range. Right-hand panels: In the $Q$ region,
where the acoustic dispersion relation is linear, the broadening of the
acoustic excitations shows a $T$-independent $Q^{2}$ behaviour.
Velocities and viscosities obtained by this method are reported in figure 3.
In the low temperature regime, the velocity of the LA modes obtained by IXS
agrees very well with the unrelaxed sound velocities $(v_{\infty})$ measured
at much lower $Q$ values by Brillouin light scattering (BLS) [16, 17] and
impulsive stimulated thermal scattering (ISTS) techniques [18]. For increasing
temperature, a change of the slope of $v(T)$ is visible close to the glass
transition of salol $T_{\text{g}}=220$ K, where a more pronounced temperature
dispersion starts to occur [19]. Interestingly, the temperature behaviour of
$v_{\text{IXS}}$ is also compatible [see linear extrapolations in figure 3
(a)] with the existence of a further transition in the liquid at
$T_{\text{A}}=348$ K. $T_{\text{A}}$ was proposed and shown as the transition
temperature of the $\alpha$-process from VFT to Arrhenius behaviour [20, 21].
Figure 3: (Colour online) Left-hand panel: Limiting high frequency
longitudinal sound velocity (squares) determined by a DHO $+$ delta function
analysis of IXS spectra compared with the literature data. Evidence is given
for two changes in the slope corresponding to $T_{\text{g}}$ and
$T_{\text{A}}$. Right-hand panel: Limiting high frequency longitudinal
kinematic viscosity.
At the highest investigated temperatures, the values of $v_{\text{BLS}}$ and
of $v_{\text{ISTS}}$ progressively approach those of the relaxed sound
velocity $v_{0}$. Conversely, the values of $v_{\text{IXS}}$ remain
considerably higher than $v_{0}$, suggesting that IXS probes the unrelaxed
sound velocity $v_{\infty}$ also in the liquid phase. This is also supported
by the temperature dependence of the linewidth of Brillouin peaks reported in
figure 3 (b). In fact, it can be seen that $\Gamma/Q^{2}$ at the two lowest
probed $Q$ values is almost constant in the whole investigated temperature
range suggesting that the mechanism responsible for the peak broadening in the
glassy phase continues to dominate also at temperatures higher than
$T_{\text{g}}$. This result can be explained by the fact that, at the probed
$Q$s, the phonon frequency is much higher than the rate of the structural
relaxation (unrelaxed condition) and, therefore, the sound waves probe the
system as “frozen” in the whole investigated temperature range. In this
regime, the broadening of the Brillouin peaks is due to the disordered
molecular structure rather than due to truly dynamic processes [22].
Figure 4: (Colour online) Temperature dependence of the effective non-
ergodicity factor $f_{Q}$ of salol for selected values of the exchanged
wavevector $Q$. The red line is a guide for eyes to indicate the critical
temperature $T_{\text{c}}=273\pm 5$ K. The solid lines are the best fits
obtained using the square-root function predicted by the mode coupling theory
[1].
An important consequence of the unrelaxed regime here probed by IXS is that
the contribution of the structural relaxation to $S(Q,\Omega)$ is all included
within the area of the quasielastic peak. In this condition, the relative
amplitude of the structural relaxation, i.e., the non-ergodicity parameter
$f_{Q}$, can be obtained from the IXS spectra as the ratio of the intensities
$f_{Q}=I_{\text{el}}(Q)/[I_{\text{el}}(Q)+I(Q)]$, and the analysis of the IXS
spectra becomes a powerful tool to test the predictions of MCT.
Figure 4 reports the temperature behaviour of $f_{Q}$ at four different values
of $Q$, showing clear evidence of the square-root singularity predicted by the
idealized MCT [1], even more clearly than in the previously reported
m-toluidine case [9], due to the improved signal-to-noise ratio of the present
IXS spectra. It is worth noting that the value of the critical temperature
$T_{\text{c}}=273\pm 5$ K is $Q$ independent within experimental error, and
that its value favourably agrees with previous estimates obtained from
dielectric spectroscopy [23] and from wide angle x-ray experiments combined
with MD simulations [21].
Figure 5: Left-hand panel: The non-ergodicity parameter $f_{Q}$ of salol at
different temperatures. Right-hand panel: $I(Q)$ obtained by wide angle x-ray
scattering measurements.
Figure 5 shows that $f_{Q}$ changes in phase with the static structure factor,
in agreement with previous observations [9]. For $T>T_{\text{c}}$, MCT
predicts a plateau in the temperature dependence of $f_{Q}$, i.e.,
$f_{Q}(T)=f_{Q}^{\text{c}}$, which is also clearly visible in figure 4. The
values of $f_{Q}^{\text{c}}$, on the whole, follow in phase the oscillations
of the static structure factor (figure 6), coherent with MCT calculations for
simple liquids.
Figure 6: (Colour online) The plateau of the non-ergodicity factor
$f^{\text{c}}_{Q}$ oscillates in phase with the static structure factor.
Based on these two evidences, consistent with the results previously obtained
in m-toluidine, we can infer that the predictions of MCT for the existence of
a square-root singularity at $T_{\text{c}}$ and for its $Q$ dependence are
robust enough to be also observed in clustering systems [24, 21].
A peculiar property of salol and m-toluidine which probably makes the cusp
behaviour of the non-ergodicity factor more visible than in other associated
liquids is their high degree of fragility. In fact, a phenomenological
analysis of IXS spectra of glass-forming systems has revealed that the
stronger is a glass former, the higher is the non-ergodicity factor around the
glass transition [25, 26]. The non-ergodicity factor can be also expressed as
the relative longitudinal modulus variation at $T_{\text{g}}$ [26]. For a
strong glass former, like SiO2, the value of $f_{Q}$ at $T_{\text{g}}$ is
close to unity, and thus the amplitude of the square-root cusp is hardly
visible within experimental error. Conversely, for fragile glass-formers such
as m-toluidine and salol, fragility favours the visibility of temperature and
$Q$-dependences of the non-ergodicity parameter, as shown here.
## 4 Conclusions
In conclusion, our findings corroborate the notion that the early stage of the
structural arrest, where the cage effect dominates the molecular dynamics of
simple liquids, shows a universal character — the square-root cusp of the non-
ergodicity factor at a critical temperature $T_{\text{c}}$ — which is also
shared by liquids with a local order [9]. This is a non-trivial result since
it suggests that the structural arrest which occurs in dense, simple liquids,
shares some common, if not universal, features with the structural arrest of
locally ordered liquids, where the coordination number is determined by local
symmetry constraints and where the mere significance of the cage effect is
questionable. To this respect, it is interesting to recall that an IXS
investigation of even more locally-structured glass formers, namely reactive
binary mixtures, has shown the same signatures of the ergodic to non-ergodic
transition predicted by the MCT [15]. In that system, in analogy to what is
found in supercooled simple liquids, the ergodicity breakdown is observed far
before the experimental glass transition, and the static structure evolving
during the covalent bond formation is found to consistently affect the wave
vector dependence of the non-ergodicity factor. It remains a current challenge
to identify the appropriate MCT model that describes the structural arrest in
presence of molecular association or even induced by chemical bonds.
## Acknowledgements
Special thanks to Giancarlo Ruocco, who largely inspired our research on
glasses.
## References
* [1] Gotze W., Sjogren L., Rep. Prog. Phys., 1992, 55, 241–376, doi:10.1088/0034-4885/55/3/001.
* [2] Benassi P., Krisch M., Masciovecchio C., Mazzacurati V., Monaco G., Ruocco G., Sette F., Verbeni R., Phys. Rev. Lett., 1996, 77, 3835–3838, doi:10.1103/PhysRevLett.77.3835.
* [3] Sette F., Krisch M.H., Masciovecchio C., Ruocco G., Monaco G., Science, 1998, 280, 1550–1555, doi:10.1126/science.280.5369.1550.
* [4] Monaco G., Masciovecchio C., Ruocco G., Sette F., Phys. Rev. Lett., 1998, 80, 2161–2164,
doi:10.1103/PhysRevLett.80.2161.
* [5] Fioretto D., Buchenau U., Comez L., Sokolov A., Masciovecchio C., Mermet A., Ruocco G., Sette F., Willner L., Frick B., Richter D., Verdini L., Phys. Rev. E, 1999, 59, 4470–4475, doi:10.1103/PhysRevE.59.4470.
* [6] Frick B., Farago B., Richter D., Phys. Rev. Lett., 1990, 64, 2921–2924, doi:10.1103/PhysRevLett.64.2921.
* [7] Fioretto D., Mattarelli M., Masciovecchio C., Monaco G., Ruocco G., Sette F., Phys. Rev. B, 2002, 65, 224205–224210, doi:10.1103/PhysRevB.65.224205.
* [8] Comez L., Masciovecchio C., Monaco G., Fioretto D., Solid State Phys., 2012, 63, 1–77,
doi:10.1016/B978-0-12-397028-2.00001-1.
* [9] Comez L., Corezzi S., Monaco G., Verbeni R., Fioretto D., Phys. Rev. Lett., 2005, 94, 155702, doi:10.1103/PhysRevLett.94.155702.
* [10] Comez L., Corezzi S., Monaco G., Verbeni R., Fioretto D., J. Non-Cryst. Solids, 2006, 352, 4531–4535, doi:10.1016/j.jnoncrysol.2006.01.164.
* [11] Comez L., Corezzi S., Fioretto D., Kriegs H., Best A., Steffen W., Phys. Rev. E, 2004, 70, 011504, doi:10.1103/PhysRevE.70.011504.
* [12] Comez L., Fioretto D., Kriegs H., Steffen W., Phys. Rev. E, 2002, 66, 032501, doi:10.1103/PhysRevE.66.032501.
* [13] Kalampounias A.G., Kirillov S.A., Steffen W., Yannopoulos S.N., J. Mol. Struct., 2003, 651, 475–483, doi:10.1016/S0022-2860(03)00128-5.
* [14] Kalampounias A., Yannopoulos S.N., Steffen W., Kirillova L.I., Kirillov S.A., J. Chem. Phys., 2003, 118, 8340–8349, doi:10.1063/1.1565325.
* [15] Corezzi S., Comez L., Monaco G., Verbeni R., Fioretto D., Phys. Rev. Lett., 2006, 96, 255702, doi:10.1103/PhysRevLett.96.255702.
* [16] Bencivenga F., Inelastic Light and X-Ray Scattering from salol in the Supercooled, Glassy and Single Crystalline Phases, Universitá degli Studi di Perugia, Tesi di Laurea, 2003.
* [17] Zhang H.P., Brodin A., Barshilia H.C., Shen G.Q., Cummins H.Z., Pick R.M., Phys. Rev. E, 2004, 70, 011502, doi:10.1103/PhysRevE.70.011502.
* [18] Yang Y., Nelson K.A., J. Chem. Phys., 1995, 103, 7732–7739, doi:10.1063/1.470294.
* [19] Masciovecchio C., Monaco G., Ruocco G., Sette F., Cunsolo A., Krisch M., Mermet A., Soltwisch M., Verbeni R., Phys. Rev. Lett., 1998, 80, 544–547, doi:10.1103/PhysRevLett.80.544.
* [20] Hansen C., Stickel F., Richert R., Fischer E.W., J. Chem. Phys., 1998, 108, 6408, doi:10.1063/1.476063.
* [21] Eckstein E., Qian J., Hentschke R., Thurn-Albrecht T., Steffen W., Fischer E.W., J. Chem. Phys., 2000, 113, 4751, doi:10.1063/1.1288907.
* [22] Ruocco G., Sette F., Di Leonardo R., Fioretto D., Krisch M., Lorenzen M., Masciovecchio C., Monaco G., Pignon F., Scopigno T., Phys. Rev. Lett., 1999, 83, 5583–5586, doi:10.1103/PhysRevLett.83.5583.
* [23] Stickel F., Fischer E.W., Richert R., J. Chem. Phys., 1995, 102, 6251, doi:10.1063/1.469071.
* [24] Baran J., Davydova N.A., Phys. Rev. E, 2010, 81, 031503, doi:10.1103/PhysRevE.81.031503.
* [25] Scopigno T., Ruocco G., Sette F., Monaco G., Science, 2003, 302, 849–852, doi:10.1126/science.1089446.
* [26] Buchenau U., Wischnewski A., Phys. Rev. B, 2004, 70, 092201, doi:10.1103/PhysRevB.70.092201.
Ukrainian 0
Íåïðóæíå ðåíòãåíâñüêå ðîçñþâàííÿ ðîçêðèâà ïåðåõä ç åðãîäèíîãî ó íååðãîäèíèé
ñòàí â ñàëîë, ðäèí ç ëîêàëüíèì âïîðÿäêóâàííÿì Ë.
Ãîìåç${}^{\textsf{\footnotesize{\mbox{?}}}}$, Ä.
Ôîðåòòî${}^{\textsf{\footnotesize{\mbox{?}}}}$, É.
Ãàïíñê${}^{\textsf{\footnotesize{\mbox{?}}}}$, Äæ.
Ìîíàêî${}^{\textsf{\footnotesize{\mbox{?}}}}$, A.
Ïàòêîâñê${}^{\textsf{\footnotesize{\mbox{?}}}}$, Â.
Øòåôôåí${}^{\textsf{\footnotesize{\mbox{?}}}}$
${}^{\textsf{\footnotesize 1}}$ IOM-CNR ïðè Ôàêóëüòåò ôçèêè ãåîëîã¿,
Óíâåðñèòåò Ïåðóäæ, 06123 Ïåðóäæà, òàëÿ ${}^{\textsf{\footnotesize 2}}$
Ôàêóëüòåò ôçèêè ãåîëîã¿, Óíâåðñèòåò Ïåðóäæ, 06123 Ïåðóäæà, òàëÿ
${}^{\textsf{\footnotesize 3}}$ Ôàêóëüòåò ôçèêè, Óíâåðñèòåò A. Miöêåâèà,
61-614 Ïîçíàíü, Ïîëüùà ${}^{\textsf{\footnotesize 4}}$ Ôçèíèé ôàêóëüòåò,
Óíâåðñèòåò Òðåíòî, 38123 Ïîâî, òàëÿ ${}^{\textsf{\footnotesize 5}}$ íñòèòóò
Ìàêñà Ïëàíêà äëÿ ïîëìåðíèõ äîñëäæåíü, ï/ñ 3148, 55128 Ìàéíö, Íìåèíà
|
# Graph-Based Analysis and Visualisation of Mobility Data
Rafael Martínez Márquez, Giuseppe Patanè
CNR-IMATI, Italy
###### Abstract
Urban mobility forecast and analysis can be addressed through grid-based and
graph-based models. However, graph-based representations have the advantage of
more realistically depicting the mobility networks and being more robust since
they allow the implementation of Graph Theory machinery, enhancing the
analysis and visualisation of mobility flows. We define two types of mobility
graphs: Region Adjacency graphs and Origin-Destination graphs. Several node
centrality metrics of graphs are applied to identify the most relevant nodes
of the network in terms of graph connectivity. Additionally, the Perron vector
associated with a strongly connected graph is applied to define a circulation
function on the mobility graph. Such node values are visualised in the
geographically embedded graphs, showing clustering patterns within the
network. Since mobility graphs can be directed or undirected, we define
several Graph Laplacian for both cases and show that these matrices and their
spectral properties provide insightful information for network analysis. The
computation of node centrality metrics and Perron-induced circulation
functions for three different geographical regions demonstrate that basic
elements from Graph Theory applied to mobility networks can lead to structure
analysis for graphs of different connectivity, size, and orientation
properties.
Keywords: Network graphs, network structure, spatial analysis, visualisation,
gravity models
###### Contents
1. 1 Introduction
2. 2 Previous work on data-driven traffic models
1. 2.1 Grid-based traffic models
1. 2.1.1 ST-ResNet - Spatio-Temporal Residual Network
2. 2.1.2 STRN - Spatio-Temporal Relation Network
3. 2.1.3 Deep Gravity Mobility Model
2. 2.2 Graph-based traffic models
1. 2.2.1 ETGCN - Evolution Temporal Graph Convolutional Network
2. 2.2.2 GTA - Graph-based Temporal Attention Framework
3. 2.2.3 G-STARIMA - Graph-based Spatio-Temporal ARIMA Model
3. 2.3 Discussion and comparison
3. 3 “Generic” and traffic graphs: definition $\&$ metrics
1. 3.1 “Generic” graphs
2. 3.2 Centrality metrics for graphs
3. 3.3 Traffic graphs and metrics
4. 4 Graph matrices for network and traffic analysis
1. 4.1 Degree matrix
2. 4.2 Transition probability matrix
3. 4.3 Graph Laplacians
1. 4.3.1 Laplacians of undirected graphs
2. 4.3.2 Laplacians of directed graphs
3. 4.3.3 Properties and discussion/comparison
4. 4.4 Graph Laplacians and traffic models
5. 5 Conclusions, data sets and future work
## 1 Introduction
The mobility of individuals is a topic of critical importance for the
development and sustainability of cities. Forecasting the flows of crowds in a
city and foreseeing possible drawbacks in the transport system can boost
public decision-making in risk mitigation and optimising traffic flow, thus
leading to public safety and citizens’ well-being. Traffic flow is affected by
various factors, such as the spatial dependencies between the different
regions of a city, the temporal dependencies from near or distant past time
intervals, and the effect of external factors, such as weather, land and road
features.
The movement of people and vehicles within a city and region is increasingly
influenced by interdependence. In particular, the optimal use of different
transport infrastructures and networks in cities and their connection to
regional destinations and between neighbouring regions has a substantial
impact on various aspects, such as chemical and noise pollution, the
inefficiency of local public transport due to a general increase in traffic
volume, the risk of accidents, and high emergency management costs. Concerning
state-of-the-art in traffic and mobility data-driven modelling, we identify
two main approaches (Sect. 2): Euclidean models that use grid partitions over
a region for forecasting the flows within them (Zhang et al.,, 2017; Liang et
al.,, 2021) or to classify the available data into training and testing sets
(Simini et al.,, 2021), and Non-Euclidean models that use graph-based
structures (Zhang et al., 2021b, ; Zhang et al.,, 2022; Liu et al.,, 2021)
that would be more suitable and realistic for mobility modelling as the
spatial dependency is not necessarily between adjacent cells in a grid-based
partition.
For both Euclidean and non-Euclidean models, visualising the predicted flows
becomes an essential tool for traffic analysis because it provides additional
interpretations to the purely theoretical ones. For instance, visualised
traffic flows can help to localise and identify traffic patterns or areas with
the highest or lowest traffic values (e.g. speed, volume). In addition to the
visualisation of mobility flows and volumes, the node centrality metrics
(i.e., Closeness centrality, betweenness centrality, and Page rank centrality)
(Sect. 3) provide further analysis tools for the structure of the system,
particularly within a graph framework. For instance, the betweenness
centrality can be a measure of the vulnerability of a particular station to
traffic disruption (Dees et al.,, 2021; Mukherjee,, 2012); the closeness
centrality can be used to find the stations that are best connected to the
urban network in terms of minimising the total distance, travelling time, or
other road parameters. The Page rank centrality can identify the most popular
roads or stations regarding the dynamic mobility flow within the urban
network. By analysing both the traffic flow and the structure of the network,
it is also possible to better understand how the network structure affects the
dynamic of the traffic; other factors are operational interventions and
seasonal factors that affect the travel frequency on the network. We introduce
node centrality metrics analysis for two types of graphs: Region Adjacency
graphs and Origin-Destination graphs. The former type of graph is induced by a
given finite partition of a city or region, where the connections represent
the adjacency of a pair of partition units. The latter type of graph is
induced by an Origin-Destination matrix, a square matrix (possibly sparse)
whose entries are positive values representing a mobility counting between a
pair of sub-regions (not necessarily geographically adjacent). We compute,
visualise, and analyse the three centrality metrics for the Region Adjacency
graphs of Genova Province, United Kingdom and New York State. Additionally, we
perform the same visual analysis for the Origin-Destination graph for the
flows predicted by the Deep Gravity mobility model (Simini et al.,, 2021).
Furthermore, through some matrices associated with a graph, e.g. the Laplacian
matrix (Sect. 4), we can investigate demographic factors that affect mobility
flows such as the distribution of population surrounding each node of a
transport network (Dees et al.,, 2021). Also, the Laplacian spectrum is
fundamental for Graph Signal Processing tools that facilitate the analysis of
the correlation of time series of traffic volumes, thus identifying data
correlation structures over different locations and Origin-Destination pairs
in an urban network. The classical graph Laplacian is the Combinatorial
Laplacian, which satisfies important properties as symmetric and positive
semi-definite. However, these features hold only for undirected graphs because
they are based on the symmetry of the adjacency matrix. We introduce the
definitions of the Combinatorial Laplacian and Normalised Laplacian for
undirected graphs. Similarly, the Combinatorial Directed Laplacian,
Symmetrized Laplacian, Combinatorial Symmetrized Laplacian, and Diplacian are
defined for directed graphs. The Symmetrized Laplacian, Combinatorial
Symmetrized Laplacian, and Diplacian are defined only for strongly connected
graphs since they are based on the Perron vector associated with the
transition probability matrix, and their existence and uniqueness are
guaranteed when the graph has only one connected component. These definitions
for the Laplacian matrix grants their characteristic properties to hold
regardless of the graph’s connectivity or if it is directed or undirected.
Overall goals we aim to present a variety of data-driven traffic and mobility
models and develop a set of methods for analysing and visualising information
about the dynamics of mobility flows. The input data used for such models is
mainly consisting of trajectories of cars, bicycles, or people (Zhang et al.,,
2017; Liang et al.,, 2021), sensors and monitoring stations (Zhang et al.,
2021b, ; Zhang et al.,, 2022; Liu et al.,, 2021), and demographical data
(Simini et al.,, 2021). The primary research and development activities
include
* •
an overview of Grid-based and Graph-based data-driven traffic models;
* •
the visualisation of centrality metric values in mobility graphs to display an
overview of the network structure to more easily identify mobility flows of
interest;
* •
the definition and properties of distinct Laplacians matrices for undirected
and directed graphs.
These tools are essential to support autonomous decision-making for traffic
management in areas of interest, e.g., to propose suggested routes to
commuters entering a city to reduce traffic in certain areas based on traffic
forecasts. We draw several conclusions, indicate possible extents of this work
(Sect. 5), and describe the used data sets.
## 2 Previous work on data-driven traffic models
The traffic forecasting problem can be formulated as follows: given a
historical collection $\mathcal{X}$ of features, e.g., average speed, road
volume, and demand levels, during $P$ different time intervals, predict a
collection $\mathcal{Y}$ of features (possibly different) for the next $Q$
time intervals. This relation is expressed by $\mathcal{Y}=f(\mathcal{X})$,
where $f$ is an unknown function. Several data-driven traffic models depend on
the available data and the predicted variable. Also, these models can be
distinguished based on how they structure the data (e.g., in a regular or
irregular domain). Some grid-based approaches are presented where the data is
partitioned into squared cells, and the predicted value is the number of flows
within their inner sub-partitions. Similarly, we introduce some graph-based
models where data partition depends on the distribution of the road network or
data collectors (e.g., sensors, monitoring stations, road junctions). These
graph-based models can predict various traffic states (e.g., speed, flow,
occupancy). If the chosen model is grid-based, the city or area of interest is
partitioned into an $I\times J$ grid with $N=IJ$ cells. If the chosen model is
graph-based, then the road network over a city or area of interest is
represented by $N$ nodes and a set of edges connecting them. For both grid-
based and graph-based models, different techniques from Deep Learning can be
used, such as convolution networks and residual units (Zhang et al.,, 2017;
Liang et al.,, 2021), or graph attention mechanisms and spectral convolution
networks (Zhang et al., 2021a, ; Li et al.,, 2021).
### 2.1 Grid-based traffic models
Several flow forecasting models use different types of mobility data and
prediction techniques. Euclidean approaches for flow prediction in traffic
modelling consist of partitioning a city, state, or country into a square or
rectangular grid to assess the number of displacements between two elements of
the grid (Zhang et al.,, 2017; Liang et al.,, 2021), or dividing available
geographical data into training, and testing data for the use of a Deep
Learning model (Simini et al.,, 2021). When partitioning the city into an
$I\times J$ grid, urban flows, traffic volumes, and other mobility values,
data can be stored in matrices in $\mathbb{R}^{I\times J}$, or tensors in
$\mathbb{R}^{L\times I\times J}$ if the data consists of more than one type of
value, where $L$ is the number of different values that compose the data. The
forecasting problem can be stated as given historical observations of mobility
data $\\{\mathbf{X}_{1},\mathbf{X}_{2},\ldots,\mathbf{X}_{N}\\}$ during $N$
time intervals, predict $\mathbf{X}_{N+1}$. External factors may vary over
time and influence mobility in a city. Indeed, the crowd flows can be
represented by tensors in $\mathbb{R}^{K\times I\times J}$, where $K$ is the
number of external influences being considered. Suppose the $I\times J$ grid
is only used to classify the available mobility data. In that case, each of
the cells should be, in addition, sub-partitioned (possibly in irregular
units) to forecast the flows from one to another within the same cell.
#### 2.1.1 ST-ResNet - Spatio-Temporal Residual Network
In the ST-ResNet model (Spatio-temporal residual network) (Zhang et al.,,
2017), the subdivision of the city into regions of interest is performed by a
squared grid partition using the longitudes and latitudes. The predicted
traffic variables are the inflows and outflows among the cells of the squared
grid. Inflow is the total traffic crowds entering a cell in the grid from any
other cell during a given time interval. Outflow denotes the total traffic of
crowds leaving a cell to any other cell in the grid during a given time
interval. The input data for the ST-ResNet model consists of trajectories of
cars, bicycles, people, etc., during different constant length time intervals
and meteorological data as external factors that impact the traffic flow in a
city. Formally, the city, state, or country is split into an $I\times J$ mesh
with $N=IJ$ cells, and we can associate the inflow and outflow of crowds to
every cell $(i,j)$ at the time interval $t$.
The ST-ResNet model applies convolution-based residual networks for dealing
with the local and global spatial dependencies that affect the flows within a
city. There are three such networks, one for each type of temporal dependency:
closeness, period, and trend. Moreover, the ST-ResNet model uses a fully
connected network to include the effect of external factors that may impact
traffic flows, such as weather conditions and holidays. All these model
elements are dynamically aggregated to produce predicted inflow and outflow
matrices with the exact dimensions of the squared grid partition of the city.
#### 2.1.2 STRN - Spatio-Temporal Relation Network
The STRN model (Spatio-Temporal Relation Network) (Liang et al.,, 2021) is
another grid-based model for flow prediction. The subdivision of the city into
a squared grid partition consists of a considerably larger number of cells $N$
(of smaller size) than with the ST-ResNet model; this type of partition is
called fine-grained or high resolution. Moreover, a secondary irregular
partition of the city into $M$ regions, with $M$ considerably smaller than
$N$, is used to assess the effects of global spatial dependencies more
efficiently, namely, by analysing the diffusion within the connectivity
network that represents the irregular partition, which is computationally
faster than using a sequence of convolution layers as in the ST-ResNet model.
The input data for the STRN model consists of trajectories as in the ST-ResNet
model.
An assignment matrix $\mathbf{B}\in\mathbb{R}^{N\times M}$ is used to connect
the $N$ cells grid structure with the $M$ region’s network, where the entry
$b_{ij}$ denotes the likelihood that the cell $i$ belongs to the region $j$.
The irregular partition can be computationally generated by additional
available information, such as the road network, the administrative divisions
of a city, or census areas. The STRN model applies a Meta Learner that
converts the features related to external factors to a representation with the
same dimensionality as the Inflows and Outflows from the three temporal
dependencies, e.g., closeness, period, and trend. Then, these four feature
representations are passed to a Backbone Network consisting of squeeze-and-
excitation networks, and convolution layers, whose output is finally passed to
the Global Relation Module (GloNet), which converts the grid-based
representations to the network structure and applies Graph Convolution
Networks to assess the diffusion of the flows within the irregular regions.
Then, the Meta Learner converts the predictions to the original grid-based
structure.
#### 2.1.3 Deep Gravity Mobility Model
The Deep Gravity mobility model (Simini et al.,, 2021) is a grid-based model
for flow prediction without using the three-time dependencies paradigm (i.e.,
closeness, period, trend), which is fundamental in the ST-ResNet and STRN
models. To use the Deep Gravity model, partitioning a city, state, or country
starts with constructing a square grid covering the whole region. Additionally
to this handmade regular partition, it must be available a secondary finer
(and irregular) partition for which there is available population data, for
instance, the tracts or areas that constitute the units of a census. Then, the
barycentre of each irregular unit is allocated in one of the square cells.
Hence, the data of a cell consists of the data of the irregular units whose
barycentre belongs to that cell. The square cells are divided into training
and testing cells in a stratified manner based on the population data in the
cells so that the two groups have the same number of cells belonging to the
various population deciles. Newton’s law of universal gravitation inspires
this method, which is why its name is inspired. A generalised version of this
attraction law, by considering the deterrence function, is defined as
$y(l_{i},l_{j})=O_{i}\dfrac{m_{j}^{\beta_{1}}f(r_{ij})}{\sum_{k}m_{k}^{\beta_{1}}f(r_{ik})},$
(1)
where $y(l_{i},l_{j})$ represents the predicted flow from a location $l_{i}$
to a location $l_{j}$, and $r_{ij}$ is the distance between them; the $m_{i}$
are the populations, and $\beta_{1}$ is a real parameter. The relation in Eq.
(1) is called a singly-constrained gravity model since it requires knowing the
total outflow $O_{i}$ for each location in advance. The Deep Gravity model
applies feed-forward neural networks to predict the number of flows from a
given area from the irregular partition to any other of the areas within the
same cell. The choice of this type of network lies in the fact that feed-
forward neural networks generalise linear models as in Eq. (1). Unlike the ST-
ResNet and STRN models, the Deep Gravity model does not use the grid structure
to predict flows between them but to split the city into training and testing
regions.
The city’s subdivision into a squared grid partition splits the available data
sets into training and testing data since the predicted flows correspond to
the flows between additional irregular partitions within each cell provided by
census areas. The input data for the Deep Gravity model consists of population
values obtained from the official census and geographic features obtained from
OpenStreetMap. The intuition of this model is that the flow between two
locations is directly proportional to their population and inversely
proportional to the distance between them.
Given two irregular units $l_{i}$ and $l_{j}$ in a cell $C$, the components of
the input vector $x_{ij}$ used to predict the flow $y(l_{i},l_{j})$ from
$l_{i}$ to $l_{j}$ consist of the population in $l_{i}$ and $l_{j}$, the
distance between them, and eighteen geographical values extracted from
OpenStreetMap (land use areas, road lengths, counting of points of interest,
etc.). The input vector $x_{ij}$ passes through a multilayer feed-forward
neural network with an output layer of dimension $1$ (a scalar $y_{ij}$) in
the range $(-\infty,\infty)$. After computing $y_{ij}$ for $j=1,\ldots,M$,
where $M$ is the number of irregular units in the cell $C$, the $M$ values are
passed through a softmax layer to convert them into an $M$-dimensional vector
with non-negative components whose sum is $1$, namely, a probability vector
that represents the probability distribution of the flows starting in location
$l_{i}$ within the cell $C$. When training the model, the predicted values
$y_{ij}$ are compared to real flows $z_{ij}$ from location $l_{i}$ to location
$l_{j}$, for instance, provided by GPS trajectory data or commuting surveys. A
standard metric used to measure the performance of a flow prediction model is
the Common part of Commuters (CPC) defined by
$CPC=\dfrac{2\displaystyle\sum_{i,j}\min(y_{ij},z_{ij})}{\displaystyle\sum_{i,j}y_{ij}+\sum_{i,j}z_{ij}},$
where $y_{ij},z_{ij}$ refer to the predicted and real flow from location
$l_{i}$ to location $l_{j}$ respectively, and the indices $i,j$ run along all
the locations in a region of interest, for instance in one of the square cells
used in the Deep Gravity model. The CPC values are in the range
$\left[0,1\right]$ with $1$ indicating a perfect flow prediction. Suppose the
total number of outflows in a region of interest coincides with the predicted
flows (as with the Deep Gravity model used by (Simini et al.,, 2021)). In that
case, the CPC value is equal to the fraction of the flows correctly predicted
by the model.
### 2.2 Graph-based traffic models
Graph-based approaches have been used for metro traffic modelling (Dees et
al.,, 2021) and statistical analysis of road networks (Mukherjee,, 2012).
Graph theory provides additional theoretical elements, namely the weight of
the edges, which can be used to represent some features in traffic models,
such as the road length and road capacity (Tian et al., 2016a, ). Moreover,
the centrality properties of a graph can provide additional insight into the
distribution and importance of nodes and edges (Henry et al.,, 2019;
Boulmakoul et al.,, 2017).
There are several approaches to creating a graph structure for traffic
modelling depending on the available data type and the forecasted variable. In
a traffic graph (Jiang and Luo,, 2022), a graph signal
$\mathbf{X}_{t}\in\mathbb{R}^{N\times d}$ is defined at every time step $t$
where $N$ is the number of nodes and $d$ is the number of traffic elements
(e.g. speed, traffic state, traffic demand) that are measured. The majority of
graph-based spatio-temporal traffic problems belong to traffic state and
traffic demand prediction (Ye et al.,, 2022), which are modelled by various
graph-based deep learning architectures (e.g. Spectral Graph Convolution
(SGCN), Gated Recurrent Unit (GRU), Graph Attention Network (GTA)). The nodes
of a traffic graph can be defined as the intersections between roads or as the
sensors along a highway. The physical road connections between them usually
give the edges, and the predicted variables can be any traffic element for
each node, namely, a graph signal.
#### 2.2.1 ETGCN - Evolution Temporal Graph Convolutional Network
The Evolution Temporal Graph Convolutional Network (ETGCN) (Zhang et al.,
2021b, ) captures the spatial and temporal correlations among the nodes in a
Road graph. The road graph is a weighted graph whose nodes are traffic sensors
distributed over a city or region. The concept of adjacency matrix involves
the fusion of three types of information (Content Similarity Adjacency Matrix,
Graph Betweenness Adjacency Matrix, and Transportation Neighborhood Adjacency
Matrix) whose entries consider the geographical position of the sensors and
their connections in the original road network, and which fusion enhances
feature learning. Given a historical time series of the registered speed at
every node, the ETGCN model aims to predict the speed at every node in the
next timestep. The architecture of the ETGCN model combines Graph
Convolutional Networks and GRU to learn the sequence of spatial and temporal
features.
#### 2.2.2 GTA - Graph-based Temporal Attention Framework
The Graph-based Temporal Attention Framework (GTA) (Zhang et al.,, 2022) model
also considers a weighted sensor graph, but in contrast to the ETGCN model,
the weights of the edges are the road network distances between the sensors,
and the predicted variable is the traffic flow, which is defined as the number
of vehicles passing through the monitoring station over a given time interval.
The architecture of the GTA model introduces an attention mechanism to
adaptively identify the relations among three temporal dependencies: monthly
pattern, weekly pattern, and current pattern. A Long Short-Term Memory (LSTM)
network is employed to extract the temporal correlation for each dependency.
#### 2.2.3 G-STARIMA - Graph-based Spatio-Temporal ARIMA Model
The Graph-based spatio-temporal ARIMA (G-STARIMA) model (Liu et al.,, 2021) is
a graph-based framework built upon statistical methods. The road network is
described by a weighted undirected graph where the nodes are the traffic
intersections, and the edges are their road connections. The G-STARIMA model
uses a historical time series of graph signals at the network nodes (i.e.,
sensor observations such as speed, occupancy, or traffic flow) to predict the
traffic state at the next instant. Since the connectivity between traffic
intersections, represented by edge weights, may vary over time because of the
dynamics of urban traffic states, it is performed a dynamic estimation of a
weighted adjacency matrix which is found by solving a convex optimisation
problem.
### 2.3 Discussion and comparison
The ST-ResNet, STRN, and Deep Gravity models are grid-based models for flow
predictions within the units of a city, state, or country partition. The ST-
Resnet and the STRN models consider the CPT Paradigm (closeness, period, and
trend time dependencies) and predict the flow values as the next value of a
sequence of flows. In contrast, the Deep Gravity model only predicts a value
for the flows from a set of given features. Regarding the network
architectures that build the models, the Deep Gravity model has a simpler
architecture since feed-forward neural networks are the elementary Deep
Learning network. In contrast, the convolution layers and residual units of
the ST-ResNet and STRN models are more refined networks. Moreover, even if the
three models are based on a grid structure, they use the square cells
differently: the ST-ResNet and the STRN models use the cells as the units
where the flows will be predicted from historical values of the same flows. In
contrast, the Deep Gravity model uses the cells only to split the available
data into training and testing data.
Regarding the spatial dependencies, the Deep Gravity model performs flow
prediction only for irregular units within the same cell. Therefore, it does
not consider global spatial dependencies as the ST-ResNet and STRN models. The
three show significant improvements concerning other existing baseline models
for flow prediction. Still, it is necessary to make a performance comparison
between the Deep Gravity model and the other two grid-based models.
The grid-based models show interesting results in traffic forecasting.
However, some features could suggest modelling mobility in a city as a graph:
the structure of the roads as graph topology that connects different
locations, i.e., graph nodes, in the city, or the information associated with
movement between two locations as graph weights. The urban road network is a
typical spatial network because of its geographical factors (Tian et al.,
2016b, ; Ye et al.,, 2022), and usually, adjacent cells of the grid do not
have flows between them during some intervals of time because there are no
direct roads connecting them.
The ETGCN and GTA models are data-driven graph-based models for traffic
predictions that consider sensors or monitoring stations as the nodes of a
weighted graph. The GTA model uses the distance between the sensors as edge
weights and predicts the number of vehicles passing through every monitoring
station. In contrast, the ETGCN model considers a fusion of different types of
adjacency matrices that enhance the learning of features and predict the speed
at every network node. Another characteristic of the GTA model is the
implementation of three-time dependencies, which is analogous to the CPT
Paradigm of the ST-ResNet and STRN grid-based models. Regarding the Deep
Learning architecture of these graph-based models, the ETGCN uses GRU networks
to learn temporal features. In contrast, the GTA model combines LSTM networks
through an attention-based mechanism. In contrast to the ETGCN and the GTA
models, the G-STARIMA model is not a Deep Learning model but a graph-based
framework based on statistical methods. In addition, this model considers the
road intersections as network nodes instead of the monitoring stations and
predicts various traffic states (e.g., speed, traffic flow).
Both grid-based and graph-based traffic models show interesting results, and
the choice of a modelling framework should be based on the type and amount of
available data and the variables that will be predicted. However, graph-based
models have additional theoretical tools to boost traffic analysis, i.e. the
centrality metrics (Sect. 3.2). For instance, the betweenness centrality was
used in (Dees et al.,, 2021) to analyse the robustness of the metro network in
London since it facilitates the identification of metro stops that might
significantly impact the whole network during a disruption. Similarly, the
betweenness metric was used in (Mukherjee,, 2012) to identify potential
congestion points in the Indian highway network. In Ye et al., (2022), various
approaches are proposed to construct a graph-based framework, different
mobility data sets, and Deep Learning networks for graph-based models.
## 3 “Generic” and traffic graphs: definition $\&$ metrics
The graph-based approaches for traffic and mobility modelling certainly build
upon the fundamental concepts from Graph Theory. Basic definitions such as the
direction and weight of an edge enable the construction of different types of
graphs. This leads to the implementation and interpretation of diverse multi-
agent systems modelling, where the connection between the agents (the nodes)
stores significant information. The sequences of consecutive edges motivate
the definitions of paths, which induces the concept of connected components.
Furthermore, all these structures combined enable the introduction of several
quantitative scores for each node according to its importance in the
connectivity and influence to the rest of the network, which is characterised
by the centrality metrics. In traffic and mobility modelling, the centrality
metrics provide valuable information for the structure analysis of the road
and geographical networks.
### 3.1 “Generic” graphs
To represent a graph $\mathcal{G}$ with nodes set $V=\\{1,\ldots,N\\}$ and
edges set $E$, we use the notation $\mathcal{G}=(V,E)$. Undirected graphs
represent only connections between nodes, so $(u,v)=(v,u)$ for every $u,v\in
V$. A graph is a directed graph (or digraph) if the couples that define the
edges in $E$ cannot be arbitrarily ordered, namely, if in general
$(u,v)\neq(v,u)$. The positive value $w_{uv}$ is the weight of the edge
$(u,v)$, and $E$ is a weighted graph. If no values are associated with a
graph’s edges, it is called unweighted. The weights of the edges of a graph
with $N$ nodes can be stored in an $N\times N$ matrix called the adjacency
matrix of $\mathcal{G}$.
We can associate positive weights to every edge of a graph with $N$ nodes and
store them in the weighted adjacency matrix $\mathbf{A}\in\mathbb{R}^{N\times
N}$, namely
$A_{ij}=\left\\{\begin{array}[]{cl}w_{ij}&\text{if }(i,j)\in E;\\\
0&\text{otherwise}.\end{array}\right.$
If $\mathcal{G}$ is an unweighted graph; then we have binary weights, i.e.,
$1$ for existing connections between nodes, and $0$ otherwise. For undirected
graphs, since $(u,v)$ and $(v,u)$ refer to the same edge, then the weight
values satisfy $\omega_{uv}=\omega_{vu}$, the corresponding adjacency matrix
is symmetric. Knowing the adjacency matrix of a graph, we know the weights of
the edges and the connections between the nodes. Indeed, we write
$\mathcal{G}=(V,E,\mathbf{A})$ to make more explicit the relationship between
the graph representation of $\mathcal{G}$ and the matrix representation given
by the adjacency matrix $\mathbf{A}$.
##### Connected and strongly connected graphs
An undirected graph is connected if, for every pair of nodes $u,v\in V$ there
exists a path $w$ from $u$ to $v$, otherwise $\mathcal{G}$ is called
disconnected. A directed graph $\mathcal{G}$ is strongly connected if for any
pair $(p_{i},p_{j})$ of $\mathcal{G}$ there exists a path (i.e., a sequence of
edges) $(p_{i},p_{l_{1}})$, $(p_{l_{2}},p_{l_{3}})$, $\ldots$,
$(p_{l_{r-1}},p_{j})$, which leads from $p_{i}$ to $p_{j}$. In this case, the
path has length $r$. The associated directed graph $\mathcal{G}(\mathbf{A})$
of a $n\times n$ matrix $\mathbf{A}$ consists of $n$ nodes
$p_{1},\ldots,p_{n}$, where an edge leads from $p_{i}$ to $p_{j}$ if and only
if $A_{ij}\neq 0$. As main properties, we recall that a matrix $\mathbf{A}$ is
irreducible if and only if the associated directed graph
$\mathcal{G}(\mathbf{A})$ is strongly connected. Otherwise, the matrix is
called _irreducible_. A matrix $\mathbf{A}\in\mathbb{R}^{n\times n}$ is said
to be _reducible_ if there exists a perturbation matrix $\mathbf{P}$ such that
$\mathbf{C}=\mathbf{P}\mathbf{A}\mathbf{P}^{\top}=\left[\begin{array}[]{cc}\mathbf{A}_{11}&\mathbf{A}_{12}\\\
\mathbf{0}&\mathbf{A}_{22}\end{array},\right]$
$\mathbf{A}_{11}\in\mathbb{R}^{n\times r}$,
$\mathbf{A}_{12}\in\mathbb{R}^{(n-r)\times(n-r)}$, and
$\mathbf{A}_{22}\in\mathbb{R}^{n\times(n-r)}$. Let $\mathbf{A}$ be an
irreducible matrix. Then, (i) $\mathbf{A}$ has a positive real eigenvalue
equal to its spectral radius $\rho(\mathbf{A})$; (ii) $\rho(\mathbf{A})$ is a
single eigenvalue of $\mathbf{A}$ and its corresponding eigenvector $\phi>0$;
(iii) $\rho(\mathbf{A})$ increases when any entry of $\mathbf{A}$ increases;
and (iv) there is no other non-negative eigenvector of $\mathbf{A}$ different
from $\phi$. Finally, a digraph is called weakly connected if its underlying
undirected graph is connected. A disconnected graph can be decomposed into
smaller subgraphs connected as an independent graph, called connected
components.
##### Graph paths
A path $w$ in a graph (or a walk) (Chung,, 2005) of length $m$ is a sequence
of different $m$ nodes in $V$, namely, $w=\\{v_{1},\ldots,v_{m}\\}$ such that
the couple $(v_{i},v_{i+1})$ belongs to the edges set $E$ for every
$i=1,\ldots,m-1$. Given the nodes $a$ and $b$ of a graph (undirected or
directed), then $w=(v_{1},\ldots,v_{m})$ is a shortest path from $a$ to $b$ if
$v_{1}=a,v_{m}=b$ and the value $d_{w}(a,b):=\sum_{i=1}^{m-1}A_{v_{i}v_{i+1}}$
attains its minimum among all the paths starting in $a$ and ending in $b$.
Since there may be more than one shortest path from a node $a$ to a node $b$,
we define the distance from $a$ to $b$ as $d(a,b)=\min\\{d_{w}(a,b)\colon
w\text{ is a path from }a\text{ to }b\\}$, which is a unique value. If for a
couple of nodes $a,b\in V$ of a disconnected graph, there is no path starting
in $a$ and ending in $b$ then we define their distance as $d(a,b)=\infty$. The
distance between two nodes is not precisely a distance in the context of
metric spaces when the graph is directed because $d(a,b)$ is not necessarily
the same as $d(b,a)$.
For instance, the Region Adjacency graphs (Sect. 3.3) are undirected graphs
that are highly sparse since they represent the geometric partition of a
geographical region where most of the units have just a few neighbours.
However, there can be some areas where the sub-regions are clustered, which
produces square sub-matrices along the diagonal of the sparsity matrix (Fig.
1).
Undirected Region Adjacency graphs
---
| |
Region Adjacency graph
| |
Sparsity matrix
a) GOA Province | b) UK | c) NY State
Figure 1: Region Adjacency graphs and sparse adjacency matrices. a) Genova
Province (GOA Province) partitioned into 137 zones b) United Kingdom (UK)
partitioned into 344 local authority districts c) New York State (NY State)
partitioned into 5410 census tracts.
### 3.2 Centrality metrics for graphs
Different centrality metrics can be defined for the nodes in a directed graph
$\mathcal{G}=(V,E)$ with $N$ nodes. The main centrality metrics are degree
centrality, closeness centrality, harmonic centrality, betweenness centrality,
and Page rank centrality. The closeness and harmonic centralities measure the
likelihood of a node reaching the rest of the nodes, which can be extended to
the concept of being reached if the graph is directed. The betweenness
centrality measures the importance of a node to create connections between the
rest of the nodes, and it is computed in the same way for undirected and
directed graphs. Finally, PageRank centrality is a metric specifically for
directed graphs since the influence of a node depends on the influence of the
incoming connections.
| |
---|---|---
(a) GOA Province | (b) UK | (c) NY
Figure 2: Normalised node centrality metrics. The normalised betweenness
centrality is the smallest for the three Region Adjacency graphs regardless of
the number of nodes.
##### Degree centrality:
The out-degree centrality and in-degree centrality of a node $v$ in a directed
graph are defined by
$D^{+}(v)=\sum_{j=1}^{N}A_{vj},\qquad D^{-}(v)=\sum_{j=1}^{N}A_{jv},$
respectively. This centrality metric represents how big that node’s outgoing
or incoming flow is compared to the corresponding flow in the rest of the
nodes.
##### Closeness centrality:
The out-closeness centrality and in-closeness centrality of a node $v$ are
defined by
$C^{+}(v)=\dfrac{N-1}{\displaystyle\sum_{i\neq v}d(v,i)},\qquad
C^{-}(v)=\dfrac{N-1}{\displaystyle\sum_{i\neq v}d(i,v)}.$
The two types of closeness centrality defined on a directed graph measure the
likelihood of a node reaching the other nodes or being reached by them. For an
undirected graph, the closeness centrality measures the likelihood of a node
being connected to the rest of the nodes in a graph. If for some nodes $v,b$
there is no path from $v$ to $b$, then we would have $C(v)=0$. The bigger the
distances from $v\in V$ to the rest of the nodes are, the smaller the value of
the closeness centrality $C(v)$ is, the nodes that are closer to the rest of
the graph have higher closeness centrality values. The closeness centrality
for a directed graph may have more nodes with zero values, compared to the
undirected case, since more cases of not symmetric paths arise, for instance,
in a not strongly connected graph.
##### Harmonic centrality:
The out-harmonic centrality and in-harmonic centrality of a node $v$ are
defined by
$H^{+}(v)=\dfrac{1}{N-1}\sum_{i\neq v}\frac{1}{d(v,i)},\qquad
H^{-}(v)=\dfrac{1}{{N-1}}\sum_{i\neq v}\frac{1}{d(i,v)}.$
The harmonic centrality in a directed graph reduces zero closeness centrality
values when the graph is not strongly connected. It simply excludes the not-
reachable nodes to provide positive values for most nodes. A node will have a
zero out-harmonic centrality value if and only if its out-degree is zero. A
node will have a zero in-harmonic centrality value if and only if its in-
degree is zero. From the inequality between the arithmetic mean and the
harmonic mean of real values, it follows $H^{+}(v)\geq C^{+}(v)$ and
$H^{-}(v)\geq C^{-}(v)$. For any undirected graph $\mathcal{G}$ and any node
$v\in V$, from the inequality between the arithmetic mean and the harmonic
mean of real values it follows $H(v)\geq C(v)$ and the harmonic centrality is
just a variation of the closeness centrality to handle the infinite values
caused by disconnected graphs since the existence of a distance
$d(v,b)=\infty$ would vanish the value of $C(v)$ discarding the possibility of
further analysis for the node $v$.
##### Betweenness centrality:
Similarly to the undirected case, the betweenness centrality of a node $v$ is
defined by
$B(v)=\underset{(a,b)\in\mathcal{P}_{v}}{\sum}\dfrac{|S_{v}(a,b)|}{|S(a,b)|},$
where $S(a,b)$ is the set of shortest paths from $a$ to $b$,
$S_{v}(a,b)=\\{w\in S(a,b)\colon v\in w\\}$ is the set of shortest paths from
$a$ to $b$ with $v$ as an intermediate node, $\mathcal{P}_{v}=\\{(a,b)\in
V\times V\colon a,b\neq v,a\neq b,S(a,b)\neq\varnothing\\}\\}$ is the set of
node couples that have $v$ as part of a shortest path, and $v\in
w=(v_{1},\ldots,v_{m})$ means $v=v_{i}$ for some $i=2,\ldots,m-1$. The
notation $|X|$ represents the cardinality of a set $X$, in other words, the
number of elements of $X$. The betweenness centrality values for the nodes in
a directed graph are generally smaller than the values in the undirected case
because $S(a,b)\neq\varnothing$ does not imply that $S(b,a)\neq\varnothing$
and therefore there might be fewer elements in the sum.
For directed graphs, the betweenness centrality estimates the importance of a
node $v$ as a connection between the rest of the graph. Namely, it measures
the proportion of shortest paths between any other couple of nodes passing
through $v$. The higher the betweenness centrality of a node is, the more
significant the proportion of shortest paths between other nodes passing
through it. A node with a high betweenness centrality value can be interpreted
as a node whose removal would considerably affect the graph since some nodes
may remain without an optimal path between them, and some nodes may even
become disconnected. If no short paths between any couple of nodes pass
through a given (and different) node, then its betweenness centrality is zero.
If for every $a,b\in V$ it holds $S(a,b)=\varnothing$, then the graph consists
of isolated nodes, and the betweenness centrality values can be set to zero
for each node.
##### PageRank centrality:
This centrality metric is particularly used with directed graphs; it measures
the influence of a node $v\in V$ depending on the influence of every other
node $u\in V$ such that $(u,v)\in E$ (incoming edges to $V$). Intuitively, if
a node has incoming edges from a relevant node, its influence in the graph
would be bigger than if those edges were from a less relevant node. The
PageRank centrality of a node $v$ originally defined in (Page et al.,, 1999),
is defined in a recursive way through
$PR(v)=(1-c)+c\left(\dfrac{PR(v_{1})}{D^{+}(v_{1})}+\ldots\dfrac{PR(v_{m})}{D^{+}(v_{m})}\right),$
where the $v_{i}$ are the nodes in $\mathcal{G}$ such that there is an edge
from $v_{i}$ to $v$, for $i=1,\ldots,m$. Moreover, $c\in(0,1)$ is a value
called the damping factor that helps to deal with dead-end nodes (without
outgoing edges). Usually, it is set to $c=0.85$. The PageRank values can be
seen as the principal eigenvector of a matrix $\mathbf{\hat{P}}^{\top}$ given
by $\mathbf{\hat{P}}=c(\mathbf{P}+\delta\cdot b^{\top})+(1-c)\mathbf{E}$,
where $\mathbf{P}$ is the transition probability matrix $\mathbf{P}$ (Sect.
4.2), $\mathbf{b}\in\mathbb{R}^{N}$ is a distribution probability called
teleportation vector, whose component $i$ denotes the probability to move
arbitrarily to the node $i$ while moving along the edges of the graph (with
direction), and $\mathbf{E}:=(1,\ldots,1)\mathbf{b}^{\top}$ (Berkhin,, 2005)).
The vector $\delta\in\mathbb{R}^{N}$ has components
$\delta_{i}:=\delta(d_{i},0)$ (Kronecker delta) for $i=1,\ldots,N$, where
$d_{i}$ is the degree of the node $i$.
Table 1: Summary of centrality metrics for the nodes of a graph Metric and definition | Undirected graphs | Directed graphs
---|---|---
| Degree centrality
---
$D(v)=\sum_{j=1}^{N}A_{vj}$
Yes | Out-degree and in-degree variations
| Closeness centrality
---
$C(v)=\dfrac{N-1}{\displaystyle\sum_{i\neq v}d(v,i)}.$
Yes | Outgoing and incoming variations
| Harmonic centrality
---
$H(v)=\dfrac{\displaystyle\sum_{i\neq v}\frac{1}{d(v,i)}}{N-1}$
Yes | Outgoing and incoming variations
| Betweenness centrality
---
$B(v)=\underset{(a,b)\in\mathcal{P}_{v}}{\sum}\dfrac{|S_{v}(a,b)|}{|S(a,b)|}$
Yes | Yes
| PageRank centrality
---
$PR(v)=(1-c)+c\displaystyle\sum_{u\rightarrow v}\dfrac{PR(u)}{d(u)}$
No | Yes
For instance, after normalising each centrality metric for the three Region
Adjacency graphs from Fig. 1, a trend in common is shown for all of them (Fig.
2), i.e. the closeness centrality has invariably the largest values. In
contrast, the betweenness centrality has the smallest ones because the maximum
value of the betweenness centrality is larger than the maximum value of the
other two centralities. Moreover, for the NY State graph, the majority of
betweenness centrality values are localised in the first decile of its
distribution as the number of nodes in the NY State graph is considerably
larger than in the other two graphs, which results in larger betweenness
centralities because every node is part of the shortest path between more node
pairs.
##### Flow and circulation of a directed graph
In a graph $\mathcal{G}$, we consider a function
$F:\mathcal{E}(\mathcal{G})\rightarrow\mathbb{R}^{+}$ that assigns to each
directed edge $(i,j)$ a non-negative value $F_{ij}$, which is said to be a
_circulation_ if at each node $i$ we have that $\sum_{j,\,j\rightarrow
i}F_{ji}=\sum_{j,\,i\rightarrow j}F_{ij}$. A circulation is said to be
_invertible_ if $F_{ij}=F_{ji}$. For a strongly connected directed graph
$\mathcal{G}$, the eigenvector $\phi$ of the transition probability matrix
$\mathbf{P}$ with eigenvalue $1$ is associated with a circulation
$F^{\phi}_{ij}:=\phi_{i}P_{ij}.$ (2)
The average node circulation of each node is defined by
$\tilde{F^{\phi}_{i}}=\sum_{j}F^{\phi}_{ij}/D^{+}(i)$. For instance, as the
Region-Adjacency graphs are connected (after possibly removing the ”islands”,
i.e. nodes with node degree equal to $1$), it is possible to define a
circulation associated to their Perron vector by using Eq. 2. The trend of the
average node circulation (Fig. 3) has an inverse behaviour with respect to the
closeness centrality values in Fig. 5. Namely, the nodes with higher average
circulation have the lowest closeness centrality values, i.e., the nodes with
less accessibility to the rest of the network. This inverse behaviour is more
evident for the GOA Province from the central subregions to the eastern ones.
The inverse behaviour is clearer for the UK from the central subregions to the
southern ones. Since the number of nodes of the NY State Region Adjacency
graph is considerably larger than the other two graphs, the average node
circulation values variation is also larger, so its visualisation does not
require a quartile partition.
Undirected Region Adjacency graphs
---
| |
(a) GOA Province | (b) UK | (c) NY State
Figure 3: Average node circulation values for Region Adjacency graphs. The
values for a) GOA Province and b) UK correspond to the quartiles partition to
enhance the visualisation.
### 3.3 Traffic graphs and metrics
##### Region Adjacency and Origin-Destination graphs
We now introduce two types of traffic analysis and simulation graphs: the
Region Adjacency graph and the Origin-Destination graph. Given a finite non-
empty compact partition $\mathcal{P}$ of a region $K\subset\mathbb{R}^{2}$,
i.e. a collection $\\{R_{1},\ldots,R_{N}\\}$ of non-empty compact subsets of
$\mathbb{R}^{2}$ such that $\bigcup_{i}R_{i}=K$, we can identify each element
$R_{i}$ of $\mathcal{P}$ with its centroid
$c_{i}=(x_{i},y_{i})\in\mathbb{R}^{2}$. The Region Adjacency graph associated
to $\mathcal{P}$ consists of the undirected unweighted graph with $N$ nodes
embedded in $V=\\{c_{1},\ldots,c_{N}\\}$, and edges
$E=\\{(c_{i},c_{j})\colon\partial R_{i}\cap\partial R_{j}\neq\varnothing\\}$,
where $\partial X$ represents the boundary of the set $X$.
Given an Origin-Destination matrix $M\in\mathbb{R}^{N\times N}$, the Origin-
Destination graph associated to $M$ consists of the directed weighted graph
with weighted adjacency matrix $M$. The $N$ nodes can be embedded in some
$\mathbb{R}^{2}$ representation associated to $M$. For instance, we can
construct an Origin-Destination graph for the New York State using the
predicted flows generated by the Deep Gravity mobility model (Sect. 2.1.3).
The entry $M_{ij}$ of the weighted adjacency matrix $M$ is the predicted flow
from the location indexed by $i$ to the location indexed by $j$. The Origin-
Destination graph $\mathcal{G}$ associated to $M$ (Fig. 4) consists of $2836$
nodes and $939888$ edges, and it is not connected since the Deep Gravity model
predicts flows only for the locations in the same regular partition, i.e. the
same square cell.
The adjacency matrices for the Region Adjacency graphs are highly sparse (Fig.
1). In contrast, the Origin-Destination graph shows more clusters with a more
significant number of connections since the Deep Gravity mobility model
predicts flows for some regions that are not adjacent geographically (Fig.
4a), thus its adjacency matrix (Fig. 4b) has many more square sub-matrices
that are related to the number of different strongly connected components
since the Deep Gravity model predicts flows only for the irregular units
within the same squared cell.
Directed Origin-Destination graph
---
|
(a) NY State Origin-Destination graph | (b) Sparsity matrix
Figure 4: Origin-Destination digraph associated to the predicted OD matrix for
NY State using the Deep Gravity mobility model.
##### Metrics on the Region Adjacency graph
The closeness centrality values of the nodes in a Region Adjacency graph can
have a clustered behaviour showing the subregions that have higher levels of
accessibility to the rest of the subregions in a given planar partition. For
instance, there is a trend of lower levels of accessibility from the
peripherical subregions. In contrast, the central subregions have higher
closeness centrality values (Fig. 5), except for NY State, which has a
significant higher values cluster in the Southeast area, which could be due to
its geometrical shape and the number of subregions in that area. The
betweenness centrality values in the same Region Adjacency graph are, in
general, low for most of the network (Fig. 5), so there are no clusters of
subregions with higher connectivity importance that serve to link other
subregions through the shortest path. Nevertheless, in the NY State Region
Adjacency graph, there are a few nodes with the highest betweenness centrality
values around the ”bottleneck” that connects the Southeast area to the rest of
the network, which means that their removal may have an impact on the shortest
paths between the subregions in the Southeast area and the ones in the rest of
the NY State. Indeed, this difference between the maximum betweenness
centrality values and the values in the rest of the nodes are the ones that
cause the low trend for the normalised betweenness centrality (Fig. 2).
Undirected Region Adjacency graphs
---
| |
Closeness centrality
| |
Betweenness centrality
(a) GOA Province | (b) UK | (c) NY State
Figure 5: Closeness and Betweenness centralities for Region Adjacency graphs.
The closeness values show significant changes depending on their geographical
position, while the betweenness values are nearly constant.
##### Metrics on Origin-Destination graphs
As with the Region Adjacency graphs, the centrality metrics values of the
nodes in an Origin-Destination graph can also reveal some information
regarding the connectivity of the subregions in a given planar partition, and
their flows represented in an Origin-Destination matrix. However, it is
possible that the available flows are not sufficient to generate a strongly
connected graph. Indeed, the Deep Gravity mobility model predicts flows only
for the irregular subregions within the same squared cell, creating as many
strongly connected components as cells exist. The higher values of the
outcloseness and betweenness centralities of the Origin-Destination graph for
New York State (Fig. 6) are localised approximately in the same subregions as
the nodes with higher closeness centrality values in the Region Adjacency
graph of New York State (Fig. 5). This trend occurs because the Origin-
Destination graph of New York State has significantly more nodes in this area,
creating a strongly connected component with a larger number of nodes and
consequently with more accessible nodes and connections, which increases
outcloseness and betweenness centralities, respectively. The Page rank
centrality exhibits a more randomised trend because several strongly connected
components have various nodes. The original values were classified in
quartiles to visualise their distribution better.
Directed Origin-Destination graph
---
| |
(a) Outcloseness centrality | (b) Betweenness centrality | (c) Page rank centrality
Figure 6: Quartiles visualisation of Centrality metrics for the NY State
Origin-Destination graph. The edges are not shown for enhancing the
visualisation of node centrality values.
## 4 Graph matrices for network and traffic analysis
The node degree centrality motivates the definition of the degree matrix,
which associates a value to each node based on the number of its neighbours.
The transition probability matrix represents the connection’s relevance for
each node’s neighbours, as its entries are values that measure the proportion
of information flow across the edges of the graph. Furthermore, the graph
Laplacian allows us to study structure properties, such as connectedness
through eigenvalue problems, and there are different types of Laplacians on a
graph $\mathcal{G}$, depending on the model represented by $\mathcal{G}$, and
also whether the graph is undirected or directed. For connected graphs and
strongly connected graphs, the transition probability matrix has an associated
Perron vector that is used to define several Laplacian matrices and a
circulation function that provides another sense of information flow between
any couple of nodes of the graph.
### 4.1 Degree matrix
If $\mathcal{G}$ is a graph with $N$ nodes (weighted or unweighted, directed
or undirected), then the degree of $v\in V$ is defined as the sum of the
elements of the $v$-th row of the adjacency matrix. Namely, the degree $d_{v}$
of a node $v\in V$ is defined by $d_{v}=\sum_{j=1}^{N}A_{vj}$. Indeed, the
degree of a node is the sum of the weights of the edges joining that node for
undirected graphs and the sum of the weights of the edges outgoing from that
node for digraphs. The degree matrix $\mathbf{D}$ is the diagonal matrix
$\mathbf{D}\in\mathbb{R}^{N\times N}$ given by $D_{ij}=d_{i}$ if $i=j$ and $0$
otherwise.
### 4.2 Transition probability matrix
Let $\mathcal{G}$ be a weighted graph and let $i\in V$ be a node with
$d_{i}>0$ outgoing edges (arriving to the nodes $v_{1},\ldots,v_{d_{i}}$). The
proportion of the flow at node $i$ that will continue through node $k$ is
given by $p_{ik}=A_{ik}/d_{i}$, for $k=v_{1},\ldots,v_{d_{i}}$. On the other
hand, if $d_{i}=0$ then the proportion of outgoing flows from $i$ to $k$ is
simply zero for every $k\in V$. This discussion motivates the definition of
the transition probability matrix $\mathbf{P}$, which defines a Markov chain
associated with random walks on the graph $\mathcal{G}$ (Li and Zhang,, 2012),
where the entry $P_{ij}$ denotes the probability of moving from node $i$ to
node $j$, for $i,j=1,\ldots,N$. However, it is required that there are no
nodes without any edge for undirected graphs or nodes without outgoing edges
(called dead-end nodes, possibly with incoming edges, for directed graphs).
The reason is that if $i\in V$ is a node with such conditions. The proportion
of the flow going from $i$ to $j$ will be zero for any node $j\in V$;
therefore, the whole $i$-th row would be zero. Still, a transition matrix of a
Markov chain must satisfy having the sum of each column equal to $1$ (row
stochastic).
For the sake of theoretical completeness in Markov chains (Li and Zhang,,
2012), we assume that a graph is strongly connected to define its transition
probability matrix. The transition probability matrix $\mathbf{P}$ of the
graph $\mathcal{G}$ is the $N\times N$ matrix that satisfies the relation
$\mathbf{P}=\mathbf{D}^{-1}\mathbf{A}$, where $\mathbf{A}$ and $\mathbf{D}$
are the adjacency and degree matrices, respectively. The entries $p_{ij}$
represent the probability of the flow moving from node $i$ to node $j$ (or the
proportion of flows that moves from $i$ to $j$), for $i,j=1,\ldots,N$. The
matrix $\mathbf{P}$ is sometimes called the normalised adjacency matrix of
$\mathcal{G}$ (Veerman and Lyons,, 2020). From the definition of the degree
matrix $\mathbf{D}$ it follows that
$\sum_{j=1}^{N}P_{ij}=\sum_{j=1}^{N}(\mathbf{D}^{-1}\mathbf{A})_{ij}=\sum_{j=1}^{N}\sum_{k=1}^{N}D^{-1}_{ik}A_{kj}=\sum_{j=1}^{N}\dfrac{1}{d_{i}}A_{ij}=\dfrac{1}{d_{i}}\sum_{j=1}^{N}A_{ij}=1,$
for every $i=1,\ldots,N$, in other words, the sum of all the entries of a row
equals $1$, for every row in $\mathbf{P}$. Indeed, the matrix $\mathbf{P}$ is
row stochastic, and there is a Markov chain associated with random walks on
the graph $\mathcal{G}$ defined by $\mathbf{P}$.
##### Perron vector
For a strongly connected directed graph, the transition probability matrix
$\mathbf{P}$ has a unique left eigenvector $\phi$ with positive components,
namely, $\phi^{\top}\mathbf{P}=\rho\phi^{\top}$. The vector $\phi$ is called
the Perron vector of $\mathbf{P}$, and in fact, it can be easily proven that
$\rho=1$. We notice that $\phi$ is the Perron vector of the transition
probability matrix $\mathbf{P}:=\mathbf{D}^{-1}\mathbf{A}$ if and only if
$\widetilde{\phi}:=\mathbf{D}^{-1}\phi$ is the generalised eigenvector of the
couple $(\mathbf{A},\mathbf{D})$ associated with the eigenvalue $1$, i.e.,
$\mathbf{A}\widetilde{\phi}=\mathbf{D}\widetilde{\phi}$. Indeed, we solve the
generalised eigenproblem and then compute
$\phi:=\mathbf{D}^{-1}\widetilde{\phi}$. Since the eigenvector associated with
the eigenvalue $1$ is unique (as the graph is strongly connected), its entries
are all positive or negative; if negative, we change their sign to guarantee
that the entries of $\phi$ are all positive. Since
$\mathbf{P}\mathbf{1}=\mathbf{1}$, we get that $\rho=1$ is an eigenvalue of
$\mathbf{P}$ and all the eigenvalues of $\mathbf{P}$ are lower than $1$. In
particular, we normalise the entries of $\phi$ such that
$\sum_{i=1}^{n}\phi(i)=1$.
### 4.3 Graph Laplacians
An essential property of the graph Laplacians of an undirected graph is the
symmetry, which follows from the symmetry of the adjacency graph and
guarantees that all its eigenvalues are real numbers. The Combinatorial
Laplacian and the Normalised Laplacian are two possible definitions when the
graph is undirected, being the latter a transformation of the former that
results in an upper bound for its real eigenvalues. For a directed graph, it
is possible to define the Combinatorial Directed Laplacian, Symmetrized
Laplacian, and the Combinatorial Symmetrized Laplacian, which have real
eigenvalues even if the adjacency matrix is not symmetric. The Diplacian is
another viable definition that involves obtaining complex eigenvalues, which,
however, coincides with the Normalised Laplacian when the graph is undirected.
#### 4.3.1 Laplacians of undirected graphs
For an undirected graph $\mathcal{G}$ with $N$ nodes the Combinatorial
Laplacian (Veerman and Lyons,, 2020), is the $N\times N$ matrix defined by
$\mathbf{L}=\mathbf{D}-\mathbf{A}.$ (3)
Since the adjacency matrix of an undirected graph is symmetric, the
Combinatorial Laplacian is also symmetric. Moreover, $\mathbf{D}$ is positive
definite, and it is possible to define a scalar product in $\mathbb{R}^{N}$ by
letting
$\langle\mathbf{x},\mathbf{y}\rangle_{\mathbf{D}}:=\mathbf{x}^{\top}\mathbf{Dy}$,
for $\mathbf{x},\mathbf{y}\in\mathbb{R}^{N}$. The Combinatorial Laplacian is
also called the Kirchhoff matrix of the graph (Caughman and Veerman,, 2006).
(Zhang et al., 2021a, ) defines the normalised graph Laplacian
$\mathbf{\hat{L}}$ by
$\mathbf{\hat{L}}=\mathbf{I}-\mathbf{D}^{-1/2}\mathbf{A}\mathbf{D}^{-1/2}$,
where $\mathbf{I}$ is the $N\times N$ identity matrix, being $N$ the number of
nodes in the graph. The normalised graph Laplacian of an undirected graph is
also symmetric, and it is also positive semidefinite since $\mathbf{D}$ is an
invertible diagonal matrix with positive entries. Furthermore,
$\mathbf{\hat{L}}=\mathbf{D}^{-1/2}\mathbf{LD}^{-1/2}$.
#### 4.3.2 Laplacians of directed graphs
Since the Combinatorial Laplacian of a directed graph is not necessarily
symmetric, (Hasanzadeh et al.,, 2017) defines the Combinatorial Directed
Laplacian of a digraph $\mathcal{G}=(V,E,\mathbf{A})$ by
$\mathbf{L_{G}}=\dfrac{1}{2}(\mathbf{D}_{out}+\mathbf{D}_{in}-\mathbf{A}-\mathbf{A}^{\top}),\quad\mathbf{D}_{out}=\sum_{j=1}^{N}A_{ij},\quad\mathbf{D}_{in}=\sum_{j=1}^{N}A_{ji},$
where $\mathbf{D}_{out},\mathbf{D}_{in}$ are the out-degree and in-degree and
$\mathbf{A}$ is the adjacency matrix. The Combinatorial Directed Laplacian is
symmetric regardless of $\mathcal{G}$ being directed or not. Moreover, if
$\mathcal{G}$ is undirected then $\mathbf{L}=\mathbf{L_{G}}$. The
Combinatorial Directed Laplacian is also positive semi-definite since it can
be seen as the Combinatorial Laplacian of an undirected with adjacency matrix
$\tilde{\mathbf{A}}=(\mathbf{A}+\mathbf{A}^{\top})/2$. Assuming that the input
graph is strongly connected, the unique stationary probability distribution
$\mathbf{\phi}$ of the transition probability matrix $\mathbf{P}$ is defined
as the unique vector $\phi\in\mathbb{R}^{N}$ with strictly positive components
such that $\phi^{\top}\mathbf{P}=\phi^{\top}$. In (Chung,, 2005), the
Symmetrized Laplacian of a strongly connected graph $\mathcal{G}$ is
$\mathbf{\mathcal{L}}=\mathbf{I}-\dfrac{\mathbf{\Phi}^{1/2}\mathbf{P}\mathbf{\Phi}^{-1/2}+\mathbf{\Phi}^{-1/2}\mathbf{P}^{\top}\mathbf{\Phi}^{1/2}}{2},$
where $\mathbf{\Phi}=\textrm{diag}(\phi_{i})$. The Symmetrised Laplacian is
indeed symmetric. (Chung,, 2005) also defines the Combinatorial Symmetrized
Laplacian by
$\mathbf{\mathcal{L}_{G}}=\mathbf{\Phi}-\dfrac{\mathbf{\Phi
P}+\mathbf{P}^{\top}\mathbf{\Phi}}{2},$
which is symmetric and positive semi-definite since it can be written as the
Combinatorial Laplacian of an undirected graph
$\tilde{G}=(V,\tilde{E},\tilde{A})$ with adjacency matrix
$\tilde{\mathbf{A}}=(\mathbf{\Phi P}+\mathbf{P^{\top}\Phi})/2$. The
Combinatorial Symmetrised Laplacian coincides with the Combinatorial Laplacian
defined in (3) when the graph is undirected. However, $\mathbf{\mathcal{L}}$
does not capture the unique characteristic of random walks on digraphs, since
different directed graphs can have the same $\mathbf{\mathcal{L}}$. To
overcome this problem, (Li and Zhang,, 2012) defines the Diplacian
$\mathbf{\Gamma}$ though
$\mathbf{\Gamma}=\mathbf{\Phi}^{1/2}(\mathbf{I}-\mathbf{P})\mathbf{\Phi}^{-1/2},$
for which the strongly connected assumption for the graph $\mathcal{G}$ still
holds since the stationary probabilities are required.
#### 4.3.3 Properties and discussion/comparison
The Laplacian matrix associated with a graph is an essential operator for
network models because it constitutes the foundation for Deep Learning
techniques on graph structures. The symmetry of the Laplacian is a desired
feature even if the graph is directed. Consequently, it is necessary to define
more Laplacian matrices in addition to the classic one in Eq. (3). Some
matrices, such as the Combinatorial Directed Laplacian, the Symmetrised
Laplacian, and the Combinatorial Symmetrised Laplacian, are always symmetric
independently of whether the graph is directed or undirected. Moreover, these
Laplacians are also positive semi-definite operators because they could be
seen as the Combinatorial Laplacian of an undirected graph
$\tilde{G}=(V,\tilde{E})$ with an appropriate choice for the adjacency matrix
$\tilde{\mathbf{A}}$, so the positive semi-definite property of $\mathbf{L}$
is inherited to the new matrices.
Table 2: Summary of the properties of different Laplacians Graph Laplacian | Real Eigenvalues | Positive Semidefinite
---|---|---
Undirected graph
| Combinatorial Laplacian
---
$\mathbf{L}=\mathbf{D}-\mathbf{A}$
Yes | Yes
| Normalised Laplacian
---
$\hat{\mathbf{L}}=\mathbf{I}-\mathbf{D}^{-1/2}\mathbf{AD}^{-1/2}$
Yes | Yes
Directed graph
| Combinatorial Directed Laplacian
---
$\mathbf{L}_{G}=\frac{1}{2}(\mathbf{D}_{out}+\mathbf{D}_{in}-\mathbf{A}-\mathbf{A}^{\top})$
Yes | Yes
| Symmetrized Laplacian
---
$\mathbf{\mathcal{L}}=\mathbf{I}-\dfrac{\mathbf{\Phi}^{1/2}\mathbf{P\Phi}^{-1/2}+\mathbf{\Phi}^{-1/2}\mathbf{P}^{\top}\mathbf{\Phi}^{1/2}}{2}$
Yes | Yes
| Combinatorial Symmetrized Laplacian
---
$\mathbf{\mathcal{L}}_{G}=\mathbf{\Phi}-\dfrac{\mathbf{\Phi}\mathbf{P}+\mathbf{P}^{\top}\mathbf{\Phi}}{2}$
Yes | Yes
| Diplacian
---
$\mathbf{\Gamma}=\mathbf{\Phi}^{1/2}(\mathbf{I}-\mathbf{P})\mathbf{\Phi}^{-1/2}$
No | -
The Combinatorial Laplacian $\mathbf{L}$ of an undirected graph is symmetric
since $\mathbf{D}$ and $\mathbf{A}$ are symmetric. $\mathbf{L}$ is also self-
adjoint since $\mathbf{L}$ is symmetric, in fact, if
$\mathbf{x},\mathbf{y}\in\mathbb{R}^{N}$ then
$\langle\mathbf{x},\mathbf{Ly}\rangle=\langle\mathbf{Lx},\mathbf{y}\rangle\Leftrightarrow\mathbf{x}^{\top}\mathbf{Ly}=(\mathbf{Lx})^{\top}\mathbf{y}=\mathbf{x}^{\top}\mathbf{L}^{\top}\mathbf{y}$.
Additionally, since $\mathbf{L}$ is positive semidefinite, then the Normalised
Laplacian $\hat{\mathbf{L}}$ of an undirected graph is also positive
semidefinite because it can be written as
$\hat{\mathbf{L}}=\mathbf{D}^{-1/2}\mathbf{LD}^{-1/2}$. Rewriting a Laplacian
matrix of a directed graph $\mathcal{G}=(V,E,\mathbf{A})$ as
$\tilde{\mathbf{D}}-\tilde{\mathbf{A}}$, namely, a diagonal matrix
$\tilde{\mathbf{D}}$ whose $i$-th element equals the sum of the $i$-th row of
a non-negative matrix $\tilde{\mathbf{A}}$, then the positive semidefinite
property follows from considering the Laplacian as the one corresponding to an
undirected graph. For instance, for the Combinatorial Directed Laplacian
$\mathbf{L}_{G}$ we can set
$\tilde{\mathbf{A}}=(\mathbf{A}+\mathbf{A}^{\top})/2$ and
$\tilde{\mathbf{D}}=(\mathbf{D}_{out}+\mathbf{D}_{in})/2$, which satisfies
that the sum of the $i$-th row equals $\tilde{D}_{ii}$. Similarly, for the
Combinatorial Symmetrised Laplacian $\mathcal{L}_{G}$ we have that
$\tilde{\mathbf{A}}=(\mathbf{\Phi}\mathbf{P}+\mathbf{P}^{\top}\mathbf{\Phi})/2$
and $\tilde{\mathbf{D}}=\mathbf{\Phi}$. The sum of the $i$-th row of
$\tilde{\mathbf{A}}$ equals $\Phi_{ii}=\phi_{i}$, in fact,
$2\sum_{j=1}^{N}\tilde{A}_{ij}=\sum_{j=1}^{N}(\Phi\mathbf{P})_{ij}+\sum_{j=1}^{N}(\mathbf{P}^{\top}\Phi)_{ij}=\sum_{j=1}^{N}\phi_{i}P_{ij}+\sum_{j=1}^{N}P_{ji}\phi_{j}=2\phi_{i},$
since $\mathbf{P}$ is row stochastic, and by definition the Perron vector
$\mathbf{\phi}$ satisfies the relation
$\mathbf{\phi}^{\top}=\mathbf{\phi}^{\top}\mathbf{P}$. The positive
semidefinite property for the Symmetrised Laplacian follows from the relation
$\mathcal{L}=\mathbf{\Phi}^{-1/2}\mathcal{L}_{G}\mathbf{\Phi}^{-1/2}$. The
graph Laplacians for directed graphs, except for the Diplacian, are generally
symmetric. Consequently, all their eigenvalues are real. Also, they are
bounded for the Symmetrised and the Combinatorial Symmetrised Laplacians (Fig.
7).
Undirected Region Adjacency graphs
---
| |
(a) GOA Province | (b) UK | (c) NY State
Figure 7: Laplacian eigenvalues of Region Adjacency Graphs. The eigenvalues of
the Combinatorial Laplacian are not bounded. The largest eigenvalue (spectral
radius) increases as the number of nodes increases in the Region Adjacency
graph. In contrast, the eigenvalues of the Normalised Laplacian are bounded
regardless of the number of nodes in the graph.
The Diplacian $\mathbf{\Gamma}$ of a directed graph is symmetric if and only
if its adjacency matrix is symmetric. Indeed, we have
$\mathbf{\Gamma}=\mathbf{I}-\mathbf{\Phi}^{1/2}\mathbf{P\Phi}^{-1/2}$.
Assuming $\mathbf{\Gamma}=\mathbf{\Gamma}^{\top}$, then
$\mathbf{\Phi}^{1/2}\mathbf{P\Phi}^{-1/2}=\mathbf{\Phi}^{-1/2}\mathbf{P}^{\top}\mathbf{\Phi}^{1/2}$.
Recalling that $\mathbf{P}=\mathbf{D}^{-1}\mathbf{A}$, $\mathbf{D}$ and
$\mathbf{\Phi}$ are diagonal matrices, using their multiplication
commutativity, and multiplying by $\mathbf{\Phi}^{1/2}$ from the left and from
the right, we obtain
$\mathbf{\Phi}\mathbf{D}^{-1}\mathbf{A}=\mathbf{A}^{\top}\mathbf{D}^{-1}\mathbf{\Phi}=\mathbf{\Phi}\mathbf{D}^{-1}\mathbf{A}^{\top}$,
Since $\mathbf{D}^{-1}$ and $\mathbf{\Phi}$ have an inverse, we get
$\mathbf{A}=\mathbf{A}^{\top}$ as a necessary condition for the symmetry of
the Diplacian. Conversely, if $\mathbf{A}=\mathbf{A}^{\top}$ then
$\mathbf{\Gamma}$ is symmetric. If $\mathbf{A}$ is symmetric, then the
Diplacian reduces to
$\mathbf{I}-\mathbf{P}=\mathbf{I}-\mathbf{D}^{-1/2}\mathbf{AD}^{-1/2}=\hat{\mathbf{L}}$,
i.e. the Normalised Laplacian.
### 4.4 Graph Laplacians and traffic models
The Graph Laplacian can be used to estimate the population of the sub-regions
in a given area for which mobility or traffic flows are known. In (Dees et
al.,, 2021) it is proposed a way to estimate the population distribution
$\phi$ surrounding the stations of the London underground network, by using
Fick’s law of diffusion $\mathbf{q}=-k\nabla\mathbf{\phi}$ which states that
the flux $\mathbf{q}$ flows from regions of high concentration to regions of
low concentration, with a magnitude proportional to the concentration gradient
$\nabla\mathbf{\phi}$, and coefficient of diffusivity $k$. Because of the
graph structure and the discrete nature of the stations, an algebraic
manipulation of the Fick’s law gives the relation
$\mathbf{q}=-k\mathbf{L\phi}$, where $\mathbf{L}$ is the combinatorial
Laplacian of the network. The flux $\mathbf{q}$ was estimated using the
average daily flow of passengers obtained from Transport for London data.
Using the pseudo-inverse matrix $\mathbf{L}^{+}$, the estimation of the
population surrounding each station is obtained through
$\hat{\mathbf{\phi}}=-(1/k)\mathbf{L}^{+}\mathbf{q}$.
## 5 Conclusions, data sets and future work
The data for defining the traffic features include sensors, GPS, rail-hailing,
and transaction data sets. Sensors can be implemented on roads and highways
and collect traffic measurements, such as speed. GPS trajectories are
generated by taxis or rented vehicles during a period. Rail-haling records
include some transport services demand, such as rented bicycles or mobility as
a service option. Transaction data sets are generated by automatic systems,
such as ticket machines in metro stations, which count the number of departing
or arriving passengers.
Graph-based structures allow the analysis of the interplay between the
different elements of the system, as well as the identification of the most
relevant or vulnerable nodes within the whole network. This type of method is
even applicable for some grid-based structures in traffic and mobility
modelling as long as Origin-Destination matrices are available, as in the case
of the Deep Gravity model. Furthermore, node centrality metrics are an
instrument of structure analysis of the network even if there are no available
mobility flows, as in the case of the Region Adjacency graphs. Similarly, the
Laplacian matrix can provide additional reasoning about external information,
e.g., the distribution of the population surrounding the nodes of an urban
network. For connected and strongly connected graphs, the existence and
uniqueness of the Perron vector associated with the transition probability
matrix enables the definition of several Laplacian matrices and constructs a
circulation function in the graph, giving more quantitative information for
analysis and visualisation.
In future work, we plan to address the monitoring of mobility data to
reconstruct traffic flows in urban, regional and inter-regional contexts
through the analysis of origin-destination data, i.e., of mobility
trajectories, considering heterogeneous, partial and uncertain data, and
integrating them with meteorological and pollution data. In particular, it
will be possible to analyse the analysis of short-term mobility patterns to
classify the mobility habits of users, e.g., assessing whether a particular
route/access is regular or sporadic. Furthermore, an implementation of a
graph-based Deep Gravity mobility model with a strongly connected structure is
desired to implement the circulations induced by the Perron vector of the
transition probability matrix, as well as the exploration of alternatives for
the circulation function when the graph is not connected or strongly
connected, for instance, based in node centrality metrics.
Data availability. The Python code, the Region Adjacency data and the
predicted flows using the Deep Gravity mobility model (Simini et al.,, 2021)
for New York State are freely available at github.com/scikit-
mobility/DeepGravity. The Region Adjacency data for the United Kingdom and
Genova Province are freely available at
https://census.ukdataservice.ac.uk/use-data/guides/boundary-data, and
https://smart.comune.genova.it/opendata, respectively.
Declarations of interest. None.
Funding. Rafael Martínez Márquez has been supported by a REACT-EU PhD fellow.
## References
* Berkhin, (2005) Berkhin, P. (2005). A Survey on PageRank Computing. Internet Mathematics, 2(1):73 – 120.
* Boulmakoul et al., (2017) Boulmakoul, B., Besri, Z., Karim, L., Boulmakoul, A., and Lbath, A. (2017). Combinatorial connectivity and spectral graph analytics for urban public transportation system. Transportation Research Procedia, 27:1154–1162. 20th EURO Working Group on Transportation Meeting, EWGT 2017, 4-6 September 2017, Budapest, Hungary.
* Caughman and Veerman, (2006) Caughman, J. and Veerman, J. (2006). Kernels of directed graph laplacians. The electronic journal of combinatorics, 13.
* Chung, (2005) Chung, F. R. K. (2005). Laplacians and the cheeger inequality for directed graphs. Annals of Combinatorics, 9:1–19.
* Dees et al., (2021) Dees, B. S., Xu, Y. L., Constantinides, A. G., and Mandic, D. P. (2021). Graph theory for metro traffic modelling. 2021 International Joint Conference on Neural Networks, pages 1–5.
* Hasanzadeh et al., (2017) Hasanzadeh, A., Liu, X., Duffield, N. G., Narayanan, K. R., and Chigoy, B. T. (2017). A graph signal processing approach for real-time traffic prediction in transportation networks. arXiv: Signal Processing.
* Henry et al., (2019) Henry, E., Bonnetain, L., Furno, A., Faouzi, N.-E. E., and Zimeo, E. (2019). Spatio-temporal correlations of betweenness centrality and traffic metrics. In 2019 6th International Conference on Models and Technologies for Intelligent Transportation Systems, pages 1–10.
* Jiang and Luo, (2022) Jiang, W. and Luo, J. (2022). Graph neural network for traffic forecasting: A survey. Expert Syst. Appl., 207:117921.
* Li et al., (2021) Li, G., Knoop, V. L., and van Lint, H. (2021). Multistep traffic forecasting by dynamic graph convolution: Interpretations of real-time spatial correlations. Transportation Research Part C: Emerging Technologies.
* Li and Zhang, (2012) Li, Y. and Zhang, Z.-L. (2012). Digraph laplacian and the degree of asymmetry. Internet Mathematics, 8(4):381–401.
* Liang et al., (2021) Liang, Y., Ouyang, K., Sun, J., Wang, Y., Zhang, J., Zheng, Y., Rosenblum, D. S., and Zimmermann, R. (2021). Fine-grained urban flow prediction. Proceedings of the Web Conference 2021.
* Liu et al., (2021) Liu, T., Jiang, A., Miao, X., Tang, Y., Zhu, Y., and Kwan, H. K. (2021). Graph-based dynamic modeling and traffic prediction of urban road network. IEEE Sensors Journal, 21(24):28118–28130.
* Mukherjee, (2012) Mukherjee, S. (2012). Statistical analysis of the road network of india. Pramana, 79(3):483–491.
* Page et al., (1999) Page, L., Brin, S., Motwani, R., and Winograd, T. (1999). The pagerank citation ranking: Bringing order to the web. Technical Report 1999-66, Stanford InfoLab.
* Simini et al., (2021) Simini, F., Barlacchi, G., Luca, M., and Pappalardo, L. (2021). A deep gravity model for mobility flows generation. Nature communications, 12.
* (16) Tian, Z., Jia, L., Dong, H., Su, F., and Zhang, Z. (2016a). Analysis of urban road traffic network based on complex network. Procedia Engineering, 137:537–546. Green Intelligent Transportation System and Safety.
* (17) Tian, Z., Jia, L., Dong, H., Su, F., and Zhang, Z. (2016b). Analysis of urban road traffic network based on complex network. Procedia Engineering, 137(C):537–546.
* Veerman and Lyons, (2020) Veerman, J. J. P. and Lyons, R. (2020). A primer on Laplacian dynamics in directed graphs. arXiv:2002.02605.
* Ye et al., (2022) Ye, J., Zhao, J., Ye, K., and Xu, C. (2022). How to build a graph-based deep learning architecture in traffic domain: A survey. IEEE Transactions on Intelligent Transportation Systems, 23(5):3904–3924.
* Zhang et al., (2017) Zhang, J., Zheng, Y., and Qi, D. (2017). Deep spatio-temporal residual networks for citywide crowd flows prediction. Proceedings of the AAAI Conference on Artificial Intelligence, 31(1).
* Zhang et al., (2022) Zhang, S., Guo, Y., Zhao, P., Zheng, C., and Chen, X. (2022). A graph-based temporal attention framework for multi-sensor traffic flow forecasting. IEEE Transactions on Intelligent Transportation Systems, 23(7):7743–7758.
* (22) Zhang, S., Zheng, H., Su, H., Yan, B., Liu, J., and Yang, S. (2021a). Gacan: Graph attention-convolution-attention networks for traffic forecasting based on multi-granularity time series. 2021 International Joint Conference on Neural Networks, pages 1–8.
* (23) Zhang, Z., Li, Y., Song, H., and Dong, H. (2021b). Multiple dynamic graph based traffic speed prediction method. Neurocomputing, 461:109–117.
|
# Pólya’s conjecture for thin products
Xiang He, Zuoqin Wang School of Mathematical Sciences
University of Science and Technology of China
Hefei, 230026
P.R. China
<EMAIL_ADDRESS>School of Mathematical Sciences
University of Science and Technology of China
Hefei, 230026
P.R. China
<EMAIL_ADDRESS>
###### Abstract.
Let $\Omega\subset\mathbb{R}^{d}$ be a bounded Euclidean domain. According to
the famous Weyl law, both its Dirichlet eigenvalue $\lambda_{k}(\Omega)$ and
its Neumann eigenvalue $\mu_{k}(\Omega)$ have the same leading asymptotics
$w_{k}(\Omega)=C(d,\Omega)k^{2/d}$ as $k\to\infty$. G. Pólya conjectured in
1954 that each Dirichlet eigenvalue $\lambda_{k}(\Omega)$ is greater than
$w_{k}(\Omega)$, while each Neumann eigenvalue $\mu_{k}(\Omega)$ is no more
than $w_{k}(\Omega)$. In this paper we prove Pólya’s conjecture for thin
products, i.e. domains of the form $(a\Omega_{1})\times\Omega_{2}$, where
$\Omega_{1},\Omega_{2}$ are Euclidean domains, and $a$ is small enough. We
also prove that the same inequalities hold if $\Omega_{2}$ is replaced by a
Riemannian manifold, and thus get Pólya’s conjecture for a class of “thin”
Riemannian manifolds with boundary.
Partially supported by National Key R and D Program of China 2020YFA0713100,
and by NSFC no. 12171446.
## 1\. Introduction
Let $\Omega\subset\mathbb{R}^{d}$ be a bounded domain. Then the Dirichlet
Laplacian on $\Omega$ has discrete spectrum which forms an increasing sequence
of positive numbers (each with finite multiplicity) that tend to infinity,
$0<\lambda_{1}(\Omega)\leq\lambda_{2}(\Omega)\leq\lambda_{3}(\Omega)\leq\cdots\nearrow+\infty,$
and the Neumann Laplacian on $\Omega$ has a similar discrete spectrum (under
suitable boundary regularity assumptions, which we always assume below without
further mentioning)
$0=\mu_{0}(\Omega)\leq\mu_{1}(\Omega)\leq\mu_{2}(\Omega)\leq\cdots\nearrow+\infty.$
Moreover, by a simple variational argument one has
$\mu_{k-1}(\Omega)<\lambda_{k}(\Omega)$ for all $k$, which was strengthened to
(1.1) $\mu_{k}(\Omega)<\lambda_{k}(\Omega),\quad\forall k$
by L. Friedlander in [15] (See also N. Filonov [6]), answering a conjecture of
L. E. Payne [27].
Starting from H. Weyl ([35]), the asymptotic behavior of the eigenvalues
$\lambda_{k}(\Omega)$ and $\mu_{k}(\Omega)$ as $k\to\infty$ has attracted a
lot of attention. In fact, both $\lambda_{k}(\Omega)$ and $\mu_{k}(\Omega)$
admit the same leading term asymptotics
$\lambda_{k}(\Omega)\sim\frac{4\pi^{2}}{(\omega_{d}|\Omega|)^{\frac{2}{d}}}k^{\frac{2}{d}}\quad\text{and}\quad\mu_{k}(\Omega)\sim\frac{4\pi^{2}}{(\omega_{d}|\Omega|)^{\frac{2}{d}}}k^{\frac{2}{d}},$
where $|\Omega|$ represents the volume of $\Omega$, and $\omega_{d}$ is the
volume of the unit ball in $\mathbb{R}^{d}$.
In his classical book [29], G. Pólya conjectured (in a slightly weaker form
for the Neumann case) that for each $k$, the $k^{\mathrm{th}}$ Dirichlet
eigenvalue
(1.2)
$\lambda_{k}(\Omega)\geq\frac{4\pi^{2}}{(\omega_{d}|\Omega|)^{\frac{2}{d}}}k^{\frac{2}{d}}$
while the $k^{\mathrm{th}}$ positive Neumann eigenvalue
(1.3)
$\mu_{k}(\Omega)\leq\frac{4\pi^{2}}{(\omega_{d}|\Omega|)^{\frac{2}{d}}}k^{\frac{2}{d}}.$
As observed by G. Pólya, these conjectured inequalities hold for all
rectangles. As for arbitrary domain, the conjecture holds for $k=1$ (the
Faber-Krahn inequality ([5], [19]) for the Dirichlet eigenvalue, and the
Szegö-Weinberger inequality ([33], [34]) for the Neumann case) and $k=2$ (the
Krahn-Szegö inequality ([20]) for the Dirichlet case, and recently proved by
D. Bucur and A. Henrot in [4] for the Neumann case).
The first major progress on the conjecture was made by G. Pólya himself in
1961 ([30]), in which he presented an elegant proof of his conjecture for
planar tiling domains (in fact G. Pólya’s proof for the Neumann eigenvalue
case was a bit complicated and assumed the tiling to be regular. The
regularity assumption was removed and the proof was simplified in 1966 by R.
Kellner [17]). Very recently, N. Filonov, M. Levitin, I. Polterovich and D.
Sher ([7]) proved that Pólya’s conjecture holds for planar disks (and for
Euclidean balls of all dimensions for the Dirichlet case), and thus gave the
first non-tiling planar domain for which Pólya’s conjecture is known to be
true.
For an arbitrary Euclidean domain $\Omega\subset\mathbb{R}^{d}$, P. Li and
S.T. Yau proved in [23] that
(1.4)
$\sum_{j=1}^{k}\lambda_{k}(\Omega)\geq\frac{d}{d+2}\frac{4\pi^{2}}{(\omega_{d}|\Omega|)^{\frac{2}{d}}}k^{\frac{d+2}{d}},$
and as a consequence, got a weaker version of Pólya’s inequality for all
Dirichlet eigenvalues,
(1.5)
$\lambda_{k}(\Omega)\geq\frac{d}{d+2}\frac{4\pi^{2}}{(\omega_{d}|\Omega|)^{\frac{2}{d}}}k^{\frac{2}{d}}.$
P. Kröger proved in [18] a similar weaker upper bound for the Neumann
eigenvalues of any Euclidean domain with piecewise smooth boundary.
Another important class of domains satisfying Pólya’s conjecture was obtained
by A. Laptev [21], in which he proved that if Pólya’s conjecture (1.2) holds
for $\Omega_{1}\subset\mathbb{R}^{d_{1}}$, where $d_{1}\geq 2$, then Pólya’s
conjecture (1.2) also holds for any domain of the form
$\Omega=\Omega_{1}\times\Omega_{2}$. One key ingredient in his proof is the
following inequality (which is a special case of Berezin-Lieb inequality ([3],
[24]) and is equivalent to Li-Yau’s inequality (1.4) above) for the Riesz
mean,
(1.6)
$\sum_{\lambda_{k}(\Omega)<\lambda}(\lambda-\lambda_{k}(\Omega))^{\gamma}\leq
L_{\gamma,d}|\Omega|\lambda^{\gamma+\frac{d}{2}},$
where $\gamma\geq 1$, and
(1.7)
$L_{\gamma,d}=\frac{\Gamma(\gamma+1)}{(4\pi)^{\frac{d}{2}}\Gamma(\gamma+1+\frac{d}{2})}.$
For Neumann eigenvalues, A. Laptev also got a similar inequality
(1.8) $\sum_{\mu_{k}(\Omega)<\lambda}(\lambda-\mu_{k}(\Omega))^{\gamma}\geq
L_{\gamma,d}|\Omega|\lambda^{\gamma+\frac{d}{2}}$
using which one can get Pólya’s conjecture (1.3) for
$\Omega=\Omega_{1}\times\Omega_{2}$ provided $\Omega_{1}$ satisfies (1.3) and
has dimension $d_{1}\geq 2$. For other recent progresses concerning Pólya’s
conjecture, we refer to [8],[11], [12], [13], [25] etc.
In this paper we will prove Pólya’s conjecture for domains of product type
that are “thin” in one component, namely regions of the form
$\Omega=a\Omega_{1}\times\Omega_{2}$
for $a$ small enough, without assuming that $\Omega_{1}$ or $\Omega_{2}$
satisfies Pólya’s conjecture. We first prove
###### Theorem 1.1.
Let $\Omega_{1}\subset\mathbb{R}^{d_{1}}$ and
$\Omega_{2}\subset\mathbb{R}^{d_{2}}$ be bounded Euclidean domains, where
$d_{1},d_{2}\geq 2$, and $\Omega_{2}$ has piecewise smooth boundary.
1. (1)
If $\Omega_{1}$ has Lipschitz boundary, then there exists $a_{0}>0$ (depends
on $\Omega_{1}$ and $\Omega_{2}$) such that for any $0<a<a_{0}$, the product
$\Omega=a\Omega_{1}\times\Omega_{2}$ satisfies the Dirichlet Pólya’s
conjecture (1.2).
2. (2)
If $\Omega_{1}$ has $C^{1}$ boundary, then there exists $a_{0}>0$ (depends on
$\Omega_{1}$ and $\Omega_{2}$) such that for any $0<a<a_{0}$, the product
$\Omega=a\Omega_{1}\times\Omega_{2}$ satisfies the Neumann Pólya’s conjecture
(1.3).
Here is the strategy of proof: Following Laptev’s argument [21], we write the
eigenvalue counting function of $a\Omega_{1}\times\Omega_{2}$ as the sum of
many eigenvalue counting functions of $\Omega_{2}$. Although we don’t have
Pólya’s inequality for $\Omega_{2}$, we do have weaker inequalities (See (2.6)
and (2.7) below) that follow from Seeley’s version of the two-term Weyl law
(which only requires $\Omega_{2}$ to have piecewise smooth boundary). Now
instead of applying Laptev’s inequalities on Riesz mean above, we apply
stronger two-term inequalities on Riesz mean, namely (2.8) obtained by R.
Frank and S. Larson in [10] (for $\Omega_{1}$ to have Lipschitz boundary) and
its Neumann analogue, (2.9), obtained by R. Frank and L. Geisinger in [9] (for
$\Omega_{1}$ to have $C^{1}$ boundary) to control the sum of both terms in
Seeley’s inequalities. By comparing what we lose from Seeley’s two-term bound
and what we gain from these two-term Riesz mean bound, we are able to prove
that for $a$ small enough, Pólya’s inequalities hold for $\lambda$ large
enough (which depends on $a$). For smaller $\lambda$, we use a simple
observation that Laptev’s inequalities above are in fact strict (See §2.2
below), and thus (by taking $a$ even smaller) give us the demanded gap to
prove Pólya’s inequality. This argument works perfectly well for $d_{2}\geq
3$, but the first part fails for $d_{2}=2$ since in this case we can’t apply
Laptev-type inequality on Riesz mean (which requires
$\gamma=\frac{d_{2}-1}{2}\geq 1$) to control the sum of the second term of
Seeley’s inequality. Fortunately, we can overcome this problem by using Li-
Yau’s estimate (1.5) above and an explicit integral computation. The Neumann
case is a bit simpler, since we only need Weyl’s law to control the second
term in Seeley’s inequality for large $\lambda$, and thus we don’t need to
distinguish the case $d_{2}=2$ with $d_{2}\geq 3$.
Note that Laptev’s argument does not work for the case $d_{1}=1$, since the
inequalities (1.6) and (1.8) require $\gamma\geq 1$. Even though the interval
$(0,1)$ tiles $\mathbb{R}$, it is still not known whether $(0,1)\times\Omega$
satisfies Pólya’s conjecture for general $\Omega$. In the second part of this
paper, we turn to study Pólya’s conjecture for thin products
$(0,a)\times\Omega$. Instead of writing the eigenvalue counting function of
$(0,a)\times\Omega$ as the sum of many eigenvalue counting functions of
$(0,a)$ (which is a tiling domain) that we have a nice control, we will write
it as the sum of many eigenvalue counting functions of $\Omega$ and apply
Seeley’s two-term inequalities. By carefully analyzing the two sums, we shall
prove that in this case, all thin products satisfy Pólya’s conjecture:
###### Theorem 1.2.
Let $\Omega\subset\mathbb{R}^{d}$ be a bounded domain with piecewise smooth
boundary, then there exists $a_{0}>0$ (depends on $\Omega$) such that for any
$0<a<a_{0}$, $(0,a)\times\Omega$ satisfies Pólya’s conjecture (1.2) and (1.3).
Since scaling will not affect Pólya’s inequalities, we immediately see that
for any bounded Euclidean domain $\Omega$, there exists a constant $C>0$ such
that for all $A>C$, $(0,1)\times A\Omega$ satisfies Pólya’s conjecture.
Unfortunately we still can’t prove Pólya’s conjecture for products of the form
$(0,1)\times a\Omega$ for small $a$, which obviously implies Pólya’s
conjecture for $(0,1)\times\Omega$.
The last part of this paper devoted to Pólya’s inequalities for Riemannian
manifolds with boundary. Although the original conjecture was only proposed
for Euclidean domains, people do study the analogous problem in the more
general Riemannian setting. For example, P. Bérard and G. Besson proved in [2]
that for a 2-dimensional hemisphere (or a quarter of a sphere, or even an
octant of a sphere), both Dirichlet eigenvalues and Neumann eigenvalues
satisfy Pólya’s inequalities above. Recently in [14], P. Freitas, J. Mao and
I. Salavessa studied the problem for hemispheres in arbitrary dimension. They
showed that (1.3) holds for Neumann eigenvalues of hemispheres in any
dimension, while (1.2) fails for Dirichlet eigenvalues when $d>2$, and they
derived sharp inequality for Dirichlet eigenvalues by adding a correction
term.
It is thus a natural problem to find out more Riemannian manifolds with
boundary satisfying Pólya’s inequalities. Note that in the proof of Theorem
1.1 and Theorem 1.2, for $\Omega_{2}$ and $\Omega$ we mainly used Seeley’s
two-term Weyl’s inequality. As a result, by literally repeating the proof one
can easily see that for any closed Riemannian manifold $M$, the Neumann
eigenvalues of the product $a\Omega\times M$ satisfy Pólya conjecture (1.3) as
long as $a$ is small enough. For the Dirichlet case, there will be one extra
term (since $0$ is an eigenvalue of $M$) in the eigenvalue counting function
of the product, namely the number of eigenvalues of $\Omega$ that is less than
$a^{2}\lambda$, which can be explicitly calculated if $d_{1}=\dim\Omega=1$ and
can be controlled via Li-Yau’s estimate (1.5) if $d_{2}\geq 2$. As a result,
we are able to prove that Pólya’s conjecture holds for such Riemannian
manifolds with boundary:
###### Theorem 1.3.
Let $\Omega\subset\mathbb{R}^{d_{1}}$ be a bounded domain with $C^{1}$
boundary and $(M,g)$ be a closed Riemannian manifold of dimension $d_{2}\geq
2$. Then there exists $a_{0}>0$ (depends on $\Omega$ and $M$) such that for
any $0<a<a_{0}$, $a\Omega\times M$ satisfies Pólya’s conjecture (1.2) and
(1.3).
The arrangement of this paper is as follows. In Section 2 we will list the
two-term inequalities for the eigenvalues counting functions and for the Riesz
means that will be used later, and also prove strict version of Laptev’s
inequalities for completeness. In Section 3 we will prove Theorem 1.1, and in
Section 4 we will prove Theorem 1.2. In Section 5 we will turn to the
Riemannian manifold setting and prove Theorem 1.3. Moreover we will explain
how to get similar results for a larger class of eigenvalue problems. Finally
in Section 6 we will give an explicit non-tiling planar domain $\Omega$ and
explicitly calculate the constant involved in the proof, and as a result, show
that the Dirichlet eigenvalues of $[0,\frac{1}{4\pi}]\times\Omega$ for that
$\Omega$ satisfies (1.2). Similarly in the Riemannian setting we show that
$(0,a)\times S^{2}$ satisfies Pólya’s conjecture for $a\leq\frac{\pi}{24}$,
but will break both Pólya’s inequalities if $a$ is large.
## 2\. Some preparations
### 2.1. Two term inequalities for the eigenvalue counting functions and the
Riesz means
For any bounded domain $\Omega\subset\mathbb{R}^{d}$, we denote the Dirichlet
eigenvalue counting function by
$\mathcal{N}^{D}_{\Omega}(\lambda):=\\#\\{n:\ \lambda_{n}(\Omega)<\lambda\\},$
and the Neumann eigenvalue counting function by
$\mathcal{N}^{N}_{\Omega}(\lambda):=\\#\\{n:\ \mu_{n}(\Omega)<\lambda\\}.$
Then the inequality (1.1) implies
$\mathcal{N}^{D}_{\Omega}(\lambda)\leq\mathcal{N}^{N}_{\Omega}(\lambda),\quad\forall\lambda>0,$
while Pólya’s conjectures (1.2) and (1.3) can be restated as
(2.1) $\mathcal{N}^{D}_{\Omega}(\lambda)\leq
C_{d}|\Omega|\lambda^{\frac{d}{2}},\qquad\forall\lambda>0,$
for all bounded domains, and
(2.2) $\mathcal{N}^{N}_{\Omega}(\lambda)\geq
C_{d}|\Omega|\lambda^{\frac{d}{2}},\qquad\forall\lambda>0,$
for all bounded domains with suitable boundary regularity, where the constant
(2.3)
$C_{d}=\frac{\omega_{d}}{(2\pi)^{d}}=\frac{1}{(4\pi)^{\frac{d}{2}}\Gamma(\frac{d}{2}+1)}=L_{0,d}.$
Since the unit balls satisfy $B^{d_{1}+d_{2}}\subset B^{d_{1}}\times
B^{d_{2}}$, one has $\omega_{d_{1}+d_{2}}<\omega_{d_{1}}\cdot\omega_{d_{2}}$
and thus
(2.4) $C_{d_{1}+d_{2}}<C_{d_{1}}\cdot C_{d_{2}}.$
It was first obtained by H. Weyl ([35]) that both eigenvalue counting
functions $\mathcal{N}^{D}_{\Omega}(\lambda)$ and
$\mathcal{N}^{N}_{\Omega}(\lambda)$ have the same leading asymptotics
(2.5)
$\mathcal{N}^{D/N}_{\Omega}(\lambda)=C_{d}|\Omega|\lambda^{\frac{d}{2}}+o(\lambda^{\frac{d}{2}})$
as $\lambda\to\infty$, and the famous Weyl’s conjecture, proven by V. Ivrii
([16]) and R. Melrose ([26]) under extra assumptions on the behavior of
billiard dynamics, claims that for $\Omega\subset\mathbb{R}^{d}$ with
piecewise smooth boundary,
$\mathcal{N}^{D}_{\Omega}(\lambda)=C_{d}|\Omega|\lambda^{\frac{d}{2}}-\frac{1}{4}C_{d-1}|\partial\Omega|\lambda^{\frac{d-1}{2}}+o(\lambda^{\frac{d-1}{2}})$
while
$\mathcal{N}^{N}_{\Omega}(\lambda)=C_{d}|\Omega|\lambda^{\frac{d}{2}}+\frac{1}{4}C_{d-1}|\partial\Omega|\lambda^{\frac{d-1}{2}}+o(\lambda^{\frac{d-1}{2}}),$
where $|\partial\Omega|$ is the surface area of $\partial\Omega$.
Although Weyl’s conjecture was not proven in its full generality, R. Seeley
([31], [32]) proved a weaker version, namely both eigenvalue counting
functions satisfy
$\mathcal{N}^{D/N}_{\Omega}(\lambda)=C_{d}|\Omega|\lambda^{\frac{d}{2}}+\mathrm{O}(\lambda^{\frac{d-1}{2}}),\qquad\text{as\
}\lambda\to\infty,$
for all bounded domains in $\mathbb{R}^{d}$ with piecewise smooth boundary. In
view of the facts $\lambda_{1}(\Omega)>0$ and $\mu_{0}(\Omega)=0$, we see that
there exists a positive constant $C(\Omega)$ such that for any $\lambda>0$,
(2.6) $\mathcal{N}^{D}_{\Omega}(\lambda)\leq
C_{d}|\Omega|\lambda^{\frac{d}{2}}+C(\Omega)\lambda^{\frac{d-1}{2}}$
and
(2.7) $\mathcal{N}^{N}_{\Omega}(\lambda)\geq
C_{d}|\Omega|\lambda^{\frac{d}{2}}-C(\Omega)\lambda^{\frac{d-1}{2}}.$
These two-term inequalities sharpen Weyl’s leading estimates and will play a
crucial role below.
We also need two-term inequalities for the Riesz mean that sharpen Laptev’s
inequalities (1.6) and (1.8). For the Dirichlet case, R. L. Frank and S.
Larson ([10]) proved that for any bounded domain $\Omega$ in $\mathbb{R}^{d}$
($d\geq 2$) with Lipschitz boundary and any $\gamma\geq 1$,
$\sum_{\lambda_{k}(\Omega)<\lambda}(\lambda-\lambda_{k}(\Omega))^{\gamma}=L_{\gamma,d}|\Omega|\lambda^{\gamma+\frac{d}{2}}-\frac{1}{4}L_{\gamma,d-1}|\partial\Omega|\lambda^{\gamma+\frac{d-1}{2}}+\mathrm{o}(\lambda^{\gamma+\frac{d-1}{2}})$
as $\lambda\to\infty$. As a consequence, for fixed $\gamma$, there exists a
positive constant $C_{1}(\Omega)$ such that if $\lambda>C_{1}(\Omega)$, one
has
(2.8)
$\sum_{\lambda_{k}(\Omega)<\lambda}(\lambda-\lambda_{k}(\Omega))^{\gamma}\leq
L_{\gamma,d}|\Omega|\lambda^{\gamma+\frac{d}{2}}-\frac{1}{5}L_{\gamma,d-1}|\partial\Omega|\lambda^{\gamma+\frac{d-1}{2}}.$
For the Neumann case, R. L. Frank and L. Geisinger ([9]) proved that for any
bounded domain $\Omega$ in $\mathbb{R}^{d}$ ($d\geq 2$) with $C^{1}$ boundary
and any $\gamma\geq 1$, one has
$\sum_{\mu_{k}(\Omega)<\lambda}(\lambda-\mu_{k}(\Omega))^{\gamma}=L_{\gamma,d}|\Omega|\lambda^{\gamma+\frac{d}{2}}+\frac{1}{4}L_{\gamma,d-1}|\partial\Omega|\lambda^{\gamma+\frac{d-1}{2}}+\mathrm{o}(\lambda^{\gamma+\frac{d-1}{2}})$
as $\lambda\to\infty$. As a consequence, for fixed $\gamma$, there exists a
positive constant $C_{2}(\Omega)$ such that if $\lambda>C_{2}(\Omega)$, one
has
(2.9) $\sum_{\mu_{k}(\Omega)<\lambda}(\lambda-\mu_{k}(\Omega))^{\gamma}\geq
L_{\gamma,d}|\Omega|\lambda^{\gamma+\frac{d}{2}}+\frac{1}{5}L_{\gamma,d-1}|\partial\Omega|\lambda^{\gamma+\frac{d-1}{2}}.$
### 2.2. Strict inequalities for Riesz means
Another ingredient in our proof is the inequality (1.6) of A. Laptev, which
follows from
$\sum_{\lambda_{k}(\Omega)<\lambda}(\lambda-\lambda_{k}(\Omega))\leq
L_{1,d}|\Omega|\lambda^{\frac{d+2}{2}}.$
By carefully analyzing his proof, one can see that the inequality is strict.
We include the proof for reader’s convenience.
###### Proposition 2.1.
Let $\Omega$ be a bounded domain in $\mathbb{R}^{d}$, then for any
$\lambda>0$, one has
$\sum_{\lambda_{k}(\Omega)<\lambda}(\lambda-\lambda_{k}(\Omega))<L_{1,d}|\Omega|\lambda^{\frac{d+2}{2}}.$
###### Proof.
We first sketch Laptev’s proof ([21, Theorem 2.1]). Let
$\\{\varphi_{k}\\}_{k=1}^{\infty}$ be the $L^{2}$ normalized eigenfunctions
associated to $\\{\lambda_{k}(\Omega)\\}_{k=1}^{\infty}$. Then
$\sum_{\lambda_{k}(\Omega)<\lambda}(\lambda-\lambda_{k}(\Omega))\leq(2\pi)^{-d}\int_{\mathbb{R}^{d}}(\lambda-|\xi|^{2})_{+}\cdot\sum_{\lambda_{k}(\Omega)<\lambda}|\hat{\varphi}_{k}(\xi)|^{2}\mathrm{d}\xi.$
Let
$e_{\xi}(x)=\begin{cases}e^{-ix\cdot\xi},&\qquad x\in\Omega,\\\ 0,&\qquad
x\notin\Omega.\end{cases}$
Then
$\sum_{\lambda_{k}(\Omega)<\lambda}|\hat{\varphi}_{k}(\xi)|^{2}=\sum_{\lambda_{k}(\Omega)<\lambda}\langle
e_{\xi},\varphi_{k}\rangle^{2}_{L^{2}(\Omega)}\\\
\leq\|e_{\xi}\|^{2}_{L^{2}(\Omega)}=|\Omega|.$
So
$\sum_{\lambda_{k}(\Omega)<\lambda}(\lambda-\lambda_{k}(\Omega))\leq(2\pi)^{-d}|\Omega|\int_{\mathbb{R}^{d}}(\lambda-|\xi|^{2})_{+}\mathrm{d}\xi=L_{1,d}|\Omega|\lambda^{\frac{d+2}{2}}.$
Now assume by contradiction that
$\sum_{\lambda_{k}(\Omega)<\lambda}(\lambda-\lambda_{k}(\Omega))=L_{1,d}|\Omega|\lambda^{\frac{d+2}{2}}$
for some $\lambda$, then
$\sum_{\lambda_{k}(\Omega)<\lambda}\langle
e_{\xi},\varphi_{k}\rangle^{2}_{L^{2}(\Omega)}=\|e_{\xi}\|^{2}_{L^{2}(\Omega)},\qquad\forall|\xi|^{2}<\lambda.$
Take $N$ such that $\lambda_{N}(\Omega)<\lambda\leq\lambda_{N+1}(\Omega)$,
then one has $e_{\xi}\in\mathrm{span}\\{\varphi_{1},\cdots,\varphi_{N}\\}$ for
any $\xi$ with $|\xi|^{2}<\lambda$. So if we take $N+1$ different points
$\\{\xi_{1},\cdots,\xi_{N+1}\\}$ with $|\xi_{i}|^{2}<\lambda$, $\forall 1\leq
i\leq N+1$, then there exists $(c_{1},\cdots,c_{N+1})\neq(0,\cdots,0)$ such
that
$\sum_{i=1}^{N+1}c_{i}e_{\xi_{i}}(x)=0,\qquad\forall x\in\Omega.$
By the analyticity of $e^{-ix\cdot\xi}$, one has
$\sum_{i=1}^{N+1}c_{i}e^{-ix\cdot\xi_{i}}=0,\qquad\forall x\in\mathbb{R}^{d},$
which is a contradiction. ∎
Together with the fact
(2.10)
$\int^{\infty}_{0}(z-\lambda-t)_{+}t^{\gamma-2}\mathrm{d}t=\frac{(z-\lambda)^{\gamma}_{+}}{\gamma(\gamma-1)},\qquad\forall\gamma>1,$
we immediately get
###### Corollary 2.2.
Let $\Omega$ be a bounded domain in $\mathbb{R}^{d}$, then for any $\lambda>0$
and $\gamma\geq 1$, one has
$\sum_{\lambda_{k}(\Omega)<\lambda}(\lambda-\lambda_{k}(\Omega))^{\gamma}<L_{\gamma,d}|\Omega|\lambda^{\gamma+\frac{d}{2}}.$
In particular, for any $0<a<b$, one has
(2.11) $K(a,b,\Omega):=\inf_{a\leq\lambda\leq
b}\frac{L_{\gamma,d}|\Omega|\lambda^{\gamma+\frac{d}{2}}-\sum_{\lambda_{k}(\Omega)<\lambda}(\lambda-\lambda_{k}(\Omega))^{\gamma}}{\lambda^{\gamma+\frac{d-1}{2}}}>0.$
We also need the Neumann version of the same (strict) inequality, which is
essentially proved by A. Laptev (c.f. [21], Theorem 3.1):
###### Proposition 2.3.
Let $\Omega$ be a bounded domain in $\mathbb{R}^{d}$ with discrete Neumann
spectrum. Then for any $\lambda\geq 0$, one has
$\sum_{\mu_{k}(\Omega)<\lambda}(\lambda-\mu_{k}(\Omega))>L_{1,d}|\Omega|\lambda^{\frac{d+2}{2}}.$
###### Proof.
Again we first sketch Laptev’s proof of the corresponding (non-strict)
inequality in [21]. Let $\\{\psi_{k}\\}_{k=0}^{\infty}$ be the $L^{2}$
normalized eigenfunctions associated to $\\{\mu_{k}(\Omega)\\}_{k=0}^{\infty}$
and denote $\varphi_{\lambda}(x)=(\lambda-x)_{+}$. Then
$\sum_{\mu_{k}(\Omega)<\lambda}(\lambda-\mu_{k}(\Omega))=(2\pi)^{-d}|\Omega|\int_{\mathbb{R}^{d}}\sum_{k=0}^{\infty}\frac{1}{|\Omega|}\langle
e_{\xi},\psi_{k}\rangle^{2}_{L^{2}(\Omega)}\cdot\varphi_{\lambda}(\mu_{k}(\Omega))\mathrm{d}\xi.$
Using the facts
$\sum_{k=0}^{\infty}\frac{1}{|\Omega|}\langle
e_{\xi},\psi_{k}\rangle^{2}_{L^{2}(\Omega)}=1\quad\text{and}\quad\sum_{k=0}^{\infty}\frac{1}{|\Omega|}\langle
e_{\xi},\psi_{k}\rangle^{2}_{L^{2}(\Omega)}\cdot\mu_{k}=|\xi|^{2}$
and the Jensen inequality (since $\varphi_{\lambda}$ is convex), Laptev proved
$\sum_{k=0}^{\infty}\frac{1}{|\Omega|}\langle
e_{\xi},\psi_{k}\rangle^{2}_{L^{2}(\Omega)}\cdot\varphi_{\lambda}(\mu_{k}(\Omega))\geq\varphi_{\lambda}(|\xi|^{2})$
and thus
$\sum_{\mu_{k}(\Omega)<\lambda}(\lambda-\mu_{k}(\Omega))\geq
L_{1,d}|\Omega|\lambda^{\frac{d+2}{2}}.$
To show that the inequality is strict, again we fix $\lambda$ and let $N$ be
the integer with $\mu_{N}(\Omega)\leq\lambda<\mu_{N+1}(\Omega)$. Take $\xi$
with $|\xi|^{2}<\lambda$ such that
$\langle e_{\xi},1\rangle_{L^{2}(\Omega)}\neq 0,\text{ and
}e_{\xi}\notin\mathrm{span}\\{\psi_{1},\cdots,\psi_{N}\\},$
then for such a $\xi$, Jensen’s inequality is strict:
$\displaystyle\sum_{k=0}^{\infty}\frac{1}{|\Omega|}\langle
e_{\xi},\psi_{k}\rangle^{2}_{L^{2}(\Omega)}\cdot\varphi_{\lambda}(\mu_{k}(\Omega))$
$\displaystyle=$
$\displaystyle\sum_{\mu_{k}(\Omega)<\lambda}\frac{1}{|\Omega|}\langle
e_{\xi},\psi_{k}\rangle^{2}_{L^{2}(\Omega)}\cdot\varphi_{\lambda}(\mu_{k}(\Omega))+\sum_{\mu_{k}(\Omega)>\lambda}\frac{1}{|\Omega|}\langle
e_{\xi},\psi_{k}\rangle^{2}_{L^{2}(\Omega)}\cdot\varphi_{\lambda}(\mu_{k}(\Omega))$
$\displaystyle=$
$\displaystyle(\sum_{\mu_{k}(\Omega)<\lambda}\frac{1}{|\Omega|}\langle
e_{\xi},\psi_{k}\rangle^{2}_{L^{2}(\Omega)})\cdot\varphi_{\lambda}(\frac{\sum_{\mu_{k}(\Omega)<\lambda}\frac{1}{|\Omega|}\langle
e_{\xi},\psi_{k}\rangle^{2}_{L^{2}(\Omega)}\cdot\mu_{k}(\Omega)}{\sum_{\mu_{k}(\Omega)<\lambda}\frac{1}{|\Omega|}\langle
e_{\xi},\psi_{k}\rangle^{2}_{L^{2}(\Omega)}})+$
$\displaystyle(\sum_{\mu_{k}(\Omega)>\lambda}\frac{1}{|\Omega|}\langle
e_{\xi},\psi_{k}\rangle^{2}_{L^{2}(\Omega)})\cdot\varphi_{\lambda}(\frac{\sum_{\mu_{k}(\Omega)>\lambda}\frac{1}{|\Omega|}\langle
e_{\xi},\psi_{k}\rangle^{2}_{L^{2}(\Omega)}\cdot\mu_{k}(\Omega)}{\sum_{\mu_{k}(\Omega)>\lambda}\frac{1}{|\Omega|}\langle
e_{\xi},\psi_{k}\rangle^{2}_{L^{2}(\Omega)}})$ $\displaystyle>$
$\displaystyle\varphi_{\lambda}(\sum_{\mu_{k}(\Omega)<\lambda}\frac{1}{|\Omega|}\langle
e_{\xi},\psi_{k}\rangle^{2}_{L^{2}(\Omega)}\cdot\mu_{k}(\Omega)+\sum_{\mu_{k}(\Omega)>\lambda}\frac{1}{|\Omega|}\langle
e_{\xi},\psi_{k}\rangle^{2}_{L^{2}(\Omega)}\cdot\mu_{k}(\Omega))$
$\displaystyle=$ $\displaystyle\varphi_{\lambda}(|\xi|^{2}).$
This completes the proof. ∎
As a consequence,
###### Corollary 2.4.
Let $\Omega$ be a bounded domain in $\mathbb{R}^{d}$ with discrete Neumann
spectrum. Then for any $\lambda\geq 0$ and $\gamma\geq 1$, one has
$\sum_{\mu_{k}(\Omega)<\lambda}(\lambda-\mu_{k}(\Omega))^{\gamma}>L_{\gamma,d}|\Omega|\lambda^{\gamma+\frac{d}{2}}.$
In particular, for any $0<a<b$, one has
$K_{1}(a,b,\Omega):=\inf_{a\leq\lambda\leq
b}\frac{\sum_{\mu_{k}(\Omega)<\lambda}(\lambda-\mu_{k}(\Omega))^{\gamma}-L_{\gamma,d}|\Omega|\lambda^{\gamma+\frac{d}{2}}}{\lambda^{\gamma+\frac{d-1}{2}}}>0.$
## 3\. Proof of Theorem 1.1
As observed by P. Freitas, J. Lagace and J. Payette in [12, Proposition 3.1],
it is enough to assume that both $\Omega_{1}$ and $\Omega_{2}$ are connected.
We divide the proof of Theorem 1.1 into three parts: the Dirichlet case with
$d_{2}\geq 3$, the Dirichlet case with $d_{2}=2$, and the Neumann case.
For the Dirichlet case, the eigenvalues of $a\Omega_{1}\times\Omega_{2}$ are
$a^{-2}\lambda_{l}(\Omega_{1})+\lambda_{k}(\Omega_{2}),\qquad\forall
l,k\in\mathbb{Z}_{>0}$
and thus
$\mathcal{N}^{D}_{a\Omega_{1}\times\Omega_{2}}(\lambda)=\sum_{l=1}^{Z^{\lambda}_{a}}\mathcal{N}^{D}_{\Omega_{2}}(\lambda-a^{-2}\lambda_{l}(\Omega_{1})),$
where
$Z^{\lambda}_{a}=\mathcal{N}^{D}_{\Omega_{1}}(a^{2}\lambda).$
By inequality (2.6), there exists a constant $C(\Omega_{2})>0$ such that
$\mathcal{N}^{D}_{\Omega_{2}}(\lambda)\leq
C_{d_{2}}|\Omega_{2}|\lambda^{\frac{d_{2}}{2}}+C(\Omega_{2})\lambda^{\frac{d_{2}-1}{2}},\qquad\forall\lambda>0.$
So we get
(3.1)
$\displaystyle\sum_{l=1}^{Z^{\lambda}_{a}}\mathcal{N}^{D}_{\Omega_{2}}(\lambda-a^{-2}\lambda_{l}(\Omega_{1}))$
$\displaystyle\leq$ $\displaystyle
C_{d_{2}}|\Omega_{2}|\sum_{l=1}^{Z^{\lambda}_{a}}(\lambda-a^{-2}\lambda_{l}(\Omega_{1}))^{\frac{d_{2}}{2}}+C(\Omega_{2})\sum_{l=1}^{Z^{\lambda}_{a}}(\lambda-a^{-2}\lambda_{l}(\Omega_{1}))^{\frac{d_{2}-1}{2}}$
$\displaystyle=$ $\displaystyle
C_{d_{2}}|\Omega_{2}|a^{-d_{2}}\sum_{l=1}^{Z^{\lambda}_{a}}(a^{2}\lambda-\lambda_{l}(\Omega_{1}))^{\frac{d_{2}}{2}}+C(\Omega_{2})a^{1-d_{2}}\sum_{l=1}^{Z^{\lambda}_{a}}(a^{2}\lambda-\lambda_{l}(\Omega_{1}))^{\frac{d_{2}-1}{2}}.$
By inequality (2.8), there exists a constant $C(\Omega_{1})>0$ such that if
$a^{2}\lambda>C(\Omega_{1})$, then
(3.2)
$\sum_{l=1}^{Z^{\lambda}_{a}}(a^{2}\lambda-\lambda_{l}(\Omega_{1}))^{\frac{d_{2}}{2}}\leq
L_{\frac{d_{2}}{2},d_{1}}|\Omega_{1}|a^{d_{1}+d_{2}}\lambda^{\frac{d_{1}+d_{2}}{2}}-\frac{1}{5}L_{\frac{d_{2}}{2},d_{1}-1}|\partial\Omega_{1}|a^{d_{1}+d_{2}-1}\lambda^{\frac{d_{1}+d_{2}-1}{2}}.$
### 3.1. The Dirichlet case with $d_{2}\geq 3$
By Corollary 2.2, one has
(3.3)
$\sum_{l=1}^{Z^{\lambda}_{a}}(a^{2}\lambda-\lambda_{l}(\Omega_{1}))^{\frac{d_{2}-1}{2}}<L_{\frac{d_{2}-1}{2},d_{1}}|\Omega_{1}|a^{d_{1}+d_{2}-1}\lambda^{\frac{d_{1}+d_{2}-1}{2}}.$
So by (3.1), (3.2), (3.3) and the fact
$C_{d_{2}}L_{\frac{d_{2}}{2},d_{1}}=C_{d_{1}+d_{2}},$
one has that if $a^{2}\lambda>C(\Omega_{1})$, then
$\displaystyle\sum_{l=1}^{Z^{\lambda}_{a}}\mathcal{N}^{D}_{\Omega_{2}}(\lambda-a^{-2}\lambda_{l}(\Omega_{1}))$
$\displaystyle\leq$ $\displaystyle
C_{d_{1}+d_{2}}a^{d_{1}}|\Omega_{1}||\Omega_{2}|\lambda^{\frac{d_{1}+d_{2}}{2}}-\frac{1}{5}C_{d_{1}+d_{2}-1}|\partial\Omega_{1}||\Omega_{2}|a^{d_{1}-1}\lambda^{\frac{d_{1}+d_{2}-1}{2}}$
$\displaystyle+L_{\frac{d_{2}-1}{2},d_{1}}|\Omega_{1}|C(\Omega_{2})a^{d_{1}}\lambda^{\frac{d_{1}+d_{2}-1}{2}}.$
Thus if we assume
$a<\frac{C_{d_{1}+d_{2}-1}|\partial\Omega_{1}||\Omega_{2}|}{5L_{\frac{d_{2}-1}{2},d_{1}}|\Omega_{1}|C(\Omega_{2})}=\frac{C_{d_{2}-1}|\partial\Omega_{1}||\Omega_{2}|}{5|\Omega_{1}|C(\Omega_{2})},$
then for any $\lambda>a^{-2}C(\Omega_{1})$, we will get the demanded
inequality
$\mathcal{N}^{D}_{a\Omega_{1}\times\Omega_{2}}(\lambda)\leq
C_{d_{1}+d_{2}}a^{d_{1}}|\Omega_{1}||\Omega_{2}|\lambda^{\frac{d_{1}+d_{2}}{2}}.$
Note that if $0<\lambda<a^{-2}\lambda_{1}(\Omega_{1})$, then we automatically
have
$\mathcal{N}^{D}_{a\Omega_{1}\times\Omega_{2}}(\lambda)=0<C_{d_{1}+d_{2}}a^{d_{1}}|\Omega_{1}||\Omega_{2}|\lambda^{\frac{d_{1}+d_{2}}{2}}.$
So it remains to consider the case
$a^{-2}\lambda_{1}(\Omega_{1})\leq\lambda\leq a^{-2}C(\Omega_{1})$ assuming
$C(\Omega_{1})>\lambda_{1}(\Omega_{1})$. Let $\mu=a^{2}\lambda$, then by
Corollary 2.2, one has
$K(\Omega_{1})=\inf_{\lambda_{1}(\Omega_{1})\leq\mu\leq
C(\Omega_{1})}\frac{L_{\frac{d_{2}}{2},d_{1}}|\Omega_{1}|\mu^{\frac{d_{1}+d_{2}}{2}}-\sum_{\lambda_{l}(\Omega_{1})<\mu}(\mu-\lambda_{l}(\Omega_{1}))^{\frac{d_{2}}{2}}}{\mu^{\frac{d_{1}+d_{2}-1}{2}}}>0$
and thus
(3.4)
$\sum_{\lambda_{l}(\Omega_{1})<\mu}(\mu-\lambda_{l}(\Omega_{1}))^{\frac{d_{2}}{2}}\leq
L_{\frac{d_{2}}{2},d_{1}}|\Omega_{1}|\mu^{\frac{d_{1}+d_{2}}{2}}-K(\Omega_{1})\mu^{\frac{d_{1}+d_{2}-1}{2}}$
for all $\lambda_{1}(\Omega_{1})\leq\mu\leq C(\Omega_{1})$. Thus by (3.1),
(3.3) and (3.4), one has that if $a^{-2}\lambda_{1}(\Omega_{1})\leq\lambda\leq
a^{-2}C(\Omega_{1})$, then
$\displaystyle\sum_{l=1}^{Z^{\lambda}_{a}}\mathcal{N}^{D}_{\Omega_{2}}(\lambda-a^{-2}\lambda_{l}(\Omega_{1}))$
$\displaystyle\leq$ $\displaystyle
C_{d_{1}+d_{2}}a^{d_{1}}|\Omega_{1}||\Omega_{2}|\lambda^{\frac{d_{1}+d_{2}}{2}}-K(\Omega_{1})C_{d_{2}}|\Omega_{2}|a^{d_{1}-1}\lambda^{\frac{d_{1}+d_{2}-1}{2}}$
$\displaystyle+L_{\frac{d_{2}-1}{2},d_{1}}|\Omega_{1}|C(\Omega_{2})a^{d_{1}}\lambda^{\frac{d_{1}+d_{2}-1}{2}}.$
So if we assume
$a<\frac{K(\Omega_{1})C_{d_{2}}|\Omega_{2}|}{L_{\frac{d_{2}-1}{2},d_{1}}|\Omega_{1}|C(\Omega_{2})}$,
then for any $a^{-2}\lambda_{1}(\Omega_{1})\leq\lambda\leq
a^{-2}C(\Omega_{1})$,
$\mathcal{N}^{D}_{a\Omega_{1}\times\Omega_{2}}(\lambda)\leq
C_{d_{1}+d_{2}}a^{d_{1}}|\Omega_{1}||\Omega_{2}|\lambda^{\frac{d_{1}+d_{2}}{2}}.$
Combining all discussions above, one has that if
$a<\min\bigg{(}\frac{C_{d_{2}-1}|\partial\Omega_{1}||\Omega_{2}|}{5|\Omega_{1}|C(\Omega_{2})},\frac{K(\Omega_{1})C_{d_{2}}|\Omega_{2}|}{L_{\frac{d_{2}-1}{2},d_{1}}|\Omega_{1}|C(\Omega_{2})}\bigg{)},$
then all Dirichlet eigenvalues of $a\Omega_{1}\times\Omega_{2}$ satisfy
Pólya’s conjecture (2.1). This complete the proof of Theorem 1.1, part (1),
for the case of $d_{2}\geq 3$.
### 3.2. The Dirichlet case with $d_{2}=2$
Since $C_{2}=\frac{1}{4\pi}$, the inequality (3.1) becomes
(3.5)
$\displaystyle\sum_{l=1}^{Z^{\lambda}_{a}}\mathcal{N}^{D}_{\Omega_{2}}(\lambda-a^{-2}\lambda_{l}(\Omega_{1}))$
$\displaystyle\leq$
$\displaystyle(4\pi)^{-1}|\Omega_{2}|a^{-2}\sum_{l=1}^{Z^{\lambda}_{a}}(a^{2}\lambda-\lambda_{l}(\Omega_{1}))+C(\Omega_{2})a^{-1}\sum_{l=1}^{Z^{\lambda}_{a}}(a^{2}\lambda-\lambda_{l}(\Omega_{1}))^{\frac{1}{2}}$
and the inequality (3.2) gives, for $a^{2}\lambda>C(\Omega_{1})$,
(3.6) $\sum_{l=1}^{Z^{\lambda}_{a}}(a^{2}\lambda-\lambda_{l}(\Omega_{1}))\leq
L_{1,d_{1}}|\Omega_{1}|a^{d_{1}+2}\lambda^{\frac{d_{1}+2}{2}}-\frac{1}{5}L_{1,d_{1}-1}|\partial\Omega_{1}|a^{d_{1}+1}\lambda^{\frac{d_{1}+1}{2}}.$
To estimate the second term in (3.5), we use Li-Yau’s lower bound (1.5),
namely
$\lambda_{l}(\Omega_{1})\geq\frac{d_{1}}{d_{1}+2}l^{\frac{2}{d_{1}}}(C_{d_{1}}|\Omega_{1}|)^{-\frac{2}{d_{1}}},$
to get
(3.7)
$\displaystyle\sum_{l=1}^{Z^{\lambda}_{a}}(a^{2}\lambda-\lambda_{l}(\Omega_{1}))^{\frac{1}{2}}\leq$
$\displaystyle\sum_{l=1}^{Z^{\lambda}_{a}}\big{(}a^{2}\lambda-\frac{d_{1}}{d_{1}+2}l^{\frac{2}{d_{1}}}(C_{d_{1}}|\Omega_{1}|)^{-\frac{2}{d_{1}}}\big{)}^{\frac{1}{2}}$
$\displaystyle\leq$
$\displaystyle\int_{0}^{\infty}\big{(}a^{2}\lambda-\frac{d_{1}}{d_{1}+2}x^{\frac{2}{d_{1}}}(C_{d_{1}}|\Omega_{1}|)^{-\frac{2}{d_{1}}}\big{)}_{+}^{\frac{1}{2}}\mathrm{d}x$
$\displaystyle=$
$\displaystyle(\frac{d_{1}+2}{d_{1}})^{\frac{d_{1}}{2}}C_{d_{1}}|\Omega_{1}|\frac{d_{1}}{2}a^{d_{1}+1}\lambda^{\frac{d_{1}+1}{2}}\int_{0}^{1}(1-s)^{\frac{1}{2}}s^{\frac{d_{1}}{2}-1}\mathrm{d}s$
$\displaystyle=$
$\displaystyle(\frac{d_{1}+2}{d_{1}})^{\frac{d_{1}}{2}}L_{\frac{1}{2},d_{1}}|\Omega_{1}|a^{d_{1}+1}\lambda^{\frac{d_{1}+1}{2}}.$
Then by (3.5), (3.6) and (3.7), one has that if $a^{2}\lambda>C(\Omega_{1})$,
then
$\displaystyle\sum_{l=1}^{Z^{\lambda}_{a}}\mathcal{N}^{D}_{\Omega_{2}}(\lambda-a^{-2}\lambda_{l}(\Omega_{1}))$
$\displaystyle\leq$ $\displaystyle
C_{d_{1}+2}a^{d_{1}}|\Omega_{1}||\Omega_{2}|\lambda^{\frac{d_{1}+2}{2}}-\frac{1}{5}C_{d_{1}+1}|\Omega_{2}||\partial\Omega_{1}|a^{d_{1}-1}\lambda^{\frac{d_{1}+1}{2}}+$
$\displaystyle(\frac{d_{1}+2}{d_{1}})^{\frac{d_{1}}{2}}C(\Omega_{2})L_{\frac{1}{2},d_{1}}|\Omega_{1}|a^{d_{1}}\lambda^{\frac{d_{1}+1}{2}}.$
So if we assume
$a<\frac{C_{d_{1}+1}|\Omega_{2}||\partial\Omega_{1}|}{5(\frac{d_{1}+2}{d_{1}})^{\frac{d_{1}}{2}}C(\Omega_{2})L_{\frac{1}{2},d_{1}}|\Omega_{1}|}=\frac{1}{5\pi}\big{(}\frac{d_{1}}{d_{1}+2}\big{)}^{\frac{d_{1}}{2}}\frac{|\Omega_{2}||\partial\Omega_{1}|}{C(\Omega_{2})|\Omega_{1}|},$
then for any $\lambda>a^{-2}C(\Omega_{1})$, one also gets the demanded
inequality
$\mathcal{N}^{D}_{a\Omega_{1}\times\Omega_{2}}(\lambda)\leq
C_{d_{1}+2}a^{d_{1}}|\Omega_{1}||\Omega_{2}|\lambda^{\frac{d_{1}+2}{2}}.$
For $\lambda<a^{-2}C(\Omega_{1})$, one just repeat the corresponding part of
the proof of the Dirichlet case with $d_{2}\geq 3$, so we omit it. This
completes the proof of Theorem 1.1, (1).
### 3.3. The Neumann case
Since the Neumann eigenvalues of $a\Omega_{1}\times\Omega_{2}$ are
$a^{-2}\mu_{l}(\Omega_{1})+\mu_{k}(\Omega_{2}),\qquad\forall
l,k\in\mathbb{Z}_{\geq 0},$
one has
$\mathcal{N}^{N}_{a\Omega_{1}\times\Omega_{2}}(\lambda)=\sum_{l=0}^{Y_{a}^{\lambda}}\mathcal{N}^{N}_{\Omega_{2}}(\lambda-a^{-2}\mu_{l}(\Omega_{1}))$
where
$Y_{a}^{\lambda}=\mathcal{N}^{N}_{\Omega_{1}}(a^{2}\lambda)-1.$
By inequality (2.7), there exists a constant $C_{1}(\Omega_{2})>0$ such that
$\mathcal{N}^{N}_{\Omega_{2}}(\lambda)\geq
C_{d_{2}}|\Omega_{2}|\lambda^{\frac{d_{2}}{2}}-C_{1}(\Omega_{2})\lambda^{\frac{d_{2}-1}{2}},\qquad\forall\lambda>0.$
So we get
(3.8)
$\displaystyle\sum_{l=0}^{Y_{a}^{\lambda}}\mathcal{N}^{N}_{\Omega_{2}}(\lambda-a^{-2}\mu_{l}(\Omega_{1}))$
$\displaystyle\geq$ $\displaystyle
C_{d_{2}}|\Omega_{2}|\sum_{l=0}^{Y_{a}^{\lambda}}(\lambda-a^{-2}\mu_{l}(\Omega_{1}))^{\frac{d_{2}}{2}}-C_{1}(\Omega_{2})\sum_{l=0}^{Y_{a}^{\lambda}}(\lambda-a^{-2}\mu_{l}(\Omega_{1}))^{\frac{d_{2}-1}{2}}$
$\displaystyle=$ $\displaystyle
C_{d_{2}}|\Omega_{2}|a^{-d_{2}}\sum_{l=0}^{Y_{a}^{\lambda}}(a^{2}\lambda-\mu_{l}(\Omega_{1}))^{\frac{d_{2}}{2}}-C_{1}(\Omega_{2})a^{1-d_{2}}\sum_{l=0}^{Y_{a}^{\lambda}}(a^{2}\lambda-\mu_{l}(\Omega_{1}))^{\frac{d_{2}-1}{2}}.$
For the first term, by (2.9), there exists a constant $C_{1}(\Omega_{1})>0$
such that if $a^{2}\lambda>C_{1}(\Omega_{1})$, then
(3.9)
$\sum_{l=0}^{Y_{a}^{\lambda}}(a^{2}\lambda-\mu_{l}(\Omega_{1}))^{\frac{d_{2}}{2}}\geq
L_{\frac{d_{2}}{2},d_{1}}|\Omega_{1}|a^{d_{1}+d_{2}}\lambda^{\frac{d_{1}+d_{2}}{2}}+\frac{1}{5}L_{\frac{d_{2}}{2},d_{1}-1}|\partial\Omega_{1}|a^{d_{1}+d_{2}-1}\lambda^{\frac{d_{1}+d_{2}-1}{2}}.$
To estimate the second term, we use (2.5) to get $L=L(\Omega_{1})>0$ such that
$\mu_{l}(\Omega_{1})\geq\frac{1}{2}l^{\frac{2}{d_{1}}}(C_{d_{1}}|\Omega_{1}|)^{-\frac{2}{d_{1}}},\text{
if }l\geq L.$
Note that if $a^{2}\lambda$ is large enough, one has
$\sum_{l=L}^{3L-1}\big{(}a^{2}\lambda-\frac{1}{2}l^{\frac{2}{d_{1}}}(C_{d_{1}}|\Omega_{1}|)^{-\frac{2}{d_{1}}}\big{)}^{\frac{d_{2}-1}{2}}\geq
L(a^{2}\lambda)^{\frac{d_{2}-1}{2}}\geq\sum_{l=0}^{L-1}(a^{2}\lambda-\mu_{l}(\Omega_{1}))^{\frac{d_{2}-1}{2}}.$
So there exists a constant $C_{2}(\Omega_{1})>0$ such that if
$a^{2}\lambda>C_{2}(\Omega_{1})$, then
(3.10)
$\displaystyle\sum_{l=0}^{Y_{a}^{\lambda}}(a^{2}\lambda-\mu_{l}(\Omega_{1}))^{\frac{d_{2}-1}{2}}$
$\displaystyle\leq$ $\displaystyle
2\sum_{l=0}^{Y_{a}^{\lambda}}\big{(}a^{2}\lambda-\frac{1}{2}l^{\frac{2}{d_{1}}}(C_{d_{1}}|\Omega_{1}|)^{-\frac{2}{d_{1}}}\big{)}^{\frac{d_{2}-1}{2}}$
$\displaystyle=$ $\displaystyle
2(a^{2}\lambda)^{\frac{d_{2}-1}{2}}+2\sum_{l=1}^{Y_{a}^{\lambda}}\big{(}a^{2}\lambda-\frac{1}{2}l^{\frac{2}{d_{1}}}(C_{d_{1}}|\Omega_{1}|)^{-\frac{2}{d_{1}}}\big{)}^{\frac{d_{2}-1}{2}}$
$\displaystyle\leq$ $\displaystyle
4\int_{0}^{\infty}\big{(}a^{2}\lambda-\frac{1}{2}x^{\frac{2}{d_{1}}}(C_{d_{1}}|\Omega_{1}|)^{-\frac{2}{d_{1}}}\big{)}_{+}^{\frac{d_{2}-1}{2}}\mathrm{d}x$
$\displaystyle=$ $\displaystyle
2^{\frac{d_{1}}{2}+2}C_{d_{1}}|\Omega_{1}|\frac{d_{1}}{2}a^{d_{1}+d_{2}-1}\lambda^{\frac{d_{1}+d_{2}-1}{2}}\int_{0}^{1}(1-s)^{\frac{d_{2}-1}{2}}s^{\frac{d_{1}}{2}-1}\mathrm{d}s$
$\displaystyle=$ $\displaystyle
2^{\frac{d_{1}}{2}+1}C_{d_{1}}B(\frac{d_{1}}{2},\frac{d_{2}+1}{2})|\Omega_{1}|d_{1}a^{d_{1}+d_{2}-1}\lambda^{\frac{d_{1}+d_{2}-1}{2}}.$
Thus by (3.8), (3.9) and (3.10), one has that if
$a^{2}\lambda>\max(C_{1}(\Omega_{1}),C_{2}(\Omega_{1}))$, then
$\displaystyle\sum_{l=0}^{Y^{\lambda}_{a}}\mathcal{N}^{N}_{\Omega_{2}}(\lambda-a^{-2}\mu_{l}(\Omega_{1}))$
$\displaystyle\geq$ $\displaystyle
C_{d_{1}+d_{2}}a^{d_{1}}|\Omega_{1}||\Omega_{2}|\lambda^{\frac{d_{1}+d_{2}}{2}}+\frac{1}{5}C_{d_{1}+d_{2}-1}|\Omega_{2}||\partial\Omega_{1}|a^{d_{1}-1}\lambda^{\frac{d_{1}+d_{2}-1}{2}}-$
$\displaystyle
C_{1}(\Omega_{2})2^{\frac{d_{1}}{2}+1}C_{d_{1}}B(\frac{d_{1}}{2},\frac{d_{2}+1}{2})|\Omega_{1}|d_{1}a^{d_{1}}\lambda^{\frac{d_{1}+d_{2}-1}{2}}.$
So if we require
$a<\frac{C_{d_{1}+d_{2}-1}|\Omega_{2}||\partial\Omega_{1}|}{5C_{1}(\Omega_{2})2^{\frac{d_{1}}{2}+1}C_{d_{1}}B(\frac{d_{1}}{2},\frac{d_{2}+1}{2})|\Omega_{1}|d_{1}}=\frac{C_{d_{2}-1}|\Omega_{2}||\partial\Omega_{1}|}{5\cdot
2^{\frac{d_{1}}{2}+2}C_{1}(\Omega_{2})|\Omega_{1}|},$
then for any $\lambda>a^{-2}\max(C_{1}(\Omega_{1}),C_{2}(\Omega_{1}))$, one
gets the demanded
$\mathcal{N}^{N}_{a\Omega_{1}\times\Omega_{2}}(\lambda)\geq
C_{d_{1}+d_{2}}a^{d_{1}}|\Omega_{1}||\Omega_{2}|\lambda^{\frac{d_{1}+d_{2}}{2}}.$
Next, we consider $0<\lambda<a^{-2}\mu_{1}(\Omega_{1})$, in which case
$\mathcal{N}^{N}_{a\Omega_{1}\times\Omega_{2}}(\lambda)=\mathcal{N}^{N}_{\Omega_{2}}(\lambda).$
By Szegö-Weinberger inequality ([33], [34]), one has
$\mu_{1}(\Omega_{1})\leq(C_{d_{1}}|\Omega_{1}|)^{-\frac{2}{d_{1}}}$
which implies that for $0<\lambda<a^{-2}\mu_{1}(\Omega_{1})$,
$\displaystyle
C_{d_{1}+d_{2}}a^{d_{1}}|\Omega_{1}||\Omega_{2}|\lambda^{\frac{d_{1}+d_{2}}{2}}$
$\displaystyle<C_{d_{1}+d_{2}}a^{d_{1}}|\Omega_{1}||\Omega_{2}|(a^{-2}\mu_{1}(\Omega_{1}))^{\frac{d_{1}}{2}}\lambda^{\frac{d_{2}}{2}}$
$\displaystyle\leq\frac{C_{d_{1}+d_{2}}}{C_{d_{1}}}|\Omega_{2}|\lambda^{\frac{d_{2}}{2}}.$
On the other hand, by (2.4) one has
$\frac{C_{d_{1}+d_{2}}}{C_{d_{1}}}<C_{d_{2}}$. So by (2.5), there exists a
constant $C_{2}(\Omega_{2})>0$ such that for $\lambda>C_{2}(\Omega_{2})$,
$\mathcal{N}^{N}_{\Omega_{2}}(\lambda)>\frac{C_{d_{1}+d_{2}}}{C_{d_{1}}}|\Omega_{2}|\lambda^{\frac{d_{2}}{2}}.$
Thus if $a^{-2}\mu_{1}(\Omega_{1})>\lambda>C_{2}(\Omega_{2})$, one gets
$\mathcal{N}^{N}_{a\Omega_{1}\times\Omega_{2}}(\lambda)=\mathcal{N}^{N}_{\Omega_{2}}(\lambda)>\frac{C_{d_{1}+d_{2}}}{C_{d_{1}}}|\Omega_{2}|\lambda^{\frac{d_{2}}{2}}\geq
C_{d_{1}+d_{2}}a^{d_{1}}|\Omega_{1}||\Omega_{2}|\lambda^{\frac{d_{1}+d_{2}}{2}}.$
For $0<\lambda\leq C_{2}(\Omega_{2})$, we only need to require
$a<\big{(}C_{d_{1}+d_{2}}|\Omega_{1}||\Omega_{2}|C_{2}(\Omega_{2})^{\frac{d_{1}+d_{2}}{2}}\big{)}^{-\frac{1}{d_{1}}}=:C(\Omega_{1},\Omega_{2}),$
to get
$C_{d_{1}+d_{2}}a^{d_{1}}|\Omega_{1}||\Omega_{2}|\lambda^{\frac{d_{1}+d_{2}}{2}}<1\leq\mathcal{N}^{N}_{a\Omega_{1}\times\Omega_{2}}(\lambda).$
It remains to consider the case $a^{-2}\mu_{1}(\Omega_{1})\leq\lambda\leq
a^{-2}\max(C_{1}(\Omega_{1}),C_{2}(\Omega_{1}))$ assuming
$\max(C_{1}(\Omega_{1}),C_{2}(\Omega_{1}))>\mu_{1}(\Omega_{1})$. Let
$\mu=a^{2}\lambda$, then by Corollary 2.4, one has
(3.11)
$K_{1}(\Omega_{1})=\inf_{\mu_{1}(\Omega_{1})\leq\mu\leq\max(C_{1}(\Omega_{1}),C_{2}(\Omega_{1}))}\frac{\underset{\mu_{l}(\Omega_{1})<\mu}{\sum}(\mu-\mu_{l}(\Omega_{1}))^{\frac{d_{2}}{2}}-L_{\frac{d_{2}}{2},d_{1}}|\Omega_{1}|\mu^{\frac{d_{1}+d_{2}}{2}}}{\mu^{\frac{d_{1}+d_{2}-1}{2}}}>0.$
Let
(3.12)
$K_{2}(\Omega_{1}):=\sup_{\mu_{1}(\Omega_{1})\leq\mu\leq\max(C_{1}(\Omega_{1}),C_{2}(\Omega_{1}))}\frac{\underset{\mu_{l}(\Omega_{1})<\mu}{\sum}(\mu-\mu_{l}(\Omega_{1}))^{\frac{d_{2}-1}{2}}}{\mu^{\frac{d_{1}+d_{2}-1}{2}}}>0.$
Then by (3.8) (3.11) and (3.12), for $a^{-2}\mu_{1}(\Omega_{1})\leq\lambda\leq
a^{-2}\max(C_{1}(\Omega_{1}),C_{2}(\Omega_{1}))$ one has
$\displaystyle\sum_{l=0}^{Y_{a}^{\lambda}}\mathcal{N}^{N}_{\Omega_{2}}(\lambda-a^{-2}\mu_{l}(\Omega_{1}))$
$\displaystyle\geq$ $\displaystyle
C_{d_{1}+d_{2}}a^{d_{1}}|\Omega_{1}||\Omega_{2}|\lambda^{\frac{d_{1}+d_{2}}{2}}+C_{d_{2}}|\Omega_{2}|K_{1}(\Omega_{1})a^{d_{1}-1}\lambda^{\frac{d_{1}+d_{2}-1}{2}}-$
$\displaystyle
C_{1}(\Omega_{2})K_{2}(\Omega_{1})a^{d_{1}}\lambda^{\frac{d_{1}+d_{2}-1}{2}}.$
Thus if we assume
$a<\frac{C_{d_{2}}|\Omega_{2}|K_{1}(\Omega_{1})}{C_{1}(\Omega_{2})K_{2}(\Omega_{1})}$
then for any $a^{-2}\mu_{1}(\Omega_{1})\leq\lambda\leq
a^{-2}\max(C_{1}(\Omega_{1}),C_{2}(\Omega_{1}))$, one gets the demanded
inequality
$\mathcal{N}^{N}_{a\Omega_{1}\times\Omega_{2}}(\lambda)\geq
C_{d_{1}+d_{2}}a^{d_{1}}|\Omega_{1}||\Omega_{2}|\lambda^{\frac{d_{1}+d_{2}}{2}}.$
Thus we conclude that for
$\displaystyle
a<\min\bigg{(}\frac{C_{d_{2}-1}|\Omega_{2}||\partial\Omega_{1}|}{5\cdot
2^{\frac{d_{1}}{2}+2}C_{1}(\Omega_{2})|\Omega_{1}|},\frac{C_{d_{2}}|\Omega_{2}|K_{1}(\Omega_{1})}{C_{1}(\Omega_{2})K_{2}(\Omega_{1})},C(\Omega_{1},\Omega_{2})\bigg{)},$
all Neumann eigenvalues of $a\Omega_{1}\times\Omega_{2}$ satisfies Pólya’s
conjecture (2.2). $\hfill\square$
## 4\. Proof of Theorem 1.2
### 4.1. Two elementary lemmas
Before proving Theorem 1.2, we give two elementary lemmas that will play
important roles later.
###### Lemma 4.1.
Let $f_{d}(x)=(\lambda-\frac{x^{2}\pi^{2}}{a^{2}})^{\frac{d}{2}}$, then
1. $\mathrm{(1)}$
$f_{d}$ is decreasing on $(0,\frac{a\sqrt{\lambda}}{\pi})$.
2. $\mathrm{(2)}$
If $d\geq 3$, $f_{d}$ is concave on
$(0,\sqrt{\frac{\lambda}{d-1}}\frac{a}{\pi})$ and is convex on
$(\sqrt{\frac{\lambda}{d-1}}\frac{a}{\pi},\frac{a\sqrt{\lambda}}{\pi})$.
3. $\mathrm{(3)}$
$\int_{0}^{\frac{a\sqrt{\lambda}}{\pi}}f_{d}(x)\mathrm{d}x=a\cdot\frac{C_{d+1}}{C_{d}}\lambda^{\frac{d+1}{2}}$.
###### Proof.
(1) is trivial. (2) follows from
$f_{d}^{\prime\prime}(x)=\frac{\pi^{2}d}{a^{2}}(\lambda-\frac{x^{2}\pi^{2}}{a^{2}})^{\frac{d}{2}-2}((d-1)\frac{\pi^{2}x^{2}}{a^{2}}-\lambda),$
and (3) is also elementary:
$\displaystyle\int_{0}^{\frac{a\sqrt{\lambda}}{\pi}}f_{d}(x)\mathrm{d}x=\lambda^{\frac{d+1}{2}}\frac{a}{\pi}\int_{0}^{1}(1-t^{2})^{\frac{d}{2}}\mathrm{d}t$
$\displaystyle=$
$\displaystyle\lambda^{\frac{d+1}{2}}\frac{a}{\pi}\int_{0}^{\frac{\pi}{2}}(\cos\theta)^{d+1}\mathrm{d}\theta=\lambda^{\frac{d+1}{2}}\frac{a}{\pi}\cdot\frac{\Gamma(\frac{1}{2})\Gamma(\frac{d}{2}+1)}{2\Gamma(\frac{d+1}{2}+1)}=a\cdot\frac{C_{d+1}}{C_{d}}\lambda^{\frac{d+1}{2}}.$
∎
The second lemma is
###### Lemma 4.2.
Let
(4.1) $M_{a}^{\lambda}=\lfloor{\frac{a\sqrt{\lambda}}{\pi}}\rfloor,$
then for $\lambda\geq\frac{\pi^{2}}{a^{2}}$ (i.e. $M_{a}^{\lambda}\geq 1$), we
have
(4.2)
$\sum_{l=1}^{M_{a}^{\lambda}}(\lambda-\frac{l^{2}\pi^{2}}{a^{2}})\leq\frac{2a\lambda^{\frac{3}{2}}}{3\pi}-\frac{\lambda}{8}-\frac{\sqrt{\lambda}\pi}{12a}$
and
(4.3)
$\sum_{l=0}^{M_{a}^{\lambda}}(\lambda-\frac{l^{2}\pi^{2}}{a^{2}})\geq\frac{2a}{3\pi}\lambda^{\frac{3}{2}}+\frac{1}{12}\lambda.$
###### Proof.
We have
$\sum_{l=1}^{M_{a}^{\lambda}}(\lambda-\frac{l^{2}\pi^{2}}{a^{2}})=\lambda
M_{a}^{\lambda}-\frac{\pi^{2}}{3a^{2}}(M_{a}^{\lambda})^{3}-\frac{\pi^{2}}{2a^{2}}(M_{a}^{\lambda})^{2}-\frac{\pi^{2}}{6a^{2}}M_{a}^{\lambda}.$
Let
(4.4) $g(x)=\lambda x-\frac{\pi^{2}}{3a^{2}}x^{3},$
then $g^{\prime}(x)=\lambda-\frac{\pi^{2}}{a^{2}}x^{2}$ which is positive if
$x\in(0,\frac{a\sqrt{\lambda}}{\pi})$. So
$g(M^{\lambda}_{a})\leq
g(\frac{a\sqrt{\lambda}}{\pi})=\frac{2a\lambda^{\frac{3}{2}}}{3\pi}.$
Combining with the fact
$\lfloor{x}\rfloor\geq\frac{x}{2},\qquad\forall x\geq 1,$
one gets (4.2).
Similarly,
$\sum_{l=0}^{M_{a}^{\lambda}}(\lambda-\frac{l^{2}\pi^{2}}{a^{2}})=\lambda+\lambda
M_{a}^{\lambda}-\frac{\pi^{2}}{3a^{2}}(M_{a}^{\lambda})^{3}-\frac{\pi^{2}}{a^{2}}\cdot\frac{3(M_{a}^{\lambda})^{2}+M_{a}^{\lambda}}{6}.$
Again consider the function $g(x)$ defined in (4.4). Since $g^{\prime}(x)$ is
positive and monotonically decreasing on $(0,\frac{a\sqrt{\lambda}}{\pi})$,
one has
$\frac{2a\lambda^{\frac{3}{2}}}{3\pi}-(\lambda
M_{a}^{\lambda}-\frac{\pi^{2}}{3a^{2}}(M_{a}^{\lambda})^{3})=g(\frac{a\sqrt{\lambda}}{\pi})-g(M_{a}^{\lambda})\leq
g^{\prime}(M_{a}^{\lambda})=\lambda-\frac{\pi^{2}}{a^{2}}(M_{a}^{\lambda})^{2},$
which implies (4.3). ∎
Now we start to prove Theorem 1.2. Again by [12, Proposition 3.1], it is
enough to assume that $\Omega$ is connected. Since any rectangle in
$\mathbb{R}^{2}$ satisfies Pólya’s conjecture, we can assume that the
dimension $d$ of $\Omega$ is at least 2. Again we argue by treating $d\geq 3$
and $d=2$ separately, and by treating Dirichlet case and Neumann case
separately.
### 4.2. The Dirichlet case with $d=2$
The Dirichlet eigenvalues of $(0,a)\times\Omega$ are
$\frac{l^{2}\pi^{2}}{a^{2}}+\lambda_{k}(\Omega),\qquad l,k\in\mathbb{Z}_{>0},$
and thus
(4.5)
$\mathcal{N}^{D}_{(0,a)\times\Omega}(\lambda)=\sum_{l=1}^{M_{a}^{\lambda}}\mathcal{N}^{D}_{\Omega}(\lambda-\frac{l^{2}\pi^{2}}{a^{2}}).$
Note that if $0<\lambda<\frac{\pi^{2}}{a^{2}}$, then
$\mathcal{N}^{D}_{(0,a)\times\Omega}(\lambda)=0<C_{3}a|\Omega|\lambda^{\frac{3}{2}}.$
So one only need to consider the case $\lambda\geq\frac{\pi^{2}}{a^{2}}$, i.e.
$M_{a}^{\lambda}\geq 1$. By inequality (2.6), for any $\lambda>0$, there
exists a constant $C(\Omega)>0$ such that
$\mathcal{N}^{D}_{\Omega}(\lambda)\leq\frac{|\Omega|}{4\pi}\lambda+C(\Omega)\lambda^{\frac{1}{2}}.$
In view of (4.2) and the fact
$\sum_{l=1}^{M_{a}^{\lambda}}(\lambda-\frac{l^{2}\pi^{2}}{a^{2}})^{\frac{1}{2}}\leq\int_{0}^{\frac{a\sqrt{\lambda}}{\pi}}(\lambda-\frac{x^{2}\pi^{2}}{a^{2}})^{\frac{1}{2}}\mathrm{d}x=\frac{a}{4}\lambda$
we get
$\displaystyle\sum_{l=1}^{M_{a}^{\lambda}}\mathcal{N}^{D}_{\Omega}(\lambda-\frac{l^{2}\pi^{2}}{a^{2}})\leq$
$\displaystyle\frac{|\Omega|}{4\pi}\sum_{l=1}^{M_{a}^{\lambda}}(\lambda-\frac{l^{2}\pi^{2}}{a^{2}})+C(\Omega)\sum_{l=1}^{M_{a}^{\lambda}}(\lambda-\frac{l^{2}\pi^{2}}{a^{2}})^{\frac{1}{2}}$
$\displaystyle\leq$
$\displaystyle\frac{a|\Omega|\lambda^{\frac{3}{2}}}{6\pi^{2}}-\frac{|\Omega|\lambda}{32\pi}+\frac{C(\Omega)a}{4}\lambda.$
Thus if we assume $a<\frac{|\Omega|}{8\pi C(\Omega)}$, then
$\mathcal{N}^{D}_{(0,a)\times\Omega}(\lambda)\leq\frac{a|\Omega|\lambda^{\frac{3}{2}}}{6\pi^{2}}=C_{3}a|\Omega|\lambda^{\frac{3}{2}}.$
This completes the proof of the Dirichlet case with $d=2$.
### 4.3. The Neumann case with $d=2$
For the Neumann case, the eigenvalues of $(0,a)\times\Omega$ are
$\frac{l^{2}\pi^{2}}{a^{2}}+\mu_{k}(\Omega),\qquad l,k\in\mathbb{Z}_{\geq 0},$
thus
(4.6)
$\mathcal{N}^{N}_{(0,a)\times\Omega}(\lambda)=\sum_{l=0}^{M_{a}^{\lambda}}\mathcal{N}^{N}_{\Omega}(\lambda-\frac{l^{2}\pi^{2}}{a^{2}}).$
By inequality (2.6), for any $\lambda>0$, there exists $C(\Omega)>0$ such that
$\mathcal{N}^{N}_{\Omega}(\lambda)\geq
C_{d}|\Omega|\lambda-C(\Omega)\lambda^{\frac{1}{2}}.$
In view of (4.3) and the fact
$\sum_{l=0}^{M_{a}^{\lambda}}(\lambda-\frac{l^{2}\pi^{2}}{a^{2}})^{\frac{1}{2}}\leq\lambda^{\frac{1}{2}}+\int_{0}^{\frac{a\sqrt{\lambda}}{\pi}}(\lambda-\frac{x^{2}\pi^{2}}{a^{2}})^{\frac{1}{2}}\mathrm{d}x=\lambda^{\frac{1}{2}}+\frac{a}{4}\lambda$
we get, for $\lambda\geq\frac{\pi^{2}}{a^{2}}$,
$\displaystyle\sum_{l=0}^{M_{a}^{\lambda}}\mathcal{N}^{N}_{\Omega}(\lambda-\frac{l^{2}\pi^{2}}{a^{2}})\geq$
$\displaystyle\frac{|\Omega|}{4\pi}\sum_{l=0}^{M_{a}^{\lambda}}(\lambda-\frac{l^{2}\pi^{2}}{a^{2}})-C(\Omega)\sum_{l=0}^{M_{a}^{\lambda}}(\lambda-\frac{l^{2}\pi^{2}}{a^{2}})^{\frac{1}{2}}$
$\displaystyle\geq$
$\displaystyle\frac{a|\Omega|\lambda^{\frac{3}{2}}}{6\pi^{2}}+\frac{|\Omega|\lambda}{48\pi}-C(\Omega)(\lambda^{\frac{1}{2}}+\frac{a\lambda}{4}).$
So if we assume $a\leq\frac{|\Omega|}{96C(\Omega)}$, then
$\lambda\geq\frac{\pi^{2}}{a^{2}}\geq(\frac{96\pi C(\Omega)}{|\Omega|})^{2}$
and thus
$\frac{C(\Omega)a\lambda}{4}\leq\frac{|\Omega|\lambda}{4\cdot
96}<\frac{|\Omega|\lambda}{96\pi}\quad\text{and}\quad
C(\Omega)\lambda^{\frac{1}{2}}\leq\frac{|\Omega|\lambda}{96\pi}.$
In other words, if we assume $a\leq\frac{|\Omega|}{96C(\Omega)}$, then for any
$\lambda\geq\frac{\pi^{2}}{a^{2}}$,
$\mathcal{N}^{N}_{(0,a)\times\Omega}(\lambda)\geq\frac{a|\Omega|\lambda^{\frac{3}{2}}}{6\pi^{2}}=C_{3}a|\Omega|\lambda^{\frac{3}{2}}.$
Next, if $0<\lambda<\frac{\pi^{2}}{a^{2}}$, then
$\mathcal{N}^{N}_{(0,a)\times\Omega}(\lambda)=\mathcal{N}^{N}_{\Omega}(\lambda),\quad\text{
and }\quad C_{3}a|\Omega|\lambda^{\frac{3}{2}}<C_{3}\pi|\Omega|\lambda.$
By (2.4), one has $C_{3}\pi<C_{2}$. So by (2.5), there exists a constant
$C_{1}(\Omega)>0$, such that if $\lambda\geq C_{1}(\Omega)$, then
$\mathcal{N}^{N}_{\Omega}(\lambda)>C_{3}\pi|\Omega|\lambda.$
Thus for $C_{1}(\Omega)\leq\lambda<\frac{\pi^{2}}{a^{2}}$, one gets
$\mathcal{N}^{N}_{(0,a)\times\Omega}(\lambda)=\mathcal{N}^{N}_{\Omega}(\lambda)>C_{3}\pi|\Omega|\lambda>C_{3}a|\Omega|\lambda^{\frac{3}{2}}.$
Finally for $0<\lambda\leq C_{1}(\Omega)$, we may require
$a\leq(C_{3}|\Omega|)^{-1}C_{1}(\Omega)^{-\frac{3}{2}}$ to get
$C_{3}a|\Omega|\lambda^{\frac{3}{2}}\leq
1\leq\mathcal{N}^{N}_{(0,a)\times\Omega}(\lambda).$
Combining all discussions above, one has that if
(4.7)
$a<\min\bigg{(}\frac{|\Omega|}{96C(\Omega)},(C_{3}|\Omega|)^{-1}C_{1}(\Omega)^{-\frac{3}{2}}\bigg{)},$
then all Neumann eigenvalues of $(0,a)\times\Omega$ satisfy Pólya’s conjecture
(1.3). Thus we complete the proof of Theorem 1.2 with $d=2$.
### 4.4. The Dirichlet case with $d\geq 3$
Again we have
$\mathcal{N}^{D}_{(0,a)\times\Omega}(\lambda)=\sum_{l=1}^{M_{a}^{\lambda}}\mathcal{N}^{D}_{\Omega}(\lambda-\frac{l^{2}\pi^{2}}{a^{2}})$
and there exists a constant $C(\Omega)>0$ such that
$\mathcal{N}^{D}_{\Omega}(\lambda)\leq
C_{d}|\Omega|\lambda^{\frac{d}{2}}+C(\Omega)\lambda^{\frac{d-1}{2}}.$
So we get
(4.8) $\mathcal{N}^{D}_{(0,a)\times\Omega}(\lambda)\leq
C_{d}|\Omega|\sum_{l=1}^{M_{a}^{\lambda}}f_{d}(l)+C(\Omega)\sum_{l=1}^{M_{a}^{\lambda}}f_{d-1}(l),$
where $f_{d}$ is defined in Lemma 4.1. We split the first sum into two parts.
Denote
$N_{a}^{\lambda}=\lfloor\sqrt{\frac{\lambda}{d-1}}\frac{a}{\pi}\rfloor.$
By concavity of $f_{d}$ (see (2) of Lemma 4.1), one has
$\int_{0}^{N_{a}^{\lambda}}f_{d}(x)\mathrm{d}x-\sum_{l=1}^{N_{a}^{\lambda}}f_{d}(l)\geq\sum_{l=0}^{N_{a}^{\lambda}-1}f_{d}(l)-\int_{0}^{N_{a}^{\lambda}}f_{d}(x)\mathrm{d}x$
which implies
$\sum_{l=1}^{N_{a}^{\lambda}}f_{d}(l)\leq\int_{0}^{N_{a}^{\lambda}}f_{d}(x)\mathrm{d}x-\frac{1}{2}\bigg{(}\lambda^{\frac{d}{2}}-f_{d}(N_{a}^{\lambda})\bigg{)}.$
First consider $\lambda\geq\frac{(d-1)\pi^{2}}{a^{2}}$, in which case
$N_{a}^{\lambda}\geq\frac{1}{2}\frac{a}{\pi}\sqrt{\frac{\lambda}{d-1}}$, and
thus
$f_{d}(N_{a}^{\lambda})\leq
f_{d}(\frac{1}{2}\frac{a}{\pi}\sqrt{\frac{\lambda}{d-1}})=(\frac{4d-5}{4d-4})^{\frac{d}{2}}\lambda^{\frac{d}{2}}.$
For simplicity, we denote
$A_{d}=\frac{1}{2}\bigg{(}1-\big{(}\frac{4d-5}{4d-4}\big{)}^{\frac{d}{2}}\bigg{)}C_{d}|\Omega|.$
Then we get, for $\lambda\geq\frac{(d-1)\pi^{2}}{a^{2}}$,
$\displaystyle\mathcal{N}^{D}_{(0,a)\times\Omega}(\lambda)\leq$ $\displaystyle
C_{d}|\Omega|\sum_{l=1}^{N_{a}^{\lambda}}f_{d}(l)+C_{d}|\Omega|\sum_{l=N_{a}^{\lambda}+1}^{M_{a}^{\lambda}}f_{d}(l)+C(\Omega)\sum_{l=1}^{M_{a}^{\lambda}}f_{d-1}(l)$
$\displaystyle\leq$ $\displaystyle
C_{d}|\Omega|\int_{0}^{{\frac{a\sqrt{\lambda}}{\pi}}}f_{d}(x)\mathrm{d}x-A_{d}\lambda^{\frac{d}{2}}+C(\Omega)\int_{0}^{{\frac{a\sqrt{\lambda}}{\pi}}}f_{d-1}(x)\mathrm{d}x$
$\displaystyle=$ $\displaystyle
C_{d+1}a|\Omega|\lambda^{\frac{d+1}{2}}-A_{d}\lambda^{\frac{d}{2}}+C(\Omega)\frac{C_{d}a}{C_{d-1}}\lambda^{\frac{d}{2}}.$
So if we assume $a\leq\frac{A_{d}\cdot C_{d-1}}{C(\Omega)\cdot C_{d}}$, then
for any $\lambda\geq\frac{(d-1)\pi^{2}}{a^{2}_{1}}$,
(4.9) $\mathcal{N}^{D}_{(0,a)\times\Omega}(\lambda)\leq
C_{d+1}a|\Omega|\lambda^{\frac{d+1}{2}}.$
Note that if $0<\lambda<\frac{\pi^{2}}{a^{2}}$, then we automatically have
$\mathcal{N}^{D}_{(0,a)\times\Omega}(\lambda)=0<C_{d+1}a|\Omega|\lambda^{\frac{d+1}{2}},$
so it remains to consider the case
$\frac{\pi^{2}}{a^{2}}\leq\lambda<\frac{(d-1)\pi^{2}}{a^{2}}$. Let
$\mu=\frac{\lambda a^{2}}{\pi^{2}}$, then $1\leq\mu<d-1$. Let
$H_{1}:=\inf_{1\leq\mu<d-1}\frac{\int^{\sqrt{\mu}}_{0}(\mu-x^{2})^{\frac{d}{2}}\mathrm{d}x-\sum_{0<l^{2}<\mu}(\mu-l^{2})^{\frac{d}{2}}}{\mu^{\frac{d}{2}}}>0,$
then
$\displaystyle\mathcal{N}^{D}_{(0,a)\times\Omega}(\lambda)\leq$ $\displaystyle
C_{d}|\Omega|\int_{0}^{{\frac{a\sqrt{\lambda}}{\pi}}}f_{d}(x)\mathrm{d}x-C_{d}|\Omega|H_{1}\lambda^{\frac{d}{2}}+C(\Omega)\int_{0}^{{\frac{a\sqrt{\lambda}}{\pi}}}f_{d-1}(x)\mathrm{d}x$
$\displaystyle=$ $\displaystyle
C_{d+1}a|\Omega|\lambda^{\frac{d+1}{2}}-C_{d}|\Omega|H_{1}\lambda^{\frac{d}{2}}+C(\Omega)\frac{C_{d}a}{C_{d-1}}\lambda^{\frac{d}{2}}.$
So if we assume $a\leq\frac{C_{d-1}|\Omega|H_{1}}{C(\Omega)}$, then for any
$\frac{\pi^{2}}{a^{2}}\leq\lambda<\frac{(d-1)\pi^{2}}{a^{2}}$,
$\mathcal{N}^{D}_{(0,a)\times\Omega}(\lambda)\leq
C_{d+1}a|\Omega|\lambda^{\frac{d+1}{2}}.$
Combining with (4.9), one gets that if
$a\leq\min\bigg{(}\frac{A_{d}\cdot C_{d-1}}{C(\Omega)\cdot
C_{d}},\frac{C_{d-1}|\Omega|H_{1}}{C(\Omega)}\bigg{)},$
then all Dirichlet eigenvalues of $(0,a)\times\Omega$ satisfy Pólya’s
conjecture (1.2).
### 4.5. The Neumann case with $d\geq 3$
Again we have
$\mathcal{N}^{N}_{(0,a)\times\Omega}(\lambda)=\sum_{l=0}^{M_{a}^{\lambda}}\mathcal{N}^{N}_{\Omega}(\lambda-\frac{l^{2}\pi^{2}}{a^{2}})$
and there exists a constant $C(\Omega)>0$ such that
$\mathcal{N}^{N}_{\Omega}(\lambda)\geq
C_{d}|\Omega|\lambda^{\frac{d}{2}}-C(\Omega)\lambda^{\frac{d-1}{2}},\qquad\forall\lambda>0.$
So we get
(4.10) $\mathcal{N}^{N}_{(0,a)\times\Omega}(\lambda)\geq
C_{d}|\Omega|\sum_{l=0}^{M_{a}^{\lambda}}f_{d}(l)-C(\Omega)\sum_{l=0}^{M_{a}^{\lambda}}f_{d-1}(l).$
By convexity of $f_{d}$ (see (2) of Lemma 4.1), one has
$\sum_{l=N_{a}^{\lambda}+1}^{M_{a}^{\lambda}}f_{d}(l)-\int_{N_{a}^{\lambda}+1}^{\frac{a\sqrt{\lambda}}{\pi}}f_{d}(x)\mathrm{d}x\geq\int_{N_{a}^{\lambda}+1}^{\frac{a\sqrt{\lambda}}{\pi}}f_{d}(x)\mathrm{d}x-\sum_{l=N_{a}^{\lambda}+2}^{M_{a}^{\lambda}}f_{d}(l)$
which implies
$\sum_{l=N_{a}^{\lambda}+1}^{M_{a}^{\lambda}}f_{d}(l)\geq\int_{N_{a}^{\lambda}+1}^{\frac{a\sqrt{\lambda}}{\pi}}f_{d}(x)\mathrm{d}x+\frac{1}{2}f_{d}(N_{a}^{\lambda}+1).$
If $\lambda\geq\frac{9\pi^{2}(d-1)}{a^{2}}$, then $N_{a}^{\lambda}\geq 3$ and
thus $N_{a}^{\lambda}+1\leq\frac{4a}{3\pi}\sqrt{\frac{\lambda}{d-1}}$, which
implies
$f_{d}(N_{a}^{\lambda}+1)\geq
f_{d}(\frac{4a}{3\pi}\sqrt{\frac{\lambda}{d-1}})\geq
3^{-{d}}\lambda^{\frac{d}{2}},$
where we used $d\geq 3$. For simplicity, we denote
$B_{d}=\frac{1}{2}3^{-d}C_{d}|\Omega|.$
Then for $\lambda\geq\frac{9\pi^{2}(d-1)}{a^{2}}$,
$\displaystyle\mathcal{N}^{N}_{(0,a)\times\Omega}(\lambda)\geq$ $\displaystyle
C_{d}|\Omega|\sum_{l=0}^{N_{a}^{\lambda}}f_{d}(l)+C_{d}|\Omega|\sum_{l=N_{a}^{\lambda}+1}^{M_{a}^{\lambda}}f_{d}(l)-C(\Omega)\sum_{l=0}^{M_{a}^{\lambda}}f_{d-1}(l)$
$\displaystyle\geq$ $\displaystyle
C_{d}|\Omega|\int_{0}^{\frac{a\sqrt{\lambda}}{\pi}}\\!\\!\\!f_{d}(x)\mathrm{d}x+B_{d}\lambda^{\frac{d}{2}}-C(\Omega)\big{(}\lambda^{\frac{d-1}{2}}+\int_{0}^{{\frac{a\sqrt{\lambda}}{\pi}}}\\!\\!\\!f_{d-1}(x)\mathrm{d}x\big{)}$
$\displaystyle=$ $\displaystyle
C_{d+1}a|\Omega|\lambda^{\frac{d+1}{2}}+B_{d}\lambda^{\frac{d}{2}}-C(\Omega)\big{(}\lambda^{\frac{d-1}{2}}+\frac{C_{d}a}{C_{d-1}}\lambda^{\frac{d}{2}}\big{)}.$
So if we assume
$a\leq\min\bigg{(}\frac{B_{d}C_{d-1}}{2C(\Omega)C_{d}},\frac{B_{d}3\pi\sqrt{d-1}}{2C(\Omega)}\bigg{)},$
then
$\lambda\geq\frac{9\pi^{2}(d-1)}{a^{2}}\geq\frac{4C(\Omega)^{2}}{B_{d}^{2}}$
and thus
$\frac{C(\Omega)C_{d}a}{C_{d-1}}\lambda^{\frac{d}{2}}\leq\frac{1}{2}B_{d}\lambda^{\frac{d}{2}}\quad\text{and}\quad
C(\Omega)\lambda^{\frac{d-1}{2}}\leq\frac{1}{2}B_{d}\lambda^{\frac{d}{2}}.$
Thus for any $\lambda\geq\frac{9\pi^{2}(d-1)}{a^{2}}$, one gets the demanded
inequality
$\mathcal{N}^{N}_{(0,a)\times\Omega}(\lambda)\geq
C_{d+1}a|\Omega|\lambda^{\frac{d+1}{2}}.$
Next if $\frac{\pi^{2}}{a^{2}}\leq\lambda\leq\frac{9\pi^{2}(d-1)}{a^{2}}$, let
$\mu=\frac{\lambda a^{2}}{\pi^{2}}$, then $1\leq\mu\leq 9(d-1)$. Let
$H_{2}=\inf_{1\leq\mu\leq 9(d-1)}\frac{\sum_{0\leq
l^{2}<\mu}(\mu-l^{2})^{\frac{d}{2}}-\int_{0}^{\sqrt{\mu}}(\mu-x^{2})^{\frac{d}{2}}\mathrm{d}x}{\mu^{\frac{d}{2}}}>0,$
then
$\displaystyle\mathcal{N}^{N}_{(0,a)\times\Omega}(\lambda)\geq$ $\displaystyle
C_{d}|\Omega|\int_{0}^{\frac{a\sqrt{\lambda}}{\pi}}\\!\\!\\!f_{d}(x)\mathrm{d}x+C_{d}|\Omega|H_{2}\lambda^{\frac{d}{2}}-C(\Omega)\big{(}\lambda^{\frac{d-1}{2}}+\int_{0}^{{\frac{a\sqrt{\lambda}}{\pi}}}\\!\\!\\!f_{d-1}(x)\mathrm{d}x\big{)}$
$\displaystyle=$ $\displaystyle
C_{d+1}a|\Omega|\lambda^{\frac{d+1}{2}}+C_{d}|\Omega|H_{2}\lambda^{\frac{d}{2}}-C(\Omega)\big{(}\lambda^{\frac{d-1}{2}}+\frac{C_{d}a}{C_{d-1}}\lambda^{\frac{d}{2}}\big{)}.$
Similar to the case $\lambda\geq\frac{9\pi^{2}(d-1)}{a^{2}}$, one can prove
that if we take $a\leq\min\big{(}\frac{|\Omega|H_{2}}{2C_{d-1}},\frac{\pi
C_{d}|\Omega|H_{2}}{2C(\Omega)}\big{)}$, then for any
$\frac{\pi^{2}}{a^{2}}\leq\lambda\leq\frac{9\pi^{2}(d-1)}{a^{2}}$, one gets
$\mathcal{N}^{N}_{(0,a)\times\Omega}(\lambda)\geq
C_{d+1}a|\Omega|\lambda^{\frac{d+1}{2}}.$
Finally for $0<\lambda<\frac{\pi^{2}}{a^{2}}$, we just repeat the
corresponding part in the proof of the Neumann case with $d=2$. In conclusion,
we get: if
$a<\min\bigg{(}\frac{B_{d}C_{d-1}}{2C(\Omega)C_{d}},\frac{B_{d}3\pi\sqrt{d-1}}{2C(\Omega)},\frac{|\Omega|H_{2}}{2C_{d-1}},\frac{\pi
C_{d}|\Omega|H_{2}}{2C(\Omega)},(C_{d+1}|\Omega|)^{-1}C_{1}(\Omega)^{-\frac{d+1}{2}}\bigg{)},$
then all Neumann eigenvalues of $(0,a)\times\Omega$ satisfy Pólya’s conjecture
(1.3). So we complete the proof of Theorem 1.2. $\hfill\square$
## 5\. Proof of Theorem 1.3
Again by [12, Proposition 3.1], it is enough to assume that both $\Omega$ and
$M$ are connected. Let the eigenvalues of $M$ be
$0=\lambda_{0}(M)<\lambda_{1}(M)\leq\cdots\nearrow\infty,$
and the counting functions for the eigenvalues of $M$ be
$\mathcal{N}_{M}(\lambda)=\\#\\{n|\ \lambda_{n}(M)<\lambda\\}.$
B. M. Levitan ([22]) and V. G. Avakumović ([1]) proved that
$\mathcal{N}_{M}(\lambda)=C_{d_{2}}|M|\lambda^{\frac{d_{2}}{2}}+\mathrm{O}(\lambda^{\frac{d-1}{2}}),\text{
as }\lambda\to\infty.$
So there exists a constant $C(M)>0$ such that
(5.1) $\mathcal{N}_{M}(\lambda)\geq
C_{d_{2}}|M|\lambda^{\frac{d_{2}}{2}}-C(M)\lambda^{\frac{d_{2}-1}{2}},\
\forall\lambda>0.$
Repeating the proof of the Neumann case of Theorem 1.1 and Theorem 1.2 word by
word, one can easily prove the Neumann case of Theorem 1.3.
For the Dirichlet case of Theorem 1.3, since 0 is an eigenvalue of $M$, one
can only get that there exists a constant $C_{1}(M)>0$ such that
(5.2) $\mathcal{N}_{M}(\lambda)\leq
C_{d_{2}}|M|\lambda^{\frac{d_{2}}{2}}+C_{1}(M)\lambda^{\frac{d_{2}-1}{2}}+1,\
\forall\lambda>0.$
So to prove the Dirichlet case of Theorem 1.3, one need to carefully handle
this extra number. Again we divided the proof into three parts: the Dirichlet
case with $d_{1}=1$ and $d_{2}=2$, the Dirichlet case with $d_{1}=1$ and
$d_{2}\geq 3$, and the Dirichlet case with $d_{1}\geq 2$.
### 5.1. The Dirichlet case with $d_{1}=1$ and $d_{2}=2$
When $d_{1}=1$, we can assume $\Omega=(0,1)$ for simplicity. The Dirichlet
eigenvalues of $(0,a)\times M$ are
$\frac{l^{2}\pi^{2}}{a^{2}}+\lambda_{k}(M),\qquad l\in\mathbb{Z}_{>0},\
k\in\mathbb{Z}_{\geq 0}.$
If $0<\lambda<\frac{\pi^{2}}{a^{2}}$, then $\mathcal{N}^{D}_{(0,a)\times
M}(\lambda)=0<C_{3}a|M|\lambda^{\frac{3}{2}}$. For
$\lambda\geq\frac{\pi^{2}}{a^{2}}$, by (4.2) we get
$\displaystyle\mathcal{N}^{D}_{(0,a)\times M}(\lambda)$
$\displaystyle=\sum_{l=1}^{M^{\lambda}_{a}}\mathcal{N}_{M}(\lambda-\frac{l^{2}\pi^{2}}{a^{2}})$
$\displaystyle\leq\frac{|M|}{4\pi}\sum_{l=1}^{M^{\lambda}_{a}}(\lambda-\frac{l^{2}\pi^{2}}{a^{2}})+C_{1}(M)\sum_{l=1}^{M^{\lambda}_{a}}(\lambda-\frac{l^{2}\pi^{2}}{a^{2}})^{\frac{1}{2}}+M^{\lambda}_{a}$
$\displaystyle\leq\frac{|M|}{4\pi}(\frac{2a\lambda^{\frac{3}{2}}}{3\pi}-\frac{\lambda}{8}-\frac{\sqrt{\lambda}\pi}{12a})+\frac{C_{1}(M)a}{4}\lambda+\frac{a\sqrt{\lambda}}{\pi}.$
Note that if $a\leq\sqrt{\frac{|M|\pi}{48}}$, then
$\frac{|M|}{4\pi}(\frac{2a\lambda^{\frac{3}{2}}}{3\pi}-\frac{\lambda}{8}-\frac{\sqrt{\lambda}\pi}{12a})+\frac{C_{1}(M)a}{4}\lambda+\frac{a\sqrt{\lambda}}{\pi}\leq\frac{a|M|\lambda^{\frac{3}{2}}}{6\pi^{2}}-\frac{|M|\lambda}{32\pi}+\frac{C_{1}(M)a}{4}\lambda.$
Thus we proved: if
(5.3) $a\leq\min\big{(}\sqrt{\frac{|M|\pi}{48}},\frac{|M|}{8\pi
C_{1}(M)}\big{)},$
then all Dirichlet eigenvalues of $(0,a)\times M$ satisfy Pólya’s conjecture
(1.2).
### 5.2. The Dirichlet case with $d_{1}=1$ and $d_{2}\geq 3$
We still have
$\displaystyle\mathcal{N}^{D}_{(0,a)\times M}(\lambda)$
$\displaystyle=\sum_{l=1}^{M^{\lambda}_{a}}\mathcal{N}_{M}(\lambda-\frac{l^{2}\pi^{2}}{a^{2}})$
$\displaystyle\leq
C_{d_{2}}|M|\sum_{l=1}^{M^{\lambda}_{a}}(\lambda-\frac{l^{2}\pi^{2}}{a^{2}})^{\frac{d_{2}}{2}}+C_{1}(M)\sum_{l=1}^{M^{\lambda}_{a}}(\lambda-\frac{l^{2}\pi^{2}}{a^{2}})^{\frac{d_{2}-1}{2}}+M^{\lambda}_{a}.$
As in §4.4, if $\lambda\geq\frac{(d_{2}-1)\pi^{2}}{a^{2}}$, one has
$\mathcal{N}^{D}_{(0,a)\times M}(\lambda)\leq
C_{d_{2}+1}a|M|\lambda^{\frac{d_{2}+1}{2}}-A_{d_{2}}\lambda^{\frac{d_{2}}{2}}+C_{1}(M)\frac{C_{d_{2}}a}{C_{d_{2}-1}}\lambda^{\frac{d_{2}}{2}}+\frac{a\sqrt{\lambda}}{\pi},$
where
$A_{d_{2}}=\frac{1}{2}(1-(\frac{4d_{2}-5}{4d_{2}-4})^{\frac{d_{2}}{2}})C_{d_{2}}|M|$.
To control the last term, we require
$a<\big{(}\frac{\pi}{2}A_{d_{2}}\big{)}^{\frac{1}{d_{2}}}\big{(}\pi^{2}(d_{2}-1)\big{)}^{\frac{d_{2}-1}{2d_{2}}}$
to get
$\frac{a\sqrt{\lambda}}{\pi}\leq\frac{1}{2}A_{d_{2}}\lambda^{\frac{d_{2}}{2}}$
for all $\lambda\geq\frac{(d_{2}-1)\pi^{2}}{a^{2}}$. Repeating §4.4, we will
get: if
$a<\min\bigg{(}\pi\big{(}\frac{1}{2}A_{d_{2}}\big{)}^{\frac{1}{d_{2}}}(d_{2}-1)^{\frac{d_{2}-1}{2d_{2}}},\frac{A_{d_{2}}C_{d_{2}-1}}{2C_{1}(M)C_{d_{2}}}\bigg{)},$
then for any $\lambda\geq\frac{(d_{2}-1)\pi^{2}}{a^{2}}$,
$\mathcal{N}^{D}_{(0,a)\times M}(\lambda)\leq
C_{d_{2}+1}a|M|\lambda^{\frac{d_{2}+1}{2}}.$
For $\lambda<\frac{(d_{2}-1)\pi^{2}}{a^{2}}$, again one only needs to consider
$\frac{\pi^{2}}{a^{2}}\leq\lambda<\frac{(d_{2}-1)\pi^{2}}{a^{2}}$. As in §4.4,
in this case one has
$\mathcal{N}^{D}_{(0,a)\times M}(\lambda)\leq
C_{d_{2}+1}a|M|\lambda^{\frac{d_{2}+1}{2}}-C_{d_{2}}|M|H_{1}\lambda^{\frac{d_{2}}{2}}+C_{1}(M)\frac{C_{d_{2}}a}{C_{d_{2}-1}}\lambda^{\frac{d_{2}}{2}}+\frac{a\sqrt{\lambda}}{\pi},$
where
$H_{1}:=\inf_{1\leq\mu\leq
d_{2}-1}\frac{\int^{\sqrt{\mu}}_{0}(\mu-x^{2})^{\frac{d_{2}}{2}}\mathrm{d}x-\sum_{0<l^{2}<\mu}(\mu-l^{2})^{\frac{d_{2}}{2}}}{\mu^{\frac{d_{2}}{2}}}>0.$
So if we assume
$a<\min\bigg{(}\pi\big{(}\frac{1}{2}C_{d_{2}}|M|H_{1}\big{)}^{\frac{1}{d_{2}}},\frac{C_{d_{2}-1}|M|H_{1}}{2C_{1}(M)}\bigg{)},$
then for all
$\frac{\pi^{2}}{a^{2}}\leq\lambda<\frac{(d_{2}-1)\pi^{2}}{a^{2}}$, one has
$\mathcal{N}^{D}_{(0,a)\times M}(\lambda)\leq
C_{d_{2}+1}a|M|\lambda^{\frac{d_{2}+1}{2}}.$
Thus if $d_{2}\geq 3$ and
$a<\min\bigg{(}\pi\big{(}\frac{1}{2}A_{d_{2}}\big{)}^{\frac{1}{d_{2}}}(d_{2}-1)^{\frac{d_{2}-1}{2d_{2}}},\frac{A_{d_{2}}C_{d_{2}-1}}{2C_{1}(M)C_{d_{2}}},\pi\big{(}\frac{1}{2}C_{d_{2}}|M|H_{1}\big{)}^{\frac{1}{d_{2}}},\frac{C_{d_{2}-1}|M|H_{1}}{2C_{1}(M)}\bigg{)},$
then all Dirichlet eigenvalues of $(0,a)\times M$ satisfy Pólya’s Conjecture
(1.2).
### 5.3. The Dirichlet case with $d_{1}\geq 2$
The Dirichlet eigenvalues of $a\Omega\times M$ are
$a^{-2}\lambda_{l}(\Omega)+\lambda_{k}(M),\qquad l\in\mathbb{Z}_{>0},\
k\in\mathbb{Z}_{\geq 0},$
which implies
$\mathcal{N}^{D}_{a\Omega\times M}(\lambda)\leq
C_{d_{2}}|M|a^{-d_{2}}\sum_{l=1}^{Z^{\lambda}_{a}}(a^{2}\lambda-\lambda_{l}(\Omega))^{\frac{d_{2}}{2}}+C_{1}(M)a^{1-d_{2}}\sum_{l=1}^{Z^{\lambda}_{a}}(a^{2}\lambda-\lambda_{l}(\Omega))^{\frac{d_{2}-1}{2}}+Z^{\lambda}_{a}$
where $Z^{\lambda}_{a}=\mathcal{N}^{D}_{\Omega}(a^{2}\lambda)$. To control the
extra $Z^{\lambda}_{a}$, we use Li-Yau’s estimate (1.5) to get
$\mathcal{N}^{D}_{\Omega}(\lambda)\leq\big{(}\frac{d_{1}+2}{d_{1}}\big{)}^{\frac{d_{1}}{2}}C_{d_{1}}|\Omega|\lambda^{\frac{d_{1}}{2}},\qquad\forall\lambda>0.$
Thus for all $a>0$ and $\lambda>0$,
$Z^{\lambda}_{a}=\mathcal{N}^{D}_{\Omega}(a^{2}\lambda)\leq\big{(}\frac{d_{1}+2}{d_{1}}\big{)}^{\frac{d_{1}}{2}}C_{d_{1}}|\Omega|a^{d_{1}}\lambda^{\frac{d_{1}}{2}}.$
If $d_{2}\geq 3$, then as in §3.1, there exists a constant $C(\Omega)>0$ such
that for $a^{2}\lambda>C(\Omega)$,
$\displaystyle\mathcal{N}^{D}_{a\Omega\times M}(\lambda)\leq$ $\displaystyle
C_{d_{1}+d_{2}}a^{d_{1}}|\Omega||M|\lambda^{\frac{d_{1}+d_{2}}{2}}-\frac{1}{5}C_{d_{1}+d_{2}-1}|\partial\Omega||M|a^{d_{1}-1}\lambda^{\frac{d_{1}+d_{2}-1}{2}}+$
$\displaystyle
L_{\frac{d_{2}-1}{2},d_{1}}|\Omega|C_{1}(M)a^{d_{1}}\lambda^{\frac{d_{1}+d_{2}-1}{2}}+(\frac{d_{1}+2}{d_{1}})^{\frac{d_{1}}{2}}C_{d_{1}}|\Omega|a^{d_{1}}\lambda^{\frac{d_{1}}{2}}.$
So if we assume
$a<\big{(}\frac{C_{d_{1}+d_{2}-1}|\partial\Omega||M|}{10C_{d_{1}}|\Omega|}\big{)}^{\frac{1}{d_{2}}}\big{(}\frac{d_{1}+2}{d_{1}}\big{)}^{-\frac{1}{2}}C(\Omega)^{\frac{d_{2}-1}{2d_{2}}},$
then for any $\lambda>a^{-2}C(\Omega)$, the extra term is controlled by
$(\frac{d_{1}+2}{d_{1}})^{\frac{d_{1}}{2}}C_{d_{1}}|\Omega|a^{d_{1}}\lambda^{\frac{d_{1}}{2}}\leq\frac{1}{10}C_{d_{1}+d_{2}-1}|\partial\Omega||M|a^{d_{1}-1}\lambda^{\frac{d_{1}+d_{2}-1}{2}}.$
If $d_{2}=2$, then as in §3.2, there exists a constant $C(\Omega)>0$ such that
if $a^{2}\lambda>C(\Omega)$, one has
$\displaystyle\mathcal{N}^{D}_{a\Omega\times M}(\lambda)\leq$ $\displaystyle
C_{d_{1}+2}a^{d_{1}}|\Omega||M|\lambda^{\frac{d_{1}+2}{2}}-\frac{1}{5}C_{d_{1}+1}|M||\partial\Omega|a^{d_{1}-1}\lambda^{\frac{d_{1}+1}{2}}+$
$\displaystyle\big{(}\frac{d_{1}+2}{d_{1}}\big{)}^{\frac{d_{1}}{2}}C_{1}(M)L_{\frac{1}{2},d_{1}}|\Omega|a^{d_{1}}\lambda^{\frac{d_{1}+1}{2}}+(\frac{d_{1}+2}{d_{1}})^{\frac{d_{1}}{2}}C_{d_{1}}|\Omega|a^{d_{1}}\lambda^{\frac{d_{1}}{2}}.$
and similarly we can control the last term.
The rest of the proof for both cases are identically the same as before, and
thus will be omitted. $\hfill\square$
### 5.4. An abstract extension
As we have seen, although the upper bound given by (5.2) is a bit weaker than
(2.6), the extra term 1 can be controlled. Of course one may replace 1 by
other number.
More generally, one may start with two increasing sequence
$0<s_{1}\leq s_{2}\leq\cdots+\infty,\qquad t_{1}\leq
t_{2}\leq\cdots\to+\infty$
and study the new increasing sequence
$\\{\nu_{k}(a)\\}_{k=1}^{\infty}=\\{a^{-2}s_{m}+t_{n}\\}$. As usual we will
denote
$\mathcal{N}_{(s_{k})}(\lambda)=\\#\\{k|s_{k}\leq\lambda\\}$
and likewise for $\mathcal{N}_{(t_{k})}(\lambda)$. By using the same idea and
modifying the proof above slightly, it is easy to prove
###### Theorem 5.1.
Suppose there exist constants $V_{t},B_{1},B_{2}>0,d\geq 2$ such that
(5.4) $\mathcal{N}_{(t_{k})}(\lambda)\leq
V_{t}C_{d}\lambda^{\frac{d}{2}}+B_{1}\lambda^{\frac{d-1}{2}}+B_{2},\qquad\forall\lambda,$
and suppose either $s_{k}=\pi^{2}k^{2}(k\geq 1)$ (in which case we take
$V_{s}=1$, $d^{\prime}=1$ below), or there exist $V_{s}>0$ and $d^{\prime}\geq
2$ such that
$\sum_{s_{k}<\lambda}(\lambda-
s_{k})<L_{1,{d^{\prime}}}V_{s}\lambda^{\frac{d^{\prime}}{2}+1},\qquad\forall\lambda>0,$
and there exist $C^{\prime}>0$ and $C_{s}>0$ such that for all
$\lambda>C_{s}$,
$\sum_{s_{k}<\lambda}(\lambda-s_{k})\leq
L_{1,d^{\prime}}V_{s}\lambda^{\frac{d^{\prime}}{2}+1}-C^{\prime}\lambda^{\frac{d^{\prime}+1}{2}},$
then there exists $a_{0}>0$ such that for any $0<a<a_{0}$,
$\nu_{k}(a)\geq\frac{4\pi^{2}}{(\omega_{d+d^{\prime}}a^{d^{\prime}}V_{s}V_{t})^{\frac{2}{d+d^{\prime}}}}k^{\frac{2}{d+d^{\prime}}},\qquad\forall
k\geq 1.$
Similarly one may write down an abstract version that extends the results for
the Neumann eigenvalues above, in which case one may relax the condition on
$\mathcal{N}_{(t_{k})}(\lambda)$ to
(5.5) $\mathcal{N}_{(t_{k})}(\lambda)\geq
V_{t}C_{d}\lambda^{\frac{d}{2}}-B_{1}\lambda^{\frac{d-1}{2}},\qquad\forall\lambda>0,$
and pose suitable conditions on $(s_{k})$ (including a Szegö-Weinberger type
condition on $s_{1}$).
As a consequence, we could get a bunch of eigenvalue problems that satisfies
Pólya inequalities. For example, let $(M,g)$ be a compact Riemannian manifold
of dimension $d\geq 2$, with piecewise smooth boundary $\partial M$. Let $(H)$
be certain boundary condition so that the Laplace-Beltrami operator on $(M,g)$
has discrete spectrum. As usual we denote the corresponding eigenvalue
counting function by $\mathcal{N}^{(H)}_{M}(\lambda)$. Then we have
###### Theorem 5.2.
Let $\Omega\subset\mathbb{R}^{d_{1}}$ be a bounded domain with $C^{1}$
boundary and consider the product manifold $a\Omega\times M$.
1. (1)
If $\mathcal{N}^{(H)}_{M}(\lambda)$ satisfies (5.4), then there exists
$a_{0}>0$ (depends on $\Omega$ and $M$) such that for any $0<a<a_{0}$, the
eigenvalues of the Laplace-Beltrami operator on $a\Omega\times M$ with the
following mixed boundary condition
$\text{ Dirichlet condition on }\partial(a\Omega)\times M,\quad\text{
condition (H) on }a\Omega\times\partial M$
satisfy Pólya’s conjecture (1.2),
2. (2)
If $\mathcal{N}^{(H)}_{M}(\lambda)$ satisfies (5.5), then there exists
$a_{0}>0$ (depends on $\Omega$ and $M$) such that for any $0<a<a_{0}$, the
eigenvalues of the Laplace-Beltrami operator on $a\Omega\times M$ with the
following mixed boundary condition
$\text{ Neumann condition on }\partial(a\Omega)\times M,\quad\text{ condition
(H) on }a\Omega\times\partial M$
satisfy Pólya’s conjecture (1.3).
For example, one may take the condition (H) to be either Dirichlet boundary
condition or Neumann boundary condition or Robin ($\frac{\partial
f}{\partial\nu}=\rho f$, with bounded $\rho$) boundary condition, and in all
these cases the inequalities (5.4) and (5.5) hold. Thus one get many
eigenvalue problems whose eigenvalues satisfy Pólya’s conjecture.
## 6\. Two examples with explicit constants
In this section, we give two examples for which one can calculate the
constants involved in the proof, and thus give explicit domains/manifolds for
which Pólya’s conjecture holds.
We first construct a planar domain $\Omega$ for which we can calculate the
constant $C(\Omega)$ in (2.6) explicitly, and thus find out the number $a_{0}$
in Theorem 1.2 for $\Omega$.
Let $S$ be a square with side length 10 and $T$ be an equilateral triangle
with side length 1. The domain $\Omega$ is constructed by placing $T$ at the
center of one side of $S$, as shown by the picture below:
Note that the angle $\theta=\frac{2\pi}{3}$, which implies that $\Omega$
cannot tile $\mathbb{R}^{2}$. In what follows we prove
###### Proposition 6.1.
For any $a\leq\frac{1}{4\pi}$, the Dirichlet eigenvalues of
$(0,a)\times\Omega$ satisfies Pólya’s conjecture (1.2).
###### Proof.
By Faber–Krahn’s inequality ([5], [19]),
$\lambda_{1}(\Omega)\geq\frac{4\pi^{2}}{\omega_{2}|\Omega|}>10^{-1}.$
So $\mathcal{N}^{D}_{\Omega}(\lambda)=0$ for $\lambda\leq 10^{-1}$.
Now suppose $\lambda>10^{-1}$. For the square $S$ we have
$\displaystyle\mathcal{N}^{N}_{S}(\lambda)$
$\displaystyle=\\#\\{(m,n)\in\mathbb{Z}_{\geq 0}^{2}|\
m^{2}+n^{2}<\frac{100\lambda}{\pi^{2}}\\}$
$\displaystyle\leq\frac{100\lambda}{4\pi}+\frac{20}{\pi}\sqrt{\lambda}+2$
$\displaystyle<\frac{100\lambda}{4\pi}+20\sqrt{\lambda}.$
For the triangle $T$, let
$P=\\{(x,2x)\\}_{x\in\mathbb{R}}\cup\\{(2x,x)\\}_{x\in\mathbb{R}}\cup\\{(x,-x)\\}_{x\in\mathbb{R}}$,
then by [28, Proposition 3], one has
$\displaystyle\mathcal{N}^{N}_{T}(\lambda)=$
$\displaystyle\frac{1}{6}\\#\\{(m,n)\in\mathbb{Z}^{2}|\ (m,n)\notin
P,3|(m+n),\frac{16\pi^{2}}{27}(m^{2}+n^{2}-mn)<\lambda\\}+$
$\displaystyle\frac{1}{3}\\#\\{(m,n)\in\mathbb{Z}^{2}|\ (m,n)\in
P,\frac{16\pi^{2}}{27}(m^{2}+n^{2}-mn)<\lambda\\}+\frac{2}{3}.$
Since
$\displaystyle\\#\\{(m,n)\in\mathbb{Z}^{2}|\ (m,n)\notin
P,3|(m+n),\frac{16\pi^{2}}{27}(m^{2}+n^{2}-mn)<\lambda\\}$ $\displaystyle=$
$\displaystyle\\#\\{(m,n)\in\mathbb{Z}^{2}|\ (m,n)\notin
P,3|(m+n),(m-\frac{n}{2})^{2}+\frac{3n^{2}}{4}<\frac{27\lambda}{16\pi^{2}}\\}$
$\displaystyle\leq$ $\displaystyle\frac{1}{3}\\#\\{(m,n)\in\mathbb{Z}^{2}|\
m^{2}+\frac{3n^{2}}{4}<\frac{27\lambda}{16\pi^{2}}\\}+2(\frac{3\sqrt{3}\sqrt{\lambda}}{4\pi}+\frac{3\sqrt{\lambda}}{2\pi})+4$
$\displaystyle<$ $\displaystyle\frac{3\sqrt{3}\lambda}{8\pi}+60\sqrt{\lambda}$
and
$\displaystyle\\#\\{(m,n)\in\mathbb{Z}^{2}|\ (m,n)\in
P,\frac{16\pi^{2}}{27}(m^{2}+n^{2}-mn)<\lambda\\}$ $\displaystyle\leq$
$\displaystyle 3\\#\\{k\in\mathbb{Z}|\
k^{2}<\frac{9\lambda}{16\pi^{2}}\\}<\frac{9\sqrt{\lambda}}{4\pi}+3<30\sqrt{\lambda},$
we get
$\mathcal{N}^{N}_{T}(\lambda)<\frac{\sqrt{3}\lambda}{16\pi}+30\sqrt{\lambda}.$
So we get
$\mathcal{N}^{D}_{\Omega}(\lambda)<\mathcal{N}^{N}_{\Omega}(\lambda)\leq\mathcal{N}^{N}_{S}(\lambda)+\mathcal{N}^{N}_{T}(\lambda)\leq\frac{1}{4\pi}(100+\frac{\sqrt{3}}{4})\lambda+50\sqrt{\lambda}.$
In other words, one may take $C(\Omega)=50$ in (2.6). It follows from the
proof in §4.2 that for any $a\leq\frac{1}{4\pi}$, all Dirichlet eigenvalues of
$(0,a)\times\Omega$ satisfy Pólya’s Conjecture (1.2). ∎
Finally we turn to the Riemannian manifold setting and consider the standard
two-sphere $M=S^{2}$. It is well known that the eigenvalues of $(S^{2},g_{0})$
are $k(k+1)$, with multiplicity $2k+1$ for all $k\in\mathbb{Z}_{\geq 0}$. It
follows that
$\mathcal{N}_{S^{2}}\big{(}k(k+1)\big{)}=k^{2},\qquad\mathcal{N}_{S^{2}}\big{(}k(k+1)+\varepsilon\big{)}=(k+1)^{2}.$
In other words, one can choose $C(S^{2})$ in (5.1) to be 1 and $C_{1}(S^{2})$
in (5.2) to be 1. Plugging into (5.3) and (4.7), we get
###### Proposition 6.2.
For any $a\leq\frac{\pi}{24}$, the manifold $(0,a)\times S^{2}$ satisfy
Pólya’s conjecture (1.2) and (1.3).
Note that in this example, if we take $a$ large,
* •
If we take $a>\sqrt{\frac{2}{3}}\pi$, then the first Dirichlet eigenvalue of
$(0,a)\times S^{2}$
$\lambda_{1}\big{(}(0,a)\times
S^{2}\big{)}=\frac{\pi^{2}}{a^{2}}<\frac{4\pi^{2}}{({4\pi\omega_{3}}a)^{\frac{2}{3}}},$
* •
If we take $\frac{\pi}{\sqrt{2}}\leq a<\sqrt{\frac{2}{3}}\pi$, then the first
nonzero Neumann eigenvalue of $(0,a)\times S^{2}$
$\mu_{1}\big{(}(0,a)\times
S^{2}\big{)}=\frac{\pi^{2}}{a^{2}}>\frac{4\pi^{2}}{({4\pi\omega_{3}}a)^{\frac{2}{3}}},$
so Pólya’s inequalities (1.2) and (1.3) will not hold for $(0,a)\times S^{2}$
when $a$ is large.
###### Remark 6.3.
We remark that in [13, Example 2.D], P. Freitas and I. Salavessa had already
observed (from a very simple tiling argument) that $(0,a)\times S^{2}$
satisfies Pólya’s inequalities for $a$ small enough but fails to satisfy
Pólya’s inequalities for $a$ large. Our method is more complicated but has the
advantage that we could give explicit estimates of $a$ for Pólya’s
inequalities to hold. We would like to thank the authors for pointing out this
fact to us.
## References
* [1] V. G. Avakumović: über die Eigenfunktionen auf geschlossenen Riemannschen Mannigfaltigkeiten. _Mathematische Zeitschrift_ , 1956, 65(1): 324-344.
* [2] P. Bérard and G. Besson: Spectres et groupes cristallographiques. II: domaines sphériques. _Annales de l’institut Fourier_ , 1980, 30(3): 237-248.
* [3] F. A. Berezin: Covariant and contravariant symbols of operators. _Mathematics of the USSR-Izvestiya_ , 1972, 6(5): 1117.
* [4] D. Bucur and A. Henrot: Maximization of the second non-trivial Neumann eigenvalue. _Acta Mathematica_ , 2019, 222(2), 337–361.
* [5] G. Faber: Beweis, dass unter allen homogenen Membranen von gleicher Fläche und gleicher Spannung die kreisförmige den tiefsten Grundton gibt. 1923.
* [6] N. Filonov: On an inequality between Dirichlet and Neumann eigenvalues for the Laplace operator. _St. Petersburg Mathematical Journal_ , 2005, 16(2): 413-416.
* [7] N. Filonov and M. Levitin, I. Polterovich and D. Sher: Pólya’s conjecture for Euclidean balls. _Inventiones mathematicae_ , 2023: 1-41.
* [8] N. Filonov: On the Pólya conjecture for the Neumann problem in planar convex domains. arXiv preprint arXiv:2309.01432, 2023.
* [9] R. L. Frank and L. Geisinger: Semi-classical analysis of the Laplace operator with Robin boundary conditions. _Bulletin of Mathematical Sciences_ , 2012, 2(2): 281-319.
* [10] R. L. Frank and S. Larson: Two-term spectral asymptotics for the Dirichlet Laplacian in a Lipschitz domain. _Journal für die reine und angewandte Mathematik (Crelles Journal)_ , 2020, 2020(766): 195-228.
* [11] P. Freitas: A remark on Pólya’s conjecture at low frequencies. _Archiv der Mathematik_ , 2019, 112: 305-311.
* [12] P. Freitas, J. Lagacé, J. Payette: Optimal unions of scaled copies of domains and Pólya’s conjecture. _Arkiv för Matematik_ , 2021, 59(1): 11-51.
* [13] P. Freitas and I. Salavessa: Families of non-tiling domains satisfying Pólya’s conjecture. _Journal of Mathematical Physics_ , 64, 121503 (2023).
* [14] P. Freitas, J. Mao and I. Salavessa: Pólya-type inequalities on spheres and hemispheres. To appear in _Annales de l’Institut Fourier_ (Grenoble), arXiv:2204.07277.
* [15] L. Friedlander: Some inequalities between Dirichlet and Neumann eigenvalues. _Archive for rational mechanics and analysis_ , 1991, 116: 153-160.
* [16] V. Ivrii: Second term of the spectral asymptotic expansion of the Laplace-Beltrami operator on manifolds with boundary. _Functional Analysis and Its Applications_ , 1980, 14(2): 98-106.
* [17] R. Kellner: On a theorem of Pólya. _The American Mathematical Monthly_ , 1966, 73(8): 856-858.
* [18] P. Kröger: Upper bounds for the Neumann eigenvalues on a bounded domain in Euclidean space. _Journal of functional analysis_ , 1992, 106(2): 353-357.
* [19] E. Krahn: Über eine von Rayleigh formulierte Minimaleigenschaft des Kreises. _Mathematische Annalen_ , 1925, 94(1): 97-100.
* [20] E. Krahn: Über Minimaleigenschaften der Kugel in drei und mehr Dimensionen. _Acta et Commentationes Universitatis Tartuensis (Dorpatensis)_ , A 9, 1-44, 1926.
* [21] A. Laptev: Dirichlet and Neumann eigenvalue problems on domains in Euclidean spaces. _Journal of Functional Analysis_ , 1997, 151(2): 531-545.
* [22] B. M. Levitan: On the asymptotic behaviour of the spectral function of the second order elliptic equation. _Izvestiya Akademii Nauk SSSR. Seriya Matematicheskaya_ 1952, 16(4): 325-352.
* [23] P. Li and S. T. Yau: On the Schrödinger equation and the eigenvalue problem. _Communications in Mathematical Physics_ , 1983, 88(3): 309-318.
* [24] E. H. Lieb: The classical limit of quantum spin systems. _Communications in Mathematical Physics_ , 1973, 31: 327-340.
* [25] F. Lin: Extremum problems of Laplacian eigenvalues and generalized Pólya conjecture. _Chinese Annals of Mathematics, Series B_ , 2017, 38(2): 497-512.
* [26] R. Melrose: Weyl asymptotics for the phase in obstacle scattering. _Communications in partial differential equations_ , 1988, 13(11): 1431-1439.
* [27] L. E. Payne: Inequalities for eigenvalues of plates and membranes. _Journal of Rational Mechanics and Analysis_ , 1955, 4: 517-529.
* [28] M. A. Pinsky:. The eigenvalues of an equilateral triangle. _SIAM Journal on Mathematical Analysis_ , 1980, 11(5): 819-827.
* [29] G. Pólya: Mathematics and Plausible Reasoning, Volume II: Patterns of Plausible Inference, _Princeton University Press_ , 1954.
* [30] G. Pólya: On the eigenvalues of vibrating membranes. _Proceedings of the London Mathematical Society_ , 1961, 11(3): 419-433.
* [31] R. Seeley: A sharp asymptotic remainder estimate for the eigenvalues of the Laplacian in a domain of $\mathbb{R}^{3}$. _Advances in Mathematics_ , 1978, 29(2): 244-269.
* [32] R. Seeley: An estimate near the boundary for the spectral function of the Laplace operator. _American Journal of Mathematics_ , 1980, 102(5): 869-902.
* [33] G. Szegö: Inequalities for certain eigenvalues of a membrane of given area. _Journal of Rational Mechanics and Analysis_ , 1954, 3: 343-356.
* [34] H. F. Weinberger: An isoperimetric inequality for the N-dimensional free membrane problem. _Journal of Rational Mechanics and Analysis_ , 1956, 5(4): 633-636.
* [35] H. Weyl: Über die asymptotische Verteilung der Eigenwerte. _Nachrichten von der Gesellschaft der Wissenschaften zu Göttingen, Mathematisch-Physikalische Klasse_ , 1911: 110-117.
|
# Multi-band Extension of the Wideband Timing Technique
Avinash Kumar Paladi1,2, Churchil Dwivedi2, Prerna Rana3, Nobleson K.4,
Abhimanyu Susobhanan5, Bhal Chandra Joshi6,7, Pratik Tarafdar8, Debabrata
Deb8, Swetha Arumugam10, A. Gopakumar3, M. A. Krishnakumar11,12, Neelam Dhanda
Batra13, Jyotijwal Debnath8,9, Fazal Kareem14,15, Paramasivan Arumugam7,
Manjari Bagchi8,9, Adarsh Bathula16, Subhajit Dandapat3, Shantanu Desai17,
Yashwant Gupta6, Shinnosuke Hisano18, Divyansh Kharbanda17, Tomonosuke
Kikunaga18, Neel Kolhe19, Yogesh Maan6, P. K. Manoharan20, Jaikhomba Singha7,
Aman Srivastava17, Mayuresh Surnis21, Keitaro Takahashi22,23
1 Joint Astronomy Programme, Indian Institute of Science, Bengaluru,
Karnataka, 560012, India
2 Department of Earth and Space Sciences, Indian Institute of Space Science
and Technology, Valiamala, Thiruvananthapuram 695547, Kerala, India
3 Department of Astronomy and Astrophysics, Tata Institute of Fundamental
Research, Homi Bhabha Road, Navy Nagar, Colaba, Mumbai 400005, India
4 Department of Physics, BITS Pilani Hyderabad Campus, Hyderabad 500078,
Telangana, India
5 Center for Gravitation Cosmology and Astrophysics, University of Wisconsin-
Milwaukee, Milwaukee, WI 53211, USA
6 National Centre for Radio Astrophysics, Pune University Campus, Pune 411007,
India
7 Department of Physics, Indian Institute of Technology Roorkee,
Roorkee-247667, India
8 The Institute of Mathematical Sciences, C. I. T. Campus, Taramani, Chennai
600113, India
9 Homi Bhabha National Institute, Training School Complex, Anushakti Nagar,
Mumbai 400094, India
10 Department of Electrical Engineering, IIT Hyderabad, Kandi, Telangana
502284, India
11 Max-Planck-Institut für Radioastronomie, Auf dem Hügel 69, 53121 Bonn,
Germany
12 Fakultät für Physik, Universität Bielefeld, Postfach 100131, 33501
Bielefeld, Germany
13 Department of Physics and Astrophysics, University of Delhi, Delhi
14 Department of Physical Sciences, Indian Institute of Science Education and
Research Kolkata, West Bengal, 741246, India
15 Center of Excellence in Space Sciences India, Indian Institute of Science
Education and Research Kolkata, West Bengal, 741246, India
16 Indian Institute of Science Education and Research, Mohali - 140306,
Punjab, India
17 Department of Physics, IIT Hyderabad, Kandi, Telangana 502284, India
18 Kumamoto University, Graduate School of Science and Technology, Kumamoto,
860-8555, Japan
19 Department of Physics, St. Xavier’s College (Autonomous), Mumbai 400001,
Maharashtra, India
20 Arecibo Observatory, University of Central Florida, Arecibo 00612, USA
21 Department of Physics, IISER Bhopal, Bhauri Bypass Road, Bhopal 462066,
India
22 Faculty of Advanced Science and Technology, Kumamoto University, 2-39-1
Kurokami, Kumamoto 860-8555, Japan
23 International Research Organization for Advanced Science and Technology,
Kumamoto University, 2-39-1 Kurokami, Kumamoto 860-8555, Japan E-mail:
<EMAIL_ADDRESS>
(Accepted XXX. Received YYY; in original form ZZZ)
###### Abstract
The wideband timing technique enables the high-precision simultaneous
estimation of pulsar Times of Arrival (ToAs) and Dispersion Measures (DMs)
while effectively modeling frequency-dependent profile evolution. We present
two novel independent methods that extend the standard wideband technique to
handle simultaneous multi-band pulsar data incorporating profile evolution
over a larger frequency span to estimate DMs and ToAs with enhanced precision.
We implement the wideband likelihood using the libstempo python interface to
perform wideband timing in the tempo2 framework. We present the application of
these techniques to the dataset of fourteen millisecond pulsars observed
simultaneously in Band 3 ($300-500$ MHz) and Band 5 ($1260-1460$ MHz) of the
upgraded Giant Metrewave Radio Telescope (uGMRT) with a large band gap of 760
MHz as a part of the Indian Pulsar Timing Array (InPTA) campaign. We achieve
increased ToA and DM precision and sub-microsecond root mean square post-fit
timing residuals by combining simultaneous multi-band pulsar observations done
in non-contiguous bands for the first time using our novel techniques.
###### keywords:
pulsars: general — galaxies: ISM — gravitational waves — methods: data
analysis
††pubyear: 2023††pagerange: Multi-band Extension of the Wideband Timing
Technique–B
## 1 Introduction
Pulsars are rotating neutron stars emitting broadband electromagnetic
radiation that is observed as periodic pulses. The rotation of a pulsar can be
tracked accurately by measuring the times of arrival (ToAs) of its pulses, and
this technique is known as pulsar timing (Hobbs et al., 2006; Edwards et al.,
2006). The pulsar signal is dispersed while propagating through the ionized
interstellar medium (IISM) by an amount that is proportional to the integrated
free electron column density along the line of sight, usually referred to as
the dispersion measure (DM), and inversely proportional to the square of the
observing frequency $\nu$ (Lorimer & Kramer, 2012). Conventionally, the rough
measurement of DM for a pulsar used to be done by splitting the data into
multiple sub-bands and correcting for the DM induced delay for each sub-band
and then adding the dispersed bands again (Lorimer & Kramer, 2012). In recent
days, many sophisticated techniques have been proposed, which not only
provides more accurate values of DM but also provide epoch to epoch variations
of DM (e.g. Ahuja et al., 2005).
Pulsar timing has traditionally been done by splitting the data into multiple
sub-bands with negligible dispersion smear and independently measuring the ToA
from each sub-band, known as narrowband timing (Taylor, 1992). The improvement
in telescope sensitivity, the advent of wideband receivers and backends (e.g.
Johnston et al., 2021; Hobbs et al., 2020; Reddy et al., 2017; Gupta et al.,
2017), and decades-long pulsar timing campaigns such as Pulsar Timing Arrays
(PTAs: Foster & Backer, 1990) have presented significant challenges to the
narrowband approach. These challenges include inadequate modeling of the pulse
profile variability as a function of frequency, difficulty in correcting for
interstellar scattering, and large data sizes. The wideband timing technique
seeks to address these issues by treating the pulse profile as a two-
dimensional entity in frequency and pulse phase (usually referred to as a
portrait) and simultaneously measuring one ToA and one DM per observation
using the full bandwidth (Pennucci et al., 2014; Pennucci, 2019).
PTA experiments, such as the Parkes Pulsar Timing Array (PPTA: Hobbs, 2013),
the European Pulsar Timing Array (EPTA: Kramer & Champion, 2013), the North
American Nanohertz Observatory for Gravitational Waves (NANOGrav: McLaughlin,
2013), the Indian Pulsar Timing Array (InPTA: Joshi et al., 2018), the Chinese
Pulsar Timing Array (CPTA: Lee, 2016), MeerKat Pulsar Timing Array (MPTA:
Miles et al., 2023), and the International Pulsar Timing Array (IPTA: Hobbs et
al., 2010; Perera et al., 2019; Verbiest et al., 2016) consortium which
combines the data and resources from various PTAs, aim to detect nanohertz
gravitational waves using an ensemble of millisecond pulsars (MSPs) as
celestial clocks. Recent wideband timing studies across a wide range of
observing frequencies have demonstrated significant improvements in ToA and DM
measurement precision (Liu et al., 2014; Fonseca et al., 2021; Nobleson et
al., 2022), and PTAs and other high-precision pulsar timing campaigns are now
increasingly adopting wideband techniques due to their advantages in dealing
with broadband observations (e.g. Alam et al., 2021; Tarafdar et al., 2022;
Curyło et al., 2023).
The InPTA experiment complements the international PTA efforts by employing
the unique features of the upgraded Giant Metrewave Radio Telescope (uGMRT:
Gupta et al., 2017). The high sensitivity of the uGMRT at low frequencies,
combined with its ability to perform simultaneous multi-band observations,
makes it an ideal instrument to characterize interstellar medium effects,
which are stronger at low frequencies (Krishnakumar et al., 2021). The
application of the wideband technique to the uGMRT data and the ToA and DM
precision improvements therefrom were demonstrated in Nobleson et al. (2022).
The recently published first data release of the InPTA (InPTA DR1: Tarafdar et
al., 2022) built on the work of Krishnakumar et al. (2021) and Nobleson et al.
(2022), has presented the results of narrowband and wideband timing of 14
pulsars observed over a time span of 3.5 years. This work included some of the
most precise DM measurements to date, estimated using both the narrowband and
the wideband techniques. Recently the InPTA collaboration has completed Single
Pulsar Noise Analysis on the DR1 pulsars using narrowband data (Srivastava et
al., 2023).
The InPTA observes pulsars in two uGMRT bands, namely the Band 3 ($300-500$
MHz) and the Band 5 ($1260-1460$ MHz). Although Nobleson et al. (2022) and
Tarafdar et al. (2022) only used Band 3 data for the wideband timing, the DM
precision achieved therein was comparable to the combined Band 3+5 narrowband
DM estimates. This raises the exciting possibility of attaining further
improvements in DM precision by combining the two uGMRT bands in the wideband
paradigm.
In this work, we develop two novel methods to combine simultaneous
observations of the same pulsar obtained at multiple bands using the wideband
technique to obtain a single ToA and DM combination per epoch across these
multiple bands. We then demonstrate the ToA and DM precision improvement
achieved with these techniques using the InPTA observations of 14 MSPs,
simultaneously observed at Band 3 and Band 5 seperated by a large band gap of
about 760 MHz, which were selected for the InPTA first data release (Tarafdar
et al., 2022). These techniques provide significant improvements in the DM
precision estimation that can be achieved using existing and future telescopes
which can perform simultaneous or quasi-simultaneous multi-frequency
observations such as the Square Kilometer Array (SKA) (Kramer & Stappers,
2015; Janssen et al., 2015). For the timing analysis, we extend the
traditional timing methodology to incorporate the wideband timing likelihood
function (Appendix B of Alam et al. (2021)) in tempo2 using libstempo.
This paper is structured as follows. We begin by providing a brief overview of
the wideband timing technique in subsection 2.1. In subsection 2.2, we present
two novel independent methods for applying the wideband technique to two
simultaneous band observations of a pulsar taken at different radio
frequencies, which can be easily extended to multiple bands. We apply our
methods to the case of PSR J1909$-$3744 and show the validation scheme and
comparisons against each other as well as against the single band (Band 3)
results in Section 3. We present the application of our novel methods to the
InPTA dataset of 14 MSPs in Section 4. We summarize our results in Section 5
and discuss avenues for future improvements and extensions in Section 6. Our
implementation of the wideband likelihood function using tempo2 and libstempo
is briefly described in Appendix B.
## 2 Multi-band extension of the Wideband Timing technique
### 2.1 Brief overview of the Wideband Technique
We begin by briefly summarising the wideband technique developed in Pennucci
et al. (2014), Pennucci (2019), and Alam et al. (2021). The total intensity
integrated pulse profile of a pulsar can be expressed as a two-dimensional
object $D(\nu,\varphi)$ which is a function of the observing frequency $\nu$
and the pulse phase $\varphi$, and is referred to as a pulse portrait. Given a
model for the observed portrait $P(\nu,\varphi)$, referred to as the template
portrait or the model portrait, $D(\nu,\varphi)$ can be written as:
$D(\nu,\varphi)=B(\nu)+a(\nu)P(\nu,\varphi-\phi(\nu))+N(\nu,\varphi)\,,$ (1)
where $B(\nu)$ is the DC offset in each frequency channel, $a(\nu)$ is an
amplitude that arises from the intrinsic power spectral density of the pulsar
emission and interstellar scintillation and also depends on the receiver
bandpass, and $N(\nu,\varphi)$ is an additive noise that is usually assumed to
be Gaussian and uncorrelated in the absence of radio frequency interference
(RFI). In practice, $D(\nu,\varphi)$ and $P(\nu,\varphi)$ are discretised in
both $\nu$ and $\varphi$, i.e., $D_{nj}\equiv D(\nu_{n},\varphi_{j})$ such
that $n$ denotes the frequency channels and $j$ corresponds to the phase bins.
The frequency dependence of the phase shift $\phi(\nu)$ arises primarily due
to the interstellar dispersion and is given by:
$\phi(\nu)=\phi_{\text{ref}}+\frac{K\times\text{DM}}{P_{s}}\left(\frac{1}{\nu^{2}}-\frac{1}{\nu_{\text{ref}}^{2}}\right)\,,$
(2)
where $\phi_{\text{ref}}$ is the achromatic phase shift, $K$ is the Dispersion
constant, $P_{s}$ is the apparent spin period of the pulsar at the epoch of
observation, and $\nu_{\text{ref}}$ is a Barycentric reference frequency.
Given $\phi_{\text{ref}}$, the ToA can be computed as
$t=t_{f}+{P_{s}\phi_{\text{ref}}}/{2\pi}$ where $t_{f}$ is the timestamp
corresponding to a fiducial phase point in the data portrait111In practice,
$t_{f}$ may be affected by instrumental delays such as those encountered in
Tarafdar et al. (2022), and one must correct for them.. $\phi_{\text{ref}}$
can be understood as the difference between the fiducial phases of the data
portrait and the template portrait.
Computing the discrete Fourier transform of equation (1) along the $\phi$
axis, applying the discrete Fourier shift theorem, and excluding the DC term,
we have:
$\widetilde{D}_{nk}=a_{n}\widetilde{P}_{nk}\;e^{2\pi
ik\phi_{n}}+\widetilde{N}_{nk}\,.$ (3)
where $\widetilde{D}_{nk}$ and $\widetilde{P}_{nk}$ denote the discrete
Fourier transform of the data portrait and the template portrait respectively,
and $\phi_{n}=\phi(\nu_{n})$. The quantities of interest $\phi_{\text{ref}}$
and DM can then be estimated by minimizing the weighted least-squares
statistic:
$\chi^{2}(\phi_{\text{ref}},\text{DM},a_{n})=\sum_{n,k}\frac{\left|\widetilde{D}_{nk}-a_{n}\widetilde{P}_{nk}e^{2\pi
ik\phi_{n}}\right|^{2}}{{\sigma_{n}^{\prime}}^{2}}\,,$ (4)
where ${\sigma_{n}^{\prime}}^{2}$ denotes the noise variance of the Fourier
coefficients $\widetilde{D}_{nk}$. It turns out that
$\chi^{2}(\phi_{\text{ref}},\text{DM},a_{n})$ can be analytically minimised
over the amplitudes $a_{n}$, and this leads to
$\chi^{2}(\phi_{\text{ref}},\text{DM})=S-\sum_{n}\frac{C_{n}^{2}}{T_{n}}\,,$
(5)
where
$\displaystyle S$
$\displaystyle=\sum_{n,k}\frac{|\widetilde{D}_{nk}|^{2}}{{\sigma_{n}^{\prime}}^{2}}\,,$
(6a) $\displaystyle T_{n}$
$\displaystyle=\sum_{k}\frac{|\widetilde{P}_{nk}|^{2}}{{\sigma_{n}^{\prime}}^{2}}\,,$
(6b) $\displaystyle C_{n}$
$\displaystyle=\Re\left\\{\sum_{k}\frac{\widetilde{D}_{nk}\widetilde{P}_{nk}^{*}\;e^{2\pi
ik\phi_{n}}}{{\sigma_{n}^{\prime}}^{2}}\right\\}\,.$ (6c)
Choosing $\nu_{\text{ref}}$ such that the covariance between
$\phi_{\text{ref}}$ and DM vanishes (see the Appendix of Pennucci et al.
(2014)), $\phi_{\text{ref}}$ and DM can be estimated by numerically minimizing
$\chi^{2}(\phi_{\text{ref}},\text{DM})$.
The template portrait $P(\nu,\varphi)$ is usually obtained from a single high
signal-to-noise ratio (S/N) portrait or an averaged portrait generated from
many observations. The mean-subtracted template portrait is decomposed into
many ‘eigenprofiles’ using principal component analysis (PCA). A smoothed
template portrait is then reconstructed from a small number of significant
eigenprofiles by spline-interpolating them (Pennucci, 2019). By linearly
combining the $n_{\text{eig}}$ significant eigenprofiles $\hat{e}_{i}$ using
the spline coefficients $B_{i}$ and adding it to the mean profile
$\widetilde{p}$, a template profile $T(\nu)$ at any frequency $\nu$ can be
created as
$T(\nu)=\sum_{i=1}^{n_{\text{eig}}}B_{i}(\nu)\;\hat{e}_{i}+\widetilde{p}\,.$
(7)
Note that the DMs estimated from the wideband technique are not derived from
ToAs unlike in the narrowband case, but rather measured simultaneously with
each ToA. Therefore, DM measurements should be treated as data points on an
equal footing with the ToAs while computing the likelihood function. In the
simple case of a pulsar with timing model parameters ($\boldsymbol{\theta}$),
ToAs ($t_{i}$), timing residuals ($r_{i}$), ToA uncertainties ($\sigma_{i}$),
DM measurements ($\text{DM}_{i}$), DM uncertainties ($\varsigma_{i}$), and DM
model ($d(t)$), the wideband log-likelihood can be written as
$\ln
L(\boldsymbol{r},\boldsymbol{\sigma},\boldsymbol{\text{DM}},\boldsymbol{\varsigma}|\boldsymbol{\theta})=\ln
L_{0}-\frac{1}{2}\sum_{i}\left[\left(\frac{r_{i}}{\sigma_{i}}\right)^{2}-\left(\frac{\text{DM}_{i}-d(t_{i})}{\varsigma_{i}}\right)^{2}\right]\,,$
(8)
where the first term is a normalization term, the second term is the usual
narrowband likelihood, and the third term is the likelihood function of the DM
measurements. A more general version of the above equation, applicable to more
rigorous noise models can be found in Appendix B of Alam et al. (2021).
### 2.2 Extending the Wideband technique for multiple bands
The standard wideband timing technique, summarized in section 2.1, has been
applied to various single band observations across a wide range of observing
frequencies (Fonseca et al., 2021; Alam et al., 2021; Nobleson et al., 2022;
Tarafdar et al., 2022; Curyło et al., 2023). In this section, we present and
demonstrate two novel independent methods namely the Combined Portrait (CP)
method and Combined Chi-squared (CC) method to combine simultaneous
observations performed in two non-contiguous frequency bands within the
paradigm of wideband technique to estimate a single ToA and DM combination per
epoch covering the entire frequency range of these bands. It is
straightforward to extend these techniques to multiple bands with simultaneous
observations, which will be part of a future work.
#### 2.2.1 The Combined-portrait (CP) method
In this method, we begin by time-collapsing the frequency-resolved profiles
obtained simultaneously in the two frequency bands using the pam command of
PSRCHIVE (Hotan et al., 2004). We then combine the profiles in the two bands
along the frequency axis using the psradd command of PSRCHIVE. This requires
both profiles to have the same number of phase bins; hence, we phase-collapse
the higher-phase resolution profile to match the lower-phase resolution one
using the pam command before appending them using psradd. The profiles of each
frequency band are also collapsed in frequency to an appropriate number of
sub-bands such that there is a reasonable signal-to-noise ratio (S/N) in each
sub-band, and there are also enough sub-bands to obtain a 2-D template
containing information of profile evolution across the band. Since the exact
start time of the observation in each band may not be identical, the profiles
are aligned by the psradd command by rotating them in phase using the pulsar
ephemeris used for folding. For generating a noise-free template portrait, we
use an epoch with high-S/N in both the bands. We first excise frequency
channels with any residual RFI from both the bands for the template epoch
using the pazi command and then obtain a combined data profile using psradd
covering the frequency of the two bands. Finally, a template portrait is
generated from this combined data profile using the ppalign and ppspline
modules of PulsePortraiture (Pennucci et al., 2014; Pennucci, 2019). Here, the
spline model is interpolated over the large frequency gap in between the two
bands. For accurate modeling of the profile evolution across the two bands, we
choose the required number of eigenprofiles and tolerance values for the
template portrait using the procedure described in section 4.2 of Tarafdar et
al. (2022). A single wideband ToA and the corresponding DM for the combined
observation of each epoch are then estimated using the ppToAs module of
PulsePortraiture.
#### 2.2.2 The Combined Chi-squared (CC) method
In this method, we treat the data portraits and the corresponding templates
for each band in their native phase resolution (without phase-collapsing or
combining them along the frequency axis) and bandwidths. We use the time-
collapsed data of two bands and partially collapse the frequency channels in
each band to maintain a reasonable S/N in each sub-band. The noise-free
templates are generated for each band separately using a high-S/N epoch after
RFI excision. Here, there is no interpolation of spline model over the large
frequency gap in between two bands, as both bands are treated separately.
While generating the templates, we take care of the phase offset between
multiple bands by rotating them appropriately. We estimate a single ToA and DM
pair for multiple bands in each epoch by minimizing a combined chi-squared
statistic defined as
$\chi^{2}(\phi_{\text{ref}},\text{DM})=\sum_{b}\left\\{S_{b}-\sum_{n}\frac{C_{bn}^{2}}{T_{bn}}\right\\}\,,$
(9)
where the index $b$ labels the different bands, and $S_{b}$, $C_{bn}$ and
$T_{bn}$ are defined by equations (6) using the data portrait $D_{bnj}$ and
the template portrait $P_{bnj}$ for each band $b$. Since the timestamp
$t_{fb}$ corresponding to the fiducial phase for different bands need not be
the same, equation (2) should be modified as follows:
$\phi_{bn}=\phi_{\text{ref}}+\frac{K\times\text{DM}}{P_{s}}\left(\frac{1}{\nu_{n}^{2}}-\frac{1}{\nu_{\text{ref}}^{2}}\right)-\delta_{b}\,,$
(10)
where
$\delta_{b}=\frac{t_{fb}-t_{f0}}{P_{s}}\,,$ (11)
and we have arbitrarily chosen the band labeled $b=0$ as the reference and
$P_{s}$ is the pulsar spin period222In this work, we are considering the
period from the center of the observation. The frequency $\nu_{\text{ref}}$ is
chosen such that the covariance between $\phi_{\text{ref}}$ and the DM implied
by equation (9) vanishes. Note that this method preserves the full phase
resolution available in each band since the number of phase bins need not be
equal for the different bands in equations (6) and (9)-(11).
#### 2.2.3 Wideband Timing with tempo2 using libstempo
The wideband likelihood was previously only available in tempo (Nice et al.,
2015) and PINT (Luo et al., 2021). In this work, We implement the wideband
likelihood using the libstempo (Vallisneri, 2020) python interface to perform
wideband timing in the tempo2 framework (refer Appendix B for details). We
considered DMEFAC and T2EFAC333DMEFAC and T2EFAC are white noise parameters
used to scale the DM and ToA uncertainties, respectively to account for the
radiometer noise contribution to the DM and ToA uncertainties, respectively.
These are estimated via a $\chi^{2}$-implementation done with libstempo and
the optimum fit parameters for various pulsars were chosen as per the InPTA
DR1 Narrowband timing (Tarafdar et al., 2022). The DMEFAC and T2EFAC values
were estimated such that the reduced $\chi^{2}$ obtained by iteratively
fitting the timing parameters is close to unity along with the post-fit
weighted RMS to be of the order of a few 100s of ns to a few $\mu$s, for each
of the Band 3, CC, and CP ToAs. In this way, ToAs obtained from the
combination of data from two non-contiguous frequency bands are timed for the
first time within the paradigm of the wideband technique.
## 3 Application on PSR J1909–3744
PSR J1909$-$3744 is a binary MSP with a rotational period $P_{s}\sim$2.95 ms.
It was discovered using the Parkes 64-m Radio Telescope in the Swinburne High
Latitude Pulsar Survey (Jacoby et al., 2003). It is one of the best pulsars
for PTA studies (Verbiest et al., 2016; Perera et al., 2019) due to its sharp
pulse profile, low-profile evolution with the radio frequency, and well-
studied timing model (Liu et al., 2020). Here, we demonstrate and validate the
CC and CP methods (§2.2) using the uGMRT Band 3 and Band 5 data of PSR
J1909$-$3744 from Cycles 37-40 (MJDs $58781-59496$), with 200 MHz bandwidth
(BW), obtained as a part of the InPTA campaign (Tarafdar et al., 2022). We
used MJD $59630$ as the template epoch obtained from InPTA observations of
Cycle 41 of the uGMRT. The details of observations and data reduction
procedures for these datasets can be found in Susobhanan et al. (2021) and
Tarafdar et al. (2022).
### 3.1 Combined Portrait (CP) method
As discussed in subsection 2.2.1, the CP method requires the phase resolution
of two bands to be the same for combining the data. The Band 5 uGMRT data of
the InPTA campaign is configured to be recorded with a smaller time resolution
than the Band 3 data, which leads to a smaller number of phase bins in Band 5
than in Band 3 when the data is folded. Hence, we phase-collapsed the Band 3
data to the same number of phase bins as those of Band 5 before appending the
two bands using psradd.
A comparison of wideband DM time series of Band 3+5 (CP) and Band 3 is shown
in figure 1, where the Band 3 DM time series is obtained while preserving the
original phase resolution. The Kendall Tau correlation coefficient (Kendall,
1938) of value 0.7188 and $p$-value $\sim\times 10^{-12}$ indicates a good
agreement between the two DM time series. We also see a slight offset between
Band 3 and Band 3+5 CP method DM time series (refer section 4.2 for a
discussion). In figure 2, we compare the DM (left panel) and ToA (right panel)
precisions of Band 3+5 (CP) and Band 3 time series. The points lying below the
dashed diagonal line indicate an improved DM or ToA precision with the CP
method compared to Band 3 results and vice versa.
We see in figure 2 that all epochs do not show an improved DM precision, and
most of the epochs show a worsened ToA precision i.e., we found a decrement in
the median precision or an increase in the median uncertainties values
($\sigma_{\text{DM}}$ and $\sigma_{\text{ToA}}$) of Band 3+5 (CP) results
compared to Band 3. This is primarily due to the decreased phase resolution of
Band 3 data used in the CP method. Hence, for combining bands, the CP method
has a disadvantage, especially for MSPs like J1909$-$3744, wherein the pulse
profile is sharp with minimal features, leading to only a few phase bins in
the pulse region of the profile upon toning down the phase resolution which
leads to a loss of information content, and thereby leading to poor template
construction as well as bad DM and ToA estimates.
Figure 1: Wideband DMs obtained from the Combined Portrait (CP) method for
Band 3+5 data along with traditional Band 3 (single-band) results for PSR
J1909$-$3744\. The top panel shows the DM time series for Band 3+5 CP (green
points) and Band 3 (blue points) overlaid for comparison. The Kendall Tau
coefficient on top right shows the correlation between the two time series
with the mentioned $p$-value. The bottom panel shows the DM differences
($\delta$DM in units of $10^{-5}$ pc cm-3) on subtracting both time series
(Band 3+5 $-$ Band 3) where the median value is shown by the dashed line and
the dash-dotted lines representing the MAD-band (equivalent to $3\sigma$
contour). Figure 2: A comparison of Wideband DM (left panel) and ToA (right
panel) uncertainties obtained from the CP method for Band 3+5 data is shown
against those obtained from the traditional Band 3 data for PSR J1909$-$3744\.
The Band 3+5 uncertainties estimated with the CP method are shown on the
vertical axis and Band 3 uncertainties on the horizontal axis (in units of
$10^{-5}\;\text{pc}\;\text{cm}^{-3}$ for DM uncertainties, and in units of
$\mu\text{s}$ for ToA uncertainties). The diagonal dash-dotted line shows the
$y=x$ curve. Points lying below the diagonal line show that the obtained
uncertainties with the CP method are lower compared to Band 3 results and vice
versa.
### 3.2 Combined Chi-squared (CC) method
The CC method preserves the native phase resolution of Band 3 and Band 5 data
as well as the template portraits, since it incorporates them within a
combined Fourier domain $\chi^{2}$-statistic as described in subsection 2.2.2.
Figure 3 shows the Band 3+5 DM time series obtained using the CC method in
comparison with the Band 3 DM time series, wherein we can see that the Band
3+5 DMs bear a high positive correlation with the Band 3 DMs, showing a good
agreement between the two. Figure 4 shows the Band 3+5 (CC) DM (left panel)
and ToA (right panel) uncertainties in comparison with the Band 3 results. We
can see that the Band 3+5 $\sigma_{\text{DM}}$ values are smaller than those
of Band 3 (all lying below the $y=x$ curve), hence showing a universal
improvement in the median DM precision after band-combination. The Band 3+5
$\sigma_{\text{ToA}}$ values are also slightly less compared to Band 3 leading
to an improvement in the median ToA precision as well.
Overall, the CC method provides significant improvements for Band 3+5 compared
to Band 3 results, especially because of preserving the native Band 3 phase
resolution, unlike the CP method. The templates are also more effectively
modelled because of applying PCA separately on each band without the need for
interpolating over a large frequency gap of $760$ MHz between Band 3 and Band
5. Similarly, when we apply PCA to model the template for the CP method on the
Band 3+5 data obtained using psradd, there is a possibility that the PCA
method may not be interpolating the profile evolution accurately due to the
wide band separation ($\sim 760$ MHz) between Band 3 and Band 5 data.
Therefore, CC method comes out as a more robust method for the band
combination.
Figure 3: Wideband DMs obtained from the Combined Chi-squared (CC) method for
Band 3+5 data along with traditional Band 3 (single-band) results for PSR
J1909$-$3744\. The top panel shows the DM time series for Band 3+5 CC (magenta
points) and Band 3 (blue points) overlaid for comparison. The Kendall Tau
coefficient on top right shows the correlation between the two time series
with the mentioned p-value. The bottom panel shows the DM differences
($\delta$DM in units of $10^{-5}\;\text{pc}\;\text{cm}^{-3}$) on subtracting
both time series (Band 3+5 $-$ Band 3), where the median value is shown by the
dashed line and the dash-dotted lines representing the MAD-band (equivalent to
$3\sigma$ contour). Figure 4: A comparison of Wideband DM (left panel) and ToA
(right panel) uncertainties obtained from the CC method for Band 3+5 data is
shown against those obtained from the traditional Band 3 data for PSR
J1909$-$3744\. The Band 3+5 uncertainties estimated with the CC method are
shown on the vertical axis and Band 3 uncertainties on the horizontal axis (in
units of $10^{-5}\;\text{pc}\;\text{cm}^{-3}$ for DM uncertainties, and in
units of $\mu\text{s}$ for ToA uncertainties). The diagonal dash-dotted line
shows the $y=x$ curve. Points lying below the diagonal line show that the
obtained uncertainties with the CC method are lower compared to Band 3 results
and vice versa.
### 3.3 Split-band test for the CC method
To validate the application of our novel CC method to combine the data of two
bands for estimating wideband DMs and ToAs, we perform a split-band test. In
this test, we consider one of the 200 MHz BW data (Band 3 is selected as it
has higher S/N than Band 5) and split it into two sub-bands each with a
bandwidth of 100 MHz using the psrsplit command of psrchive. We then estimate
the DM time series obtained by applying the CC method on these two sub-bands
and compare it with the wideband DM estimates obtained for the full 200 MHz BW
data. The split-band test results for PSR J1909-3744 are shown in figure 5,
where we can see that the DM values are in close agreement with Kendall Tau
value $\tau\sim 0.94$ and $p\sim 2\times 10^{-20}$ implying strong (positive)
correlation with the single band result for Band 3. The strong (positive)
correlation with negligible offsets indicates that the CC method for combining
bands, within the regime of the wideband technique, is working well. Hence,
the split-band test serves as a litmus test for validating the new technique.
Figure 5: The split-band test results for PSR J1909$-$3744\. The top panel
shows the wideband DM time series for Band 3 alone (blue points), and by
splitting it into two 100 MHz bands and using the CC method on the two sub-
bands (yellow points) overlaid for comparison. The Kendall Tau coefficient on
top right shows the correlation between the two time series with the mentioned
p-value. The bottom panel shows the DM differences ($\delta$DM in units of
$10^{-5}\;\text{pc}\;\text{cm}^{-3}$) on subtracting both time series (Band 3
split $-$ Band 3 single-band) with the median value shown by the dashed line
and the dash-dotted lines representing the MAD-band (equivalent to $3\sigma$
contour).
### 3.4 Wideband Timing results for PSR J1909$-$3744
The wideband timing results for PSR J1909$-$3744 obtained by implementing the
wideband likelihood with tempo2 using libstempo are shown in figure 6. The
DMEFAC and T2EFAC values are estimated for each of the Band 3, CC, and CP
ToAs. We then incorporate the T2EFAC and DMEFAC values to generate a global
timing solution. The post-fit timing residuals obtained from this procedure
are shown in figure 6. The post-fit weighted RMS ToA residual values for the
Band 3, CC, and CP timing residuals are obtained to be 0.235 $\mu$s, 0.326
$\mu$s, and 0.471 $\mu$s respectively, and are consistent with each other. We
fit the same parameters as fitted in the InPTA DR1 (Tarafdar et al., 2022)
narrowband timing method, which are F0 and F1 for the case of J1909$-$3744.
Figure 6: Timing results for PSR J1909-3744 obtained by implementing the wideband likelihood inside tempo2 using libstempo. The DMEFAC and T2EFAC values are estimated separately for the cases of Band 3 (blue points), CC (magenta points) and CP (green points) ToAs with only $200$ MHz data. The error bars shown in the plot have T2EFAC values incorporated. The post-fit weighted RMS ToA residual values for the Band 3, CC, and CP timing residuals are shown in the top right corner and the dashed horizontal line corresponds to zero residual level. | Parameter
---
Name (unit)
| Band 3
---
| Value | Uncertainty
---|---
| Band 3+5 CC
---
| Value | Uncertainty
---|---
| Band 3+5 CP
---
| Value | Uncertainty
---|---
F0 ($\text{s}^{-1}$) | | $339.31568666042$ | $3.1\times 10^{-11}$
---|---
| $339.31568666042$ | $2.6\times 10^{-11}$
---|---
| $339.31568666042$ | $1.8\times 10^{-11}$
---|---
F1 ($\text{s}^{-2}$) | | $-1.615\times 10^{-15}$ | $1.3\times 10^{-18}$
---|---
| $-1.615\times 10^{-15}$ | $1.1\times 10^{-18}$
---|---
| $-1.6158\times 10^{-15}$ | $8.1\times 10^{-19}$
---|---
Table 1: Table of fitted parameters obtained from the wideband timing of PSR
J1909$-$3744 using tempo2. The first column specifies the fitted parameters.
The second, third, and fourth columns list the Band 3, Band 3+5 CC, and Band
3+5 CP timing results respectively, enlisting the parameter values and their
uncertainties.
The post-fit RMS values for Band 3 and Band 3+5 (both CC and CP) are obtained
to be very close, while the fitted parameters estimated using the Band 3+5
methods are having better precision compared to the former, as shown in table
1. This clearly highlights the improvement in timing precision attained with
Band 3+5 wideband timing, especially the CC method.
## 4 Application on InPTA Data
We now present the results obtained by implementing the CP and CC methods, and
the wideband timing technique using tempo2 on the InPTA dual-band data (Band
3: 300$-$500 MHz and Band 5: 1260$-$1460 MHz) of 14 MSPs. The same data was
used for the first data release of InPTA (InPTA DR1: Tarafdar et al. (2022)).
| | DM uncertainties ($\text{pc}\;\text{cm}^{-3}$) |
---|---|---|---
| Pulsar
---
Name
| Band 3
---
| Median | Minimum
---|---
| Band 3+5 CC
---
| Median | Minimum
---|---
| Band 3+5 CP
---
| Median | Minimum
---|---
J0437$-$4715 | | $1.8\times 10^{-4}$ | $1.2\times 10^{-4}$
---|---
| $4.2\times 10^{-5}$ | $3.7\times 10^{-5}$
---|---
| $4.1\times 10^{-5}$ | $3.7\times 10^{-5}$
---|---
J0613$-$0200 | | $7.8\times 10^{-5}$ | $3.2\times 10^{-5}$
---|---
| $5.6\times 10^{-5}$ | $2.9\times 10^{-5}$
---|---
| $6.0\times 10^{-5}$ | $3.1\times 10^{-5}$
---|---
J0751$+$1807 | | $4.1\times 10^{-4}$ | $1.9\times 10^{-4}$
---|---
| $2.0\times 10^{-4}$ | $1.0\times 10^{-4}$
---|---
| $2.1\times 10^{-4}$ | $1.0\times 10^{-4}$
---|---
J1012$+$5307 | | $5.9\times 10^{-5}$ | $1.9\times 10^{-5}$
---|---
| $3.8\times 10^{-5}$ | $1.8\times 10^{-5}$
---|---
| $4.0\times 10^{-5}$ | $2.4\times 10^{-5}$
---|---
J1022$+$1001 | | $1.1\times 10^{-4}$ | $4.8\times 10^{-5}$
---|---
| $9.8\times 10^{-5}$ | $4.8\times 10^{-5}$
---|---
| $1.0\times 10^{-4}$ | $4.8\times 10^{-5}$
---|---
J1600$-$3053 | | $2.1\times 10^{-4}$ | $1.3\times 10^{-4}$
---|---
| $7.8\times 10^{-5}$ | $6.1\times 10^{-5}$
---|---
| $8.4\times 10^{-5}$ | $5.7\times 10^{-5}$
---|---
J1643$-$1224 | | $1.2\times 10^{-4}$ | $6.4\times 10^{-5}$
---|---
| $6.2\times 10^{-5}$ | $3.6\times 10^{-5}$
---|---
| $6.3\times 10^{-5}$ | $3.4\times 10^{-5}$
---|---
J1713$+$0747 | | $7.3\times 10^{-5}$ | $2.8\times 10^{-5}$
---|---
| $3.2\times 10^{-5}$ | $1.8\times 10^{-5}$
---|---
| $4.1\times 10^{-5}$ | $2.2\times 10^{-5}$
---|---
J1744$-$1134 | | $2.6\times 10^{-5}$ | $1.5\times 10^{-5}$
---|---
| $1.9\times 10^{-5}$ | $8.7\times 10^{-6}$
---|---
| $2.7\times 10^{-5}$ | $1.2\times 10^{-5}$
---|---
J1857$+$0943 | | $2.0\times 10^{-4}$ | $7.2\times 10^{-5}$
---|---
| $8.7\times 10^{-5}$ | $3.5\times 10^{-5}$
---|---
| $9.1\times 10^{-5}$ | $3.6\times 10^{-5}$
---|---
J1909$-$3744 | | $1.6\times 10^{-5}$ | $6.7\times 10^{-6}$
---|---
| $1.3\times 10^{-5}$ | $6.2\times 10^{-6}$
---|---
| $1.8\times 10^{-5}$ | $1.2\times 10^{-5}$
---|---
J1939$+$2134 | | $2.8\times 10^{-6}$ | $1.1\times 10^{-6}$
---|---
| $2.7\times 10^{-6}$ | $1.1\times 10^{-6}$
---|---
| $1.7\times 10^{-5}$ | $2.6\times 10^{-6}$
---|---
J2124$-$3358 | | $1.3\times 10^{-4}$ | $2.0\times 10^{-5}$
---|---
| $1.1\times 10^{-4}$ | $2.0\times 10^{-5}$
---|---
| $1.3\times 10^{-4}$ | $2.2\times 10^{-5}$
---|---
J2145$-$0750 | | $3.3\times 10^{-5}$ | $1.0\times 10^{-5}$
---|---
| $2.5\times 10^{-5}$ | $1.0\times 10^{-5}$
---|---
| $2.5\times 10^{-5}$ | $1.0\times 10^{-5}$
---|---
Table 2: Table of DM uncertainties (in units of $\text{pc}\;\text{cm}^{-3}$).
The first column specifies the pulsar names. The second column lists the
median and minimum errors in the DM estimation using the standard Wideband
technique on Band 3 (single-band). The third and fourth columns enlist the
median and minimum DM errors using the wideband CC and CP methods
respectively. All listed values are calculated by including both 100 MHz and
200 MHz bandwidth data.
### 4.1 Description of InPTA DR1
The InPTA DR1 (Tarafdar et al., 2022) constitutes 3.5 years of data
corresponding to the observations of 14 MSPs obtained using the uGMRT Gupta et
al. (2017). The data spans from 2018 to 2021 and has a typical cadence of two
weeks, carried out during uGMRT observing cycles 34$-$35 and 37$-$40\. These
observations were performed by dividing the 30 uGMRT antennae into multiple
phased subarrays which were used to observe the same source in multiple
frequency bands simultaneously. The data were recorded in total intensity mode
(Joshi et al., 2022). The GMRT Wideband Backend (GWB: Reddy et al., 2017) was
used to record the channelized time series data in binary format, and then
RFI-mitigated and reduced to PSRFITS archives using the pinta pipeline
Susobhanan et al. (2021). During cycles 34$-$35 we observed MSPs
simultaneously in Band 3 (400$-$500 MHz), Band 4 (650$-$750 MHz) and Band 5
(1360$-$1460 MHz) of uGMRT with 100 MHz bandwidth in each band. During cycles
37$-$40, we performed simultaneous observations only in Band 3 (300$-$500 MHz)
and Band 5 (1260$-$1460 MHz) with 200 MHz bandwidth. The Band 3 data in all
cycles as well as the Band 5 data in cycles 34$-$35 (except observations
between Oct. 20, 2018 and Nov. 14, 2018) were coherently dedispersed using a
real-time pipeline (De & Gupta, 2016) to the known DM of each pulsar. uGMRT
can perform coherent dedispersion on a total bandwidth of 200 MHz only, so in
cycles 34$-$35, observations were made with 100 MHz bandwidth in each band so
that both Band 3 and Band 5 data can be coherently dedispersed (Tarafdar et
al., 2022).
The Global Positioning System (GPS) was used to measure the narrowband ToAs
and the hydrogen maser clock at the uGMRT provided a local topocentric
frequency standard. The narrowband timing residuals in the InPTA DR1 were
obtained using tempo2 Hobbs et al. (2006). The timing residuals were also
generated from the wideband likelihood method Pennucci et al. (2014); Pennucci
(2019); Alam et al. (2021); Nobleson et al. (2022) using TEMPO Nice et al.
(2015) for Band 3 data only. The epoch-wise DM corrections were introduced in
the fit. The DMX parameters were calculated from the DM time series estimated
using DMcalc (Krishnakumar et al., 2021) for the narrowband timing from low-
frequency uGMRT data obtained in Band 3 and Band 5 simultaneously. Similarly,
DMX parameters were estimated using the wideband likelihood method for
wideband timing from Band 3 data of the uGMRT.
### 4.2 DM time series
Figure 7: Consolidated DM time series for 14 InPTA DR1 pulsars. The vertical
axes depict the difference ($\Delta\text{DM}$ in units of
$10^{-4}\;\text{pc}\;\text{cm}^{-3}$) between the fiducial DM and the
corresponding estimated DMs for each pulsar estimated by applying the standard
wideband technique on Band 3 data (blue points), using the wideband CC method
(magenta points) and the wideband CP method (green points) on Band 3+5 data.
The horizontal axes depict epochs in terms of the Modified Julian Date. The
estimated DM precision of 200 MHz bandwidth data is higher than that of 100
MHz bandwidth data, hence the horizontal axes are split into two parts at MJD
58600 with dashed vertical lines, with 100 MHz bandwidth epochs on the left
side and 200 MHz bandwidth epochs on the right side of the dashed line. The
vertical axes in each panel are scaled differently for 100 MHz (left axis) and
200 MHz (right axis) epochs such that the DM variations are clearly visible.
Pulsar names and their respective fiducial DM values are mentioned at the
bottom of each respective panel.
In this section, we present the wideband DM time series obtained for 14 InPTA
DR1 pulsars using the CC and CP methods, described in sections 2.2.1 and
2.2.2, for the combination of Band 3 and Band 5 InPTA data. We compare our
Band 3+5 combination results with the Band 3 (single-band) DM time series of
these pulsars estimated using the standard wideband method (described in
section 2.1). The template epochs used for ToA and DM estimation are the same
as those used in the InPTA DR1 analysis for all pulsars, which are high-S/N
epochs selected from Cycle 41 of the uGMRT for the respective pulsars. We also
keep the template epoch to be the same for Band 3 alone, the CC, and the CP
analysis to maintain consistency. In table 2, we have listed the median and
minimum uncertainties in DMs estimated for (i) Band 3, (ii) Band 3+5 CC, and
(iii) Band 3+5 CP, including both 100 MHz and 200 MHz data. It is evident from
the listed uncertainty values that there is a significant improvement in the
DM precision when Band 3 and Band 5 data is combined using the CC method.
However, for the CP method we find that the median DM precision goes slightly
down compared to Band 3 for PSRs J1744$-$1134, J1909$-$3744, and J1939$+$2134,
while remains same for PSRs J2124$-$3358\. Overall, there is nearly two times
increment in median DM precision using CC method for most of the pulsars.
A consolidated DM time-series plot illustrating the epoch-by-epoch DM
variations for all 14 InPTA DR1 pulsars is presented in figure 7. The plot
shows the Band 3, CC, and CP method results, where the vertical axes in both
panels depict the differences between the estimated DMs and the fiducial DMs.
The DM precisions estimated from 100 MHz bandwidth (BW) data are lower as
compared to those obtained from the 200 MHz bandwidth (BW) data, hence the
scaling of the vertical axes is made separately for these two cases to make
the DM variations over both 100 MHz and 200 MHz bandwidth epochs clearly
visible. The epochs having these two different bandwidths are separated along
the horizontal axis with a vertical dashed line at MJD 58600. The fiducial DM
value for each pulsar is mentioned inside the respective panel of the figure.
We have taken fiducial DM from the InPTA DR1 analysis. Refer (Tarafdar et al.,
2022) for more details.
In both CC and CP methods, we are combining over a large gap in frequency that
can cause differences in template portrait computations which are reflected as
systematic DC offsets in DM time-series of Band3, CC, and CP methods, as seen
in figure 7. A similar DM offset was also seen in the InPTA DR1 DM time-series
estimated from the narrowband and wideband analysis, which was found to be
caused by different templates used in the techniques (Tarafdar et al., 2022).
Here, in the CP method, we first psradd Band 3 and Band 5 data and then create
an analytic template using the standard wideband technique, which means that
the spline interpolation is done over a band-gap of $\sim$760 MHz. Whereas in
the CC method, we supply separate analytic wideband templates of Band 3 and
Band 5 which are internally used within the combined chi-square metric to
estimate DMs and ToAs. This leads to the selection of different number of
eigenprofiles and tolerance values (Pennucci et al., 2014; Pennucci, 2019) in
the CC and CP methods, leading to different analytical templates.
We have provided a series of plots for 14 InPTA DR1 pulsars in appendix A to
show a comparison between DM uncertainties estimated for Band 3 alone and Band
3+5 data with CC and CP methods. As the data of 100 and 200 MHz bandwidth have
different sensitivities, they have different scales of corresponding
uncertainties, hence we have presented them in different panels for each
pulsar. In the case of 100 MHz bandwidth data, we see a significant
improvement in the median DM precision for all pulsars with Band 3+5 data
using both CC and CP methods compared to Band 3 alone results. For the 200 MHz
bandwidth data, the CC method shows much higher improvement in the median DM
precision than the CP method for all pulsars except J0437$-$4715\. For PSRs
J1744$-$1134, J1909$-$3744, and J1939$+$2134, we find a decrement in the
median DM precision using the CP method compared to Band 3 alone for 200 MHz
data, whereas the CC method shows improvement for these pulsars also. Such
decrement in DM precision using CP method is expected due to reduced phase
resolution in Band 3 which affects pulsars with sharp pulse profiles as
explained in subsection 3.1. There is also a frequency gap of $\sim 750$ MHz
between Band 3 and Band 5 data which affects the modeling of profile evolution
across band edges in CP method, hence altering the results of pulsars with
high profile evolution with radio frequency. For PSR J1643$-$1224, we observe
that the trend in DM timeseries is not in agreement betweeen Band 3 and CC or
CP methods. This effect can be explained in terms of scattering variations.
PSR J1643-1224 has a highly scattered profile, especially at low radio
frequencies. At widely separated radio frequencies, scattered pulses sample
different path lengths through the ISM, which manifests as distinct variations
in DMs (McKee et al., 2018; Singha et al., 2023; Cordes et al., 2016;
Krishnakumar et al., 2019) estimated for Band 3 and combination of Band 3+5
using CC or CP methods as seen in figure 7.
### 4.3 ToAs and Timing residuals
We show a comparison of ToA uncertainties estimated for Band 3 alone and Band
3+5 data with CC and CP methods in a series of plots for 14 InPTA DR1 pulsars
in appendix A. Similar to DM precision, we see a significant improvement in
the median ToA precision for 100 MHz bandwidth data of all pulsars with Band
3+5 data using both CC and CP methods compared to Band 3 alone data. In the
case of 200 MHz bandwidth data, there is improvement in median ToA precision
using the CC method for PSRs J0751$+$1807, J1012$+$5307, J1600$-$3053,
J1643$-$1224, J1713$+$0747, J1744$-$1134, J1857$+$0943 and J2145$-$0750 while
it stays at par with Band 3 results for other pulsars. As the ToA precision
depends on the S/N, and as Band 5 S/N is comparatively lesser than Band 3 S/N,
therefore ToAs obtained after band combination, i.e. CC or CP ToAs, are not
able to achieve a significant improvement in ToA uncertainty for 200 MHz data
of all pulsars. The CP method shows improvement in median ToA precision than
Band 3 alone for PSRs J0751$+$1807, J1012$+$5307, J1022$+$1001, J1600$-$3053,
J1643$-$1224, and J2145$-$0750, whereas it decreases ToA precision for all
other pulsars which can be attributed to the aforementioned reasons.
A consolidated wideband timing residual plot obtained from Band 3, CC, and CP
ToAs for all the 14 InPTA DR1 pulsars is shown in figure 8. The timing
procedure that we followed is the same as that described in sections 2.2.3 and
3.4. The DMEFAC and T2EFAC values are estimated separately for $100$ MHz and
$200$ MHz BW data (as they have different sensitivities) for each of the Band
3, CC, and CP ToAs. We then incorporate the T2EFAC and DMEFAC values along
with combining the 100 MHz and 200 MHz BW data ToAs to generate a global
timing solution. The details of timing parameter estimates obtained after
wideband timing using Band 3, CC, and CP method ToAs for all 14 pulsars are
mentioned in table 3, where the fit parameters are chosen as per InPTA DR1
narrowband timing (Tarafdar et al., 2022). We find that the precision of the
fitted parameters are improved when the timing is done on Band 3+5 data using
both CC and CP ToAs for most of the pulsars.
Figure 8: Consolidated wideband timing residuals for 14 InPTA DR1 pulsars. The
post-fit wideband timing residuals (in units of $\mu\text{s}$) obtained from
Band 3 ToAs (blue points), and using the wideband CC (magenta points) and CP
(green points) methods on the Band 3+5 combination ToAs by implementing
wideband timing technique in tempo2, are plotted against the corresponding
epochs. Pulsar names and their respective post-fit weighted RMS residuals for
Band 3, CC, and CP methods (Band 3+5) are mentioned at the bottom of the
respective panels. The vertical axes in each panel are scaled differently for
100 MHz (left axis) and 200 MHz (right axis) epochs and a vertical dashed line
is added such that the residual variations are clearly visible. Epochs are
depicted in terms of the Modified Julian Date on the consolidated horizontal
axes.
Pulsar Method Timing Parameters RA (hh:mm:ss) DEC ($\deg$:mm:ss) PMRA (mas/yr)
A1 (lt s) F0 ($\mathrm{s}^{-1}$) F1 ($\mathrm{s}^{-2}$) PB (s) Value Error
($\mathrm{s}$) Value Error ($\mathrm{s}$) Value Error ((mas/yr)) Value Error
($\mathrm{lt}\ \mathrm{s}$) Value Error ($\mathrm{s}^{-1}$) Value Error
($\mathrm{s}^{-2}$) Value Error ($\mathrm{s}$) J0437$-$4715 Band 3
$4:37:16.0432$ $5.9\times 10^{-4}$ $-47:15:09.991$ $9.5\times 10^{-3}$ —— ——
$3.36673$ $1.2\times 10^{-5}$ $173.6879451409$ $3.7\times 10^{-10}$ —— ——
$5.741045808$ $5.0\times 10^{-9}$ CC $4:37:16.04316$ $8.3\times 10^{-5}$
$-47:15:09.991$ $1.1\times 10^{-3}$ —— —— $3.366759$ $1.6\times 10^{-6}$
$173.68794514023$ $4.2\times 10^{-11}$ —— —— $5.7410458046$ $8.1\times
10^{-10}$ CP $4:37:16.04316$ $9.1\times 10^{-5}$ $-47:15:09.991$ $1.2\times
10^{-3}$ —— —— $3.366759$ $1.8\times 10^{-6}$ $173.68794514020$ $4.5\times
10^{-11}$ —— —— $5.74104580486$ $8.8\times 10^{-10}$ J0613$-$0200 —— —— —— ——
—— —— —— —— —— —— —— —— —— —— —— J0751$+$1807 —— —— —— —— —— —— —— —— —— —— ——
—— —— —— —— J1012$+$5307 Band 3 —— —— —— —— —— —— —— —— $190.26783422728$
$1.0\times 10^{-11}$ —— —— —— —— CC —— —— —— —— —— —— —— —— $190.267834227285$
$4.8\times 10^{-12}$ —— —— —— —— CP —— —— —— —— —— —— —— —— $190.267834227289$
$4.8\times 10^{-12}$ —— —— —— —— J1022+1001 —— —— —— —— —— —— —— —— —— —— ——
—— —— —— —— J1600$-$3053 Band 3 —— —— —— —— —— —— —— —— $277.937706735932$
$6.8\times 10^{-11}$ —— —— —— —— CC —— —— —— —— —— —— —— ——
$277.9377067359737$ $8.9\times 10^{-12}$ —— —— —— —— CP —— —— —— —— —— —— ——
—— $277.937706735978$ $8.8\times 10^{-12}$ —— —— —— —— J1643$-$1224 Band 3 ——
—— —— —— $5.5$ $1.0$ $25.072598$ $1.7\times 10^{-6}$ $216.37333684399$
$1.2\times 10^{-11}$ $-8.599\times 10^{-16}$ $8.2\times 10^{-19}$
$147.01739775$ $4.0\times 10^{-8}$ CC —— —— —— —— $7.1$ $0.4$ $25.0725981$
$8.9\times 10^{-7}$ $216.373336843944$ $6.3\times 10^{-12}$ $-8.637\times
10^{-16}$ $3.2\times 10^{-19}$ $147.01739786$ $2.1\times 10^{-8}$ CP —— —— ——
—— $7.2$ $0.4$ $25.0725978$ $9.3\times 10^{-7}$ $216.373336843935$ $6.5\times
10^{-12}$ $-8.635\times 10^{-16}$ $3.3\times 10^{-19}$ $147.01739786$
$2.3\times 10^{-8}$ J1713$+$0747 Band 3 —— —— —— —— —— —— $32.3424310$
$5.7\times 10^{-7}$ —— —— —— —— $67.825130884$ $2.5\times 10^{-9}$ CC —— —— ——
—— —— —— $32.3424298$ $2.1\times 10^{-7}$ —— —— —— —— $67.8251308817$
$8.9\times 10^{-10}$ CP —— —— —— —— —— —— $32.3424297$ $2.5\times 10^{-7}$ ——
—— —— —— $67.825130881$ $1.0\times 10^{-9}$ J1744$-$1134 Band 3 —— —— —— —— ——
—— —— —— $245.42611950378$ $3.0\times 10^{-11}$ —— —— —— —— CC —— —— —— —— ——
—— —— —— $245.42611950381$ $2.6\times 10^{-11}$ —— —— —— —— CP —— —— —— —— ——
—— —— —— $245.42611950381$ $1.6\times 10^{-11}$ —— —— —— —— J1857$+$0943 Band
3 —— —— —— —— —— —— —— —— $186.49407816357$ $1.5\times 10^{-11}$ $-6.22\times
10^{-16}$ $1.2\times 10^{-18}$ —— —— CC —— —— —— —— —— —— —— ——
$186.494078163546$ $3.6\times 10^{-12}$ $-6.205\times 10^{-16}$ $1.6\times
10^{-19}$ —— —— CP —— —— —— —— —— —— —— —— $186.494078163547$ $2.8\times
10^{-12}$ $-6.204\times 10^{-16}$ $1.4\times 10^{-19}$ —— —— J1909$-$3744 Band
3 —— —— —— —— —— —— —— —— $339.31568666040$ $1.0\times 10^{-11}$
$-1.6142\times 10^{-15}$ $4.4\times 10^{-19}$ —— —— CC —— —— —— —— —— —— —— ——
$339.315686660410$ $2.7\times 10^{-12}$ $-1.6149\times 10^{-15}$ $1.7\times
10^{-19}$ —— —— CP —— —— —— —— —— —— —— —— $339.315686660412$ $2.3\times
10^{-12}$ $-1.6152\times 10^{-15}$ $1.4\times 10^{-19}$ —— —— J1939$+$2134
Band 3 —— —— —— —— —— —— —— —— $641.92820961498$ $1.2\times 10^{-11}$
$-4.33096\times 10^{-14}$ $7.5\times 10^{-19}$ —— —— CC —— —— —— —— —— —— ——
—— $641.928209615009$ $6.7\times 10^{-12}$ $-4.33091\times 10^{-14}$
$4.1\times 10^{-19}$ —— —— CP —— —— —— —— —— —— —— —— $641.928209615108$
$6.7\times 10^{-12}$ $-4.33058\times 10^{-14}$ $2.6\times 10^{-19}$ —— ——
J2124-3358 —— —— —— —— —— —— —— —— —— —— —— —— —— —— —— J2145$-$0750 Band 3 ——
—— —— —— —— —— —— —— $62.295887797432$ $1.7\times 10^{-12}$ $-1.155\times
10^{-16}$ $1.0\times 10^{-19}$ $6.8389026151$ $1.0\times 10^{-10}$ CC —— —— ——
—— —— —— —— —— $62.2958877974360$ $9.1\times 10^{-13}$ $-1.1565\times
10^{-16}$ $4.5\times 10^{-20}$ $6.838902615$ $5.5\times 10^{-11}$ CP —— —— ——
—— —— —— —— —— $62.29588779744$ $1.0\times 10^{-12}$ $-1.1547\times 10^{-16}$
$4.2\times 10^{-20}$ $6.8389026152$ $4.4\times 10^{-11}$
Table 3: Table of timing parameters for 14 InPTA DR1 pulsars. The first column
lists the pulsar names. The second column lists the methodology used to obtain
ToAs which are then used for estimating the timing parameters in separate
rows, namely the Band 3, CC, and CP methods. Columns three to nine represent
various fitted pulsar timing parameters, their units, and their uncertainties.
The choice of the fitted parameters for each pulsar is consistent with the
timing analysis of Tarafdar et al. (2022), where no timing parameters are fit
for PSRs J0613$-$0200, J0751+1807, J1022+1001, and J2124$-$3358.
## 5 Summary and Conclusions
Table of wRMS and $\chi^{2}$
---
| Pulsar Name | Method | ToA $\chi^{2}$ | DM $\chi^{2}$ | DOF | Total red. $\chi^{2}$ | wRMS
---|---|---|---|---|---|---
J0437$-$4715 | B3 | 0.087 | 10.196 | 9 | 1.143 | 0.201
CC | 0.631 | 9.777 | 9 | 1.156 | 0.217
CP | 0.585 | 8.928 | 9 | 1.057 | 0.225
J0613$-$0200 | B3 | 7.034 | 17.621 | 21 | 1.174 | 0.963
CC | 2.729 | 31.737 | 21 | 1.641 | 0.555
CP | 7.271 | 21.607 | 21 | 1.375 | 0.972
J0751$+$1807 | B3 | 0.845 | 22.384 | 23 | 1.01 | 0.654
CC | 0.785 | 27.978 | 23 | 1.251 | 0.348
CP | 1.582 | 35.636 | 23 | 1.618 | 0.302
J1012$+$5307 | B3 | 6.088 | 28.914 | 22 | 1.591 | 0.77
CC | 6.631 | 24.177 | 22 | 1.4 | 0.53
CP | 2.558 | 21.554 | 22 | 1.096 | 0.306
J1022$+$1001 | B3 | 0.398 | 21.283 | 24 | 0.903 | 0.264
CC | 0.448 | 20.798 | 24 | 0.885 | 0.214
CP | 0.734 | 32.475 | 24 | 1.384 | 0.286
J1600$-$3053 | B3 | 1.72 | 19.029 | 21 | 0.988 | 1.018
CC | 9.953 | 16.848 | 21 | 1.276 | 1.804
CP | 13.781 | 13.766 | 21 | 1.312 | 2.178
J1643$-$1224 | B3 | 9.984 | 164.109 | 63 | 2.763 | 0.869
CC | 4.317 | 63.795 | 63 | 1.081 | 0.257
CP | 3.799 | 63.037 | 63 | 1.061 | 0.269
J1713$+$0747 | B3 | 5.203 | 49.358 | 43 | 1.269 | 0.399
CC | 18.088 | 48.758 | 43 | 1.555 | 0.417
CP | 25.567 | 111.641 | 43 | 3.191 | 0.516
J1744$-$1134 | B3 | 3.859 | 11.284 | 14 | 1.082 | 0.49
CC | 3.413 | 11.203 | 14 | 1.044 | 0.434
CP | 2.348 | 10.309 | 14 | 0.904 | 0.255
J1857$+$0943 | B3 | 1.315 | 31.023 | 34 | 0.951 | 0.323
CC | 6.395 | 38.517 | 34 | 1.321 | 0.352
CP | 11.967 | 33.109 | 34 | 1.326 | 0.589
J1909$-$3744 | B3 | 5.944 | 50.301 | 56 | 1.004 | 0.235
CC | 11.395 | 45.588 | 56 | 1.018 | 0.326
CP | 19.304 | 35.852 | 56 | 0.985 | 0.471
J1939$+$2134 | B3 | 2.827 | 61.999 | 63 | 1.029 | 0.148
CC | 9.856 | 69.861 | 63 | 1.265 | 0.253
CP | 23.987 | 74.188 | 63 | 1.558 | 0.329
J2124$-$3358 | B3 | 4.094 | 63.051 | 39 | 1.722 | 0.515
CC | 4.697 | 67.579 | 39 | 1.853 | 0.324
CP | 5.501 | 73.437 | 39 | 2.024 | 0.54
J2145$-$0750 | B3 | 2.284 | 40.514 | 44 | 0.973 | 0.231
CC | 11.141 | 60.259 | 44 | 1.623 | 0.335
CP | 19.252 | 92.857 | 44 | 2.548 | 0.472
Table 4: Table of reduced chi-squares of 14 InPTA DR1 pulsars. The first
column lists the pulsar names. The second column lists the methodology used to
obtain ToAs and DM, namely the Band 3, CC, and CP methods. Columns three to
seven represent ToA component of the chi-square, DM component of the chi-
square, degrees of freedom (DOF), the total reduced chi-square and the wRMS of
the timing residuals.
In this work, we have developed two independent novel techniques, namely the
Combined Portrait (CP) and Combined Chi-squared (CC) methods, to combine data
simultaneously recorded in two non-contiguous frequency bands within the
paradigm of wideband technique (Pennucci et al., 2014; Pennucci, 2019) to
obtain a single DM and ToA per epoch encapsulating information contained in
both the bands. In the CP method, we create an auxiliary dataset by combining
the data of two frequency bands to create a single 2-dimensional analytic
template containing the information on pulse profile evolution with frequency.
This template is then used for cross-correlation with other epochs to obtain
wideband DMs and ToAs. In the CC method, we create separate 2-dimensional
analytic templates for both the bands, and these are integrated within a
combined Fourier-domain $\chi^{2}$-statistic and perform a global fit over the
whole frequency space to generate a single wideband DM and ToA per epoch. We
have applied these two techniques to 14 millisecond pulsars observed under the
InPTA campaign using uGMRT in Band 3 and Band 5 frequency bands
simultaneously, and they are included in the first data release of the InPTA
(Tarafdar et al., 2022).
We obtained high-precision DMs and ToAs for Band 3+5 data using these
techniques. We observe that combining the data having 100 MHz bandwidth in
each band showed consistent improvement in DM and ToA precision for all 14
pulsars and that both CC and CP are performing equally well. However, the
combination of data having 200 MHz bandwidth in each band shows
inconsistencies using the CP method. This is due to the reduction in the
number of phase bins in Band 3 which is essential to combine it with Band 5
data to create a single analytic template of Band 3+5. Another caveat of the
CP method is the band gap of $\sim 760$ MHz, which needs to be interpolated
over, between two bands leading to probable imperfections in the modeling of
profile evolution with frequency across the bands. The combination of data
having 200 MHz bandwidth in each band using the CC method shows much higher
improvement in DM and ToA precision than the CP method and Band 3 alone. We
plan to extend these techniques further to combine simultaneously recorded
data of multiple non-contiguous bands in future work.
We have also incorporated the wideband likelihood in tempo2 using libstempo
for the first time. We perform the wideband timing analysis on ToAs obtained
from the CC and CP methods along with Band 3 ToAs for comparison. We achieved
the weighted RMS ToA residuals in the range of 214 ns to 1.8 $\mu$s for ToAs
obtained from the CC method, while in the range of 225 ns to 2.1 $\mu$s for
ToAs obtained from the CP method for the whole spectrum of InPTA DR1 pulsars.
We observe an improvement in the precision of fitted timing parameters with
Band 3+5 combination compared to Band 3 alone for all pulsars. Since we are
combining data of multiple frequency bands, we may require frequency-dependent
parameters to obtain a better fit for our timing solutions. This will be
explored in future work.
We observe that the DM chi-square, obtained from the DM part of the
likelihood, is larger than the ToA chi-square (see Table 4). We suspect that
this could be related to the way we estimate DMEFAC and T2EFAC parameters. We
plan to investigate this further in future work where we will apply Bayesian
methods to estimate optimum DMEFAC and T2EFAC parameter values.
## 6 Discussion and Future directions
The extension of the wideband technique to multiple non-contiguous frequency
bands demonstrated in this work, is likely to be useful in largely removing
chromatic noise sources, such as variations in the pulse profile, DM and
scattering, in precision timing experiments like pulsar timing arrays. This
technique not only improves the ToA precision significantly by accumulating
the signal over the entire frequency range of combined bands, it also takes
care of DM noise across the bands by incorporating DM-chromatic noise
measurements in the timing likelihood naturally (see Appendix B). This
restricts the noise analysis of PTA data to just the time-independent and
time-correlated achromatic and scattering noise sources, greatly simplifying
and constraining these noise models. This has implications both for the
computational needs as well as the sensitivity of PTA data for a GW search.
Other precision timing experiments targeted at measuring timing noise,
parameters of relativistic binary systems and tests of General Theory of
Relativity are also likely to benefit from this extension of the standard
wideband technique. With large upcoming and future telescopes, such as the SKA
(Kramer & Stappers, 2015; Janssen et al., 2015) and DSA (Hallinan et al.,
2019), likely to employ simultaneous observations over multiple bands with
frequency coverage as large as 5 GHz, we expect this extended technique or its
variants to be widely used in the future.
## Software
RFIclean (Maan et al., 2021), DSPSR (Straten & Bailes, 2011), PSRCHIVE (Hotan
et al., 2004), pinta (Susobhanan et al., 2021), PulsePortraiture (Pennucci et
al., 2014; Pennucci, 2019), tempo2 (Hobbs et al., 2006; Edwards et al., 2006),
libstempo (Vallisneri, 2020), tempo (Nice et al., 2015), numpy (Harris et al.,
2020), scipy (Virtanen et al., 2020), matplotlib (Hunter, 2007)
## Acknowledgements
We thank the staff of the GMRT who made our observations possible. GMRT is
operated by the National Centre for Radio Astrophysics of the Tata Institute
of Fundamental Research. AKP is supported by CSIR fellowship Grant number
$09/0079(15784)/2022$-EMR-I. BCJ acknowledges support from Raja Ramanna Chair
(Track - I) grant from the Department of Atomic Energy, Government of India.
KN is supported by the Birla Institute of Technology and Science Institute
fellowship. AS is supported by the NANOGrav NSF Physics Frontiers Center
(awards 1430284 and 2020265). DD acknowledges the support from the Department
of Atomic Energy, Government of India through ‘Apex Project - Advance Research
and Education in Mathematical Sciences at IMSc’. MB acknowledges the support
from the Department of Atomic Energy, Government of India through ‘Apex
Project - Advance Research and Education in Mathematical Sciences at IMSc’. YG
and BCJ acknowledges support from the Department of Atomic Energy, Government
of India, under project number 12-R&D-TFR-5.02-0700. TK is partially supported
by the JSPS Overseas Challenge Program for Young Researchers. AmS is supported
by CSIR fellowship Grant number $09/1001(12656)/2021$-EMR-I and DST-ICPS
T-641. KT is partially supported by JSPS KAKENHI Grant Numbers 20H00180,
21H01130, and 21H04467 and the ISM Cooperative Research Program
(2023-ISMCRP-2046). We thank Scott Ransom for his suggestions that improved
the manuscript.
## Data Availability
The python scripts used for the analysis are available in
https://github.com/AvinashKumarPaladi/Multiband-extension-of-Wideband-Timing-
Technique. The data underlying this article will be shared on reasonable
request to the corresponding author.
## References
* Ahuja et al. (2005) Ahuja A. L., Gupta Y., Mitra D., Kembhavi A. K., 2005, MNRAS, 357, 1013
* Alam et al. (2021) Alam M. F., et al., 2021, ApJS, 252, 5
* Cordes et al. (2016) Cordes J. M., Shannon R. M., Stinebring D. R., 2016, ApJ, 817, 16
* Curyło et al. (2023) Curyło M., et al., 2023, ApJ, 944, 128
* De & Gupta (2016) De K., Gupta Y., 2016, Experimental Astronomy, 41, 67
* Edwards et al. (2006) Edwards R. T., Hobbs G. B., Manchester R. N., 2006, Monthly Notices of the Royal Astronomical Society, 372, 1549
* Fonseca et al. (2021) Fonseca E., et al., 2021, ApJ, 915, L12
* Foster & Backer (1990) Foster R. S., Backer D. C., 1990, The Astrophysical Journal, 361, 300
* Gupta et al. (2017) Gupta Y., et al., 2017, Current Science, 113, 707
* Hallinan et al. (2019) Hallinan G., et al., 2019, in Bulletin of the American Astronomical Society. p. 255 (arXiv:1907.07648), doi:10.48550/arXiv.1907.07648
* Harris et al. (2020) Harris C. R., et al., 2020, Nature, 585, 357
* Hobbs (2013) Hobbs G., 2013, Classical and Quantum Gravity, 30, 224007
* Hobbs et al. (2006) Hobbs G. B., Edwards R. T., Manchester R. N., 2006, Monthly Notices of the Royal Astronomical Society, 369, 655
* Hobbs et al. (2010) Hobbs G., et al., 2010, Classical and Quantum Gravity, 27, 084013
* Hobbs et al. (2020) Hobbs G., et al., 2020, Publ. Astron. Soc. Australia, 37, e012
* Hotan et al. (2004) Hotan A. W., van Straten W., Manchester R. N., 2004, Publications of the Astronomical Society of Australia, 21, 302
* Hunter (2007) Hunter J. D., 2007, Computing in Science & Engineering, 9, 90
* Jacoby et al. (2003) Jacoby B. A., Bailes M., van Kerkwijk M. H., Ord S., Hotan A., Kulkarni S. R., Anderson S. B., 2003, ApJ, 599, L99
* Janssen et al. (2015) Janssen G., et al., 2015, in Advancing Astrophysics with the Square Kilometre Array (AASKA14). p. 37 (arXiv:1501.00127), doi:10.22323/1.215.0037
* Johnston et al. (2021) Johnston S., et al., 2021, MNRAS, 502, 1253
* Joshi et al. (2018) Joshi B. C., et al., 2018, Journal of Astrophysics and Astronomy, 39, 51
* Joshi et al. (2022) Joshi B. C., et al., 2022, Journal of Astrophysics and Astronomy, 43, 98
* Kendall (1938) Kendall M. G., 1938, Biometrika, 30, 81
* Kramer & Champion (2013) Kramer M., Champion D. J., 2013, Classical and Quantum Gravity, 30, 224009
* Kramer & Stappers (2015) Kramer M., Stappers B., 2015, in Advancing Astrophysics with the Square Kilometre Array (AASKA14). p. 36 (arXiv:1507.04423), doi:10.22323/1.215.0036
* Krishnakumar et al. (2019) Krishnakumar M. A., Maan Y., Joshi B. C., Manoharan P. K., 2019, ApJ, 878, 130
* Krishnakumar et al. (2021) Krishnakumar M. A., et al., 2021, A&A, 651, A5
* Lee (2016) Lee K. J., 2016, in Qain L., Li D., eds, Astronomical Society of the Pacific Conference Series Vol. 502, Frontiers in Radio Astronomy and FAST Early Sciences Symposium 2015. p. 19
* Liu et al. (2014) Liu K., et al., 2014, MNRAS, 443, 3752
* Liu et al. (2020) Liu K., et al., 2020, MNRAS, 499, 2276
* Lorimer & Kramer (2012) Lorimer D. R., Kramer M., 2012, Handbook of Pulsar Astronomy. Cambridge University Press
* Luo et al. (2021) Luo J., et al., 2021, ApJ, 911, 45
* Maan et al. (2021) Maan Y., van Leeuwen J., Vohl D., 2021, Astronomy & Astrophysics, 650, A80
* McKee et al. (2018) McKee J. W., Lyne A. G., Stappers B. W., Bassa C. G., Jordan C. A., 2018, MNRAS, 479, 4216
* McLaughlin (2013) McLaughlin M. A., 2013, Classical and Quantum Gravity, 30, 224008
* Miles et al. (2023) Miles M. T., et al., 2023, MNRAS, 519, 3976
* Nice et al. (2015) Nice D., et al., 2015, Tempo: Pulsar timing data analysis, http://tempo.sourceforge.net/
* Nobleson et al. (2022) Nobleson K., et al., 2022, Monthly Notices of the Royal Astronomical Society, 512, 1234
* Pennucci (2019) Pennucci T. T., 2019, The Astrophysical Journal, 871, 34
* Pennucci et al. (2014) Pennucci T. T., Demorest P. B., Ransom S. M., 2014, The Astrophysical Journal, 790, 93
* Perera et al. (2019) Perera B. B. P., et al., 2019, MNRAS, 490, 4666
* Reddy et al. (2017) Reddy S. H., et al., 2017, Journal of Astronomical Instrumentation, 06, 1641011
* Singha et al. (2023) Singha J., Indian Pulsar Timing Array Collaboration Indian Pulsar Timing Array Collaboration 2023, in prep.
* Srivastava et al. (2023) Srivastava A., et al., 2023, arXiv e-prints, p. arXiv:2303.12105
* Straten & Bailes (2011) Straten W. V., Bailes M., 2011, Publications of the Astronomical Society of Australia, 28, 1
* Susobhanan et al. (2021) Susobhanan A., et al., 2021, Publications of the Astronomical Society of Australia, 38, e017
* Tarafdar et al. (2022) Tarafdar P., et al., 2022, Publ. Astron. Soc. Australia, 39, e053
* Taylor (1992) Taylor J. H., 1992, Philosophical Transactions of the Royal Society of London. Series A: Physical and Engineering Sciences, 341, 117
* Vallisneri (2020) Vallisneri M., 2020, libstempo: Python wrapper for Tempo2, Astrophysics Source Code Library, record ascl:2002.017 (ascl:2002.017)
* Verbiest et al. (2016) Verbiest J. P. W., et al., 2016, MNRAS, 458, 1267
* Virtanen et al. (2020) Virtanen P., et al., 2020, Nature Methods, 17, 261
## Appendix A DM and ToA Uncertainty Comparison Plots
The DM and ToA uncertainty ($\sigma_{\text{DM}}$ and $\sigma_{\text{ToA}}$)
comparison between Band 3 and Band 3+5 data – CC and CP methods is presented
here for all the 14 InPTA DR1 pulsars. Since the sensitivity of both
bandwidths is different, it is not visually feasible to plot all of them on
the same scale, hence we have shown them in different panels.
Figure 9: A comparison of Wideband DM (left panel) and ToA (right panel)
uncertainties obtained from the CC and CP methods for Band 3+5 data is shown
against those obtained from the traditional Band 3 data for PSR J0437$-$4715\.
This pulsar has only 100MHz InPTA data before Cycle 40 of the uGMRT. The Band
3+5 uncertainties estimated with the CC (magenta points) and CP (green points)
methods are shown on the vertical axis and Band 3 uncertainties on the
horizontal axis (in units of $10^{-5}\;\text{pc}\;\text{cm}^{-3}$ for DM
uncertainties, and in units of $\mu\text{s}$ for ToA uncertainties). The
diagonal dashed-dotted line shows the $y=x$ while the magenta and green dashed
lines indicate median DM (left panel) and ToA (right panel) uncertainties for
CC and CP method respectively. Points lying below the diagonal line indicate
that the obtained precision with Band 3+5 combination is better than Band 3
(single-band) results and vice versa. Hence, we obtain an overall increase in
the DM and ToA precisions from both CP and CC methods for PSR J0437$-$4715.
Figure 10: A comparison of Wideband DM (left panel) and ToA (right panel)
uncertainties obtained from the CC and CP methods for Band 3+5 data is shown
against those obtained from the traditional Band 3 data for PSR J0613$-$0200\.
The results are shown for both 100 MHz and 200 MHz bandwidth data. The curves
and colors displayed in the figure conform to the details mentioned in the
legend at the bottom. Figure 11: A comparison of Wideband DM (left panel) and
ToA (right panel) uncertainties obtained from the CC and CP methods for Band
3+5 data is shown against those obtained from the traditional Band 3 data for
PSR J0751$+$1807\. The results are shown for both 100 MHz and 200 MHz
bandwidth data. The curves and colors displayed in the figure conform to the
details mentioned in the legend at the bottom. Figure 12: A comparison of
Wideband DM (left panel) and ToA (right panel) uncertainties obtained from the
CC and CP methods for Band 3+5 data is shown against those obtained from the
traditional Band 3 data for PSR J1012$+$5307\. The results are shown for both
100 MHz and 200 MHz bandwidth data. The curves and colors displayed in the
figure conform to the details mentioned in the legend at the bottom. Figure
13: A comparison of Wideband DM (left panel) and ToA (right panel)
uncertainties obtained from the CC and CP methods for Band 3+5 data is shown
against those obtained from the traditional Band 3 data for PSR J1022$+$1001\.
The results are shown for both 100 MHz and 200 MHz bandwidth data. The curves
and colors displayed in the figure conform to the details mentioned in the
legend at the bottom. Figure 14: A comparison of Wideband DM (left panel) and
ToA (right panel) uncertainties obtained from the CC and CP methods for Band
3+5 data is shown against those obtained from the traditional Band 3 data for
PSR J1600$-$3053\. The results are shown for both 100 MHz and 200 MHz
bandwidth data. The curves and colors displayed in the figure conform to the
details mentioned in the legend at the bottom. Figure 15: A comparison of
Wideband DM (left panel) and ToA (right panel) uncertainties obtained from the
CC and CP methods for Band 3+5 data is shown against those obtained from the
traditional Band 3 data for PSR J1643$-$1224\. The results are shown for both
100 MHz and 200 MHz bandwidth data. The curves and colors displayed in the
figure conform to the details mentioned in the legend at the bottom. Figure
16: A comparison of Wideband DM (left panel) and ToA (right panel)
uncertainties obtained from the CC and CP methods for Band 3+5 data is shown
against those obtained from the traditional Band 3 data for PSR J1713$+$0747\.
The results are shown for both 100 MHz and 200 MHz bandwidth data. The curves
and colors displayed in the figure conform to the details mentioned in the
legend at the bottom. Figure 17: A comparison of Wideband DM (left panel) and
ToA (right panel) uncertainties obtained from the CC and CP methods for Band
3+5 data is shown against those obtained from the traditional Band 3 data for
PSR J1744$-$1134 with only 200 MHz data as it wasn’t observed in earlier
cycles. The curves and colors displayed in the figure conform to the details
mentioned in the legend at the bottom. Figure 18: A comparison of Wideband DM
(left panel) and ToA (right panel) uncertainties obtained from the CC and CP
methods for Band 3+5 data is shown against those obtained from the traditional
Band 3 data for PSR J1857$+$0943\. The results are shown for both 100 MHz and
200 MHz bandwidth data. The curves and colors displayed in the figure conform
to the details mentioned in the legend at the bottom. Figure 19: A comparison
of Wideband DM (left panel) and ToA (right panel) uncertainties obtained from
the CC and CP methods for Band 3+5 data is shown against those obtained from
the traditional Band 3 data for PSR J1909$-$3744\. The results are shown for
both 100 MHz and 200 MHz bandwidth data. The curves and colors displayed in
the figure conform to the details mentioned in the legend at the bottom.
Figure 20: A comparison of Wideband DM (left panel) and ToA (right panel)
uncertainties obtained from the CC and CP methods for Band 3+5 data is shown
against those obtained from the traditional Band 3 data for PSR J1939$+$2134\.
The results are shown for both 100 MHz and 200 MHz bandwidth data. The curves
and colors displayed in the figure conform to the details mentioned in the
legend at the bottom. Figure 21: A comparison of Wideband DM (left panel) and
ToA (right panel) uncertainties obtained from the CC and CP methods for Band
3+5 data is shown against those obtained from the traditional Band 3 data for
PSR J2124$-$3358\. The results are shown for both 100 MHz and 200 MHz
bandwidth data. The curves and colors displayed in the figure conform to the
details mentioned in the legend at the bottom. Figure 22: A comparison of
Wideband DM (left panel) and ToA (right panel) uncertainties obtained from the
CC and CP methods for Band 3+5 data is shown against those obtained from the
traditional Band 3 data for PSR J2145$-$0750\. The results are shown for both
100 MHz and 200 MHz bandwidth data. The curves and colors displayed in the
figure conform to the details mentioned in the legend at the bottom.
## Appendix B Implementing the wideband likelihood using tempo2 and libstempo
The wideband timing residuals $\delta t$ can be modeled as
$\delta t=M\epsilon+r$ (12)
The product of the timing model design matrix $M$ with small offsets in the
timing model parameters $\epsilon$ describes the systematic residuals from
subtracting the timing model. $r$ represents the uncorrelated noise in the
residuals.
The Narrowband likelihood for the timing residuals is given by
$p\left(\delta
t|\epsilon,\phi\right)=\frac{\exp\left(-\frac{1}{2}r^{T}N^{-1}r\right)}{\sqrt{|2\pi
N|}}$ (13) $N_{ij}=(E_{k(i)}^{2}\sigma_{i}^{2}+Q_{k(i)}^{2})\delta_{ij}$ (14)
where $\phi$ comprises of EFAC $E_{k(i)}$ and EQUAD $Q_{k(i)}$. $\sigma_{i}$
are the uncertainties in ToAs.
In wideband timing we have an additional likelihood term that includes the DMX
priors,
$p\left(\epsilon^{DMX}\;\delta
D,E^{DM}\right)=\frac{e^{-\frac{1}{2}\left(\left(\epsilon^{DMX}-\delta
D\right)^{T}N^{DM^{-1}}\left(\epsilon^{DMX}-\delta
D\right)\right)}}{\sqrt{|2\pi N^{DM}|}}$ (15)
$N_{ij}^{DM}=(E_{k(i)}^{DM}\sigma_{i}^{DM})^{2}\delta_{ij}$ (16)
where $E^{DM}$ is the DM EFAC and $\sigma_{i}^{DM}$ is the DM error.
$\epsilon^{DMX}$ represents subset of timing model offsets $\epsilon$ that
describe the piece-wise constant DMX model. $\delta D$ is the vector
containing difference of DM measurements with respect to the fiducial dm.
The complete wideband timing likelihood is given by the product of both
narrowband likelihood and likelihood containing DM priors, Alam et al. (2021)
$p\left(\epsilon,\phi,E^{DM}|\delta t,\delta D\right)\propto p\left(\delta
t|\epsilon,\phi\right)\times p\left(\epsilon^{DMX}|\delta D,E^{DM}\right)\ $
(17)
This wideband likelihood is implemented in python using libstempo, a python
wrapper for tempo2. We obtained the design matrix $M$ from libstempo by giving
the par file and ToAs for a particular pulsar as inputs. This design matrix is
then extended to account for $\delta D$ and $\epsilon^{DMX}$. Using the
extended design matrix, we estimate the timing residuals and DMX parameters
from equation 17 using the Generalized Least Squares (GLS) method.
|
0}\int_{0}^{T}\|f^{\delta}(\varphi^{\delta})\|_{L^{2}(\Gamma(t))}^{2}\leq C,$
and hence $\varphi(\varphi)\in L^{2}_{L^{2}}$. Now by the uniqueness of weak
limits and a suitable variant of the dominated convergence theorem for
evolving surfaces (see [18], Theorem B.2), one finds that
$f(\varphi)=\tilde{f}$.
## 6 Proof of uniqueness
### 6.1 Uniqueness for the regular potential
In this section we prove the uniqueness of solutions to (4.1)-(4.3). As a
preliminary result, we note that by elliptic regularity theory one has that
$\displaystyle\int_{0}^{T}\|\varphi\|_{H^{2}(\Gamma(t))}^{2}\leq
C\int_{0}^{T}\left(\|\mu\|_{L^{2}(\Gamma(t))}^{2}+\|F^{\prime}(\varphi)\|_{L^{2}(\Gamma(t))}^{2}\right).$
(6.1)
This $L^{2}_{H^{2}}$ regularity of $\varphi$ is invaluable for proving the
uniqueness of solutions, as it allows one to eliminate $\mu$ from (4.1). To
see this we notice that for almost all $t\in[0,T]$, and all
$\boldsymbol{\phi}\in\mathbf{V}_{\sigma}(t)$
$\mathbf{c}_{2}(\mu,\varphi,\boldsymbol{\phi})=\int_{\Gamma(t)}\mu\nabla_{\Gamma}\varphi\cdot\boldsymbol{\phi}=\int_{\Gamma(t)}\left(-\varepsilon\Delta_{\Gamma}\varphi+\frac{1}{\varepsilon}F^{\prime}(\varphi)\right)\nabla_{\Gamma}\varphi\cdot\boldsymbol{\phi}.$
Now, as observed in the derivation, one formally calculates that
$-\varepsilon\Delta_{\Gamma}\varphi\nabla_{\Gamma}\varphi=-\varepsilon\nabla_{\Gamma}\cdot(\nabla_{\Gamma}\varphi\otimes\nabla_{\Gamma}\varphi)+\frac{\varepsilon}{2}\nabla_{\Gamma}|\nabla_{\Gamma}\varphi|^{2}-\varepsilon\left(\nabla_{\Gamma}\varphi\cdot\mathbb{H}\nabla_{\Gamma}\varphi\right)\boldsymbol{\nu},$
almost everywhere on $\Gamma(t)$. Thus one finds
$\mathbf{c}_{2}(\mu,\varphi,\boldsymbol{\phi})=\int_{\Gamma(t)}\nabla_{\Gamma}\left(\frac{\varepsilon}{2}|\nabla_{\Gamma}\varphi|^{2}+\frac{1}{\varepsilon}F(\varphi)\right)\cdot\boldsymbol{\phi}-\varepsilon\nabla_{\Gamma}\cdot(\nabla_{\Gamma}\varphi\otimes\nabla_{\Gamma}\varphi)\cdot\boldsymbol{\phi},$
where the normal term has vanished as $\boldsymbol{\phi}$ is tangential. Now
using integration by parts, and the fact that $\boldsymbol{\phi}$ is
solenoidal, it is clear that
$\displaystyle\int_{\Gamma(t)}\nabla_{\Gamma}\left(\frac{\varepsilon}{2}|\nabla_{\Gamma}\varphi|^{2}+\frac{1}{\varepsilon}F(\varphi)\right)\cdot\boldsymbol{\phi}=\int_{\Gamma(t)}\nabla_{\Gamma}\cdot\left(\frac{\varepsilon}{2}|\nabla_{\Gamma}\varphi|^{2}\boldsymbol{\phi}+\frac{1}{\varepsilon}F(\varphi)\boldsymbol{\phi}\right)-\int_{\Gamma(t)}\left(\frac{\varepsilon}{2}|\nabla_{\Gamma}\varphi|^{2}+\frac{1}{\varepsilon}F(\varphi)\right)\nabla_{\Gamma}\cdot\boldsymbol{\phi}=0,$
$\displaystyle-\varepsilon\int_{\Gamma(t)}\nabla_{\Gamma}\cdot(\nabla_{\Gamma}\varphi\otimes\nabla_{\Gamma}\varphi))\cdot\boldsymbol{\phi}=\varepsilon\int_{\Gamma(t)}(\nabla_{\Gamma}\varphi\otimes\nabla_{\Gamma}\varphi):\nabla_{\Gamma}\boldsymbol{\phi}.$
Hence we find
$\mathbf{c}_{2}(\mu,\varphi,\boldsymbol{\phi})=\varepsilon\mathbf{c}_{3}(\varphi,\varphi,\boldsymbol{\phi}),$
where the trilinear form $\mathbf{c}_{3}$ is defined as
$\mathbf{c}_{3}(t;\phi,\psi,\boldsymbol{\chi}):=\int_{\Gamma(t)}(\nabla_{\Gamma}\phi\otimes\nabla_{\Gamma}\psi):\nabla_{\Gamma}\boldsymbol{\chi}.$
The structure of this proof is similar to that in [29], with relevant
modifications for an evolving surface - as discussed in Appendix B.
###### Theorem 6.1.
Let $\Gamma(t),F$ be such that the assumptions in Theorem 4.1 hold. Moreover,
assume $F_{2}^{\prime}$ is Lipschitz continuous. Then the solution triple,
$(\varphi,\mu,\mathbf{u}_{T})$ solving (4.1)-(4.3) is unique.
The first step is to observe that if we have two solution triples,
$(\varphi^{i},\mu^{i},\mathbf{u}_{T}^{i})$, $i=1,2$, and defining
$\displaystyle\bar{\varphi}:=\varphi^{1}-\varphi^{2},$
$\displaystyle\bar{\mu}:=\mu^{1}-\mu^{2},$
$\displaystyle\bar{\mathbf{u}_{T}}:=\mathbf{u}_{T}^{1}-\mathbf{u}_{T}^{2},$
then these solve the system
$\displaystyle\begin{split}\mathbf{m}_{*}\left(\partial^{\circ}{\bar{\mathbf{u}_{T}}},\boldsymbol{\phi}\right)+\hat{\mathbf{a}}(\eta(\varphi^{1}),\mathbf{u}_{T}^{1},\boldsymbol{\phi})-\hat{\mathbf{a}}(\eta(\varphi^{2}),\mathbf{u}_{T}^{2},\boldsymbol{\phi})+\mathbf{c}_{1}(\mathbf{u}_{T}^{1},\mathbf{u}_{T}^{1},\boldsymbol{\phi})-\mathbf{c}_{1}(\mathbf{u}_{T}^{2},\mathbf{u}_{T}^{2},\boldsymbol{\phi})+\mathbf{l}(\bar{\mathbf{u}_{T}},\boldsymbol{\phi})+\mathbf{d}_{1}(\bar{\mathbf{u}_{T}},\boldsymbol{\phi})\\\
+\mathbf{d}_{2}(\eta(\varphi^{1})-\eta(\varphi^{2}),\boldsymbol{\phi})=\varepsilon\mathbf{c}_{3}(\varphi^{1},\varphi^{1},\boldsymbol{\phi})-\varepsilon\mathbf{c}_{3}(\varphi^{2},\varphi^{2},\boldsymbol{\phi})\end{split},$
(6.2) $\displaystyle
m_{*}\left(\partial^{\circ}{\bar{\varphi}},\phi\right)+a(\bar{\mu},\phi)+\mathbf{c}_{2}(\phi,\varphi^{1},\mathbf{u}_{T}^{1})-\mathbf{c}_{2}(\phi,\varphi^{2},\mathbf{u}_{T}^{2})+\mathbf{c}_{2}(\phi,\bar{\varphi},\widetilde{\mathbf{u}_{T}})=0,$
(6.3) $\displaystyle m(\bar{\mu},\phi)=\varepsilon
a(\bar{\varphi},\phi)+\frac{1}{\varepsilon}m(F^{\prime}(\varphi^{1})-F^{\prime}(\varphi^{2}),\phi),$
(6.4)
for almost all $t\in[0,T]$, and all $\phi\in
H^{1}(\Gamma(t)),\boldsymbol{\phi}\in\mathbf{V}_{\sigma}(t)$. This proof
firstly requires obtaining bounds for
$\bar{\varphi},P_{\mathcal{K}}\bar{\mathbf{u}_{T}},P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}$
in appropriate norms. We refer to Appendix B for definitions and properties of
the operators $P_{\mathcal{K}},P_{\mathcal{K}^{\perp}}$.
The proof relies on proving the following differential inequalities.
###### Lemma 6.2.
$P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}$ is such that
$\frac{1}{2}\frac{d}{dt}\|P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}\|_{\perp}^{2}+\frac{\eta_{*}}{2}\|P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}\|_{\mathbf{L}^{2}(\Gamma(t))}^{2}\leq
C\|P_{\mathcal{K}}\bar{\mathbf{u}_{T}}\|_{\mathbf{L}^{2}(\Gamma(t))}^{2}+\frac{\varepsilon}{4}\|\nabla_{\Gamma}\bar{\varphi}\|_{L^{2}(\Gamma(t))}\\\
+K_{1}(t)\|P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}\|_{\mathbf{L}^{2}(\Gamma(t))}^{2}+C_{\log}(\bar{\varphi},\mathbf{u}_{T}^{1})\|P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}\|_{\perp},$
(6.5)
where
$K_{1}(t)=C\left(\|\mathbb{E}(\mathbf{u}_{T}^{1})\|_{\mathbf{L}^{2}(\Gamma(t))}^{2}+\|\mathbf{u}_{T}^{1}\|_{\mathbf{L}^{4}(\Gamma(t))}^{4}+\|\mathbf{u}_{T}^{2}\|_{\mathbf{L}^{4}(\Gamma(t))}^{4}+\|\varphi^{1}\|_{H^{1,4}(\Gamma(t))}^{4}+\|\varphi^{2}\|_{H^{1,4}(\Gamma(t))}^{4}\right),$
and
$C_{\log}(\bar{\varphi},\mathbf{u}_{T}^{1})=C\log\left(\frac{C_{2}(\|\nabla_{\Gamma}\bar{\varphi}\|_{L^{2}(\Gamma(t))}+\|\bar{\varphi}\|_{H^{2}(\Gamma(t))})}{\|\nabla_{\Gamma}\bar{\varphi}\|_{L^{2}(\Gamma(t))}^{2}}\right)^{\frac{1}{2}}\|\nabla_{\Gamma}\bar{\varphi}\|_{L^{2}(\Gamma(t))}\|\mathbb{E}(\mathbf{u}_{T}^{1})\|_{\mathbf{L}^{2}(\Gamma(t))}.$
$P_{\mathcal{K}}\bar{\mathbf{u}_{T}}$ is such that
$\displaystyle\frac{1}{2}\frac{d}{dt}\|P_{\mathcal{K}}\bar{\mathbf{u}_{T}}\|_{\mathbf{L}^{2}(\Gamma(t))}^{2}\leq\frac{\eta_{*}}{4}\|P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}\|_{\mathbf{L}^{2}(\Gamma(t))}^{2}+\frac{\varepsilon}{4}\|\nabla_{\Gamma}\bar{\varphi}\|_{L^{2}(\Gamma(t))}^{2}+K_{2}(t)\|P_{\mathcal{K}}\bar{\mathbf{u}_{T}}\|_{\mathbf{L}^{2}(\Gamma(t))}^{2},$
(6.6)
where
$K_{2}(t)=C\left(1+\|\widetilde{\mathbf{u}_{T}}\|_{\mathbf{L}^{4}(\Gamma(t))}^{2}+\|\mathbf{u}_{T}^{1}\|_{\mathbf{L}^{4}(\Gamma(t))}^{2}+\|\mathbf{u}_{T}^{2}\|_{\mathbf{L}^{4}(\Gamma(t))}^{2}+\|\varphi^{1}\|_{H^{1,4}(\Gamma(t))}^{4}+\|\varphi^{2}\|_{H^{1,4}(\Gamma(t))}^{4}\right).$
Lastly, $\bar{\varphi}$ is such that
$\frac{1}{2}\frac{d}{dt}\|\bar{\varphi}\|_{-1}^{2}+\varepsilon\|\nabla_{\Gamma}\bar{\varphi}\|_{L^{2}(\Gamma(t))}^{2}\leq\frac{\varepsilon}{4}\|\nabla_{\Gamma}\bar{\varphi}\|_{L^{2}(\Gamma(t))}^{2}+\frac{\eta_{*}}{4}\|P_{\mathcal{K}}\bar{\mathbf{u}_{T}}\|_{\mathbf{L}^{2}(\Gamma(t))}^{2}+\frac{\eta_{*}}{4}\|P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}\|_{\mathbf{L}^{2}(\Gamma(t))}^{2}\\\
+K_{3}(t)\|\bar{\varphi}\|_{-1}^{2},$ (6.7)
where
$K_{3}(t)=C\left(1+\|\widetilde{\mathbf{u}_{T}}\|_{\mathbf{L}^{\infty}(\Gamma(t))}^{2}+\|\mathbf{u}_{T}^{1}\|_{\mathbf{L}^{4}(\Gamma(t))}^{2}+\|\varphi^{2}\|_{L^{\infty}(\Gamma(t))}^{2}\right).$
###### Proof of uniqueness.
With these bounds we are now in a position to show uniqueness. Taking the sum
of (6.5), (6.6), (6.7) one finds
$\frac{1}{2}\frac{d}{dt}\left(\|P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}\|_{\perp}^{2}+\|P_{\mathcal{K}}\bar{\mathbf{u}_{T}}\|_{\mathbf{L}^{2}(\Gamma(t))}^{2}+\|\bar{\varphi}\|_{-1}^{2}\right)+\frac{\varepsilon}{4}\|\nabla_{\Gamma}\bar{\varphi}\|_{L^{2}(\Gamma(t))}^{2}\\\
\leq
K(t)\left(\|P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}\|_{\perp}^{2}+\|P_{\mathcal{K}}\bar{\mathbf{u}_{T}}\|_{\mathbf{L}^{2}(\Gamma(t))}^{2}+\|\bar{\varphi}\|_{-1}^{2}\right)\\\
+C_{\log}(\bar{\varphi},\mathbf{u}_{T}^{1})\left(\|P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}\|_{\perp}^{2}+\|P_{\mathcal{K}}\bar{\mathbf{u}_{T}}\|_{\mathbf{L}^{2}(\Gamma(t))}^{2}+\|\bar{\varphi}\|_{-1}^{2}\right)^{\frac{1}{2}},$
where
$K(t)=K_{1}(t)+K_{2}(t)+K_{3}(t).$
Now recall that we have $\mathbf{u}_{T}^{i}\in L^{\infty}_{\mathbf{L}^{2}}\cap
L^{2}_{\mathbf{H}^{1}},\varphi^{i}\in L^{\infty}_{H^{1}}\cap L^{2}_{H^{2}}$,
and in particular this implies $\varphi^{i}\in L^{4}_{H^{1,4}}$ as
$\displaystyle\int_{0}^{T}\|\varphi^{i}\|_{H^{1,4}(\Gamma(t))}^{4}$
$\displaystyle\leq
C\int_{0}^{T}\|\varphi^{i}\|_{H^{1}(\Gamma(t))}^{2}\|\varphi^{i}\|_{H^{2}(\Gamma(t))}^{2}$
$\displaystyle\leq
C\left(\sup_{t\in[0,T]}\|\varphi^{i}\|_{H^{1}(\Gamma(t))}^{2}\right)\int_{0}^{T}\|\varphi^{i}\|_{H^{2}(\Gamma(t))}^{2}<\infty.$
Moreover
$\|P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}(0)\|_{\perp}^{2}+\|P_{\mathcal{K}}\bar{\mathbf{u}_{T}}(0)\|_{\mathbf{L}^{2}(\Gamma(t))}+\|\bar{\varphi}(0)\|_{-1}^{2}=0$
by definition, we see that we may use Lemma 3.10 to see that
$\|P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}\|_{\perp},\|P_{\mathcal{K}}\bar{\mathbf{u}_{T}}\|_{\mathbf{L}^{2}(\Gamma(t))},\|\bar{\varphi}\|_{-1}$
all vanish on $[0,T]$. In particular, since $\|\cdot\|_{\perp}$ and
$\|\cdot\|_{-1}$ are norms on $\mathcal{K}^{\perp}$ and $\\{z\in
H^{-1}(\Gamma(t))\mid m_{*}(z,1)=0\\}$ respectively we see that
$P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}=0,\quad
P_{\mathcal{K}}\bar{\mathbf{u}_{T}}=0,\quad\bar{\varphi}=0,$
and as
$\bar{\mathbf{u}_{T}}=P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}+P_{\mathcal{K}}\bar{\mathbf{u}_{T}}$
we see $\bar{\mathbf{u}_{T}}$ vanishes on $[0,T]$ also. From this one can
readily show that $\bar{\mu}=0$ on $[0,T]$ and hence determine uniqueness of
weak solutions. ∎
#### 6.1.1 Proof of Lemma 6.2
###### Proof.
Firstly we establish the most complicated inequality, (6.5). Testing (6.2)
with $\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}}$, as defined in Appendix B, and
rewriting terms in a suitable way we find that
$\mathbf{m}_{*}\left(\partial^{\circ}{\bar{\mathbf{u}_{T}}},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}}\right)+\hat{\mathbf{a}}(\eta(\varphi^{1})-\eta(\varphi^{2}),\mathbf{u}_{T}^{1},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}})+\hat{\mathbf{a}}(\eta(\varphi^{2}),\bar{\mathbf{u}_{T}},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}})+\mathbf{c}_{1}(\bar{\mathbf{u}_{T}},\mathbf{u}_{T}^{1},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}})+\mathbf{c}_{1}(\mathbf{u}_{T}^{2},\bar{\mathbf{u}_{T}},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}})\\\
+\mathbf{l}(\bar{\mathbf{u}_{T}},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}})+\mathbf{d}_{1}(\bar{\mathbf{u}_{T}},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}})+\mathbf{d}_{2}(\eta(\varphi^{1})-\eta(\varphi^{2}),\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}})=\varepsilon\mathbf{c}_{3}(\bar{\varphi},\varphi^{1},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}})+\varepsilon\mathbf{c}_{3}(\varphi^{2},\bar{\varphi},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}}).$
(6.8)
Firstly, we claim that
$\mathbf{m}_{*}\left(\partial^{\circ}{\bar{\mathbf{u}_{T}}},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}}\right)=\frac{1}{2}\frac{d}{dt}\|P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}\|_{\perp}^{2}+\frac{1}{2}\mathbf{b}(\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}})-\mathbf{m}(P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}HV_{N},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}})\\\
+\sum_{i=1}^{N_{K}}\mathbf{m}(\bar{\mathbf{u}_{T}},\boldsymbol{\kappa}_{i})\mathbf{m}(\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}},\partial^{\circ}\boldsymbol{\kappa}_{i}),$
(6.9)
where we are using the notation from Appendix B. To see this we write
$\mathbf{m}_{*}\left(\partial^{\circ}{\bar{\mathbf{u}_{T}}},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}}\right)=\frac{d}{dt}\mathbf{m}(\bar{\mathbf{u}_{T}},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}})-\mathbf{m}\left(\bar{\mathbf{u}_{T}},\partial^{\circ}{\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}}}\right)-\mathbf{m}\left(\bar{\mathbf{u}_{T}}HV_{N},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}}\right).$
Since $\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}}\in\mathcal{K}^{\perp}(t)$ we
see that
$\mathbf{m}(\bar{\mathbf{u}_{T}},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}})=\mathbf{m}(P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}})$,
and hence by the definition of $\mathcal{S}^{\perp}$
$\frac{d}{dt}\mathbf{m}(\bar{\mathbf{u}_{T}},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}})=\frac{d}{dt}\|P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}\|_{\perp}^{2},$
and so all that remains is to consider
$\mathbf{m}\left(\bar{\mathbf{u}_{T}},\partial^{\circ}{\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}}}\right)$.
Firstly, we split this term into two parts
$\mathbf{m}\left(\bar{\mathbf{u}_{T}},\partial^{\circ}{\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}}}\right)=\mathbf{m}\left(P_{\mathcal{K}}\bar{\mathbf{u}_{T}},\partial^{\circ}{\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}}}\right)+\mathbf{m}\left(P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}},\partial^{\circ}{\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}}}\right).$
We write this first term as
$\mathbf{m}\left(P_{\mathcal{K}}\bar{\mathbf{u}_{T}},\partial^{\circ}{\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}}}\right)=\frac{d}{dt}\mathbf{m}\left(P_{\mathcal{K}}\bar{\mathbf{u}_{T}},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}}\right)-\mathbf{m}\left(P_{\mathcal{K}}\bar{\mathbf{u}_{T}}HV_{N},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}}\right)-\mathbf{m}_{*}\left(\partial^{\circ}P_{\mathcal{K}}\bar{\mathbf{u}_{T}},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}}\right),$
where clearly
$\mathbf{m}\left(P_{\mathcal{K}}\bar{\mathbf{u}_{T}},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}}\right)=0$.
Next we use (B.5) to see that
$\mathbf{m}_{*}\left(\partial^{\circ}P_{\mathcal{K}}\bar{\mathbf{u}_{T}},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}}\right)=\mathbf{m}_{*}\left(\partial^{\circ}\bar{\mathbf{u}_{T}},P_{\mathcal{K}}\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}}\right)\\\
+\sum_{i=1}^{N_{K}}\left[\mathbf{m}(\bar{\mathbf{u}_{T}},\partial^{\circ}\boldsymbol{\kappa}_{i})+\mathbf{m}(\bar{\mathbf{u}_{T}}HV_{N},\boldsymbol{\kappa}_{i})\right]\mathbf{m}(\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}},\boldsymbol{\kappa}_{i})+\mathbf{m}(\bar{\mathbf{u}_{T}},\boldsymbol{\kappa}_{i})\mathbf{m}(\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}},\partial^{\circ}\boldsymbol{\kappa}_{i}),$
which simplifies to
$\mathbf{m}_{*}\left(\partial^{\circ}P_{\mathcal{K}}\bar{\mathbf{u}_{T}},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}}\right)=\sum_{i=1}^{N_{K}}\mathbf{m}(\bar{\mathbf{u}_{T}},\boldsymbol{\kappa}_{i})\mathbf{m}(\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}},\partial^{\circ}\boldsymbol{\kappa}_{i}).$
Next we write
$\mathbf{m}\left(P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}},\partial^{\circ}{\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}}}\right)=\mathbf{a}(\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}},\partial^{\circ}\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}})=\frac{1}{2}\frac{d}{dt}\|P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}\|_{\perp}^{2}-\frac{1}{2}\mathbf{b}(\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}}).$
Combining all of this together one obtains (6.9).
Next we rewrite the second $\hat{\mathbf{a}}$ term in (6.8) by using
integration by parts. To do this we note that
$\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}}\in L^{2}_{H^{2}}$ and we have the
bound (B.4). Integration by parts yields
$\hat{\mathbf{a}}(\eta(\varphi^{2}),\bar{\mathbf{u}_{T}},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}})=\int_{\Gamma(t)}\eta(\varphi^{2})\mathbb{E}(\bar{\mathbf{u}_{T}}):\mathbb{E}(\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}})=-\int_{\Gamma(t)}\bar{\mathbf{u}_{T}}\cdot\mathbb{P}\nabla_{\Gamma}\cdot\left(\eta(\varphi^{2})\mathbb{E}(\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}})\right)\\\
=-\int_{\Gamma(t)}\eta^{\prime}(\varphi^{2})\bar{\mathbf{u}_{T}}\cdot\mathbb{E}(\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}})\nabla_{\Gamma}\varphi^{2}+\int_{\Gamma(t)}\eta(\varphi^{2})\bar{\mathbf{u}_{T}}\cdot
P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}},$
where we have used the fact that
$-\mathbb{P}\nabla_{\Gamma}\cdot\mathbb{E}(\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}})=P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}$
a.e. on $\Gamma(t)$. Hence one obtains
$\hat{\mathbf{a}}(\eta(\varphi^{2}),\bar{\mathbf{u}_{T}},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}})=-\int_{\Gamma(t)}\eta^{\prime}(\varphi^{2})\bar{\mathbf{u}_{T}}\cdot\mathbb{E}(\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}})\nabla_{\Gamma}\varphi^{2}+\int_{\Gamma(t)}\eta(\varphi^{2})P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}\cdot
P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}\\\
+\int_{\Gamma(t)}\eta(\varphi^{2})P_{\mathcal{K}}\bar{\mathbf{u}_{T}}\cdot
P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}.$ (6.10)
Next we bound the other $\hat{\mathbf{a}}$ term. To do this we write
$\hat{\mathbf{a}}(\eta(\varphi^{1})-\eta(\varphi^{2}),\mathbf{u}_{T}^{1},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}})=\int_{\Gamma(t)}\left(\int_{0}^{1}\eta^{\prime}(su^{1}+(1-s)\varphi^{2})\bar{\varphi}\,ds\right)\mathbb{E}(\mathbf{u}_{T}^{1}):\mathbb{E}(\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}}),$
and noting the boundedness of $\eta^{\prime}$ we obtain the bound
$\hat{\mathbf{a}}(\eta(\varphi^{1})-\eta(\varphi^{2}),\mathbf{u}_{T}^{1},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}})\leq
C\|\bar{\varphi}\|_{L^{\infty}(\Gamma(t))}\|\mathbb{E}(\mathbf{u}_{T}^{1})\|_{\mathbf{L}^{2}(\Gamma(t))}\|P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}\|_{\perp}.$
Now we use Lemma 3.11, and Poincaré’s inequality, to see that
$\|\bar{\varphi}\|_{L^{\infty}(\Gamma(t))}\leq
C\|\nabla_{\Gamma}\bar{\varphi}\|_{L^{2}(\Gamma(t))}\left(1+\log\left(1+\frac{C\|\bar{\varphi}\|_{H^{2}(\Gamma(t))}}{\|\nabla_{\Gamma}\bar{\varphi}\|_{L^{2}(\Gamma(t))}}\right)^{\frac{1}{2}}\right),$
where we note from (6.1) that $\bar{\varphi}$ is sufficiently smooth for this.
Now from the $L^{\infty}_{H^{1}}$ bounds for $\varphi_{1},\varphi_{2}$ we see
that there is a constant, $C_{1}$, such that
$\|\nabla_{\Gamma}\bar{\varphi}\|_{L^{2}(\Gamma(t))}\leq C_{1}$ for almost all
$t\in[0,T]$. Hence one can find a sufficiently large constant $C_{2}$ so that
$\log\left(1+\frac{C\|\bar{\varphi}\|_{H^{2}(\Gamma(t))}}{\|\nabla_{\Gamma}\bar{\varphi}\|_{L^{2}(\Gamma(t))}}\right)\leq\log\left(\frac{C_{2}(\|\nabla_{\Gamma}\bar{\varphi}\|_{L^{2}(\Gamma)}+\|\bar{\varphi}\|_{H^{2}(\Gamma(t))})}{\|\nabla_{\Gamma}\bar{\varphi}\|_{L^{2}(\Gamma(t))}^{2}}\right).$
Moreover for $C_{2}>1$ this logarithmic term is positive, and we will
ultimately be able to apply Lemma 3.10. All in all this gives us the bound
$|\hat{\mathbf{a}}(\eta(\varphi^{1})-\eta(\varphi^{2}),\mathbf{u}_{T}^{1},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}})|\leq\frac{\varepsilon}{8}\|\nabla_{\Gamma}\bar{\varphi}\|_{L^{2}(\Gamma(t))}^{2}+C\|\mathbb{E}(\mathbf{u}_{T}^{1})\|_{\mathbf{L}^{2}(\Gamma(t))}^{2}\|P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}\|_{\perp}^{2}\\\
+C\log\left(\frac{C_{2}(\|\nabla_{\Gamma}\bar{\varphi}\|_{L^{2}(\Gamma(t))}+\|\bar{\varphi}\|_{H^{2}(\Gamma(t))})}{\|\nabla_{\Gamma}\bar{\varphi}\|_{L^{2}(\Gamma(t))}^{2}}\right)^{\frac{1}{2}}\|\nabla_{\Gamma}\bar{\varphi}\|_{L^{2}(\Gamma(t))}\|\mathbb{E}(\mathbf{u}_{T}^{1})\|_{\mathbf{L}^{2}(\Gamma(t))}\|P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}\|_{\perp},$
(6.11)
where we have used Young’s inequality where appropriate.
Now we bound the $\mathbf{c}_{1}$ terms. Firstly we note that
$|\mathbf{c}_{1}(\bar{\mathbf{u}_{T}},\mathbf{u}_{T}^{1},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}})|=|\mathbf{c}_{1}(\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}},\mathbf{u}_{T}^{1},\bar{\mathbf{u}_{T}})|\leq\|\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}}\|_{\mathbf{H}^{1,4}(\Gamma(t))}\|\mathbf{u}_{T}^{1}\|_{\mathbf{L}^{4}(\Gamma(t))}\|\bar{\mathbf{u}_{T}}\|_{\mathbf{L}^{2}(\Gamma(t))}.$
We recall the interpolation inequality777See [10], and note that a $C^{3}$
surface is sufficiently smooth for this to hold. This can be extended to
evolving surfaces with a time independent constant as in [45] Lemma 3.4.,
$\|\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}}\|_{\mathbf{H}^{1,4}(\Gamma(t))}\leq
C\|\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}}\|_{\mathbf{H}^{1}(\Gamma(t))}^{\frac{1}{2}}\|\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}}\|_{\mathbf{H}^{2}(\Gamma(t))}^{\frac{1}{2}},$
and use (B.1), (B.4) to see that
$\|\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}}\|_{\mathbf{H}^{1,4}(\Gamma(t))}\leq
C\|P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}\|_{\perp}^{\frac{1}{2}}\|P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}\|_{\mathbf{L}^{2}(\Gamma(t))}^{\frac{1}{2}}\leq
C\|P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}\|_{\perp}^{\frac{1}{2}}\|\bar{\mathbf{u}_{T}}\|_{\mathbf{L}^{2}(\Gamma(t))}^{\frac{1}{2}}.$
Hence we observe that
$|\mathbf{c}_{1}(\bar{\mathbf{u}_{T}},\mathbf{u}_{T}^{1},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}})|\leq
C\|\mathbf{u}_{T}^{1}\|_{\mathbf{L}^{4}(\Gamma(t))}\|P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}\|_{\perp}^{\frac{1}{2}}\|\bar{\mathbf{u}_{T}}\|_{\mathbf{L}^{2}(\Gamma(t))}^{\frac{3}{2}},$
and hence Young’s inequality yields
$\displaystyle|\mathbf{c}_{1}(\bar{\mathbf{u}_{T}},\mathbf{u}_{T}^{1},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}})|\leq
C\|\mathbf{u}_{T}^{1}\|_{\mathbf{L}^{4}(\Gamma(t))}^{4}\|P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}\|_{\perp}^{2}+\frac{\eta_{*}}{14}\|\bar{\mathbf{u}_{T}}\|_{\mathbf{L}^{2}(\Gamma(t))}^{2}.$
(6.12)
An identical argument yields
$\displaystyle|\mathbf{c}_{1}(\mathbf{u}_{T}^{2},\bar{\mathbf{u}_{T}},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}})|\leq
C\|\mathbf{u}_{T}^{2}\|_{\mathbf{L}^{4}(\Gamma(t))}^{4}\|P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}\|_{\perp}^{2}+\frac{\eta_{*}}{14}\|\bar{\mathbf{u}_{T}}\|_{\mathbf{L}^{2}(\Gamma(t))}^{2}.$
(6.13)
We now turn to the contributions from the evolution of the surface, that is
the terms involving $\mathbf{l},\mathbf{d}_{1},\mathbf{d}_{2}$, which would
vanish for a stationary surface. The simplest of these terms is
$\displaystyle\mathbf{l}(\bar{\mathbf{u}_{T}},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}})\leq
C\|\bar{\mathbf{u}_{T}}\|_{\mathbf{L}^{2}(\Gamma(t))}\|\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}}\|_{\mathbf{L}^{2}(\Gamma(t))}\leq\frac{\eta_{*}}{14}\|\bar{\mathbf{u}_{T}}\|_{\mathbf{L}^{2}(\Gamma(t))}^{2}+C\|P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}\|_{\perp}^{2},$
(6.14)
where we have use Young’s inequality and (B.1). Next we look at
$\displaystyle\mathbf{d}_{2}(\eta(\varphi^{1})-\eta(\varphi^{2}),\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}})\leq
C\|\bar{\varphi}\|_{L^{2}(\Gamma(t))}\|P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}\|_{\perp}\leq\frac{\varepsilon}{12}\|\nabla_{\Gamma}\bar{\varphi}\|_{L^{2}(\Gamma(t))}^{2}+C\|P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}\|_{\perp}^{2},$
(6.15)
which follows similarly to the above inequality, but we have also use the
Lipschitz continuity of $\eta(\cdot)$, Poincaré’s inequality, and the uniform
bounds on $\widetilde{\mathbf{u}_{T}}$. Finally to bound the $\mathbf{d}_{1}$
term we see
$\mathbf{d}_{1}(\bar{\mathbf{u}_{T}},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}})=\mathbf{c}_{1}(\bar{\mathbf{u}_{T}},\widetilde{\mathbf{u}_{T}},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}})+\mathbf{c}_{1}(\widetilde{\mathbf{u}_{T}},\bar{\mathbf{u}_{T}},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}})\\\
=-\mathbf{c}_{1}(\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}},\widetilde{\mathbf{u}_{T}},\bar{\mathbf{u}_{T}})-\mathbf{c}_{1}(\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}},\bar{\mathbf{u}_{T}},\widetilde{\mathbf{u}_{T}})+2\mathbf{m}(\bar{\mathbf{u}_{T}}HV_{N},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}}),$
where we have used the antisymmetry of $\mathbf{c}_{1}$ (for solenoidal
functions), and the extra term comes from the fact that
$\nabla_{\Gamma}\cdot\widetilde{\mathbf{u}_{T}}=-HV_{N}$. From this one
readily sees that
$|\mathbf{d}_{1}(\bar{\mathbf{u}_{T}},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}})|\leq
C\|P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}\|_{\perp}\|\bar{\mathbf{u}_{T}}\|_{\mathbf{L}^{2}(\Gamma(t))}\|\widetilde{\mathbf{u}_{T}}\|_{\mathbf{L}^{\infty}(\Gamma(t))}+C\|\bar{\mathbf{u}_{T}}\|_{\mathbf{L}^{2}(\Gamma(t))}\|P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}\|_{\perp},$
where we have used (B.1). This clearly yields
$\displaystyle|\mathbf{d}_{1}(\bar{\mathbf{u}_{T}},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}})|\leq
C\|\widetilde{\mathbf{u}_{T}}\|_{\mathbf{L}^{\infty}(\Gamma(t))}^{2}\|P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}\|_{\perp}^{2}+\frac{\eta_{*}}{14}\|\bar{\mathbf{u}_{T}}\|_{\mathbf{L}^{2}(\Gamma(t))}^{2}.$
(6.16)
The last terms for us to bound are the $\mathbf{c}_{3}$ contributions, from
which one readily sees from Young’s inequality that
$\displaystyle\varepsilon|\mathbf{c}_{3}(\bar{\varphi},\varphi^{1},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}})|+\varepsilon|\mathbf{c}_{3}(\varphi^{2},\bar{\varphi},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}})|\leq\frac{\varepsilon}{12}\|\nabla_{\Gamma}\bar{\varphi}\|_{L^{2}(\Gamma(t))}^{2}+C\left(\|\varphi^{1}\|_{H^{1,4}(\Gamma(t))}^{2}+\|\varphi^{2}\|_{H^{1,4}(\Gamma(t))}^{2}\right)\|\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}}\|_{\mathbf{H}^{1,4}(\Gamma(t))}^{2}.$
Again using the above interpolation inequality we see that
$\|\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}}\|_{\mathbf{H}^{1,4}(\Gamma(t))}^{2}\leq
C\|P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}\|_{\perp}\|\bar{\mathbf{u}_{T}}\|_{\mathbf{L}^{2}(\Gamma(t))}.$
Hence using Young’s inequality we find
$\varepsilon|\mathbf{c}_{3}(\bar{\varphi},\varphi^{1},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}})|+\varepsilon|\mathbf{c}_{3}(\varphi^{2},\bar{\varphi},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}})|\leq\frac{\varepsilon}{12}\|\nabla_{\Gamma}\bar{\varphi}\|_{L^{2}(\Gamma(t))}^{2}+C\left(\|\varphi^{1}\|_{H^{1,4}(\Gamma(t))}^{4}+\|\varphi^{2}\|_{H^{1,4}(\Gamma(t))}^{4}\right)\|P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}\|_{\perp}^{2}\\\
+\frac{\eta_{*}}{14}\|\bar{\mathbf{u}_{T}}\|_{\mathbf{L}^{2}(\Gamma(t))}^{2},$
(6.17)
where we note that $\varphi^{i}\in L^{4}_{H^{1,4}}$ as shown above.
Now we use (6.9)-(6.17) in (6.8) to see that
$\frac{1}{2}\frac{d}{dt}\|P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}\|_{\perp}^{2}+\int_{\Gamma(t)}\eta(\varphi^{2})P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}\cdot
P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}\leq\left|\int_{\Gamma(t)}\eta^{\prime}(\varphi^{2})\bar{\mathbf{u}_{T}}\cdot\mathbb{E}(\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}})\nabla_{\Gamma}\varphi^{2}\right|+\left|\int_{\Gamma(t)}\eta(\varphi^{2})P_{\mathcal{K}}\bar{\mathbf{u}_{T}}\cdot
P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}\right|\\\
+\frac{\varepsilon}{4}\|\nabla_{\Gamma}\bar{\varphi}\|_{L^{2}(\Gamma(t))}^{2}+C\|\mathbb{E}(\mathbf{u}_{T}^{1})\|_{\mathbf{L}^{2}(\Gamma(t))}^{2}\|P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}\|_{\perp}^{2}+C\left(\|\mathbf{u}_{T}^{1}\|_{\mathbf{L}^{4}(\Gamma(t))}^{4}+\|\mathbf{u}_{T}^{2}\|_{\mathbf{L}^{4}(\Gamma(t))}^{4}\right)\|P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}\|_{\perp}^{2}+\frac{6\eta_{*}}{14}\|\bar{\mathbf{u}_{T}}\|_{\mathbf{L}^{2}(\Gamma(t))}^{2}\\\
+\frac{\eta_{*}}{14}\|P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}\|_{\mathbf{L}^{2}(\Gamma(t))}^{2}+C\left(\|\varphi^{1}\|_{H^{1,4}(\Gamma(t))}^{4}+\|\varphi^{2}\|_{H^{1,4}(\Gamma(t))}^{4}\right)\|P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}\|_{\perp}^{2}\\\
+C\log\left(\frac{C_{2}(\|\nabla_{\Gamma}\bar{\varphi}\|_{L^{2}(\Gamma(t))}+\|\bar{\varphi}\|_{H^{2}(\Gamma(t))})}{\|\nabla_{\Gamma}\bar{\varphi}\|_{L^{2}(\Gamma(t))}^{2}}\right)^{\frac{1}{2}}\|\nabla_{\Gamma}\bar{\varphi}\|_{L^{2}(\Gamma(t))}\|\mathbb{E}(\mathbf{u}_{T}^{1})\|_{\mathbf{L}^{2}(\Gamma(t))}\|P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}\|_{\perp},$
where we have also bounded the extra terms in (6.9) as
$\frac{1}{2}|\mathbf{b}(\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}})|+|\mathbf{m}(P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}HV_{N},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}})|+\sum_{i=1}^{N_{K}}|\mathbf{m}(\bar{\mathbf{u}_{T}},\boldsymbol{\kappa}_{i})\mathbf{m}(\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}},\partial^{\circ}\boldsymbol{\kappa}_{i})|\leq
C\|P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}\|_{\perp}^{2}\\\
+\frac{\eta_{*}}{14}\|P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}\|_{\mathbf{L}^{2}(\Gamma(t))}^{2}+\frac{\eta_{*}}{14}\|\bar{\mathbf{u}_{T}}\|_{\mathbf{L}^{2}(\Gamma(t))}^{2},$
where we use (B.1) and the uniform bounds on
$H,V_{N},\boldsymbol{\kappa}_{i},\partial^{\circ}\boldsymbol{\kappa}_{i}$
appropriately. Now recalling that $\eta_{*}\leq\eta(\cdot)\leq\eta^{*}$,
$\eta^{\prime}$ is bounded, and that by construction
$\displaystyle\|\bar{\mathbf{u}_{T}}\|_{\mathbf{L}^{2}(\Gamma(t))}^{2}=\|P_{\mathcal{K}}\bar{\mathbf{u}_{T}}\|_{\mathbf{L}^{2}(\Gamma(t))}^{2}+\|P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}\|_{\mathbf{L}^{2}(\Gamma(t))}^{2},$
(6.18)
one can use Young’s inequality to arrive at (6.5).
Now we establish (6.6). Testing (6.2) with
$P_{\mathcal{K}}\bar{\mathbf{u}_{T}}$, we see that the $\hat{\mathbf{a}}$ and
$\mathbf{d}_{2}$ terms vanish and one has
$\mathbf{m}_{*}\left(\partial^{\circ}{\bar{\mathbf{u}_{T}}},P_{\mathcal{K}}\bar{\mathbf{u}_{T}}\right)+\mathbf{c}_{1}(\bar{\mathbf{u}_{T}},\mathbf{u}_{T}^{1},P_{\mathcal{K}}\bar{\mathbf{u}_{T}})+\mathbf{c}_{1}(\mathbf{u}_{T}^{2},\bar{\mathbf{u}_{T}},P_{\mathcal{K}}\bar{\mathbf{u}_{T}})+\mathbf{l}(\bar{\mathbf{u}_{T}},P_{\mathcal{K}}\bar{\mathbf{u}_{T}})+\mathbf{d}_{1}(\bar{\mathbf{u}_{T}},P_{\mathcal{K}}\bar{\mathbf{u}_{T}})\\\
=\varepsilon\mathbf{c}_{3}(\bar{\varphi},\varphi^{1},P_{\mathcal{K}}\bar{\mathbf{u}_{T}})+\varepsilon\mathbf{c}_{3}(\varphi^{2},\bar{\varphi},P_{\mathcal{K}}\bar{\mathbf{u}_{T}}),$
(6.19)
where we have rewritten the nonlinear terms in a more appropriate way. As
before we pull out the derivative term so that
$\mathbf{m}_{*}\left(\partial^{\circ}{\bar{\mathbf{u}_{T}}},P_{\mathcal{K}}\bar{\mathbf{u}_{T}}\right)=\frac{1}{2}\frac{d}{dt}\|P_{\mathcal{K}}\bar{\mathbf{u}_{T}}\|_{\mathbf{L}^{2}(\Gamma(t))}^{2}-\mathbf{m}(\bar{\mathbf{u}_{T}}HV_{N},P_{\mathcal{K}}\bar{\mathbf{u}_{T}})+\frac{1}{2}\mathbf{m}(P_{\mathcal{K}}\bar{\mathbf{u}_{T}}HV_{N},P_{\mathcal{K}}\bar{\mathbf{u}_{T}})\\\
-\sum_{i=1}^{N_{K}}\mathbf{m}(\bar{\mathbf{u}_{T}},\boldsymbol{\kappa}_{i})\mathbf{m}(P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}},\partial^{\circ}\boldsymbol{\kappa}_{i}).$
(6.20)
To see this, we see that
$\mathbf{m}_{*}\left(\partial^{\circ}\bar{\mathbf{u}_{T}},P_{\mathcal{K}}\bar{\mathbf{u}_{T}}\right)=\frac{d}{dt}\|P_{\mathcal{K}}\bar{\mathbf{u}_{T}}\|_{\mathbf{L}^{2}(\Gamma(t))}^{2}-\mathbf{m}(\bar{\mathbf{u}_{T}}HV_{N},P_{\mathcal{K}}\bar{\mathbf{u}_{T}})-\mathbf{m}_{*}\left(\partial^{\circ}P_{\mathcal{K}}\bar{\mathbf{u}_{T}},\bar{\mathbf{u}_{T}}\right).$
Then one writes
$\mathbf{m}_{*}\left(\partial^{\circ}P_{\mathcal{K}}\bar{\mathbf{u}_{T}},\bar{\mathbf{u}_{T}}\right)=\mathbf{m}_{*}\left(\partial^{\circ}P_{\mathcal{K}}\bar{\mathbf{u}_{T}},P_{\mathcal{K}}\bar{\mathbf{u}_{T}}\right)+\mathbf{m}_{*}\left(\partial^{\circ}P_{\mathcal{K}}\bar{\mathbf{u}_{T}},P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}\right),$
where the transport theorem lets us see that
$\mathbf{m}_{*}\left(\partial^{\circ}P_{\mathcal{K}}\bar{\mathbf{u}_{T}},P_{\mathcal{K}}\bar{\mathbf{u}_{T}}\right)=\frac{1}{2}\frac{d}{dt}\|P_{\mathcal{K}}\bar{\mathbf{u}_{T}}\|_{\mathbf{L}^{2}(\Gamma(t))}^{2}-\frac{1}{2}\mathbf{m}(P_{\mathcal{K}}\bar{\mathbf{u}_{T}}HV_{N},P_{\mathcal{K}}\bar{\mathbf{u}_{T}}).$
The term involving $P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}$ is then dealt
with by using (B.5) so that
$\displaystyle\mathbf{m}_{*}\left(\partial^{\circ}P_{\mathcal{K}}\bar{\mathbf{u}_{T}},P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}\right)=\sum_{i=1}^{N_{K}}\mathbf{m}(\bar{\mathbf{u}_{T}},\boldsymbol{\kappa}_{i})\mathbf{m}(P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}},\partial^{\circ}\boldsymbol{\kappa}_{i}),$
and combining all this together yields (6.20).
Next we consider the $\mathbf{c}_{1}$ terms. We notice that
$|\mathbf{c}_{1}(\bar{\mathbf{u}_{T}},\mathbf{u}_{T}^{1},P_{\mathcal{K}}\bar{\mathbf{u}_{T}})|=|\mathbf{c}_{1}(P_{\mathcal{K}}\bar{\mathbf{u}_{T}},\mathbf{u}_{T}^{1},\bar{\mathbf{u}_{T}})|\leq\|P_{\mathcal{K}}\bar{\mathbf{u}_{T}}\|_{\mathbf{H}^{1,4}(\Gamma(t))}\|\mathbf{u}_{T}^{1}\|_{\mathbf{L}^{4}(\Gamma(t))}\|\bar{\mathbf{u}_{T}}\|_{\mathbf{L}^{2}(\Gamma(t))},$
and by using (B.3) one can deduce
$|\mathbf{c}_{1}(\bar{\mathbf{u}_{T}},\mathbf{u}_{T}^{1},P_{\mathcal{K}}\bar{\mathbf{u}_{T}})|\leq
C\|P_{\mathcal{K}}\bar{\mathbf{u}_{T}}\|_{\mathbf{L}^{2}(\Gamma(t))}\|\mathbf{u}_{T}^{1}\|_{\mathbf{L}^{4}(\Gamma(t))}\|\bar{\mathbf{u}_{T}}\|_{\mathbf{L}^{2}(\Gamma(t))}.$
The same calculations for the other $\mathbf{c}_{1}$ term and Young’s
inequality yields
$\displaystyle|\mathbf{c}_{1}(\bar{\mathbf{u}_{T}},\mathbf{u}_{T}^{1},P_{\mathcal{K}}\bar{\mathbf{u}_{T}})|+|\mathbf{c}_{1}(\mathbf{u}_{T}^{2},\bar{\mathbf{u}_{T}},P_{\mathcal{K}}\bar{\mathbf{u}_{T}})|\leq\frac{\eta_{*}}{12}\|\bar{\mathbf{u}_{T}}\|_{\mathbf{L}^{2}(\Gamma(t))}^{2}+C\left(\|\mathbf{u}_{T}^{1}\|_{\mathbf{L}^{4}(\Gamma(t))}^{2}+\|\mathbf{u}_{T}^{2}\|_{\mathbf{L}^{4}(\Gamma(t))}^{2}\right)\|P_{\mathcal{K}}\bar{\mathbf{u}_{T}}\|_{\mathbf{L}^{2}(\Gamma(t))}^{2}.$
(6.21)
We then turn to the terms involving $\mathbf{l},\mathbf{d}_{1}$. As seen
before, it is straightforward to see that
$\displaystyle\mathbf{l}(\bar{\mathbf{u}_{T}},P_{\mathcal{K}}\bar{\mathbf{u}_{T}})\leq\frac{\eta_{*}}{12}\|\bar{\mathbf{u}_{T}}\|_{\mathbf{L}^{2}(\Gamma(t)}^{2}+C\|P_{\mathcal{K}}\bar{\mathbf{u}_{T}}\|_{\mathbf{L}^{2}(\Gamma(t))}^{2}.$
(6.22)
We then express the $\mathbf{d}_{1}$ term as
$\mathbf{d}_{1}(\bar{\mathbf{u}_{T}},P_{\mathcal{K}}\bar{\mathbf{u}_{T}})=\mathbf{c}_{1}(\bar{\mathbf{u}_{T}},\widetilde{\mathbf{u}_{T}},P_{\mathcal{K}}\bar{\mathbf{u}_{T}})+\mathbf{c}_{1}(\widetilde{\mathbf{u}_{T}},\bar{\mathbf{u}_{T}},P_{\mathcal{K}}\bar{\mathbf{u}_{T}})\\\
=-\mathbf{c}_{1}(P_{\mathcal{K}}\bar{\mathbf{u}_{T}},\widetilde{\mathbf{u}_{T}},\bar{\mathbf{u}_{T}})-\mathbf{c}_{1}(P_{\mathcal{K}}\bar{\mathbf{u}_{T}},\bar{\mathbf{u}_{T}},\widetilde{\mathbf{u}_{T}})+2\mathbf{m}(\bar{\mathbf{u}_{T}}HV_{N},P_{\mathcal{K}}\bar{\mathbf{u}_{T}}),$
and so arguing as we did for the previous $\mathbf{d}_{1}$ term we find
$\displaystyle\mathbf{d}_{1}(\bar{\mathbf{u}_{T}},P_{\mathcal{K}}\bar{\mathbf{u}_{T}})\leq\frac{\eta_{*}}{12}\|\bar{\mathbf{u}_{T}}\|_{\mathbf{L}^{2}(\Gamma(t))}^{2}+C\|\widetilde{\mathbf{u}_{T}}\|_{\mathbf{L}^{4}(\Gamma(t))}^{2}\|P_{\mathcal{K}}\bar{\mathbf{u}_{T}}\|_{\mathbf{L}^{2}(\Gamma(t))}^{2}.$
(6.23)
Lastly we bound the $\mathbf{c}_{3}$ terms as before, where we find
$|\mathbf{c}_{3}(\bar{\varphi},\varphi^{1},P_{\mathcal{K}}\bar{\mathbf{u}_{T}})|\leq\|\nabla_{\Gamma}\bar{\varphi}\|_{L^{2}(\Gamma(t))}\|\varphi^{1}\|_{H^{1,4}(\Gamma(t))}\|P_{\mathcal{K}}\bar{\mathbf{u}_{T}}\|_{\mathbf{H}^{1,4}(\Gamma(t))}\leq
C\|\nabla_{\Gamma}\bar{\varphi}\|_{L^{2}(\Gamma)}\|\varphi^{1}\|_{H^{1,4}(\Gamma(t))}\|P_{\mathcal{K}}\bar{\mathbf{u}_{T}}\|_{\mathbf{L}^{2}(\Gamma(t))},$
where we have used (B.3) in the last inequality. From this, using Young’s
inequality it is straightforward to see
$\varepsilon|\mathbf{c}_{3}(\bar{\varphi},\varphi^{1},P_{\mathcal{K}}\bar{\mathbf{u}_{T}})|+\varepsilon|\mathbf{c}_{3}(\varphi^{2},\bar{\varphi},P_{\mathcal{K}}\bar{\mathbf{u}_{T}})|\leq\frac{\varepsilon}{4}\|\nabla_{\Gamma}\bar{\varphi}\|_{L^{2}(\Gamma(t))}^{2}\\\
+C\left(\|\varphi^{1}\|_{H^{1,4}(\Gamma(t))}^{2}+\|\varphi^{2}\|_{H^{1,4}(\Gamma(t))}^{2}\right)\|P_{\mathcal{K}}\bar{\mathbf{u}_{T}}\|_{\mathbf{L}^{2}(\Gamma(t))}^{2}.$
(6.24)
Hence using (6.20)-(6.24) and (6.18) in (6.19) yields (6.6).
The final inequality to show is (6.7). To do this we test (6.3) with
$\mathcal{G}\bar{\varphi}$, which we note is well defined as
$\mathchoice{{\vbox{\hbox{$\textstyle-$
}}\kern-7.83337pt}}{{\vbox{\hbox{$\scriptstyle-$
}}\kern-5.90005pt}}{{\vbox{\hbox{$\scriptscriptstyle-$
}}\kern-4.75003pt}}{{\vbox{\hbox{$\scriptscriptstyle-$
}}\kern-4.25003pt}}\\!\int_{\Gamma(t)}\bar{\varphi}=0$ for almost all
$t\in[0,T]$, which yields
$\displaystyle
m_{*}\left(\partial^{\circ}{\bar{\varphi}},\mathcal{G}\bar{\varphi}\right)+a(\bar{\mu},\mathcal{G}\bar{\varphi})+\mathbf{c}_{2}(\mathcal{G}\bar{\varphi},\bar{\varphi},\mathbf{u}_{T}^{1})+\mathbf{c}_{2}(\mathcal{G}\bar{\varphi},\varphi^{2},\bar{\mathbf{u}_{T}})+\mathbf{c}_{2}(\mathcal{G}\bar{\varphi},\bar{\varphi},\widetilde{\mathbf{u}_{T}})=0.$
(6.25)
The first term can be expressed as
$\displaystyle
m_{*}\left(\partial^{\circ}{\bar{\varphi}},\mathcal{G}\bar{\varphi}\right)=\frac{1}{2}\frac{d}{dt}\|\bar{\varphi}\|_{-1}^{2}-m(\bar{\varphi}HV_{N},\mathcal{G}\bar{\varphi})+\frac{1}{2}b(\mathcal{G}\bar{\varphi},\mathcal{G}\bar{\varphi}).$
(6.26)
To see this we express this term as
$m_{*}\left(\partial^{\circ}{\bar{\varphi}},\mathcal{G}\bar{\varphi}\right)=\frac{d}{dt}m\left(\bar{\varphi},\mathcal{G}\bar{\varphi}\right)-m\left(\bar{\varphi},\partial^{\circ}{\mathcal{G}\bar{\varphi}}\right)-m(\bar{\varphi}HV_{N},\mathcal{G}\bar{\varphi}),$
and note
$m(\bar{\varphi},\mathcal{G}\bar{\varphi})=\|\bar{\varphi}\|_{-1}^{2}$, and
that
$m\left(\bar{\varphi},\partial^{\circ}{\mathcal{G}\bar{\varphi}}\right)=a\left(\mathcal{G}\bar{\varphi},\partial^{\circ}{\mathcal{G}\bar{\varphi}}\right)=\frac{1}{2}\frac{d}{dt}\|\bar{\varphi}\|_{-1}^{2}-\frac{1}{2}b(\mathcal{G}\bar{\varphi},\mathcal{G}\bar{\varphi}).$
To bound the second term, we see from the definition of the inverse Laplacian
that
$a(\bar{\varphi},\mathcal{G}\bar{\varphi})=m(\bar{\varphi},\bar{\varphi})$,
and hence testing (6.3) with $\bar{\varphi}$ we see
$a(\bar{\mu},\mathcal{G}\bar{\varphi})=\varepsilon
a(\bar{\varphi},\bar{\varphi})+\frac{1}{\varepsilon}m(F^{\prime}(\varphi^{1})-F^{\prime}(\varphi^{2}),\bar{\varphi}).$
We recall that $F=F_{1}+F_{2}$ where $F_{1}$ is convex, so that
$m(F^{\prime}(\varphi^{1})-F^{\prime}(\varphi^{2}),\bar{\varphi})\geq
m(F_{2}^{\prime}(\varphi^{1})-F_{2}^{\prime}(\varphi^{2}),\bar{\varphi}).$
By using the Lipschitz continuity of $F_{2}^{\prime}$, and the definition of
the inverse Laplacian, one readily sees that
$\displaystyle|m(F_{2}^{\prime}(\varphi^{1})-F_{2}^{\prime}(\varphi^{2}),\bar{\varphi})|\leq
C\|\bar{\varphi}\|_{L^{2}(\Gamma(t))}^{2}=Ca(\bar{\varphi},\mathcal{G}\bar{\varphi})\leq\frac{\varepsilon}{16}\|\nabla_{\Gamma}\bar{\varphi}\|_{L^{2}(\Gamma(t))}+C\|\bar{\varphi}\|_{-1}^{2}.$
(6.27)
It remains to bound the various $\mathbf{c}_{2}$ terms. Firstly we find that
$|\mathbf{c}_{2}(\mathcal{G}\bar{\varphi},\bar{\varphi},\mathbf{u}_{T}^{1})|\leq\|\mathcal{G}\bar{\varphi}\|_{L^{4}(\Gamma(t))}\|\nabla_{\Gamma}\bar{\varphi}\|_{L^{2}(\Gamma(t))}\|\mathbf{u}_{T}^{1}\|_{\mathbf{L}^{4}(\Gamma(t))},$
and by using the embedding $H^{1}(\Gamma(t))\hookrightarrow L^{4}(\Gamma(t))$,
Poincaré’s inequality and Young’s inequality we find that
$\displaystyle|\mathbf{c}_{2}(\mathcal{G}\bar{\varphi},\bar{\varphi},\mathbf{u}_{T}^{1})|\leq\frac{\varepsilon}{16}\|\nabla_{\Gamma}\bar{\varphi}\|_{L^{2}(\Gamma(t))}^{2}+C\|\mathbf{u}_{T}^{1}\|_{\mathbf{L}^{4}(\Gamma(t))}^{2}\|\bar{\varphi}\|_{-1}^{2}.$
(6.28)
The other $\mathbf{c}_{2}$ term is similar, but now we use the antisymmetry of
$\mathbf{c}_{2}$ in the first two arguments so that
$|\mathbf{c}_{2}(\mathcal{G}\bar{\varphi},\varphi^{2},\bar{\mathbf{u}_{T}})|=|\mathbf{c}_{2}(\varphi^{2},\mathcal{G}\bar{\varphi},\bar{\mathbf{u}_{T}})|\leq\|\varphi^{2}\|_{L^{\infty}(\Gamma(t))}\|\bar{\varphi}\|_{-1}\|\bar{\mathbf{u}_{T}}\|_{\mathbf{L}^{2}(\Gamma(t))}.$
It is then clear from Young’s inequality that
$\displaystyle|\mathbf{c}_{2}(\mathcal{G}\bar{\varphi},\varphi^{2},\bar{\mathbf{u}_{T}})|\leq\frac{\eta_{*}}{4}\|\bar{\mathbf{u}_{T}}\|_{\mathbf{L}^{2}(\Gamma(t))}^{2}+C\|\varphi^{2}\|_{L^{\infty}(\Gamma(t))}^{2}\|\bar{\varphi}\|_{-1}^{2}.$
(6.29)
Finally we bound the term involving $\widetilde{\mathbf{u}_{T}}$. To do this,
we observe that
$|\mathbf{c}_{2}(\mathcal{G}\bar{\varphi},\bar{\varphi},\widetilde{\mathbf{u}_{T}})|\leq|\mathbf{c}_{2}(\bar{\varphi},\mathcal{G}\bar{\varphi},\widetilde{\mathbf{u}_{T}})|+|m(\bar{\varphi}HV_{N},\mathcal{G}\bar{\varphi})|,$
so that by similar arguments to the above
$\displaystyle|\mathbf{c}_{2}(\mathcal{G}\bar{\varphi},\bar{\varphi},\widetilde{\mathbf{u}_{T}})|\leq\frac{\varepsilon}{16}\|\nabla_{\Gamma}\bar{\varphi}\|_{L^{2}(\Gamma(t))}^{2}+C\left(1+\|\widetilde{\mathbf{u}_{T}}\|_{\mathbf{L}^{\infty}(\Gamma(t))}^{2}\right)\|\bar{\varphi}\|_{-1}^{2}.$
(6.30)
Hence using (6.26)-(6.30) and (6.18) in (6.25) one obtains (6.7), where we
have also bounded the extra terms in (6.26) as
$|m(\bar{\varphi}HV_{N},\mathcal{G}\bar{\varphi})|+|b(\mathcal{G}\bar{\varphi},\mathcal{G}\bar{\varphi})|\leq\frac{\varepsilon}{16}\|\nabla_{\Gamma}\bar{\varphi}\|_{L^{2}(\Gamma(t))}^{2}+C\|\bar{\varphi}\|_{-1}^{2}.$
∎
Next we show a stability result for the case of constant viscosity. This also
provides a simpler proof for uniqueness in this special case, where we no
longer require Lemma 3.11.
###### Proposition 6.3.
Let $(\varphi^{i},\mu^{i},\mathbf{u}_{T}^{i})$ denote the solution triple
corresponding to some choice of initial data $\varphi_{0}^{i}\in
H^{1}(\Gamma_{0}),\mathbf{u}_{T,0}^{i}\in\mathbf{H}_{\sigma}$, for $i=1,2$,
where $\mathchoice{{\vbox{\hbox{$\textstyle-$
}}\kern-7.83337pt}}{{\vbox{\hbox{$\scriptstyle-$
}}\kern-5.90005pt}}{{\vbox{\hbox{$\scriptscriptstyle-$
}}\kern-4.75003pt}}{{\vbox{\hbox{$\scriptscriptstyle-$
}}\kern-4.25003pt}}\\!\int_{\Gamma_{0}}\varphi_{0}^{1}=\mathchoice{{\vbox{\hbox{$\textstyle-$
}}\kern-7.83337pt}}{{\vbox{\hbox{$\scriptstyle-$
}}\kern-5.90005pt}}{{\vbox{\hbox{$\scriptscriptstyle-$
}}\kern-4.75003pt}}{{\vbox{\hbox{$\scriptscriptstyle-$
}}\kern-4.25003pt}}\\!\int_{\Gamma_{0}}\varphi_{2}^{1}$. Then, under the same
assumptions as the preceding theorem, we have
$\|P_{\mathcal{K}^{\perp}}(\mathbf{u}_{T}^{1}(t)-\mathbf{u}_{T}^{2}(t))\|_{\perp}^{2}+\|P_{\mathcal{K}}(\mathbf{u}_{T}^{1}(t)-\mathbf{u}_{T}^{2}(t))\|_{\mathbf{L}^{2}(\Gamma(t))}^{2}+\|\varphi^{1}(t)-\varphi^{2}(t)\|_{-1}^{2}\\\
\leq
C\left(\|P_{\mathcal{K}^{\perp}}(\mathbf{u}_{T,0}^{1}-\mathbf{u}_{T,0}^{2})\|_{\perp}^{2}+\|P_{\mathcal{K}}(\mathbf{u}_{T,0}^{1}-\mathbf{u}_{T,0}^{2})\|_{\mathbf{L}^{2}(\Gamma_{0})}^{2}+\|\varphi_{0}^{1}-\varphi_{0}^{2}\|_{-1}^{2}\right),$
(6.31)
for a constant $C$ which depends on $t,\Gamma$ and the initial data.
###### Proof.
This proof is largely the same as that of the previous theorem, where we have
some minor modifications. We use the same notation as before, except now we
denote $(\varphi^{i},\mu^{i},\mathbf{u}_{T}^{i})$ as the solution
corresponding to some choice of initial data $\varphi_{0}^{i}\in
H^{1}(\Gamma),\mathbf{u}_{T,0}^{i}\in\mathbf{H}_{\sigma}.$ We define
$(\bar{\varphi},\bar{\mu},\bar{\mathbf{u}_{T}})$ as before, and note that
instead of (6.2) we find that $(\bar{\varphi},\bar{\mu},\bar{\mathbf{u}_{T}})$
solves
$\mathbf{m}_{*}\left(\partial^{\circ}{\bar{\mathbf{u}_{T}}},\boldsymbol{\phi}\right)+\eta\mathbf{a}(\bar{\mathbf{u}_{T}},\boldsymbol{\phi})+\mathbf{c}_{1}(\mathbf{u}_{T}^{1},\mathbf{u}_{T}^{1},\boldsymbol{\phi})-\mathbf{c}_{1}(\mathbf{u}_{T}^{2},\mathbf{u}_{T}^{2},\boldsymbol{\phi})+\mathbf{l}(\bar{\mathbf{u}_{T}},\boldsymbol{\phi})+\mathbf{d}_{1}(\bar{\mathbf{u}_{T}},\boldsymbol{\phi})\\\
=\varepsilon\mathbf{c}_{3}(\varphi^{1},\varphi^{1},\boldsymbol{\phi})-\varepsilon\mathbf{c}_{3}(\varphi^{2},\varphi^{2},\boldsymbol{\phi}),$
(6.32)
but (6.3) and (6.4) are still satisfied. As before we test (6.32) with
$\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}}$ to see
$\frac{1}{2}\frac{d}{dt}\|P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}\|_{\perp}^{2}+\eta\|P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}\|_{\mathbf{L}^{2}(\Gamma(t))}^{2}+\mathbf{c}_{1}(\bar{\mathbf{u}_{T}},\mathbf{u}_{T}^{1},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}})+\mathbf{c}_{1}(\mathbf{u}_{T}^{2},\bar{\mathbf{u}_{T}},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}})+\mathbf{l}(\bar{\mathbf{u}_{T}},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}})+\mathbf{d}_{1}(\bar{\mathbf{u}_{T}},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}})\\\
=\varepsilon\mathbf{c}_{3}(\bar{\varphi},\varphi^{1},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}})+\varepsilon\mathbf{c}_{3}(\varphi^{2},\bar{\varphi},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}})-\frac{1}{2}\mathbf{b}(\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}})+\mathbf{m}(P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}HV_{N},\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}})-\sum_{i=1}^{N_{K}}\mathbf{m}(\bar{\mathbf{u}_{T}},\boldsymbol{\kappa}_{i})\mathbf{m}(\mathcal{S}^{\perp}\bar{\mathbf{u}_{T}},\partial^{\circ}\boldsymbol{\kappa}_{i})$
where we have used (6.9) and the definition of $\mathcal{S}^{\perp}$. Now by
arguing as we did for (6.12)-(6.17) it is straightforward to see that
$\displaystyle\frac{1}{2}\frac{d}{dt}\|P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}\|_{\perp}^{2}+\eta\|P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}\|_{\mathbf{L}^{2}(\Gamma(t))}^{2}\leq\frac{\eta}{2}\|\bar{\mathbf{u}_{T}}\|_{\mathbf{L}^{2}(\Gamma(t))}^{2}+\frac{\varepsilon}{4}\|\nabla_{\Gamma}\bar{\varphi}\|_{L^{2}(\Gamma(t))}^{2}+K_{1}(t)\|P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}\|_{\perp}^{2},$
(6.33)
where
$K_{1}(t)=C\left(\|\mathbf{u}_{T}^{1}\|_{\mathbf{L}^{4}(\Gamma(t))}^{4}+\|\mathbf{u}_{T}^{2}\|_{\mathbf{L}^{4}(\Gamma(t))}^{4}+\|\varphi^{1}\|_{H^{1,4}(\Gamma(t))}^{4}+\|\varphi^{2}\|_{H^{1,4}(\Gamma(t))}^{4}+\|\widetilde{\mathbf{u}_{T}}\|_{\mathbf{L}^{\infty}(\Gamma(t))}^{2}\right).$
Notice that the requirement $\mathchoice{{\vbox{\hbox{$\textstyle-$
}}\kern-7.83337pt}}{{\vbox{\hbox{$\scriptstyle-$
}}\kern-5.90005pt}}{{\vbox{\hbox{$\scriptscriptstyle-$
}}\kern-4.75003pt}}{{\vbox{\hbox{$\scriptscriptstyle-$
}}\kern-4.25003pt}}\\!\int_{\Gamma}\varphi_{0}^{1}=\mathchoice{{\vbox{\hbox{$\textstyle-$
}}\kern-7.83337pt}}{{\vbox{\hbox{$\scriptstyle-$
}}\kern-5.90005pt}}{{\vbox{\hbox{$\scriptscriptstyle-$
}}\kern-4.75003pt}}{{\vbox{\hbox{$\scriptscriptstyle-$
}}\kern-4.25003pt}}\\!\int_{\Gamma}\varphi_{0}^{2}$ allows us to define
$\mathcal{G}\bar{\varphi}$, and so related calculations from the preceding
theorem still hold. By summing (6.6), (6.7), and (6.33) one finds
$\frac{1}{2}\frac{d}{dt}\left(\|P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}\|_{\perp}^{2}+\|P_{\mathcal{K}}\bar{\mathbf{u}_{T}}\|_{\mathbf{L}^{2}(\Gamma(t))}^{2}+\|\bar{\varphi}\|_{-1}^{2}\right)\leq
K(t)\left(\|P_{\mathcal{K}^{\perp}}\bar{\mathbf{u}_{T}}\|_{\perp}^{2}+\|P_{\mathcal{K}}\bar{\mathbf{u}_{T}}\|_{\mathbf{L}^{2}(\Gamma(t))}^{2}+\|\bar{\varphi}\|_{-1}^{2}\right),$
where $K\in L^{1}([0,T])$. An application of Grönwall’s inequality then yields
(6.31). ∎
###### Remark 6.4.
One may notice that this proof can be somewhat simplified by defining
$\mathcal{S}\boldsymbol{\phi}\in\mathbf{V}_{\sigma}(t)$, where
$\boldsymbol{\phi}\in\mathbf{L}^{2}(\Gamma(t))$, to be the unique solution of
$\mathbf{a}(\mathcal{S}\boldsymbol{\phi},\boldsymbol{\psi})+\mathbf{m}(\mathcal{S}\boldsymbol{\phi},\boldsymbol{\psi})=\mathbf{m}(\boldsymbol{\phi},\boldsymbol{\psi}),$
for all $\boldsymbol{\psi}\in\mathbf{L}^{2}(\Gamma(t))$. From (3.9) it is
clear that this operator is coercive, and so one does not have to quotient out
a kernel as we did above. Proceeding as before one defines a corresponding
norm, and the proof should follow but without the Killing vector splitting. We
have not chosen to do this, as this obscures the fact that the Killing vector
component of $\mathbf{u}_{T}$ has better stability properties than the non-
Killing component. Indeed, by using (B.3) and the
$\mathcal{K}-\mathcal{K}^{\perp}$ splitting as in our proof one observes that
we have $\mathbf{H}^{1,p}-\mathbf{V}_{\sigma}^{\prime}$ stability, whereas
this simplified proof would only give $\mathbf{V}_{\sigma}^{\prime}$ stability
for both components. Similarly if one considers a surface with boundary and
prescribes Dirichlet boundary conditions on $\mathbf{u}_{T}$ then one can
define the inverse Stokes operator in a way that doesn’t involve this Killing
vector splitting.
### 6.2 Uniqueness for the logarithmic potential
It is clear that (6.1) still holds, that is $\varphi\in L^{2}_{H^{2}}$, since
we know $\mu,f(\varphi)\in L^{2}_{L^{2}},$ and so we may use the
$\mathbf{c}_{3}$ bilinear form as before. With this at hand, the proofs of
Theorem 6.1 and Proposition 6.3 follow. That is, for $\Gamma(t)$ be a $C^{3}$
evolving surface, and initial data
$\varphi_{0}\in\mathcal{I}_{0},\mathbf{u}_{T,0}\in\mathbf{H}_{\sigma}(0)$ the
solution triple $(\varphi,\mu,\mathbf{u}_{T})$ solving (4.4)-(4.6) is unique.
Moreover, for a constant viscosity, and $\varphi_{0}^{i}\in\mathcal{I}_{0}$
such that where $\mathchoice{{\vbox{\hbox{$\textstyle-$
}}\kern-7.83337pt}}{{\vbox{\hbox{$\scriptstyle-$
}}\kern-5.90005pt}}{{\vbox{\hbox{$\scriptscriptstyle-$
}}\kern-4.75003pt}}{{\vbox{\hbox{$\scriptscriptstyle-$
}}\kern-4.25003pt}}\\!\int_{\Gamma_{0}}\varphi_{0}^{1}=\mathchoice{{\vbox{\hbox{$\textstyle-$
}}\kern-7.83337pt}}{{\vbox{\hbox{$\scriptstyle-$
}}\kern-5.90005pt}}{{\vbox{\hbox{$\scriptscriptstyle-$
}}\kern-4.75003pt}}{{\vbox{\hbox{$\scriptscriptstyle-$
}}\kern-4.25003pt}}\\!\int_{\Gamma_{0}}\varphi_{0}^{2}$, then we have
stability bound similar to (6.31)
###### Remark 6.5.
The results for the logarithmic potential extend to more general singular
potentials of the form $F(r)=F_{1}(r)-\frac{\theta}{2}r^{2}$, with $F_{1}\in
C^{2}((a,b))\cap C^{0}([a,b])$ for some $a,b\in\mathbb{R}$ under some
necessary assumptions we do not expand upon. Potentials of this form are
treated on a Euclidean domain in [1, 29], but here we have only covered the
thermodynamically relevant logarithmic potential - which is still illustrative
of the general case.
## 7 Reintroducing the surface pressure
We end our discussion by reintroducing the surface pressure and the correct
divergence condition. We now consider the mixed formulation, with a regular
potential, where one finds a solution $(\varphi,\mu,\mathbf{u}_{T},p)$, with
$\varphi\in H^{1}_{H^{-1}}\cap L^{2}_{H^{1}},\mu\in
L^{2}_{H^{1}},\mathbf{u}_{T}\in H^{1}_{\mathbf{H}^{-1}}\cap
L^{2}_{\mathbf{H}^{1}},p\in L^{2}_{L^{2}}$, solving
$\displaystyle\begin{split}\langle\partial^{\circ}\mathbf{u}_{T},\boldsymbol{\phi}\rangle_{\mathbf{H}^{-1}(\Gamma(t)),\mathbf{H}^{1}(\Gamma(t))}+\hat{\mathbf{a}}(\eta(\varphi),\mathbf{u}_{T},\boldsymbol{\phi})+\mathbf{c}_{1}(\mathbf{u}_{T},\mathbf{u}_{T},\boldsymbol{\phi})+\mathbf{l}(\mathbf{u}_{T},\boldsymbol{\phi})+m(p,\nabla_{\Gamma}\cdot\boldsymbol{\phi})+\mathbf{d}_{1}(\mathbf{u}_{T},\boldsymbol{\phi})\\\
+\mathbf{d}_{2}(\eta(\varphi),\boldsymbol{\phi})=\mathbf{m}(\mathbf{B},\boldsymbol{\phi})+\mathbf{c}_{2}(\mu,\varphi,\boldsymbol{\phi})\end{split},$
(7.1) $\displaystyle m(q,\nabla_{\Gamma}\cdot\mathbf{u}_{T})=0,$ (7.2)
$\displaystyle
m_{*}(\partial^{\circ}\varphi,\phi)+a(\mu,\phi)+\mathbf{c}_{2}(\phi,\varphi,\mathbf{u}_{T})+\mathbf{c}_{2}(\phi,\varphi,\widetilde{\mathbf{u}_{T}})=0,$
(7.3) $\displaystyle m(\mu,\phi)=\varepsilon
a(\varphi,\phi)+\frac{1}{\varepsilon}m(F^{\prime}(\varphi),\phi),$ (7.4)
for all $q\in L^{2}(\Gamma(t)),\phi\in
H^{1}(\Gamma(t)),\boldsymbol{\phi}\in\mathbf{H}^{1}(\Gamma(t))$ for almost all
$t\in[0,T]$. Here the initial data is $\varphi_{0}\in
H^{1}(\Gamma_{0}),\mathbf{u}_{T,0}\in\mathbf{H}_{\sigma}(0)$, so one has
$\varphi(0)=\varphi_{0},\mathbf{u}_{T}(0)=\mathbf{u}_{T,0}$ almost everywhere
on $\Gamma_{0}$.
Before proving the existence and uniqueness of this system, we recall the
uniform inf-sup condition of [45].
###### Lemma 7.1 ([45], Lemma 3.3).
There exists a constant, $C$, independent of time such that for all $q\in
L_{0}^{2}(\Gamma(t)):=\left\\{\phi\in
L^{2}(\Gamma(t))\mid\mathchoice{{\vbox{\hbox{$\textstyle-$
}}\kern-7.83337pt}}{{\vbox{\hbox{$\scriptstyle-$
}}\kern-5.90005pt}}{{\vbox{\hbox{$\scriptscriptstyle-$
}}\kern-4.75003pt}}{{\vbox{\hbox{$\scriptscriptstyle-$
}}\kern-4.25003pt}}\\!\int_{\Gamma(t)}\phi=0\right\\}$,
$\displaystyle\|\nabla_{\Gamma}q\|_{\mathbf{H}^{-1}(\Gamma(t))}:=\sup_{\boldsymbol{\phi}\in\mathbf{H}^{1}(\Gamma(t))\setminus\\{0\\}}\frac{\int_{\Gamma(t)}q\nabla_{\Gamma}\cdot\boldsymbol{\phi}}{\|\boldsymbol{\phi}\|_{\mathbf{H}^{1}(\Gamma(t))}}\geq
C\|q\|_{L^{2}(\Gamma(t))}.$ (7.5)
###### Theorem 7.2.
There exists a unique solution, $(\varphi,\mu,\mathbf{u}_{T},p)$, with
$\varphi\in H^{1}_{H^{-1}}\cap L^{2}_{H^{1}},\mu\in
L^{2}_{H^{1}},\mathbf{u}_{T}\in H^{1}_{\mathbf{H}^{-1}}\cap
L^{2}_{\mathbf{H}^{1}},p\in L^{2}_{L^{2}}$, of (7.1)-(7.4).
###### Proof.
To begin, let $(\varphi,\mu,\mathbf{u}_{T})$ be the unique solution of
(4.1)-(4.3), and define $\mathcal{F}(t)$ by
$\langle\mathcal{F}(t),\boldsymbol{\phi}\rangle_{\mathbf{H}^{-1}(\Gamma(t)),\mathbf{H}^{1}(\Gamma(t))}:=\langle\partial^{\circ}\mathbf{u}_{T},\boldsymbol{\phi}\rangle_{\mathbf{H}^{-1}(\Gamma(t)),\mathbf{H}^{1}(\Gamma(t))}+\hat{\mathbf{a}}(\eta(\varphi),\mathbf{u}_{T},\boldsymbol{\phi})+\mathbf{c}_{1}(\mathbf{u}_{T},\mathbf{u}_{T},\boldsymbol{\phi})+\mathbf{l}(\mathbf{u}_{T},\boldsymbol{\phi})+m(p,\nabla_{\Gamma}\cdot\boldsymbol{\phi})\\\
+\mathbf{d}_{1}(\mathbf{u}_{T},\boldsymbol{\phi})+\mathbf{d}_{2}(\eta(\varphi),\boldsymbol{\phi})-\mathbf{m}(\mathbf{B},\boldsymbol{\phi})-\mathbf{c}_{2}(\mu,\varphi,\boldsymbol{\phi}),$
for $\boldsymbol{\phi}\in\mathbf{H}^{1}(\Gamma(t))$. We claim that
$\mathcal{F}\in L^{2}_{\mathbf{H}^{-1}}$. To see this, we recall the
equivalence,
$\partial^{\circ}\boldsymbol{\phi}\in
L^{2}_{\mathbf{V}_{\sigma}^{\prime}}\Leftrightarrow\partial^{\circ}\boldsymbol{\phi}\in
L^{2}_{\mathbf{H}^{-1}},\ \boldsymbol{\phi}\in L^{2}_{\mathbf{V}_{\sigma}}$
from [45], and repeat various estimates we have used throughout. We elaborate
on the estimates for
$\mathbf{c}_{1}(\mathbf{u}_{T},\mathbf{u}_{T},\boldsymbol{\phi}),\mathbf{c}_{2}(\mu,\varphi,\boldsymbol{\phi})$,
but skip further calculations. As
$\boldsymbol{\phi}\in\mathbf{H}^{1}(\Gamma(t))$, and not necessarily
$\mathbf{V}_{\sigma}(t)$, we cannot use the properties,
$\mathbf{c}_{1}(\mathbf{u}_{T},\mathbf{u}_{T},\boldsymbol{\phi})=-\mathbf{c}_{1}(\boldsymbol{\phi},\mathbf{u}_{T},\mathbf{u}_{T})$
and
$\mathbf{c}_{2}(\mu,\varphi,\boldsymbol{\phi})=-\mathbf{c}_{2}(\varphi,\mu,\boldsymbol{\phi})$.
However, by using the divergence theorem and the fact that
$\partial\Gamma(t)=\emptyset$, one finds
$\displaystyle
c_{1}(\mathbf{u}_{T},\mathbf{u}_{T},\boldsymbol{\phi})=-c_{1}(\boldsymbol{\phi},\mathbf{u}_{T},\mathbf{u}_{T})-\int_{\Gamma(t)}(\mathbf{u}_{T}\cdot\mathbf{u}_{T})\nabla_{\Gamma}\cdot\boldsymbol{\phi},$
$\displaystyle
c_{2}(\mu,\varphi,\boldsymbol{\phi})=-c_{2}(\varphi,\mu,\boldsymbol{\phi})-\int_{\Gamma(t)}\varphi\mu\nabla_{\Gamma}\cdot\boldsymbol{\phi},$
and hence
$\displaystyle\int_{0}^{T}|c_{1}(\mathbf{u}_{T},\mathbf{u}_{T},\boldsymbol{\phi})|\leq
C\sup_{t\in[0,T]}\|\mathbf{u}_{T}\|_{\mathbf{L}^{2}(\Gamma(t))}\left(\int_{0}^{T}\|\mathbf{u}_{T}\|_{\mathbf{H}^{1}(\Gamma(t))}\right)^{\frac{1}{2}}\left(\int_{0}^{T}\|\boldsymbol{\phi}\|_{\mathbf{H}^{1}(\Gamma(t))}\right)^{\frac{1}{2}},$
$\displaystyle\int_{0}^{T}|c_{2}(\mu,\varphi,\boldsymbol{\phi})|\leq
C\sup_{t\in[0,T]}\|\varphi\|_{H^{1}(\Gamma(t))}\left(\int_{0}^{T}\|\mu\|_{H^{1}(\Gamma(t))}^{2}\right)^{\frac{1}{2}}\left(\int_{0}^{T}\|\boldsymbol{\phi}\|_{\mathbf{H}^{1}(\Gamma(t))}^{2}\right)^{\frac{1}{2}},$
where we have used Sobolev embeddings and (3.8) as appropriate. The other
terms follow similar, but simpler, arguments.
We now observe that from the inf-sup condition (7.5) that the distributional
divergence,
$\nabla_{\Gamma}:L_{0}^{2}(\Gamma(t))\rightarrow\mathbf{H}^{-1}(\Gamma(t)),$
has a closed range $R(\nabla_{\Gamma})\subset\mathbf{H}^{-1}(\Gamma(t))$. This
follows from (7.5) and continuity of $\nabla_{\Gamma}$ as an operator. Now by
the closed range theorem, see for example [52] VII.5, we find that
$R(\nabla_{\Gamma})=\ker(\nabla_{\Gamma}^{*})^{\perp},\text{ where
}\ker(\nabla_{\Gamma}^{*})=\mathbf{V}_{\sigma}(t),$
where $\nabla_{\Gamma}^{*}$ is the adjoint of $\nabla_{\Gamma}$.
Since $(\varphi,\mu,\mathbf{u}_{T})$ solves (4.1)-(4.3) for almost all
$t\in[0,T]$, we see that
$\langle\mathcal{F}(t),\boldsymbol{\phi}\rangle_{\mathbf{H}^{-1}(\Gamma(t)),\mathbf{H}^{1}(\Gamma(t))}=0$
for all $\boldsymbol{\phi}\in\mathbf{V}_{\sigma}(t).$ Hence from the above we
see that $\mathcal{F}(t)\in R(\nabla_{\Gamma})$ for almost all $t\in[0,T]$.
Thus there exists some $p\in L^{2}_{0}(\Gamma(t))$ such that
$\nabla_{\Gamma}p=\mathcal{F}(t)$ in the distributional sense. The map
$t\mapsto\|p\|_{L^{2}(\Gamma(t))}$ is measurable by the same logic as in the
proof of [45], Theorem 4.2. Moreover, by using (7.5) we see that $p$ is unique
and one has
$\int_{0}^{T}\|p\|_{L^{2}(\Gamma(t))}^{2}\leq
C\int_{0}^{T}\|\mathcal{F}\|_{\mathbf{H}^{-1}(\Gamma(t))}^{2},$
where the latter term can be expressed in terms of
$\varphi,\mu,\mathbf{u}_{T}$. ∎
Lastly we want to return to the setting of non-solenoidal vectors. Letting
$(\varphi,\mu,\widehat{\mathbf{u}_{T}},p)$ be the solution from the previous
theorem, then by our construction of $\widetilde{\mathbf{u}_{T}}$ and the
bilinear forms $\mathbf{d}_{1},\mathbf{d}_{2}$, it is clear that
$\mathbf{u}_{T}:=\widehat{\mathbf{u}_{T}}-\widetilde{\mathbf{u}_{T}}$ is such
that
$\displaystyle\langle\partial^{\circ}\mathbf{u}_{T},\boldsymbol{\phi}\rangle_{\mathbf{H}^{-1}(\Gamma(t)),\mathbf{H}^{1}(\Gamma(t))}+\hat{\mathbf{a}}(\eta(\varphi),\mathbf{u}_{T},\boldsymbol{\phi})+\mathbf{c}_{1}(\mathbf{u}_{T},\mathbf{u}_{T},\boldsymbol{\phi})+\mathbf{l}(\mathbf{u}_{T},\boldsymbol{\phi})+m(p,\nabla_{\Gamma}\cdot\boldsymbol{\phi})=\mathbf{m}(\mathbf{F}_{T},\boldsymbol{\phi})+\mathbf{c}_{2}(\mu,\varphi,\boldsymbol{\phi}),$
$\displaystyle m(q,\nabla_{\Gamma}\cdot\mathbf{u}_{T})=-m(q,HV_{N}),$
$\displaystyle
m_{*}(\partial^{\circ}\varphi,\phi)+a(\mu,\phi)+\mathbf{c}_{2}(\phi,\varphi,\mathbf{u}_{T})=0,$
$\displaystyle m(\mu,\phi)=\varepsilon
a(\varphi,\phi)+\frac{1}{\varepsilon}m(F^{\prime}(\varphi),\phi),$
for all $q\in L^{2}(\Gamma(t)),\phi\in
H^{1}(\Gamma(t)),\boldsymbol{\phi}\in\mathbf{H}^{1}(\Gamma(t))$ for almost all
$t\in[0,T]$. Moreover we find that $\varphi\in H^{1}_{H^{-1}}\cap
L^{2}_{H^{1}},\mu\in L^{2}_{H^{1}},\mathbf{u}_{T}\in
H^{1}_{\mathbf{H}^{-1}}\cap L^{2}_{\mathbf{H}^{1}},p\in L^{2}_{L^{2}}$.
The initial condition for $\varphi$ is unchanged, but the initial condition
for $\mathbf{u}_{T}$ is required to be such that
$\mathbf{u}_{T}(0)=\mathbf{u}_{T,0}\in\widetilde{\mathbf{u}_{T}}+\mathbf{H}_{\sigma}(0)$.
One deduces the appropriate regularity for $\mathbf{u}_{T}$ from the
regularity of $\widehat{\mathbf{u}_{T}},\widetilde{\mathbf{u}_{T}}$. The above
arguments also work for the logarithmic potential, but we omit further
details.
###### Remark 7.3.
In this section we have not discriminated between the pressure, $p$, and the
modified pressure, $\tilde{p}$, as it is largely beside the point - that is
the existence of some Lagrange multiplier enforcing the divergence condition.
The distinction between these two pressures is discussed in Section 2.
Moreover, it is straightforward to establish that $p\in
L^{2}_{L^{2}}\Leftrightarrow\tilde{p}\in L^{2}_{L^{2}}$.
## 8 Concluding remarks
We have derived a system coupling the Navier-Stokes equations with the Cahn-
Hilliard equations on an evolving surface, and shown the well-posedness for a
prescribed, sufficiently smooth normal evolution. There is still much work to
be done on this topic, which we expound upon here.
Firstly, for the (evolving surface) Cahn-Hilliard equations with a logarithmic
potential one observes a “separation from the pure phases” where after some
small time the solution, $\varphi$, is such that $|\varphi|<1-\xi$ for some
small $\xi$ \- as was shown in [19]. This has been established for a Navier-
Stokes-Cahn-Hilliard system on a stationary domain in [29], and so it seems
reasonable it would extend to our setting.
If one does not prescribe the normal component of the velocity then the system
(1.1)-(1.4) also contains a geometric evolution equation, (2.8), which one
must solve. Unlike more standard geometric evolution equations, for example
mean curvature flow, this flow is essentially second order in time as one
considers the material derivative of the normal velocity. Indeed, even if one
ignores the Cahn-Hilliard component of (1.1)-(1.4) there are, to the authors’
knowledge, no results on the well-posedness of the evolving surface Navier-
Stokes equations (with unknown normal component) as discussed in [16, 45].
Moreover, the model we have considered is a diffuse interface model - and
depends strongly on the choice of the interface width, $\varepsilon$. It is
known that, in the sharp interface limit, $\varepsilon\rightarrow 0$, the
zero-level set of the solution of the Cahn-Hilliard equation (with a constant
mobility) converges in a suitably weak sense to the Mullins-Sekerka system,
also known as the Hele-Shaw system - see [5]. Likewise, it is known that the
analogous zero-level set from the Navier-Stokes-Cahn-Hilliard system converges
to a coupled Navier-Stokes-Mullins-Sekerka system - see for instance [3, 4].
However such results, or even formal asymptotics, have not been obtained for
the corresponding systems on an evolving surface - or even on a stationary
surface, to our knowledge. In particular, it would be interesting to study the
sharp interface limit of (1.1)-(1.4), as the limiting system should consist of
a coupling been a Navier-Stokes type equation for the surface velocity coupled
with the Mullins-Sekerka problem.
Lastly, there is interest in the numerical simulation of the system we have
considered (with or without a prescribed normal velocity). There has recently
(see [43]) been some numerical analysis of the tangential Navier-Stokes
equations, where the authors discretise by using the TraceFEM method - but
this has not yet been considered for the system (1.5)-(1.8). It therefore
would be interesting to see how existing results for a stationary domain, for
instance [48], adapt to an evolving surface.
### Acknowledgments
Thomas Sales is supported by the Warwick Mathematics Institute Centre for
Doctoral Training, and gratefully acknowledges funding from the University of
Warwick and the UK Engineering and Physical Sciences Research Council (Grant
number: EP/TS1794X/1). For the purpose of open access, the author has applied
a Creative Commons Attribution (CC BY) licence to any Author Accepted
Manuscript version arising from this submission. The authors would also like
to thank Achilles Mavrakis for discussions which resulted in Remark 6.4.
## Appendix A Laplace’s equation on an evolving surface
In this appendix we consider the regularity the solution of Laplace’s equation
on an evolving domain. For $t\in[0,T]$ we define $\Psi(t)$ to be the unique
weak solution of
$-\Delta_{\Gamma}\Psi(t)=H(t)V_{N}(t),$
on $\Gamma(t)$, subject to the constraint
$\mathchoice{{\vbox{\hbox{$\textstyle-$
}}\kern-7.83337pt}}{{\vbox{\hbox{$\scriptstyle-$
}}\kern-5.90005pt}}{{\vbox{\hbox{$\scriptscriptstyle-$
}}\kern-4.75003pt}}{{\vbox{\hbox{$\scriptscriptstyle-$
}}\kern-4.25003pt}}\\!\int_{\Gamma(t)}\Psi=0$. We note that this is well
defined since
$\int_{\Gamma(t)}H(t)V_{N}(t)=-2\int_{\Gamma(t)}\partial^{\circ}1=0.$
Here we denote the normal pushforward map as
$\Phi_{t}^{n}:\Gamma_{0}\rightarrow\Gamma(t),$ and $\Phi_{-t}^{n}$ denotes its
inverse. As these are $C^{2}$ diffeomorphisms the differentials
$D\Phi_{t}^{n}(p):T_{p}\Gamma_{0}\rightarrow T_{\Phi_{t}^{n}(p)}\Gamma(t)$ are
invertible. We recall the notation $J(p,t)=\det(D\Phi_{t}^{n}(p))$,
$J^{-1}(x,t)=\det(D\Phi_{-t}^{n}(x))=J(t,\Phi_{-t}^{n}x)^{-1}$,
$\mathbb{D}(p,t)=D\Phi_{t}^{n}(p))\mathbb{P}(p,0)$, and
$\mathbb{D}^{-1}(x,t)=D\Phi_{-t}^{n}(x))\mathbb{P}(x,t)$. These matrices are
such that $\mathbb{D}\mathbb{D}^{-1}=\mathbb{D}^{-1}\mathbb{D}=\mathbb{P}.$
###### Lemma A.1.
Let $\Psi$ be as above, and $\Gamma(t)$ be a $C^{3}$ evolving surface. Then
$\Psi\in C^{0}_{H^{3,p}}\cap C^{1}_{H^{1,p}},$ for all $p\in[1,\infty).$
###### Proof.
Let $\chi\in H^{1}(\Gamma_{0})$, then by the weak formulation of the above PDE
and the compatibility of $(H^{1}(\Gamma(t)),\Phi_{t}^{n})$ we find that
$\int_{\Gamma(t)}\nabla_{\Gamma}\Psi(t)\cdot\nabla_{\Gamma}\Phi_{t}^{n}\chi=\int_{\Gamma{t}}H(t)V_{N}(t)\Phi_{t}^{n}\chi.$
Hence by pulling back the integrals onto $\Gamma_{0}$ we see that
$\int_{\Gamma(t)}J(t)\mathbb{D}\nabla_{\Gamma}\Phi_{-t}^{n}\Psi(t)\cdot\mathbb{D}\nabla_{\Gamma}\chi=\int_{\Gamma_{0}}J(t)\Phi_{-t}^{n}(HV_{N})\chi,$
where the operators now are $\nabla_{\Gamma_{0}}.$ Similarly, the mean value
condition transforms as
$0=\int_{\Gamma(t)}\Psi(t)=\int_{\Gamma_{0}}J(t)\Phi_{-t}^{n}\Psi(t),$
and as such we focus on the function $\psi(t):=J(t)\Phi_{-t}^{n}\Psi(t)$,
where we see $\psi(t)\in H^{1}(\Gamma_{0})$ for all $t\in[0,T]$. We similarly
write $f(t):=J(t)\Phi_{-t}^{n}(HV_{n})\in H^{1}(\Gamma_{0}).$ It is then clear
that $\psi(t)$ solves the PDE
$\displaystyle\int_{\Gamma_{0}}\tilde{\mathbb{D}}(t)\nabla_{\Gamma}\psi(t)\cdot\nabla_{\Gamma}\chi+\psi(t)\boldsymbol{\omega}(t)\cdot\nabla_{\Gamma}\chi=\int_{\Gamma_{0}}f(t)\chi,$
(A.1)
for all $\chi\in H^{1}(\Gamma_{0})$, where
$\tilde{\mathbb{D}}=\mathbb{D}^{T}\mathbb{D},\qquad\boldsymbol{\omega}=J(t)\tilde{\mathbb{D}}\nabla_{\Gamma}(J(t)^{-1}).$
We note that clearly $\tilde{\mathbb{D}}$ is positive definite, and the
uniqueness of $\Psi$ implies uniqueness of $\psi$.
Then our assumptions on $\Phi_{t}^{n}$ imply we have sufficient smoothness so
that we may apply elliptic regularity theory to see that
$\|\psi(t)\|_{H^{3,p}(\Gamma_{0})}\leq C\|f(t)\|_{H^{1,p}(\Gamma_{0})},$
for $p\in[1,\infty)$ and $C$ depends on
$p,\Gamma_{0},\tilde{\mathbb{D}}(t),\boldsymbol{\omega}(t)$. It is
straightforward to see that by considering (A.1) at two times $t,s\in[0,T]$,
and noting that $\tilde{\mathbb{D}},\boldsymbol{\omega}$ are $C^{2}$ in $t$
and $f$ is $C^{1}$ in $t$, that the map
$t\mapsto\|\psi(t)\|_{H^{3,p}(\Gamma_{0})}$ is continuous on $[0,T]$. We omit
further details on this calculation.
Next we show that $\psi$ has a strong derivative. By considering (A.1) at
times $t\in[0,T)$ and $t+h$ for some small $h>0$ so that $t+h\in(0,T)$ we find
that
$\frac{1}{h}\int_{\Gamma_{0}}\left(\tilde{\mathbb{D}}(t+h)\nabla_{\Gamma}\psi(t+h)-\tilde{\mathbb{D}}(t)\nabla_{\Gamma}\psi(t)\right)\cdot\nabla_{\Gamma}\chi+\left(\psi(t+h)\boldsymbol{\omega}(t+h)-\psi(t)\boldsymbol{\omega}(t)\right)\cdot\nabla_{\Gamma}\chi\\\
=\frac{1}{h}\int_{\Gamma_{0}}\left(f(t+h)-f(t)\right)\chi,$
for all $\chi\in H^{1}(\Gamma_{0})$. We write this in terms of difference
quotients as
$\int_{\Gamma_{0}}\left(\Delta_{h}\tilde{\mathbb{D}}(t)\nabla_{\Gamma}\psi(t+h)+\tilde{\mathbb{D}}(t)\nabla_{\Gamma}\Delta_{h}\psi(t)\right)\cdot\nabla_{\Gamma}\chi+\left(\psi(t+h)\Delta_{h}\boldsymbol{\omega}(t)+\Delta_{h}\psi(t)\boldsymbol{\omega}(t)\right)\cdot\nabla_{\Gamma}\chi=\int_{\Gamma_{0}}\Delta_{h}f(t)\chi,$
where $\Delta_{h}X(t)=\frac{X(t+h)-X(t)}{h}$ for some quantity $X$. Now by
letting $h,h^{\prime}>0$ be sufficiently small one readily finds that
$\displaystyle\langle
L(t)(\Delta_{h}\psi(t)-\Delta_{h^{\prime}}\psi(t)),\Delta_{h}\psi(t)-\Delta_{h^{\prime}}\psi(t)\rangle$
$\displaystyle\leq
C\|\Delta_{h^{\prime}}\tilde{\mathbb{D}}(t)\nabla_{\Gamma}\psi(t+h^{\prime})-\Delta_{h}\tilde{\mathbb{D}}(t)\nabla_{\Gamma}\psi(t+h)\|_{L^{2}(\Gamma_{0})}^{2}$
$\displaystyle+C\|\psi(t+h^{\prime})\Delta_{h^{\prime}}\boldsymbol{\omega}(t)-\psi(t+h)\Delta_{h}\boldsymbol{\omega}(t)\|_{\mathbf{L}^{2}(\Gamma_{0})}^{2}$
$\displaystyle+C\|\Delta_{h}f(t)-\Delta_{h^{\prime}}f(t)\|_{L^{2}(\Gamma_{0})}^{2}+\gamma\|\nabla_{\Gamma}(\Delta_{h}\psi(t)-\Delta_{h^{\prime}}\psi(t))\|_{L^{2}(\Gamma_{0})}^{2},$
for some small $\gamma$ to be determined. Here
$L(t)\in\mathcal{L}(H^{1}(\Gamma_{0})\cap
L_{0}^{2}(\Gamma_{0}),(H^{1}(\Gamma_{0})\cap L_{0}^{2}(\Gamma_{0}))^{\prime})$
is the operator defined so that
$\langle
L(t)\zeta,\chi\rangle=\int_{\Gamma_{0}}\tilde{\mathbb{D}}(t)\nabla_{\Gamma}\zeta\cdot\nabla_{\Gamma}\chi+\zeta\boldsymbol{\omega}(t)\cdot\nabla_{\Gamma}\chi,$
where $L^{2}_{0}(\Gamma_{0})$ is the subspace of $L^{2}(\Gamma_{0})$
containing elements such that $\mathchoice{{\vbox{\hbox{$\textstyle-$
}}\kern-7.83337pt}}{{\vbox{\hbox{$\scriptstyle-$
}}\kern-5.90005pt}}{{\vbox{\hbox{$\scriptscriptstyle-$
}}\kern-4.75003pt}}{{\vbox{\hbox{$\scriptscriptstyle-$
}}\kern-4.25003pt}}\\!\int_{\Gamma_{0}}\phi=0$. By pushing the integral
forward onto $\Gamma(t)$, in the reverse to the beginning of the proof, we can
observe $L(t)$ is elliptic by the ellipticity of $-\Delta_{\Gamma(t)}$ on
$H^{1}(\Gamma(t))\cap L_{0}^{2}(\Gamma(t)),$ and moreover the ellipticity
constant is independent of $t$. Thus there exists some constant $\kappa$ such
that
$\langle
L(t)(\Delta_{h}\psi(t)-\Delta_{h^{\prime}}\psi(t)),\Delta_{h}\psi(t)-\Delta_{h^{\prime}}\psi(t)\rangle\geq\kappa\|\nabla_{\Gamma}(\Delta_{h}\psi(t)-\Delta_{h^{\prime}}\psi(t))\|_{L^{2}(\Gamma_{0})}^{2},$
and hence choosing $\gamma=\frac{\kappa}{2}$, and using Poincaré’s inequality
on $\Gamma_{0}$ we see
$\displaystyle\|\Delta_{h}\psi(t)-\Delta_{h^{\prime}}\psi(t)\|_{H^{1}(\Gamma_{0})}^{2}$
$\displaystyle\leq
C\|\Delta_{h^{\prime}}\tilde{\mathbb{D}}(t)\nabla_{\Gamma}\psi(t+h^{\prime})-\Delta_{h}\tilde{\mathbb{D}}(t)\nabla_{\Gamma}\psi(t+h)\|_{L^{2}(\Gamma_{0})}^{2}$
$\displaystyle+C\|\psi(t+h^{\prime})\Delta_{h^{\prime}}\boldsymbol{\omega}(t)-\psi(t+h)\Delta_{h}\boldsymbol{\omega}(t)\|_{\mathbf{L}^{2}(\Gamma_{0})}^{2}$
$\displaystyle+C\|\Delta_{h}f(t)-\Delta_{h^{\prime}}f(t)\|_{L^{2}(\Gamma_{0})}^{2},$
for some constants $C(t)$ depending on $\Gamma_{0}$, and the ellipticity of
$L(t)$. Now, by the differentiability of
$\tilde{\mathbb{D}},\boldsymbol{\omega},f$, and the continuity of $\psi$ it is
clear that by taking $h,h^{\prime}$ sufficiently small that we can make
$\|\Delta_{h}\psi(t)-\Delta_{h^{\prime}}\psi(t)\|_{H^{1}(\Gamma_{0})}$
arbitrarily small. Thus $\Delta_{h}\psi(t)$ is a Cauchy sequence in
$H^{1}(\Gamma_{0})$ and a right time derivative of $\psi$ exists at
$t\in[0,T).$ A similar calculation verifies that a left time derivative exists
too.
Differentiating (A.1) in time we find
$\displaystyle\int_{\Gamma_{0}}\frac{\partial\tilde{\mathbb{D}}}{\partial
t}(t)\nabla_{\Gamma}\psi(t)\cdot\nabla_{\Gamma}\chi+\tilde{\mathbb{D}}(t)\nabla_{\Gamma}\frac{\partial\psi}{\partial
t}(t)\cdot\nabla_{\Gamma}\chi+\frac{\partial\psi}{\partial
t}(t)\boldsymbol{\omega}(t)\cdot\nabla_{\Gamma}\chi+\psi(t)\frac{\partial\boldsymbol{\omega}}{\partial
t}(t)\cdot\nabla_{\Gamma}\chi=\int_{\Gamma_{0}}\frac{\partial f}{\partial
t}(t)\chi,$
for all $\chi\in H^{1}(\Gamma_{0})$, $t\in[0,T]$. As above, by noting that
$\tilde{\mathbb{D}},\boldsymbol{\omega}$ are $C^{2}$ in $t$ and $f$ is $C^{1}$
in $t$, one can now readily observe that the map
$t\mapsto\|\frac{\partial\psi}{\partial t}(t)\|_{H^{1}(\Gamma_{0})}$ is
continuous on $[0,T]$. Applying elliptic regularity theory we find that
$\psi\in C^{0}([0,T];H^{3,p}(\Gamma_{0}))\cap
C^{1}([0,T];H^{1,p}(\Gamma_{0}))$, and using the hence the compatibility of
$(H^{3,p}(\Gamma(t)),\Phi_{t}^{n})$ (and uniform bounds on $J(t)$ where
needed) it follows that $\Psi\in C^{0}_{H^{1,p}}\cap C^{1}_{H^{1,p}}.$ ∎
## Appendix B Killing Vectors and the inverse Stokes operator
In this appendix we discuss a solution operator related to the surface Stokes
equation. We refer the reader to [15, 34] for further details. For
$t\in[0,T]$, and a given $\boldsymbol{\phi}\in\mathbf{H}_{\sigma}(t)$ we are
interested in finding a solution
$\mathcal{S}\boldsymbol{\phi}\in\mathbf{V}_{\sigma}(t)$ solving
$-\mathbb{P}\nabla_{\Gamma}\cdot(2\mathbb{E}(\mathcal{S}\boldsymbol{\phi}))=\boldsymbol{\phi},\text{
on }\Gamma(t),$
in a weak sense. However, for general
$\boldsymbol{\phi}\in\mathbf{H}_{\sigma}(t)$ there is not a unique solution as
$\mathbf{E}(\cdot)$ has a non-trivial kernel. However one can obtain a unique
solution by filtering out Killing vectors, that is
$\boldsymbol{\psi}\in\mathcal{K}(t):=\\{\boldsymbol{\phi}\in\mathbf{V}_{\sigma}(t)\mid\mathbb{E}(\boldsymbol{\phi})=0\\}.$
This is analogous to considering the subspace $L^{2}_{0}(\Gamma(t)\leq
L^{2}(\Gamma(t))$ when solving the Laplace equation. Clearly we have that
$\mathcal{K}(t)$ is a closed subspace of $\mathbf{V}_{\sigma}(t)$, and it is
known to have dimension $\leq 3$ (see [47], Proposition III.6.5). We then
define an orthogonal space $\mathcal{K}^{\perp}$, with respect to the
$\mathbf{L}^{2}$ inner product, so that
$\mathbf{H}_{\sigma}(t)=\mathcal{K}(t)\oplus\mathcal{K}^{\perp}(t)$.
For $\boldsymbol{\phi}\in\mathcal{K}^{\perp}(t)$ we now define the inverse
Stokes’ operator to be the unique solution,
$\mathcal{S}\boldsymbol{\phi}\in\mathbf{V}_{\sigma}(t)\cap\mathcal{K}^{\perp}(t)$,
of
$\mathbf{a}(\mathcal{S}\boldsymbol{\phi},\boldsymbol{\psi})=\mathbf{m}(\boldsymbol{\phi},\boldsymbol{\psi}),$
for all $\boldsymbol{\psi}\in\mathbf{V}_{\sigma}(t)$. The well-posedness of
this owes to the following Korn-type inequality.
###### Lemma B.1 ([34], Lemma 4.1).
There exists a constant, $C$, independent of $t$, such that for
$\boldsymbol{\phi}\in\mathbf{V}_{\sigma}(t)\cap\mathcal{K}^{\perp}(t)$ we have
$\displaystyle\|\boldsymbol{\phi}\|_{\mathbf{H}^{1}(\Gamma(t))}\leq
C\|\mathbb{E}(\boldsymbol{\phi})\|_{\mathbf{L}^{2}(\Gamma(t))}.$ (B.1)
The fact this constant is independent of $t$ follows from the same logic as in
the proof of [45] Lemma 3.2. One also obtains an inequality for
$\boldsymbol{\phi}\in\mathcal{K}(t)$,
$\displaystyle\|\boldsymbol{\phi}\|_{\mathbf{H}^{1}(\Gamma(t))}\leq
C\|\boldsymbol{\phi}\|_{\mathbf{L}^{2}(\Gamma(t))},$ (B.2)
which follows from (3.9) and the definition of $\mathcal{K}(t)$. In fact,
using (3.10), the Sobolev embedding
$\mathbf{H}^{1}(\Gamma(t))\hookrightarrow\mathbf{L}^{p}(\Gamma(t))$, and the
above inequality one finds
$\displaystyle\|\boldsymbol{\phi}\|_{\mathbf{H}^{1,p}(\Gamma(t))}\leq
C\|\boldsymbol{\phi}\|_{\mathbf{L}^{2}(\Gamma(t))},$ (B.3)
for $p\in[1,\infty)$.
As in [15], we have a sufficiently smooth surface, $\Gamma(t)$, so that one
has improved regularity
$\displaystyle\|\mathcal{S}\boldsymbol{\phi}\|_{\mathbf{H}^{2}(\Gamma(t))}\leq
C\|\boldsymbol{\phi}\|_{\mathbf{L}^{2}(\Gamma(t)},$ (B.4)
and the constant $C$ is independent of $t$ by the usual arguments.
With this inverse Stokes operator, and (B.1), one can define a norm on
$\mathcal{K}^{\perp}(t)$ given by
$\|\boldsymbol{\phi}\|_{\perp}^{2}:=\mathbf{a}(\mathcal{S}\boldsymbol{\phi},\mathcal{S}\boldsymbol{\phi})=\mathbf{m}(\boldsymbol{\phi},\mathcal{S}\boldsymbol{\phi}),$
and it is straightforward to see that $\|\boldsymbol{\phi}\|_{\perp}\leq
C\|\boldsymbol{\phi}\|_{L^{2}(\Gamma(t))},$ for the constant $C$ in (B.1).
To extend $\mathcal{S}$ to general elements of $\mathbf{H}_{\sigma}(t)$ we use
projections. For $\boldsymbol{\phi}\in\mathbf{V}_{\sigma}(t)$ we define the
projections
$P_{\mathcal{K}}\boldsymbol{\phi}\in\mathcal{K}(t),P_{\mathcal{K}^{\perp}}\boldsymbol{\phi}\in\mathcal{K}^{\perp}(t)$
to be the unique solutions of
$\displaystyle\mathbf{m}(P_{\mathcal{K}}\boldsymbol{\phi},\boldsymbol{\psi})=\mathbf{m}(\boldsymbol{\phi},\boldsymbol{\psi}),\text{
for all }\boldsymbol{\psi}\in\mathcal{K}(t),$
$\displaystyle\mathbf{m}(P_{\mathcal{K}^{\perp}}\boldsymbol{\phi},\boldsymbol{\psi})=\mathbf{m}(\boldsymbol{\phi},\boldsymbol{\psi}),\text{
for all }\boldsymbol{\psi}\in\mathcal{K}^{\perp}(t),$
respectively. It is clear that
$\|P_{\mathcal{K}}\boldsymbol{\phi}\|_{\mathbf{L}^{2}(\Gamma(t))}\leq\|\boldsymbol{\phi}\|_{\mathbf{L}^{2}(\Gamma(t))}$
and
$\|P_{\mathcal{K}^{\perp}}\boldsymbol{\phi}\|_{\mathbf{L}^{2}(\Gamma(t))}\leq\|\boldsymbol{\phi}\|_{\mathbf{L}^{2}(\Gamma(t))}$.
Using this we now extend $\mathcal{S}$ to $\mathbf{V}_{\sigma}(t)$ as follows.
For $\boldsymbol{\phi}\in\mathbf{V}_{\sigma}(t)$ we define
$\mathcal{S}^{\perp}\boldsymbol{\phi}\in\mathcal{K}^{\perp}(t)$ to be the
unique solution of
$\mathbf{a}(\mathcal{S}^{\perp}\boldsymbol{\phi},\boldsymbol{\psi})=\mathbf{m}(P_{\mathcal{K}^{\perp}}\boldsymbol{\phi},\boldsymbol{\psi}),$
for all $\boldsymbol{\psi}\in\mathbf{V}_{\sigma}(t)$. This is just the
composition $\mathcal{S}^{\perp}=\mathcal{S}\circ P_{\mathcal{K}^{\perp}}$,
but it is convenient to have a shorthand notation for this. It is clear that
from the above properties of the projections that
$\|P_{\mathcal{K}^{\perp}}\boldsymbol{\phi}\|_{\perp}\leq
C\|\boldsymbol{\phi}\|_{\mathbf{L}^{2}(\Gamma(t))}.$
These operators arise in our proof of uniqueness, Theorem 6.1. In order to
discuss uniqueness, we need to consider the differentiability of
$P_{\mathcal{K}}\boldsymbol{\phi},P_{\mathcal{K}^{\perp}}\boldsymbol{\phi}$ in
time. We denote the killing vectors by
$\boldsymbol{\kappa}_{1},...,\boldsymbol{\kappa}_{N_{K}}$ which are determined
by $\Gamma(t)$, where we have used the notation $N_{K}:=\dim\mathcal{K}(t)\leq
3.$ We note that $N_{K}$ is independent of time, as the surfaces $\Gamma(t)$
are all diffeomorphic. We choose Killing vectors, $\boldsymbol{\kappa}_{i}$,
such that they are orthonormal with respect to the $\mathbf{L^{2}}(\Gamma(t))$
inner product. The projection $P_{\mathcal{K}}$ is then seen to be such that
$P_{\mathcal{K}}\boldsymbol{\phi}=\sum_{i=1}^{N_{K}}\mathbf{m}(\boldsymbol{\phi},\boldsymbol{\kappa}_{i})\boldsymbol{\kappa}_{i}.$
Thus, heuristically, one expects that for sufficiently smooth
$\boldsymbol{\phi}$,
$\partial^{\circ}P_{\mathcal{K}}\boldsymbol{\phi}=P_{\mathcal{K}}\partial^{\circ}\boldsymbol{\phi}+\text{lower
order terms in }\boldsymbol{\phi},$
and likewise an analogous result for
$P_{\mathcal{K}^{\perp}}=\text{id}-P_{\mathcal{K}}$.
We will show this in a later lemma, but first need to see that we can map from
$\mathcal{K}(0)\rightarrow\mathcal{K}(t)$ in a sufficiently smooth way.
###### Lemma B.2.
There exists a $C^{1}_{\mathbf{H}^{1}}$ mapping from $\mathcal{K}(0)$ onto
$\mathcal{K}(t)$.
###### Proof.
Firstly we note that our definition of a Killing vector is equivalent to the
more standard definition of a Killing vector on a Riemannian manifold, defined
in terms of the Lie derivative. That is for some
$\boldsymbol{\kappa}\in\mathcal{K}(t),$
$\mathbb{E}(\boldsymbol{\kappa})=0\text{ on
}\Gamma(t)\Leftrightarrow\mathcal{L}_{\boldsymbol{\kappa}}g(t)=0,$
where $g(t)$ is the associated metric for $\Gamma(t)$. To see this we refer to
reader to [20], Appendix A. As $\Gamma(t)$ is an embedded surface in
$\mathbb{R}^{3}$, the natural metric on $\Gamma(t)$ is given by
$g(t)=x(t)^{*}\mathfrak{e},$ the pullback of the Euclidean metric,
$\mathfrak{e}$, on $\mathbb{R}^{3}$ by a $C^{3}$ embedding
$x(t):\Gamma_{0}\rightarrow\Gamma(t)$. Here, we consider $\Gamma(t)$ to be
parametrised over $\Gamma_{0}$, with the embedding
$x:\Gamma_{0}\times[0,T]\rightarrow\mathcal{G}_{T}$ given by the unique
solution of
$\frac{dx}{dt}=V_{N}(x(t),t),\qquad x(0)=x_{0},$
for all points $x_{0}\in\Gamma_{0}$. In particular this lets us see that
$g(t)=x(t)^{*}g(0)$, and hence
$x(t)^{*}g(0)=x(t)^{*}(x(0)^{*}\mathfrak{e})=(x(t)\circ
x(0))^{*}\mathfrak{e}=x(t)^{*}\mathfrak{e}=g(t).$
Now it is known that the Lie derivative commutes with pullbacks, as can be
seen from the fact that the exterior derivative commutes with pullbacks and
writing the Lie derivative in terms of the exterior derivative - see for
example [35], Theorem 2.3.4. Hence for $\boldsymbol{\kappa}\in\mathcal{K}(0)$
we see $\mathcal{L}_{\boldsymbol{\kappa}}g(0)=0$, and so
$0=x(t)^{*}\left(\mathcal{L}_{\boldsymbol{\kappa}}g(0)\right)=\mathcal{L}_{x(t)^{*}\boldsymbol{\kappa}}x(t)^{*}g(0)=\mathcal{L}_{x(t)^{*}\boldsymbol{\kappa}}g(t).$
Hence the map $\boldsymbol{\kappa}\mapsto x(t)^{*}\boldsymbol{\kappa}$ is such
that $\mathcal{K}(0)\rightarrow\mathcal{K}(t)$. Moreover we see that the
choice of embedding we used that
$x(t)^{*}\boldsymbol{\kappa}=\Phi_{t}^{n}\boldsymbol{\kappa}$, and
$\Phi_{t}^{n}\boldsymbol{\kappa}$ clearly has a strong material time
derivative (which vanishes by definition). ∎
Given an orthonormal basis (with respect to $\mathbf{L}^{2}(\Gamma_{0})$),
$\boldsymbol{\kappa}_{1},...,\boldsymbol{\kappa}_{N_{K}}$, of
$\mathcal{K}(0)$, the above calculations let us see that
$\Phi_{t}^{n}\boldsymbol{\kappa}_{1},...,\Phi_{t}^{n}\boldsymbol{\kappa}_{N_{K}}$
form a basis of $\mathcal{K}(t)$. However this new basis is not necessarily
orthonormal with respect to $\mathbf{L}^{2}(\Gamma(t))$, but using the Gram-
Schmidt procedure on
$\Phi_{t}^{n}\boldsymbol{\kappa}_{1},...,\Phi_{t}^{n}\boldsymbol{\kappa}_{N_{K}}$,
we obtain an orthonormal basis which we label
$\boldsymbol{\kappa}_{1}(t),...,\boldsymbol{\kappa}_{N_{K}}(t)$. In using the
Gram-Schmidt process we no longer retain the fact that the material time
derivative vanishes, but it is straightforward to show that
$\partial^{\circ}\boldsymbol{\kappa}_{i}\in C^{0}_{\mathbf{H}^{1}}$.
We now show the desired differentiability of the Killing projections.
###### Lemma B.3.
For $\Gamma(t)$ a $C^{3}$ evolving surface we have the following.
1. 1.
If $\boldsymbol{\phi}\in L^{2}_{\mathbf{V}_{\sigma}}\cap
H^{1}_{\mathbf{V}_{\sigma}^{\prime}}$ then
$P_{\mathcal{K}}\boldsymbol{\phi}\in L^{2}_{\mathbf{V}_{\sigma}}\cap
H^{1}_{\mathbf{V}_{\sigma}^{\prime}}$ with
$\displaystyle\int_{0}^{T}\left\|\partial^{\circ}P_{\mathcal{K}}\boldsymbol{\phi}\right\|_{\mathbf{V}_{\sigma}(t)^{\prime}}^{2}$
$\displaystyle\leq
C\int_{0}^{T}\left\|\partial^{\circ}\boldsymbol{\phi}\right\|_{\mathbf{V}_{\sigma}(t)^{\prime}}^{2}+C\int_{0}^{T}\|\boldsymbol{\phi}\|_{L^{2}(\Gamma(t)}^{2},$
$\displaystyle\int_{0}^{T}\left\|\partial^{\circ}P_{\mathcal{K}^{\perp}}\boldsymbol{\phi}\right\|_{\mathbf{V}_{\sigma}(t)^{\prime}}^{2}$
$\displaystyle\leq
C\int_{0}^{T}\left\|\partial^{\circ}\boldsymbol{\phi}\right\|_{\mathbf{V}_{\sigma}(t)^{\prime}}^{2}+C\int_{0}^{T}\|\boldsymbol{\phi}\|_{L^{2}(\Gamma(t)}^{2}$
2. 2.
If $\boldsymbol{\phi}\in L^{2}_{\mathbf{V}_{\sigma}}\cap
H^{1}_{\mathbf{V}_{\sigma}^{\prime}}$ is such that
$\boldsymbol{\phi}(t)\in\mathcal{K}^{\perp}$ for almost all $t\in[0,T]$ then
$\mathcal{S}\boldsymbol{\phi}\in H^{1}_{\mathbf{V}_{\sigma}}$ with
$\int_{0}^{T}\left\|\partial^{\circ}\mathcal{S}\boldsymbol{\phi}\right\|_{\mathbf{H}^{1}(\Gamma(t))}^{2}\leq
C\int_{0}^{T}\left(\|\mathcal{S}\boldsymbol{\phi}\|_{\mathbf{H}^{1}(\Gamma(t))}^{2}+\|\partial^{\circ}\boldsymbol{\phi}\|_{\mathbf{V}_{\sigma}(t)^{\prime}}^{2}+\|\boldsymbol{\phi}\|_{\mathbf{L}^{2}(\Gamma(t))}^{2}\right).$
###### Proof.
For the first part of this one considers
$\boldsymbol{\phi},\boldsymbol{\psi}\in H^{1}_{\mathbf{V}_{\sigma}}$ and by
using the definition of $P_{\mathcal{K}}$ we see
$\mathbf{m}(P_{\mathcal{K}}\boldsymbol{\phi},\boldsymbol{\psi})=\sum_{i=1}^{N_{K}}\mathbf{m}(\boldsymbol{\phi},\boldsymbol{\kappa}_{i})\mathbf{m}(\boldsymbol{\psi},\boldsymbol{\kappa}_{i}),$
where $\boldsymbol{\kappa}_{i}$ are an orthonormal basis of $\mathcal{K}(t)$
with respect to the $\mathbf{L}^{2}(\Gamma(t))$ inner product. Differentiating
the above yields
$\mathbf{m}(\partial^{\circ}P_{\mathcal{K}}\boldsymbol{\phi},\boldsymbol{\psi})+\mathbf{m}(P_{\mathcal{K}}\boldsymbol{\phi},\partial^{\circ}\boldsymbol{\psi})+\mathbf{m}(P_{\mathcal{K}}\boldsymbol{\phi},\boldsymbol{\psi}HV_{N})=\sum_{i=1}^{N_{K}}\left[\mathbf{m}(\partial^{\circ}\boldsymbol{\phi},\boldsymbol{\kappa}_{i})+\mathbf{m}(\boldsymbol{\phi},\partial^{\circ}\boldsymbol{\kappa}_{i})+\mathbf{m}(\boldsymbol{\phi}HV_{N},\boldsymbol{\kappa}_{i})\right]\mathbf{m}(\boldsymbol{\psi},\boldsymbol{\kappa}_{i})\\\
+\sum_{i=1}^{N_{K}}\mathbf{m}(\boldsymbol{\phi},\boldsymbol{\kappa}_{i})\left[\mathbf{m}(\partial^{\circ}\boldsymbol{\psi},\boldsymbol{\kappa}_{i})+\mathbf{m}(\boldsymbol{\psi},\partial^{\circ}\boldsymbol{\kappa}_{i})+\mathbf{m}(\boldsymbol{\psi}HV_{N},\boldsymbol{\kappa}_{i})\right],$
from which, by expanding terms on the left similarly, one finds
$\displaystyle\mathbf{m}(\partial^{\circ}P_{\mathcal{K}}\boldsymbol{\phi},\boldsymbol{\psi})=\mathbf{m}(\partial^{\circ}\boldsymbol{\phi},P_{\mathcal{K}}\boldsymbol{\psi})+\sum_{i=1}^{N_{K}}\left[\mathbf{m}(\boldsymbol{\phi},\partial^{\circ}\boldsymbol{\kappa}_{i})+\mathbf{m}(\boldsymbol{\phi}HV_{N},\boldsymbol{\kappa}_{i})\right]\mathbf{m}(\boldsymbol{\psi},\boldsymbol{\kappa}_{i})+\mathbf{m}(\boldsymbol{\phi},\boldsymbol{\kappa}_{i})\mathbf{m}(\boldsymbol{\psi},\partial^{\circ}\boldsymbol{\kappa}_{i}).$
From this one can use an approximation argument to extend to
$\boldsymbol{\phi}\in L^{2}_{\mathbf{V}_{\sigma}}\cap
H^{1}_{\mathbf{V}_{\sigma}^{\prime}},\boldsymbol{\psi}\in
L^{2}_{\mathbf{V}_{\sigma}}$, and see that more generally
$\displaystyle\mathbf{m}_{*}(\partial^{\circ}P_{\mathcal{K}}\boldsymbol{\phi},\boldsymbol{\psi})=\mathbf{m}_{*}(\partial^{\circ}\boldsymbol{\phi},P_{\mathcal{K}}\boldsymbol{\psi})+\sum_{i=1}^{N_{K}}\left[\mathbf{m}(\boldsymbol{\phi},\partial^{\circ}\boldsymbol{\kappa}_{i})+\mathbf{m}(\boldsymbol{\phi}HV_{N},\boldsymbol{\kappa}_{i})\right]\mathbf{m}(\boldsymbol{\psi},\boldsymbol{\kappa}_{i})+\mathbf{m}(\boldsymbol{\phi},\boldsymbol{\kappa}_{i})\mathbf{m}(\boldsymbol{\psi},\partial^{\circ}\boldsymbol{\kappa}_{i}),$
(B.5)
from which one obtains
$|\mathbf{m}_{*}(\partial^{\circ}P_{\mathcal{K}}\boldsymbol{\phi},\boldsymbol{\psi})|=\|\partial^{\circ}\boldsymbol{\phi}\|_{\mathbf{V}_{\sigma}^{\prime}}\|P_{\mathcal{K}}\boldsymbol{\psi}\|_{\mathbf{H}^{1}(\Gamma(t))}\\\
+\|\boldsymbol{\phi}\|_{\mathbf{L}^{2}(\Gamma(t)}\|\boldsymbol{\psi}\|_{\mathbf{L}^{2}(\Gamma(t))}\sum_{i=1}^{N_{K}}\left[2\|\partial^{\circ}\boldsymbol{\kappa}_{i}\|_{\mathbf{L}^{2}(\Gamma(t))}\|\boldsymbol{\kappa}_{i}\|_{\mathbf{L}^{2}(\Gamma(t))}+\|HV_{N}\|_{L^{\infty}(\Gamma(t))}\|\boldsymbol{\kappa}_{i}\|_{\mathbf{L}^{2}(\Gamma(t))}^{2}\right].$
Now by using (B.2) we see
$\|P_{\mathcal{K}}\boldsymbol{\psi}\|_{\mathbf{H}^{1}(\Gamma(t))}\leq
C\|\boldsymbol{\psi}\|_{\mathbf{H}^{1}(\Gamma(t))}$, from which one can now
readily deduce the $L^{2}_{\mathbf{V}_{\sigma}^{\prime}}$ bound for
$\partial^{\circ}P_{\mathcal{K}}\phi$ by integrating over $[0,T]$. The
argument for bounding
$\partial^{\circ}P_{\mathcal{K}^{\perp}}\boldsymbol{\phi}$ is similar, where
we exploit the fact that
$P_{\mathcal{K}^{\perp}}\boldsymbol{\phi}=\boldsymbol{\phi}-P_{\mathcal{K}}\boldsymbol{\phi}$.
We omit the details.
For the second part of this result, one differentiates the equation
$\mathbf{a}(\mathcal{S}\boldsymbol{\phi},\boldsymbol{\psi})=\mathbf{m}(\boldsymbol{\phi},\boldsymbol{\psi}),$
where we take $\boldsymbol{\psi}\in H^{1}_{\mathbf{V}_{\sigma}}$. This yields
$\mathbf{a}\left(\partial^{\circ}\mathcal{S}\boldsymbol{\phi},\boldsymbol{\psi}\right)+\mathbf{a}\left(\mathcal{S}\boldsymbol{\phi},\partial^{\circ}\boldsymbol{\psi}\right)+\mathbf{b}(\mathcal{S}\boldsymbol{\phi},\boldsymbol{\psi})=\mathbf{m}_{*}\left(\partial^{\circ}\boldsymbol{\phi},\boldsymbol{\psi}\right)+\mathbf{m}\left(\boldsymbol{\phi},\partial^{\circ}\boldsymbol{\psi}\right)+\mathbf{m}(\boldsymbol{\phi},\boldsymbol{\psi}HV_{N}),$
and we note that by definition
$\mathbf{a}\left(\mathcal{S}\boldsymbol{\phi},\partial^{\circ}\boldsymbol{\psi}\right)=\mathbf{m}\left(\boldsymbol{\phi},\partial^{\circ}\boldsymbol{\psi}\right).$
Hence we find $\partial^{\circ}\mathcal{S}\boldsymbol{\phi}$ solves
$\displaystyle\mathbf{a}\left(\partial^{\circ}\mathcal{S}\boldsymbol{\phi},\boldsymbol{\psi}\right)=\mathbf{m}_{*}\left(\partial^{\circ}\boldsymbol{\phi},\boldsymbol{\psi}\right)+\mathbf{m}(\boldsymbol{\phi},\boldsymbol{\psi}HV_{N})-\mathbf{b}(\mathcal{S}\boldsymbol{\phi},\boldsymbol{\psi}),$
(B.6)
for all $\boldsymbol{\psi}\in L^{2}_{\mathbf{V}_{\sigma}}$, where we have
extended to $L^{2}_{\mathbf{V}_{\sigma}}$ by density. One can verify that for
a fixed $t$, say $t^{*}$, and $\boldsymbol{\psi}\in\mathcal{K}(t^{*})$ that
$\mathbf{m}_{*}\left(\partial^{\circ}\boldsymbol{\phi},\boldsymbol{\psi}\right)+\mathbf{m}(\boldsymbol{\phi},\boldsymbol{\psi}HV_{N})-\mathbf{b}(\mathcal{S}\boldsymbol{\phi},\boldsymbol{\psi})=0,$
by considering the function $\Phi_{t}^{n}\Phi_{-t^{*}}^{n}\boldsymbol{\psi}$
which has vanishing strong material derivative so that
$\mathbf{m}_{*}\left(\partial^{\circ}\boldsymbol{\phi},\boldsymbol{\psi}\right)+\mathbf{m}(\boldsymbol{\phi},\boldsymbol{\psi}HV_{N})=\frac{d}{dt}\mathbf{m}(\boldsymbol{\phi},\boldsymbol{\psi})=0.$
Likewise $\mathbf{b}(\mathcal{S}\boldsymbol{\phi},\boldsymbol{\psi})=0$ as
$\mathbb{E}(\boldsymbol{\psi})=0$. Hence one observes that the right-hand side
of (B.7) is a bounded linear functional acting on $\mathbf{V}_{\sigma}(t)$,
which vanishes on $\mathcal{K}(t)$. Hence using the coercivity result (B.1)
and the Lax-Milgram theorem one finds there is a unique solution of (B.6) in
$\mathcal{K}^{\perp}(t)$, and from the above we see this must be
$P_{\mathcal{K}^{\perp}}\partial^{\circ}\mathcal{S}\boldsymbol{\phi}$.
Moreover we see that this is bounded by
$\displaystyle\|P_{\mathcal{K}^{\perp}}\partial^{\circ}\mathcal{S}\boldsymbol{\phi}\|_{\mathbf{H}^{1}(\Gamma(t))}\leq
C\left(\|\mathcal{S}\boldsymbol{\phi}\|_{\mathbf{H}^{1}(\Gamma(t))}^{2}+\|\partial^{\circ}\boldsymbol{\phi}\|_{\mathbf{V}_{\sigma}(t)^{\prime}}^{2}+\|\boldsymbol{\phi}\|_{\mathbf{L}^{2}(\Gamma(t))}^{2}\right).$
(B.7)
It remains to bound
$P_{\mathcal{K}}\partial^{\circ}\mathcal{S}\boldsymbol{\phi}$. Firstly, we use
a duality argument to bound
$P_{\mathcal{K}}\partial^{\circ}\mathcal{S}\boldsymbol{\phi}$, writing
$\|P_{\mathcal{K}}\partial^{\circ}\mathcal{S}\boldsymbol{\phi}\|_{\mathbf{L}^{2}(\Gamma(t))}=\sup_{\boldsymbol{\psi}\in\mathbf{L}^{2}(\Gamma(t))\setminus\\{0\\}}\frac{|\mathbf{m}(P_{\mathcal{K}}\partial^{\circ}\mathcal{S}\boldsymbol{\phi},\boldsymbol{\psi})|}{\|\boldsymbol{\psi}\|_{\mathbf{L}^{2}(\Gamma(t))}},$
and from (B.5) one finds
$\mathbf{m}(P_{\mathcal{K}}\partial^{\circ}\mathcal{S}\boldsymbol{\phi},\boldsymbol{\psi})=-\sum_{i=1}^{N_{K}}\left[\mathbf{m}(\mathcal{S}\boldsymbol{\phi},\partial^{\circ}\boldsymbol{\kappa}_{i})+\mathbf{m}(\mathcal{S}\boldsymbol{\phi}HV_{N},\boldsymbol{\kappa}_{i})\right]\mathbf{m}(\boldsymbol{\psi},\boldsymbol{\kappa}_{i})-\mathbf{m}(\mathcal{S}\boldsymbol{\phi},\boldsymbol{\kappa}_{i})\mathbf{m}(\boldsymbol{\psi},\partial^{\circ}\boldsymbol{\kappa}_{i}),$
where we have used the fact that
$P_{\mathcal{K}}\mathcal{S}\boldsymbol{\phi}=0$. Combining these facts, we see
$\|P_{\mathcal{K}}\partial^{\circ}\mathcal{S}\boldsymbol{\phi}\|_{\mathbf{L}^{2}(\Gamma(t))}\leq
C\|\mathcal{S}\boldsymbol{\phi}\|_{\mathbf{L}^{2}(\Gamma(t))},$
for a constant $C$ which depends on the relevant geometric quantities. Then by
using (B.2) it follows that we have in fact that
$\displaystyle\|P_{\mathcal{K}}\partial^{\circ}\mathcal{S}\boldsymbol{\phi}\|_{\mathbf{H}^{1}(\Gamma(t))}\leq
C\|P_{\mathcal{K}}\partial^{\circ}\mathcal{S}\boldsymbol{\phi}\|_{\mathbf{L}^{2}(\Gamma(t))}\leq
C\|\mathcal{S}\boldsymbol{\phi}\|_{\mathbf{L}^{2}(\Gamma(t))},$ (B.8)
and combining (B.7), (B.8) yields the result. ∎
From this one also deduces $\mathcal{S}^{\perp}\boldsymbol{\phi}\in
H^{1}_{\mathbf{V}_{\sigma}}$ for $\boldsymbol{\phi}\in
L^{2}_{\mathbf{V}_{\sigma}}\cap H^{1}_{\mathbf{V}_{\sigma}^{\prime}}$, which
we use in our uniqueness result.
## References
* [1] Abels, H. On a diffuse interface model for two-phase flows of viscous, incompressible fluids with matched densities. Archive for rational mechanics and analysis 194, 2 (2009), 463–506.
* [2] Abels, H., Garcke, H., and Grün, G. Thermodynamically consistent, frame indifferent diffuse interface models for incompressible two-phase flows with different densities. Mathematical Models and Methods in Applied Sciences 22, 03 (2012), 1150013.
* [3] Abels, H., and Röger, M. Existence of weak solutions for a non-classical sharp interface model for a two-phase flow of viscous, incompressible fluids. Annales de l’Institut Henri Poincaré C 26, 6 (2009), 2403–2424.
* [4] Abels, H., and Wilke, M. Well-posedness and qualitative behaviour of solutions for a two-phase Navier–Stokes-Mullins–Sekerka system. Interfaces and Free Boundaries 15, 1 (2013), 39–75.
* [5] Alikakos, N. D., Bates, P. W., and Chen, X. Convergence of the cahn-hilliard equation to the Hele-Shaw model. Archive for rational mechanics and analysis 128 (1994), 165–205.
* [6] Alphonse, A., D, C., Djurdjevac, A., and Elliott, C. M. Function spaces, time derivatives and compactness for evolving families of Banach spaces with applications to PDEs. Journal of Differential Equations 353 (2023), 268–338.
* [7] Alphonse, A., Elliott, C. M., and Stinner, B. An abstract framework for parabolic PDEs on evolving spaces. Portugaliae Mathematica 71, 1 (2015), 1–46.
* [8] Anderson, D. M., McFadden, G. B., and Wheeler, A. A. Diffuse-interface methods in fluid mechanics. Annual review of fluid mechanics 30, 1 (1998), 139–165.
* [9] Antanovskii, L. K. A phase field model of capillarity. Physics of fluids 7, 4 (1995), 747–753.
* [10] Aubin, T. Nonlinear analysis on manifolds. Monge-Ampère equations, vol. 252 of Grundlehren der mathematischen Wissenschaften,. Springer-Verlag, Berlin and New York,, 1982.
* [11] Bachini, E., Krause, V., and Voigt, A. Derivation and simulation of a two-phase fluid deformable surface model. arXiv preprint arXiv:2305.15147 (2023).
* [12] Bachini, E., Krause, V., and Voigt, A. The interplay of geometry and coarsening in multicomponent lipid vesicles under the influence of hydrodynamics. Physics of Fluids 35, 4 (2023).
* [13] Barrett, J. W., Garcke, H., and Nürnberg, R. Finite element approximation for the dynamics of fluidic two-phase biomembranes. ESAIM: Mathematical Modelling and Numerical Analysis 51, 6 (2017), 2319–2366.
* [14] Bihari, I. A generalization of a lemma of Bellman and its application to uniqueness problems of differential equations. Acta Mathematica Hungarica 7, 1 (1956), 81–94.
* [15] Bonito, A., Demlow, A., and Licht, M. A divergence-conforming finite element method for the surface Stokes equation. SIAM Journal on Numerical Analysis 58, 5 (2020), 2764–2798.
* [16] Brandner, P., Reusken, A., and Schwering, P. On derivations of evolving surface Navier-Stokes equations. arXiv preprint arXiv:2110.14262 (2021).
* [17] Brezis, H., and Gallouet, T. Nonlinear Schrödinger evolution equations. Mathematics Research Center, University of Wisconsin, 1979.
* [18] Caetano, D., and Elliott, C. M. Cahn–Hilliard equations on an evolving surface. European Journal of Applied Mathematics 32, 5 (2021), 937–1000.
* [19] Caetano, D., Elliott, C. M., Grasselli, M., and Poiatti, A. Regularization and separation for evolving surface Cahn–Hilliard equations. SIAM Journal on Mathematical Analysis 55, 6 (2023), 6625–6675.
* [20] Chow, B., and Knopf, D. The Ricci Flow: An Introduction, vol. 1. American Mathematical Soc., 2004.
* [21] Ciarlet, P. G. On Korn’s inequality. Chinese Annals of Mathematics, Series B 31 (2010), 607–618.
* [22] Deckelnick, K., Dziuk, G., and Elliott, C. M. Computation of geometric partial differential equations and mean curvature flow. Acta Numerica 14 (2005), 139–232.
* [23] Deserno, M. Fluid lipid membranes: From differential geometry to curvature stresses. Chemistry and physics of lipids 185 (2015), 11–45.
* [24] Djurdjevac, A., Gräser, C., and Herbert, P. J. An evolving space framework for Oseen equations on a moving domain. ESAIM: Mathematical Modelling and Numerical Analysis 57, 5 (2023), 3113–3138.
* [25] Elliott, C., and Luckhaus, S. A generalized equation for phase separation of a multi-component mixture with interfacial free energy. preprint SFB 256 (1991), 195.
* [26] Elliott, C. M., and Ranner, T. Evolving surface finite element method for the Cahn–Hilliard equation. Numerische Mathematik 129, 3 (2015), 483–534.
* [27] Evans, L. C., and Gariepy, R. F. Measure Theory and Fine Properties of Functions. Chapman and Hall/CRC, 2015.
* [28] Fan, J., Han, T., and Haataja, M. Hydrodynamic effects on spinodal decomposition kinetics in planar lipid bilayer membranes. The Journal of chemical physics 133, 23 (2010).
* [29] Giorgini, A., Miranville, A., and Temam, R. Uniqueness and regularity for the Navier–Stokes–Cahn–Hilliard system. SIAM Journal on Mathematical Analysis 51, 3 (2019), 2535–2574.
* [30] Gorka, P. Brézis-Wainger inequality on Riemannian manifolds. Journal of Inequalities and Applications 2008 (2008), 1–6.
* [31] Gurtin, M. E., Polignone, D., and Vinals, J. Two-phase binary fluids and immiscible fluids described by an order parameter. Mathematical Models and Methods in Applied Sciences 6, 06 (1996), 815–831.
* [32] Hatcher, L. Phase field models for small deformations of biomembranes arising as Helfrich energy equilibria. PhD thesis, University of Warwick, 2020.
* [33] Hebey, E. Nonlinear Analysis on Manifolds: Sobolev Spaces and Inequalities: Sobolev Spaces and Inequalities, vol. 5. American Mathematical Soc., 2000.
* [34] Jankuhn, T., Olshanskii, M. A., and Reusken, A. Incompressible fluid problems on embedded surfaces: modeling and variational formulations. Interfaces and Free Boundaries 20, 3 (2018), 353–377.
* [35] Jost, J. Riemannian geometry and geometric analysis, vol. 42005. Springer, 2008.
* [36] Jülicher, F., and Lipowsky, R. Shape transformations of vesicles with intramembrane domains. Physical Review E 53, 3 (1996), 2670.
* [37] Lam, K. F., and Wu, H. Thermodynamically consistent Navier–Stokes–Cahn–Hilliard models with mass transfer and chemotaxis. European Journal of Applied Mathematics 29, 4 (2018), 595–644.
* [38] Li, J., and Titi, E. S. A tropical atmosphere model with moisture: global well-posedness and relaxation limit. Nonlinearity 29, 9 (2016), 2674.
* [39] McMahon, H. T., and Gallop, J. L. Membrane curvature and mechanisms of dynamic cell membrane remodelling. Nature 438, 7068 (2005), 590–596.
* [40] Miura, T.-H. On singular limit equations for incompressible fluids in moving thin domains. arXiv preprint arXiv:1703.09698 (2017).
* [41] Miura, T.-H. Zero width limit of the heat equation on moving thin domains. Interfaces and Free Boundaries 19, 1 (2017), 31–77.
* [42] Olshanskii, M., Palzhanov, Y., and Quaini, A. A comparison of Cahn–Hilliard and Navier–Stokes–Cahn–Hilliard models on manifolds. Vietnam Journal of Mathematics 50, 4 (2022), 929–945.
* [43] Olshanskii, M. A., Reusken, A., and Schwering, P. An eulerian finite element method for tangential Navier-Stokes equations on evolving surfaces. arXiv preprint arXiv:2302.00779 (2023).
* [44] Olshanskii, M. A., Reusken, A., and Zhiliakov, A. Tangential Navier-Stokes equations on evolving surfaces: Analysis and simulations. Mathematical Models and Methods in Applied Sciences 14 (2022), 2817–2852.
* [45] Olshanskii, M. A., Reusken, A., and Zhiliakov, A. Tangential Navier-Stokes equations on evolving surfaces: Analysis and simulations. arXiv preprint arXiv:2203.01521 (2022).
* [46] Palzhanov, Y., Zhiliakov, A., Quaini, A., and Olshanskii, M. A decoupled, stable, and linear fem for a phase-field model of variable density two-phase incompressible surface flow. Computer Methods in Applied Mechanics and Engineering 387 (2021), 114167.
* [47] Sakai, T. Riemannian geometry, vol. 149. American Mathematical Soc., 1996.
* [48] Styles, V., Kay, D., and Welford, R. Finite element approximation of a Cahn–Hilliard–Navier–Stokes system. Interfaces and free Boundaries 10, 1 (2008), 15–43.
* [49] Sun, M., Xiao, X., Feng, X., and Wang, K. Modeling and numerical simulation of surfactant systems with incompressible fluid flows on surfaces. Computer Methods in Applied Mechanics and Engineering 390 (2022), 114450.
* [50] Temam, R. Navier-Stokes equations: theory and numerical analysis, vol. 343. American Mathematical Soc., 2001.
* [51] Yang, J., and Kim, J. A phase-field model and its efficient numerical method for two-phase flows on arbitrarily curved surfaces in 3d space. Computer Methods in Applied Mechanics and Engineering 372 (2020), 113382.
* [52] Yosida, K. Functional analysis. Springer Science & Business Media, 2012.
|
from the instances fenerated from SYNTCOMP benchmark suite are listed in Table
LABEL:table:FPvsOddfairzlk. On the Odd-fair instances with $50\%-$liveness
generated from the SYNTCOMP benchmark suite, there are 204 instances where
neither of the algorithms OF-FP, OF-ZL, N-FP or N-ZL timed out. On these
instances, OF-ZL gives an average computation time of $4.6$ seconds while OF-
FP took $122.7$ seconds on average. On the same examples, N-ZL takes on
average $3.6$ seconds to compute while N-FP gives an average of $45.2$
seconds. For the PGSolver dataset OF-FP timed out on all generated instances,
whereas OF-ZL took $24.9$ seconds on average to terminate.
#### A.4.2 Sensitivity
To monitor the sensitivity of OF-ZL to the change in number of priorities as
well as the percentage of live edges in the game, we picked $12$ parity game
instances from the SYNTCOMP dataset which did not timeout (after one hour).
With priorities $3-4-5-6$ and liveness degrees 0$\%$777regular parity
game-30$\%$-50$\%$-80$\%$ we get 192 different Odd-fair parity instances. Fig.
5 shows the runtime of OF-ZL on these instances.
We can see that the runtimes of instances with different priority and liveness
percentages are distributed in a seemingly random manner. This tells us that
Odd-fair Zielonka’s algorithm is highly insensitive to a change in the
percentage of live edges and the number of priorities. This observation is
inline with the known insensitivity of Zielonka’s algorithm for the number of
priorities.
Figure 5: Runtime of OF-ZL on the 192 Odd-fair parity instances generated from
12 fixed parity examples through changing their priorities and liveness
degrees. Different shapes indicate the number of prioirities an instance has,
and the $x-$axis denotes their liveness percentages. At each coloumn we view
48 different instances of the 12 examples with varying colours.
#### A.4.3 Comparative Evaluation
In order to validate the computational advantage of OF-ZL over OF-FP, we have
run both algorithms on all 50$\%$-liveness instances generated from the
SYNTCOMP benchmark dataset. On 58 of these instances, both algorithms time
out. The run-times for all other instances are depicted in Fig. 6 (right), 7
(right) and 8 (right). The left plots in Fig. 6-8 show the same comparison for
the ?normal? parity algorithms N-ZL and N-FP. In both cases, Fig. 7 shows the
zoomed-in version of the respective plot in Fig. 6. Fig. 8 shows the data-
points from the respective plot in Fig. 7 as a scatter plot in log-scale. The
examples on which only x-FP times out, can be seen as the dots on the ceiling
of the plots in Fig. 6. In all plots, points above the diagonal correspond to
instances where Zielonka’s algorithm outperforms the fixed-point algorithm.
We clearly see in Fig. 6-8 that Zielonka’s algorithm performs significantly
better than the fixed-point version, both in the Odd-fair (right) and in the
normal (left) case. More importantly, the overall performance comparison
between OF-ZL over OF-FP (right plots) mimics the comparison between N-ZL over
N-FP. This allows us to conclude that our new Odd-fair Zielonka’s algorithm
retains the computational advantages of Zielonka’s algorithm.
In addition, Table LABEL:table:FPvsOddfairzlk shows that OF-ZL results in
almost the same run-time as N-ZL, showing that our changes in the algorithm
incur almost no computational disadvantages over the original algorithm. This
allows us to handle transition fairness for almost free in practice.
Figure 6: (Zoomed out version) A comparison of N-FP vs. N-ZL in regular parity
games (left), and OF-FP vs. OF-ZL on fair parity games (right)
Figure 7: (Zoomed in version) A comparison of N-FP vs. N-ZL in regular parity
games (left), and OF-FP vs. OF-ZL on fair parity games (right)
Figure 8: A comparison of N-FP vs. N-ZL in regular parity games (left), and
OF-FP vs. OF-ZL on fair parity games (right) in terms of log-scale plots where
the timeouts are removed.
Figure 9: A comparison of N-ZL vs. OF-ZL over examples that do not timeout on
both. Right hand side plot visualizes the same data in logscale.
Conclusion: The results show that Zielonka’s algorithm is significantly faster
in solving Odd-fair parity games compared to the calculation performed by the
fixed-point algorithm, as is the case in normal parity games. The fixed-point
algorithm started timing out as soon as the examples became more complex,
being especially sensitive to the increase in the number of priorities.
Whereas, Zielonka’s algorithm preserves its performance considerably in the
face of the increase in the same parameters. These outcomes match the known
comparison results between the naive fixed-point calculation versus Zielonka’s
algorithm, on normal parity games.
Table 1: Detailed run-time comparison of N-FP and N-ZL on the original parity game instances (yellow rows) with OF-FP and OF-ZL on their respective $30\%$\- and $50\%$-liveness Odd-fair parity game instances (white rows). The instance name is taken from the original benchmark suite. Name | $\\#$ | $\\#$ | $\\#$ | FP | ZL
---|---|---|---|---|---
| nodes | edges | priorities | (sec.) | (sec.)
EscalatorCountingInit | 99 | 148 | 3 | 0.064 | 0.012
$30\%$-EscalatorCountingInit | 99 | 148 | 3 | 0.075 | 0.018
$50\%$-EscalatorCountingInit | 99 | 148 | 3 | 0.072 | 0.02
KitchenTimerV1 | 80 | 124 | 3 | 0.055 | 0.008
$30\%$-KitchenTimerV1 | 80 | 124 | 3 | 0.068 | 0.012
$50\%$-KitchenTimerV1 | 80 | 124 | 3 | 0.21 | 0.009
KitchenTimerV6 | 4099 | 6560 | 3 | 87 | 11
$30\%$-KitchenTimerV6 | 4099 | 6560 | 3 | 88 | 11
$50\%$-KitchenTimerV6 | 4099 | 6560 | 3 | 352 | 18
MusicAppSimple | 344 | 562 | 3 | 0.488 | 0.073
$30\%$-MusicAppSimple | 344 | 562 | 3 | 0.496 | 0.082
$50\%$-MusicAppSimple | 344 | 562 | 3 | 0.799 | 0.089
TwoCountersRefinedRefined | 1933 | 3140 | 3 | 14.9 | 2.5
$30\%$-TwoCountersRefinedRefined | 1933 | 3140 | 3 | 15 | 1.2
$50\%$-TwoCountersRefinedRefined | 1933 | 3140 | 3 | 74 | 3.72
Zoo5 | 479 | 768 | 3 | 0.96 | 0.135
$30\%$-Zoo5 | 479 | 768 | 3 | 0.981 | 0.152
$50\%$-Zoo5 | 479 | 768 | 3 | 1.57 | 0.172
amba$\\_$decomposed$\\_$lock$\\_$3 | 1558 | 2336 | 3 | 72 | 1.5
$30\%$-amba$\\_$decomposed$\\_$lock$\\_$3 | 1558 | 2336 | 3 | 73 | 1.5
$50\%$-amba$\\_$decomposed$\\_$lock$\\_$3 | 1558 | 2336 | 3 | 56 | 2.9
full$\\_$arbiter$\\_$2 | 204 | 324 | 3 | 0.59 | 0.049
$30\%$-full$\\_$arbiter$\\_$2 | 204 | 324 | 3 | 0.602 | 0.047
$50\%$-full$\\_$arbiter$\\_$2 | 204 | 324 | 3 | 5 | 0.059
full$\\_$arbiter$\\_$3 | 1403 | 2396 | 3 | 21.18 | 2
$30\%$-full$\\_$arbiter$\\_$3 | 1403 | 2396 | 3 | 21.5 | 2
$50\%$-full$\\_$arbiter$\\_$3 | 1403 | 2396 | 3 | 93 | 3.46
lilydemo06 | 369 | 548 | 3 | 8.1 | 0.18
$30\%$-lilydemo06 | 369 | 548 | 3 | 8.13 | 0.206
$50\%$-lilydemo06 | 369 | 548 | 3 | 18 | 0.212
lilydemo07 | 78 | 108 | 3 | 0.27 | 0.01
$30\%$-lilydemo07 | 78 | 108 | 3 | 0.284 | 0.017
$50\%$-lilydemo07 | 78 | 108 | 3 | 0.33 | 0.008
simple$\\_$arbiter$\\_$unreal1 | 2178 | 3676 | 3 | 22.8 | 3
$30\%$-simple$\\_$arbiter$\\_$unreal1 | 2178 | 3676 | 3 | 23 | 3
$50\%$-simple$\\_$arbiter$\\_$unreal1 | 2178 | 3676 | 3 | 254 | 7
amba$\\_$decomposed$\\_$arbiter$\\_$2 | 141 | 212 | 4 | 0.72 | 0.03
$30\%$-amba$\\_$decomposed$\\_$arbiter$\\_$2 | 141 | 212 | 4 | 0.73 | 0.06
$50\%$-amba$\\_$decomposed$\\_$arbiter$\\_$2 | 141 | 212 | 4 | 1 | 0.035
loadfull3 | 1159 | 2030 | 4 | 5.62 | 0.609
$30\%$-loadfull3 | 1159 | 2030 | 4 | 5 | 0.614
$50\%$-loadfull3 | 1159 | 2030 | 4 | 5 | 0.754
ltl2dba01 | 101 | 152 | 4 | 0.074 | 0.031
$30\%$-ltl2dba01 | 101 | 152 | 4 | 0.075 | 0.030
$50\%$-ltl2dba01 | 101 | 152 | 4 | 1.4 | 0.028
ltl2dba14 | 97 | 144 | 4 | 0.18 | 0.016
$30\%$-ltl2dba14 | 97 | 144 | 4 | 0.181 | 0.013
$50\%$-ltl2dba14 | 97 | 144 | 4 | 0.574 | 0.012
ltl2dba22 | 21 | 30 | 4 | 0.037 | 0.002
$30\%$-ltl2dba22 | 21 | 30 | 4 | 0.036 | 0.002
$50\%$-ltl2dba22 | 21 | 30 | 4 | 0.03 | 0.0009
prioritized$\\_$arbiter$\\_$unreal2 | 851 | 1412 | 4 | 15.8 | 0.73
$30\%$-prioritized$\\_$arbiter$\\_$unreal2 | 851 | 1412 | 4 | 16 | 0.759
$50\%$-prioritized$\\_$arbiter$\\_$unreal2 | 851 | 1412 | 4 | 126 | 1.2
lilydemo17 | 3102 | 5334 | 7 | 1237 | 41
$30\%$-lilydemo17 | 3102 | 5334 | 7 | Timeout | 41
$50\%$-lilydemo17 | 3102 | 5334 | 7 | Timeout | 24
lilydemo18 | 449 | 728 | 9 | 220 | 0.6
$30\%$-lilydemo18 | 449 | 728 | 9 | 224 | 0.621
$50\%$-lilydemo18 | 449 | 728 | 9 | Timeout | 0.552
### A.5 Additional material for Ex. 4.3
Below we present an extended version of the fixed-point calculation in (8),
$\displaystyle Y_{4}^{0}=\emptyset$ $\displaystyle\quad X_{3}^{0,0}=V$
$\displaystyle\quad\quad Y_{2}^{0,0,0}=\emptyset$
$\displaystyle\quad\quad\quad X_{1}^{0,0,0,0}=V$ $\displaystyle\quad\quad\quad
X_{1}^{0,0,0,1}=\Phi^{Y_{4}^{0},X_{3}^{0,0},Y_{2}^{0,0,0},X_{1}^{0,0,0,0}}=C_{3}\cup
C_{1}$ $\displaystyle\quad\quad\quad
X_{1}^{0,0,0,2}=\Phi^{Y_{4}^{0},X_{3}^{0,0},Y_{2}^{0,0,0},X_{1}^{0,0,0,1}}=C_{3}\cup(C_{1}\cap\mathsf{Npre}(Y_{2}^{0,0,0},X_{1}^{0,0,0,1}))=C_{3}$
$\displaystyle\quad\quad\quad
X_{1}^{0,0,0,3}=\Phi^{Y_{4}^{0},X_{3}^{0,0},Y_{2}^{0,0,0},X_{1}^{0,0,0,1}}=C_{3}\cup(C_{1}\cap\mathsf{Npre}(Y_{2}^{0,0,0},X_{1}^{0,0,0,2}))=C_{3}$
$\displaystyle\quad\quad Y_{2}^{0,0,1}=X_{1}^{0,0,0,\infty}=C_{3}$
$\displaystyle\quad\quad\quad X_{1}^{0,0,1,0}=V$ $\displaystyle\quad\quad\quad
X_{1}^{0,0,1,1}=\Phi^{Y_{4}^{0},X_{3}^{0,0},Y_{2}^{0,0,1},X_{1}^{0,0,0,0}}=C_{3}\cup
C_{1}\cup\\{2b\\}$ $\displaystyle\quad\quad\quad
X_{1}^{0,0,1,2}=\Phi^{Y_{4}^{0},X_{3}^{0,0},Y_{2}^{0,0,1},X_{1}^{0,0,0,1}}=C_{3}\cup\\{2b\\}$
$\displaystyle\quad\quad\quad
X_{1}^{0,0,1,3}=\Phi^{Y_{4}^{0},X_{3}^{0,0},Y_{2}^{0,0,1},X_{1}^{0,0,0,2}}=C_{3}\cup\\{2b\\}$
$\displaystyle\quad\quad Y_{2}^{0,0,2}=X_{1}^{0,0,1,\infty}=C_{3}\cup\\{2b\\}$
$\displaystyle\quad\quad\quad X_{1}^{0,0,2,0}=V$ $\displaystyle\quad\quad\quad
X_{1}^{0,0,2,1}=\Phi^{Y_{4}^{0},X_{3}^{0,0},Y_{2}^{0,0,2},X_{1}^{0,0,0,0}}=C_{3}\cup
C_{1}\cup\\{2b,2c\\}$ $\displaystyle\quad\quad\quad
X_{1}^{0,0,2,2}=\Phi^{Y_{4}^{0},X_{3}^{0,0},Y_{2}^{0,0,2},X_{1}^{0,0,0,1}}=C_{3}\cup\\{2b,2c\\}$
$\displaystyle\quad\quad\quad
X_{1}^{0,0,2,3}=\Phi^{Y_{4}^{0},X_{3}^{0,0},Y_{2}^{0,0,2},X_{1}^{0,0,0,2}}=C_{3}\cup\\{2b,2c\\}$
$\displaystyle\quad\quad
Y_{2}^{0,0,3}=X_{1}^{0,0,2,\infty}=C_{3}\cup\\{2b,2c\\}$
$\displaystyle\quad\quad\ldots$ $\displaystyle\quad\quad
Y_{2}^{0,0,4}=X_{1}^{0,0,3,\infty}=C_{3}\cup\\{2b,2c\\}$ $\displaystyle\quad
X_{3}^{0,1}=Y_{2}^{0,0,\infty}=C_{3}\cup\\{2b,2c\\}$ $\displaystyle\quad\quad
Y_{2}^{0,1,0}=\emptyset\quad$ $\displaystyle\quad\quad
Y_{2}^{0,1,1}=X_{1}^{0,1,0,\infty}=\\{3b\\}\quad$ $\displaystyle\quad\quad
Y_{2}^{0,1,2}=X_{1}^{0,1,1,\infty}=\\{2b,3b\\}\ \quad$
$\displaystyle\quad\quad
Y_{2}^{0,1,3}=Y_{2}^{0,1,4}=X_{1}^{0,1,2,\infty}=X_{1}^{0,1,3,\infty}=\\{2b,2c,3b\\}$
$\displaystyle\quad X_{3}^{0,2}=Y_{2}^{0,1,\infty}=\\{2b,2c,3b\\}$
$\displaystyle\quad\ldots$ $\displaystyle\quad
X_{3}^{0,3}=Y_{2}^{0,2,\infty}=\\{2b,2c,3b\\}$ $\displaystyle
Y_{4}^{1}=X_{3}^{0,\infty}=\\{2b,2c,3b\\}$ $\displaystyle\quad X_{3}^{1,0}=V$
$\displaystyle\quad\quad Y_{2}^{1,0,0}=\emptyset$ $\displaystyle\quad\quad
Y_{2}^{1,0,1}=X_{1}^{1,0,0,\infty}=C_{3}\cup C_{4}$ $\displaystyle\quad\quad
Y_{2}^{1,0,2}=X_{1}^{1,0,1,\infty}=C_{3}\cup C_{4}\cup\\{2b\\}$
$\displaystyle\quad\quad Y_{2}^{1,0,4}=Y_{2}^{1,0,3}=C_{1}\cup C_{3}\cup
C_{4}\cup\\{2b,2c\\}$ $\displaystyle\quad
X_{3}^{1,1}=Y_{2}^{1,0,\infty}=C_{1}\cup C_{3}\cup C_{4}\cup\\{2b,2c\\}$
$\displaystyle\quad\quad Y_{2}^{1,1,0}=\emptyset$ $\displaystyle\quad\quad
Y_{2}^{1,1,1}=C_{3}\cup C_{4}$ $\displaystyle\quad\quad
Y_{2}^{1,1,2}=C_{3}\cup C_{4}\cup\\{2b\\}$ $\displaystyle\quad\quad
Y_{2}^{1,1,3}=Y_{2}^{1,1,4}=C_{1}\cup C_{3}\cup C_{4}\cup\\{2b,2c\\}$
$\displaystyle\quad X_{3}^{1,2}=Y_{2}^{1,1,\infty}=C_{1}\cup C_{3}\cup
C_{4}\cup\\{2b,2c\\}$ $\displaystyle Y_{4}^{2}=X_{3}^{1,\infty}=C_{1}\cup
C_{3}\cup C_{4}\cup\\{2b,2c\\}$ $\displaystyle\ldots$ $\displaystyle
Y_{4}^{3}=C_{1}\cup C_{3}\cup C_{4}\cup\\{2b,2c\\}$
And finally,
$\mathcal{W}_{Odd}=Y_{4}^{\infty}=C_{1}\cup C_{3}\cup
C_{4}\cup\\{2b,2c\\}=V\setminus\\{2a\\}$
|
1,2,3,4]Salvatore Capozziello<EMAIL_ADDRESS>5,2]Maurizio Capriolo<EMAIL_ADDRESS>5,6]Gaetano Lambiase<EMAIL_ADDRESS>
[1]Dipartimento di Fisica "E. Pancini", Università di Napoli “Federico II”, Compl. Univ. di Monte S. Angelo, Edificio G, Via Cinthia, I-80126, Napoli, Italy
[2]Istituto Nazionale di Fisica Nucleare, Sezione di Napoli, Compl. Univ. di Monte S. Angelo, Edificio G, Via Cinthia, I-80126, Napoli, Italy
[3]Scuola Superiore Meridionale, Largo S. Marcellino 10, I-80138, Napoli, Italy
[4]Department of Mathematics, Faculty of Civil Engineering,VSB-Technical University of Ostrava, Ludvika Podeste 1875/17, 708 00 Ostrava-Poruba,
Czech Republic
[5]Dipartimento di Fisica "E. R. Caianiello", Università degli Studi di Salerno, via Giovanni Paolo II, 132 I-84084 Fisciano, Salerno, Italy
[6]Istituto Nazionale di Fisica Nucleare, Sezione di Napoli, Gruppo Collegato di Salerno, via Giovanni Paolo II, 132 I-84084 Fisciano, Salerno, Italy
Energy-Momentum Complex in Higher Order Curvature-Based Local Gravity
An unambiguous definition of gravitational energy remains one of the unresolved issues of physics today. This problem is related to the non-localization of gravitational energy density. In General Relativity, there have been many proposals for defining the gravitational energy density, notably those proposed by Einstein, Tolman, Landau and Lifshitz, Papapetrou, Møller, and Weinberg. In this review, we firstly explored the energy–momentum complex in an $n^{th}$ order gravitational Lagrangian $L=L\left(g_{\mu\nu}, g_{\mu\nu,i_{1}}, g_{\mu\nu,i_{1}i_{2}},g_{\mu\nu,i_{1}i_{2}i_{3}},\cdots, g_{\mu\nu,i_{1}i_{2}i_{3}\cdots i_{n}}\right)$ and then in a gravitational Lagrangian as . Its gravitational part was obtained by invariance of gravitational action under infinitesimal rigid translations using Noether's theorem. We also showed that this tensor, in general, is not a covariant object but only an affine object, that is, a pseudo-tensor. Therefore, the pseudo-tensor $\tau^{\eta}_{\alpha}$ becomes the one introduced by Einstein if we limit ourselves to General Relativity and its extended corrections have been explicitly indicated. The same method was used to derive the energy–momentum complex in $ f\left (R \right) $ gravity both in Palatini and metric approaches. Moreover, in the weak field approximation the pseudo-tensor $\tau^{\eta}_{\alpha}$ to lowest order in the metric perturbation $h$ was calculated. As a practical application, the power per unit solid angle $\Omega$ emitted by a localized source carried by a gravitational wave in a direction $\hat{x}$ for a fixed wave number $\mathbf{k}$ under a suitable gauge was obtained, through the average value of the pseudo-tensor over a suitable spacetime domain and the local conservation of the pseudo-tensor. As a cosmological application, in a flat Friedmann–Lemaître–Robertson–Walker spacetime, the gravitational and matter energy density in $f(R)$ gravity both in Palatini and metric formalism was proposed. The gravitational energy–momentum pseudo-tensor could be a useful tool to investigate further modes of gravitational radiation beyond two standard modes required by General Relativity and to deal with non-local theories of gravity involving $\Box^{-k}$ terms.
Keywords: Energy–Momentum Complex; Pseudo-Tensor; Gravitational Energy
§ INTRODUCTION
A widely accepted definition of gravitational energy density and its localization in curved spacetime are serious problems that afflict the general relativity. Several prescriptions for gravitational contribution to energy–momentum density and more generally for energy–momentum complex have been suggested by Einstein, Tolman, Landau and Lifshitz, Papapetrou, Møller, and Weinberg [8, 9, 10, 11, 12, 13, 14, 15, 16, 17]. These attempts are based on the introduction of a super-potential or through the expansion of the Ricci tensor in the metric perturbation $h$. Thus, the gravitational part of the energy–momentum density transforms as an affine tensor not as a covariant tensor, and for this reason, it is not really a tensor but a pseudo-tensor. This affine property of the gravitational stress–energy tensor
makes the gravitational energy–momentum density not localizable. However, integrating the density over a suitable spatial region at a certain time such as over an asymptotically flat spacetime, viable for isolated systems, the gravitational energy–momentum becomes a four-vector, as meaning that changes in right way under asymptotically flat coordinate transformations. Over all space it becomes quasi independent of the coordinate system, that is, the gravitational energy–momentum of the spacetime exists, but it cannot be localized. In this review a generalization of Einstein's pseudo-tensor to Extended Theories of Gravity [18, 19] is proposed by imposing the invariance of the higher order gravitational Lagrangian under an infinitesimal rigid translation and by using Noether's theorem. Then, thanks to a continuity equation, a Noether current and a Noether charge were derived that correspond to a gravitational energy–momentum pseudo-tensor and gravitational energy–momentum, respectively, both locally conserved.
By weakly perturbing the metric tensor around the Minkowskian metric, a weak-field limit, in a suitable gauge, the gravitational energy–momentum pseudo-tensor for a Lagrangian of $n^{th}$ order appears an object easier to handle. Then, by averaging of the pseudo-tensor over a suitable spacetime domain, it is possible to calculate the power emitted by some localized astrophysical source carried away by the gravitational waves. This approach could be relevant for searching for polarization states of gravitational waves in addition to the two standards of general relativity [20, 21]. Finally, after deriving the gravitational energy–momentum pseudo-tensor in $F(R)$ gravity formulated in Palatini and metric formalism, some cosmological applications were discussed, wherein a flat FLRW metric the total energy density was obtained in both approaches [22, 23].
For more details on the issue of energy–momentum localization in modified theories of gravity such as $f(R)$, ${f(R,\Box R,\dots, \Box^{k} R)}$ [23, 24], teleparallel gravity and its extended version $f(T)$, see Ref. [25]. Meanwhile, for a study of wavelike solutions of modified teleparallel gravity necessary for future applications of the pseudo-tensor, see references [29, 30].
The review is organized as follows. Firstly in Sec. <ref> some definitions of gravitational pseudo-tensors in general relativity are listed. In Sec. <ref> we derived the gravitational energy–momentum pseudo-tensor for a general Lagrangian of $n^{th}$ order through two procedures: the first method uses a variational principle under rigid transformations via Noether's theorem and the second adopts the Landau–Lifshitz procedure [8] without the introduction of the super-potential. Hence, in Sec. <ref>, we proved that a stress–energy object is transformed in the correct manner under linear transformations but not under diffeomorphisms and, therefore, is a pseudo-tensor and not a covariant tensor. In Sec. <ref>, we calculated the Euler–Lagrange equations and the gravitational energy–momentum pseudo-tensor for $f(R)$ gravity, always using Noether's theorem applied to a particular one-parameter group of diffeomorphisms given by rigid translations. Therefore, in all models of gravity we obtained the continuity equation for an energy–momentum complex. In Sec. <ref>, we derived the gravitational energy–momentum pseudo-tensor of a gravitation field for a particular Lagrangian $L_{g}=(\overline{R}+a_{0}R^{2}+\sum_{k=1}^{p} a_{k}R\Box^{k}R)\sqrt{-g}$. Sec. <ref>, is devoted to the weak-field limit of the gravitational stress–energy pseudo-tensor expanded to lowest order in a small perturbation $h$, i.e., up to $h^2$ order, and we have shown two simple cases where the index $p$ is equal to zero and one. Hence, in Sec. <ref>, we averaged the pseudo-tensor over an suitable region containing the isolated sources and then we found the emitted power carried by the gravitational radiation. Afterward, in Sec. <ref>, in Palatini $f({\cal R})$ gravity, related field equations and related gravitational energy–momentum pseudo-tensor were obtained. Therefore in Sec. <ref>, by adopting a flat FLRW spacetime, an explicit calculus of an energy density complex for power law cosmological solutions was performed, also in the metric formalism of $f(R)$. Conclusions are summarized in Sec. <ref>. Finally in Appendix <ref>, we proved that the additive terms related to the symmetries of $g_{\mu\nu}$ and its derivatives yield a mean of zero, i.e., $\langle\left(A_{p}\right)_{\alpha}^{\eta}\rangle=\langle\left(B_{p}\right)_{\alpha}^{\eta}\rangle=0$. While in Appendix <ref>, we explicitly showed the six polarization tensors associated with the gravitational waves present in higher-order theories.
§ SEVERAL DEFINITIONS OF GRAVITATIONAL ENERGY–MOMENTUM PSEUDO-TENSOR IN GENERAL RELATIVITY
Here are some of the most important definitions of gravitational energy–momentum pseudo-tensor in general relativity in the scientific literature, for details see [31].
§.§ Einstein energy–momentum complex
In special relativity the law of conservation of energy and momentum is given by
\begin{equation}\label{conserveinongrav}
\frac{\partial T^{\mu\nu}}{\partial x^{\mu}}=0\ ,
\end{equation}
with $T^{\mu\nu}$ the energy–momentum tensor of matter and non-gravitational fields. In general relativity this principle becomes for general covariance
\begin{equation}\label{conserveigrav}
\nabla_{\mu}T^{\mu\nu}=0\ ,
\end{equation}
which does not correspond to any law of conservation of physical quantities. Einstein therefore formulated the conservation law in the following way
\begin{equation}\label{conserveigravespl}
\frac{\partial \theta_{\mu}^{\phantom{\mu}\nu}}{\partial x^{\nu}}=\frac{\partial}{\partial x^{\nu}}\left(\sqrt{-g}\left(T_{\mu}^{\phantom{\mu}\nu}+t_{\mu}^{\phantom{\mu}\nu}\right)\right)=0\ ,
\end{equation}
where $t_{\mu}^{\phantom{\mu}\nu}$ is an pseudo-tensor. So what is conserved is not only the tensor of non-gravitational fields and matter $T_{\mu}^{\phantom{\mu}\nu}$ but a pseudo-tensor $t_{\mu}^{\phantom{\mu}\nu}$ must be added to it. This pseudo-tensor added can be interpreted as associated with the gravitational field and the energy due to the sum of the contributions of the gravitational fields plus those due to the matter is conserved. However, the pseudo-tensoriality behaviour of $t_{\mu}^{\phantom{\mu}\nu}$ makes it dependent on coordinates and the gravitational energy becomes non localizable. In order to write the Eq. (<ref>) in the form of an ordinary divergence equation Eq. (<ref>), Einstein starting from the following Lagrangian density which is a non-covariant scalar density
\begin{equation}\label{LagrEinst}
L=\sqrt{-g} g^{\mu\nu}\left(\Gamma^{\sigma}_{\mu\nu}\Gamma^{\rho}_{\sigma\rho}-\Gamma^{\sigma}_{\mu\rho}\Gamma^{\rho}_{\nu\sigma}\right)\ ,
\end{equation}
introduced a pseudo-tensor defined by the relation
\begin{equation}
\sqrt{-g}t_{\mu}^{\phantom{\mu}\nu}=\frac{1}{16\pi}\left(\delta^{\nu}_{\mu}L-\frac{\partial L}{\partial g^{\rho\sigma}_{\phantom{\rho\sigma},\nu}}g^{\rho\sigma}_{\phantom{\rho\sigma},\mu}\right)\ .
\end{equation}
§.§ Landau–Lifshitz energy–momentum pseudo-tensor
The gravitational energy–momentum pseudo-tensor defined by Landau–Lifshitz has the great advantage of being symmetric unlike Einstein's, which in general is not.This allows defining the angular momentum and therefore the related conservation law. We adopt a system of geodetic coordinates where the first derivatives of the metric tensor $g^{\mu\nu}$ vanish. Then, the Eq. (<ref>) is reduced to (<ref>) which can be written in terms of the following antisymmetric quantity in the last two indices $\eta^{\mu\nu\sigma}=-\eta^{\mu\sigma\nu}$
\begin{equation}
T^{\mu\nu}=\frac{\partial \eta^{\mu\nu\sigma}}{\partial x^{\sigma}}\ .
\end{equation}
Since the Levi–Civita connection $\Gamma$ vanishes at one point, in such coordinate system it is possible using Einstein’s equations in the presence of matter written in such coordinates, to express the stress–energy tensor of matter $T^{\mu\nu}$ as
\begin{equation}
T^{\mu\nu}=\frac{1}{\left(-g\right)}\frac{\partial}{\partial x^{\sigma}}\left\{\frac{1}{16\pi}\frac{\partial}{\partial x^{\rho}}\left[\left(-g\right)\left(g^{\mu\nu}g^{\sigma\rho}-g^{\mu\sigma}g^{\nu\rho}\right)\right]\right\}\ ,
\end{equation}
where indicating the term in braces with the antisymmetric quantity in the last two indices
$h^{\mu\nu\sigma}=-h^{\mu\sigma\nu}$, we get
\begin{equation}
\frac{\partial h^{\mu\nu\sigma}}{\partial x^{\sigma}}-\left(-g\right)T^{\mu\nu}=0\ .
\end{equation}
Returning to an arbitrary coordinate system the previous difference does not cancel anymore so we can indicate it with $\left(-g\right)t^{\mu\nu}$ or
\begin{equation}\label{LLECCG}
\left(-g\right)\left(T^{\mu\nu}+t^{\mu\nu}\right)=\frac{\partial h^{\mu\nu\sigma}}{\partial x^{\sigma}}\ .
\end{equation}
Quantities $t^{\mu\nu}$ are symmetric but are not the components of a covariant tensor but affine. Using Einstein’s field equations again it is possible from Eq. (<ref>) to get an explicit expression of $t^{\mu\nu}$, defined as the energy–momentum pseudo-tensor of the gravitational field, by means of the derivatives of the components of the metric tensor, that is
\begin{equation}\label{LLEMPT}
\begin{split}
16\pi\left(-g\right)t^{\mu\nu}=&\mathfrak{g}^{\mu\nu}_{\phantom{\mu\nu},\rho}\mathfrak{g}^{\rho\sigma}_{\phantom{\rho\sigma},\sigma}-\mathfrak{g}^{\mu\rho}_{\phantom{\mu\rho},\rho}\mathfrak{g}^{\nu\sigma}_{\phantom{\nu\sigma},\sigma}+\frac{1}{2}g^{\mu\nu}g_{\rho\sigma}\mathfrak{g}^{\rho\alpha}_{\phantom{\rho n},\beta}\mathfrak{g}^{\beta\sigma}_{\phantom{\beta\sigma},\alpha}\\
&-\left(g^{\mu\rho}g_{\sigma\alpha}\mathfrak{g}^{\nu\alpha}_{\phantom{\nu n},\beta}\mathfrak{g}^{\sigma\beta}_{\phantom{\sigma\beta},\rho}+g^{\nu\rho}g_{\sigma \alpha}\mathfrak{g}^{\mu\alpha}_{\phantom{\mu\alpha},\beta}\mathfrak{g}^{\sigma\beta}_{\phantom{\sigma\beta},\rho}\right)+g_{\rho\sigma}g^{\alpha\beta}\mathfrak{g}^{\mu\rho}_{\phantom{\mu\rho},\alpha}\mathfrak{g}^{\nu\sigma}_{\phantom{\nu m},\beta}\\
&+\frac{1}{8}\left(2g^{\mu\rho}g^{\nu\sigma}-g^{\mu\nu}g^{\rho\sigma}\right)\left(2g_{\alpha\beta}g_{\gamma\lambda}-g_{\beta \gamma}g_{\alpha\lambda}\right)\mathfrak{g}^{\alpha\lambda}_{\phantom{\alpha\lambda},\rho}\mathfrak{g}^{\beta\gamma}_{\phantom{\beta\gamma},\sigma}\ ,
\end{split}
\end{equation}
where $\mathfrak{g}^{\mu\nu}=\sqrt{-g}g^{\mu\nu}$.
§.§ Møller energy–momentum complex
The energy–momentum pseudo-tensors $t^{\mu\nu}$ of both Einstein and Landau–Lifshitz besides having the flaw of being tensors only affine and not covariant also depend on the choice of coordinates. Then, Møller looked for an expression for energy and gravitational momentum independent of the particular coordinate system. To do this Møller exploited the fact that the pseudo-tensor including matter plus gravity $\theta^{\mu\nu}=T^{\mu\nu}+t^{\mu\nu}$ can be defined at less than a magnitude $S^{\mu\nu}$ at zero divergence $\partial_{\mu} S^{\mu\nu}=0$. In 1958 Møller proposed the following complex tensor of energy–momentum complex $\mathcal{T}_{\mu}^{\phantom{\mu}\nu}=\theta_{\mu}^{\phantom{\mu}\nu}+S_{\mu}^{\phantom{\mu}\nu}$ looking for the $S_{\mu}^{\phantom{\mu}\nu}$ such that $\mathcal{T}_{\mu}^{\phantom{\mu}\nu}$ transformed as a tensor for only spatial transformations
\begin{equation}\label{EMTGM}
\mathcal{T}_{\mu}^{\phantom{\mu}\nu}=\frac{1}{8\pi}\partial_{\rho}\left[\sqrt{-g}\left(g_{\mu\sigma,\lambda}-g_{\mu\lambda,\sigma}\right)g^{\lambda\nu}g^{\sigma\rho}\right]\ ,
\end{equation}
where the expression in square brackets is the antisymmetric super-potential $U_{\mu}^{\phantom{\mu}\nu\rho}=-U_{\mu}^{\phantom{\mu}\rho\nu}$ such that
\begin{equation}
\partial_{\nu}\mathcal{T}_{\mu}^{\phantom{\mu}\nu}=0\ .
\end{equation}
§.§ Papapetrou energy–momentum pseudo-tensor
Papapetrou in 1948 used the generalized Belifante method to derive his pseudo-tensor due to the complex of energy–momentum contributions using Tolman’s expression of Einstein's total pseudo-tensor $\theta_{\mu}^{\phantom{\mu}\nu}$ (<ref>) i.e.
\begin{equation}
\theta_{\mu}^{\phantom{\mu}\nu}=\frac{1}{8\pi}\frac{\partial}{\partial x^{\rho}}\left(-\mathfrak{g}^{\nu\sigma}\frac{\partial L}{\partial \mathfrak{g}^{\mu\sigma}_{\phantom{\mu\sigma},\rho}}+\frac{1}{2}\delta_{\mu}^{\nu}\mathfrak{g}^{\alpha\beta}\frac{\partial L}{\partial \mathfrak{g}^{\alpha\beta}_{\phantom{\alpha\beta},\rho}}\right)\ ,
\end{equation}
where $L$ is Einstein Lagrangian give by (<ref>) and $\mathfrak{g}^{\nu\sigma}$ have been defined in Eq. (<ref>). Belifante’s method consists in finding a symmetric quantity $\Omega^{\mu\nu}=\Omega^{\nu\mu}$ divergence free which differs by $\eta^{\mu\rho}\theta_{\rho}^{\phantom{\rho}\nu}$ only for an antisymmetric quantity divergence in the first two indices $B^{\mu\nu\rho}=-B^{\mu\nu\rho}$ i.e.
\begin{equation}
\Omega^{\mu\nu}=\eta^{\mu\rho}\theta_{\rho}^{\phantom{\rho}\nu}+\frac{\partial}{\partial x^{\rho}}B^{\mu\nu\rho}\ ,
\end{equation}
such that
\begin{equation}\label{coserlocpapa}
\frac{\partial}{\partial x^{\nu}}\Omega^{\mu\nu}=0\ ,
\end{equation}
with $\eta_{\mu\nu}=diag\left(1,-1,-1,-1\right)$. Expressing $B^{\mu\nu\rho}$ in terms of of the field spin density $S^{\mu\nu\rho}$
\begin{equation}
B^{\mu\nu\rho}=-\frac{1}{2}\left(S^{\mu\nu\rho}+S^{\rho\mu\nu}+S^{\rho\nu\mu}\right)\ ,
\end{equation}
you get after a few counts the expression for the total pseudo-tensor $\Omega^{\mu\nu}$
\begin{equation}
\Omega^{\mu\nu}=\frac{1}{16\pi}\frac{\partial^{2}}{\partial x^{\rho}x^{\sigma}}\left[\sqrt{-g}\left(g^{\mu\nu}\eta^{\rho\sigma}-g^{\mu\rho}\eta^{\nu\sigma}-g^{\rho\sigma}\eta{\mu\nu}-g^{\nu\sigma}\eta^{\mu\rho}\right)\right]\ .
\end{equation}
This geometric object is symmetric with respect to the first two indices$\mu$ e $\nu$.
§.§ Weinberg gravitational energy–momentum pseudo-tensor
Weinberg [32] derived the gravitational energy–momentum pseudo-tensor by adopting a quasi-minkowskian coordinate system. In this system the metric tensor $g_{\mu\nu}$ tends to that of Minkowski $\eta_{\mu\nu}$ at great distances from a localized material system. We write the metric $g_{\mu\nu}$ as the sum of the metric of Minkowski $\eta_{\mu\nu}$ plus $h_{\mu\nu}$ that goes to zero to infinity
\begin{equation}
g_{\mu\nu}=\eta_{\mu\nu}+h_{\mu\nu}\ .
\end{equation}
We linearize Einstein equations $G_{\mu\nu}=-8\pi G T_{\mu\nu}$, expanding Ricci tensor $R_{\mu\nu}$ in terms of powers of $h_{\mu\nu}$ as
\begin{equation}\label{EELFIRST}
R^{\left(1\right)}_{\phantom{\left(1\right)}\mu\nu}-\frac{1}{2}\eta_{\mu\nu}R^{\left(1\right)}=-8\pi G\left[T_{\mu\nu}+t_{\mu\nu}\right]\ ,
\end{equation}
\begin{equation}
t_{\mu\nu}=\frac{1}{8\pi G}\left(R_{\mu\nu}-\frac{1}{2}g_{\mu\nu}R-R^{\left(1\right)}_{\phantom{\left(1\right)}\mu\nu}+\frac{1}{2}\eta_{\mu\nu}R^{\left(1\right)}\right)\ ,
\end{equation}
is the gravitational energy–momentum pseudo-tensor. So in the Eq. (<ref>) you see that reading the equation from right to left, $t_{\mu\nu}$ assumes the meaning of the source of the linearized curvature together with the tensor of the non-gravitational fields and of the matter $T_{\mu\nu}$. From the linearized Bianchi law to which quantity $R^{\left(1\right)}_{\phantom{\left(1\right)}\mu\nu}$ must satisfy, we get the following local conservation law
\begin{equation}
\frac{\partial}{\partial x^{\nu}}\left(T^{\mu\nu}+t^{\mu\nu}\right)=0\ .
\end{equation}
The pseudo-tensor $t_{\mu\nu}$ to the second order in $h$ is
\begin{equation}
t_{\mu\nu}=\frac{1}{8\pi G}\left(-\frac{1}{2}h_{\mu\nu}R^{\left(1\right)}+\frac{1}{2}\eta_{\mu\nu}\eta^{\rho\sigma}R^{\left(1\right)}_{\phantom{\left(1\right)}\rho\sigma}+R^{\left(2\right)}_{\phantom{\left(2\right)}\mu\nu}-\frac{1}{2}\eta_{\mu\nu}\eta^{\rho\sigma}R^{\left(2\right)}_{\phantom{\left(2\right)}\rho\sigma}\right)+\mathcal{O}\left(h^{3}\right)\ ,
\end{equation}
where $R_{\mu\nu}$ to first order in $h$ is
\begin{equation}
R^{\left(1\right)}_{\phantom{\left(1\right)}\mu\nu}=\frac{1}{2}\left(\frac{\partial^{2}h^{\lambda}_{\phantom{\lambda}\lambda}}{\partial x^{\mu}\partial x^{\nu}}-\frac{\partial^{2}h^{\lambda}_{\phantom{\lambda}\mu}}{\partial x^{\lambda}\partial x^{\nu}}-\frac{\partial^{2}h^{\lambda}_{\phantom{\lambda}\nu}}{\partial x^{\lambda}\partial x^{\mu}}+\frac{\partial^{2}h_{\mu\nu}}{\partial x^{\lambda}\partial x_{\lambda}}\right)\ ,
\end{equation}
while to second order $h$ becomes
\begin{equation}
\begin{split}
R^{\left(2\right)}_{\phantom{\left(2\right)}\mu\nu}=&-\frac{1}{2}h^{\lambda\rho}\left(\frac{\partial^{2}h_{\lambda\rho}}{\partial x^{\nu}\partial x^{\mu}}-\frac{\partial^{2}h_{\mu\rho}}{\partial x^{\nu}\partial x^{\lambda}}-\frac{\partial^{2}h_{\lambda\nu}}{\partial x^{\rho}\partial x^{\mu}}+\frac{\partial^{2}h_{\mu\nu}}{\partial x^{\rho}\partial x^{\lambda}}\right)\\
&+\frac{1}{4}\left(2\frac{\partial h^{\rho}_{\phantom{\rho}\sigma}}{\partial x^{\rho}}-\frac{\partial h^{\rho}_{\phantom{\rho}\rho}}{\partial x^{\sigma}}\right)\left(\frac{\partial h^{\sigma}_{\phantom{\sigma}\mu}}{\partial x^{\nu}}+\frac{\partial h^{\sigma}_{\phantom{\sigma}\nu}}{\partial x^{\mu}}-\frac{\partial h_{\mu\nu}}{\partial x_{\sigma}}\right)\\
&-\frac{1}{4}\left(\frac{\partial h_{\sigma\nu}}{\partial x^{\lambda}}+\frac{\partial h_{\sigma\lambda}}{\partial x^{\nu}}-\frac{\partial h_{\lambda\nu}}{\partial x^{\sigma}}\right)\left(\frac{\partial h^{\sigma}_{\phantom{\sigma}\mu}}{\partial x_{\lambda}}+\frac{\partial h^{\sigma\lambda}}{\partial x^{\mu}}-\frac{\partial h^{\lambda}_{\phantom{\lambda}\mu}}{\partial x_{\sigma}}\right)
\end{split}\ .
\end{equation}
§ ENERGY–MOMENTUM COMPLEX IN CURVATURE BASED GRAVITY
§.§ The gravitational energy–momentum ”tensor” of $n^{th}$ order Lagrangian
Let us examine the energy–momentum complex for a fourth order gravitational Lagrangian, that is, which depends up to fourth derivatives of the metric tensor $g_{\mu\nu}$ as $L=L(g_{\mu\nu}, g_{\mu\nu,\rho}, g_{\mu\nu,\rho\lambda},g_{\mu\nu,\rho\lambda\xi}, g_{\mu\nu,\rho\lambda\xi\sigma})$, whose field equations, in general, are of eighth order in metric formalism (see also [33, 34]). In this manner we include all possible curvature invariants, not only $\Box$ operators, into the gravitational action. Then, we will generalize the approach to a gravitational Lagrangian of $n$-th order, i.e., which depends up to $n^{th}$ derivatives of metric tensor. We will derive the energy–momentum tensor using the Noether's theorem, imposing that gravitational action is invariant under global translations [8]. In this review the metric signature of $g_{\mu\nu}$ adopted is $(+\ \ , -\ \ , -\ \ , -)$, while Ricci tensor is defined as
$R_{\mu\nu}=R_{\ \ \mu\rho\nu}^{\rho}$ and Riemann tensor as $R_{\ \ \beta\mu\nu}^{\alpha}=\Gamma_{\beta\nu,\mu}^{\alpha}+\ldots$.
Let us vary the gravitational action with respect to metric $g_{\mu\nu}$ and coordinates $x^{\mu}$ [18, 35, 36]
\begin{equation}\label{1}
I=\int_{\Omega}d^{4}x L \rightarrow \tilde{\delta} I=\int_{\Omega^{\prime}} d^{4}x^{\prime} L^{\prime}-\int_{\Omega}d^{4}x L=\int_{\Omega}d^{4}x \left[{\delta}L+\partial_{\mu}\left(L\delta x^{\mu}\right)\right]\ ,
\end{equation}
where $\tilde{\delta}$ stands for the local variation while $\delta$ means the total variation, that is, keeping the value of coordinate $x$ fixed. By infinitesimal transformations as
\begin{equation}\label{2}
x^{\prime\mu}=x^{\mu}+\epsilon^{\mu}\left(x\right)\ ,
\end{equation}
the total variation of the metric tensor reads
\begin{equation}\label{3}
\delta g_{\mu\nu}=g^{\prime}_{\mu\nu}\left(x\right)-g_{\mu\nu}\left(x\right)=-\epsilon^{\alpha}\partial_{\alpha}g_{\mu\nu}-g_{\mu\alpha}\partial_{\nu}\epsilon^{\alpha}-g_{\nu\alpha}\partial_{\mu}\epsilon^{\alpha}\ .
\end{equation}
Under global transformation, $\partial_{\lambda}\epsilon^{\mu}=0$, the functional variation of the metric becomes $\delta g_{\mu\nu}=-\epsilon^{\alpha}\partial_{\alpha}g_{\mu\nu}$. If we also require that the action to be invariant under this transformation, that is, $\tilde{\delta I} =0$, from arbitrariness of domain of integration $\Omega$, we have
\begin{equation}\label{4}
\begin{split}
0=\delta L +\partial_{\mu}\left(L\delta x^{\mu}\right)=\biggl(\frac{\partial L}{\partial g_{\mu\nu}}-\partial_{\rho}\frac{\partial L}{\partial g_{\mu\nu,\rho}}+\partial_{\rho}\partial_{\lambda}\frac{\partial L}{\partial g_{\mu\nu,\rho\lambda}}-\partial_{\rho}\partial_{\lambda}\partial_{\xi}
\frac{\partial L}{\partial g_{\mu\nu,\rho\lambda\xi}}\\+\partial_{\rho}\partial_{\lambda}\partial_{\xi}\partial_{\sigma}\frac{\partial L}{\partial g_{\mu\nu,\rho\lambda\xi\sigma}}\biggr)\delta g_{\mu\nu}+\partial_{\eta}\left(2\chi\sqrt{-g}\tau_{\alpha}^{\eta}\right)\epsilon^{\alpha}\ ,
\end{split}
\end{equation}
where the explicit expression of gravitational energy–momentum tensor, that we will see being a pseudo-tensor or affine tensor, is
\begin{multline}\label{7}
\tau_{\alpha}^{\eta}=\frac{1}{2\chi\sqrt{-g}}\biggl[\left(\frac{\partial L}{\partial g_{\mu\nu,\eta}}-\partial_{\lambda}\frac{\partial L}{\partial g_{\mu\nu,\eta\lambda}}+\partial_{\lambda}\partial_{\xi}\frac{\partial L}{\partial g_{\mu\nu,\eta\lambda\xi}}-\partial_{\lambda}\partial_{\xi}\partial_{\sigma}\frac{\partial L}{\partial g_{\mu\nu,\eta\lambda\xi\sigma}}\right)g_{\mu\nu,\alpha}\\
+\left(\frac{\partial L}{\partial g_{\mu\nu,\rho\eta}}-\partial_{\xi}\frac{\partial L}{\partial g_{\mu\nu,\rho\eta\xi}}+\partial_{\xi}\partial_{\sigma}\frac{\partial L}{\partial g_{\mu\nu,\rho\eta\xi\sigma}}\right)g_{\mu\nu,\alpha\rho}
+\left(\frac{\partial L}{\partial g_{\mu\nu,\rho\lambda\eta}}-\partial_{\sigma}\frac{\partial L}{\partial g_{\mu\nu,\rho\lambda\eta\sigma}}\right)g_{\mu\nu,\rho\lambda\alpha}\\
+\frac{\partial L}{\partial g_{\mu\nu,\rho\lambda\eta\sigma}}g_{\mu\nu,\rho\lambda\xi\alpha}-\delta^{\eta}_{\alpha}L\biggr]\ .
\end{multline}
If the metric tensor $g_{\mu\nu}$ satisfies the Euler–Lagrange equations for our gravitational Lagrangian
\begin{equation}\label{5}
\frac{\delta L}{\delta g_{\mu\nu}}=\frac{\partial L}{\partial g_{\mu\nu}}-\partial_{\rho}\frac{\partial L}{\partial g_{\mu\nu,\rho}}+\partial_{\rho}\partial_{\lambda}\frac{\partial L}{\partial g_{\mu\nu,\rho\lambda}}-\partial_{\rho}\partial_{\lambda}\partial_{\xi}\frac{\partial L}{\partial g_{\mu\nu,\rho\lambda\xi}}+\partial_{\rho}\partial_{\lambda}\partial_{\xi}\partial_{\sigma}\frac{\partial L}{\partial g_{\mu\nu,\rho\lambda\xi\sigma}}=0\ ,
\end{equation}
for an arbitrary $\epsilon^{\alpha}$, we get a local continuity equation for our Noether current
\begin{equation}\label{6}
\partial_{\eta}\left(\sqrt{-g} \tau_{\alpha}^{\eta}\right)=0\ .
\end{equation}
In a more compact form, the gravitational energy–momentum tensor takes the following form
\begin{multline}\label{tensem4}
\tau_{\alpha}^{\eta}=\frac{1}{2\chi\sqrt{-g}}\biggl[\sum_{m=0}^{3}\left(-1\right)^{m}\left(
\frac{\partial L}{\partial g_{\mu\nu,\eta i_{0}\cdots i_{m}}}\right)_{,i_{0}\cdots i_{m}}g_{\mu\nu,\alpha}
\\ +\sum_{j=0}^{2}\sum_{m=j+1}^{3}\left(-1\right)^{j}\left(
\frac{\partial L}{\partial g_{\mu\nu,\eta i_{0}\cdots i_{m}}}\right)_{,i_{0}\cdots i_{j}}g_{\mu\nu,i_{j+1}\cdots i_{m}\alpha}-\delta_{\alpha}^{\eta}L\biggr]\ ,
\end{multline}
where we used the following notation
\begin{equation*}
\left(\right)_{,i_{0}}=\mathbb{I} ; \qquad \left(\right)_{,i_{0}\cdots i_{m}}=
\begin{cases}
\left(\right)_{,i_{1}}& \quad \text{if} \quad m=1\\
\left(\right)_{,i_{1}i_{2}}& \quad \text{if} \quad m=2\\
\left(\right)_{,i_{1}i_{2}i_{3}}& \quad \text{if} \quad m=3\\
\text{and so on}&
\end{cases}
;\qquad \left(\right)_{,i_{k}\.i_{k}}=\left(\right)_{,i_{k}}
\end{equation*}
Let us now generalize our approach considering a general Lagrangian density depending up to $n^{th}$ derivative of $g_{\mu\nu}$, that is, $L=L\left(g_{\mu\nu}, g_{\mu\nu,i_{1}}, g_{\mu\nu,i_{1}i_{2}},g_{\mu\nu,i_{1}i_{2}i_{3}},\cdots, g_{\mu\nu,i_{1}i_{2}i_{3}\cdots i_{n}}\right)$. Total variation of Lagrangian $L$ and its Euler–Lagrange equations yield
\begin{equation}\label{8}
\delta L=\sum_{m=0}^{n}\frac{\partial L}{\partial g_{\mu\nu,i_{0}\cdots i_{m}}}\delta g_{\mu\nu,i_{0}\cdots i_{m}}=\sum_{m=0}^{n}\frac{\partial L}{\partial g_{\mu\nu,i_{0}\cdots i_{m}}}\partial_{i_{0}\cdots i_{m}}\delta g_{\mu\nu}\ ,
\end{equation}
\begin{equation}\label{9}
\frac{\delta L}{\delta g_{\mu\nu}}=\sum_{m=0}^{n}\left(-1\right)^{m}\partial_{i_{0}\cdots i_{m}}\frac{\partial L}{\partial g_{\mu\nu,i_{0}\cdots i_{m}}}=0\ ,
\end{equation}
where $\delta/\delta g_{\mu\nu}$ is the functional derivative, while it is possible to exchange the variation $\delta$ with the derivatives $\delta g_{\mu\nu,i_{0}\cdots i_{m}}=\partial_{i_{0}\cdots i_{m}}\delta g_{\mu\nu}$, because we are varying keeping $x$ fixed. So, we can find a most general local continuity equation which allows us to define the energy–momentum pseudo-tensor (which is an affine tensor as it will be proved later) for the gravitational field of $2n^{th}$ order gravity
\begin{multline}\label{tensemN}
\tau_{\alpha}^{\eta}=\frac{1}{2\chi\sqrt{-g}}\biggl[\sum_{m=0}^{n-1}\left(-1\right)^{m}\left(
\frac{\partial L}{\partial g_{\mu\nu,\eta i_{0}\cdots i_{m}}}\right)_{,i_{0}\cdots i_{m}}g_{\mu\nu,\alpha}\\+\Theta_{\left[2,+\infty\right[}\left(n\right)\sum_{j=0}^{n-2}\sum_{m=j+1}^{n-1}\left(-1\right)^{j}\left(
\frac{\partial L}{\partial g_{\mu\nu,\eta i_{0}\cdots i_{m}}}\right)_{,i_{0}\cdots i_{j}}g_{\mu\nu,i_{j+1}\cdots i_{m}\alpha}-\delta_{\alpha}^{\eta}L\biggr]\ ,
\end{multline}
where $\Theta$ is the Heaviside function
\begin{equation}
\Theta_{\left[a,+\infty\right[}\left(n\right)=
\begin{cases}
1& \quad \text{if} \quad n\in \left[a,+\infty\right[\\
0& \quad \text{otherwise}
\end{cases}\ .
\end{equation}
If fields and its derivatives vanish on boundary of our spatial region or rapidly decreasing to the spatial infinite on an infinity spacelike hypersurface, the gravitational energy–momentum tensor is totally conserved and satisfies a more general conservation law.
An alternative way to obtain the tensor (<ref>) is the procedure developed by Landau [8]. For example, we start by deriving the tensor (<ref>), because its generalization to higher order Lagrangians is the same. First of all, let us impose the stationary condition and vary the action with respect to the metric to find the field equations under the hypothesis that both $\delta g_{\mu\nu}$ and the variation of derivative $\delta \partial^{n}g$ vanish on the boundary of integration domain, canceling the surface integrals. Hence, the following occurs:
\begin{gather}\label{15}
\delta I=\delta\int_{\Omega}d^{4}x L\left(g_{\mu\nu}, g_{\mu\nu,\rho}, g_{\mu\nu,\rho\lambda},g_{\mu\nu,\rho\lambda\xi}, g_{\mu\nu,\rho\lambda\xi\sigma}\right) =0\ , \\
\updownarrow\nonumber\\
\frac{\partial L}{\partial g_{\mu\nu}}-\partial_{\rho}\frac{\partial L}{\partial g_{\mu\nu,\rho}}+\partial_{\rho}\partial_{\lambda}\frac{\partial L}{\partial g_{\mu\nu,\rho\lambda}}-\partial_{\rho}\partial_{\lambda}\partial_{\xi}\frac{\partial L}{\partial g_{\mu\nu,\rho\lambda\xi}}+\partial_{\rho}\partial_{\lambda}\partial_{\xi}\partial_{\sigma}\frac{\partial L}{\partial g_{\mu\nu,\rho\lambda\xi\sigma}}=0\ .
\end{gather}
Now, we perform the derivative of Lagrangian respect to metric tensor and then we put it into the field equations (<ref>). We obtain
\begin{multline}\label{16}
\frac{\partial L}{\partial x^{\alpha}} =\frac{\partial L}{\partial g_{\mu\nu}}\frac{\partial g_{\mu\nu}}{\partial x^{\alpha}}+\frac{\partial L}{\partial g_{\mu\nu,\rho}}\frac{\partial g_{\mu\nu,\rho}}{\partial x^{\alpha}}+\frac{\partial L}{\partial g_{\mu\nu,\rho\lambda}}\frac{\partial g_{\mu\nu,\rho\lambda}}{\partial x^{\alpha}}\\
+\frac{\partial L}{\partial g_{\mu\nu,\rho\lambda\xi}}\frac{\partial g_{\mu\nu,\rho\lambda\xi}}{\partial x^{\alpha}}+\frac{\partial L}{\partial g_{\mu\nu,\rho\lambda\xi\sigma}}\frac{\partial g_{\mu\nu,\rho\lambda\xi\sigma}}{\partial x^{\alpha}}\\
=\partial_{\rho} \frac{\partial L}{\partial g_{\mu\nu,\rho}}g_{\mu\nu,\alpha}-\partial_{\rho}\partial_{\lambda}\frac{\partial L}{\partial g_{\mu\nu,\rho\lambda}}g_{\mu\nu,\alpha}+\partial_{\rho}\partial_{\lambda}\partial_{\xi}\frac{\partial L}{\partial g_{\mu\nu,\rho\lambda\xi}}g_{\mu\nu,\alpha}-\partial_{\rho}\partial_{\lambda}\partial_{\xi}\partial_{\sigma}\frac{\partial L}{\partial g_{\mu\nu,\rho\lambda\xi\sigma}}g_{\mu\nu,\alpha}\\
+\frac{\partial L}{\partial g_{\mu\nu,\rho}} g_{\mu\nu,\rho\alpha}+\frac{\partial L}{\partial g_{\mu\nu,\rho\lambda}}g_{\mu\nu,\rho\lambda\alpha}+\frac{\partial L}{\partial g_{\mu\nu,\rho\lambda\xi}}g_{\mu\nu,\rho\lambda\xi\alpha}+\frac{\partial L}{\partial g_{\mu\nu,\rho\lambda\xi\sigma}} g_{\mu\nu,\rho\lambda\xi\sigma\alpha}\\
=\partial_{\rho}\left(\frac{\partial L}{\partial g_{\mu\nu,\rho}}g_{\mu\nu,\alpha}\right)-\partial_{\rho}\left(\partial_{\lambda}\frac{\partial L}{\partial g_{\mu\nu,\rho\lambda}}g_{\mu\nu,\alpha}\right)+\partial_{\lambda}\left(\frac{\partial L}{\partial g_{\mu\nu,\rho\lambda}}g_{\mu\nu,\rho\alpha}\right)\\
+\partial_{\rho}\left(\partial_{\lambda}\partial_{\xi}\frac{\partial L}{\partial g_{\mu\nu,\rho\lambda\xi}}g_{\mu\nu,\alpha}\right)
+\partial_{\lambda}\left(\frac{\partial L}{\partial g_{\mu\nu,\rho\lambda\xi}}g_{\mu\nu,\rho\xi\alpha}\right)\\
-\partial_{\xi}\left(\partial_{\lambda}\frac{\partial L}{\partial g_{\mu\nu,\rho\lambda\xi}}g_{\mu\nu,\alpha\rho}\right)-\partial_{\rho}\left(\partial_{\lambda}\partial_{\xi}\partial_{\sigma}\frac{\partial L}{\partial g_{\mu\nu,\rho\lambda\xi\sigma}}g_{\mu\nu,\alpha}\right)\\
+\partial_{\lambda}\left(\frac{\partial L}{\partial g_{\mu\nu,\rho\lambda\xi\sigma}}g_{\mu\nu,\rho\xi\sigma\alpha}\right)-\partial_{\xi}\left(\partial_{\lambda}\frac{\partial L}{\partial g_{\mu\nu,\rho\lambda\xi\sigma}}g_{\mu\nu,\rho\sigma\alpha}\right)
\\+\partial_{\sigma}\left(\partial_{\xi}\partial_{\lambda}\frac{\partial L}{\partial g_{\mu\nu,\rho\lambda\xi\sigma}}g_{\mu\nu,\rho\alpha}\right)\ .
\end{multline}
Grouping together terms and renaming dumb indices, we obtain
\begin{equation}\label{17}
\partial_{\eta}\left(\sqrt{-g}\tau^{\eta}_{\alpha}\right)=0\ ,
\end{equation}
that is, the pseudo-tensor is locally conserved, where $\tau^{\eta}_{\alpha}$ is the tensor defined in (<ref>).
The energy–momentum complex, instead, can be derived considering the material Lagrangian $L_{m}=2\chi\sqrt{-g}\mathcal{L}_{m}$ with stress–energy tensor given by
\begin{equation}\label{18}
T^{\eta\alpha}=\frac{2}{\sqrt{-g}}\frac{\delta \left(\sqrt{-g}\mathcal{L}_{m}\right)}{\delta g_{\eta\alpha}}\ .
\end{equation}
Thus, we use the field equations in presence of matter, namely
\begin{equation}\label{19}
P^{\eta\alpha}=\chi T^{\eta\alpha}\ ,
\end{equation}
\begin{equation}\label{20}
P^{\eta\alpha}=-\frac{1}{\sqrt{-g}}\frac{\delta L_{g}}{\delta g_{\eta\alpha}}\,\qquad \mbox{with the coupling} \quad \chi=\frac{8\pi G}{c^{4}}\ .
\end{equation}
By field equations (<ref>), we obtain
\begin{multline}\label{21}
\left(2\chi\sqrt{-g}\tau^{\eta}_{\alpha}\right)_{,\eta}=-\sqrt{-g}P^{\rho\sigma}g_{\rho\sigma,\alpha}=-\chi\sqrt{-g}T^{\rho\sigma}g_{\rho\sigma,\alpha}\\
=2\chi\sqrt{-g} T^{\eta}_{\alpha;\eta}-\left(2\chi\sqrt{-g} T^{\eta}_{\alpha}\right)_{,\eta}\ ,
\end{multline}
\begin{equation}\label{22}
\partial_{\eta}\left[\sqrt{-g}\left(\tau^{\eta}_{\alpha}+ T^{\eta}_{\alpha}\right)\right]=\sqrt{-g}T^{\eta}_{\alpha;\eta}\ ,
\end{equation}
\begin{multline}\label{23}
\delta L +\partial_{\mu}\left(L\delta x^{\mu}\right)=-P^{\mu\nu}\sqrt{-g}\delta g_{\mu\nu}+\partial_{\eta}\left(2\chi\sqrt{-g}\tau^{\eta}_{\alpha}\right)\epsilon^{\alpha}\\
=\left[\sqrt{-g}P^{\mu\nu}g_{\mu\nu,\alpha}+\partial_{\eta}\left(2\chi\sqrt{-g}\tau^{\eta}_{\alpha}\right)\right]\epsilon^{\alpha}=0\ ,
\end{multline}
and because from symmetry of tensor $T^{\eta}_{\alpha}$, one gets
\begin{equation}\label{24}
\sqrt{-g}T^{\eta}_{\alpha;\eta}=\left(\sqrt{-g}T^{\eta}_{\alpha}\right)_{,\eta}-\frac{1}{2}g_{\rho\sigma,\alpha}T^{\rho\sigma}\sqrt{-g}\ .
\end{equation}
The relation (<ref>) tells us that the conservation law of the energy–momentum complex, i.e., the sum of two stress–energy tensors due to matter plus gravitational fields, is related to the covariant derivative of the only matter part. From contracted Bianchi identities we get the total conservation law and conversely
\begin{equation}\label{eqcontgeneralizzata}
G^{\eta\alpha}_{;\eta}=0 \leftrightarrow P^{\eta\alpha}_{;\eta}=0 \leftrightarrow T^{\eta\alpha}_{;\eta}=0 \leftrightarrow \partial_{\eta}\left[\sqrt{-g}\left(\tau^{\eta}_{\alpha}+ T^{\eta}_{\alpha}\right)\right]=0\ ,
\end{equation}
where ${\displaystyle G^{\eta\alpha}=R^{\eta\alpha}-\frac{1}{2}g^{\eta\alpha}R}$ is the Einstein tensor and the locally conserved energy–momentum complex is given by
\begin{equation}
\mathcal{T}_{\alpha}^{\eta}=\sqrt{-g}\left(\tau^{\eta}_{\alpha}+ T^{\eta}_{\alpha}\right)\ .
\end{equation}
In a nutshell, the contracted Bianchi identities lead to the local conservation of energy–momentum complex or, viceversa, the local conservation of matter and gravitational fields involves the contracted Bianchi identities (see also [37] for a detailed discussion in modified gravity).
From the local continuity equation (<ref>), it is possible to derive some conserved quantities, Noether charges, such as the total 4-momentum of matter plus gravitational field. If we require that the metric tensor derivatives up to the $n^{th}$ order vanish on the 3-dimensional space-domain $\Sigma$, the surface integrals are zero over the boundary $\partial\Sigma$, that is
\begin{equation}\label{25}
\partial_{0}\int_{\Sigma}d^{3}x \sqrt{-g}\left(T^{\mu0}+\tau^{\mu0}\right)=-\int_{\partial\Sigma}d\sigma_{i} \sqrt{-g}\left(T^{\mu i}+\tau^{\mu i}\right)=0\ ,
\end{equation}
where $\Sigma$ is a slice of 4-dimensional manifold of spacetime at $t$ fixed and $\partial\Sigma$ its boundary. Such conditions are fulfilled by when we are in the presence of localized objects, where we can take a spatial domain that becomes flat to infinity, i.e, a asymptotically flat spacetime. So, the energy and total momentum conserved become Ref. [38]
\begin{equation}\label{26}
P^{\mu}=\int_{\Sigma}d^{3}x \sqrt{-g}\left(T^{\mu0}+\tau^{\mu0}\right)\ .
\end{equation}
These quantities are very useful in astrophysical applications [39].
§.§ Non-covariance of gravitational energy–momentum tensor
We will prove that the tensor $\tau^{\eta}_{\alpha}$ is not a covariant object but affine, that is, it is changes like a tensor under affine transformations [40], i.e., a pseudo-tensor. We will limit ourselves first to a particular case, $n=2$, where the tensor (<ref>) reads
\begin{equation}\label{27}
\tau^{\eta}_{\alpha}=\frac{1}{2\chi\sqrt{-g}}\left[\left(\frac{\partial L}{\partial g_{\mu\nu,\eta}}-\partial_{\lambda}\frac{\partial L}{\partial g_{\mu\nu,\eta\lambda}}\right)g_{\mu\nu,\alpha}+\frac{\partial L}{\partial g_{\mu\nu,\eta\xi}}g_{\mu\nu,\xi\alpha}-\delta^{\eta}_{\alpha} L\right]\ ,
\end{equation}
We will show that, while under a general diffeomorphism transformation $x^{\prime}=x^{\prime}\left(x\right)$, the tensor changes as
\begin{equation}\label{28}
\tau^{\prime\eta}_{\ \alpha}\left(x^{\prime}\right) \neq \text{J}^{\eta}_{\sigma}\text{J}^{-1\tau}_{\ \ \ \alpha}\tau^{\sigma}_{\tau}\left(x\right)\ ,
\end{equation}
with Jacobian matrix and determinant defined as
\begin{equation}\label{29}
\text{J}^{\eta}_{\sigma}=\frac{\partial x^{\prime\eta}}{\partial x^{\sigma}}\qquad \text{J}^{-1\tau}_{\ \ \ \alpha}=\frac{\partial x^{\tau}}{\partial x^{\prime\alpha}}\qquad \text{det}\left(\text{J}^{\alpha}_{\beta}\right)=\vert J \vert =\frac{1}{\text{J}^{-1}}\ ,
\end{equation}
under the following affine transformations
\begin{equation}\label{30}
x^{\prime\mu}=\Lambda^{\mu}_{\nu}x^{\nu}\qquad \text{J}^{\mu}_{\nu}=\Lambda^{\mu}_{\nu} \qquad \vert \Lambda \vert \neq 0\ ,
\end{equation}
the tensor is transformed as
\begin{equation}\label{31}
\tau^{\prime\eta}_{\ \alpha}\left(x^{\prime}\right)=\Lambda^{\eta}_{\sigma}\Lambda^{-1\tau}_{\ \ \ \alpha}\tau^{\sigma}_{\tau}\left(x\right)\ .
\end{equation}
Generally, following identities occur
\begin{equation*}
\begin{split}
\sqrt{-g^{\prime}}&=\sqrt{-g}\qquad\qquad \ \ \qquad\quad \, \,\text{where $g$ is a scalar density of weight $w=-2$ }\ ,\\
L^{\prime}&=\text{J}^{-1}L\qquad\qquad \ \qquad \qquad \text{where $L$ is a scalar density of weight $w=-1$ }\ ,\\
g^{\prime}_{\mu\nu,\alpha}\left(x^{\prime}\right)&=\text{J}^{-1 a}_{\ \ \ \mu}\text{J}^{-1 b}_{\ \ \ \nu}\text{J}^{-1 c}_{\ \ \ \alpha}g_{ab,c}\left(x\right)+\partial^{\prime}_{\alpha}\left[\text{J}^{-1 a}_{\ \ \ \mu}\text{J}^{-1 b}_{\ \ \ \nu}\right]g_{ab}\left(x\right)\ ,\\
\frac{\partial g_{\gamma\rho,\tau}}{\partial g^{\prime}_{\mu\nu,\eta}}&=\frac{1}{2}\left[\left(\delta_{a}^{\mu}\delta_{b}^{\nu}+\delta_{a}^{\nu}\delta_{b}^{\mu}\right)\delta_{c}^{\eta}\right]\text{J}_{\gamma}^{a}\text{J}_{\rho}^{b}\text{J}_{\tau}^{c}=\text{J}^{(\mu}_{\gamma}\text{J}^{\nu)}_{\rho}\text{J}^{\eta}_{\tau}\ ,\\
\frac{\partial L^{\prime}}{\partial g^{\prime}_{\mu\nu,\eta}}&=\text{J}^{-1}\text{J}^{(\mu}_{\gamma}\text{J}^{\nu)}_{\rho}\text{J}^{\eta}_{\tau}\frac{\partial L}{\partial g_{\gamma\rho,\tau}}=\text{J}^{-1}\text{J}^{\mu}_{\gamma}\text{J}^{\nu}_{\rho}\text{J}^{\eta}_{\tau}\frac{\partial L}{\partial g_{\gamma\rho,\tau}}\ \\
& \qquad\qquad\qquad\qquad\qquad\qquad \text{tensorial density (3,0) of weight $w=-1$}\ ,\\
g^{\prime}_{\mu\nu,\xi\alpha}\left(x^{\prime}\right)&=\text{J}^{-1 a}_{\ \ \ \mu}\text{J}^{-1 b}_{\ \ \ \nu}\text{J}^{-1 c}_{\ \ \ \alpha}\text{J}^{-1 d}_{\ \ \ \xi}g_{ab,cd}\left(x\right)+\partial^{\prime 2}_{\xi\alpha}\left[\text{J}^{-1 a}_{\ \ \ \mu}\text{J}^{-1 b}_{\ \ \ \nu}\right]g_{ab}\left(x\right)\\
&+\partial^{\prime}_{\alpha}\left[\text{J}^{-1 a}_{\ \ \ \mu}\text{J}^{-1 b}_{\ \ \ \nu}\right]\text{J}^{-1 d}_{\ \ \ \xi}g_{ab,d}\left(x\right)+\partial^{\prime}_{\xi}\left[\text{J}^{-1 a}_{\ \ \ \mu}\text{J}^{-1 b}_{\ \ \ \nu}\text{J}^{-1 c}_{\ \ \ \alpha}\right]g_{ab,c}\left(x\right)\ ,\\
\frac{\partial g_{\gamma\rho,\tau\epsilon}}{\partial g_{\mu\nu,\eta\xi}^{\prime}}&=\left(\delta_{a}^{(\mu}\delta_{b}^{\nu)}\delta_{c}^{(\eta}\delta_{d}^{\xi)}\right)\text{J}_{\gamma}^{a}\text{J}_{\rho}^{b}\text{J}_{\tau}^{c}\text{J}_{\epsilon}^{d}=\text{J}_{\gamma}^{(\mu}\text{J}_{\rho}^{\nu)}\text{J}_{\tau}^{(\eta}\text{J}_{\epsilon}^{\xi)}\ ,\\
\frac{\partial L^{\prime}}{\partial g^{\prime}_{\mu\nu,\eta\xi}}&=\text{J}^{-1}\text{J}^{(\mu}_{\gamma}\text{J}^{\nu)}_{\rho}\text{J}^{(\eta}_{\tau}\text{J}^{\xi)}_{\epsilon}\frac{\partial L}{\partial g_{\gamma\rho,\tau\epsilon}}=\text{J}^{-1}\text{J}^{\mu}_{\gamma}\text{J}^{\nu}_{\rho}\text{J}^{\eta}_{\tau}\text{J}^{\xi}_{\epsilon}\frac{\partial L}{\partial g_{\gamma\rho,\tau\epsilon}}\\
& \qquad\qquad\qquad\qquad\qquad\qquad\text{tensorial density (4,0) of weight $w=-1$}\ ,\\
\partial^{\prime}_{\lambda}\frac{\partial L^{\prime}}{\partial g^{\prime}_{\mu\nu,\eta\lambda}}&=\text{J}^{-1}\text{J}^{\mu}_{\gamma}\text{J}^{\nu}_{\rho}\text{J}^{\eta}_{\tau}\text{J}^{\lambda}_{\epsilon}\text{J}^{-1 \sigma}_{\ \ \ \lambda}\partial_{\sigma}\frac{\partial L}{\partial g_{\gamma\rho,\tau\epsilon}}+\partial^{\prime}_{\lambda}\left[\text{J}^{-1}\text{J}^{\mu}_{\gamma}\text{J}^{\nu}_{\rho}\text{J}^{\eta}_{\tau}\text{J}^{\lambda}_{\epsilon}\right]\frac{\partial L}{\partial g_{\gamma\rho,\tau\epsilon}}\ ,
\end{split}
\end{equation*}
and by symmetry of $B_{\alpha\beta}$, i.e., $B_{\alpha\beta}=B_{\beta\alpha}$ follows that $A^{(\alpha\beta)}B_{\alpha\beta}=A^{\alpha\beta}B_{\alpha\beta}$.
Then we have
\begin{equation*}
\frac{\partial L^{\prime}}{\partial g^{\prime}_{\mu\nu,\eta}}g^{\prime}_{\mu\nu,\alpha}=\text{J}^{-1}\text{J}^{\eta}_{\tau}\text{J}^{-1\pi}_{\ \ \ \alpha}\frac{\partial L}{\partial g_{\gamma\rho,\tau}}g_{\gamma\rho,\pi}\left(x\right)+\frac{\partial}{\partial x^{\prime\alpha}}\left[\text{J}^{-1a}_{\ \ \ \mu}\text{J}^{-1b}_{\ \ \ \nu}\right]g_{ab}\left(x\right)\text{J}^{-1}\text{J}^{\mu}_{\gamma}\text{J}^{\nu}_{\rho}\text{J}^{\eta}_{\tau}\frac{\partial L}{\partial g_{\gamma\rho,\tau}}\ ,\\
\end{equation*}
\begin{multline*}
\partial^{\prime}_{\lambda}\frac{\partial L^{\prime}}{\partial g^{\prime}_{\mu\nu,\eta\lambda}}g^{\prime}_{\mu\nu,\alpha}\left(x^{\prime}\right)=\text{J}^{-1}\text{J}^{\eta}_{\tau}\text{J}^{-1c}_{\ \ \ \alpha}\partial_{\sigma}\frac{\partial L}{\partial g_{ab,\tau\sigma}}g_{ab,c}+\partial^{\prime}_{\lambda}\left[\text{J}^{-1}\text{J}^{\mu}_{\gamma}\text{J}^{\nu}_{\rho}\text{J}^{\eta}_{\tau}\text{J}^{\lambda}_{\epsilon}\right]
\partial^{\prime}_{\alpha}\left[\text{J}^{-1a}_{\ \ \ \mu}\text{J}^{-1b}_{\ \ \ \nu}\right]g_{ab}\left(x\right)\frac{\partial L}{\partial g_{\gamma\rho,\tau\epsilon}}\\
+\text{J}^{-1}\text{J}^{\mu}_{\gamma}\text{J}^{\nu}_{\rho}\text{J}^{\eta}_{\tau}\partial_{\sigma}\frac{\partial L}{\partial g_{\gamma\rho,\tau\sigma}}\partial^{\prime}_{\alpha}\left[\text{J}^{-1a}_{\ \ \ \mu}\text{J}^{-1b}_{\ \ \ \nu}\right]g_{ab}+\partial^{\prime}_{\lambda}\left[\text{J}^{-1}\text{J}^{\mu}_{\gamma}\text{J}^{\nu}_{\rho}\text{J}^{\eta}_{\tau}\text{J}^{\lambda}_{\epsilon}\right]\text{J}^{-1a}_{\ \ \ \mu}\text{J}^{-1b}_{\ \ \ \nu}\text{J}^{-1c}_{\ \ \ \alpha}\frac{\partial L}{\partial g_{\gamma\rho,\tau\epsilon}}g_{ab,c} \ ,
\end{multline*}
\begin{multline*}
\frac{\partial L^{\prime}}{\partial g^{\prime}_{\mu\nu,\eta\xi}}g^{\prime}_{\mu\nu,\xi\alpha}\left(x^{\prime}\right)=\text{J}^{-1}\text{J}^{\eta}_{\tau}\text{J}^{-1\omega}_{\ \ \ \alpha}\frac{\partial L}{\partial g_{\gamma\rho,\tau\epsilon}}g_{\gamma\rho,\omega\epsilon}\left(x\right)+\text{J}^{-1}\partial^{\prime 2}_{\xi\alpha}\left[\text{J}^{-1a}_{\ \ \ \mu}\text{J}^{-1b}_{\ \ \ \nu}\right]g_{ab}\left(x\right)\text{J}^{\mu}_{\gamma}\text{J}^{\nu}_{\rho}\text{J}^{\eta}_{\tau}\text{J}^{\xi}_{\epsilon}\frac{\partial L}{\partial g_{\gamma\rho,\tau\epsilon}}\\
+\text{J}^{-1}\partial^{\prime}_{\alpha}\left[\text{J}^{-1a}_{\ \ \ \mu}\text{J}^{-1b}_{\ \ \ \nu}\right]g_{ab,d}\left(x\right)\text{J}^{\mu}_{\gamma}\text{J}^{\nu}_{\rho}\text{J}^{\eta}_{\tau}\frac{\partial L}{\partial g_{\gamma\rho,\tau d}}+\text{J}^{-1}\text{J}^{\mu}_{\gamma}\text{J}^{\nu}_{\rho}\text{J}^{\eta}_{\tau}\text{J}^{\xi}_{\epsilon}\partial^{\prime}_{\xi}\left[\text{J}^{-1a}_{\ \ \ \mu}\text{J}^{-1b}_{\ \ \ \nu}\text{J}^{-1c}_{\ \ \ \alpha}\right]g_{ab,c}\left(x\right)\frac{\partial L}{\partial g_{\gamma\rho,\tau\epsilon}},
\end{multline*}
Finally, taking into account previous relations we get
\begin{equation}\label{affinitatensore}
\tau^{\prime\eta}_{\ \alpha}\left(x^{\prime}\right)=\text{J}^{\eta}_{\sigma}\text{J}^{-1\tau}_{\ \ \ \alpha}\tau^{\sigma}_{\tau}\left(x\right)+\left\{\text{terms containing }\frac{\partial^{2} x}{\partial x^{\prime 2}},\frac{\partial^{3} x}{\partial x^{\prime 3}} \right\}\ .
\end{equation}
Extra terms that include derivatives of order equal to or greater than two vanish for each non-singular affine transformation but not for generic diffeomorphisms. This proves that gravitational stress–energy tensor is non-covariant but affine, that is, it is invariant under affine transformations due to non-covariance of the derivatives of the metric tensor $g_{\mu\nu}$, that make it at least affine.
Generalizing to $n$-th order Lagrangian, metric tensor derivatives change as
\begin{multline}\label{32}
g^{\prime}_{\mu\nu,i_{1}\cdots i_{m}\alpha}\left(x^{\prime}\right)=\text{J}^{-1\alpha}_{\ \ \ \mu}\text{J}^{-1\beta}_{\ \ \ \nu}\text{J}^{-1 j_{1}}_{\ \ \ i_{1}}\cdots\text{J}^{-1 j_{m}}_{\ \ \ i_{m}}\text{J}^{-1\tau}_{\ \ \ \alpha}g_{\alpha\beta,j_{1}\cdots j_{m}\tau}\left(x\right)\\
+\left\{\text{containing terms}\;\frac{\partial^{2} x}{\partial x^{\prime 2}},\cdots,\frac{\partial^{m+2}x}{\partial x^{\prime m+2}} \right\},
\end{multline}
and derivatives of Lagrangian as
\begin{equation*}
\frac{\partial L^{\prime}}{\partial g^{\prime}_{\mu\nu,\eta i_{0}\cdots i_{m}}}=\text{J}^{-1}\text{J}^{\mu}_{\gamma}\text{J}^{\nu}_{\rho}\text{J}^{\eta}_{\tau}\text{J}^{i_{1}}_{j_{1}}\cdots \text{J}^{i_{m}}_{j_{m}} \frac{\partial L}{\partial g_{\gamma\rho,\tau j_{1}\cdots j_{m}}}\quad\text{tensorial density (m+3,0) of weight $w=-1$}\ ,
\end{equation*}
so that the non covariance of tensor $\tau^{\eta}_{\alpha}$ appears. Otherwise, we obtain for affine transformations
\begin{equation*}
\frac{\partial^{2}x}{\partial x^{\prime 2}}=\cdots=\frac{\partial^{m+2}x}{\partial x^{\prime m+2}}=0\ ,
\end{equation*}
\begin{equation*}
\tau^{\prime\eta}_{\ \alpha}\left(x^{\prime}\right)=\Lambda^{\eta}_{\sigma}\Lambda^{-1\tau}_{\ \ \ \alpha}\tau^{\sigma}_{\tau}\left(x\right)\ ,
\end{equation*}
that is, the energy–momentum tensor of gravitational field is a pseudo-tensor. This result generalize to Extended Theories of Gravity the result in [8]. The affine character of the stress–energy tensor $\tau ^ {\eta} _ {\alpha}$ is a exhibition of the non localizability of gravitational energy density. Specifically, the gravitational energy in a finite-dimensional space, at a given time, depends on the choice of coordinate system [41, 38].
It is worth highlighting that the existence of particular Lagrangians for which extra terms in Eq. (<ref>) vanish cannot be excluded a priori. This is because terms depending on derivatives in the bracket (<ref>) such as ${\frac{\partial^{2} x}{\partial x^{\prime 2}},\cdots,\frac{\partial^{m+2}x}{\partial x^{\prime m+2}}}$, can cancel each other out. Consequently, the energy–momentum pseudo-tensor ${\tau_{\alpha}^{\eta}}$ become a covariant tensor. However, due to the structure of (<ref>), in general, ${\tau_{\alpha}^{\eta}}$ is a pseudo-tensor.
§.§ The gravitational energy–momentum pseudo-tensor of $f\left(R\right)$ gravity
Let us examine the gravitational stress–energy tensor in the $f\left(R\right)$ gravity. Now, the gravitational action is given by
\begin{equation}\label{33}
\mathcal{S}_{f\left(R\right)}=\frac{1}{2\kappa^{2}}\int_{\Omega}d^{4}x\sqrt{-g}f\left(R\right)\,
\end{equation}
with the coupling $\kappa^{2}=8\pi G/c^{4}$. We perform the variation $\tilde{\delta}$ with respect to the metric $g^{\mu\nu}$ and coordinates $x^{\mu}$ for a generic infinitesimal transformation
\begin{equation}\label{34}
x^{\prime\mu}=x^{\mu}+\delta x^{\mu}\,\qquad g^{\prime\mu\nu}\left(x^{\prime}\right)= g^{\mu\nu}\left(x\right)+\tilde{\delta}g^{\mu\nu}\,\qquad g^{\prime\mu\nu}\left(x\right)= g^{\mu\nu}\left(x\right)+\delta g^{\mu\nu}\,
\end{equation}
\begin{equation}\label{varloc}
\tilde{\delta}\mathcal{S}_{f\left(R\right)}=\frac{1}{2\kappa^{2}}\int_{\Omega}d^{4}x\left[\delta\left(\sqrt{-g}f\left(R\right)\right)+\partial_{\mu}\left(\sqrt{-g}f\left(R\right) \delta x^{\mu}\right)\right]\,
\end{equation}
where $\delta$ is the global variation keeping $x$ fixed.
Thus, we get [18, 37, 42, 43, 44, 45]
\begin{multline}\label{varlocfR}
\tilde{\delta}\mathcal{S}_{f\left(R\right)}=\frac{1}{2\kappa^{2}}\int_{\Omega}d^{4}x\sqrt{-g}\Biggl[f'\left(R\right)R_{\mu\nu}-\frac{1}{2}g_{\mu\nu}f\left(R\right)-\nabla_{\mu}\nabla_{\nu}f'\left(R\right)\\
+g_{\mu\nu}\Box f'\left(R\right)\Biggr]\delta g^{\mu\nu}+\int_{\Omega}d^{4}x\partial_{\alpha}\Biggl\{\frac{\sqrt{-g}}{2\kappa^{2}}\biggl[\partial_{\beta}f'\left(R\right)\left(g^{\eta\rho}g^{\alpha\beta}-g^{\alpha\eta}g^{\rho\beta}\right)\delta g_{\eta\rho}\\
+f'\left(R\right)\Bigl[\left(\stackrel{\circ}\Gamma{}^{\rho\eta\alpha}-\stackrel{\circ}\Gamma{}^{\eta\sigma}{}_{\sigma}g^{\alpha\rho}\right)\delta g_{\eta\rho}+\left(g^{\alpha\eta}g^{\tau\rho}-g^{\eta\rho}g^{\alpha\tau}\right)\delta g_{\eta\rho,\tau}\Bigr]+f\left(R\right)\delta_{\lambda}^{\alpha}\delta x^{\lambda}\biggr]\Biggr\}\,
\end{multline}
where $f'\left(R\right)=\partial f/\partial R$. By the condition of stationarity of the action at $x$ fixed, that is, $\delta\mathcal{S}_{f\left(R\right)}=0$, in a given domain $\Omega$ where the total variation of both metric and its first derivatives are zero on the boundary, that is, $\delta g_{\mu\nu}\vert_{\partial\Omega}=0$ and $\delta \left(\partial_{\alpha}g_{\mu\nu}\right)\vert_{\partial\Omega}=0$, the field equations in vacuum become
\begin{multline}\label{ECFRV}
P^{f\left(R\right)}_{\mu\nu}=\frac{2\kappa^{2}}{\sqrt{-g}}\frac{\delta L_{f\left(R\right)}}{\delta g^{\mu\nu}}\\
=f'\left(R\right)R_{\mu\nu}-\frac{1}{2}g_{\mu\nu}f\left(R\right)-\nabla_{\mu}\nabla_{\nu}f'\left(R\right)+g_{\mu\nu}\Box f'\left(R\right)=0\,
\end{multline}
where $2\kappa^{2}L_{f\left(R\right)}=\sqrt{-g}f\left(R\right)$. For an infinitesimal transformation such as a rigid translation, one gets
\begin{equation}\label{TrRig}
x^{\prime\mu}=x^{\mu}+\epsilon^{\mu}\Rightarrow \delta g_{\mu\nu}=-\epsilon^{\lambda}g_{\mu\nu,\lambda}\,
\end{equation}
because $\partial_{\mu} \epsilon^{\mu}=0$. When the local variation of the action vanishes and the field $g_{\mu\nu}$ fulfils the field equations, we obtain the continuity equation
\begin{equation}\label{35}
\tilde{\delta}\mathcal{S}_{f\left(R\right)}=0 \Rightarrow \partial_{\sigma}\left(\sqrt{-g}\tau^{\sigma}_{\phantom{\sigma}{\lambda | f\left(R\right)}}\right)=0\,
\end{equation}
where the gravitational energy–momentum pseudo-tensor of $f\left(R\right)$ gravity is defined as
\begin{multline}\label{36}
2\kappa^{2}\tau^{\sigma}{}_{\lambda | f\left(R\right)}=2\partial_{\beta}f'\left(R\right)g^{\eta[\rho}g^{\sigma]\beta}g_{\eta\rho,\lambda}\\+f'\left(R\right)\Bigl[\bigl(\stackrel{\circ}{\Gamma}{}^{\rho\eta\sigma}
-\stackrel{\circ}{\Gamma}{}^{\eta\alpha}_{\phantom{\eta\alpha}{\alpha}}g^{\sigma\rho}\bigr)g_{\eta\rho,\lambda}+2g^{\sigma[\eta}g^{\tau]\rho}g_{\eta\rho,\tau\lambda}\Bigr]-f\left(R\right)\delta_{\lambda}^{\sigma}\ ,
\end{multline}
with $\stackrel{\circ}{\Gamma}{}^{\rho\eta\sigma}=g^{\eta\epsilon}g^{\sigma\varphi}\stackrel{\circ}{\Gamma}{}^{\rho}_{\phantom{\eta}{\epsilon\varphi}}$, and $\stackrel{\circ}{\Gamma}{}^{\eta\alpha}_{\phantom{\eta\alpha}{\alpha}}=g^{\alpha\epsilon}\stackrel{\circ}{\Gamma}{}^{\eta}_{\phantom{\eta}{\epsilon\alpha}}$.
Now to derive an equation of continuity for energy–momentum complex, we must also include matter fields, as in matter action
\begin{equation}\label{37}
\mathcal{S}_{m}=\int_{\Omega}d^{4}x L_{m}\,
\end{equation}
where $L_{m}$ depends, at most, on first derivatives of metric $g_{\mu\nu}$. Varying the matter action (<ref>), it gets
\begin{equation}\label{38}
\delta\mathcal{S}_{m}=\int_{\Omega}d^{4}x \frac{\delta L_{m}}{\delta g^{\mu\nu}}\delta g^{\mu\nu}=\int_{\Omega}d^{4}x \left(\frac{\sqrt{-g}}{2}\right)T^{\left(m\right)}_{\mu\nu}\delta g^{\mu\nu}\,
\end{equation}
where the energy–momentum tensor of matter fields $T^{\left(m\right)}_{\mu\nu}$ is defined as
\begin{equation}\label{39}
T^{\left(m\right)}_{\mu\nu}=-\frac{2}{\sqrt{-g}}\frac{\delta L_{m}}{\delta g^{\mu\nu}}\,.
\end{equation}
So minimizing the total action $\mathcal{S_{T}}=\mathcal{S}_{f\left(R\right)}+\mathcal{S}_{m}$ and imposing suitable boundary conditions, field equations in presence of matter take the following form
\begin{equation}\label{ECFRM}
\end{equation}
According to contracted Bianchi identities and following formula
\begin{equation}\label{trinabla}
\nabla^{\nu}\nabla_{\mu}\nabla_{\nu} f\left(R\right)=R^{\alpha}_{\phantom{\alpha}\mu}\nabla_{\alpha}f\left(R\right)+\nabla_{\mu}\Box f\left(R\right)
\end{equation}
we derive equivalences
\begin{equation}\label{IBFR}
\nabla^{\nu}G_{\mu\nu}=0\leftrightarrow\nabla^{\nu} \left({}^{f\left(R\right)}P_{\mu\nu}\right)=0\leftrightarrow\nabla^{\nu}T^{\left(m\right)}_{\mu\nu}=0\,
\end{equation}
The variation (<ref>) of gravitational action, the rigid translation (<ref>) and the matter field equations (<ref>) give
\begin{align}\label{41}
\delta L_{f\left(R\right)}+\partial_{\sigma}\left(L_{f\left(R\right)}\delta x^{\sigma}\right)&=\frac{\sqrt{-g}}{2\kappa^{2}}P_{{f\left(R\right)}}^{\mu\nu}\delta g_{\mu\nu}-\partial_{\sigma}\left(\sqrt{-g}\tau^{\sigma}_{\phantom{\sigma}{\lambda}}\right)\epsilon^{\lambda}\nonumber\\&=\left[-\frac{1}{2}\sqrt{-g}T_{\left(m\right)}^{\mu\nu}g_{\mu\nu,\lambda}-\partial_{\sigma}\left(\sqrt{-g}\tau^{\sigma}_{\phantom{\sigma}{\lambda}}\right)\right]\epsilon^{\lambda}\,.
\end{align}
Taking into account Eq. (<ref>), the expression Eq. (<ref>) yields
\begin{equation}\label{42}
\delta L_{f\left(R\right)}+\partial_{\sigma}\left(L_{f\left(R\right)}\delta x^{\sigma}\right)=\left[-\partial_{\sigma}\left(\sqrt{-g}T^{\sigma}_{\phantom{\sigma}{\lambda}}\right)+\sqrt{-g}T^{\sigma}_{\phantom{\sigma}{\lambda;\sigma}}-\partial_{\sigma}\left(\sqrt{-g}\tau^{\sigma}_{\phantom{\sigma}{\lambda}}\right)\right]\epsilon^{\lambda}\,.
\end{equation}
Imposing the local variation to zero, under rigid translations, we have
\begin{equation}\label{43}
\delta L_{f\left(R\right)}+\partial_{\sigma}\left(L_{f\left(R\right)}\delta x^{\sigma}\right)=0 \rightarrow \partial_{\sigma}\left[\sqrt{-g}\left(\tau^{\sigma}_{\ \lambda}+T^{\sigma}{}_{\lambda}\right)\right]=\sqrt{-g}\nabla_{\sigma}T^{\sigma}{}_{\lambda}\,.
\end{equation}
From the contracted Bianchi identities (<ref>), we derive local conservation law for the energy–momentum complex $\mathcal{T}^{\sigma}_{\phantom{\sigma}\lambda}$ in $f(R)$ gravity
\begin{equation}\label{44}
\partial_{\sigma}\left[\sqrt{-g}\left(\tau^{\sigma}_{\phantom{\sigma}{\lambda | f\left(R\right)}}+T^{\sigma}_{\phantom{\sigma}{\lambda}}\right)\right]=0\,.
\end{equation}
\begin{equation}\label{45}
\mathcal{T}^{\sigma}_{\phantom{\sigma}\lambda}=\sqrt{-g}\left(\tau^{\sigma}_{\phantom{\sigma}{\lambda | f\left(R\right)}}+T^{\sigma}_{\phantom{\sigma}{\lambda}}\right)\ .
\end{equation}
§.§ The gravitational energy–momentum pseudo-tensor of higher order gravity
Let us now address theories of gravity of order higher than fourth, where terms containing $\Box$ operators occur in the action up to $n$ times. In supergravity and, more broadly, in gauge theories concerning with gravity [46, 47, 48], these theories are not only effective field theories, but also fundamental theories. Actually, there is at least a subclass of local higher derivative theories, the so called Lee–Wick theories, that are unitary and super-renormalizable or finite at quantum level as demonstrated in [49, 50]. Then, we consider the linear and quadratic part of the Ricci scalar $R$, the first $\overline{R}$ depends only on first derivative of metric tensor $g_{\mu\nu}$ and the second $R^{\star}$ depends linearly on second derivative of metric tensor, as follows [8, 40, 32]
\begin{equation}\label{46}
R=R^{\star}+\overline{R}\ ,
\end{equation}
\begin{equation}\label{47}
R^{\star}=g^{\mu\nu}\left(\Gamma^{\rho}_{\mu\nu,\rho}-\Gamma^{\rho}_{\mu\rho,\nu}\right)\ ,
\end{equation}
\begin{equation}\label{48}
\overline{R}=g^{\mu\nu}\left(\Gamma_{\mu\nu}^{\sigma}\Gamma_{\sigma\rho}^{\rho}-\Gamma_{\mu\sigma}^{\rho}\Gamma_{\nu\rho}^{\sigma}\right)\ .
\end{equation}
Hence, we want to derive the energy–momentum pseudo-tensor $\tau_{\alpha}^{\eta}$ for a gravitational Lagrangian given by
\begin{equation}
\label{higher}
L_{g}=(\overline{R}+a_{0}R^{2}+\sum_{k=1}^{p} a_{k}R\Box^{k}R)\sqrt{-g}\ ,
\end{equation}
that has been first considered in [51].
Therefore, for the purpose of derive the pseudo-tensor $\tau^{\eta}_{\alpha}$, we have first to calculate derivatives present into the Eq. (<ref>), namely
\begin{align}
\frac{\partial L}{\partial g_{\mu\nu,\eta}}&=\sqrt{-g}\left[\frac{\partial\overline{R}}{\partial g_{\mu\nu,\eta}}+\left(2a_{0}R+\sum_{k=1}^{p}a_{k}\Box^{k}R\right)\frac{\partial R}{\partial g_{\mu\nu,\eta}}+\sum_{k=1}^{p}a_{k}R\frac{\partial\Box^{k}R}{\partial g_{\mu\nu,\eta}}\right]\ ,\\
-\partial_{\lambda}\left(\frac{\partial L}{\partial g_{\mu\nu,\eta\lambda}}\right)&=-\partial_{\lambda}\left(\sqrt{-g}\left[\left(2a_{0}R+\sum_{k=1}^{p}a_{k}\Box^{k}R\right)\frac{\partial R}{\partial g_{\mu\nu,\eta\lambda}}+\sum_{k=1}^{p}a_{k}R\frac{\partial\Box^{k}R}{\partial g_{\mu\nu,\eta\lambda}}\right]\right)\ ,
\end{align}
\begin{multline}\bar{49}
\sum_{m=2}^{n-1}\left(-1\right)^{m}\left(\frac{\partial L}{\partial g_{\mu\nu,\eta i_{0}\cdots i_{m}}}\right)_{,i_{0}\cdots i_{m}}=\sum_{m=2}^{n-1}\sum_{k=1}^{p}\left(-1\right)^{m}\partial_{i_{0}\cdots i_{m}}\left[\sqrt{-g}a_{k}R\frac{\partial\Box^{k}R}{\partial g_{\mu\nu,\eta i_{0}\cdots i_{m}}}\right]\\
=\sum_{k=1}^{p}\sum_{m=2}^{2p+3}\left(-1\right)^{m}\partial_{i_{0}\cdots i_{m}}\left[\sqrt{-g}a_{k}R\frac{\partial\Box^{k}R}{\partial g_{\mu\nu,\eta i_{0}\cdots i_{m}}}\right]\\
=\sum_{k=1}^{p}\sum_{m=2}^{2k+1}\left(-1\right)^{m}\partial_{i_{0}\cdots i_{m}}\left[\sqrt{-g}a_{k}R\frac{\partial\Box^{k}R}{\partial g_{\mu\nu,\eta i_{0}\cdots i_{m}}}\right]\ ,
\end{multline}
where $\lambda=i_{1}$, $n=2p+4$ and
\begin{equation}\label{50}
\frac{\partial\Box^{k}R}{\partial g_{\mu\nu,\eta i_{0}\cdots i_{m}}}=0 \qquad \text{if}\quad m>2k+1\ .
\end{equation}
Then, after algebraic manipulations, one have
\begin{multline}\label{51}
\sum_{j=0}^{n-2}\sum_{m=j+1}^{n-1}\left(-1\right)^{j}\left(\frac{\partial L}{\partial g_{\mu\nu,\eta i_{0}\cdots i_{m}}}\right)_{,i_{0}\cdots i_{j}}\\
=\sum_{h=1}^{p}\sum_{j=0}^{2p+2}\sum_{m=j+1}^{2p+3}\left(-1\right)^{j}\left(\sqrt{-g}a_{h}R\frac{\partial\Box^{h}R}{\partial g_{\mu\nu,\eta i_{0}\cdots i_{m}}}\right)_{,i_{0}\cdots i_{j}}\ .
\end{multline}
Thereby, after observing that $j+1\leq m \leq 2h+1$ $\rightarrow$ $j\leq 2h$, we finally get
\begin{equation*}
\sum_{j=0}^{n-2}\sum_{m=j+1}^{n-1}\left(-1\right)^{j}\left(\frac{\partial L}{\partial g_{\mu\nu,\eta i_{0}\cdots i_{m}}}\right)_{,i_{0}\cdots i_{j}}=\sum_{h=1}^{p}\sum_{j=0}^{2h}\sum_{m=j+1}^{2h+1}\left(-1\right)^{j}\left(\sqrt{-g}a_{h}R\frac{\partial\Box^{h}R}{\partial g_{\mu\nu,\eta i_{0}\cdots i_{m}}}\right)_{,i_{0}\cdots i_{j}}\ .
\end{equation*}
By inserting these expressions into (<ref>), we obtain the gravitational energy–momentum pseudo-tensor for the Lagrangian (<ref>)
\begin{equation}\label{fulltensor}%MDPI: We removed the box foramt in the Equation (87), please confirm.
\begin{split}
\tau_{\alpha}^{\eta}=\tau_{\alpha\vert GR}^{\eta}+&\frac{1}{2\chi\sqrt{-g}}\Biggl\{\sqrt{-g}\left(2a_{0}R+\sum_{k=1}^{p}a_{k}\Box^{k}R\right)\left[\frac{\partial R}{\partial g_{\mu\nu,\eta}}g_{\mu\nu,\alpha}+\frac{\partial R}{\partial g_{\mu\nu,\eta\lambda}}g_{\mu\nu,\lambda\alpha}\right]\\
&-\partial_{\lambda}\left[\sqrt{-g}\left(2a_{0}R+\sum_{k=1}^{p}a_{k}\Box^{k}R\right)\frac{\partial R}{\partial g_{\mu\nu,\eta\lambda}}\right]g_{\mu\nu,\alpha}\\
&+\Theta_{\left[1,+\infty\right[}\left(p\right)\sum_{h=1}^{p}\Biggl\{\sum_{q=0}^{2h+1}\left(-1\right)^{q}\partial_{i_{0}\cdots i_{q}}\biggl[\sqrt{-g}a_{h}R\frac{\partial \Box^{h}R}{\partial g_{\mu\nu,\eta i_{0}\cdots i_{q}}}\biggl]g_{\mu\nu,\alpha}\\
&+\sum_{j=0}^{2h}\sum_{m=j+1}^{2h+1}\left(-1\right)^{j}\partial_{i_{0}\cdots i_{j}}\biggl[\sqrt{-g}a_{h}R\frac{\partial \Box^{h}R}{\partial g_{\mu\nu,\eta i_{0}\cdots i_{m}}}\biggl]g_{\mu\nu,i_{j+1}\cdots i_{m}\alpha}\Biggr\}\\
&-\delta_{\alpha}^{\eta}\left(a_{0}R^{2}+\sum_{k=1}^{p} a_{k}R\Box^{k}R\right)\sqrt{-g}\Biggr\}
\end{split}
\end{equation}
where the notation
\partial_{i_{0}}=\mathbb{I}
$ is the identity operator and $\tau_{\alpha\vert GR}^{\eta}$ indicates the energy–momentum pseudo-tensor of general relativity [41] defined as
\begin{equation}\label{tensoreGR}
\tau_{\alpha\vert GR}^{\eta}=\frac{1}{2\chi}\left(\frac{\partial \overline{R}}{\partial g_{\mu\nu,\eta}}g_{\mu\nu,\alpha}-\delta^{\eta}_{\alpha} \overline{R}\right)\ .
\end{equation}
Given that only $\overline{R}$ contributes to the field equations we can replace scalar density $\sqrt{-g}R$ with $\sqrt{-g}\overline{R}$,which is not a scalar density. This makes the gravitational pseudo-tensor easier to manipulate and for a straightforward generalization of results see in Ref. [35].
An important extension of local Lagrangian (<ref>) to non-local Lagrangian is possible allowing $p \rightarrow \infty$.
Let $D^{p}$ be a linear differential operator defined by
\begin{equation}\label{52}
\end{equation}
If the weak or strong convergence is guaranteed under suitable assumptions for the coefficients $a_{k}$ (e.g. $\sum_{k=0}^{\infty}\vert a_{k} \vert<\infty$ ) and for the domain of the operator $ D^{p}$, we obtain the following non-local operator $F\left(\Box\right)$
\begin{equation}
\lim_{p\rightarrow\infty}\sum_{k=0}^{p}a_{k}\Box^{k}=F\left(\Box\right)\,
\end{equation}
and also our local action becomes non local, i.e.
\begin{equation}
\end{equation}
Accordingly integral operator acts as
\begin{equation}
\Phi\left(x\right)=\int_{\Omega}d^{4}yF\left(x-y\right)R\left(x\right)=F\left(\Box\right)R\left(x\right)\ .
\end{equation}
Let us carry out now the limit $n\rightarrow \infty$ for the energy–momentum pseudo-tensor of $n$-order Lagrangian (<ref>), we may obtain the non-local pseudo-tensor, that is
\begin{equation}\label{limitinftens}
\lim_{n\rightarrow\infty}\tau_{\alpha}^{\eta}\left(x\right)=\overline{\tau}_{\alpha}^{\eta}\left(x\right)\ .
\end{equation}
Whereas $\tau_{\alpha}^{\eta}\left(x\right)$ transforms as an affine tensor, we could show that also its limit for $n\rightarrow\infty$, i.e., $\overline{\tau}_{\alpha}^{\eta}\left(x\right)$, is an affine tensor. For an linear transformation
\begin{equation}
x^{\prime\mu}=\Lambda^{\mu}_{\nu}x^{\nu}\qquad \vert \Lambda \vert \neq 0
\end{equation}
the following affine pseudo-tensor changes as
\begin{equation}\label{trasfafftens}
\tau^{\eta}_{\alpha}\left(x\right)=\Lambda^{-1\eta}_{\ \ \ \sigma}\Lambda^{\tau}_{\alpha}\tau^{\prime\sigma}_{\tau}\left(x^{\prime}\right)\,.
\end{equation}
Substituting (<ref>) in (<ref>), we have
\begin{equation}
\overline{\tau}_{\alpha}^{\eta}\left(x\right)=\lim_{n\rightarrow\infty}\Lambda^{-1\eta}_{\ \ \ \sigma}\Lambda^{\tau}_{\alpha}\tau^{\prime\sigma}_{\tau}\left(x^{\prime}\right)=\Lambda^{-1\eta}_{\ \ \ \sigma}\Lambda^{\tau}_{\alpha}\lim_{n\rightarrow\infty}\tau^{\prime\sigma}_{\tau}\left(x^{\prime}\right)=\Lambda^{-1\eta}_{\ \ \ \sigma}\Lambda^{\tau}_{\alpha}\overline{\tau}^{\prime\sigma}_{\tau}\left(x^{\prime}\right)
\end{equation}
which implies that $\overline{\tau}^{\sigma}_{\tau}\left(x\right)$ transforms as an affine object also in the limit $n\rightarrow\infty$.
§.§ The weak-field limit of energy–momentum pseudo-tensor
The gravitational energy–momentum pseudo-tensor (<ref>) related to Lagrangian (<ref>) in weak field approximation can be performed perturbing weakly spacetime metric around the Minkowski metric $\eta_{\mu\nu}$ as
\begin{equation}\label{53}
g_{\mu\nu}=\eta_{\mu\nu}+h_{\mu\nu}\qquad\mbox{being}\quad |h_{\mu\nu}|\ll 1\ ,
\end{equation}
where $h=\eta^{\mu\nu}h_{\mu\nu}$ is the trace of perturbation.
Thus, we expand the energy–momentum pseudo-tensor to lower order in $h$, namely, retaining terms up to $h^2$. Let's see what becomes the weakly perturbed pseudo-tensor (<ref>) in harmonic coordinates where $g^{\mu\nu}\Gamma^{\sigma}_{\mu\nu}=0$. The quadratic part of the Ricci scalar $\overline{R}$ yields
\begin{equation}\label{54}
\overline{R}=-g^{\mu\nu}\left(\Gamma^{\rho}_{\mu\sigma}\Gamma^{\sigma}_{\nu\rho}\right)\ ,
\end{equation}
that is
\begin{equation}\label{55}
\overline{R}=-\frac{1}{4}g^{\mu\nu}g^{\sigma\lambda}g^{\rho\epsilon}\left(g_{\epsilon\mu,\sigma}+g_{\epsilon\sigma,\mu}-g_{\mu\sigma,\epsilon}\right)\left(g_{\lambda\nu,\rho}+g_{\lambda\rho,\nu}-g_{\nu\rho,\lambda}\right)\ .
\end{equation}
Keeping terms up to second order in $h^{2}$, we get
\begin{equation}\label{56}
\left(\frac{\partial\overline{R}}{\partial g_{\alpha\beta,\gamma}}\right)^{\left(1\right)}\left(g_{\alpha\beta,\delta}\right)^{\left(1\right)}\stackrel{h^{2}}=\left(\frac{1}{2}h^{\alpha\beta\ \gamma}_{\ \ ,}h_{\alpha\beta,\delta}-h^{\gamma\alpha\ \beta}_{\ \ ,}h_{\alpha\beta,\delta}\right)\ ,
\end{equation}
according to
\begin{multline}\label{57}
\frac{\partial\overline{R}}{\partial g_{\alpha\beta,\gamma}}g_{\alpha\beta,\delta}=-\frac{1}{4}\biggl\{\left(g^{\mu\beta}g^{\sigma\alpha}g^{\epsilon\gamma}+g^{\mu\gamma}g^{\sigma\alpha}g^{\beta\epsilon}-g^{\mu\alpha}g^{\sigma\gamma}g^{\beta\epsilon}\right)\left(g_{\epsilon\mu,\sigma}+g_{\epsilon\sigma,\mu}-g_{\sigma\mu,\epsilon}\right)\\
+\left(g^{\beta\nu}g^{\gamma\lambda}g^{\rho\alpha}+g^{\gamma\nu}g^{\beta\lambda}g^{\rho\alpha}-g^{\alpha\lambda}g^{\beta\nu}g^{\rho\gamma}\right)\left(g_{\lambda\nu,\rho}+g_{\lambda\rho,\nu}-g_{\nu\rho,\lambda}\right)\biggr\}g_{\alpha\beta,\delta}\ ,
%\end{split}\ ,
\end{multline}
and also
\begin{equation}\label{58}
\overline{R}^{\left(2\right)}=-\frac{1}{4}\left(h^{\sigma\lambda}_{\ \ ,\rho}h_{\lambda\sigma,}^{\ \ \ \rho}-2h^{\sigma\lambda}_{\ \ ,\rho}h^{\rho}_{\ \lambda,\sigma}\right)\ .
\end{equation}
Hence, when we put these terms into (<ref>) , the stress–energy pseudo-tensor in general relativity up to order $h^{2}$ takes the form
\begin{equation}\label{59}
\tau_{\alpha\vert GR}^{\eta}=\frac{1}{2\chi}\left[\frac{1}{2}h^{\mu\nu,\eta}h_{\mu\nu,\alpha}-h^{\eta\mu,\nu}h_{\mu\nu,\alpha}-\frac{1}{4}\delta_{\alpha}^{\eta}\left(h^{\sigma\lambda}_{\ \ ,\rho}h_{\lambda\sigma}^{\ \ ,\rho}-2h^{\sigma\lambda}_{\ \ ,\rho}h^{\rho}_{\ \lambda,\sigma}\right)\right]\ .
\end{equation}
Now, we have to expand to second order in $h$ the corrections of the pseudo-tensor (<ref>) due to extended gravity terms. To lower order in $h$ we consider the following expansions
\begin{multline}\label{60}
\left(\frac{\partial R}{\partial g_{\mu\nu,\eta\lambda}}\right)^{\left(0\right)}=\frac{1}{2}\left(g^{\mu\eta}g^{\nu\lambda}+g^{\mu\lambda}g^{\nu\eta}-2g^{\mu\nu}g^{\eta\lambda}\right)^{\left(0\right)}\\
=\frac{1}{2}\left(\eta^{\mu\eta}\eta^{\nu\lambda}+\eta^{\mu\lambda}\eta^{\nu\eta}-2\eta^{\mu\nu}\eta^{\eta\lambda}\right)\ ,
\end{multline}
\begin{equation}\label{61}
\left(\frac{\partial R}{\partial g_{\mu\nu,\eta\lambda}}\right)^{\left(0\right)}\left(g_{\mu\nu,\lambda\alpha}\right)^{\left(1\right)}=\left(h^{\lambda\eta}_{\ \ ,\lambda\alpha}-h^{,\eta}_{\ \ \alpha}\right)=\left(h^{\lambda\eta}-\eta^{\eta\lambda}h\right)_{,\lambda\alpha}\stackrel{\text{h.g.}}{=}-\frac{1}{2}h^{,\eta}_{\ \ \alpha}\ ,
\end{equation}
\begin{equation}\label{62}
\left(\frac{\partial R}{\partial g_{\mu\nu,\eta\lambda}}\right)^{\left(0\right)}\left(g_{\mu\nu,\alpha}\right)^{\left(1\right)}=\left(h^{\lambda\eta}-\eta^{\eta\lambda}h\right)_{,\alpha}\ ,
\end{equation}
\begin{multline}\label{derivatsupnonsimm}
\left(\frac{\partial\Box^{h}R}{\partial g_{\mu\nu,\eta i_{0}\cdots i_{m}}}\right)^{\left(0\right)}=\left(\frac{\partial\Box^{h}R}{\partial g_{\mu\nu,\eta i_{0}\cdots i_{q}}}\right)^{\left(0\right)}=\left(\frac{\partial\Box^{h}R}{\partial g_{\mu\nu,\eta i_{0}\cdots i_{2h+1}}}\right)^{\left(0\right)}\\=\eta^{i_{2}i_{3}}\cdots \eta^{i_{2h}i_{2h+1}}\left(\eta^{\mu i_{1}}\eta^{\nu\eta}-\eta^{\mu\nu}\eta^{\eta i_{1}}\right)+\cdots\ .
\end{multline}
Then, we take into account only the terms up to $h^{2}$ in harmonic gauge, as
\begin{equation}\label{primoterm}
\left(2a_{0}R+\sum_{k=1}^{p}a_{k}\Box^{k}R\right)\frac{\partial R}{g_{\mu\nu,\eta\lambda}}g_{\mu\nu,\lambda\alpha}\stackrel{h^{2}}{\stackrel{\text{h.g.}}{=}}\frac{1}{4}\left(\sum_{k=0}^{p}a_{k}\Box^{k+1}h\right)h^{,\eta}_{\ \ \alpha}+\frac{1}{4}a_{0}h^{,\eta}_{\ \ \alpha}\Box h\ ,
\end{equation}
\begin{multline}\label{secondoterm}
-\partial_{\lambda}\left[\sqrt{-g}\left(2a_{0}R+\sum_{k=1}^{p}a_{k}\Box^{k}R\right)\frac{\partial R}{\partial g_{\mu\nu,\eta\lambda}}\right]g_{\mu\nu,\alpha}\stackrel{h^{2}}{\stackrel{\text{h.g.}}{=}}a_{0}\Box h_{,\lambda}\left(h^{\lambda\eta}-\eta^{\eta\lambda}h\right)_{,\alpha}\\
+\frac{1}{2}\sum_{k=1}^{p}a_{k}\Box^{k+1}h_{,\lambda}\left(h^{\lambda\eta}-\eta^{\lambda\eta}h\right)_{,\alpha}\ ,
\end{multline}
\begin{multline}\label{formuladermax}
\sum_{h=1}^{p}\sum_{q=0}^{2h+1}\left(-1\right)^{q}\partial_{i_{0}\cdots i_{q}}\left[\sqrt{-g}a_{h}R\frac{\partial\Box^{h}R}{\partial g_{\mu\nu,\eta i_{0}\cdots i_{q}}}\right]g_{\mu\nu,\alpha}\\\stackrel{h^{2}}{\stackrel{\text{h.g.}}{=}}\frac{1}{2}\sum_{h=1}^{p}a_{h}\Box^{h+1}h_{,\lambda}\left(h^{\eta\lambda}-\eta^{\eta\lambda}h\right)_{,\alpha}+\left(A_{p}\right)_{\alpha}^{\eta}\ ,
\end{multline}
\begin{multline}\label{formuladerivatesup}
\sum_{h=1}^{p}\sum_{j=0}^{2h}\sum_{m=j+1}^{2h+1}\left(-1\right)^{j}\partial_{i_{0}\cdots i_{j}}\left[\sqrt{-g}a_{h}R\frac{\partial\Box^{h}R}{\partial g_{\mu\nu,\eta i_{0}\cdots i_{m}}}\right]g_{\mu\nu,i_{j+1}\cdots i_{m}\alpha}\stackrel{h^{2}}{\stackrel{\text{h.g.}}{=}}\frac{1}{4}\sum_{h=1}^{p}a_{h}\Box h \Box^{h} h^{,\eta}_{\ \ \alpha}\\
+\frac{1}{2}\sum_{h=0}^{1}\sum_{j=h}^{p-1+h}\sum_{m=j+1-h}^{p}\left(-1\right)^{h}a_{m}\Box^{m-j}\left(h^{\eta\lambda}-\eta^{\eta\lambda}h\right)_{,i_{h}\alpha}\Box^{j+1-h}h_{,\lambda}^{\ \ i_{h}}+\left(B_{p}\right)_{\alpha}^{\eta}\ .
\end{multline}
In Eqs. (<ref>), (<ref>) and (<ref>), we have disregarded the index permutations ($\mu\nu$) and $\left(\eta i_{1}\cdots i_{2h+1}\right)$ because $\left(A_{p}\right)_{\alpha}^{\eta}$ and $\left(B_{p}\right)_{\alpha}^{\eta}$ terms, averaged on a suitable spacetime region, vanish, according to Appendix (<ref>). Hence we calculated only the term deriving from (<ref>) without considering the index permutations ($\mu\nu$) and $\left(\eta i_{1}\cdots i_{2h+1}\right)$. This because, taking into account terms obtained from permutations in $\left(A_{p}\right)_{\alpha}^{\eta}$ and $\left(B_{p}\right)_{\alpha}^{\eta}$, averaged on a suitable spacetime region, we obtain that are equal to zero as we will see below in Appendix <ref>. This mathematical trick is essential to calculated the averaged gravitational energy–momentum pseudo-tensor and the power emitted by a source.
So, by inserting equalities (<ref>), (<ref>), (<ref>) and (<ref>) into (<ref>), we find the extra term of pseudo-tensor $\tau^{\eta}_{\alpha}$ to second order owing to extension of general relativity , that we call $\tilde{\tau}^{\eta}_{\alpha}$, that is
\begin{multline}\label{total}
\tilde{\tau}_{\alpha}^{\eta}\stackrel{h^{2}}{=}\frac{1}{2\chi}\Biggl\{\frac{1}{4}\left(\sum_{k=0}^{p}a_{k}\Box^{k+1}h\right)h^{,\eta}_{\ \ \alpha}+\frac{1}{2}\sum_{t=0}^{p}a_{t}\Box^{t+1}h_{,\lambda}\left(h^{\eta\lambda}-\eta^{\eta\lambda}h\right)_{,\alpha}\\
+\frac{1}{2}\sum_{h=0}^{1}\sum_{j=h}^{p}\sum_{m=j}^{p}\left(-1\right)^{h}a_{m}\Box^{m-j}\left(h^{\eta\lambda}-\eta^{\eta\lambda}h\right)_{,\alpha i_{h}}\Box^{j+1-h}h_{,\lambda}^{\ \ i_{h}}\\
+\frac{1}{4}\sum_{l=0}^{p} a_{l}\Box^{l}\left(h^{,\eta}_{\ \ \alpha}-\Box h\delta_{\alpha}^{\eta}\right)\Box h+\Theta_{\left[1,+\infty\right[}\left(p\right)\left[\left(A_{p}\right)_{\alpha}^{\eta}+\left(B_{p}\right)_{\alpha}^{\eta}\right]\Biggr\}\ ,
\end{multline}
where conventions used are
\begin{equation*}
\left(\right)_{,\alpha i_{0}}=\left(\right)_{,\alpha} \qquad h_{,\lambda}^{\ \ i_{0}}=h_{,\lambda}\ .
\end{equation*}
In summary, we can split the gravitational energy–momentum pseudo-tensor in the general relativity part and in the Extended Gravity part, that is
\begin{equation}\label{63}
\tau_{\alpha}^{\eta}\stackrel{h^{2}}=\tau_{\alpha\vert GR}^{\eta}+\tilde{\tau}_{\alpha}^{\eta}\ .
\end{equation}
Now in the particular case when $p$ is equal to $0$ and $1$, extended corrections of the pseudo-tensor $\tilde{\tau}^{\eta}_{\alpha}$ was derived. Then, for $p=0$, that is, $L_{g}=\left(\overline{R}+a_{0}R^{2}\right)\sqrt{-g}$ as in the case discussed in [35], we obtain
\begin{equation*}
\tau_{\alpha}^{\eta}\stackrel{h^{2}}=\tau_{\alpha\vert GR}^{\eta}+\tilde{\tau}_{\alpha}^{\eta}\ ,
\end{equation*}
\begin{equation}\label{64}
\tilde{\tau}_{\alpha}^{\eta}\stackrel{h^{2}}=\frac{a_{0}}{2\chi}\left(\frac{1}{2}h^{,\eta}_{\ \ \alpha}\Box h+ h^{\eta}_{\ \lambda,\alpha}\Box h^{,\lambda}-h_{,\alpha}\Box h^{,\eta}-\frac{1}{4}\left(\Box h\right)^{2}\delta_{\alpha}^{\eta}\right)\ .
\end{equation}
While for $p=1$, that is $L_{g}=\left(\overline{R}+a_{0}R^{2}+a_{1}R\Box R\right)\sqrt{-g}$, one has
\begin{equation*}
\tau_{\alpha}^{\eta}\stackrel{h^{2}}=\tau_{\alpha\vert GR}^{\eta}+\tilde{\tau}_{\alpha}^{\eta}\ ,
\end{equation*}
where extended corrections to pseudo-tensor are
\begin{multline}\label{65}
\tilde{\tau}_{\alpha}^{\eta}\stackrel{h^{2}}=\frac{1}{2\chi}\Biggl\{\frac{1}{4}\left(2a_{0}\Box h+a_{1}\Box^{2}h\right)h^{,\eta}_{\ \ \alpha}+\frac{1}{2 }\left(2a_{0}\Box h_{,\lambda}+a_{1}\Box^{2}h_{,\lambda}\right)\left(h^{\eta\lambda}-\eta^{\eta\lambda}h\right)_{,\alpha}\\
+\frac{1}{2}a_{1}\Box\left(h^{\eta\lambda}-\eta^{\eta\lambda}h\right)_{,\alpha}\Box h_{,\lambda}+\frac{1}{2}a_{1}\left(h^{\eta\lambda}-\eta^{\eta\lambda}h\right)_{,\alpha}\Box^{2}h_{,\lambda}-\frac{1}{2}a_{1}\left(h^{\eta\lambda}-\eta^{\eta\lambda}h\right)_{,\sigma\alpha}\Box h_{,\lambda}^{\ \ \sigma}\\
+\frac{1}{4}a_{1}\Box h^{,\eta}_{\ \ \alpha}\Box h-\frac{1}{4}\delta_{\alpha}^{\eta}\left[a_{0}\left(\Box h\right)+a_{1}\left(\Box^{2} h\right)\right]\Box h+\left(A_{1}\right)_{\alpha}^{\eta}+\left(B_{1}\right)_{\alpha}^{\eta}\Biggr\}\ .
\end{multline}
The iteration can be performed to every $p$ introducing new contributions into dynamics.
§ POWER EMITTED CARRIED BY A GRAVITATIONAL WAVE
We wish to calculate the power emitted in the form of gravitational waves by an isolated massive system considering the local conservation of the energy–momentum pseudo-tensor (<ref>).
§.§ The average of the energy–momentum pseudo-tensor
Let us now regard the wavelike solutions of the linearized field equations in vacuum associated with Lagrangian (<ref>), for details see Ref. [52]. Gravitational waves solutions can be expressed as
\begin{equation}
\label{wave}
h_{\mu\nu}\left(x\right)=\sum_{m=1}^{p+2}\int_{\Omega}\frac{d^{3}\mathbf{k}}{\left(2\pi\right)^{3}}\left(B_{m}\right)_{\mu\nu}\left(\mathbf{k}\right)e^{i\left(k_{m}\right)_{\alpha}x^{\alpha}}+c.c.\ ,
\end{equation}
\begin{equation}\label{66}
\left(B_{m}\right)_{\mu\nu}\left(\mathbf{k}\right)=
\begin{cases}
C_{\mu\nu}\left(\mathbf{k}\right)& \quad \text{for}\quad m=1 \\
\frac{1}{3}\left[\frac{\eta_{\mu\nu}}{2}+\frac{\left(k_{m}\right)_{\mu}\left(k_{m}\right)_{\nu}}{k_{\left(m\right)}^{2}}\right]\text{A}_{m}\left(\mathbf{k}\right)&\quad \text{for} \quad m\geq2
\end{cases}\ ,
\end{equation}
with $C_{\mu\nu}\left(\mathbf{k}\right)$ related to transverse-traceless polarization tensor typical of general relativity and $\text{A}_{m}\left(\mathbf{k}\right)$ the amplitude of wave at $\mathbf{k}$ fixed.
Here "c.c." stands for the complex conjugate.
The trace of tensor (<ref>) is
\begin{equation}\label{67}
\left(B_{m}\right)_{\lambda}^{\lambda}\left(\mathbf{k}\right)=
\begin{cases}
C_{\lambda}^{\lambda}\left(\mathbf{k}\right)&\quad \text{for}\quad m=1 \\
\text{A}_{m}\left(\mathbf{k}\right)&\quad \text{for} \quad m\geq2
\end{cases}\ ,
\end{equation}
and the $k_{m}^{\mu}=\left(\omega_{m}, \mathbf{k}\right)$ is the wave vector with $k_{m}^{2}=\omega_{m}^{2}-\vert \mathbf{k} \vert ^{2}=\text{M}^{2}$ where $k_{1}^{2}=0$ and $k_{m}^{2}\neq 0$ for $m\geq 2$.
Keeping $\mathbf{k}$ fixed, we derive the following relations
\begin{align}\label{68}
h^{\ \ \eta}_{,\alpha}=&2Re\left\{\sum_{j=1}^{p+2}\left(-1\right)\left(k_{j}\right)_{\alpha}\left(k_{j}\right)^{\eta}A_{j}e^{i k_{j}x}\right\}\ ,\\
\Box^{m}h_{,\lambda}=&2 Re\left\{\left(-1\right)^{m}i\sum_{j=1}^{p+2}\left(k_{j}\right)_{\lambda}\left(k_{j}^{2}\right)^{m}A_{j}e^{ik_{j}x}\right\} \ ,\\
\Box^{q}\left(h^{\eta\lambda}-\eta^{\eta\lambda}h\right)_{,\alpha}=&2Re\left\{\left(-1\right)^{q}i\sum_{l=1}^{p+2}\left(k_{l}\right)_{\alpha}\left(k_{l}^{2}\right)^{q}\left[\left(B_{l}\right)^{\eta\lambda}-\eta^{\eta\lambda}\left(B_{l}\right)_{\rho}^{\rho}\right]e^{ik_{l}x}\right\}\ ,\\
\Box^{m}h_{,\lambda}^{\ \ \sigma}=&2 Re\left\{\left(-1\right)^{m+1}\sum_{j=1}^{p+2}\left(k_{j}\right)_{\lambda}\left(k_{j}\right)^{\sigma}\left(k_{j}^{2}\right)^{m}A_{j}e^{ik_{j}x}\right\}\ , \\
\Box^{q}\left(h^{\eta\lambda}-\eta^{\eta\lambda}h\right)_{,\sigma\alpha}=&2Re\left\{\left(-1\right)^{q+1}\sum_{l=1}^{p+2}\left(k_{l}\right)_{\sigma}\left(k_{l}\right)_{\alpha}\left(k_{l}^{2}\right)^{q}\left[\left(B_{l}\right)^{\eta\lambda}-\eta^{\eta\lambda}\left(B_{l}\right)_{\rho}^{\rho}\right]e^{ik_{l}x}\right\}\ , \\
\Box^{n}h=&2Re\left\{\left(-1\right)^{n}\sum_{r=2}^{p+2}\left(k_{r}^{2}\right)^{n}A_{r}e^{ik_{r}x}\right\}\ .
\end{align}
Now, we choose a domain of the spacetime $\Omega$ such that $\vert \Omega \vert \gg \frac{1}{\vert k\vert}$ [32]. Then, we can perform the average of the gravitational energy–momentum pseudo-tensor $\tau_{\alpha}^{\eta}$ over our region and all integrals, including terms such as $e^{i\left(k_{i}-k_{j}\right)_{\alpha}x^{\alpha}}$, tend to zero, by means of following identities
\begin{equation}\label{69}
Re\{f\}Re\{g\}=\frac{1}{2}Re\{fg\}+\frac{1}{2}Re\{f\bar{g}\}\ ,
\end{equation}
\begin{equation}\label{70}
\left(k_{l}\right)_{\lambda}\left[\left(B_{l}\right)^{\eta\lambda}-\eta^{\eta\lambda}\left(B_{l}\right)_{\rho}^{\rho}\right]=-\frac{\left(k_{l}\right)^{\eta}}{2}A_{l}\ .
\end{equation}
In the harmonic gauge, after averaging and some algebraic manipulations, we find (see Appendix <ref>)
\begin{align}\label{medievarie}
\left\langle\Box^{m}h_{,\lambda}\Box^{q}\left(h^{\eta\lambda}-\eta^{\eta\lambda}h\right)_{,\alpha}\right\rangle=&\left(-1\right)^{m+q+1}\sum_{l=2}^{p+2}\left(k_{l}\right)_{\alpha}\left(k_{l}\right)^{\eta} \left(k_{l}^{2}\right)^{\left(m+q\right)}\vert A_{l}\vert^{2}\ , \nonumber\\
\left\langle\Box^{m}h_{,\lambda}^{\ \sigma}\Box^{q}\left(h^{\eta\lambda}-\eta^{\eta\lambda}h\right)_{,\sigma\alpha}\right\rangle=&\left(-1\right)^{m+q+1}\sum_{l=2}^{p+2}\left(k_{l}\right)_{\alpha}\left(k_{l}\right)^{\eta} \left(k_{l}^{2}\right)^{\left(m+q\right)+1}\vert A_{l}\vert^{2}\ ,\nonumber\\
\left\langle \Box^{q}h_{\ \alpha}^{,\eta}\Box^{m}h\right\rangle=&2\left(-1\right)^{m+q+1}\sum_{r=2}^{p+2}\left(k_{r}\right)_{\alpha}\left(k_{r}\right)^{\eta} \left(k_{r}^{2}\right)^{\left(m+q\right)}\vert A_{r}\vert^{2}\ ,\nonumber\\
\left\langle\Box^{m}h\Box h\right\rangle=&2\left(-1\right)^{m+1}\sum_{j=2}^{p+2}\left(k_{j}^{2}\right)^{m+1}\vert A_{j}\vert^{2}\ ,\nonumber\\
\langle\left(A_{p}\right)_{\alpha}^{\eta}\rangle=&\langle\left(B_{p}\right)_{\alpha}^{\eta}\rangle=0\ .
\end{align}
A set of polarization tensors forming a basis for the linearized solutions $h_{\mu\nu}$ is given in Appendix <ref>.
According to equalities (<ref>), we can calculate the average value of the energy–momentum pseudo-tensor as
\begin{multline}\label{MEMT}
\left\langle\tau_{\alpha}^{\eta}\right\rangle=\frac{1}{2\chi}\left[\left(k_{1}\right)^{\eta}\left(k_{1}\right)_{\alpha}\left(C^{\mu\nu}C_{\mu\nu}^{*}-\frac{1}{2}\vert C_{\lambda}^{\lambda}\vert^{2}\right)\right]\\
+\frac{1}{2\chi}\left[\left(-\frac{1}{6}\right)\sum_{j=2}^{p+2}\left(\left(k_{j}\right)^{\eta}\left(k_{j}\right)_{\alpha}-\frac{1}{2}k_{j}^{2}\delta_{\alpha}^{\eta}\right)\vert A_{j}\vert^{2}\right]\\
+\frac{1}{2\chi}\Biggl\{\Biggl[\sum_{l=0}^{p}\left(l+2\right)\left(-1\right)^{l}a_{l}\sum_{j=2}^{p+2}\left(k_{j}\right)^{\eta}\left(k_{j}\right)_{\alpha}\left(k_{j}^{2}\right)^{l+1}\vert A_{j}\vert^{2}\Biggr]\\
-\frac{1}{2}\sum_{l=0}^{p}\left(-1\right)^{l}a_{l}\sum_{j=2}^{p+2}\left(k_{j}^{2}\right)^{l+2}\vert A_{j}\vert^{2}\delta_{\alpha}^{\eta}\Biggr\}\ ,
\end{multline}
with gravitational coupling ${\chi=\frac{8\pi G}{c^{4}}}$.
In TT gauge for the first mode associated with $k_{1}$ and only in harmonic gauge for residual modes $k_{m}$, in the momentum space, it gets
\begin{equation}\label{71}
\begin{cases}
\left(k_{1}\right)_{\mu}C^{\mu\nu}=0 \quad \land \quad C_{\lambda}^{\lambda}=0&\quad \text{if}\quad m=1\\
\left(k_{m}\right)_{\mu}\left(B_{m}\right)^{\mu\nu}=\frac{1}{2}\left(B_{m}\right)_{\lambda}^{\lambda}k^{\nu}& \quad \text{if} \quad m\geq2
\end{cases}\ .
\end{equation}
We now explore a gravitational wave propagating in the $+z$-direction at $\mathbf{k}$ fixed, with 4-wave vector given by $k^{\mu}=\left(\omega,0,0,k_{z}\right)$ where $\omega_{1}^{2}=k_{z}^{2}$ if $k_{1}^{2}=0$ and $ k_{m}^{2}=m^{2}=\omega_{m}^{2}-k_{z}^{2}$ otherwise with $k_{z}>0$. Accordingly the averaged time-space tensorial component which can be seen as flux of gravitational energy along the $z$ axis through the surface that delimits our domain $\Omega$, reads
\begin{multline}\label{72}
\left\langle\tau_{0}^{3}\right\rangle=\frac{c^{4}}{8\pi G}\omega_{1}^{2}\left(C_{11}^{2}+C_{12}^{2}\right)+\frac{c^{4}}{16\pi G}\Biggl[\left(-\frac{1}{6}\right)\sum_{j=2}^{p+2}\omega_{j}k_{z}\vert A_{j}\vert^{2}\\
+\sum_{l=0}^{p}\left(l+2\right)\left(-1\right)^{l}a_{l}\sum_{j=2}^{p+2}\omega_{j}k_{z}m_{j}^{2\left(l+1\right)}\vert A_{j}\vert^{2}\Biggr]\ .
\end{multline}
Finally, we can calculate the emitted power per unit solid angle $\Omega$, radiated by the localized sources, in a direction $\hat{x}$ at $\mathbf{k}$ fixed. By choosing of the suitable gauge, for the local conservation of the energy–momentum pseudo-tensor (<ref>), the power is given by
\begin{equation}\label{73}
\frac{dP}{d\Omega}=r^2\hat{x}^{i}\left\langle\tau_{0}^{i}\right\rangle\ .
\end{equation}
By ranging the index $p$ of the pseudo-tensor (<ref>) over $\{0,1,2\}$, we obtain the following three cases
for p=0
\begin{gather}\label{74}
\left\langle\tau_{0}^{3}\right\rangle=\frac{c^4\omega_{1}^{2}}{8\pi G}\left[C_{11}^{2}+C_{12}^{2}\right]+\frac{c^{4}}{16\pi G}\biggl\{\left(-\frac{1}{6}\right)\omega_{2}\vert A_{2}\vert^{2}k_{z}+2a_{0}\omega_{2}m_{2}^{2}\vert A_{2}\vert^{2}k_{z}\biggr\}\ ,
\end{gather}
for p=1
\begin{multline}\label{75}
\left\langle\tau_{0}^{3}\right\rangle=\frac{c^4\omega_{1}^{2}}{8\pi G}\left[C_{11}^{2}+C_{12}^{2}\right]+\frac{c^{4}}{16\pi G}\biggl\{\left(-\frac{1}{6}\right)\left(\omega_{2}\vert A_{2}\vert^{2}+\omega_{3}\vert A_{3}\vert^{3}\right)k_{z}\\
+2a_{0}\left[\left(\omega_{2}m_{2}^{2}\vert A_{2}\vert^{2}+
\omega_{3}m_{3}^{2}\vert A_{3}\vert^{2}\vert^{2}\right)k_{z}\right] -3a_{1}\left[\left(\omega_{2}m_{2}^{4}\vert A_{2}\vert^{2}+\omega_{3}m_{3}^{4}\vert A_{3}\vert^{2}\right)k_{z}\right]\biggr\}\ ,
\end{multline}
and for p=2
\begin{multline}\label{76}
\left\langle\tau_{0}^{3}\right\rangle=\frac{c^4\omega_{1}^{2}}{8\pi G}\left[C_{11}^{2}+C_{12}^{2}\right]+\frac{c^{4}}{16\pi G}\biggl\{\left(-\frac{1}{6}\right)\left(\omega_{2}\vert A_{2}\vert^{2}+\omega_{3}\vert A_{3}\vert^{3}+\omega_{4}\vert A_{4}\vert^{2}\right)k_{z}\\+2a_{0}\left[\left(\omega_{2}m_{2}^{2}\vert A_{2}\vert^{2}+\omega_{3}m_{3}^{2}\vert A_{3}\vert^{2}+\omega_{4}m_{4}^{2}\vert A_{4}\vert^{2}\right)k_{z}\right]\\
-3a_{1}\left[\left(\omega_{2}m_{2}^{4}\vert A_{2}\vert^{2}+\omega_{3}m_{3}^{4}\vert A_{3}\vert^{2}+\omega_{4}m_{4}^{4}\vert A_{4}\vert^{2}\right)k_{z}\right]\\
+4a_{2}\left[\left(\omega_{2}m_{2}^{6}\vert A_{2}\vert^{2}+\omega_{3}m_{3}^{6}\vert A_{3}\vert^{2}+\omega_{4}m_{4}^{6}\vert A_{4}\vert^{2}\right)\right]\biggr\}\ ,
\end{multline}
where the gravitational coupling $\chi$ has been explicitly indicated. By formulas (<ref>), (<ref>) and (<ref>) it is obvious that the first term comes out of general relativity and the corrections strongly depends on $p$. In any context where corrections to general relativity can be investigated, this approach could constitute a paradigm to search for higher order effects.
§ ENERGY–MOMENTUM COMPLEX OF $F({\CAL R})$ GRAVITY IN PALATINI APPROACH.
§.§ The gravitational pseudo-tensor of $f({\cal R})$ gravity in Palatini formulation.
In Palatini approach the metric tensor $g_{\mu\nu}$ and the connection $\Gamma^\alpha_{\mu\nu}$ are independent, that means that we do not assume any relation between the metric and the connection, and Riemann and Ricci tensors are, in general, defined as
\begin{align}\label{78}
{\cal R}_{\mu\nu}(\Gamma)=& \partial_\alpha\Gamma^\alpha_{\mu\nu}-\partial_\nu\Gamma^\alpha_{\mu\alpha}+\Gamma^\alpha_{\mu\nu}\,\Gamma^\sigma_{\alpha\sigma} -\Gamma^\alpha_{\nu\lambda} \, \Gamma^\lambda_{\mu\alpha},
\\
{\cal R}(g,\Gamma) =& {\cal R}_{\mu\nu}(\Gamma)\, g^{\mu \nu}.
\end{align}
So, the Palatini gravitational action of $f({\cal R})$ appears as [53]
\begin{align} \label{actionf}
{\cal S} = \frac{1}{2 \kappa^2} \int {\rm d}^4x \,\sqrt{-g}\, f({\cal R}),
\end{align}
with the coupling $\kappa^2=8\pi G /c^4$ and $g$ the determinant of metric tensor $g_{\mu\nu}$. By varying the metric $g^{\mu\nu}$ and the connection $\Gamma^\alpha_{\mu\nu}$, for a general infinitesimal transformation coordinate $x^\mu$ it gets
\begin{align}
x^{\prime \mu} =& x^\mu + \delta x^\mu,
\\
g^{\prime \mu\nu}(x^\prime) =& g^{\mu\nu}(x) + \tilde{\delta}g^{\mu\nu} ,&
g^{\prime \mu\nu}(x) =& g^{\mu\nu}(x) + \delta g^{\mu\nu},
\\
\Gamma^{\prime \alpha}_{\mu\nu}(x^\prime) =& \Gamma^\alpha_{\mu\nu}(x) + \tilde{\delta} \Gamma^\alpha_{\mu\nu},&
\Gamma^{\prime\alpha}_{\mu\nu}(x) =& \Gamma^\alpha_{\mu\nu}(x) + \delta \Gamma^\alpha_{\mu\nu},
\end{align}
where $\tilde{\delta}$ is the local variation and $\delta$ is the variation that keeps the coordinates $x$ fixed.
The variation of the gravitational action with respect to the metric $g^{\mu\nu}$ and the connection $\Gamma^{\alpha}_{\beta\gamma}$ yield
\begin{multline}\label{79}
\tilde{\delta} {\cal S} = \frac{1}{2\kappa^2} \int {\rm d}^4x \Bigg\{ \sqrt{-g}\biggl[\left(f_{\cal R} {\cal R}_{\mu\nu}-\frac{1}{2} g_{\mu\nu} f\right) \, \delta g^{\mu\nu} \\
+ f_{\cal R} g^{\mu\nu} \, \delta {\cal R}_{\mu\nu}\biggl] +\partial_\mu \left(\sqrt{-g} f \, \delta x^\mu\right) \Bigg\},
\end{multline}
where $f_{\cal R}:={\rm d} f({\cal R})/{\rm d}{\cal R}$. According to the following Palatini identity
\begin{align}\label{80}
\delta {\cal R}_{\mu\nu} = \nabla_\alpha\left(\delta \Gamma^\alpha_{\mu\nu}\right)-\nabla_\nu\left(\delta\Gamma^\alpha_{\alpha\mu}\right).
\end{align}
the action (<ref>) takes the form
\begin{multline} \label{action1}
\tilde{\delta} {\cal S} = \frac{1}{2\kappa^2 } \int {\rm d}^4x \, \Bigg\{ \sqrt{-g}\left(f_{\cal R} {\cal R}_{\mu\nu}-\frac{1}{2} g_{\mu\nu} f\right) \, \delta g^{\mu\nu} \\
+\delta \Gamma^\lambda_{\phantom{\lambda}\nu\mu}\Bigl[-\nabla_\lambda\left(\sqrt{-g}g^{\nu\mu}f_{\cal R}\right)+\nabla_\alpha\left(\sqrt{-g}g^{\mu\alpha} \delta^\nu_\lambda f_{\cal R}\right)\Bigr]
\\
\partial_\lambda\left[\sqrt{-g}f_{\cal R} \left(g^{\mu\nu} \delta^\lambda_\alpha-g^{\mu\lambda} \delta^{\nu}_\alpha \right) \delta \Gamma^\alpha_{\phantom{\alpha}\mu\nu} +\sqrt{-g} f \, \delta x^\lambda \right]
\Bigg\}.
\end{multline}
By the principle of least action or stationary action (<ref>), by imposing that the variation of metric and its derivatives vanish at the boundary, we obtain field equations for the metric tensor and the connection in vacuum, i.e.,
\begin{align}\label{FieldEq1}
f_{\cal R} {\cal R}_{(\mu\nu)}-\frac{1}{2} g_{\mu\nu} f =& 0,
\\ \label{FieldEq2}
\nabla_\lambda\left(\sqrt{-g}g^{\nu\mu}f_{\cal R}\right) =& 0.
\end{align}
Given that we adopting an arbitrary non-compatible connection, the symmetric part of the Ricci tensor, ${\cal R}_{(\mu\nu)}$, enter in the Eq. (<ref>) and then the Ricci tensor is non symmetric, that is
\begin{equation}\label{81}
{\cal R}_{\mu\nu}={\cal R}_{\nu\mu}+{\cal R}^{\lambda}_{\phantom{\lambda}\lambda\mu\nu}\ ,
\end{equation}
being Riemann tensor ${\cal R}^{\sigma}_{\phantom{\sigma}\lambda\mu\nu}$ no longer antisymmetric on its first two indices, i.e., the term ${\cal R}^{\lambda}_{\phantom{\lambda}\lambda\mu\nu}$ does not vanishes. For a generic infinitesimal transformation, the metric tensor and the connection change as
\begin{align}\label{82}
x^{\prime \mu} =& \, x^\mu + \xi^\mu,
\\
g^{\prime \mu\nu}(x^\lambda) \simeq& \, g^{\mu\nu}(x^\lambda) - \xi^\lambda \frac{\partial g^{\mu\nu}}{\partial x^\lambda} + g^{\mu\alpha} \frac{\partial \xi^\nu}{\partial x^\alpha} + g^{\nu\alpha} \frac{\partial \xi^\mu}{\partial x^\alpha},
\\
\Gamma^{\prime \alpha}_{\phantom{\alpha}\mu\nu} (x^\lambda) \simeq& \, \Gamma^{\alpha}_{\phantom{\alpha}\mu\nu} (x^\lambda) - \xi^\lambda \frac{\partial \Gamma^{\alpha}_{\phantom{\alpha}\mu\nu}}{\partial x^\lambda} + \Gamma^{\rho}_{\phantom{\rho}\mu\nu} \frac{\partial \xi^\alpha}{\partial x^\rho} -\Gamma^{\alpha}_{\phantom{\alpha}\sigma\nu} \frac{\partial \xi^\sigma}{\partial x^\mu} -\Gamma^{\alpha}_{\phantom{\alpha}\mu\sigma} \frac{\partial \xi^\sigma}{\partial x^\nu} - \frac{\partial^2 \xi^\alpha}{\partial x^\mu \, \partial x^\nu},
\end{align}
where we have neglected terms of higher order in $\xi^\mu$ in the series expansion.
Under a rigid infinitesimal translation, that is, $\partial_\mu \xi^\nu =0$, we obtain
\begin{align}
g^{\prime \mu\nu}(x^\lambda) \simeq \, & g^{\mu\nu}(x^\lambda) - \xi^\lambda \frac{\partial g^{\mu\nu}}{\partial x^\lambda},
\\
\Gamma^{\prime \alpha}_{\phantom{\alpha}\mu\nu} (x^\lambda) \simeq \, & \Gamma^{\alpha}_{\phantom{\alpha}\mu\nu} (x^\lambda) - \xi^\lambda \frac{\partial \Gamma^{\alpha}_{\phantom{\alpha}\mu\nu}}{\partial x^\lambda} .
\end{align}
Therefore, the Palatini action (<ref>) becomes
\begin{multline}\label{83}
\tilde{\delta} {\cal S}_{\rm g} = \frac{1}{2\kappa^2} \int {\rm d}^4x \Bigg\{ - \sqrt{-g}\left(f_{\cal R} {\cal R}_{\mu\nu}-\frac{1}{2} g_{\mu\nu} f\right) \, \xi^\lambda \, g^{\mu\nu}_{\phantom{\mu\nu},\lambda}\\
-\xi^\lambda \, \Gamma^\beta_{\phantom{\beta}\nu\mu,\lambda}\left[-\nabla_\beta\left(\sqrt{-g}g^{\nu\mu}f_{\cal R}\right)+\nabla_\alpha\left(\sqrt{-g}g^{\mu\alpha} \delta^\nu_\beta f_{\cal R}\right)\right]
\\
\partial_\lambda\left[-\sqrt{-g}f_{\cal R} \left(g^{\mu\nu} \delta^\lambda_\alpha-g^{\mu\lambda} \delta^\nu_\alpha \right) \, \xi^\beta \, \Gamma^\alpha_{\phantom{\alpha}\mu\nu,\beta} +\sqrt{-g} f \, \xi^\lambda \right]
\Bigg\}.
\end{multline}
If the metric $g^{\mu\nu}$ and the Palatini connection $\Gamma^{\alpha}_{\beta\gamma}$ are solution of equations (<ref>) and (<ref>), the stationary of the local variation of the action (<ref>), gives the local conservation of gravitational energy–momentum pseudo-tensor $\tau^\lambda_{\phantom{\lambda}\beta}$ of Palatini $f({\cal R})$ gravity, namely
\begin{align}\label{84}
\partial_\lambda\left(\sqrt{-g} \, \tau^\lambda_{\phantom{\lambda}\beta} \right) = 0,
\end{align}
where $\tau^\lambda_{\phantom{\lambda}\beta}$ is defined as
\begin{align}\label{85}
\tau^\lambda_{\phantom{\lambda}\beta}=\frac{1}{2\kappa^{2}} \left[f\left({\cal R}\right) \, \delta^\lambda_\beta -f_{\cal R}\left({\cal R}\right)\, \left(g^{\mu\nu} \delta^\lambda_\alpha-g^{\mu\lambda} \delta^\nu_\alpha \right) \, \Gamma^\alpha_{\phantom{\alpha}\mu\nu,\beta} \right].
\end{align}
It is worth noting that the pseudo-tensor defined in Eq. (<ref>) has the opposite sign of the one defined above. In order to derive the energy–momentum complex, let us analyze the action containing the matter part, that is
\begin{align}\label{86}
{\cal S}_{\rm m} = \int {\rm d}^4x \, \sqrt{-g}\,{\cal L}_{\rm m}.
\end{align}
Generally, the matter Lagrangian ${\cal L}_{\rm m}$ depends on the connection as, for example, occurs in presence of fermion fields.
Here, we consider only material Lagrangian which does not depend on the affine connection $\Gamma$. Then, the matter energy–momentum tensor is defined as in (<ref>).
Hence, field equations for metric and connection, i.e., Eqs. (<ref>) and (<ref>), in presence of matter yield
\begin{align}\label{Einstein and Palatini Eqs.}
f_{\cal R} {\cal R}_{(\mu\nu)}-\frac{1}{2} g_{\mu\nu} f =& \kappa^2 T_{\mu\nu},
\\
\nabla_\lambda\left(\sqrt{-g}g^{\nu\mu}f_{\cal R}\right) =& 0 \label{bimetric}.
\end{align}
As already pointed out above the connection can be non compatible with the metric $g_{\mu\nu}$, i.e., $\nabla_{\lambda}g_{\mu\nu}\neq 0$. In compact form, we can define a new metric, conformally related to the metric $g_{\mu\nu}$, as
\begin{align}\label{87}
h_{\mu\nu}:=f_{\cal R}g_{\mu\nu}.
\end{align}
so that Eq. (<ref>) becomes
\begin{align}\label{88}
\nabla_\lambda \left(\sqrt{h}h^{\mu\nu}\right)=0.
\end{align}
Thus the Palatini connection $\Gamma^\alpha_{\phantom{\alpha}\mu\nu} $ appears as the Christoffel connection for the new metric $h_{\mu\nu}$, i.e.,
\begin{align}\label{PalatiniConnection}
\Gamma^\alpha_{\phantom{\alpha}\mu\nu}=
\frac{1}{2\, f_{\cal R}\left({\cal R}\right)}g^{\alpha \beta} \left[\partial_\mu\left( f_{\cal R}\left({\cal R}\right) g_{\nu\beta}\right)+\partial_\nu\left( f_{\cal R}\left({\cal R}\right) g_{\mu\beta}\right)-\partial_\beta\left( f_{\cal R}\left({\cal R}\right) g_{\mu\nu}\right)\right]\ .
\end{align}
The Palatini connection $\Gamma^\alpha_{\phantom{\alpha}\mu\nu} $ and Levi–Civita connection $\stackrel{\circ}{\Gamma}{}^\alpha_{\phantom{\alpha}\mu\nu} $ are related as
\begin{equation}\label{89}
\Gamma^\alpha_{\phantom{\alpha}\mu\nu}=\,\stackrel{\circ}{\Gamma^{\alpha}}_{\mu\nu}+\delta^{\alpha}_{\mu}A_{\nu}+\delta^{\alpha}_{\nu}A_{\mu}-g_{\mu\nu}A^{\alpha}\ ,
\end{equation}
where the four-vector $A_{\mu}$ is defined as
\begin{equation}\label{91}
A_{\mu}:=\frac{1}{2f_{\cal R}}\nabla_{\mu}f_{\cal R}\ .
\end{equation}
For $f({\cal R})={\cal R}$, we recover the Christoffel symbols constructed by the metric $g_{\mu\nu}$, that is
\begin{align}\label{92}
\Gamma^\alpha_{\phantom{\alpha}\mu\nu}=\stackrel{\circ}{\Gamma^\alpha}_{\phantom{\alpha}\mu\nu} = \frac{1}{2}g^{\alpha\beta}\left( g_{\beta\mu,\nu}+g_{\beta\nu,\mu}-g_{\mu\nu,\beta}\right)\,
\end{align}
this means that in general relativity no difference results in metric and Palatini formalism. The Ricci tensor ${\cal R}_{\mu\nu}$ in Palatini formalism and that in metric formalism $R_{\mu\nu}$, are related as follows
\begin{multline} \label{Rconf}
{\cal R}_{\mu\nu} = R_{\mu\nu} +\frac{3}{2} \frac{1}{\left(f_{\cal R}({\cal R})\right)^2} \left(\stackrel{\circ}{\nabla}_\mu f_{\cal R}({\cal R}) \right) \, \left(\stackrel{\circ}{\nabla}_\mu f_{\cal R}({\cal R})\right) \\
- \frac{1}{ f_{\cal R}({\cal R})} \left( \stackrel{\circ}{\nabla}_\mu \stackrel{\circ}{\nabla}_\nu-\frac{1}{2}g_{\mu\nu} \stackrel{\circ}{\square}\right) f_{\cal R}({\cal R}),
\end{multline}
where $\stackrel{\circ}{\square} := \stackrel{\circ}{\nabla}{}^\mu \stackrel{\circ}{\nabla}_\mu$ and $\stackrel{\circ}{\nabla}$ denotes the covariant derivative associated with the Levi–Civita connection. Contracting tensorial equality (<ref>) with $g^{\mu\nu}$, we obtain the relation between ${\cal R}$ and $R$, that is, the Ricci scalar in both approach
\begin{align}\label{93}
{\cal R} = R + \frac{3}{2\left(f_{\cal R}({\cal R})\right)^2} \left(\stackrel{\circ}{\nabla}_\mu f_{\cal R}({\cal R}) \right)\; \left(\stackrel{\circ}{\nabla}{ }^\mu f_{\cal R}({\cal R})\right) + \frac{3}{ f_{\cal R}({\cal R})} \stackrel{\circ}{\square} f_{\cal R}({\cal R}).
\end{align}
Adopting the Palatini connection $\Gamma^\alpha_{\phantom{\alpha}\mu\nu}$ (<ref>), the symmetry of Ricci tensor is restored on account of the relation
\begin{equation}\label{94}
\Gamma_{\lambda}=\frac{\partial_{\lambda}{\left(f_{\cal R}^{2}\sqrt{-g}\right)}}{f_{\cal R}^{2}\sqrt{-g}},
\end{equation}
which implies
\begin{equation}\label{95}
{\cal R}_{[\mu\nu]}=\partial_{[\mu}\Gamma_{\nu]}=0.
\end{equation}
Furthermore the connection is non compatible with metric $g_{\mu\nu}$ being
\begin{equation}\label{96}
\nabla_{\lambda}g_{\mu\nu}=-\frac{g_{\mu\nu}}{f_{\cal R}}\nabla_{\lambda}f_{\cal R}.
\end{equation}
Despite this, the covariant derivatives associated with Palatini connection commute each other, as displayed below
\begin{equation}\label{97}
\left[\nabla_{\rho},\nabla_{\lambda}\right]g_{\mu\nu}=0\,.
\end{equation}
Thus, we restore the antisymmetry on the first two indices of Riemann tensor, namely
\begin{equation}\label{99}
{\cal R}_{\mu\nu\lambda\rho}=-{\cal R}_{\nu\mu\lambda\rho}.
\end{equation}
by the definition of Riemann tensor for an arbitrary tensor $J_{\mu\nu}$
\begin{equation}\label{98}
\left[\nabla_{\rho},\nabla_{\lambda}\right]J_{\mu\nu}=-{\cal R}^{\alpha}_{\phantom{\alpha}\mu\rho\lambda}J_{\alpha\nu}-{\cal R}^{\alpha}_{\phantom{\alpha}\nu\rho\lambda}J_{\mu\alpha}.
\end{equation}
In addition, the contracted Bianchi identities are fulfilled, that is
\begin{equation}\label{100}
\nabla_{\mu}\left({\cal R}^{\mu\nu}-\frac{1}{2}g^{\mu\nu}{\cal R}\right)=0.
\end{equation}
According to the Palatini connection Eq. (<ref>) and from the symmetry of energy–momentum tensor $T_{\mu\nu}$, taking into account that for the new metric $h_{\mu\nu}$ we have
\begin{align}\label{101}
\Gamma_{\lambda}=\frac{\partial_{\lambda}{\sqrt{-h}}}{\sqrt{-h}}\ ,
\end{align}
\begin{align}\label{102}
\Gamma_{\mu\nu\lambda}+\Gamma_{\nu\mu\lambda}=\frac{1}{f_{{\cal R}}}\partial_{\lambda}h_{\mu\nu},
\end{align}
so we derive the following useful expression
\begin{align}\label{103}
\sqrt{-h}\nabla_{\sigma}T^{\sigma}_{\phantom{\sigma}\nu}=\partial_{\sigma}\left(\sqrt{-h}T^{\sigma}_{\phantom{\sigma}\nu}\right)-\frac{1}{2f_{\cal R}}T^{\lambda\rho}\partial_{\nu}h_{\lambda\rho}\sqrt{-h}.
\end{align}
Field equations in matter (<ref>) lead to
\begin{align}\label{104}
0=\frac{\sqrt{-h}}{2f_{\cal R}^{2}}T^{\mu\nu}g_{\mu\nu,\beta}\xi^{\beta}+\partial_{\lambda}\left\{\sqrt{-g}\frac{1}{2\kappa^{2}}\left[%\frac{1}{f_{\cal R}\left(\cal R\right)}\,
f\left({\cal R}\right) \, \delta^\lambda_\beta -f_{\cal R}\left(g^{\mu\nu} \delta^\lambda_\alpha-g^{\mu\lambda} \delta^\nu_\alpha \right) \, \Gamma^\alpha_{\phantom{\alpha}\mu\nu,\beta} \right]\xi^{\beta}\right\},
\end{align}
and from Eq. (<ref>), after some algebraic manipulations, we get the following 4-divergence of energy–momentum complex not vanishing
\begin{align}\label{105}
\partial_{\sigma}\left[\sqrt{-g}\left(T^{\sigma}_{\phantom{\sigma}\beta}+t^{\sigma}_{\phantom{\sigma}\beta}\right)\right]=\frac{\sqrt{-h}}{f_{\cal R}^2} \, \nabla_{\lambda}T^{\lambda}_{\phantom{\lambda}\beta}+\frac{2\sqrt{-h}}{f_{\cal R}^{3}}T^{\lambda}_{\phantom{\lambda}\beta} \, \nabla_{\lambda}f_{\cal R}-\frac{\sqrt{-h}}{2f_{\cal R}^{3}}T \, \nabla_{\beta}f_{\cal R}.
\end{align}
From contracted Bianchi identities and the field equations, the following relations are satisfied
\begin{equation}\label{106}
\left[\nabla_{\mu},\nabla_{\nu}\right]\nabla^{\mu}f_{\cal R}={\cal R}^{\alpha}_{\phantom{\alpha}\nu}\nabla_{\alpha}f_{\cal R},
\end{equation}
\begin{equation}\label{107}
\kappa^{2}\nabla_{\mu}T^{\mu}_{\phantom{\mu}\nu}={\cal R}^{\alpha}_{\phantom{\alpha}\nu}\nabla_{\alpha}f_{\cal R}=\left[\nabla_{\mu},\nabla_{\nu}\right]\nabla^{\mu}f_{\cal R}\ .
\end{equation}
The trace of Eqs. (<ref>) gives the so called structural equation of space-time [54], that is
\begin{equation}\label{108}
T=\frac{1}{\kappa^{2}}\left[f_{\cal R}{\cal R}-2f\left({\cal R}\right)\right],
\end{equation}
where $T=T_{\mu\nu}g^{\mu\nu}$. For a given $f({\cal R})$, we can, in principle, solve this equation and get a relation ${\cal R}={\cal R}(T)$. Thanks to Eq. (<ref>), considering $T=0$, the theory reduces to GR with a cosmological constant. Substituting Eqs. (<ref>) and (<ref>) into Eq. (<ref>), we get
\begin{equation} \label{109}
\partial_{\sigma}\left[\sqrt{-g}\left(T^{\sigma}_{\phantom{\sigma}\beta}+\tau^{\sigma}_{\phantom{\sigma}\beta}\right)\right]=-\frac{\sqrt{-g}}{\kappa^{2}}G^{\lambda}_{\phantom{\lambda}\beta}\nabla_{\lambda}f_{\cal R},
\end{equation}
where $G^{\lambda}_{\phantom{\lambda}\beta}$ is the Einstein tensor.
After some algebraic manipulations, we find the following expression
\begin{equation}\label{110}
G^{\lambda}_{\phantom{\lambda}\beta}\nabla_{\lambda}f_{\cal R}=-\kappa^{2}\stackrel{\circ}{\nabla}_{\mu}T^{\mu}_{\phantom{\mu}\beta}\ .
\end{equation}
The right hand side of Eq. (<ref>) vanishes [55, 56, 57] and then, according to Eqs. (<ref>) and (<ref>), the energy–momentum complex for Palatini $f({\cal R})$ gravity, $\mathcal{T}^{\sigma}_{\phantom{\sigma}\beta}$, is locally conserved, namely
\begin{equation}
\partial_{\sigma}\left[\sqrt{-g}\left(T^{\sigma}_{\phantom{\sigma}\beta}+\tau^{\sigma}_{\phantom{\sigma}\beta}\right)\right]=0,
\end{equation}
\begin{equation}\label{111}
\mathcal{T}^{\sigma}_{\phantom{\sigma}\beta}=\sqrt{-g}\left(T^{\sigma}_{\phantom{\sigma}\beta}+\tau^{\sigma}_{\phantom{\sigma}\beta}\right)\ .
\end{equation}
§ COSMOLOGICAL APPLICATIONS BOTH IN PALATINI AND METRIC APPROACH IN $F(R)$ GRAVITY
§.§ Palatini formalism
We consider a flat FLRW spacetime whose metric is
\begin{align}
{\rm d}s^2= -{\rm d}t^2+ a^2(t)\, \left({\rm d}x^2+{\rm d}y^2+{\rm d}z^2\right),
\end{align}
with scale factor $a(t)$ and cosmic time $t$. From the relation (<ref>) and the field equations (<ref>), we obtain
\begin{align}
2\kappa^2 T^0_{\phantom{0}0} = & -f +6f_{\cal R} \left(\dot{H}+H^2\right) +\ddot{f}_{\cal R} -3\frac{\dot{f}_{\cal R}^2}{f_{\cal R}} -3H\dot{f}_{\cal R},
\end{align}
where $H=\dot{a}/a$ is the Hubble parameter and dots stands for derivatives with respect to the cosmic time $t$. The gravitational energy density $\tau^0_{\phantom{0}0}$ is defined as
\begin{align}
2\kappa^2 \tau^0_{\phantom{0}0} = & f -6f_{\cal R} \left(\dot{H}+H^2\right) -3\ddot{f}_{\cal R} +3\frac{\dot{f}_{\cal R}^2}{f_{\cal R}} -3H\dot{f}_{\cal R}.
\end{align}
So, the energy density complex is
\begin{align}
\kappa^2(\tau^0_{\phantom{0}0}+T^0_{\phantom{0}0}) = -\ddot{f}_{\cal R} -3H \dot{f}_{\cal R}.
\end{align}
In general relativity , i.e., $f({\cal R})={\cal R}$, we obtain a null energy density complex
\begin{align}
\left(t^0_{\phantom{0}0}+T^0_{\phantom{0}0}\right)\Big|_{\rm GR}=0\,.
\end{align}
We postulate that perfect fluids including radiation and non-relativistic dust describe the matter and that the components of the energy–momentum tensor are
\begin{align}
\left(T^\mu_{\phantom{\mu}\nu} \right)_{\rm r} =& {\rm diag}\left(-\rho_{\rm r},p_{\rm r},p_{\rm r},p_{\rm r} \right) & \mbox{ with equation of state}& &p_{\rm r}=&\frac{1}{3} \rho_{\rm r},
\\
\left(T^\mu_{\phantom{\mu}\nu} \right)_{\rm m} =& {\rm diag}\left(-\rho_{\rm m},p_{\rm m},p_{\rm m},p_{\rm m} \right), & \mbox{ with equation of state }& &p_{\rm m}=&0,
\end{align}
where $\rho_{\rm i}$ and $p_{\rm i}$ are the energy density and pressure of each fluid component.
From the conservation of energy–momentum tensor, we obtain, respectively,
\begin{align} \label{c1}
\dot{\rho}_{\rm r}+4 H \rho_{\rm r}=&0,
\\ \label{c2}
\dot{\rho}_{\rm m}+3 H \rho_{\rm m}=&0.
\end{align}
Choosing a form for $f({\cal R})$, we can solve the structure equation Eq. (<ref>) and then explicit ${\cal R}$ as a function of $T$. Now, let us assume a polynomial form as $f({\cal R}) = {\cal R}+\alpha {\cal R}^2$, which is a model extensively studied in Palatini formalism, see for example [58, 59]. Thus, the solution of structural equation (<ref>) becomes
\begin{align}
{\cal R} = -\kappa^2 T.
\end{align}
This model implies power law cosmological solutions [60] as
\begin{align} \label{powerlaw}
a(t) = a_0 \, t^m,
\end{align}
where $m>0$ is a real number. From Eqs. (<ref>) and (<ref>), we get
\begin{align}
\rho_{\rm tot}(t) = \rho_{\rm m}(t)+\rho_{\rm r}(t)=\rho_{\rm m0}t^{-3m}+\rho_{\rm r0}t^{-4m}\,
\end{align}
with $\rho_{\rm m0}$ and $\rho_{\rm r0}$ initial values.
Therefore, we obtain the gravitational energy density
\begin{multline}
2 \kappa^2 \, \tau^0_{\phantom{0}0} = \frac{6m(1-m)}{t^2} +\kappa^2 \rho_{\rm m0}t^{-3m} +6m (5-2m) \alpha \kappa^2 \rho_{\rm m0} t^{-3m-2} \\
+\frac{108m^2 \alpha^2 \kappa^4 \rho_{\rm m0}t^{-6m-2}}{1+2m^2\kappa^2\rho_{\rm m0} t^{-3m}},
\end{multline}
and the energy density complex
\begin{align}
\tau^0_{\phantom{0}0}+T^0_{\phantom{0}0} =-6\alpha m \rho_{\rm m0} t^{-(3m+2)}.
\end{align}
The total energy density of gravitational and non-gravitational fields is then
\begin{equation}\label{300}
\sqrt{-g}(\tau^0_{\phantom{0}0}+T^0_{\phantom{0}0})=-6\alpha m \rho_{\rm m0} t^{-2},
\end{equation}
that tends to zero as the inverse square of cosmic time.
§.§ Metric approach
We consider also in this case a flat FLRW spacetime but in metric formalism. We can explicitly write the time-time components of the gravitational energy–momentum $\tau^\mu_{\phantom{\mu}\nu}$ and the matter energy–momentum $T^\mu_{\phantom{\mu}\nu}$, respectively, as follows
\begin{align}
\kappa^2\, \tau^0_{\phantom{0}0} =& \frac{1}{2}f(R)-3\left(H^2+\dot{H}\right)\, f_R(R) +3H\dot{R} \, f_{RR}(R),
\\ \label{f1}
\kappa^2 T^0_{\phantom{0}0} =& - \frac{1}{2}f(R)+3\left(H^2+\dot{H}\right)\, f_R(R) - 3H\dot{R} \, f_{RR}(R).
\end{align}
Subsequently, the total energy of the gravitation and matter vanishes for FLRW spacetime, i.e.
\begin{align}
\tau^0_{\phantom{0}0}+T^0_{\phantom{0}0} =0,
\end{align}
unlike Palatini approach where energy complex does not vanish Eq.(<ref>). Now, we can assume a power-law evolution for matter and radiation fluids such as Eq. (<ref>). We have
\begin{align} \tau^0_{\phantom{0}0} =&\rho_{\rm m}(t)+\rho_{\rm r}(t)\nonumber \\ =&\rho_{\rm m0} t^{-3m}+\rho_{\rm r0}t^{-4m}\,.
\end{align}
The Ricci curvature scalar, in this case, reads
\begin{equation}
R = 12H^2+6\dot{H} = 6m(2m-1) t^{-2}\,
\end{equation}
while the Friedman equation is reduced to
\begin{multline}\label{200}
\frac{f_{RR}\, R^2}{(2m-1)}+\frac{m-1}{2(2m-1)} f_R \, R -\frac{1}{2} f +\kappa^2 \rho_{\rm m0} \left( \frac{R}{6m(2m-1)}\right)^{\frac{3}{2}m}\\
+\kappa^2 \rho_{\rm r0} \left( \frac{R}{6m(2m-1)}\right)^{2m} =0.
\end{multline}
From this equation (<ref>), we get the explicit form of $f(R)$ that shows a power law behaviour, that is
\begin{align} \label{f}
f(R) =& -\frac{4\kappa^2 \rho_{\rm m0} (2m-1)}{12m-11} \left( \frac{R}{6m(2m-1)}\right)^{\frac{3}{2}m} -\frac{2\kappa^2 \rho_{\rm r0} (2m-1)}{10m^2-8m+1} \left( \frac{R}{6m(2m-1)}\right)^{2m}
\nonumber \\
&+C_1 R^{\frac{3}{4}-\frac{m}{4}-\frac{1}{4}\sqrt{m^2+10m+1}} +C_2 R^{\frac{3}{4}-\frac{m}{4}+\frac{1}{4}\sqrt{m^2+10m+1}}.
\end{align}
When $m=2/3$ and $\rho_{\rm r0}/\rho_{\rm m0} \ll 1$, occurs $f(R)\sim R$ and GR is restored.
§ CONCLUSIONS
Attempts to extend the general relativity through corrections to the Hilbert–Einstein Lagrangian, by introducing curvature, torsion and non-metricity invariants, both local and non-local, have increased in recent years. All of this is to address gravitational divergences at ultraviolet and infrared scales, and more generally to deal with cosmological and astrophysical issues such as current and early cosmic acceleration or the structure formation, without introducing exotic components such as dark energy and dark matter. For a detailed discussion on infinite derivative theories, see Ref. [61, 62, 63, 64, 65, 66, 67], while for non-local wavelike solutions, see Ref. [68, 69, 70]. However, most of the main features of general relativity should be retained to obtain self-consistent theories. In particular, a thorough study of the properties of the gravitational energy–momentum pseudo-tensor are indispensable in view of both the foundation and applications of any gravitational theory.
This review is devoted to generalizing the gravitational energy–momentum pseudo-tensor $\tau^{\eta}_{\alpha}$ to general $n^{th}$ order Lagrangian of the form
\begin{equation*}
L=L(g_{\mu\nu}, g_{\mu\nu,i_{1}}, g_{\mu\nu,i_{1}i_{2}},g_{\mu\nu,i_{1}i_{2}i_{3}}, \ldots, g_{\mu\nu,i_{1}i_{2}i_{3}\cdots i_{n}})\ ,
\end{equation*}
showing that in this model gravity a local conservation of energy–momentum complex is fulfilled.
Specifically, we considered Lagrangians such as $L_{g}=(\overline{R}+a_{0}R^{2}+\sum_{k=1}^{p} a_{k}R\Box^{k}R)\sqrt{-g}$ and $L=F(R)$, both in Palatini and metric approach. It has been shown that $\tau^{\eta}_{\alpha}$ is an affine and not covariant object because it changes as a tensor, under linear transformations but not under general coordinate transformations. The pseudo-tensor of higher order gravity has been weakly perturbed up to the order $h^2$, thus obtaining the weak field limit of the gravitational energy–momentum density. After averaging this object over a suitable four-dimensional domain under suitable gauge, by local conservation of pseudo-tensor, the power emitted by a gravitational source was found. Hence, the gravitational wave (<ref>) associated with higher order Lagrangian express, under the chosen gauge for a wave propagating along the $+z$-direction, in terms of six polarization tensors (see Appendix <ref>) reads as
\begin{multline}
\label{GW1}
+\text{A}^{\left(1\right)}\left(t-v_{G_{m}}z\right)\epsilon_{\mu\nu}^{\left(1\right)}+\text{A}^{\left(L\right)}\left(t-v_{G_{m}}z\right)\epsilon_{\mu\nu}^{\left(L\right)}\ ,
\end{multline}
where $v_{G_{m}}$ is the group velocity of the $m_{th}$ massive mode (see also [20, 33]). Thanks to these solutions, it was possible to derive an expression of the power emitted in terms of amplitudes of the waves $\text{A}_{j}\left(\mathbf{k}\right)$, $C_{11}\left(\mathbf{k}\right)$ and $C_{22}\left(\mathbf{k}\right)$, and the free parameters $a_{m}$. Three special cases for $p$ equal to $0,1,$ and $2$ have been shown where the extended corrections to the power are clearly visible. It was given a cosmological application of the pseudo-tensor in $f(R)$ gravity in both Palatini and metric formulation. Therefore, in a flat FLRW spacetime, we have derived that while the energy density complex vanishes in the metric formalism, in general, it does not vanish in the Palatini approach.
The analysis of gravitational waves and gravitational energy–momentum pseudo-tensor
are two indispensable tools for finding the viable theory of gravitation. Indeed, by wavelike solutions of linearized theory of gravity and by the locally conserved pseudo-tensor, it is possible to calculate the emitted power by isolated system. Then, from the local conservation of the energy–momentum complex, it is also possible to take into account the energy–momentum content of the source, which, through a multipole expansion, could also allow us to derive a generalized formula of the quadrupole formula. This procedure could lead us to fix the order of theory [51, 71], to investigate additional polarization states of gravitational wave and to establish the range of the masses $m_{j}$ of modes.
§ ACKNOWLEDGMENTS
S.C., M.C. and G.L. acknowledgment the Istituto Nazionale di Fisica Nucleare (INFN) Sez. di Napoli, Iniziative Specifiche QGSKY, and the Istituto Nazionale di Alta Matematica (INdAM), gruppo GNFM, for the support.
§ APPENDIX
§.§ The average of $\langle\left(A_{p}\right)_{\alpha}^{\eta}\rangle$ and $\langle\left(B_{p}\right)_{\alpha}^{\eta}\rangle$ terms
Let us now demonstrate the last two relations in (<ref>), that is $\langle\left(A_{p}\right)_{\alpha}^{\eta}\rangle=\langle\left(B_{p}\right)_{\alpha}^{\eta}\rangle=0$.
The general formula for $\Box^{h} R$ -derivative, according to symmetries of $g_{\mu\nu}$ and its derivatives, is [71]:
\begin{multline}\label{derivordsupsimm}
\frac{\partial \Box^{h}R}{\partial g_{\mu\nu,\eta i_{1}\cdots i_{2h+1}}}=g^{j_{2}j_{3}}\cdots g^{j_{2h}j_{2h+1}}g^{ab}g^{cd}\biggl\{\delta_{a}^{(\mu}\delta_{d}^{\nu)}\delta_{c}^{(\eta}\delta_{b}^{i_{1}}\delta_{j_{2}}^{i_{2}}\cdots\delta_{j_{2h}}^{i_{2h}}\delta_{j_{2h+1}}^{i_{2h+1})}\\
\end{multline}
We have to verify that $\langle \left(B_{p}\right)_{\alpha}^{\eta}\rangle=0$ holds. Inserting Eq. (<ref>) in the l.h.s. of Eq. (<ref>) that, in the weak field limit up to the order $h^{2}$ becomes
\begin{multline}\label{bassen}
\sum_{h=1}^{p}\sum_{j=0}^{2h}\sum_{m=j+1}^{2h+1}\left(-1\right)^{j}\partial_{i_{0}\cdots i_{j}}\left[\sqrt{-g}a_{h}R\frac{\partial\Box^{h}R}{\partial g_{\mu\nu,\eta i_{1}\cdots i_{m}}}\right]g_{\mu\nu,i_{j+1}\cdots i_{m}\alpha}\\
\stackrel{h^{2}} =\sum_{h=1}^{p}\sum_{j=0}^{2h}\left(-1\right)^{j}\sqrt{-g}^{\left(0\right)}a_{h}\partial_{i_{0}\cdots i_{j}}R^{\left(1\right)}\eta^{j_{2}j_{3}}\cdots \eta^{j_{2h}j_{2h+1}}\eta^{ab}\eta^{cd}\biggl\{\delta_{a}^{(\mu}\delta_{d}^{\nu)}\delta_{c}^{(\eta}\delta_{b}^{i_{1}}\delta_{j_{2}}^{i_{2}}\cdots\delta_{j_{2h}}^{i_{2h}}\delta_{j_{2h+1}}^{i_{2h+1})}\\
-\delta_{a}^{(\mu}\delta_{b}^{\nu)}\delta_{c}^{(\eta}\delta_{d}^{i_{1}}\delta_{j_{2}}^{i_{2}}\cdots\delta_{j_{2h}}^{i_{2h}}\delta_{j_{2h+1}}^{i_{2h+1})}\biggr\}h_{\mu\nu,i_{j+1}\cdots i_{2h+1}\alpha}\\
=\sum_{h=1}^{p}\sum_{j=0}^{2h}\left(-1\right)^{j}a_{h}\partial_{i_{0}\cdots i_{j}}R^{\left(1\right)}Q_{\left(\mu\nu\right)}^{\ \ \ \left(\eta i_{1}\cdots i_{2h+1}\right)}h^{\mu\nu}_{\ \ ,i_{j+1}\cdots i_{2h+1}\alpha}
\end{multline}
\begin{equation*}
Q_{\left(\mu\nu\right)}^{\ \ \ \left(\eta i_{1}\cdots i_{2h+1}\right)}=\frac{1}{2!\left(2h+2\right)!}\sum_{ \substack{\mu\nu\in \sigma\left({\mu\nu}\right) \\ \eta i_{1}\cdots i_{2h+1}\in\sigma\left(\eta i_{1}\cdots i_{2h+1}\right)}}Q_{\mu\nu}^{\ \ \ \eta i_{1}\cdots i_{2h+1}}
\end{equation*}
\begin{equation*}
Q_{\left(\mu\nu\right)}^{\ \ \ \left(\eta i_{1}\cdots i_{2h+1}\right)}=\delta_{(\mu}^{(\eta}\delta_{\nu)}^{i_{1}}\eta^{i_{2}i_{3}}\cdots\eta^{i_{2h}i_{2h+1})}-\eta_{(\mu\nu)}\eta^{(\eta i_{1}}\eta^{i_{2}i_{3}}\cdots\eta^{i_{2h}i_{2h+1})}
\end{equation*}
where $\sigma{\left(\mu\nu\right)}$ and $\sigma{\left(\eta i_{1}\cdots i_{2h+1}\right)}$ represent the set of index permutations in the brackets. Averaging Eq. (<ref>) by fixing $\mathbf{k}$ over a suitable spacetime region adopting a harmonic gauge, we get
\begin{multline}\label{media2}
\langle\sum_{h=1}^{p}\sum_{j=0}^{2h}\left(-1\right)^{j}a_{h}\partial_{i_{0}\cdots i_{j}}R^{\left(1\right)}Q_{\left(\mu\nu\right)}^{\ \ \ \left(\eta i_{1}\cdots i_{2h+1}\right)}h^{\mu\nu}_{\ \ ,i_{j+1}\cdots i_{2h+1}\alpha}\rangle\\
=\sum_{h=1}^{p}\sum_{j=0}^{2h}\frac{1}{2!\left(2h+2\right)!}\left(-1\right)^{j}a_{h}\sum_{ \substack{\mu\nu\in \sigma\left({\mu\nu}\right) \\ \eta i_{1}\cdots i_{2h+1}\in\sigma\left(\eta i_{1}\cdots i_{2h+1}\right)}}\langle\partial_{i_{0}\cdots i_{j}}R^{\left(1\right)}Q_{\mu\nu}^{\ \ \ \eta i_{1}\cdots i_{2h+1}}h^{\mu\nu}_{\ \ ,i_{j+1}\cdots i_{2h+1}\alpha}\rangle
\end{multline}
The average of Eq. (<ref>) is independent of index permutations in the lower and upper cases of $Q_{\mu\nu}^{\ \ \ \eta i_{1}\cdots i_{2h+1}}$, that is
\begin{equation}\label{media3}
\langle\partial_{i_{0}\cdots i_{j}}\left(-\frac{1}{2}\Box h\right)Q_{\mu\nu}^{\ \ \ \eta i_{1}\cdots i_{2h+1}}h^{\mu\nu}_{\ \ ,i_{j+1}\cdots i_{2h+1}\alpha}\rangle=\frac{1}{2}\sum_{m=2}^{p+2}\left(-1\right)^{j+h}\left(k_{m}^{2}\right)^{h+1}\left(k_{m}\right)^{\eta}\left(k_{m}\right)_{\alpha}\vert A_{m}\vert^{2}
\end{equation}
By substituting Eq. (<ref>) in Eq. (<ref>), we get
\begin{multline}\label{media4}
\langle\sum_{h=1}^{p}\sum_{j=0}^{2h}\sum_{m=j+1}^{2h+1}\left(-1\right)^{j}\partial_{i_{0}\cdots i_{j}}\left[\sqrt{-g}a_{h}R\frac{\partial\Box^{h}R}{\partial g_{\mu\nu,\eta i_{1}\cdots i_{m}}}\right]g_{\mu\nu,i_{j+1}\cdots i_{m}\alpha}\rangle\\
\stackrel{h^{2}}=\sum_{h=1}^{p}\sum_{j=0}^{2h}\left(-1\right)^{j}a_{h}\sum_{m=2}^{p+2}\left(-1\right)^{j+h}\left(k_{m}^{2}\right)^{h+1}\left(k_{m}\right)^{\eta}\left(k_{m}\right)_{\alpha}\vert A_{m}\vert^{2}\\
=\sum_{h=1}^{p}\sum_{m=2}^{p+2}\left(h+\frac{1}{2}\right)a_{h}\left(-1\right)^{h}\left(k_{m}^{2}\right)^{h+1}\left(k_{m}\right)^{\eta}\left(k_{m}\right)_{\alpha}\vert A_{m}\vert^{2}
\end{multline}
Averaging the right term in Eq. (<ref>), we have
\begin{multline}\label{media5}
\langle\frac{1}{4}\sum_{h=1}^{p}a_{h}\Box h \Box^{h} h^{,\eta}_{\ \ \alpha}+\frac{1}{2}\sum_{h=0}^{1}\sum_{j=h}^{p-1+h}\sum_{m=j+1-h}^{p}\left(-1\right)^{h}a_{m}\Box^{m-j}\left(h^{\eta\lambda}-\eta^{\eta\lambda}h\right)_{,i_{h}\alpha}\Box^{j+1-h}h_{,\lambda}^{\ \ i_{h}}\rangle\\
=\sum_{h=1}^{p}\sum_{m=2}^{p+2}\left(h+\frac{1}{2}\right)a_{h}\left(-1\right)^{h}\left(k_{m}^{2}\right)^{h+1}\left(k_{m}\right)^{\eta}\left(k_{m}\right)_{\alpha}\vert A_{m}\vert^{2}
\end{multline}
Finally, by averaging in the weak field limit Eq. (<ref>) and from Eqs. (<ref>) and (<ref>), we obtain:
\begin{equation}
\langle \left(B_{p}\right)_{\alpha}^{\eta}\rangle=0
\end{equation}
A similar argument gives $\langle \left(A_{p}\right)_{\alpha}^{\eta}\rangle=0$. It is
\begin{multline}
\langle\sum_{h=1}^{p}\sum_{q=0}^{2h+1}\left(-1\right)^{q}\partial_{i_{0}\cdots i_{q}}\left[\sqrt{-g}a_{h}R\frac{\partial\Box^{h}R}{\partial g_{\mu\nu,\eta i_{1}\cdots i_{q}}}\right]g_{\mu\nu,\alpha}\rangle\\
\stackrel{h^{2}}=\frac{1}{2}\sum_{h=1}^{p}\sum_{m=2}^{p+2}a_{h}\left(-1\right)^{h+1}\left(k_{m}^{2}\right)^{h+1}\left(k_{m}\right)^{\eta}\left(k_{m}\right)_{\alpha}\vert A_{m}\vert^{2}
\end{multline}
\begin{multline}
\langle\frac{1}{2}\sum_{h=1}^{p}a_{h}\Box^{h+1}h_{,\lambda}\left(h^{\eta\lambda}-\eta^{\eta\lambda}h\right)_{,\alpha}\rangle\stackrel{h^{2}}=\frac{1}{2}\sum_{h=1}^{p}\sum_{m=2}^{p+2}a_{h}\left(-1\right)^{h+1}\left(k_{m}^{2}\right)^{h+1}\left(k_{m}\right)^{\eta}\left(k_{m}\right)_{\alpha}\vert A_{m}\vert^{2}
\end{multline}
and then averaging Eq. (<ref>) on the l.h.s. and r.h.s., in the weak field limit, we have
\begin{equation}
\langle \left(A_{p}\right)_{\alpha}^{\eta}\rangle=0
\end{equation}
that completes our demonstration.
§.§ The polarizations of gravitational waves
The six polarizations in the solution (<ref>) can be defined in a suitable matrix base. That is [52]
\begin{align*}
\epsilon_{\mu\nu}^{\left(+\right)}&=\frac{1}{\sqrt{2}}
\begin{pmatrix}
0 & 0 & 0 & 0\\0 & 1 & 0 & 0\\0 & 0 & -1 & 0\\0 & 0 & 0 & 0
\end{pmatrix}&
\epsilon_{\mu\nu}^{\left(\times\right)}&=\frac{1}{\sqrt{2}}
\begin{pmatrix}
0 & 0 & 0 & 0\\0 & 0 & 1 & 0\\0 & 1 & 0 & 0\\0 & 0 & 0 & 0
\end{pmatrix}\\
\epsilon_{\mu\nu}^{\left(\text{TT}\right)}&=\qquad
\begin{pmatrix}
1 & 0 & 0 & 0\\0 & 0 & 0 & 0\\0 & 0 & 0 & 0\\0 & 0 & 0 & 0
\end{pmatrix}&
\epsilon_{\mu\nu}^{\left(\text{TS}\right)}&=\frac{1}{\sqrt{2}}
\begin{pmatrix}
0 & 0 & 0 & 1\\0 & 0 & 0 & 0\\0 & 0 & 0 & 0\\1 & 0 & 0 & 0
\end{pmatrix}\\
\epsilon_{\mu\nu}^{\left(1\right)}&=\frac{1}{\sqrt{2}}
\begin{pmatrix}
0 & 0 & 0 & 0\\0 & 1 & 0 & 0\\0 & 0 & 1 & 0\\0 & 0 & 0 & 0
\end{pmatrix}&
\epsilon_{\mu\nu}^{\left(L\right)}&=\qquad
\begin{pmatrix}
0 & 0 & 0 & 0\\0 & 0 & 0 & 0\\0 & 0 & 0 & 0\\0 & 0 & 0 & 1
\end{pmatrix}
\end{align*}
The $+$ and $\times$ are the two standard of general relativity. The other are related to the position of non-null terms with respect to the trace (T).
See also [33] for another derivation in fourth order gravity.
[8]Landau, L. D.; Lifshitz, E. M. The Classical Theory of Fields; Pergamon Press, Oxford, 1971.
[9] Einstein, A. Zur Allgemeinen Relativitätstheorie. Sitzungsber. Preus. Akad. Wiss. Berlin (Math. Phys.) 1915, 47, 778-779.
[10]
Hestenes, D. Energy–Momentum Complex in General Relativity and Gauge Theory. Adv. Appl. Clifford Algebra 2021, 31, 51.
[11]Goldberg, J.N. Conservation Laws in General Relativity. Phys. Rev. 1958, 111, 315.
[12]Dongsu Bak; Cangemi, D.;Jackiw, R. Energy-momentum conservation in gravity theories. Phys. Rev. D 1994, 49, 5173.
[13]Lee, D.L.; Lightman, A.P.; Ni, W.T. Conservation laws and variational principles in metric theories of gravity. Phys. Rev. D 1974, 10, 1685.
[14]Rosen, N. The Energy of the Universe. Gen. Rel. Grav. 1994, 26, 319.
[15]Lessner, G. Møller's energy-momentum complex — Once again. Gen Relat Gravit 1996, 28, 527.
[16]Palmer, T.N. Gravitational energy-momentum: The Einstein pseudotensor reexamined. Gen Relat Gravit 1980, 12, 149.
[17]Ferraris, M.; Francaviglia, M. Covariant first-order Lagrangians, energy-density and superpotentials in general relativity. Gen Relat Gravit 1990, 22, 965–985.
[18]Capozziello, S.; De Laurentis, M. Extended Theories of Gravity. Phys. Rept 2011, 509, 167.
[19] Capozziello, S. ; Francaviglia, M. Extended Theories of Gravity and their Cosmological and Astrophysical Applications. Gen. Rel. Grav. 2008, 40, 357.
[20] Bogdanos, Ch.; Capozziello, S.; De Laurentis, M.; Nesseris, S.
Massive, massless and ghost modes of gravitational waves from higher-order gravity. Astropart. Phys. 2010, 34, 236.
[21] Canfora, F.; Vilasi, G.; Vitale, P. Nonlinear gravitational waves and their polarization. Phys. Lett. B 2002, 545, 373.
[22]Mikhail, F.I.; Wanas, M.I. ; Hindawi, A. ; Lashin, E.I. Energy-Momentum Complex in Møller's
Tetrad Theory Of Gravitation. Int. J. Theor. Phys. 1993, 32, 1627. https://doi.org/10.1007/BF00672861
[23]
Multamaki, T.; Putaja, A.; Vilja, I.; Vagenas, E.C.
Energy-momentum complexes in $f(R)$ theories of gravity. Class. Quant. Grav. 2008, 25, 075017.
[24]Capozziello, S.; Capriolo, M.; Transirico, M. The gravitational energy-momentum pseudo-tensor of higher order theories of gravity, Ann. Phys. 2017, 525, 1600376.
[25]
Capozziello, S.; Capriolo, M.; Transirico, M.
The gravitational energy-momentum pseudo-tensor: the cases of $f(R)$ and $f(T)$ gravity. Int. J. Geom. Meth. Mod. Phys. 2018, 15, 1850164.
[arXiv:1804.08530 [gr-qc]].
[29]Capozziello, S.; Capriolo, M.; Caso, L. Weak field limit and gravitational waves in f(T, B) teleparallel gravity. Eur. Phys. J. C 2020, 80, 156.
[30]Capozziello, S.; Capriolo, M.; Caso, L. Gravitational waves in higher order teleparallel gravity. Class. Quantum Grav. 2020, 37, 235013.
[31]Xulu S. S., The Energy-Momentum Problem in General Relativity, PhD Thesis,
arXiv (2003), https://doi.org/10.48550/arXiv.hep-th/0308070.
[32] Weinberg, S. Gravitation and Cosmology; Wiley, New York, 1972.
[33] Capozziello, S.; Stabile, A. Gravitational waves in fourth order gravity. Astrophys. Space Sci. 2015, 358, 27.
[34] Stelle, K.S. Classical Gravity with Higher Derivatives. Gen. Rel. Grav. 1978, 9, 353.
[35] Capozziello, S.; Stabile, A.; Troisi, A. The Post-Minkowskian Limit of f(R)-gravity. Int. J. Theor. Phys. 2010, 49, 1251.
[36] Greiner, R.; Reinhardt, J. Field Quantization; Springer-Verlag, Berlin, 1996.
[37]Capozziello, S.; Faraoni, V. Beyond Einstein Gravity; FTP 170, Springer, New York, 2011.
[38] Misner, C.W.; Thorne, K.S; Wheeler, J.A. Gravitation; Freeman and Co., New York, 1971.
[39] Straumann, N. General Relativity; Springer, New York, 2013.
[40]Pauli, W. Theory of Relativity; Pergamon Press, Oxford, 1958.
[41] Dirac, P.A.M. General Theory of Relativity; Princeton Univ. Press, Princeton, 1996.
[42]Hawking, S.W.; Ellis, G.F.R. The Large Scale Structure of Spacetime; Cambridge University Press, Cambridge, 1973.
[43]Schutz, B. A First Course in General Relativity; Cambridge University Press, Cambridge, 2009.
[44]Ohanian, H.C.; Ruffini, R. Gravitation and Spacetime; Cambridge University Press, Cambridge, 2013.
[45] Stephani, H. General Relativity; Cambridge University Press, Cambridge, 1990.
[46] Modesto, L.; Rachwal, L. Universally finite gravitational and gauge theories.
Nucl. Phys. B 2015, 900, 147.
[47] Giaccari, S.; Modesto, L. Classical and Quantum Nonlocal Supergravity, 2016, e-Print: arXiv:1605.03906 [hep-th] (2016).
[48] Briscese, F.; Pucheu, M.L. Palatini formulation of non-local gravity. Int. J. Geom. Methods Mod. Phys. 2017, 14, 1750019.
[49]
Modesto, L.; Shapiro, I.L. Super-renormalizable quantum gravity with complex ghosts. Phys.Lett. B 2016, 755, 279.
[50]Modesto, L. Super-renormalizable or finite Lee-Wick quantum gravity. Nucl.Phys. B 2016 909, 584.
[51] Quandt, I.; Schmidt, H.J. The Newtonian limit of fourth and higher order gravity. Astron. Nachr. 1991, 312, 97.
[52]Capozziello, S.; Capriolo, M.; Caso, L. Weak field limit and gravitational waves in higher order
gravity. Int. J. Geom. Methods Mod. Phys. 2019, 16 No.03, 1950047.
[53]Abedi, H.; Capozziello, S.; Capriolo, M.; Abbassi, A.M. Gravitational energy–momentum pseudo-tensor in Palatini and metric f(R) gravity. Annals of Physics 2022, 439, 168796.
[54]
Allemandi, G.; Capone, M.; Capozziello, S.; Francaviglia, M.
Conformal aspects of Palatini approach in extended theories of gravity. Gen. Rel. Grav. 2006, 38, 33–60.
[arXiv:hep-th/0409198 [hep-th]].
[55]
Dick, R.
Covariant conservation laws from the Palatini formalism. Int. J. Theor. Phys. 1993, 32, 109–120.
[56]
Barraco, D.E.; Dominguez, E.; Guibert, R.
Conservation laws, symmetry properties, and the equivalence principle in a class of alternative theories of gravity. Phys. Rev. D 1999, 60, 044012.
[57]
Koivisto, T.
Covariant conservation of energy momentum in modified gravities.
Class. Quant. Grav. 2006, 23, 4289–4296.
[arXiv:gr-qc/0505128 [gr-qc]].
[58]
Barragan, C.; Olmo, G.J.; Sanchis-Alepuz, H.
|
# Deep learning of experimental electrochemistry for battery cathodes across
diverse compositions
Peichen Zhong<EMAIL_ADDRESS>Department of Materials Science and
Engineering, University of California, Berkeley, California 94720, United
States Materials Sciences Division, Lawrence Berkeley National Laboratory,
California 94720, United States Bowen Deng Department of Materials Science
and Engineering, University of California, Berkeley, California 94720, United
States Materials Sciences Division, Lawrence Berkeley National Laboratory,
California 94720, United States Tanjin He Department of Materials Science and
Engineering, University of California, Berkeley, California 94720, United
States Materials Sciences Division, Lawrence Berkeley National Laboratory,
California 94720, United States Zhengyan Lun Department of Materials Science
and Engineering, University of California, Berkeley, California 94720, United
States School of Chemical Sciences, University of Chinese Academy of
Sciences, Beijing 100049, China Gerbrand Ceder<EMAIL_ADDRESS>Department
of Materials Science and Engineering, University of California, Berkeley,
California 94720, United States Materials Sciences Division, Lawrence
Berkeley National Laboratory, California 94720, United States
###### Abstract
Artificial intelligence (AI) has emerged as a tool for discovering and
optimizing novel battery materials. However, the adoption of AI in battery
cathode representation and discovery is still limited due to the complexity of
optimizing multiple performance properties and the scarcity of high-fidelity
data. In this study, we present a machine-learning model (DRXNet) for battery
informatics and demonstrate the application in the discovery and optimization
of disordered rocksalt (DRX) cathode materials. We have compiled the
electrochemistry data of DRX cathodes over the past five years, resulting in a
dataset of more than 19,000 discharge voltage profiles on diverse chemistries
spanning 14 different metal species. Learning from this extensive dataset, our
DRXNet model can automatically capture critical features in the cycling curves
of DRX cathodes under various conditions. Illustratively, the model gives
rational predictions of the discharge capacity for diverse compositions in the
Li–Mn–O–F chemical space as well as for high-entropy systems. As a universal
model trained on diverse chemistries, our approach offers a data-driven
solution to facilitate the rapid identification of novel cathode materials,
accelerating the development of next-generation batteries for carbon
neutralization.
###### pacs:
## I Introduction
Figure 1: Discharge voltage profiles from experiments. (A) The discharge
voltage profile illustrates the relationship between capacity $Q$ and voltage
$V$, which is conditioned on the composition of the cathode material. (B) The
derivative quantity $dQ/dV$ is used to quantify the redox potentials of the
TM. The experimental discharge voltage profiles of Li1.2Mn0.2Cr0.2Ti0.4O2.0
DRX with (C) multi-rate tests from 20 – 1000 mA/g and (D) multi-cycle tests
from the first to the 30th cycles. (E) The parsed discharge profile is stored
in a voltage array $\\{V_{i}^{N}\\}$, and a capacity array $\\{Q_{i}^{N}\\}$,
where the subscript $i$ represents a point (state) on a discharge profile, and
the superscript $N$ represents the cycle number.
The pursuit of carbon neutrality has become a global imperative in the face of
climate change, driving the transition to renewable energy sources and the
widespread adoption of electric vehicles [1, 2, 3]. High-performance battery
cathode materials with large energy density, high-rate performance, and long
cycle life are central to these advancements. The development of new cathode
materials is essential to meeting the increasing demand for energy storage and
advancing the electrification of transportation systems [4].
Artificial intelligence (AI) has emerged as a potential tool in the discovery
and optimization of novel battery materials [5, 6]. By leveraging vast amounts
of experimental and computational data, AI-assisted techniques can accelerate
the design and synthesis of battery materials by identifying promising
candidates within large chemical spaces [7, 8, 9], uncovering hidden
structure-property relationships via machine-learned atomistic modeling [10],
predicting the remaining lifespan of batteries [11, 12, 13, 14, 15], and
optimizing protocols for fast charge/discharge protocol [16]. These efforts
significantly reduce the time and cost required for conventional trial-and-
error approaches. Most recently, a battery data genome initiative has been
proposed to use AI assistance to accelerate the discovery and optimization of
battery materials [17].
Despite these advancements, current machine-learning efforts in battery
research primarily focus on predicting the lifespan for a simple chemistry or
within a limited chemical space, such as NMC (Ni–Mn–Co) or LFP (LiFePO4). The
development of exploratory machine learning for representing comprehensive
compositional effects in a multi-dimensional chemical space remains
underdeveloped due to the challenges associated with simultaneously optimizing
multiple electrochemical properties (e.g., rate capability, cyclability, and
various test voltage windows) [18]. Moreover, the scarcity of high-fidelity
data further hinders the progress of AI in the battery field.
Disordered rocksalt (DRX) materials have emerged as promising cathode
materials that make use of earth-abundant precursors to enable scaling of Li-
ion energy storage to several TWh/year production [19]. Owing to the nearly
unlimited compositional design space and considerably more complex structure-
property relationship of DRX cathodes compared with conventional layered
cathodes (Figure 1A), their rational design requires the extensive involvement
of advanced characterization techniques (e.g., pair-distribution function
analysis [20], spherical-aberration-corrected transmission electron microscopy
[21], solid-state nuclear magnetic resonance spectroscopy [22]) as well as
sophisticated computational tools (e.g., high-dimensional cluster expansion
and Monte Carlo simulation [23, 24]). Data-driven methods offer alternative
means of compositional design and optimization of materials without having to
fully construct their structure-property relationships.
In light of these challenges, we developed DRXNet, an exploratory machine-
learning model for the discovery and optimization of battery cathode
materials. DRXNet uses composition, test current density, working voltage
window, and cycle number as inputs to predict entire discharge voltage
profiles. By training and testing over 19,000 experimental discharge voltage
profiles of DRX materials comprising various metal species, we show that the
model accurately captures the cathode electrochemistry under different test
conditions. Notably, DRXNet captures accessible discharge capacity in diverse
Li–Mn–O–F compositions and makes rational predictions for several high-entropy
systems. As a universal model trained on diverse chemistries, DRXNet offers a
data-driven solution to facilitate the rapid identification of novel cathode
materials with improved energy-storage capabilities.
Figure 2: Description of the collected experimental dataset and model design:
(A) The elemental distribution of collected experimental electrochemistry
data. The dataset contains 7,898 discharge profiles collected from DRX oxides
and 11,604 discharge profiles from oxyfluorides. The color-coding of the boxes
indicates the number of discharge profiles (cycles) on compounds that contain
that specific element. The number within each elemental box represents the
number of compounds with that element on which experiments were conducted. (B)
A histogram of the number of cycles ($N_{\text{cycle}}$) and current density
(rate) for all the individual electrochemical tests. (C) An end-to-end
pipeline that maps $Q_{i}=\mathcal{F}(V_{i}|\mathcal{O})$, which consists of
the electrochemical condition network $\mathcal{O}$ (left) and the state
prediction network $\mathcal{F}$ (right). The electrochemical condition
network encodes the DRX composition, current density rate, and cycle
information. The three encoded vectors are synthesized through gated-MLPs with
soft attention to obtain the condition vector $\vec{X}_{\mathcal{O}}$ [25].
The state prediction is approximated as a forward deep neural network that
takes the voltage state $V_{i}$ and cycling voltage window
$V_{\text{low}},V_{\text{high}}$ as inputs. The encoded condition vector
$\vec{X}_{\mathcal{O}}$ is element-wise added in the hidden layer of
$\mathcal{F}$. The circled symbols are all element-wise operations. The
message-passing graph neural network (GNN) is used for compositional encoding
of DRX, adapted from the Roost model [26].
## II Results
### II.1 Parsing discharge profiles
Unlike conventional NMC-based layered cathodes [27, 28], DRX materials exhibit
more diverse electrochemical behavior due to the significantly larger chemical
space over which they can exist and their more subtle structure involving
various forms of cation short-range order [29]. A prototype DRX cathode
(Li1+xM’aM”bO2-yFy) is composed of three primary compositional parameters: (1)
the redox-active species M’; (2) the inert high-valent transition metal M”,
which charge-compensates for the Li excess and stabilizes disordered
structures [30]; (3) fluorine, which enhances the cyclability and accommodates
more Li excess without losing TM redox by reducing the anion valence [31]. In
addition, other compositional modifications are often made to enhance
capacity, rate, or cyclability. For instance, Mg doping in Mn-based
oxyfluoride DRX can increase the discharge capacity while retaining a similar
voltage-profile shape [32]; Cr doping in Li1.2Mn0.4Ti0.4O2.0 results in
comparable low-rate capacity but significantly improves the high-rate
performance near the top of charge [33]. These non-linear effects arising from
compositional changes make both material design and machine-learning modeling
challenging, thereby necessitating a comprehensive, high-fidelity dataset to
address such issues.
Figure 1A introduces the typical discharge-voltage profile in battery tests.
The profile shape is tied to various factors, such as the DRX composition,
applied current density rate, and degradation that may have occurred in prior
cycles. Figure 1C and D show the multi-rate tests (the first cycle) and multi-
cycle tests (of 20 mA/g and 1000 mA/g) of Li1.2Mn0.2Cr0.2Ti0.4O2.0 cathode as
an example. The capacity $Q$ is measured in experiments by determining the
cumulative charge transferred in a galvanostatic test. Taking the derivative
of $Q$ with respect to $V$, the $dQ/dV$ value can be evaluated for a given
voltage profile, which is a crucial physical quantity for analyzing
characteristic redox potentials from different TMs [34].
### II.2 DRX Battery Test Dataset
We have compiled the electrochemical test data related to DRX compounds by
mining electronic experimental notebooks in our research group over the past
five years to construct the DRX Test Dataset (DRX-TD). The dataset contains
not only results on successful materials published in several papers [35, 36,
29, 37, 38, 39, 40, 33, 32] but also data on less well-performing DRX
compounds. This endeavor yielded a comprehensive dataset containing 19,000
discharge profiles across 16 different elements (14 metal species + O and F)
from lab experiments and published literature (see Methods). An individual
electrochemical test is defined as a group of $N_{\text{cycle}}$ discharge
profiles with a fixed current density rate, where $N_{\text{cycle}}$ is the
number of cycles conducted in such a test, corresponding to the results
obtained from one coin-cell. The distribution of elements in the DRX-TD is
shown in Figure 2A, where the number in each element’s box represents the
number of compounds with that element present for which an electrochemical
test is present. The box’s color indicates the total number of discharge
profiles for compounds containing that element. Comprising 7,898 discharge
profiles of DRX oxides and 11,604 discharge profiles of oxyfluorides, the
dataset offers extensive coverage of major redox-active TMs. Figure 2B
displays histograms for the number of cycles, $N_{\text{cycle}}$, and the
current rates at which experiments were performed. As is typical for
exploratory research programs in a research laboratory, most of the
electrochemical tests were conducted at a low current rate (20 mA/g) and for
10-100 cycles.
For each discharge profile, 100 points were uniformly sampled from the values
of $V$ and $Q$, resulting in a voltage series
$\boldsymbol{V}=\left[V_{1},V_{2},...,V_{i},...\right]$ and a capacity series
$\boldsymbol{Q}=\left[Q_{1},Q_{2},...,Q_{i},...\right]$. The $dQ/dV$ curve was
then calculated by differentiating $\boldsymbol{Q}$ with $\boldsymbol{V}$. As
$dQ/dV$ is a more intrinsic property for battery materials, including this
value in the modeling allows for a more representative analysis of the
electrochemical performance of DRX compounds under various conditions (see
Methods).
### II.3 DRXNet architecture
DRXNet aims to draw a connection between chemistry and cathode performance by
establishing a mapping between $\boldsymbol{V}$ and $\boldsymbol{Q}$ for
arbitrary cathode compositions under various test conditions. This idea can be
conceptualized as identifying a function $\mathcal{F}$ that maps cathode
parameters and the voltage state $V_{i}$ to produce the capacity state $Q_{i}$
as an output. The function $\mathcal{F}$ is conditionally defined by the
parameters $\mathcal{O}$, which consider the electrode composition, current
rate, and cycle number
$Q_{i}=\mathcal{F}(V_{i}|\mathcal{O}).$ (1)
We designed DRXNet with two main components, as shown in Figure 2C: (1) An
electrochemical condition network that generates a feature vector
$\vec{X}_{\mathcal{O}}$ based on the compound’s composition and
electrochemical test information; (2) A state prediction network to
approximate the discharge state of the cathode as a function of the voltage
state, $Q_{i}=\mathcal{F}(V_{i}|\mathcal{O})$, given the electrochemical
conditional encoding of $\mathcal{O}$. For instance, Algorithm 1 demonstrates
how DRXNet predicts the first-cycle discharge profile of
$\text{Li}_{1.2}\text{Mn}_{0.2}\text{Cr}_{0.2}\text{Ti}_{0.4}\text{O}_{2}$ at
a current rate of 20 mA/g between 1.5 and 4.8 V.
Condition Inputs:
$\mathcal{O}=\begin{cases}&\textbf{composition}=\text{Li}_{1.2}\text{Mn}_{0.2}\text{Cr}_{0.2}\text{Ti}_{0.4}\text{O}_{2}\\\
&\textbf{rate}=20~{}\text{mA/g},\\\ &\textbf{cycle}=1\end{cases}$
Condition Outputs:
$\vec{X}_{\mathcal{O}_{1}}=\vec{X}_{\text{comp}}+\sigma_{f_{1}}(\vec{X}_{\text{comp}}||\vec{X}_{\text{rate}})\cdot
f_{1}(\vec{X}_{\text{comp}}||\vec{X}_{\text{rate}})$
$\displaystyle\vec{X}_{\mathcal{O}_{N}}=\vec{X}_{\mathcal{O}_{1}}+\sigma_{f_{2}}(\vec{X}_{\mathcal{O}_{1}}||\vec{X}_{\text{cycle}})$
$\displaystyle\cdot f_{2}(\vec{X}_{\mathcal{O}_{1}}||\vec{X}_{\text{cycle}})$
$\displaystyle\cdot\boldsymbol{W}_{n}(N-1)$
Inputs: $\boldsymbol{V}=\left[1.5,...,V_{i},...,4.8\right]\rightarrow N$
series
for _$i=1$ to $N$_ do
Compute $Q_{i}=\mathcal{F}(V_{i}|\vec{X}_{\mathcal{O}_{N}})$
end for
Outputs: $\boldsymbol{Q}=\left[Q_{1},...,Q_{i},...,Q_{N}\right]$
Algorithm 1 The workflow of DRXNet with an example of
$\text{Li}_{1.2}\text{Mn}_{0.2}\text{Cr}_{0.2}\text{Ti}_{0.4}\text{O}_{2}$
Initially, three condition inputs (composition, rate, cycle) are encoded to
represent $\mathcal{O}$. We use Roost, a graph neural network model proposed
by Goodall and Lee [26], for compositional encoding. Roost takes elements as
graph nodes and updates the correlation between elements through weighted
message passing based on each element’s fractional concentration. The nodes
are initialized with elemental embedded vectors $\vec{h}_{s}$ ($s$: species)
from mat2vec to capture as much prior chemical information as possible through
text mining of previously published literature [41]. Moreover, we consider
only the cation species as independent nodes in Roost, treating the anion-
species information (fluorine) as a mean-field background, i.e.,
$\vec{h}^{\prime}_{\text{Li}}=\vec{h}_{\text{Li}}+c_{\text{F}}\cdot\vec{h}_{\text{F}}$,
where $c_{\text{F}}$ is the fractional concentration of fluorine and
$\vec{h}_{\text{Li/F}}$ is the embedded vector of Li/F. Rate and cycle
information is encoded using multi-layer perceptrons (MLPs).
Figure 3: Error and model variance analysis of DRXNet in compositional space:
The prediction error of discharge capacity between 2.0 and 4.4 V ($y$-axis)
vs. cycle number ($x$-axis). The model variance is represented by
$\sigma_{Q}$, a standard deviation of the ensemble of the models’ prediction,
which is plotted as scaled colored dots. (A)–(B): Predictions on 3TM/HE using
models trained on the 2TM dataset. (C)–(D): Predictions on 3TM/HE using models
trained on both the 2TM dataset and the first cycles of the 3TM/HE dataset.
(E)–(F): Predictions on 3TM/HE using models trained on the 2+3TM dataset.
Because the rate and cycle properties are intrinsically affected by the
composition, we used gated MLPs with soft attention for electrochemical
condition encoding via a hierarchical network structure [25]. The
$\vec{X}_{\mathcal{O}_{1}}=\vec{X}_{\text{comp}}+\sigma_{f_{1}}(\vec{X}_{\text{comp}}||\vec{X}_{\text{rate}})\cdot
f_{1}(\vec{X}_{\text{comp}}||\vec{X}_{\text{rate}})$ is a rate-informed
feature vector, where $\sigma_{f}$ and $f$ represent MLPs with different
activation functions and $||$ denotes the concatenation operation. In
addition, the cycle-informed vector
$\vec{X}_{\mathcal{O}_{N}}=\vec{X}_{\mathcal{O}_{1}}+\sigma_{f_{2}}(\vec{X}_{\mathcal{O}_{1}}||\vec{X}_{\text{cycle}})\cdot
f_{2}(\vec{X}_{\mathcal{O}_{1}}||\vec{X}_{\text{cycle}})\cdot\boldsymbol{W}_{n}(N-1)$
is linearly dependent on the cycle number with a trainable weight
$\boldsymbol{W}_{n}$. As such, the feature vector $\vec{X}_{\mathcal{O}_{1}}$
is used to represent the 1st cycle and $\vec{X}_{\mathcal{O}_{N}}$ is used to
represent the $N$-th cycle, respectively.
Lastly, we used several MLPs to construct the state prediction network
$\mathcal{F}$, as shown in Figure 2C. $\mathcal{F}$ takes the voltage state
$V_{i}$ and working window $V_{\text{low}},V_{\text{high}}$ as inputs, and the
$\vec{X}_{\mathcal{O}}$ is element-wise added to the hidden layer of
$\mathcal{F}$ to inform $\mathcal{F}$ of conditions $\mathcal{O}$ (see
Methods). As such, the state prediction network $\mathcal{F}$ is constructed
as a simple function mapping from the voltage state $V_{i}$ to the capacity
$Q_{i}$. In addition, $(dQ/dV)_{i}$ is obtained by auto-differentiation of
$\mathcal{F}$.
### II.4 Applicability domain
We explore the scope of DRXNet’s applicability in the realm of composition
space. Determination of the applicability domain in battery machine-learning
models can be challenging due to the unavailability of sufficient test data,
as generating new data necessitates the synthesis of new materials or
conducting battery cycling tests for weeks to months [18, 42]. Simply
separating the sequence of voltage and capacity signals $\\{V_{i},Q_{i}\\}$
into training and test sets can result in data leakage and a failure to
represent the expected error in real applications. To evaluate the
expressibility and generalization of DRXNet, we designed several experiments
by partitioning the dataset based on compositions. The electrochemical tests
with no more than two metal species (2TM, excluding Li) were designated as the
training set, whereas the tests with three metal species (3TM) and higher
numbers of TM components (high-entropy, HE) were assigned as test sets. For
each test, an ensemble of five independent models was trained to enhance the
overall prediction accuracy and robustness and to quantify the model variance.
The average value is used for the prediction, and the standard deviation of
the prediction from the ensemble of five DRXNet models ($\sigma_{Q}$) is used
to represent the model variance as an approximation of how uncertain the
predictions are.
A rational design of battery cathodes typically focuses on the capacity that
can be delivered within a certain voltage window. Therefore, we used DRXNet to
compute the voltage profiles with electrochemical test parameters in the test
set and compared the delivered capacity between 2.0 – 4.4 V of experiments and
predictions within 50 cycles (see Figure 3). The voltage range of 2.0 – 4.4 V
(vs. Li+/Li) is reasonable for current electrolytes, and most commercialized
cathodes such as LiFePO4, LiCoO2, and NMC operate within this voltage range.
Our choice of this voltage range for testing model performance is aligned with
these industry norms. The average voltages ($\bar{V}=\sum_{i}V_{i}\Delta
Q_{i}/\sum_{i}\Delta Q_{i}$) between 2.0 – 4.4 V were subsequently computed.
As a baseline, the mean absolute deviation (MAD) of average voltage is
0.16/0.21 V for 3TM/HE, and the MAD of discharge capacity is 36.59/38.54 mAh/g
for 3TM/HE. Figure 3A and B demonstrate the performance of the DRXNet models
trained on the 2TM dataset and tested on the 3TM and HE datasets. Mean
absolute errors (MAEs) of 0.1/0.13 V for the average voltage and 23.38/29.97
mAh/g for the capacity were obtained for the 3TM/HE test datasets,
respectively. It is found that large prediction errors already occur for the
first cycle and propagate into the subsequent cycles. Notably, a systematic
underestimation of capacity is observed for the HE compounds (Figure 3B),
which can be rationalized by the fact that 2TM compounds cannot capture the
improved performance arising from the novel high-entropy physics [40, 43].
For practical applications, new data points can be continuously collected as
experiments progress, enabling on-the-fly training with incoming data to
improve predictive performance. To evaluate possible improvement with
additional information specific to the system being tested, we evaluated the
improvement when DRXNet is trained on a dataset containing all 2TM data and is
provided with the first cycle data from 3TM/HE materials. The knowledge of
just the first cycle data results in a reduction of the mean capacity error
from 23.38/29.97 mAh/g to 14.84/17.58 mAh/g for 3TM/HE (Figure 3C and D). The
enhanced performance achieved by explicitly training with the first cycle
indicates that the model can better generalize cycling performance, even when
experiments for a specific composition are not extensively sampled. This
capability has the potential to significantly reduce the month-long timeframe
typically required for electrochemical testing to identify whether a new
cathode material has a desired cyclability or rate capability. Training the
model with first-cycle data led to a substantial decrease in both the
prediction error and model variance for the initial few cycles, although the
model variance increased subsequently with the cycle number for untrained
domains (Figures 3C and D).
To examine how data augmentation could improve the performance of DRXNet, we
further trained models on the 2+3TM dataset where chemical information in
addition to 2TM interactions is included. Figure 3E and F display the
predictions on the 3TM (MAE: 6.0 mAh/g) and HE (MAE: 19.63 mAh/g) datasets. It
is important to note that the models trained on 2+3TM data show an error
reduction of around 10 mAh/g for the HE capacity prediction compared to the
results obtained when training the 2TM model (Figures 3B and F), along with a
significant reduction on the model variance. This finding suggests that the
2TM dataset is inadequate for extracting relevant information and generalizing
it to other compositions. The scaling to electrode material with a high number
of components necessitates capturing more than 2TM correlations or
interactions in training the graph neural network. Failure to do so may lead
to systematic prediction errors, as demonstrated in Figure 3B. When the model
is able to acquire sufficient chemical domain knowledge (e.g., 2+3TM-model),
it becomes feasible to extrapolate the electrochemical properties of high-
component electrodes, which is evidenced in Figure 3F with reduced prediction
error as well as model variance, and only a few outlier experiments exhibit
large errors.
Figure 4: Illustration of predictions of discharge capacity in Li–Mn–O–F DRX
systems: (A) Compositional design principle includes the optimization of Li-
excess content, TM-redox and Li-F short-range order (SRO) [39].
(B, C) Prediction of discharge capacity in the Li–Mn–O–F chemical space for
the 1st (B) and 30th (C) cycle between 1.5 – 4.8 V at a current density rate
of 20 mA/g. The blue stars indicate the compositions included in the training
set.
### II.5 Recover design principles in Li–Mn–O–F chemical space
We present several examples to illustrate how DRXNet learns the underlying
cathode chemistry and assists in designing new materials, where the models
used for these applications are pretrained on all discharge profiles. As an
attractive earth-abundant, non-precious TM, Mn is of considerable interest for
designing next-generation cathode materials [35]. Lun _et al._ [39] proposed
three primary design degrees of freedom for Mn-based DRX (Figure 4A): (1) the
Li-excess content, which controls the presence of a percolating network
facilitating Li diffusion; (2) the Mn content, as achieving high capacity with
a low amount of Mn requires a large amount of oxygen redox, leading to poor
cyclability; and (3) the fluorine content, which lowers the total cation
valence and provides greater freedom to optimize the Li and Mn content.
Fluorine modifies cation short-range order (SRO) through the strong Li–F
attraction and lowers the initial capacity [44, 32], but can increase
stability at high voltage charging [21]. These principles are highly
correlated and exert non-linear effects on performance.
We used DRXNet to predict the discharge capacity of DRX compounds between 1.5
and 4.8 V at a current rate of 20 mA/g for the 1st and 30th cycles. The
results, as a function of Li and F content, are shown in Figures 4B and C. The
Mn content and valence follow directly from the Li and F content. The effect
of fluorine on performance, extensively characterized experimentally, is well
captured by our model: A higher F content ($y$ in O2-yFy) results in a lower
discharge capacity for the 1st cycle but a higher capacity for the 30th cycle,
consistent with its documented role in promoting surface stability [21]. In
particular, cation-disordered Li1.333Mn0.667O2 (bottom right corner of Figure
4C) is predicted to have the highest capacity ($>320$ mAh/g) for the first
cycle but the lowest capacity for the 30th cycle. In this compound, the
valence of Mn is 4+, and all capacity originates from oxygen. Such a large
amount of O-redox leads to rapid capacity fade consistent with the
experimental observations on disordered Li2MnO3 reported in Ref. [45].
To provide some context for the extrapolation capability of DRXNet, we have
illustrated the compositions in the training dataset with blue stars in
Figures 4B and C. From this, it can be observed that even with a limited
distribution of training points on the composition map, DRXNet offers
reasonably consistent predictions that seem to be in line with the
experimental observations beyond the training points. As DRXNet is trained on
various compositions beyond the Li–Mn–O–F chemical space, the ability to
extrapolate to other domains can be attributed to the transfer learning from
other F- and non-F-containing compounds. The example in this section
demonstrates how practitioners can generalize the design principles from a
data-driven perspective purely starting from the data mined from experiments.
### II.6 Exploratory search for high-entropy cathodes
Figure 5: Predicted discharge voltage profiles of two high-entropy DRX
materials. (A) Li1.2Mn0.1Mg0.1Cr0.3Ti0.2Nb0.1O1.8F0.2 (HE-1) and (B)
Li1.2Mn0.1Mg0.1Cr0.15V0.15Ti0.2Nb0.1O1.8F0.2 (HE-2) with various current
densities (from 20 mA/g to 1000 mA/g) between voltage window of 1.5 –4.8 V.
The inset displays the cycled discharge capacity at a current density of 20
mA/g. HE-2 with various current densities (from 10 mA/g to 10 A/g) between
voltage windows of (C) 2.0 –4.4 V and (D) 2.0 – 4.0 V.
High-entropy DRXs are composed of many species and present a vast chemical
space to explore for battery materials discovery. We used DRXNet for virtual
high-throughput screening considering redox-compatible species from the
bivalent (Mn2+, Fe2+, Ni2+, Mg2+) and trivalent group (Mn3+, Cr3+, V3+, Fe3+).
Two case studies of predicted high-entropy DRXs are presented:
Li1.2Mn0.1Mg0.1Cr0.3Ti0.2Nb0.1O1.8F0.2 (HE-1) and
Li1.2Mn0.1Mg0.1Cr0.15V0.15Ti0.2Nb0.1O1.8F0.2 (HE-2). The discharge profiles
predicted with DRXNet under various current densities are shown in Figures 5A
and B. A more comprehensive collection of predictions for other compositions
is included in Figure S6.
For HE-1, DRXNet predicts a discharge capacity of 276 mAh/g at a current rate
of 20 mA/g. The compound delivers its largest discharge capacity near 3V and
transitions to a higher voltage slope below 3V, a phenomenon that has been
widely observed in Mn redox and/or Cr redox-based DRXs [46, 33, 40]. HE-1 is
predicted to have an unusually high rate capability for a DRX compound when
discharging. A capacity of 196 mAh/g is estimated at 1000 mA/g, which is
$71\%$ of the capacity at 20 mA/g. Previous work has demonstrated that multi-
elemental substitution (i.e., high-entropy strategy) frustrates the
unfavorable short-range order that leads to poor Li kinetics. In addition, the
incorporation of Cr and its migration as Cr6+ at high voltage creates a more
extended 0-TM network for Li transport. Both of these features improve the Li
diffusion kinetics [33, 40]. DRXNet clearly learns those benefits and
extrapolates rationally into electrochemistry prediction of the high-entropy
compositions.
As a comparison to HE-1, we formulated HE-2 with partial V3+ to Cr3+
substitution. The change in the shape of the voltage profile due to the low
potential of V5+/V3+ reduction is well captured by DRXNet as shown in Figure
5B and $dQ/dV$ curves in SI. It is clearly demonstrated that with V3+
incorporation, a nearly constant slope can be observed down to the low-voltage
region, which is characteristic for reported V-based DRX cathodes [47, 48].
Nevertheless, similar to Cr6+, V5+ can migrate into the tetrahedral sites to
enhance Li transport, which benefits the rate capability [47]. Consistently,
with this concept, HE-2 is predicted to retrain 171 mAh/g capacity at 1000
mA/g ($64\%$ of the 266 mAh/g capacity at 20 mA/g), which is superior to the
majority of the DRX cathodes reported to date.
The inset plots in Figures 5A and B show the predicted discharge capacity of
HE-1 and HE-2 for 20 cycles. The capacity drop in the first five cycles is
predicted to slow down upon further cycling. This result is in full agreement
with experimental findings, which indicate some of the irreversibility in the
initial cycles, such as cathode–electrolyte interface formation [49]. These
examples illustrate how practitioners can effectively use DRXNet to navigate
the extensive chemical space of high-entropy DRXs and identify promising
candidates for cathode design and optimization.
### II.7 Electrochemical conditions
We further tested the depth and transferability of DRXNet’s predictive
capabilities by varying the HE-2 discharge voltage window and cycling rate,
which are typical parameters varied in the investigation of a new cathode
material. Figure 5C displays the discharge profiles between 2.0 – 4.4 V, with
two additional rates tested (10 mA/g for a low rate and 104 mA/g for an
extremely high rate). These conditions are infrequently incorporated into our
training data. The 10 mA/g exhibits a discharge profile very similar to that
obtained at 20 mA/g, which is entirely consistent with typical experimental
findings, as the discharge process at such a low rate exhibits a reduced
overpotential and is closer to the equilibrium. The 10 A/g rate discharge
profile demonstrates a sharp drop in voltage, reasonably indicating poor
performance at this extremely high rate. Some unphysical predictions start to
appear when the model is tested to predict the discharge profiles between 2.0
– 4.0 V. As Figure 5D shows a small non-zero offset $\sim 6$ mAh/g for the 20
mA/g rate profile appears at the onset of discharge (4.0 V). Since the start
of the capacity curve at the upper level of the voltage cutoff is not formally
enforced to zero by the model, but emerges from the linear embeddings of the
voltage state $V_{i}$ with the voltage window
$[V_{\text{low}},V_{\text{high}}]$, an offset can be created when there is not
enough data for that specific voltage window.
Based on the tests, our primary conclusion is that DRXNet exhibits a
reasonable ability to learn the cathode material’s chemical information in the
latent space and generalize to test conditions that are included in the
dataset. However, for test conditions that the model has not or rarely
encountered (e.g., experiments with $V_{\text{high}}<4.0$ V), discrepancies or
unphysical profiles may still arise. This highlights the data scarcity issue,
which is typical for human-generated experimental conditions, which are biased
toward what is needed to demonstrate performance rather than what is optimal
for model training [50].
## III Discussion
Most machine learning approaches predicting battery performance have been
focused on predictions for a specific chemistry or limited chemical space of
commercialized cathodes, typically, the remaining useful life forecasted from
the initial cycles [11, 51, 52]. However, the nature of battery cathode
material discovery and optimization lies in a broad domain of chemistries,
which is more challenging for AI as it needs to capture the direct (e.g.,
voltage) and indirect effect (e.g., cycle life) of chemical changes [18].
Recent studies have demonstrated the feasibility of building universal models
for atomistic modeling by harnessing more than ten years of ab-initio
calculations spanning the periodic table [53, 54, 10, 55, 56]. It becomes a
logical extension to envision universal models for the experimental discovery
of battery materials by leveraging the wealth of both ab-initio calculation
and experimental data generated on cathode materials worldwide [17, 57]. In
this work, we propose an end-to-end training pipeline to encode and learn the
(electro)chemical information of cathode materials from voltage profiles.
Focused on DRX cathodes, we data-mined years of lab-generated experimental
discharge voltage profiles and trained a universal machine-learning model
(DRXNet) to make predictions across diverse compositions. This was achieved
through a novel model design consisting of an electrochemical condition
network $\mathcal{O}$ and a state prediction network $\mathcal{F}$.
The design of the two networks promotes modularity in the architecture,
streamlining the optimization and interpretation of each network individually
and their learned features. For instance, composition is an intrinsic property
of the synthesized cathode materials, and the encoding of such features is
independent of other factors such as current density and cycle status,
rationalizing our approach to first extract the composition-only feature
$\vec{X}_{\text{comp}}$ via a GNN. Although it remains a challenge that the
composition may change as a function of current density and cycle status due
to TM dissolution and the irreversible reaction of lithium outside the
cathode, DRXNet encompasses these factors into the rate- and cycle-informed
feature vector representations. By leveraging a ResNet-inspired architecture
using skip connections [58], we achieve a more effective synthesis of the
feature vector within the latent space. This design allows for a direct
connection between the rate-informed feature, $\vec{X}_{\mathcal{O}_{1}}$, and
the prediction of the first cycle capacity. Such architecture has been proven
to boost model training and alleviate the well-known gradient vanishing
issues.
Given the inherent sequential nature of battery testing data – where
possessing information from the $N$-th cycle implies the availability of data
from the first cycle – it becomes crucial to design features that reflect this
causality. This insight leads to the formulation of the cycle-informed
feature, $\vec{X}_{\mathcal{O}_{N}}$. This feature accentuates the difference
between the first and the $N$-th cycles, guiding the prediction for the $N$-th
cycle capacity, as detailed in Eq.(6). Consequently, our loss function is
constructed for multi-task learning with both terms for the first and $N$-th
cycle capacities, ensuring the causal relationships in cycle-dependent
capacity predictions (refer to Eq.(12)). Through an ablation study on whether
to include the first cycle term, $\ell(Q^{1})$, in the loss function or not,
we found that the model without $\ell(Q^{1})$ tends to be underfitted (more
details in Figure S8). Our incorporation of loss terms for both the first and
$N$-th cycle capacities enhances the model expressibility, which is a crucial
factor in the optimization of battery materials.
In addition, the modular design of the electrochemical condition network
$(\mathcal{O})$ provides flexibility for the feature representation when
expanding the model to include other information. The training dataset, being
derived from our own experimental results, does not encompass testing
parameters such as particle size, electrolyte type, synthesis variations, etc.
Since the battery electrodes were fabricated in our laboratory using
standardized recipes and methodologies, these factors have been coarsely
integrated into the compositional model and are treated as constants across
our dataset. Currently, the model does not include features to capture
structural information (crystal structure, short-range order, etc.). In DRX
compounds, short-range order is known to influence performance and to the
extent that this is not a direct consequence of composition, but modified by
synthesis parameters its effects are not accounted for [59]. In principle,
researchers can choose to include such factors to design the electrochemical
feature vector, depending on the specific problem they are addressing. Given
the vast amount and complexity of these properties, a synthetic data
collection approach is necessary. Data-mining techniques, such as text mining
and figure mining, can automatically retrieve valuable experimental
information from decades of published literature [60, 61], though a challenge
with aggregating diverse data from literature is the numerous hidden and
unspecified variables relevant to materials synthesis and electrochemical
testing. Looking forward, automated labs can address both data scarcity and
transparency issues by enabling more extensive exploration of the experimental
space and even better collect data from ”failed” experiments [62, 63, 64, 65].
In conclusion, DRXNet represents a step forward in developing machine-learning
models for battery materials research. By continuously refining the model and
incorporating additional data and parameters, we anticipate that such a
machine-learning framework will play an increasingly critical role in
discovering and optimizing next-generation battery materials.
## IV Methods
### IV.1 Data collection
We collected coin-cell electrochemical test data from our lab starting in 2016
and converted them into a digital format (.json). Each .json file contains
information on one individual electrochemical test, including the electrode
composition, electrode mass (g), active mass (g), test current rate (mA/g),
low and high voltage value of the working window (V), and charge/discharge
profiles of $N_{\text{cycle}}$ collected cycles. The compositions used for the
model training were taken as the targeted composition in experiments. For the
fraction of our data set which was previously published, the composition
values were typically confirmed by Inductively Coupled Plasma (ICP) analysis.
For these compounds, the feature vectors of the targeted compositions and ICP-
analyzed composition exhibit $\geq 99.7\%$ in cosine similarity as shown in
Supplementary Information, which supports using the targeted composition for
the general prediction purpose. Nonetheless, minor variations between the
actual composition and the target composition can be a source of noise in the
data.
For the in-house battery tests, the CR2032 coin cells were assembled using
commercial 1 M LiPF6 in an ethylene carbonate and dimethyl carbonate solution
(volume ratio 1:1) as the electrolyte, glass microfiber filters (Whatman) as
separators, and Li-metal foil (FMC) as the anode. The coin cells were tested
on an Arbin battery cycler at room temperature. The cathode consisted of a
mixture of active material (DRX), Super C65 carbon black, and
polytetrafluoroethylene (PTFE). The capacity signal, collected in units of Ah
from the Arbin battery cycler, was normalized to mAh/g using the mass of the
active material (active mass). The data from the failed tests (e.g., Arbin
cycler breakdown, electrolyte failure, strong signal fluctuations, etc.) were
removed from the dataset (see Figure S1 for examples).
To enhance the generalization and expressibility of DRXNet, we expanded the
dataset by figure mining published voltage profiles in related systems not
covered by our lab tests (see Table S1), which was accomplished using the
WebPlotDigitizer [66]. We used the UnivariateSpline method to denoise and
resample the experimental profiles and compute the $dQ/dV$ curves. One hundred
points were uniformly sampled to form the voltage series
$\boldsymbol{V}=\left[V_{0},V_{1},...,V_{i},...\right]$ for each discharge
profile, and the capacity series and $dQ/dV$ series were calculated
accordingly from $\boldsymbol{V}$.
### IV.2 Model design
#### IV.2.1 Preliminaries
A linear layer with trainable weight $\boldsymbol{W}$ and bias
$\boldsymbol{b}$ is defined as
$L(\vec{X})=\vec{X}\boldsymbol{W}+\boldsymbol{b}.$ (2)
For simplicity of notion, each $L$ represents different trainable weights in
the following equations.
#### IV.2.2 Compositional encoding
For elemental information, each element is first embedded into a
200-dimensional vector using mat2vec [41]. The Roost (Representation Learning
from Stoichiometry) model is used for compositional encoding [26], which is a
graph neural network (GNN) with message passings as follows:
$\displaystyle\vec{h}_{i}^{t+1}$
$\displaystyle=\vec{h}_{i}^{t}+\sum_{j,m}a_{i,j}^{t,m}\cdot\sigma_{g}\circ
L_{c}\left(\vec{h}_{i}^{t}||\vec{h}_{j}^{t}\right),$ (3) $\displaystyle
a_{i,j}^{t,m}$
$\displaystyle=\frac{w_{j}\exp(e_{i,j}^{t,m})}{\sum_{k}w_{k}\exp(e_{i,k}^{t,m})},~{}e_{i,k}^{t,m}=\sigma_{g}\circ
L_{a}\left(\vec{h}_{i}^{t}||\vec{h}_{j}^{t}\right).$
In these equations, $\vec{h}_{i}^{t}$ represents the $t$-th hidden layer for
the $i$-th element; $||$ denotes the concatenation operation; and the soft-
attention coefficient $a_{i,j}^{t,m}$ describes the interaction between
elements $i$ and $j$, with $m$ as the index of multi-head attention. $L_{c}$
and $L_{a}$ denote the linear layer for the core and attention layer,
respectively. The fractional concentration $w_{j}$ of element $j$ depends on
the specific compound (e.g., $w_{j}=0.6/0.1/0.1/0.2$ for Li/Mn/Cr/Ti in
Li1.2Mn0.2Cr0.2Ti0.4O2.0). $\sigma_{g}$ is the SiLu activation function. After
$n$ graph convolution layers, the encoded composition vector
$\vec{X}_{\text{comp}}$ is obtained by average pooling over the elements with
weighted attention
$\vec{X}_{\text{comp}}=\text{Pooling}\left[\frac{w_{i}\exp\left(\sigma_{g}\circ
L_{a}(\vec{h}_{i}^{n})\right)}{\sum_{k}\exp\left(\sigma_{g}\circ
L_{a}(\vec{h}_{i}^{n})\right)}\cdot\left(\sigma_{g}\circ
L_{c}(\vec{h}_{i}^{n})\right)\right]$ (4)
#### IV.2.3 Electrochemical condition encoding
The electrochemical test primarily involves two pieces of information: the
current density rate and cycle number. We use MLPs to encode the rate and
cycle number:
$\vec{X}_{\text{rate}}=\sigma_{g}\circ
L(\text{rate}),~{}\vec{X}_{\text{cycle}}=\sigma_{g}\circ L(\text{cycle}).$ (5)
As the actual rate and cycle performance are strongly correlated with cathode
materials, the relationship between the composition, rate, and cycle is
synthesized using gated-MLPs with soft attention[25]:
$\displaystyle\vec{X}_{\mathcal{O}_{1}}$
$\displaystyle=\vec{X}_{\text{comp}}+\sigma_{f_{1}}(\vec{X}_{\text{comp}}||\vec{X}_{\text{rate}})\cdot
f_{1}(\vec{X}_{\text{comp}}||\vec{X}_{\text{rate}})$ (6)
$\displaystyle\vec{X}_{\mathcal{O}_{N}}$
$\displaystyle=\vec{X}_{\mathcal{O}_{1}}+\sigma_{f_{2}}(\vec{X}_{\mathcal{O}_{1}}||\vec{X}_{\text{cycle}})\cdot
f_{2}(\vec{X}_{\mathcal{O}_{1}}||\vec{X}_{\text{cycle}})$
$\displaystyle\quad\cdot\boldsymbol{W}_{n}(N-1)$
where $\sigma_{f}=\sigma_{s}\circ B\circ L$ is an MLP, $\sigma_{s}$ is the
Sigmoid activation function, and $f=\sigma_{g}\circ B\circ L$ is an MLP with
SiLu activation function $\sigma_{g}$. The BatchNormalization layer $B$ is
added before the activation function. In this equation,
$\vec{X}_{\mathcal{O}_{1}}$ is a feature vector jointly determined by the
composition and rate information, which is used to predict the first cycle
property. $\vec{X}_{\mathcal{O}_{N}}$ is a feature vector jointly determined
by the composition, rate, and cycle information, which is used to predict the
$N$-th cycle property. The difference between $\vec{X}_{\mathcal{O}_{1}}$ and
$\vec{X}_{\mathcal{O}_{N}}$ is linearly dependent on the number of cycles with
a trainable weight $\boldsymbol{W}_{n}$, allowing the model to learn cycle
performance contrastively.
#### IV.2.4 State prediction network
The state prediction network ($\mathcal{F}$) takes the inputs of voltage state
($V_{i}$) and outputs the discharge-capacity state ($Q_{i}$)
$Q_{i}=\mathcal{F}\left(V_{i}|\mathcal{O}\right).$ (7)
In practice, the voltage profile is measured within the applied voltage window
[$V_{\text{low}},V_{\text{high}}$]. To accommodate the voltage window in the
discharge state prediction, the first layer in $\mathcal{F}$ is encoded via an
MLP:
$\displaystyle\vec{Z}_{i}^{0}$
$\displaystyle=L\circ\sigma_{\mathcal{F}}\circ\left[L(V_{\text{low}},V_{\text{high}})+L(V_{i})\right],$
(8)
where $\sigma_{\mathcal{F}}(\cdot)$ is the $\mathtt{Softplus}$ activation
function. The test-condition information is element-wise added to the state
prediction network [58]
$\vec{Z}_{i}^{N}=\sigma_{\mathcal{F}}\circ
L\left(\vec{Z}_{i}^{0}+\vec{X}_{\mathcal{O}_{N}}\right)$ (9)
The state of capacity is obtained by
$Q^{N}_{i}=\sigma_{\mathcal{F}}\circ L\circ\sigma_{\mathcal{F}}\circ
L(\vec{Z}_{i}^{N})$ (10)
where $Q^{N}_{i}$ is the capacity for the $N$-th cycle (including the first
cycle). Because the discharge capacity is always positive,
$\sigma_{\mathcal{F}}$ is added to constrain the predicted capacity to be
positive and accelerate the training process. $dQ/dV$ for the redox potential
can be obtained via PyTorch auto-differentiation [67]
$\left.\frac{dQ}{dV}\right|_{i}=\text{AutoDiff}(Q_{i},V_{i}).$ (11)
### IV.3 Model training
The model is trained to minimize the sum of multi-task losses for the capacity
of the first cycle, the $n$-th cycle, and $dQ/dV$:
$\mathcal{L}=w_{Q}\ell(Q_{i}^{N})+w_{dQ}\ell(\frac{dQ^{N}}{dV_{i}})+w_{Q_{1}}\ell(Q_{i}^{1})+\mathcal{R}.$
(12)
A MSE loss function is used for $\ell(Q_{i}^{N})$ and
$\ell(\frac{dQ^{N}}{dV_{i}})$, whereas a MAE loss function is employed for the
first cycle as a contrastive term $\ell(Q_{i}^{1})$. The weights for
$Q_{i}^{N}$, $dQ/dV$, and $Q_{i}^{1}$ are set to $w_{Q}$ = 1, $w_{dQ}$ = 1,
and $w_{Q_{1}}$ = 5. The term $\mathcal{R}$ represents regularization, which
consists of two parts: (1) an $\ell_{2}$-norm regularization of the network’s
parameters $||\boldsymbol{\theta}||_{2}$ and (2) a smoothing term
$||dQ/d\textbf{c}||_{2}$ to avoid large, unphysical performance fluctuations
(c denotes the fractional concentration of elements). The weight of
regularization is $10^{-4}$.
To make predictions, an ensemble of five independent models was trained to
make predictions. Each model was trained with a batch size of 1024 within 30
epochs. The Adam optimizer was used with $10^{-3}$ as the initial learning
rate. The ExponentialLR scheduler was used to adjust the learning rate with a
decay of 0.9 per epoch.
## V Acknowledgments
This work was primarily supported by the U.S. Department of Energy, Office of
Science, Office of Basic Energy Sciences, Materials Sciences and Engineering
Division under Contract No. DE-AC0205CH11231 (Materials Project program
KC23MP). The data collection in this work was supported by the Assistant
Secretary for Energy Efficiency and Renewable Energy, Vehicle Technologies
Office, under the Advanced Battery Materials Research (BMR) Program of the US
Department of Energy (DOE) under contract No. DE-AC0205CH11231. The
computational modeling in this work was supported by the computational
resources provided by the Extreme Science and Engineering Discovery
Environment (XSEDE), supported by National Science Foundation grant number
ACI1053575; the National Energy Research Scientific Computing Center (NERSC);
and the Lawrencium computational cluster resource provided by the IT Division
at the Lawrence Berkeley National Laboratory. The authors thank Huiwen Ji,
Jianping Huang, and Zijian Cai for their help in experimental data collection,
and Yifan Chen for valuable discussions.
## VI Availability
The codes of DRXNet are open-sourced at https://github.com/zhongpc/DRXNet and
https://doi.org/10.5281/zenodo.10719829. The open-source dataset is available
at https://doi.org/10.6084/m9.figshare.25328578.v1 for public access, which
contains 12,688 experimental discharge voltage profiles excluding the Mn-rich
and Ti-based DRX. The open-source dataset is not identical to, but rather a
part of, the DRX-TD that was used for the pretrained models in the paper.
## VII Supplementary Information
Supplemental information can be found online at
https://doi.org/10.1016/j.joule.2024.03.010.
## VIII Author Contributions
P.Z. and G.C. conceived the initial idea. P.Z. collected the dataset and
developed the code base with help from B.D. and T.H.. Z.L. and G.C. offered
insight into the project. P.Z. and G.C. prepared the manuscript. All authors
contributed to discussions and approved the manuscript.
## References
* Olivetti _et al._ [2017] E. A. Olivetti, G. Ceder, G. G. Gaustad, and X. Fu, Lithium-Ion Battery Supply Chain Considerations: Analysis of Potential Bottlenecks in Critical Metals, Joule 1, 229 (2017).
* Xie _et al._ [2021] L. Xie, C. Singh, S. K. Mitter, M. A. Dahleh, and S. S. Oren, Toward carbon-neutral electricity and mobility: Is the grid infrastructure ready?, Joule 5, 1908 (2021).
* Tian _et al._ [2021] Y. Tian, G. Zeng, A. Rutt, T. Shi, H. Kim, J. Wang, J. Koettgen, Y. Sun, B. Ouyang, _et al._ , Promises and Challenges of Next-Generation “Beyond Li-ion” Batteries for Electric Vehicles and Grid Decarbonization, Chemical Reviews 121, 1623 (2021).
* Goodenough and Kim [2010] J. B. Goodenough and Y. Kim, Challenges for rechargeable Li batteries, Chemistry of Materials 22, 587 (2010).
* Lv _et al._ [2022] C. Lv, X. Zhou, L. Zhong, C. Yan, M. Srinivasan, Z. W. Seh, C. Liu, H. Pan, S. Li, Y. Wen, and Q. Yan, Machine Learning: An Advanced Platform for Materials Development and State Prediction in Lithium‐Ion Batteries, Advanced Materials 34, 2101474 (2022).
* Wang _et al._ [2023] H. Wang, T. Fu, Y. Du, W. Gao, K. Huang, Z. Liu, P. Chandak, S. Liu, P. Van Katwyk, _et al._ , Scientific discovery in the age of artificial intelligence, Nature 620, 47 (2023).
* Ahmad _et al._ [2018] Z. Ahmad, T. Xie, C. Maheshwari, J. C. Grossman, and V. Viswanathan, Machine Learning Enabled Computational Screening of Inorganic Solid Electrolytes for Suppression of Dendrite Formation in Lithium Metal Anodes, ACS Central Science 4, 996 (2018).
* Liow _et al._ [2022] C. H. Liow, H. Kang, S. Kim, M. Na, Y. Lee, A. Baucour, K. Bang, Y. Shim, J. Choe, _et al._ , Machine learning assisted synthesis of lithium-ion batteries cathode materials, Nano Energy 98, 107214 (2022).
* Chen _et al._ [2024] C. Chen, D. T. Nguyen, S. J. Lee, N. A. Baker, A. S. Karakoti, L. Lauw, C. Owen, K. T. Mueller, B. A. Bilodeau, V. Murugesan, _et al._ , Accelerating computational materials discovery with artificial intelligence and cloud high-performance computing: from large-scale screening to experimental validation, arXiv preprint arXiv:2401.04070 (2024).
* Deng _et al._ [2023] B. Deng, P. Zhong, K. Jun, J. Riebesell, K. Han, C. J. Bartel, and G. Ceder, CHGNet as a pretrained universal neural network potential for charge-informed atomistic modelling, Nature Machine Intelligence 5, 1031 (2023).
* Severson _et al._ [2019] K. A. Severson, P. M. Attia, N. Jin, N. Perkins, B. Jiang, Z. Yang, M. H. Chen, M. Aykol, P. K. Herring, _et al._ , Data-driven prediction of battery cycle life before capacity degradation, Nature Energy 4, 383 (2019).
* Jones _et al._ [2022] P. K. Jones, U. Stimming, and A. A. Lee, Impedance-based forecasting of lithium-ion battery performance amid uneven usage, Nature Communications 13, 1 (2022).
* Aykol _et al._ [2021] M. Aykol, C. B. Gopal, A. Anapolsky, P. K. Herring, B. van Vlijmen, M. D. Berliner, M. Z. Bazant, R. D. Braatz, W. C. Chueh, and B. D. Storey, Perspective—Combining Physics and Machine Learning to Predict Battery Lifetime, Journal of The Electrochemical Society 168, 030525 (2021).
* Hu _et al._ [2020] X. Hu, L. Xu, X. Lin, and M. Pecht, Battery lifetime prognostics, Joule 4, 310 (2020).
* Sulzer _et al._ [2021] V. Sulzer, P. Mohtat, A. Aitio, S. Lee, Y. T. Yeh, F. Steinbacher, M. U. Khan, J. W. Lee, J. B. Siegel, _et al._ , The challenge and opportunity of battery lifetime prediction from field data, Joule 5, 1934 (2021).
* Attia _et al._ [2020] P. M. Attia, A. Grover, N. Jin, K. A. Severson, T. M. Markov, Y.-H. Liao, M. H. Chen, B. Cheong, N. Perkins, _et al._ , Closed-loop optimization of fast-charging protocols for batteries with machine learning, Nature 578, 397 (2020).
* Ward _et al._ [2022] L. Ward, S. Babinec, E. J. Dufek, D. A. Howey, V. Viswanathan, M. Aykol, D. A. Beck, B. Blaiszik, B.-R. Chen, _et al._ , Principles of the Battery Data Genome, Joule 6, 2253 (2022).
* Sendek _et al._ [2022] A. D. Sendek, B. Ransom, E. D. Cubuk, L. A. Pellouchoud, J. Nanda, and E. J. Reed, Machine Learning Modeling for Accelerated Battery Materials Design in the Small Data Regime, Advanced Energy Materials 12, 2200553 (2022).
* Clément _et al._ [2020] R. J. Clément, Z. Lun, and G. Ceder, Cation-disordered rocksalt transition metal oxides and oxyfluorides for high energy lithium-ion cathodes, Energy & Environmental Science 13, 345 (2020).
* Key _et al._ [2011] B. Key, M. Morcrette, J.-M. Tarascon, and C. P. Grey, Pair Distribution Function Analysis and Solid State NMR Studies of Silicon Electrodes for Lithium Ion Batteries: Understanding the (De)lithiation Mechanisms, Journal of the American Chemical Society 133, 503 (2011).
* Li _et al._ [2021] L. Li, Z. Lun, D. Chen, Y. Yue, W. Tong, G. Chen, G. Ceder, and C. Wang, Fluorination‐Enhanced Surface Stability of Cation‐Disordered Rocksalt Cathodes for Li‐Ion Batteries, Advanced Functional Materials 31, 2101888 (2021).
* Clément _et al._ [2018] R. J. Clément, D. Kitchaev, J. Lee, and Gerbrand Ceder, Short-Range Order and Unusual Modes of Nickel Redox in a Fluorine-Substituted Disordered Rocksalt Oxide Lithium-Ion Cathode, Chemistry of Materials 30, 6945 (2018).
* Zhong _et al._ [2022] P. Zhong, T. Chen, L. Barroso-Luque, F. Xie, and G. Ceder, An $\ell_{0}\ell_{2}$-norm regularized regression model for construction of robust cluster expansions in multicomponent systems, Physical Review B 106, 024203 (2022).
* Barroso-Luque _et al._ [2022] L. Barroso-Luque, P. Zhong, J. H. Yang, F. Xie, T. Chen, B. Ouyang, and G. Ceder, Cluster expansions of multicomponent ionic materials: Formalism and methodology, Physical Review B 106, 144202 (2022).
* Xie and Grossman [2018] T. Xie and J. C. Grossman, Crystal Graph Convolutional Neural Networks for an Accurate and Interpretable Prediction of Material Properties, Physical Review Letters 120, 145301 (2018).
* Goodall and Lee [2020] R. E. A. Goodall and A. A. Lee, Predicting materials properties without crystal structure: deep representation learning from stoichiometry, Nature Communications 11, 6280 (2020).
* Mizushima _et al._ [1980] K. Mizushima, P. Jones, P. Wiseman, and J. Goodenough, LixCoO2 ($0<x<1$): A new cathode material for batteries of high energy density, Materials Research Bulletin 15, 783 (1980).
* Manthiram [2020] A. Manthiram, A reflection on lithium-ion battery cathode chemistry, Nature Communications 11, 1550 (2020).
* Ji _et al._ [2019a] H. Ji, A. Urban, D. A. Kitchaev, D.-H. Kwon, N. Artrith, C. Ophus, W. Huang, Z. Cai, T. Shi, _et al._ , Hidden structural and chemical order controls lithium transport in cation-disordered oxides for rechargeable batteries, Nature Communications 10, 592 (2019a).
* Urban _et al._ [2017] A. Urban, A. Abdellahi, S. Dacek, N. Artrith, and G. Ceder, Electronic-Structure Origin of Cation Disorder in Transition-Metal Oxides, Physical Review Letters 119, 1 (2017).
* Lee _et al._ [2017] J. Lee, J. K. Papp, R. J. Clément, S. Sallis, D.-H. Kwon, T. Shi, W. Yang, B. D. McCloskey, and G. Ceder, Mitigating oxygen loss to improve the cycling performance of high capacity cation-disordered cathode materials, Nature Communications 8, 981 (2017).
* Zhong _et al._ [2020] P. Zhong, Z. Cai, Y. Zhang, R. Giovine, B. Ouyang, G. Zeng, Y. Chen, R. Clément, Z. Lun, and G. Ceder, Increasing Capacity in Disordered Rocksalt Cathodes by Mg Doping, Chemistry of Materials 32, 10728 (2020).
* Huang _et al._ [2021] J. Huang, P. Zhong, Y. Ha, D.-h. Kwon, M. J. Crafton, Y. Tian, M. Balasubramanian, B. D. McCloskey, W. Yang, and G. Ceder, Non-topotactic reactions enable high rate capability in Li-rich cathode materials, Nature Energy 6, 706 (2021).
* Newman and Balsara [2021] J. Newman and N. P. Balsara, _Electrochemical systems_ (John Wiley & Sons, 2021).
* Lee _et al._ [2018] J. Lee, D. A. Kitchaev, D.-h. Kwon, C.-w. Lee, J. K. Papp, Y.-s. Liu, Z. Lun, R. J. Clément, T. Shi, B. D. McCloskey, J. Guo, M. Balasubramanian, and G. Ceder, Reversible Mn2+/Mn4+ double redox in lithium-excess cathode materials, Nature 556, 185 (2018).
* Kitchaev _et al._ [2018] D. A. Kitchaev, Z. Lun, W. D. Richards, H. Ji, R. J. Clément, M. Balasubramanian, D.-H. Kwon, K. Dai, J. K. Papp, T. Lei, B. D. McCloskey, W. Yang, J. Lee, and G. Ceder, Design principles for high transition metal capacity in disordered rocksalt Li-ion cathodes, Energy & Environmental Science 11, 2159 (2018).
* Lun _et al._ [2019] Z. Lun, B. Ouyang, D. A. Kitchaev, R. J. Clément, J. K. Papp, M. Balasubramanian, Y. Tian, T. Lei, T. Shi, _et al._ , Improved Cycling Performance of Li-Excess Cation-Disordered Cathode Materials upon Fluorine Substitution, Advanced Energy Materials 9, 1802959 (2019).
* Ji _et al._ [2019b] H. Ji, D. A. Kitchaev, Z. Lun, H. Kim, E. Foley, D.-H. Kwon, Y. Tian, M. Balasubramanian, M. Bianchini, _et al._ , Computational Investigation and Experimental Realization of Disordered High-Capacity Li-Ion Cathodes Based on Ni Redox, Chemistry of Materials 31, 2431 (2019b).
* Lun _et al._ [2020] Z. Lun, B. Ouyang, Z. Cai, R. J. Clément, D.-H. Kwon, J. Huang, J. K. Papp, M. Balasubramanian, Y. Tian, _et al._ , Design Principles for High-Capacity Mn-Based Cation-Disordered Rocksalt Cathodes, Chem 6, 153 (2020).
* Lun _et al._ [2021] Z. Lun, B. Ouyang, D.-h. Kwon, Y. Ha, E. E. Foley, T.-Y. Huang, Z. Cai, H. Kim, M. Balasubramanian, _et al._ , Cation-disordered rocksalt-type high-entropy cathodes for Li-ion batteries, Nature Materials 20, 214 (2021).
* Tshitoyan _et al._ [2019] V. Tshitoyan, J. Dagdelen, L. Weston, A. Dunn, Z. Rong, O. Kononova, K. A. Persson, G. Ceder, and A. Jain, Unsupervised word embeddings capture latent knowledge from materials science literature, Nature 571, 95 (2019).
* Sutton _et al._ [2020] C. Sutton, M. Boley, L. M. Ghiringhelli, M. Rupp, J. Vreeken, and M. Scheffler, Identifying domains of applicability of machine learning models for materials science, Nature Communications 11, 4428 (2020).
* Zhou _et al._ [2023] S. Zhou, Y. Sun, T. Gao, J. Liao, S. Zhao, and G. Cao, Enhanced Li + Diffusion and Lattice oxygen Stability by the High Entropy Effect in Disordered‐Rocksalt Cathodes, Angewandte Chemie International Edition 100084, 10.1002/anie.202311930 (2023).
* Ouyang _et al._ [2020] B. Ouyang, N. Artrith, Z. Lun, Z. Jadidi, D. A. Kitchaev, H. Ji, A. Urban, and G. Ceder, Effect of Fluorination on Lithium Transport and Short‐Range Order in Disordered‐Rocksalt‐Type Lithium‐Ion Battery Cathodes, Advanced Energy Materials 10, 1903240 (2020).
* Kataoka _et al._ [2018] R. Kataoka, T. Kojima, and N. Takeichi, Electrochemical Property of Li-Mn Cation Disordered Li-Rich Li 2 MnO 3 with NaCl Type Structure, Journal of The Electrochemical Society 165, A291 (2018).
* Ren _et al._ [2015] S. Ren, R. Chen, E. Maawad, O. Dolotko, A. A. Guda, V. Shapovalov, D. Wang, H. Hahn, and M. Fichtner, Improved Voltage and Cycling for Li+ Intercalation in High-Capacity Disordered Oxyfluoride Cathodes, Advanced Science 2, 1500128 (2015).
* Nakajima and Yabuuchi [2017] M. Nakajima and N. Yabuuchi, Lithium-Excess Cation-Disordered Rocksalt-Type Oxide with Nanoscale Phase Segregation: Li1.25Nb0.25V0.5O2, Chemistry of Materials 29, 6927 (2017).
* Chen _et al._ [2015] R. Chen, S. Ren, M. Yavuz, A. A. Guda, V. Shapovalov, R. Witter, M. Fichtner, and H. Hahn, Li+ intercalation in isostructural Li2VO3 and Li2VO2F with O2- and mixed O2-/F- anions, Physical Chemistry Chemical Physics 17, 17288 (2015).
* Edström _et al._ [2004] K. Edström, T. Gustafsson, and J. Thomas, The cathode–electrolyte interface in the Li-ion battery, Electrochimica Acta 50, 397 (2004).
* Raccuglia _et al._ [2016] P. Raccuglia, K. C. Elbert, P. D. F. Adler, C. Falk, M. B. Wenny, A. Mollo, M. Zeller, S. A. Friedler, J. Schrier, and A. J. Norquist, Machine-learning-assisted materials discovery using failed experiments, Nature 533, 73 (2016).
* Zhu _et al._ [2022] J. Zhu, Y. Wang, Y. Huang, R. Bhushan Gopaluni, Y. Cao, M. Heere, M. J. Mühlbauer, L. Mereacre, H. Dai, _et al._ , Data-driven capacity estimation of commercial lithium-ion batteries from voltage relaxation, Nature Communications 13, 2261 (2022).
* Roman _et al._ [2021] D. Roman, S. Saxena, V. Robu, M. Pecht, and D. Flynn, Machine learning pipeline for battery state-of-health estimation, Nature Machine Intelligence 3, 447 (2021).
* Jain _et al._ [2013] A. Jain, S. P. Ong, G. Hautier, W. Chen, W. D. Richards, S. Dacek, S. Cholia, D. Gunter, D. Skinner, G. Ceder, and K. A. Persson, Commentary: The Materials Project: A materials genome approach to accelerating materials innovation, APL Materials 1, 011002 (2013).
* Chen and Ong [2022] C. Chen and S. P. Ong, A universal graph deep learning interatomic potential for the periodic table, Nature Computational Science 2, 718 (2022).
* Merchant _et al._ [2023] A. Merchant, S. Batzner, S. S. Schoenholz, M. Aykol, G. Cheon, and E. D. Cubuk, Scaling deep learning for materials discovery, Nature 624, 80 (2023).
* Batatia _et al._ [2023] I. Batatia, P. Benner, Y. Chiang, A. M. Elena, D. P. Kovács, J. Riebesell, X. R. Advincula, M. Asta, W. J. Baldwin, N. Bernstein, _et al._ , A foundation model for atomistic materials chemistry, arXiv preprint arXiv:2401.00096 (2023).
* Hong _et al._ [2021] S. Hong, C. H. Liow, J. M. Yuk, H. R. Byon, Y. Yang, E. Cho, J. Yeom, G. Park, H. Kang, _et al._ , Reducing Time to Discovery: Materials and Molecular Modeling, Imaging, Informatics, and Integration, ACS Nano 15, 3971 (2021).
* He _et al._ [2016] K. He, X. Zhang, S. Ren, and J. Sun, Deep Residual Learning for Image Recognition, in _2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)_, Vol. 2016-Decem (IEEE, 2016) pp. 770–778.
* Cai _et al._ [2022] Z. Cai, Y. Zhang, Z. Lun, B. Ouyang, L. C. Gallington, Y. Sun, H. Hau, Y. Chen, M. C. Scott, and G. Ceder, Thermodynamically Driven Synthetic Optimization for Cation‐Disordered Rock Salt Cathodes, Advanced Energy Materials 12, 2103923 (2022).
* Kononova _et al._ [2021] O. Kononova, T. He, H. Huo, A. Trewartha, E. A. Olivetti, and G. Ceder, Opportunities and challenges of text mining in materials research, iScience 24, 102155 (2021).
* Baibakova _et al._ [2022] V. Baibakova, M. Elzouka, S. Lubner, R. Prasher, and A. Jain, Optical emissivity dataset of multi-material heterogeneous designs generated with automated figure extraction, Scientific Data 9, 589 (2022).
* Szymanski _et al._ [2021] N. J. Szymanski, Y. Zeng, H. Huo, C. J. Bartel, H. Kim, and G. Ceder, Toward autonomous design and synthesis of novel inorganic materials, Materials Horizons 8, 2169 (2021).
* Stach _et al._ [2021] E. Stach, B. DeCost, A. G. Kusne, J. Hattrick-Simpers, K. A. Brown, K. G. Reyes, J. Schrier, S. Billinge, T. Buonassisi, _et al._ , Autonomous experimentation systems for materials development: A community perspective, Matter 4, 2702 (2021).
* Szymanski _et al._ [2023] N. J. Szymanski, B. Rendy, Y. Fei, R. E. Kumar, T. He, D. Milsted, M. J. McDermott, M. Gallant, E. D. Cubuk, _et al._ , An autonomous laboratory for the accelerated synthesis of novel materials, Nature 624, 86 (2023).
* Lunt _et al._ [2024] A. Lunt, H. Fakhruldeen, G. Pizzuto, L. Longley, A. White, N. Rankin, R. Clowes, B. Alston, L. Gigli, _et al._ , Modular, Multi-Robot Integration of Laboratories: An Autonomous Workflow for Solid-State Chemistry, Chemical Science 10.1039/D3SC06206F (2024).
* Rohatgi [2022] A. Rohatgi, Webplotdigitizer: Version 4.6 (2022).
* Paszke _et al._ [2019] A. Paszke, S. Gross, F. Massa, A. Lerer, J. Bradbury, G. Chanan, T. Killeen, Z. Lin, N. Gimelshein, L. Antiga, _et al._ , Pytorch: An imperative style, high-performance deep learning library, Advances in neural information processing systems 32 (2019).
|
# PAIRING SYMMETRY AND PAIRING STATE IN FERROPNICTIDES: THEORETICAL OVERVIEW
I.I. Mazina and J. Schmalianb Code 6391, Naval Research Laboratory,
Washington, DC 20375 Iowa State University and Ames Laboratory, Ames, IA,
50011
(February 17, 2009)
###### Abstract
We review the main ingredients for an unconventional pairing state in the
ferropnictides, with particular emphasis on interband pairing due to magnetic
fluctuations. Summarizing the key experimental prerequisites for such pairing,
the electronic structure and nature of magnetic excitations, we discuss the
properties of the $s^{\pm}$ state that emerges as a likely candidate pairing
state for these materials and survey experimental evidence in favor of and
against this novel state of matter.
One fist of iron, the other of steel
If the right one don’t get you, then the left one will
Merle Travis, 16 tons
## 1 Introduction
The discovery of cuprate superconductors has changed our mentality in many
ways. In particular, the question that would have sounded moot to most before
1988, what is the symmetry of the superconducting state, is now the first
question to be asked when a new superconductor has been discovered. The pool
of potential candidates, before considered at best a mental Tetris for
theorists, had acquired a practical meaning. It has been demonstrated that
superconductivity in cuprates is $d$-wave, while in MgB2 it is multi-gap
$s$-wave with a large gap disparity. There is considerable evidence that
Sr2RuO4 is a $p$-wave material. Other complex order parameters are routinely
discussed for heavy fermion systems or organic charge transfer salts. It is
likely that the newly discovered ferropnictides represent another
superconducting state, not encountered in experiment before.
Besides the general appreciation that pairing states may be rather nontrivial,
it has also been recognized that unconventional pairing is likely due, at
least to some extent, to electronic (Coulomb or magnetic) mechanisms and,
conversely, electronic mechanisms are much more likely to produce
unconventional pairing symmetries than the standard uniform-gap $s$-wave. It
has been appreciated that the actual symmetry is very sensitive to the
momentum dependence of the pairing interaction, as well as to the underlying
electronic structure (mostly, fermiology).
Therefore we have structured this overview so that it starts with a layout of
prerequisites for a meaningful discussion of the pairing symmetry. First of
all, we shall describe the gross features of the fermiology according to
density-functional (DFT) calculations, as well as briefly assess verification
of such calculations via ARPES and quantum oscillations experiments. Again,
detailed discussion of these can be found elsewhere in this volume. We will
also point out where one may expect caveats in using the DFT band structure:
it is in our view misleading to assume that these compounds are uncorrelated.
While not necessarily of the same nature as in cuprates, considerable
electron-electron interaction effects cannot be excluded and are even
expected.
We will then proceed to discuss the role of magnetic fluctuations as well as
other excitations due to electron-electron interactions. We discuss the
special role the antiferromagnetic (AFM) ordering vector plays for the pairing
symmetry and address the on-site Coulomb (Hubbard correlations), to the extent
of their possible effect on the pairing symmetry, and possible overscreeining
(Ginzburg-Little) interactions. We also discuss puzzling issues that are
related to the magnetoelastic interaction in these systems. As for a
discussion of the electron-phonon interaction we refer to the article by Boeri
et al in this volume. The final part of this review consists of a summary of
theoretical aspects of the pairing state, along with a discussion of its
experimental manifestations.
## 2 Prerequisites for addressing the Cooper pairing
### 2.1 Electronic structure and fermiology
#### 2.1.1 Density functional calculations
The two families of the Fe-based superconductors are $1111$ systems ROFeAs
with rare earth ions R[1, 2] and the $122$ systems AFe2As2 with alkaline earth
element A[3]. Both families have been studied in much detail by first
principles DFT calculations. Here and below, unless specifically indicated, we
use a 2D unit cell with two Fe per cell, and the corresponding reciprocal
lattice cell; the $x$ and $y$ directions are along the next-nearest-neighbor
Fe-Fe bond. It appears that all materials share the same common motif: two or
more hole-like Fermi surfaces near the $\Gamma$ point [$\mathbf{k=}(0,0)$],
and two electron-like surfaces near the M point [$\mathbf{k=}(\pi,\pi)$] (Fig.
1-5). This is true, however, in strictly non-magnetic calculations only, when
the magnetic moment on each Fe is restricted to zero. As discussed below, this
is not necessarily a correct picture.
Figure 1: (color online) The Fermi surface of the non-magnetic LaAsFeO for 10%
e-doping [4] Figure 2: (color online) The Fermi surface of the non-magnetic
BaFe2As2 for 10% e-doping (Co doping, virtual crysatl approximation)[4]
If, however, we neglect this potential caveat, and concentrate on the two best
studied systems, 1111 and 122, the following relevant characteristics can be
pointed out: First, the density of states (DOS) for holes and electrons is
comparable for undoped materials; with doping, respectively one or the other
becomes dominant. For instance, for Ba0.6K0.4Fe2As2 the calculated DOS (in the
experimental structure) for the three hole bands varies between $1.1$
st/eV/f.u. and $1.3$ st/eV/f.u., the inner cylinder having, naturally, the
smallest DOS and the outer the largest. For the electron bands the total DOS
is $1.2$ st/eV/f.u., that is, two to three times smaller than the total for
the hole bands[4]. We shall see later that this is important. Another
interesting effect is that in the 122 family doping in either direction
strongly reduces the dimensionality compared to undoped compounds (in the 1111
family this effect exists, but is much less pronounced), see Fig. 4. This
suggests that the reason that doping destroys the long-range magnetic order
(it is believed by many that such a destruction is prerequisite for
superconductivity in ferropnictides) is not primarily due to the change in the
2D electronic structure, as it was initially anticipated[5], but rather due to
the destruction of magnetic coupling between the layers. Indeed the most
striking difference between the undoped 1111 and undoped 122 electronic
structure is quasi two-dimensionality of the former and a more 3D character of
the latter (the difference is clear already in the paramagnetic calculations,
but is particularly drastic in the antiferromagnetic state), while at the same
time the observed magnetism in the 122 family is at least three times stronger
than in LaFeAsO (in the mean-field DFT calculation the difference is quite
small).
Figure 3: (color online) The Fermi surface of the non-magnetic BaFe2As2 for
10% h-doping (20% Cs doping, virtual crysatl approximation.[4]
The fact that the nesting is very imperfect is crucial from the point of view
of an SDW instability, making the material stable against infinitesimally
small magnetic perturbation. For superconductivity, however, it is less
important, as discussed later in the paper.
Figure 4: (color online) The Fermi surface of BaFe2As2 for 20% h-doping
(corresponding to Ba1.6K0.4Fe2As2, calculated as 40% Cs doping in the virtual
crystal approximation) [4]. Figure 5: (color online) The Fermi surface of
undoped nonmagnetic FeTe. [4]
#### 2.1.2 Experimental evidence
Experimental evidence regarding the band structure and fermiology of these
materials comes, basically, from two sources: Angular resolved photoemission
spectroscopy (ARPES) and quantum oscillations measurements. The former has an
additional advantage of being capable of probing the electronic structure in
the superconducting state, assessing the amplitude and angular variation of
the superconducting gap. A potential disadvantage is that it is a surface
probe, and pnictides, especially the 122 family, are much more three-
dimensional than cuprates. This means that, first, the in-plane bands as
measured by ARPES, strongly depend on the normal momentum, $k_{\perp},$ and,
second, there is a bigger danger of surface effects in the electronic
structure than in the cuprates. There are indications that the at least in
1111 compounds the surface is charged, that is to say, the doping level in the
bulk is different from that on the surface. Additionally, LDA calculations
suggest that in the magnetic prototypes, the band structure depends
substantially on interlayer magnetic ordering, again, not surprisingly, mostly
in the 122 compounds, as Fig.6 illustrates. Of course, there is no guarantee
that the last two layers order in the same way as the bulk (or even with the
same moment).
Figure 6: (color online) Band structure of the orthorhombic antiferromagnetic
BaFe2As2 calculated for two different interlayer ordering pattern: the
experimental antiferromagnetic one (space group #66, broken green) and the
hypothetical ferromagnetic (still antiferromagnetic in plane, space group #67,
solid red). In both cases the magnetic moment on Fe was artifically suppressed
to 1 $\mu_{B}$ by aplying a fictitious negative Hubbard U [4]. The point N is
above the point Y.
These caveats notwithstanding, ARPES has already provided invaluable
information. ARPES measurements have been performed for both 1111[6, 7] and
122 materials[8, 9, 10, 11]. These measurements demonstrated the existence of
a well-defined Fermi surface that consists of hole and electron pockets, in
qualitative agreement with the predictions of electronic structure
calculations. Thus, one can say that the topology of the Fermi surface,
including the location and the relative size of the individual Fermi surface
sheets agrees with the LDA expectation — which is most important for the
pairing models. Similarly, it is rather clear that the ARPES bandwidth is
reduced from the LDA one by a factor of 2–2.5, similar to materials with
strong itinerant magnetic fluctuations (cf., for instance, Sr2RuO4 near a
magnetic quantum critical point[12]). These findings are also consistent with
the deduced normal state linear specific heat coefficient in 1111 materials
(e.g., $4-6$ mJ/mol K2 in Ref. [13]) corresponding to a factor 1–2 compared to
the bare LDA value[14]. However, in the 122 compound a specific heat
coefficient 63 mJ/mol K2 was reported[13], to be compared with roughly
11.5mJ/mol K2 from the LDA calculations[4]. While a renormalization of 5.5 is
not consistent with either ARPES or quantum oscillations, consistency among
different experimental publications for the 122 systems is lacking as well
[15, 13].
Another experimental probe of the electronic structure is based on quantum
oscillations that measure extremal cross-section areas of the FS (ideally, for
different directions of the applied field) and the effective masses. Such
measurements are very sensitive to the sample quality, therefore so far only a
handful of results are available. However, data on the P-based 1111 compound
agree reasonably well with band structure calculations[16], and indicate the
same mass renormalization as ARPES[17]
Importantly, quantum oscillations measurements on AFM 122 compounds[18, 19]
indicate that even the undoped pnictides are well defined Fermi liquids, even
though a significant portion of the Fermi surface disappears due to the
opening of a magnetic gap. The frequencies of the magneto-oscillations then
suggest that the ordered magnetic state has small Fermi surface pockets
consistent with the formation of a spin-density wave. Thus, the electronic
structure of the pnictides is consistent with a metallic state with well
defined Fermi surfaces.
Besides determining the overall shape of the Fermi surface sheets, ARPES is
able to yield crucial information about the momentum dependence of the
superconducting gap. Several groups performed high quality ARPES measurements
of this effect[7, 8, 9, 10]. In some cases significant differences in the size
of the gap amplitude for different Fermi surface sheets have been observed.
However, there seems to be a consensus between all ARPES groups that the gap
amplitude on an individual Fermi surface sheet depends weakly on the
direction. While this seems to favor a pairing state without nodes, one has to
keep in mind that all measurements so far have been done for fixed values of
the momentum $k_{\perp}$, perpendicular to the planes. While it might be
premature to place too much emphesis on the relative magnitude of the gaps
observed in different bands in ARPES experiments, it is worth noting that most
experimentalists agree that in the hole-doped 122 material the inner hole
barrel and the electron barrel have comparable (and large) superconducting
gaps, while the outer hole barrel has about twice smaller gap. On the other
hand, there are first data[20] indicating that in the electron doped
BaFe1.85Co0.15As2 the hole and the electron bands have about the same gap
despite the hole pockets shrinking, and electron pocket extending. Even more
interesting, the most natural interpretation of the measured fermiology is
that the hole FS in BaFe1.85Co0.15As2 actually corresponds to the outer
($xz/yz)$ barrel in Ba0.6K0.4Fe2As2 that has a small gap in that compound.
#### 2.1.3 Role of spin fluctuations in electronic structure
As is clear from the above discussion, strong spin fluctuations have a
substantial effect upon the band structure. First of all, they dress one-
electron excitations providing mass renormalization, offering an explanation
for the factor 2–2.5. This is in fact a relatively modest renormalization: it
is believed that, for instance, in He3 or in Sr2RuO4 itinerant spin
fluctuations provide renormalization of a factor of 4 or larger. However, it
is likely that the effect goes beyond simple mass renormalziation. As will be
discussed in detail below, there is overwhelming evidence of large local
moments on Fe, mostly from the fact that the Fe-As bond length corresponds to
a fully magnetic (large) Fe ion. There is also evidence that the in-plane
moments are rather well correlated in the planes, and the apparent loss of the
long-range ordering above $T_{N}$ is mainly due to a loss of 3D coherency
between the planes[21]. It is only natural to expect a similar situation to be
true when magnetism is suppressed by doping.
If that is the case, the electronic structure in the paramagnetic parts of the
phase diagram, at least in the vicinity of the transition, should not be
viewed as dressed nonmagnetic band, but rather as an average between the bands
corresponding to various magnetic 3D stackings (cf. Fig. 6). Fig. 6,
corresponding to the $T=0$ magnetic moment of 1 $\mu_{B},$ is probably
exaggerating this effect, but it is still likely that in a considerable range
of temperatures and doping near the observed magnetic phase boundary a
nonmagnetic band structure is not a good starting point, and a theory based on
magnetic precursors is needed. More experiments, particularly using diffuse
scattering, and more theoretical work are needed to clarify the issue. A
discussion to this effect may be found in Ref. [22]. See also Section 2.3
below.
### 2.2 Magnetic excitations
#### 2.2.1 Experimental evidence
Compared to cuprates and other similar compounds, two peculiarities strike the
eye. First, the parent compounds of the pnictide superconductors assume an
antiferromagnetic structure, where neighboring Fe moments are parallel along
one direction withinin the FeAs plane and antiparallel along the other.
Neutron scattering data yield ordered moments per Fe of $0.35\mu_{B}$ for
LaFeAsO[23], $0.25\mu_{B}$ for NdFeAsO[24], $0.8\mu_{B}$ for CeFeAsO[25], and
$0.9$ $\mu_{B}$ for BaFe2As2[26]. Intriguingly, in NdFeAsO the ordered moment
at very low temperatures increases by a factor of 3 to 4 at the temperature
corresponding to the ordering of Nd-spins[27]. Note that the correct magnetic
structure has been theoretically predicted by DFT calculations[5, 28], which,
moreover, consistently overestimated the tendency to magnetism (as opposed to
the cuprates). Second, the magnetically ordered state remains metallic. As
opposed to cuprates or other transition metal oxides, the undoped systems
exhibit a small but well established Drude conductivity[29], display magneto-
oscillations[18] and have Fermi surface sheets of a partially gapped metallic
antiferromagnetic state[30]. Above the magnetic ordering temperature a sizable
Drude weight, not untypical for an almost semimetal has been observed.
Further, the ordered Fe magnetic moment in the 1111 systems depends
sensitively on the rare earth ion, very different from YBa2Cu3O6 where yttrium
can be substituted by various rare earth elements with hardly any effect on
the Cu moment. Note that the rare earth sites project onto the centers of the
Fe plaquettes and thus do not exchange-couple with the latter by symmetry.
Finally, the magnetic susceptibility of BaFe2As2 single crystals[31] above the
magnetic transition shows no sign for an uncoupled local moment behavior.
#### 2.2.2 Itinerant versus local magnetism
The vicinity of superconductivity to a magnetically ordered state is the key
motivation to consider pairing mechanisms in the doped systems that are linked
to magnetic degrees of freedom. Similar to cuprate superconductors, proposals
for magnetic pairing range from quantum spin fluctuations of localized
magnetic moments to fluctuations of paramagnons as expected in itinerant
electron systems. To judge whether the magnetism of the parent compounds is
localized or itinerant (or located in the crossover regime between these two
extremes) is therefore crucial for the development of the correct description
of magnetic excitations and possibly the pairing interactions in the doped
systems.
In our view the case at hand is different from such extreme cases as undoped
cuprate on one end and weak itinerant magnets like ZrZn2 on the other. While
being metals with partially gapped Fermi surface, there is evidence that Fe
ions are in a strongly magnetic states with strong Hund rule coupling for Fe.
This results in a large magnetic moment — but only for some particular
ordering patterns (for comparison, in FeO and similar materials LDA produce
large magnetic moment regardless of the imposed long range order). While it is
obvious that ferropnictides are not Mott insulators with localized spins,
interacting solely with near neighbors, a noninteracting electron system may
be not a perfect starting approximation either. To make progress we have to
decide what is the lesser of two evils and use it, even realizing the problems
with the selected approach. Given the above mentioned experimental facts, our
preference is that these systems are still on the itinerant side.
A feature that has attracted much interest is the quasi-nesting between the
electron and the hole pockets. The word “quasi” is instrumental here: even the
arguably most nested undoped LaFeAsO is very far from the ideal nesting and
even worse in the (more magnetic) BaFe2As${}_{2}.$ Indeed, it has been
observed that in the LDA calculations the nonmagnetic structure in either
compound is stable with respect to an infinitesimally small AFM perturbations,
but strongly unstable with respect to finite amplitude perturbations. This can
be understood from the point of view of the Stoner theory, applied to a finite
wave vector Q: the renormalized static spin susceptibility (in the DFT the RPA
approximation is formally exact) can be written as
$\chi_{LDA}(\mathbf{Q)}=\frac{\chi_{0}(\mathbf{Q})}{1-I\chi_{0}(\mathbf{Q})},$
(1)
where $I$ is the Stoner factor of iron, measuring the intra-atomic Hund
interaction (in the DFT, it is defined by the second variation of the
exchange-correlation functional with respect to the spin density). While the
denominator in Eq. 1 provides a strong enhancement of $\chi$, albeit not
exactly at $\mathbf{Q=}(\pi,\pi)$, but at a range of the wave vectors near
$\mathbf{Q}$), it does not by itself generate an instability. One can say that
an infinitesimally weak magnetization can only open a gap over a very small
fraction of the Fermi surface. However, a large-amplitude spin density wave
opens a gap of the order of the exchange splitting, $IM$, where $M$ is the
magnetic moment on iron, and, obviously, affects most of the conducting
electrons. In other words, the magnetism itself is generated by the strong
Hund rule coupling on Fe (just as in the metal iron), but the topology of the
Fermi surface helps select the right ordering pattern. Formation of the
magnetic moments is local; arranging them into a particular pattern is
itinerant.
There are several corollaries of this fact that are important for pairing and
superconductivity. First, despite the fact that the overall physics of these
materials is more on the itinerant side than on the localized side (see a
discussion to this effect later in the paper), it is more appropriate to
consider magnetic moments on Fe as local rather than itinerant (as for
instance in the classical spin-Peierls theory). Note that the same is true for
the metal iron as well. Second, the interaction among these moments is not
local, as for instance in superexchange systems (it appears impossible to map
the energetics of the DFT calculations onto a two nearest neighbor Heisenberg
model[32]). The AFM vector is not determined by local interactions in real
space (as for instance in the $J_{1}+J_{2}$ models, see below), but by the
underlying electronic structure in reciprocal space. Third, since the energy
gain due to formation of the SDW mainly occurs at finite (and large, $IM$ is
on the order of eV) energies, looking solely at the FS may be misleading.
Indeed, FeTe is one compound where the Fe moments apparently do not order into
a $\mathbf{Q=}(\pi,\pi)$ SDW, but in a more complex structure corresponding to
a different ordering vector[33], despite the fact that the FS shows about the
same degree of nesting (Fig.5) as LaFeAsO and a noticeably better nesting than
BaFe2As${}_{2}.$ DFT calculations correctly identify the ground state in all
these cases, and the origin can be traced down again to the opening of a
partial gap: in both 1111 and 122 compounds the $\mathbf{Q=}(\pi,\pi)$ is
about the only pattern that opens such a gap around the Fermi level, while in
FeTe comparable pseudogaps open in both magnetic structures (and the
calculated energies are very close, the actual experimental structure being
slightly lower[34]).
#### 2.2.3 Perturbative itinerant approach
Even if one accepts the point of view that the magnetism in the Fe-pnictides
is predominantly itinerant, the development of an adequate theory for the
magnetic fluctuation spectrum is still highly nontrivial. As pointed out
above, there are strong arguments that the driving force for magnetism is not
Fermi surface nesting but rather a significant local Hund’s and exchange
coupling. This can be quantitatively described in terms of a multiband Hubbard
type interaction of the Fe-$3d$ states
$\displaystyle H_{int}$
$\displaystyle=U\sum_{i,a}n_{ia\uparrow}n_{ia\downarrow}+U^{\prime}\sum_{i,a>b}n_{ia}n_{ib}$
$\displaystyle-
J_{H}\sum_{i,a>b}\left(2\mathbf{s}_{ia}\cdot\mathbf{s}_{ib}+\frac{1}{2}n_{ia}n_{ib}\right)$
$\displaystyle+J\sum_{i,a>b,\sigma}d_{ia\sigma}^{\dagger}d_{ia\overline{\sigma}}^{\dagger}d_{ib\overline{\sigma}}d_{ib\sigma},$
(2)
with intra- and inter-orbital Coulomb interaction $U$ and $U^{\prime}$, Hund’s
coupling $J_{H}$ and exchange coupling $J$, respectively. Here $a$, $b$ refer
to the orbitals in a Wannier type orbital at site $i$. $X$-ray absorption
spectroscopy measurements support large values for the Hund’s couplings that
lead to a preferred high spin configuration,[35] leading to larger values of
$J_{H}$. The importance of the Hund coupling for the normal state behavior of
the pnictides was recently stressed in Ref.[36].
Weak coupling expansions in these interaction parameters may not capture
quantitative aspects of the magnetism in the pnictides. Nevertheless, it is
instructive to summarize the main finding of the result of weak coupling
expansions, in particular as they demonstrate the very interesting and
nontrivial aspects that results from interband interactions with almost nested
hole and electron Fermi-surfaces[37, 38, 39]. For an ideal semimetal (two
identical hole and electron bands with the Fermi energies $E_{h}$ and $E_{e})$
all susceptibilities at the nesting vector Q diverge as $\log|E_{h}/E_{e}-1|$.
Depending on the details of electron-electron interaction this signals an
instability, at $E_{h}=E_{e},$ to a spin density wave state or to a
superconducting state for infinitesimal interaction. The corresponding
interference between particle-hole and particle-particle scattering events can
be analyzed by using a renormalization group approach. For $J_{H}=J=0$, the
authors of Ref.[38] find that at low energies the interactions are dominated
by Cooper pair-hopping between the two bands, favoring an
$s^{\pm}$-superconducting state that is fully gapped on each Fermi surface
sheet, but with opposite sign on the two sheets. It is worth pointing out that
this pairing mechanism is due to very generic interband scattering, not
necessarily due to _spin-fluctuations_ , as all particle-hole and particle-
particle scattering events enter in essentially the same matter. An
$s^{\pm}$-state was also obtained using a functional renormalization group
approach[37], where the authors argue that the pairing mechanism is due to
collective spin fluctuations that generate a pairing interaction at low
energies. The appeal of these calculations is clearly that controlled and thus
robust conclusions can be drawn. On the other hand, as discussed below, the
Fermi surface nesting is less crucial as is implied by these calculations.
Attempts to include sizable electron-electron interactions within an itinerant
electron theory are based on the partial summation of ladder and bubble
diagrams, in the spirit of Eq.1. This leads to the RPA type theory of Ref.[40,
41, 42, 43] and the fluctuation exchange approximation of multiband
systems[44, 45]. RPA calculations yield a magnetic susceptibility that is
peaked at or near $\mathbf{Q=}\left(\pi,\pi\right)$. For parameters where the
Fermi surface around $\Gamma$ is present, the dominant pairing channel is
again the $s^{\pm}$-state, while $d$-wave pairing occurs as one artificially
eliminates this sheet of the Fermi surface. The exchange of paramagnons
between Fermi surface sheets is shown to be an efficient mechanism for spin
fluctuation induced pairing. The fluctuation exchange (FLEX) approach is to
some extent a self consistent version of the RPA theory[46]. While the method
is not very reliable to address high energy features, the description of the
low energy dynamics spin response, the low energy electronic band
renormalization and, the nature of the pairing instabilityare rather reliable.
The fact that several orbitals matter in the FeAs systems is also of help as
FLEX type approaches can be formulated as theories that become exact in the
limit of large fermion flavor[47]. Refs.[44, 45] performed FLEX calculations
for the FeAs systems and find once again that the dominant pairing state is an
$s^{\pm}$-state, even though Ref.[44] also find a $d$-wave state in a regime
where the magnetic fluctuation spectrum is peaked at vectors away from
$\mathbf{Q=}\left(\pi,\pi\right)$. These authors find a solution that is
numerically close to a compact form
$\Delta\left(\mathbf{k}\right)=\Delta_{0}\cos(ak_{x})\cos(ak_{y}),$ (3)
but this form is neither required by symmetry nor can be consistently deduced
from any low-energy theory (where pairing occurs at or near the Fermi
surface). We will come back to this issue later in this review.
To summarize, numerous calculations that start from an itinerant description
of the magnetic interactions yield an $s^{\pm}$ pairing state caused by the
exchange of collective interband scattering or paramagnons.
#### 2.2.4 J1-J2 model
The initially assumed (although later refuted by the experiment[49]) absence
of the Drude weight in undoped ferropnictides has been taken as evidence for
the fact that they are in the vicinity of a Mott transition and should be
considered as bad metals with significant incoherent excitations[48]. If
correct, it is clearly appropriate to start from a theory of localized spins,
analogous to what is believed to be correct in the cuprate superconductors[50,
51] (it is worth noting that proximity to a Mott transition is a sufficient,
but not necessary condition for existence of local moments). If the dominant
magnetic interactions are between nearest and next nearest neighbor Fe-spins,
the following model describes the localized spins:
$H=J_{1}\sum_{\left\langle
i,j\right\rangle}\mathbf{S}_{i}\cdot\mathbf{S}_{j}+J_{2}\sum_{\left\langle\left\langle
i,j\right\rangle\right\rangle}\mathbf{S}_{i}\cdot\mathbf{S}_{j}$ (4)
Here, $J_{1}$ and $J_{2}$ are the superexchange interactions between two
nearest-neighbor and next-nearest-neighbor Fe sites, respectively. A
geometrical argument can be made[52, 48] that indeed the two superexchange
paths $via$ As have comparable strength (however, this argument fails to
recognize that the direct overlap between Fe orbitals in pnictides is very
large[53], thus leading to a strong enhancement of the nearest neighbor
antiferromagnetic exchange in the localized picture[54], and that in metals
superexchange is not the only and usually not the most important magnetic
interaction). When $J_{1}>2J_{2}$ the conventional Neel state has the lowest
energy, when $J_{1}<2J_{2}$ the stripe order emerging in the experiment is the
lowest magnetic state. The system is frustrated if $J_{1}=2J_{2}.$
Upon doping the poor metal (strictly the insulator) described by Eq. 4 with
charge carriers can be investigated for superconductivity, with pairing
stabilized by strong quantum spin fluctuations. In Ref.[55] a single band of
carriers was investigated leading to either $d_{x^{2}-y^{2}}+id_{xy}$ or
$d_{xy}$-pairing, depending on the carrier concentration and the precise ratio
of $J_{1}$ and $J_{2}$. A more realistic theory for the pairing in the
$J_{1}$-$J_{2}$ model in the pnictides must of course include at least two
bands and was developed in Ref.[56]. For sufficiently large $J_{2}$, the
$s^{\pm}$-state is once again the dominating pairing state. It may seem
strange that this strong coupling theory based upon the (unlikely, from the
experimental point of view) proximity to a Mott transition has essentially the
same pairing solutions ($d$-wave for one Fermi surface sheet and
$s^{\pm}$-wave for two Fermi surface sheets separated by $\mathbf{Q}$), as the
RPA calculation of [40]. In Section 3 we will explain that this is not
surprising at all and that even a totally unphysical theory may lead to
perfectly sensible results for superconductivity, as long as it has the same
structure of magnetic excitations in the reciprocal space.
### 2.3 Magneto-elastic coupling
The parent compounds exhibit a structural and a magnetic transition, strongly
suggesting that magnetoelastic coupling plays a role in the physics of
pnictides in general and in superconductivity in particular. Electronic
structure calculations for a non-magnetic state indicate that the electron-
phonon interaction in the pnictides is rather modest and definitely not
sufficient to explain superconducting transition temperatures of $50$ K[57,
5]. However, as these calculations were based on the nonmagnetic electronic
structure, effects of local magnetism on iron were entirely neglected. Indeed,
the equilibrium position of As calculated under this assumption are quite
incorrect and the force constant for the Fe-As bond is $30\%$ higher than it
should be. On the other hand, fully magnetic AFM calculations, while
overestimating the ordered moment, produce highly accurate equilibrium
structures and the force constant in agreement with experiment[22]. It was
pointed out that including soft magnetism in the calculation, i.e. magnetism
with directional and amplitude fluctuations, may substantially enhance the
electron-phonon coupling[58]. The emphasis is on “soft” : additional reduction
of the force constants of the Fe-As bonds does not come from the fact that the
moment exists, but from the fact that the amplitude of the moment depends on
the bond length. Intriguingly, in the 1111 systems the AFM transition occurs
somewhat below a structural phase transition. Both transitions seem to be of
the second order, or of very weakly first order[59]. In 122 compounds the
structural and magnetic orders emerge simultaneously through a strong first
order transition[60, 61].
In the ordered state, Fe spins are parallel along one direction and
antiparallel along the other. Since we expect the bond length for parallel and
antiparallel Fe-spin polarization to be distinct, magnetism couples strongly
to the shear strain
$\varepsilon_{\mathrm{shear}}=\varepsilon_{xy}-\varepsilon_{yx}$. Thus,
$\varepsilon_{\mathrm{shear}}\neq 0$ should invariably occur below the Neel
temperature. Experiment finds that the ferromagnetic bonds are shorter than
antiferromagnetic bonds. From the point of view of superexchange interaction
it seems somewhat surprising that ferromagnetic bonds shorten and the
superexchange-satisfied bonds expand. Yet this behavior is exactly the same as
the DFT calculations had predicted[52], and it can be traced down to one-
electron energy (the observed sign of the orthorhombic distortion simply
lowers the one-electron DOS at the Fermi level)[101].
What remains puzzling is however why in the 1111 family the structural
transition occurs above $T_{N}.$ Naively, this fact could be taken as evidence
for a hypothesis that elastic degrees of freedom are the driving force and
that magnetism is secondary. There are strong quantitative and qualitative
arguments against this view. First, numerous DFT calculations[62, 63, 22]
converge to the correct orthorhombic structure (with correct sign and
magnitude of the distortion), if performed with AFM magnetic ordering, and to
a tetragonal solution if done without magnetism. On the other hand, the
antiferromagnetism is obtained even without allowing for a structural
distortion. In other words, magnetism is essential for the distortion, but the
distortion is not needed for the magnetism.
There exists also a very general argument that demonstrates that the magnetism
is indeed primary and the structural distortion secondary. Historically the
relevant physics was first encountered in the 2D $J_{1}$-$J_{2}$ model[64],
and applied to ferropnictides in Refs.[50, 51]. Below we will reformulate this
argument form a general point of view. We begin with a unit cell that contains
two Fe sites (just as the actual cristallographic unit cell for the FeAs
trilayer). The most natural choice of the origin is in the middle between
these two Fe cites (Fig. 7a ). The coordinates of the atoms are
$\mathbf{r}_{ij}^{+}=\mathbf{R}_{ij}+\mathbf{d}$,
$\mathbf{r}_{ij}^{-}=\mathbf{R}_{ij}-\mathbf{d,}$
$\mathbf{d}=(\frac{1}{4},\frac{1}{4}),$ where $\mathbf{R}_{ij}$ ($i,j$
integer) are the coordinates of the centers of the unit cells. This naturally
implies partitioning the entire lattice into two sublattives, shown as open
and solid dots in Fig. 7a.
Both ferro- and antiferromagnetic checkerboard orderings correspond to a
$\mathbf{Q}=(0,0)$ perturbation of the uniform state, since in both cases all
unit cells remain identical. The Fourier transform of either patter contains
only momenta corresponding to the reciprocal lattice vectors. Conversely, a
spin density wave with the quasi-momentum $\mathbf{Q}=(\pi,\pi)$ corresponds
to flipping all spins in every other unit cell, as illustrated in Fig. 7b,c by
shading colors (blue cells have the magnetization density opposite to that of
the pink cells). It is evident from Fig. 7b and c that this imposes no
requirement upon the mutual orientation of the two sublattices. Again, one can
say that the susceptibility as a function of quasimomentum $\mathbf{q}$ inside
the first Brillouin zone does not describe fluctuations of the magnetic moment
of two ions in the same unit cell with respect to each other, for that purpose
one needs to know the linear response at all momenta $\mathbf{q+G}$, where
$\mathbf{G}$ is an arbitrary reciprocal lattice vector.
Figure 7: (color online) (a)Fe2 lattice with the fully symmetric unit cells
shown. The full circles denote one sublattice, the hollow ones the other.
Shading shows ordering corersponding to the vector
$\mathbf{Q=}\left(\pi,\pi\right)$ in the Fe2 lattice; for each ssublattice,
spins in the pink unit cells are opposite to the spins in the blue cells, but
relative orientation of the two sublattices is arbitrary. (b) Ordered state
with $\mathbf{Q=}\left(\pi,\pi\right)$ and with parallel orientation of the
spins in the unit cell ($\sigma=1$). (c) Same ordering vector
$\mathbf{Q=}\left(\pi,\pi\right)$, but with antiparallel orientation of the
spins in the unit cell ($\sigma=-1$).
Let us assume that the most stable mean field phase corresponds to Néel order
in each of the two sublattices. In the $J_{1}$-$J_{2}$ language that
corresponds to $J_{2}>J_{1}/2,$ in the itinerant language to an instability in
$\chi$ at $\mathbf{Q}=(\pi,\pi)$. Moreover, it is obvious from Fig. 7b,c that
in the classical ground state one sublattice does not exchange-couple at all
to the other, so the classical ground state is infinitely degenerate. this is
however not important for the following discussion, what matters is that the
two extreme cases are always degenerate, the one where two spin in the same
cell are parallel (Fig. 7b) or antiparallel (Fig. 7c). In the $J_{1}+J_{2}$
model the infinite degeneracy is reduced by quantum fluctuations, but the
double degeneracy remains, while in the LDA it is only double degenerate
already on the mean-field level[65].
It is instructive [64] to introduce two order parameters corresponding to the
Neel (checkerboard) ordering for each sublattice,
$\mathbf{m}_{\pm}=\sum_{ij}(-1)^{i+j}\mathbf{M}_{ij}^{\pm},$where
$\mathbf{M}_{ij}^{\pm}$ are the magnetic moments of the two Fe’s in the unit
cell $ij.$ Following Ref. [64] one can introduce the third (scalar) order
parameter,
$\sigma=\sum_{ij}\sigma_{ij}=\sum_{ij}\mathbf{M}_{ij}^{+}\cdot\mathbf{M}_{ij}^{-}$.
Now $\sigma>0$ corresponds to parallel orientation of the magnetization inside
the unit cell (Fig. 7b) while $\sigma<0$ refers to antiparallel orientation
(Fig. 7c). In the former case $\sigma>0$, neighboring Fe spins are parallel
along the diagonal and antiparallel along the counter-diagonal. The situation
is reversed for $\sigma<0$. These two configurations are degenerate and
correspond to the frequently discussed ’stripe’ magnetic order. In two
dimensions, according to the Mermin-Wagner theorem, $\sigma$ is the only order
parameter that can be finite at finite temperature. Therefore the presumably
largest energy scale of the system, the mean field transition temperature of
each sublattice, $T^{\ast}$ ($\sim J_{2}$ in the local model, and the energy
difference $E_{FM}-E_{AFM}$ in the itinerant picture), does not generate any
phase transition, but rather starts a crossover regime where the correlation
length $\xi_{m}$ for the $\mathbf{m}_{\pm}$ order parameter becomes much
longer that the lattice parameter.
In this regime, one can investigate a possibility of a phase transition
corresponding to the $\sigma$ order parameter. It is important to realize that
$\sigma$ does not have to change sign along a domain wall of the
magnetization. This ensures that $\sigma$ can order even though the sublattice
magnetization vanishes. $\sigma$ does couple to the (long-range) fluctuations
of $\mathbf{m;}$ integrating these fluctuations out one will obtain an
effective Hamiltonian coupling $\sigma_{ij}$ and
$\sigma_{i^{\prime}j^{\prime}}$ as far as $\xi_{m},$ meaning that even very
small coupling between $\mathbf{m}_{+}$ and $\mathbf{m}_{-}$ will produce a
phase transition to a finite $\sigma$ at a temperature $T_{s}\sim
J_{1}\xi_{m}^{2}(T_{s})\sim J_{1}\exp(J_{2}/T_{s})$. Solving this for $T_{s}$,
one gets $T_{S}\sim J_{2}/\log(J_{2}/J_{1})$. Note that here again $J_{1}$ and
$J_{2}\sim T^{\ast}$ just characterize the relevant energy scales and by no
means require the validity of the $J_{1}+J_{2}$ model.
As mentioned above $\sigma$ is positive (negative) for ferromagnetic
(antiferromagnetic) bonds, see Fig.8. Thus $\sigma$ couples bilinearly to the
order parameter of the orthorhombic structural transition
$F_{c}=\gamma\varepsilon_{\mathrm{shear}}\sigma.$ (5)
When the expectation value of $\sigma$ is nonzero below a transition
temperature $T_{s}$, the tetragonal symmetry is spontaneously broken leading
to $\varepsilon_{\mathrm{shear}}\neq 0$. We see that $T_{s}$ is suppressed
from $T^{\ast}$ rather weakly (logarithmically) and that even a weak coupling
between the two sublattices would produce a structural phase transition.
Figure 8: (color online) Magnetoelastic coupling: The two atoms per unit cell
are denoted by filled and open circles. A ferromagnetic bond leads to a
shortening of the nearest neighbor lattice constant (bold dashed lines), while
an antiferromagnetic bond leads to a longer lattice constanti (thin dashed
lines). Depending on the relative orientation of the two sublattices (i.e. the
sign of $\sigma$), two distortions with opposite sign of
$\varepsilon_{\mathrm{shear}}$ are possible.
The third energy scale existing in the problem is set by the interlayer
magnetic coupling, $J_{\perp}.$ In the DFT we found $J_{\perp}\lesssim 1$ meV
in LaFeAsO and $J_{\perp}\sim 16$ meV in BaFe2As2[4]. This huge difference
defines the different behavior of these two compounds. In the former the Neel
transition temperature for a sublattice ordering is on the order of
$T^{\ast}/\log(T^{\ast}/J_{\perp}),$ logarithmically smaller than $T_{s},$
while in the latter one expects a much larger $T_{N}$, and likely larger than
the $T_{s}$ for an individual FeAs plane.
The phase between $T_{N}$ and $T_{s},$ if $T_{s}>T_{N},$ was dubbed “nematic”
in Refs. [50, 51], as the order parameter $\left\langle\sigma\right\rangle\neq
0$ even though $\left\langle\mathbf{M}_{ij}\right\rangle=0$, as expected for
an axial, as opposed to vectorial order parameter. The first order nature of
the transition in the 122 systems is then likely a consequence of the coupling
to soft elastic degrees of freedom, and/or of nonlinear interactions. A more
rigorous treatment of the described physics will be published elsewhere[66].
There is another interesting experimental evidence for the unconventional
nature of the magneto-elastic coupling in these systems. In the 122 systems
the structural distortion $\propto\varepsilon_{\mathrm{shear}}$ and the
sublattice magnetization seem to be proportional to each other.[67] At a
second order transition, symmetry arguments imply however that the former
should be proportional to the square of the sublattice magnetization. At a
first order transition, no such strict connection can be established, however
one expects that the generic behavior is recovered as the strength of the
first order transition gets smaller, realizable via alcaline earth
substitution. Experiments show that the mentioned linear behavior is similar
for Ca, Ba or Sr[68]. In our view this behavior is evidence for the fact that
the first order transition in the 122 systems is never close to being weak.
Arguments that the first order character of the magneto-elastic phase
transition originates from the lattice instabilities near the onset of spin-
density wave order were recently given in Ref.[69]. However, further
discussion clearly goes beyond the limit of this review.
The fact that at the structural transition (and even above), magnetic
correlations in plane are already well established, with large correlation
lengths, explains many otherwise mysterious observations. A more detailed
discussion can be found in Ref. [22].
This picture is not without ramifications for superconductivity. First and
foremost, it implies that at superconducting composition ferropnictides,
especially the 1111 family, are not really paramagnetic, bat rather systems
with a large in-plane magnetic correlation length, much larger than the
lattice parameter and likely much larger than the superconducting correlation
length. Second, the excitation structure in such a system is unusual and
cannot be entirely described in terms of $\chi(\mathbf{Q),}$ where
$\mathbf{Q}=(\pi,\pi),$ since such a description loses the physics associated
with the parameter $\sigma.$ Finally, it implies that the lattice and spin
degrees of freedom do not fluctuate independently and are naturally connected
to each other. Therefore a detailed quantitative theory for the pairing state
will have to include lattice vibrations. Conversely, experiments that find
evidence for a lattice contribution to the pairing mechanism should not be
considered as evidence against magnetic pairing.
### 2.4 Other excitations
While everybody’s attention is attracted to magnetic pairing mechanisms and
spin fluctuations, it would be premature and preposterous to exclude any other
excitations from consideration. First of all, it might be still too early to
discard the venerable phonons. While there is no question that the
calculations performed so far [57, 5] were accurate and the linear response
technique used had proved very reliable before (MgB${}_{2},$ CaC6 $etc.),$
these calculation by definition do not take into account any effects of the
magnetism. As discussed above, it is very likely that the ground state even in
the so-called nonmagnetic region of the phase diagram is characterized by an
AFM correlation length long enough compared to the inverse Fermi vector. In
this case, the amplitude of the magnetic moment of Fe (even though its
direction fluctuates in time) is nonzero and the electronic structure is
sensitive to it. Calculations suggest that a phonon stretching the Fe-As bond
will strongly modulate this magnetic moment and thus affect the electronic
structure at the Fermi level more than for a nonmagnetic compound (or, for
that matter, a magnetic compound with a hard magnetic moment). Softness of the
Fe moments, variationally, provides an additional route for electron-phonon
coupling and should therefore always enhance the overall coupling constant.
Whether this is a weak or a strong effect, and whether the resulting coupling
is stronger in the intraband channel (enhancing the $s_{\pm}$
superconductivity) or in the interband channel (with the opposite effect), is
an open question. Only preliminary results are available[58].
Besides the phonons and the spin fluctuation, charge (polarization)
fluctuations can also, in principle, be pairing agents. To the great surprise
of the current authors, nobody has yet suggested an acoustic plasmon mechanism
for ferropnictides, a mechanism that was unsuccessfully proposed for cuprates,
for MgB2 and for CaC${}_{6}.$ Presumably the apparent lack of strong transport
anisotropy in 122 and the absence of carriers with largely disparate mass
prevented these usual suspects from being discussed.
It is not only the harsh condition on the very existence of acoustic plasmons,
but a very general malady (better known in the case of acoustic plasmons, but
generally existing for any sort of exciton pairing) that prevents plasmonic
superconductivity in most realistic cases: lattice stability. Basically,
efficient pairing of electrons via charge excitations of electronic origin
requires overscreening of electrostatic repulsion — which by itself does not
constitute a problem. But since the ion-ion interaction is screened by the
same polarization operator as electron-electron interaction, there is an
imminent danger that the former is overscreened as well. This is an
oversimplified picture (electron-electron susceptibility differs from the
response to an external field on the level of vertex corrections), but it
captures the essential physics.
This danger was appreciated by the early proponents of the excitonic
superconductivity, W. Little[70] and V. Ginzburg[71], therefore they proposed
space separation between a highly polarizable insulating media, providing
excitons, and a metallic layer or string where the superconducting electrons
live. The sandwich structure of the As-Fe-As trilayer reminds us of the
Ginzburg’s “sandwich” (“Ginzburger” ) and tempts to revisit his old proposal.
This was done recently by Sawatzky and collaborators[72] who pointed out that
As is a large ion (Pauling radius for As4- is 2.2 Å) and ionic polarizability
grows with the radius cube. Since the conducting electrons are predominantly
of Fe origin, they suggested pairing of Fe d electrons $via$ polarization of
As ions. So far, this proposal was received with a skepticism that can be
summarized as follows. (1) Analyzing the muffin-tin projected character of the
valence bands, as it was done in Ref. [72] is generally considered to be an
unreliable way to estimate the hybridization between different ions; indeed
the largest part of the electronic wave function refers to the interstitial
space, which is naturally identified as mostly As-like. (2) Removal of the As
orbitals from the basis leads to a strong reduction of the valence band width,
indicating that hybridization between Fe and As is about as strong as direct
Fe-Fe hopping. (3) When Bloch functions are projected upon the Fe-only Wannier
functions, the latter come out very diffuse and extend way beyond the Fe ionic
radius. That is to say, negligible hybridization between Fe and As, that is
prerequisite for the scenario promoted in Ref. [72], appears to be a rather
questionable proposition. Besides, above-mentioned calculations of the phonon
spectra and electron-phonon coupling implicitly account for the large
susceptibility of the As-4 ions (which comes mostly from the outer, valence
shell) yet they find no manifestation of strong As polarization: neither
particular phonon softening nor strong coupling with any phonon.
## 3 Pairing symmetry: general considerations
### 3.1 Geometrical consideration: excitation vectors and Fermi surface
Given such disparate views that different researchers hold about the origin of
magnetism in ferropnictides and of the character of spin fluctuations there,
it may seem strange that a great majority of model calculations predict the
same pairing symmetry, $s_{\pm},$ with full gaps in both electron and hole
bands, but with the opposite signs of the order parameters between the two. In
fact, this is not surprising at all. To begin with, let us point out that the
sign of the interaction mediated by boson exchange is always positive
(attraction) for charge excitations (phonons, plasmons, polarization
excitons), since the components of a Cooper pair have the same charge, but can
be either positive (for triplet pairing, where the electrons in the pair have
the same spin) or negative (repulsion) for singlet pairing, for spin
excitations. That is to say, exchange of spin fluctuations mediates repulsion.
A quick glance at the anisotropic BCS equation reveals that repulsive
interactions can be pairing when, and only when the wave vector of such a
fluctuation spans parts of the Fermi surface(s) with opposite signs of the
order parameter (equivalently, one can say that an interaction that is
repulsive everywhere in the momentul space, can be partially attractive in the
real space, for instance, for electrons located an nearest lattice sites).
This can be illustrated on a popular model of high-$T_{c}$ cuprates, which
considers a simplified cylindrical Fermi surface nearly touching the edge of
the Brillouin zone and superexchange-driven spin fluctuations with the wave
vector $(\pi,0)$. As Fig. 9a illustrates, such an interaction is pairing in
the $d_{x^{2}-y^{2}}$ symmetry, because it spans nearly perfectly the lobes of
the order parameter with the opposite signs.
Figure 9: (color online) (a) A cartoon illustrating how a repulsive
interaction corresponding to superexchange spin fluctuations $Q=(\pi,\pi$) may
generate $d$-wave pairing in cuprates. (b) The same, for an $s_{\pm}$ state
and spin fluctuations with $Q=(\pi,0)$ (in a Brillouin zone corresponding to
one Fe per cell). (c) If the central hole pocket is absent, the superexchange
interaction favors a nodeiless $d$ state.
Most models used for ferropnictides assume a simplified fermiology with one or
more hole FSs and one or more electron FSs displaced by the SDW vector
($\pi,0$) (in this Section, we use the notations corresponding to the
Brillouin zone with one Fe per cell). Any spin-fluctuation induced interaction
with this wave vector, no matter what the origin of these fluctuations (FS
nesting, frustrated superexchange, or anything else) unavoidably leads to a
superconducting state with the opposite signs of the order parameter for the
electrons and for the holes. Depending on the details of the model the ground
state maybe isotropic or anisotropic and the gap magnitudes on the different
sheets may be the same or may be different, but the general extended $s$
symmetry with the sign-reversal of the order parameter (an $s_{\pm}$ state) is
predetermined by the fermiology and the spin fluctuation wave vector (Fig.
9b).
It is worth noting that while most (but not all) models consider spin
fluctuations corresponding to the observed instability to be the leading
pairing agent, some include spin fluctuations of different nature [for
instance, nearest neighbor superexchange or nesting between the “X” and “Y”
electron pockets, both corresponding to the same wave vector, ($\pi,\pi)$ in
the unfolded zone and $(0,0)$ in the conventional zone], or phonons, or direct
Coulomb repulsion; these additional interactions may modify the gap ratios and
anisotropies (in extreme cases, creating nodes on some surfaces), but, for a
realistic choice of parameters, unlikely to change the symmetry.
Moreover, if the radius of the largest FS pocket is larger than the magnetic
vector, spin fluctuations start to generate an intraband pair-breaking
interaction, which by itself will lead to an angular anisotropy and possible
gap nodes.
The above reasoning, however, is heavily relying upon an assumption that the
topology predicted by the DFT is correct. So far, as discussed above, the
evidence from ARPES and from quantum oscillations has been favorable. It is
still of interest to imagine, for instance, electron-doped compounds not
having hole pockets at all or having them so small that the pairing energy for
them is negligible. It was pointed out[40, 73] that in this case spin
fluctuations with different momentum vectors dominate and create a nodeless
$d$-wave state in the electron pockets, as Fig. 9c illustrates.
### 3.2 General properties of the $s_{\pm}$ state
Since the $s_{\pm}$ states constitute the most popular candidate for the
superconducting symmetry of pnictides, it is worth recapitulating the physics
of this state. Let us start with the simplest possible case: two bands (two
Fermi surfaces) and interband repulsive interaction between the two. Let the
interaction strength be $-V,$ and the DOSs $N_{1}\neq N_{2}.$ To be specific,
let $N_{2}=\alpha N_{1},$ $\alpha\geq 1.$ Then in the weak coupling limit the
BCS equations read
$\displaystyle\Delta_{1}$ $\displaystyle=-\int
d\epsilon\frac{N_{2}V\Delta_{2}\tanh(E_{2}/2k_{B}T)}{2E_{2}}$
$\displaystyle\Delta_{2}$ $\displaystyle=-\int
d\epsilon\frac{N_{1}V\Delta_{1}\tanh(E_{1}/2k_{B}T)}{2E_{1}}$ (6)
where $E_{i}$ is the usual quasiparticle energy in band $i$ given by
$\sqrt{(\epsilon-\mu)^{2}+\Delta_{i}^{2}}.$ Near $T_{c}$ linearization gives
$\displaystyle\Delta_{1}$
$\displaystyle=\Delta_{2}\lambda_{12}\log(1.136\omega_{c}/T_{c})$
$\displaystyle\Delta_{2}$
$\displaystyle=\Delta_{1}\lambda_{21}\log(1.136\omega_{c}/T_{c}),$ (7)
where $\lambda_{12}=N_{2}V$, the dimensionless coupling constant, with a
similar expression for $\lambda_{21}.$ These equations readily yield
$\lambda_{eff}=\sqrt{\lambda_{12}\lambda_{21}}$ and
$-\Delta_{1}/\Delta_{2}=\sqrt{N_{2}/N_{1}}\equiv\sqrt{\alpha}.$ Note that the
Fermi surface with the larger DOS has a smaller gap. It can also be shown that
the gap ratio at zero temperature in the weak coupling limit is also given by
$\sqrt{N_{2}/N_{1}},$ and strong coupling effects tend to reduce the disparity
between the gaps.
The situation becomes more interesting for more than two orbitals with
distinct gaps. Let us consider a model for the hole-doped 122 compound. The
calculated FS (Fig.4) shows three sets of sheets: Two e-pockets at the corner
of the zone, two outer h-pockets, formed by the $xz$ and $yz$ orbitals
(degenerate at $\Gamma$ without the spin-orbit), and the inner pocket formed
by $x^{2}-y^{2}.$ In the DFT calculations all three hole cylinders are
accidentally close to each other, however, ARPES shows two distinct sets, the
inner barrel, one of which presumably corresponding to $x^{2}-y^{2}$ band, and
the outer one, presumably $xz/yz.$ The pairing interaction between the
e-pockets and the two different types of the h-pockets need not be the same
(by virtue of the the matrix elements). Using the same partial DOS as listed
above for Ba1.6K0.6Fe2As2 (both total and individual DOS depend weakly on the
position of the Fermi level, reflecting the 2D character of the band structure
at this doping), roughly 1.2 st/eV for each hole band and the same for the two
e-band together, we get the coupling matrix
$\left(\begin{array}[]{ccc}0&0&-\lambda_{1}\nu_{1}\\\
0&0&-\lambda_{2}\nu_{2}\\\ -\lambda_{1}&-\lambda_{2}&0\end{array}\right),$ (8)
where $\nu_{1,2}$ is the ratio of DOS of the first ($xz/yz)$ and the second
($x^{2}-y^{2})$ hole bands to that of the electron bands. Note that
$\nu_{1}\sim 2$ and $\nu_{2}\sim 1.$ Diagonalizing this matrix we find the gap
ratios to be
$\Delta_{1}:\Delta_{2}:\Delta_{e}=\lambda_{1}:\lambda_{2}:\sqrt{\lambda_{1}^{2}\nu_{1}+\lambda_{2}^{2}\nu_{2}.}$
The latest ARPES measurements[11] imply that $\Delta_{i}:\Delta_{o}\approx
2:1,$ where $i$ and $o$ stand for the inner and outer sets of hole Fermi
surfaces. This would mean that the two coupling constants are twice larger
that the other (although we do not know which), which is fairly possible.
However, that implies that the electron FS has a gap that is larger than that
of the largest hole band by at least a factor of $\sqrt{1.5}=1.22$ (assuming
that the outer FSs in the calculations, are formed by the $xz/yz$ bands; the
opposite assumptions leads to an even larger electron-band gap). This is in
some disagreement with the ARPES data that suggest that $\Delta_{e}$ is on the
order of $\Delta_{i}$ or slightly smaller. However, this is a small
discrepancy, which can be easily corrected by introducing small intraband
electron-phonon coupling for the hole bands, and/or taking into account
possible gap suppression by impurities in the electron band. It is also worth
noting that the spread of the measured values, depending on the sample and on
the location on the FS, is on the order of 10%.
### 3.3 Coulomb avoidance
It was realized quite some time ago that a $d$-wave pairing has an additional
advantage compared to an $s$-wave, namely that the electrons in a Cooper pair
avoid each other (the pair wave function has zero amplitude at
$\mathbf{r-r}^{\prime}=0$), strongly reducing their local Coulomb repulsion.
The leading contribution to the pairing interaction in the single band Hubbard
model $U\sum_{\mathbf{k}}\left\langle
c_{\mathbf{k\uparrow}}c_{-\mathbf{k\downarrow}}\right\rangle$ is repulsive,
but vanishes as $\sum_{\mathbf{k}}\Delta_{\mathbf{k}}=0$ due to the symmetry
of the $d$-wave state. Thus, a contact Coulomb repulsion does not affect
$d$-wave superconductivity.
The simplest possible $s^{\pm}$-wave function is given by Eq.3. In this case,
the sum over the Brillouin zone vanishes again due to nodes at $\pm ak_{x}\pm
ak_{y}=\pi/2$. This description is however somewhat misleading because it may
produce a false impression that there is a symmetry reason for the vanishing
of the Coulomb repulsion in the $s^{\pm}$state, or that this particular
functional form is essential for avoiding the Coulomb repulsion. To illustrate
that this is not the case, it is instructive to consider a toy problem in
reciprocal space. In the weak coupling regime, the effective coupling matrix
$\Lambda_{\mathbf{kk}^{\prime}}$ (note that the band index is uniquely defined
by the wave vector) is
$\Lambda_{\mathbf{kk}^{\prime}}=\lambda_{\mathbf{kk}^{\prime}}-\mu_{\mathbf{kk}^{\prime}}^{\ast},$
(9)
where $\lambda$ is the original coupling matrix in orbital space and
$\mu_{\mathbf{kk}^{\prime}}^{\ast}$ is the renormalized Coulomb
pseudopotential. The critical temperature is determined by the largest
eigenvalue of the matrix $\Lambda,$ and the $\mathbf{k}$ dependence of the
order parameter $\Delta_{\mathbf{k}}$ is given by the corresponding
eigenvector. If $\mu^{\ast}$ is a constant and
$\sum_{\mathbf{k}}\Delta_{\mathbf{k}}=0$ (as in the $d$-wave case), any
eigenvector of the matrix $\lambda$ is also an eigenvector of $\Lambda,$ with
the same eigenvalue. This proves that Coulomb avoidance takes place for any
superconductor where the order parameter averages to zero over the entire FS,
and not only for the $d$-wave symmetry.
Let us now consider a specific $s^{\pm}$ superconductor. For simplicity, let
us take two bands with the same DOS, $N_{1}=N_{2}=N$ and with an interband
coupling only:
$\lambda_{ij}=\left(\begin{array}[]{cc}0&-VN\\\ -VN&0\end{array}\right).$ (10)
We shall also assume that the Coulomb repulsion $U$ is a contact interaction,
so that $\mu_{ij}^{\ast}=UN$ is the same for all matrix elements. The maximal
eigenvalue of $\Lambda$, which corresponds to the effective coupling constant
$\lambda_{\mathrm{eff}}$, is indeed simply $VN$ and _independent_ of $U$. The
corresponding eigenvector is $\Delta_{1}=-\Delta_{2}$, i.e. the $s^{\pm}$
state. The Coulomb interaction is irrelevant, just like in case of $d$-wave
pairing. The effect is however a consequence of the assumed symmetry of the
two bands. In general, unlike the d-wave, no symmetry requires that
$\sum_{\mathbf{k}}\Delta_{\mathbf{k}}=0$. This can already be seen if one
considers a model with distinct densities of states: $N_{2}=\alpha
N_{1}=\alpha N$. We have
$\lambda_{ij}=\left(\begin{array}[]{cc}0&-\alpha VN\\\
-VN&0\end{array}\right).$ (11)
and the weak-coupling gap ratio near $T_{c}$ is $\sqrt{\alpha}$. Now the
effect of the Coulomb repulsion is not nullified, but is still strongly
suppressed. The eigenvalues are easily determined. The key result is that the
maximal eigenvalue remains positive for all finite $\alpha$. Even the extreme
limit $\lambda_{\mathrm{eff}}^{\pm}(U\rightarrow\infty)=2VN\alpha/(1+\alpha)$
is for realistic $\alpha$ only somewhat reduced compared to
$\lambda_{\mathrm{eff}}^{\pm}(U=0)=\sqrt{\alpha}VN$. This is qualitatively
different from the regular ($s_{++})$ interband-only pairing with an
attractive interband interaction of the same strength. In this case,
$\lambda_{\mathrm{eff}}^{++}(U>V/2)<0$, and the Coulomb interaction dominates
over the attractive interband pairing interaction. In the linear in $UN$
regime, the suppression rate of $\lambda_{eff}(U)$ is $(\sqrt{\alpha}-1)/2$
for $s^{\pm}$ and $(\sqrt{\alpha}+1)/2$ for $s^{++}$ pairing. For example, for
the DOSs ratio of $4$ (the gap ratio is then $2$) $\mu^{\ast}\approx
0.25\lambda_{eff}\left(U=0\right)$ will suppress an $s^{++}$ superconductivity
entirely, while in the $s^{\pm}$ case the effective coupling will be reduced
only by 8%.
The efficiency of the Coulomb avoidance is neither limited to the assumption
of a uniform Coulomb interaction among and within the bands, nor is a result
of the weak coupling approach. Strong coupling FLEX type calculations also
find pairing states with very small repulsive contribution due to Coulomb
interaction[44, 45].
## 4 Pairing symmetry: experimental manifestations
### 4.1 Parity
Since we want to review the experimental situation regarding the pairing
symmetry, the first question to ask is, whether superconductivity is singlet
or triplet? Fortunately, this question can be answered relatively confidently.
Measurements of the Knight shift on single crystals of the Co-doped BaFe2As2
superconductor[74] clearly indicate full suppression of spin susceptibility in
the superconducting state in all directions, incompatible with a triplet
pairing in a tetragonal crystal. For other compounds only polycrystalline,
direction-averaged data exist, but they fully agree with the above result,
virtually excluding triplet superconductivity. This leaves, of all possible
scenarios, essentially three: conventional $s$ (presumably multigap),
$s_{\pm}$ and $d$.
### 4.2 Gap amplitude
All experiments that distinguish between different pairing states can be,
roughly speaking, grouped into two classes: those probing the gap amplitude
and those probing the gap symmetry. The advantage of the former is that they
are comparatively easier to perform. The temperature dependence of any
observable sensitive to the excitation gap is sensitive to the presence of
nodes or multiple gaps. The disadvantage is that only a measurement of the
relative phase of the wave function will unambiguously determine the pairing
state, including its symmetry.
Important and very transparent probes of the gap amplitude are thermodynamic
measurements. The early reports of the specific heat leaned towards power-law
behavior characteristic of nodal superconductivity. The latest data [13, 15]
suggest a fully gapped superconductivity, or a dominant fully gapped component
with possible small admixture of a nodal state. While the experimental
situation is still far from consensus, especially regarding the 1111 family, a
few observations may be in place: (i) The specific heat jump in the h-doped
BaFe2As2 is strong and sharp, and in 1111 compounds is weak and poorly
expressed. This cannot be ascribed to a difference in calculated band
structures. This is either due to sample quality issues or possibly to the
more isotropic character of superconducting and magnetic properties in 122
systems. (ii) In no case can specific heat temperature dependence be fitted
with one gap. Multiple gap fits, having more parameters, are of course less
reliable. (iii) Another, usually more reliable signature of nodal
superconductivity is a square-root dependence of the specific heat coefficient
on the magnetic field. Existing reports[13] however show a clear linear
dependence, characteristic of a fully gapped superconductor.
Another popular probe is temperature dependence of the NMR relaxation rate.
Extensive studies have been done in this aspect (see other articles in this
volume). In all studied systems, the relaxation rate is non-exponential. The
initial impression was that the relaxation rate is cubic in temperature,
$1/T_{1}\propto T^{3},$ consistent with nodal lines[75, 76]. Later it was
argued that the data cannot be described by a single power law as in the
cuprates[77, 78]. These results were obtained for the 1111 systems. The
situation with the 122 family is even less clear. Published data[79, 74] do
not show exponential decay either, but the results are equally far from any
single power law behavior. Even more puzzling, the only paper reporting on the
low-$T_{c}$ LaFePO superconductor claims that the relaxation rate does not
decrease below $T_{c}$ at all[80].
The third relevant experiment is measuring the London penetration depth.
Reports are again contradictory. For instance, in Pr-based 1111 compound the
penetration depth was found[81] to barely change between $\approx 0.05T_{c}$
and $T^{\ast}\approx 0.35T_{c},$ and than increase roughly as
$(T-T^{\ast})^{2}$ between $T^{\ast}$ and $\approx 0.65T_{c},$ a picture
roughly consistent with a multi-gap nodeless superconductor. Malone $et$
$al$[82] measured Sm-based 1111 and were able to fit their data very well in
the entire interval from $T_{c}/30$ and $T_{c}$ using two full gaps. In Nd-
based 1111 the penetration depth was measured at $T>0.1T_{c}$ and fitted with
a single anisotropic gap for $0.1T_{c}<T<T_{c}/3$,[83] however, the latest
result from the same authors, taken at lower temperature, can be better fitted
with a quadratic law[84]. Similar quadratic behavior has been clearly seen in
the 122 compounds[85]. At the same time, the low-$T_{c}$ LaFePO is again odd:
it shows a linear behavior[86].
To summarize, the thermodynamic data on average lean towards a nodal
superconductivity. However, some data are not consistent with the gap nodes,
and there is no clear correlation with the sample quality either way.
Moreover, while some data suggest line nodes, others are consistent only with
point nodes, in the clean limit. One can say with a reasonable degree of
confidence that the entire corpus of the data cannot be described by any one
scenario in the clean limit. On the other hand, essentially any temperature
dependence of thermodynamic characteristics can be fitted if a particular
distribution of impurity scattering is assumed in an intermediate regime
between the Born and the unitary scattering, and a particular relation between
the intra- and interband scattering (there have been a number of paper doing
exactly that for the NMR relaxation rate, for instance, Ref. [87], or for the
penetration depth, for instance, Ref. [88]). However, the fact that all these
papers rely upon specific combinations of parameters, while the phenomena they
seek to describe are rather universal, calls for caution. Besides, except in
the pure unitary regime, scattering is accompanied by a $T_{c}$ suppression
and most papers do not find any correlation between thermodynamic probes and
$T_{c}$ among different samples. Another possibility is that required
scattering is provided not by impurities, but by intrinsic defects that are
thermodynamically or kinetically necessarily present in all samples (for
example, dynamic domain walls introduced in Ref. [22]). More measurements at
the lower temperature and on clean samples will probably clarify the matter.
At the moment one cannot consider this problem solved.
Close to the thermodynamic measurements are tunneling type experiments. As of
now, these have been nearly exclusively point-contact Andreev reflection
probes. Here, again, the experimental reports are quite inconsistent,
moreover, the situation is in some sense worse than in thermodynamic probes,
since uncontrollable surface properties enter the picture. Interpretation
generally includes fitting one curve with a large number of parameters, and
the procedure is not always well defined. Generally speaking, three types of
results have been reported: $d$-wave like, single full gap-like, and multigap.
Interpretation is particularly difficult because within the $s^{\pm}$ picture
formation of subgap Andreev bound states was predicted (e.g., Refs. [89, 90])
that can be easily mistaken for multiple gaps.
### 4.3 Phase-sensitive probes
In view of all that, experiments directly probing the gap symmetry are highly
desirable. The paramagnetic Meissner effect, also known as Wohlleben effect,
occurs in a polycrystalline sample when inter-grain weak links have random
order parameter phase shifts, $0$ or $\pi.$ It has been routinely observed in
cuprates and is considered a key signature of $d$-wave superconductivity. The
effect does not exist in conventional, even anisotropic and multi-gap
superconductors, even though sometimes it can be emulated by impurity effects
in the junctions. For $d-$wave superconductors without pronounced
crystallographic texture the Wohlleben effect is expected, and its absence can
be taken as evidence against $d$-wave. Finally, in the $s^{\pm}$ scenario the
phase is the same by symmetry for $(100)$ and $(010)$ grain boundaries, and
there are good reasons to expect the same phase for $(110)$ boundaries as
well. There may or may not be a $\pi$ phase shift for phase boundaries at some
specific orientation, likely for a narrow range of angles[91], but probably
not enough to produce a measurable Wohlleben effect. The absence of the effect
in experiment[92] is a significant argument against $d$-wave, but hardly helps
to distinguish $s$ from $s^{\pm}.$
Similarly, the $c$-axis tunneling provides evidence against the $d$-wave,
where the Josephson current strictly parallel to the crystallographic $c$
direction vanishes by symmetry. Experimentally a sizable current was
found[93].
Recalling the cuprates again, the ultimate argument in favor of the $d$-wave
was provided by the corner Josephson junction experiments that probe directly
the phase shift between two separate junctions; in cuprates, with their
$d_{x^{2}-y^{2}}$ symmetry, these junction were to be along the $(100)$ and
$(010)$ directions. Similarly, a potential $d_{xy}$ state could be detected by
the combination of $(110)$ and $(\bar{1}10)$ directions. On the other hand, a
conventional $s$ state would not produce a phase shift for any combination of
contacts. Again, the case of $s_{\pm}$ superconductivity is nontrivial. While
symmetry does not mandate a $\pi$ shift for any direction, it can be shown
that, depending on the electronic structure parameters and properties of the
interface, there may exist intermediate angles (between $0$ and $45^{o})$
where a $\pi$ shift is possible[91]. It also may be possible if the two
junctions have different tunneling properties, so that one of them filters
through only hole-pocket electrons, and the other only electron-pockets. It is
not as bizarre as it may seem, and some possibilities were discussed in Ref.
[91]. Probably the most promising design involves “sandwiches” of various
geometries. The first proposal of that kind was by Tsoi et al[90], who
suggested an $s/s^{\pm}/s^{\prime}$ trilayer, where $s$ is a conventional
quai-2D superconductor with a large Fermi surface that has no overlap with the
hole FS of the $s^{\pm}$ layer (equivalently, a superconductor with small
Fermi surfaces centered around the M points), and $s^{\prime}$ is a
conventional superconductor with a small FS centered around $\Gamma.$ This was
followed by another proposal of a bilayer of hole-doped and electron-doped 122
materials[91]. In both cases the idea is that the current through the top of
the sandwich will be dominated by the electron FS, and through the bottom by
the hole one. Both proposals require momentum conservation in the interfacial
plane, that is, basically, epitaxial or very high quality interface. The
former proposal has an additional disadvantage of requiring two high-quality
interfaces with very special conventional superconductors, particularly the
one that should filter through the electron FS is rather difficult to find. As
of now, no experiments have been reported pursuing any of the above
suggestions, but with better single crystals and thin films it should become
increasingly doable. It should be stressed, however, that in this case, unlike
the cuprates, an absence of the $\pi$ shifts in any of the proposed geometries
does not disprove the $s^{\pm}$ scenario, since the effect here is
quantitative rather than qualitative, but the presence of the sought effect
would be a very strong argument in favor of it. On the other hand, standard
90o corner junction experiments similar to cuprates are also important, as
they could prove unambiguously that the symmetry is not $d$-wave (even though
they cannot distinguish between $s$ and $s^{\pm}).$
Further properties of interfaces between an $s^{\pm}$ superconductor and
normal metal or conventional superconductor are now actively being studied
theoretically, encouraging further experimental research. Probably we will see
first results within the next year.
### 4.4 Coherence factor effects
Other signatures of the $s^{\pm}$ state are based on the fact, previously
pointed out by many in connection with the cuprates, that the coherence
factors are “reversed” for electronic transitions involving order parameters
of the opposite sign. In the conventional BCS theory, as is well known,
coherence factors of two kinds appear. The first kind, sometimes called “Type
I” or “minus” coherence factor, is given by the expression
$(1-\Delta_{\mathbf{k}}\Delta_{\mathbf{k}^{\prime}}/E_{\mathbf{k}}E_{\mathbf{k}^{\prime}}),$
where
$E_{\mathbf{k}}=\sqrt{\Delta_{\mathbf{k}}^{2}+\varepsilon_{\mathbf{k}}^{2}},$
and $\varepsilon_{\mathbf{k}}$ in the normal state excitation. The other kind,
Type II or the “plus” coherence factor has the opposite sign in front of the
fraction. If both order parameters entering this formula have the same sign,
the Type I factor is destructive, in the sense that it goes to zero when
$\varepsilon\rightarrow 0,$ and cancels out the peak in the superconducting
DOS. Type I factors appear, for instance, in the polarization operator, and as
a result there are no coherence peaks in phonon renormalization (as measured
by ultrasound attenuation, for instance) and in spin susceptibility (including
the Knight shift). Type II factors appear, for instance, in the NMR relaxation
rate, and they are constructive, resulting in the famous Hebel-Slichter peak
below $T_{c}.$
Obviously, if $\Delta_{\mathbf{k}}$ and $\Delta_{\mathbf{k}^{\prime}}$ have
opposite signs, the meaning of the coherence factors is reversed; the Type I
factors are now constructive and the Type II destructive. There are several
straightforward ramifications of that. For instance, as it was pointed out
already in the first paper proposing the $s^{\pm}$ scenario[5], the spin
susceptibility at the SDW wave vector should show resonance enhancement just
below $T_{c}$. For explicit calculations of this effect see for example
Refs.[94, 95]. There are indeed some reports of this effect, as measured by
neutron scattering[96]. In principle, one can expect a similar effect in the
phonon line-width, for the phonons with the same wave vector, just below
$T_{c},$ but this is really hard to observe.
Less straightforward are cases of the quantities that involve averaging over
the entire Brillouin zone, in which case the answer, essentially, depends on
which processes play a more dominant role in the measured quantity, those
involving intra-, or interband scattering. The answer usually depends on
additional assumptions about the matrix elements involved, which can rarely be
calculated easily from first principles. An example is electronic Raman
scattering; a possibility of a resonant enhancement in some symmetries has
been discussed recently[97].
## 5 Role of impurities
Impurity and defect scattering is believed to play an important role in
pnictide superconductors. Proximity to a magnetic instability implies that
ordinary defects may induce static magnetic moments on the neighboring Fe
sites and thus trigger magnetic scattering. If, as is nearly universally
believed, an order parameter with both signs is present, nonmagnetic
impurities are also pair-breaking. Thus the anticipation is that in regular
samples, and maybe in samples of much higher quality, impurity-induced pair
breaking will play a role.
Our intuition regarding the impurity effects in superconductors is largely
based upon the Abrikosov-Gorkov theory of Born-scattering impurities in BCS
superconductors. There was an observation at that time that folklore ascribes
to Mark Azbel: Soviet theorists do what can be done as good as it should be
done, and American ones do what shall be done as good as it could be done. For
many years the approach to the impurity effects in superconductors was largely
Soviet: most researchers refine the Abrikosov-Gorkov theory, applying it to
anisotropic gaps and to unconventional superconductors, and relatively little
has been done beyond the Born limit — despite multiple indications that most
interesting superconductors, from cuprates to MgB2 to pnictides are in the
unitary limit or in an intermediate regime.
The physics of the nonmagnetic scattering in the two different limits is quite
different. In the Born limit, averaging over all scattering events yields a
spatially uniform superconducting state and tries to reduce the variation of
the order parameter over the FS. Ultimately, for sufficiently strong
scattering, the order parameter becomes a constant, corresponding to the DOS-
weighted average over the FS. Note that unless this average is zero by
symmetry (like in d-wave) the suppression of $T_{c},$ while linear at small
concentrations, is never complete. As pointed out by Mishra $et$ $al.$ [98],
this effect should manifest itself most clearly in an extended s-wave pairing
with accidental nodes in the order parameter. Indeed, while in $d$-wave
superconductors impurities broadens nodes into finite gapless spots, in an
extended $s$ case it is likely that the order parameter of one particular sign
dominates a given FS pocket, in which case Born impurities will first make the
parts of the FS with the “wrong” order parameter gapless, and then lead to a
fully gapped superconductivity. Of course, this only holds for nonmagnetic
impurities. Isotropic magnetic impurities will be just pair-breaking as they
are in conventional superconductors, with the only interesting new physics
being that magnetic impurities cease being pair-breakers if they scatter a
pair such that the sign of the order parameter is flipping. The rule of thumb
is that a scattering path for which magnetic scattering is pair-breaking (no
change of sign of the order parameter), nonmagnetic scattering will not be
pair-breaking, and $vice$ $versa.$
The physics of the unitary limit is quite different. In that limit, the
concentration of impurities is relatively low, but the scattering potential of
an individual impurity is strong, $N(0)v_{imp}\gg 1.$ In that case rather than
suppressing superconductivity uniformly each impurity creates a bound state at
the chemical potential, thus creating a zero energy peak in the density of
states, without substantial suppression of the bulk superconductivity.
Increasing the impurity concentration broadens the peak, while increasing its
strength barely has any effect at all [99]. In an intermediate case between
the Born limit and the unitary limit, the bound state is formed inside the gap
at a finite energy and is the broader the closer it is to the gap (that is,
closer to the Born limit).
The principal difference from the point of view of the experiment is that the
unitary or intermediate scattering can create subgap density of states at
arbitrary low energy at any temperature, without a drastic suppression of
$T_{c}.$ It was shown in Ref. [87] that any standard code for solving the
Eliashberg equations in the Born limit can be easily modified, with minor
changes, to treat the unitary limit, as well as any intermediate regime.
Therefore we anticipate an imminent shift in the community from the “Soviet”
approach to the “Western” approach, with more quantitative understanding of
the effect beyond the Born approximation.
## 6 Conclusions
In this article we presented a brief overview of some proposals that have been
made for the pairing state in the Fe-pnictide superconductors. In particular,
we summarized arguments that support the view that the vicinity of
superconductivity and magnetism in these systems is not accidental. The
obvious appeal of this, and essentially any other electronic pairing mechanism
is, of course, that the involved energy scales, and thus $T_{c}$, can in
principle be larger if compared to pairing due to electron-phonon interaction.
Electronic mechanisms also promise a new level of versatility in the design of
new superconductors.
At this early stage in the research on the iron pnictide family, experiments
have not conclusively determined the pairing symmetry, the detailed pairing
state or the microscopic pairing mechanism. Still, in our view a plausible
picture emerges where superconductivity is caused by magnetic fluctuations.
Only two ingredients are vital to arrive at a rather robust conclusion for the
pairing state. First, pnictides need to have Fermi surface sheets of two
kinds, one near the center of the Brillouin zone, and the other near the
corner. Second, the typical momentum for the magnetic fluctuations should be
close to the ordering vectors $\mathbf{Q=}\left(\pi,\pi\right)$ of the parent
compounds. Then, magnetic interactions lead quite naturally to an efficient
inter-band coupling that yields an $s^{\pm}$ pairing state. This result is
general in the sense that it is obtained regardless of whether one develops a
theory based on localized quantum magnetism or itinerant paramagnons. There is
evidence that the two needed ingredients are present in the pnictides. Fermi
surface sheets at the appropriate locations have been predicted in non-
magnetic LDA calculations and seen in ARPES experiments. The magnetic ordering
vector has been determined via neutron scattering, even though we have to
stress that a clear identification of magnetic fluctuations for
superconducting systems without long range magnetic order is still lacking.
The resulting $s^{\pm}$ pairing state has a number of interesting properties.
As far as the a group theoretic classification is concerned, its symmetry is
the same as that for a conventional $s$-wave pairing, where the gap-function
has same sign on all sheets of the Fermi surface. However, there are
significant differences between the two states. The sign change in the gap
affects the coherence factors, leading to the resonance peak in the dynamic
spin susceptibility and the absence of a Hebel-Slichter peak in NMR.
Nonmagnetic impurities affect the $s^{\pm}$-state just like magnetic
impurities do in an ordinary $s$-wave state, i.e. here a behavior more akin to
$d$-wave superconductors. Another implication of the sign change in the
$s^{\pm}$-state leads to rather efficient Coulomb avoidance.
The presence of nodes in the superconducting gap in still an open issue. In
$d$-wave or $p$-wave pairing states, nodal lines or points are fixed by
symmetry. This is different for the $s^{\pm}$-state. In its most elementary
version, the sign change of the gap corresponds to a node located between two
Fermi surface sheets. This is the case for the $\Delta\left(\mathbf{k}\right)$
given in Eq.3. Energetic arguments favor such a gapless state as long as the
momentum transfer $\mathbf{Q}$ couples efficiently to large parts of distinct
Fermi surface sheets and Coulomb avoidance is efficient. However, as there is
no symmetry constraint for the location of the nodes, it is in principle
possible that there are nodes on some Fermi surface sheets.
Next to the nature of the pairing state, the microscopic understanding of the
magnetism of the Fe-pnictides is one of the most interesting aspects of these
materials. Are these systems made up of localized spins that interact via
short ranged, nearest neighbor exchange interactions or, are they better
described in terms of itinerant magnetism? While we emphasized that many
aspects of the pairing state emerge regardless of which of these points of
view is correct, this is really only true for the most elementary aspects of
the theory. As our understanding of these materials deepens, dynamical aspects
of the pairing state will become more and more important, and the details of
the magnetic degrees of freedom will matter. In our view, the most sensible
description starts from itinerant electrons, however with significant
electron-electron interaction. In detail, we find numerous arguments that
emphasize the role of magneto-elastic couplings and that favor a sizable Hund
coupling, i.e. the multi orbital character and the corresponding local multi-
orbital interactions are important to understand the magnetism and
superconductivity alike. Regardless of whether this specific point of view is
correct or not, it is already evident that the ferropnictides make up a whole
new class of materials that stubbornly refuse to behave according to one of
the simple minded categories of condensed matter theory.
## 7 Acknowledgements
This research was supported by the Ames Laboratory, operated for the U.S.
Department of Energy by Iowa State University under Contract No. DE-
AC02-07CH11358 (J.S.), and by the Office of Naval Research (I.I.M.). The
authors wish to thank all their friends and collaborators, without whom this
works could not be accomplished, and their numerous colleagues who read the
manuscript and sent us many useful and insightful comments.
## References
* [1] X. Marchan and W. Jeitschko, J. Solid State Chem 24 351 (1978).
* [2] Y. Kamihara, T.Watanabe, M. Hirano, and H. Hosono. J. Am. Chem. Soc. 130, 3296 (2008).
* [3] M. Rotter, M. Tegel, and D. Johrendt, Phys. Rev. Lett. 101, 107006 (2008)
* [4] All original calculations used in this article were performed using the standard full-potential LAPW package WIEN2k. The details can be found in Refs. [5, 100]. Unless stated otherwise, the experimental crystal structure is used.
* [5] I.I. Mazin, D.J. Singh, M.D. Johannes and M.H. Du, Phys. Rev. Lett. 101, 057003 (2008).
* [6] C. Liu, T. Kondo, M. E. Tillman, R. Gordon, G. D. Samolyuk, Y. Lee, C. Martin, J. L. McChesney, S. Bud’ko, M. A. Tanatar, E. Rotenberg, P. C. Canfield, R. Prozorov, B. N. Harmon, and A. Kaminski, arXiv:0806.2147.
* [7] T. Kondo, A. F. Santander-Syro, O. Copie, Chang Liu, M. E. Tillman, E. D. Mun, J. Schmalian, S. L. Bud’ko, M. A. Tanatar, P. C. Canfield, and A. Kaminski, Phys. Rev. Lett. 101, 147003 (2008).
* [8] H. Ding, P. Richard, K. Nakayama, K. Sugawara, T. Arakane, Y. Sekiba, A. Takayama, S. Souma, T. Sato, T. Takahashi, Z. Wang, X. Dai, Z. Fang, G. F. Chen, J. L. Luo, and N. L. Wang, Europhys. Lett. 83 47001 (2008).
* [9] L. Zhao, H. Liu, W. Zhang, J. Meng, X. Jia, G. Liu, X. Dong, G. F. Chen, J. L. Luo, N. L. Wang, G. Wang, Y. Zhou, Y. Zhu, X. Wang, Z. Zhao, Z. Xu, C. Chen, and X. J. Zhou, Chin. Phys. Lett. 25, 4402 (2008).
* [10] L. Wray, D. Qian, D. Hsieh, Y. Xia, L. Li, J. G. Checkelsky, A. Pasupathy, K. K. Gomes, C. V. Parker, A. V. Fedorov, G. F. Chen, J. L. Luo, A. Yazdani, N. P. Ong, N. L. Wang, and M. Z. Hasan, Phys. Rev. B 78, 184508 (2008).
* [11] K. Nakayama, T. Sato, P. Richard, Y.-M. Xu, Y. Sekiba, S. Souma,G. F. Chen, J. L. Luo, N. L. Wang, H. Ding, and T. Takahashi, arXiv:0812.0663
* [12] A. P. Mackenzie, S. R. Julian, A. J. Diver, G. G. Lonzarich, N. E. Hussey, Y. Maeno, S. Nishizaki, and T. Fujita, Physica C 263 510 (1996).
* [13] G. Mu, H. Luo, Z. Wang, Z. Ren, L. Shan, C. Ren, and H.-H. Wen, arXiv:0812.1188
* [14] The experiment was performed at the nominal doping of 0.1$e;$ in the calculations[4], the DOS drops sharply from 2.1 st/eV f.u. to 1.5 st./eV f.u. at $x\approx 0.15.$ The former value corresponds to renormalization between 1 and 1.2, and the latter to 1.5–2.
* [15] J. K. Dong, L. Ding, H. Wang, X. F. Wang, T. Wu, X. H. Chen, and S. Y. Li, arXiv:0806.3573
* [16] A.I. Coldea, J.D. Fletcher, A. Carrington, J.G. Analytis, A.F. Bangura, J.-H. Chu, A.S. Erickson, I.R. Fisher, N.E. Hussey, and R.D. McDonald, arXiv:0807.4890
* [17] D. H. Lu, M. Yi, S.-K. Mo, A. S. Erickson, J. Analytis, J.-H. Chu, D. J. Singh, Z. Hussain, T. H. Geballe, I. R. Fisher, and Z.-X. Shen, Nature 455, 81 (2008 ).
* [18] S.E. Sebastian, J. Gillett, N. Harrison, P. H. C. Lau, C. H. Mielke, and G. G. Lonzarich, J. Phys.: Condens. Matter 20 422203 (2008), and unpublished
* [19] J.G. Analytis, I.R. Fisher $et$ $al,$ unpublished
* [20] K. Terashima, Y. Sekiba, J. H. Bowen, K. Nakayama, T. Kawahara, T. Sato, P. Richard, Y.-M. Xu, L. J. Li, G. H. Cao, Z.-A. Xu, H. Ding, and T. Takahashi, arXiv:0812.3704
* [21] K. Matan1, R. Morinaga, K. Iida, and T. J. Sato, arXiv:0810.4790
* [22] I. I. Mazin and M. D. Johannes, Nature Physics (2009).
* [23] M. A. McGuire, A. D. Christianson, A. S. Sefat, B. C. Sales, M. D. Lumsden, R. Jin, E. A. Payzant, D. Mandrus, Y. Luan, V. Keppens, V. Varadarajan, J. W. Brill, R.P. Hermann, M. T. Sougrati, F. Grandjean, and G. J. Long, Phys. Rev. B 78, 094517 (2008).
* [24] Y. Chen, J. W. Lynn, J. Li, G. Li, G. F. Chen, J. L. Luo, N. L. Wang, P. Dai, C. dela Cruz, and H. A. Mook, Phys. Rev. B 78, 064515 (2008).
* [25] J. Zhao, Q. Huang, Cl. de la Cruz, S. Li, J. W. Lynn, Y. Chen, M. A. Green, G. F. Chen, G. Li, Z. Li, J. L. Luo, N. L. Wang, and P. Dai, Nature Materials 7, 953 (2008).
* [26] Q. Huang, Y. Qiu, W. Bao, J.W. Lynn, M.A. Green, Y. Chen, T. Wu, G. Wu, and X.H. Chen, Phys. Rev. Lett. 101, 257003 (2008).
* [27] Y. Qiu, W. Bao, Q. Huang, T. Yildirim, J. Simmons, J.W. Lynn, Y.C. Gasparovic, J. Li, M. Green, T. Wu, G. Wu, and X.H. Chen, Phys. Rev. Lett. 101, 257002 (2008).
* [28] J. Dong, H. J. Zhang, G. Xu, Z. Li, G. Li, W. Z. Hu, D. Wu, G. F. Chen, X. Dai, J. L. Luo, Z. Fang, and N. L. Wang, Europhysics Letters, 83, 27006 (2008).
* [29] W. Z. Hu, J. Dong, G. Li, Z. Li, P. Zheng, G. F. Chen, J. L. Luo, N. L. Wang, Phys. Rev. Lett. 101, 257005 (2008).
* [30] D. Hsieh, Y. Xia, L. Wray, D. Qian, K. Gomes, A. Yazdani, G.F. Chen, J.L. Luo, N.L. Wang, and M.Z. Hasan, arXiv:0812.2289.
* [31] X. F. Wang, T. Wu, G. Wu, H. Chen, Y. L. Xie, J. J. Ying, Y. J. Yan, R. H. Liu, and X. H. Chen, arXiv:0806.2452.
* [32] Not only the energetics of collinear LDA calculations cannot be mapped upon a $J_{1}+J_{2}$ model[100] (which required forced FM calculations, it was also directly demonstrated by noncollinear LDA calculations by Stuttgart group[102].
* [33] W. Bao, Y. Qiu, Q. Huang, M.A. Green, P. Zajdel, M.R. Fitzsimmons, M. Zhernenkov, M. Fang, B. Qian, E.K. Vehstedt, J. Yang, H.M. Pham, L. Spinu, and Z.Q. Mao, arXiv:0809.2058.
* [34] DFT calculations[4] yield the following magnetic stabilization energies for FeTe: 332 (227) meV/Fe with (without) spin-orbit coupling in the experimental P21/m double-stripe structure, and 281 (205) meV/Fe with in the single-stripe Pccm structure (with full optimization of atomic positions in each structure). Thus, the experimental structure appears to be below the competing structure by 22 (or 49, with spin- orbit) meV/Fe. Exactly the same calculation for LaFeAsO[100] render the double-stripe structure well above the single-stripe one. The difference can be traced down to the fact that for LaFeAsO only the single stripe structure opens a pseudogap in DOS at the Fermi level, while for FeTe both AFM structures open comparable pseudogaps. Further details will be published elsewhere[101].
* [35] T. Kroll, S. Bonhommeau, T. Kachel, H.A. Dürr, J. Werner, G. Behr, A.Koitzsch, R. Hübel, S. Leger, R. Schönfelder, A. Ariffin, R. Manzke, F.M.F. de Groot, J. Fink, H. Eschrig, B. Büchner, and M. Knupfer, Phys. Rev. B 78, 220502(R) (2008).
* [36] K. Haule, G. Kotliar, arXiv:0805.0722v1.
* [37] F. Wang, H. Zhai, D.Lee arXiv:0810.2320.
* [38] A.V. Chubukov, D. Efremov, and I. Eremin, Phys. Rev. B 78, 134512 (2008).
* [39] V. Stanev, J. Kang, and Z. Tesanovic, Phys. Rev. B 78, 184509 (2008).
* [40] K. Kuroki, S. Onari, R. Arita, H. Usui, Y. Tanaka, H. Kontani, and H. Aoki, Phys. Rev. Lett. 101, 087004 (2008).
* [41] S. Graser, G.R. Boyd, C. Cao, H.-P. Cheng, P.J. Hirschfeld, and D.J. Scalapino, Phys. Rev. B 77, 180514(R) (2008).
* [42] S. Graser, T. A. Maier, P. J. Hirschfeld, D. J. Scalapino, arXiv:0901.2653, to appear in: New Journal of Physics (2009).
* [43] X.-L. Qi, S. Raghu, C.-X. Liu, D. J. Scalapino, and S.-C. Zhang, arXiv:0804.4332.
* [44] Z.-J. Yao, J.-X. Li, and Z. D. Wang, arXiv:0804.4166.
* [45] R. Sknepnek, G. Samolyuk, Y. Lee, J. Schmalian, arXiv:0807.4566, to appear in Phys. Rev. B 79 (2009).
* [46] N. E. Bickers, D. J. Scalapino and S. R. White, Phys. Rev. Lett. 62, 961 (1989).
* [47] Ar. Abanov, A. V. Chubukov, and J. Schmalian, Advances in Physics 52, 119 (2003).
* [48] Q. Si and E. Abrahams, Phys. Rev. Lett. 101, 076401 (2008).
* [49] Experimentally observed Drude weight is sizeable above the Neel temperature ($\omega_{p}>1.5$ eV). Some reserachers report substantial reduction of the Drude weight in the antiferromagnetic state, consistent with Fermi surface gapping, (W. Z. Hu, Q. M. Zhang, and N. L. Wang, arXiv:0902.0435), while other do not see such large reduction (F. Pfuner, J.G. Analytis, J.-H. Chu, I.R. Fisher, and L. Degiorgi, arXiv:0811.2195).
* [50] C. Fang, H. Yao, W.-F. Tsai, J. Hu, S. A. Kivelson, Phys. Rev. B 77, 224509 (2008).
* [51] C. Xu, M. Mueller, and S. Sachdev, Phys. Rev. B 78, 020501(R) (2008).
* [52] T. Yildirim, Phys. Rev. Lett. 101, 057010 (2008).
* [53] D.J. Singh andM.H. Dur, Phys. Rev. Lett. 100 237003 (2008).
* [54] See, for instance, I.I. Mazin, Phys. Rev. B 75, 094407 (2007).
* [55] S. Sachdev, Physica A 313, 252 (2002).
* [56] K. Seo, B. A. Bernevig, and J. Hu, Phys. Rev.Lett. 101, 206404 (2008).
* [57] L. Boeri, O.V. Dolgov, and A.A. Golubov, Phys. Rev. Lett. 101, 026403 (2008).
* [58] F. Yndurain and J.M. Soler, arXiv:0810.2474
* [59] M. A. McGuire, A. D. Christianson, A. S. Sefat, B. C. Sales, M. D. Lumsden, R. Jin, E. A. Payzant, D. Mandrus, Y. Luan, V. Keppens, V. Varadarajan, J. W. Brill, R. P. Hermann, M. T. Sougrati, F. Grandjean, and G. J. Long, Phys. Rev. B 78, 094517 (2008).
* [60] M. Rotter, M. Tegel, D. Johrendt, I. Schellenberg, W. Hermes, and R. Poettgen, Phys. Rev. B 78, 020503(R) (2008).
* [61] N. Ni, S. Nandi, A. Kreyssig, A. I. Goldman, E. D. Mun, S. L. Bud’ko, and P. C. Canfield, Phys. Rev. B 78, 014523 (2008).
* [62] A. Jesche, N. Caroca-Canales, H. Rosner, H. Borrmann, A. Ormeci, D. Kasinathan, H. H. Klauss, H. Luetkens, R. Khasanov, A. Amato, A. Hoser, K. Kaneko, C. Krellner, and C. Geibel, Phys. Rev. B 78, 180504(R) (2008)
* [63] G. D. Samolyuk and V. P. Antropov, arXiv:0810.1445.
* [64] P. Chandra, P. Coleman, and A. I. Larkin, Phys. Rev.Lett 64, 88 (1990).
* [65] In reality there is no such degeneracy even on the mean-field level, reflecting the fact that, as discussed above, actual DFT calculations cannot be mapped upon a J1+J2 model. As shown by Yaresko et al[102], mutually orthogonal sublattices calculate to a have much larger energy than parallel.
* [66] R. Fernandes and J. Schmalian, to be published.
* [67] K. Kaneko, A. Hoser, N. Caroca-Canales, A. Jesche, C. Krellner, O. Stockert, C. Geibel, Phys. Rev. B 78, 212502 (2008).
* [68] C. Geibel, private communication.
* [69] V. Barzykin and L. P. Gor’kov, arXiv:0812.4277.
* [70] W. A. Little, Phys. Rev. 134A, 1416 (1964); 156, 396 (1967).
* [71] V. L. Ginzburg and D. A. Kirzhnits, Soviet Phys.—JETP 19, 269 (1964)].
* [72] M. Berciu, I. Elfimov, and G. A. Sawatzky, arXiv:0811.0214
* [73] D. F. Agterberg, V. Barzykin, and L. P. Gor’kov, Phys. Rev. B 60, 14868 (1999).
* [74] F. Ning, K. Ahilan, T. Imai, A. S. Sefat, R. Jin, M. A. McGuire, B. C. Sales, and D. Mandrus, J. Phys. Soc. Jap. 77, 103705 (2008)
* [75] N. Terasaki, H. Mukuda, M. Yashima, Y. Kitaoka, K. Miyazawa, P. Shirage, H. Kito, H. Eisaki, and A. Iyo, arXiv:0809.5155
* [76] Y. Nakai , K. Ishida, Y. Kamihara, M. Hirano, and H. Hosono, J. Phys. Soc. Jap. 77, 073701 (2008).
* [77] S. Kawasaki, K. Shimada, G. F. Chen, J. L. Luo, N. L. Wang, and G.-q. Zheng, arXiv:0810.1818
* [78] Y. Kobayashi, A. Kawabata, S. C. Lee, T. Moyoshi, and M. Sato, arXiv:0901.2830
* [79] H. Fukazawa, T. Yamazaki, K. Kondo, Y. Kohori, N. Takeshita, P. M. Shirage, K. Kihou, K. Miyazawa, H. Kito, H. Eisaki, and A. Iyo, arXiv:0901.0177
* [80] Y. Nakai, K. Ishida, Y. Kamihara, M. Hirano, and H. Hosono, Phys. Rev. Lett. 101, 077006 (2008).
* [81] K. Hashimoto, T. Shibauchi, T. Kato, K. Ikada, R. Okazaki, H. Shishido,M. Ishikado, H. Kito, A. Iyo, H. Eisaki, S. Shamoto, and Y. Matsuda, arXiv:0806.3149
* [82] L. Malone, J.D. Fletcher, A. Serafin, and A. Carrington, N.D. Zhigadlo, Z. Bukowski, S. Katrych, and J. Karpinski, arXiv:0806.3908
* [83] C. Martin, R. T. Gordon, M. A. Tanatar, M. D. Vannette, M. E. Tillman, E. D. Mun, P. C. Canfield, V. G. Kogan, G. D. Samolyuk, J. Schmalian, and R. Prozorov, arXiv: 0812.3683, and unpublished.
* [84] R. Prozorov, M. A. Tanatar, R. T. Gordon, C. Martin, H. Kim, V. G. Kogan, N. Ni, M. E. Tillman, S. L. Bud’ko, and P. C. Canfield, arXiv:0901.3698.
* [85] R. T. Gordon, C. Martin, H. Kim, N. Ni, M. A. Tanatar, J. Schmalian, I. I. Mazin S. L. Bud’ko, and P. C. Canfield, and R. Prozorov, arXiv:0812.3683
* [86] J.D. Fletcher, A. Serafin, L. Malone, J. Analytis, J-H Chu, A.S. Erickson, I.R. Fisher, and A. Carrington, arXiv:0812.3858.
* [87] D. Parker, O.V. Dolgov, M.M. Korshunov, A.A. Golubov, and I.I. Mazin, Phys. Rev. B 78, 134524 (2008).
* [88] A. B. Vorontsov, M. G. Vavilov, and A. V. Chubukov, arXiv:0901.0719
* [89] A.A. Golubov, A. Brinkman, O.V. Dolgov, I.I. Mazin, and Y. Tanaka, arXiv:0812.5057
* [90] W.-F. Tsai, D.-X. Yao, B. A. Bernevig, and J.P. Hu, arXiv:0812.0661
* [91] D. Parker and I.I. Mazin, arXiv:0812.4416
* [92] C.W. Hicks, T.M. Lippman, M.E. Huber, Z.A. Ren, Z.X. Zhao, and K.A. Moler, arXiv:0807.0467.
* [93] X. Zhang, Y. S. Oh, Y. Liu, L. Yan, K. H. Kim, R. L. Greene, I. Takeuchi, arXiv:0812.3605.
* [94] T. Maier and D. J. Scalapino, Phys. Rev. B 78, 20514(R) (2008).
* [95] M. M. Korshunov and I. Eremin, Europhys. Lett. 83, 67003 (2008).
* [96] A. D. Christianson, E. A. Goremychkin, R. Osborn, S. Rosenkranz, M. D. Lumsden, C. D. Malliakas, S. Todorov, H. Claus, D. Y. Chung, M. G. Kanatzidis, R. I. Bewley, and T. Guidi, arXiv:0807.3932; M. D. Lumsden, A. D. Christianson, D. Parshall, M. B. Stone, S. E. Nagler, H. A. Mook, K. Lokshin, T. Egami, D. L. Abernathy, E. A. Goremychkin, R. Osborn, M. A. McGuire, A. S. Sefat, R. Jin, B. C. Sales, D. Mandrus, arXiv:0811.4755
* [97] A. V. Chubukov, I. Eremin, and M. M. Korshunov, arXiv:0901.2102
* [98] V. Mishra, G. Boyd, S. Graser, T. Maier, P.J. Hirschfeld, and D.J. Scalapino, arXiv:0901.2653
* [99] G. Preosti and P. Muzikar, Phys. Rev. B 54, 3489 (1996)
* [100] I.I. Mazin, M.D. Johannes, L. Boeri, K. Koepernik, and D.J. Singh, Phys. Rev. B78, 085104 (2008).
* [101] I.I. Mazin, M.D. Johannes, to be published.
* [102] A. N. Yaresko, G.-Q. Liu, V. N. Antonov, O.K. Andersen, arXiv:0810.4469.
|
# Irreversibility of Structure Tensors of Modules
Maciej Wojtala
###### Abstract
Determining the matrix multiplication exponent $\omega$ is one of the greatest
open problems in theoretical computer science. We show that it is impossible
to prove $\omega=2$ by starting with structure tensors of modules of fixed
degree and using arbitrary restrictions. It implies that the same is
impossible by starting with $1_{A}$-generic non-diagonal tensors of fixed size
with minimal border rank. This generalizes the work of Bläser and Lysikov [3].
Our methods come from both commutative algebra and complexity theory.
Keywords: matrix multiplication complexity, minimal border rank tensors,
structure tensors for modules.
## 1 Introduction
Determining the matrix multiplication exponent $\omega$ is one of the most
important problems in theoretical computer science. Trivial bounds are
$2\leq\omega\leq 3$. In the classical paper V. Strassen proved a non-trivial
bound $\omega\leq\log_{2}7<2.81$ [15]. D. Coppersmith and S. Winograd proved
the bound $\omega<2.376$ [9]. This result was recently slightly improved [10,
16, 13] resulting with the best known upper bound $\omega<2.373$ with rounding
to the third decimal place.
No non-trivial lower bound is known and the conjecture states that $\omega=2$.
However since Coppersmith-Winograd there was very little progress in inventing
more efficient algorithms and obtaining better upper bounds. Recent papers
give an explanation for the phenomenon - many currently used approaches cannot
result with an algorithm giving $\omega=2$. For instance the laser method used
for big Coppersmith-Winograd tensors cannot show $\omega=2$, in fact it cannot
even show $\omega\leq 2.30$ [1, 2]. Also the framework proposed by Umans and
Cohn using reducing matrix multiplication to group algebra multiplication
cannot show $\omega=2$ for abelian groups and certain non-abelian groups [4,
5]. M. Christiandl, P. Vrana and J. Zuiddam introduced a quantity called
irreversibility and proved that it is impossible to show $\omega=2$ using
arbitrary restrictions starting with irreversible tensors (i.e. with
irreversibility greater than one) [8].
M. Bläser and V. Lysikov showed in their paper [3], that one cannot prove
$\omega=2$ over $\mathbb{C}$ using arbitrary restrictions and starting with
powers of structure tensors of non-semisimple algebras with bounded dimension.
This result is quite general, since the class of tensors that are structural
tensors of some algebra is quite large, larger than previously considered
classes, and using arbitrary restrictions is not a restrictive assumption. For
instance, Coppersmith-Winograd tensors are in this class. Let us name the
three coordinates of our tensors by $V_{1},\>V_{2},\>V_{3}$, i.e. our tensors
belong to the space $V_{1}\otimes V_{2}\otimes V_{3}$ where
$V_{i}\simeq\mathbb{C}^{n}$. Bläser and Lysikov use this result to conclude
that it is impossible to show $\omega=2$ using arbitrary restrictions and
starting with tensors that are both $1_{V_{1}}$\- and $1_{V_{2}}$-generic (so-
called binding tensors) with minimal border rank and being non-diagonal. This
result is really interesting since small border rank is believed to be
desirable in obtaining fast matrix multiplication. An important note is that
one could still try to prove $\omega=2$ by taking sequence of dimensions going
to infinity (since in the result there is an assumption that dimensions are
bounded).
This paper is an extension of results obtained by Bläser and Lysikov. We
consider an arbitrary algebraically closed field $\mathbb{K}$ (of arbitrary
characteristic) and tensors from
$\mathbb{K}^{n}\otimes\mathbb{K}^{n}\otimes\mathbb{K}^{n}$. From the
perspective of commutative algebra a natural generalisation is to consider
modules instead of algebras. In this paper we show such a generalisation. To
achieve that we define a structure tensor of a module. The generalisation
theorem is the main result of this paper:
###### Theorem 1.
For bounded $n$ it is impossible to prove $\omega=2$ over $\mathbb{K}$ using
arbitrary restrictions and starting with powers of tensors of size $n$ that
are isomorphic to some structure tensor of a non-semisimple module.
This allows us also to extend the corollary obtained by Bläser and Lysikov:
###### Corollary 2.
For bounded $n$ it is impossible to prove $\omega=2$ over $\mathbb{K}$ using
arbitrary restrictions and starting with powers of tensors of size $n$ that
are $1_{V_{1}}$-generic, have minimal border rank and have rank larger than
$n$.
It means that we only need to assume $1_{V_{1}}$-genericity - assuming
$1_{V_{2}}$-genericity is not necessary. There are plenty of tensors that are
$1_{V_{1}}$-generic but not $1_{V_{2}}$\- or $1_{V_{3}}$-generic; our
generalization applies to them as well.
The approach in the proof is based on the one in Bläser and Lysikov paper [3],
but there are several issues. The main part focuses on showing that a
structure tensor of a non-semisimple module is $0$-subtight-unstable. To
achieve this we use the quotient ring $A$ obtained by dividing the polynomial
ring by the annihilator of the considered module. We show that ring $A$ is
Artinian. It allows to create a filtration of the module induced by powers of
a nilradical of $A$. We also induce the filtration of the polynomial ring (by
taking the preimage of the filtration of $A$) and of its subset consisting of
forms of degree at most one (by restricting the filtration of the polynomial
ring). Then we will show that it suffices to prove that the last filtration is
non-trivial. To prove that this filtration is indeed non-trivial we will
analyze the spectrum of the quotient ring and by Hilbert’s Nullstellensatz use
it to analyze affine functions which correspond to forms of degree at most
one.
It is unclear if the assumption of $1_{V_{1}}$-genericity can be replaced with
conciseness; the following question remains open:
###### Question 3.
Do there exist concise minimal border rank (with rank greater than border
rank) tensors that are stable?
### Acknowledgements
During preparation of this publication the author was part of the Szkoła Orłów
programme. The publication was created under the supervision of Joachim
Jelisiejew, whose help and support were invaluable. The author also wants to
thank Markus Bläser, Vladimir Lysikov, Joseph Landsberg and the anonymous
referee for helpful comments.
## 2 Definitions
Let $\mathbb{K}$ be an algebraically closed field and let
$S=\mathbb{K}[x_{1},\>x_{2},\>\ldots,\>x_{n-1}]$.
We denote by $S_{\leq 1}$ the $\mathbb{K}$-linear subspace of polynomials of
degree at most one.
Let us consider an $S$-module $M$. Observe that structure of multiplication in
this module is uniquely determined by multiplication by elements of $S_{\leq
1}$. Furthermore, multiplication by $S_{\leq 1}$ is uniquely determined by the
multiplication by a basis of $S_{\leq 1}$. We define the structure tensor of
$M$ as $\sigma_{M}\in S_{\leq 1}^{*}\otimes M^{*}\otimes M$ or equivalently
$\sigma_{M}\colon S_{\leq 1}\otimes M\rightarrow M$. For a choice of
coordinates on the first, second and third factors
($\\{A_{i}\\},\>\\{B_{j}\\},\\{C_{k}\\}$ respectively), $\sigma_{M}$ simply
encodes the result of multiplying $A_{i}$ by $B_{j}$ in $\\{C_{k}\\}$ basis.
Note that using the polynomial algebra is a quite general assumption. Indeed,
let $A$ be a commutative unital algebra with $\dim_{\mathbb{K}}A=n$, let $M$
be a module over $A$ of dimension $n$ and $t$ be the tensor corresponding to
the bilinear map $A\otimes M\to M$. Let $(1,\>g_{1},\>g_{2},\ldots,\>g_{n-1})$
span the algebra as a $\mathbb{K}$-linear space. Consider the unique
surjection $S\to A$ that sends $x_{i}$ to $g_{i}$. Then $M$ becomes an
$S$-module and $t$ identifies with the structure tensor of the $S$-module $M$.
In particular, after putting $M=A$ we deduce that the structure tensor of any
commutative unital algebra with $\dim_{\mathbb{K}}A=n$ is isomorphic to the
structure tensor of the obtained $S$-module. Note that using the unity is
needed to assure $1_{V_{1}}$-genericity. In comparison to [3] we assume that
the algebra $A$ is commutative, however due to [12, Lemma 2.6] it is satisfied
when its structure tensor is of minimal border rank (so Corollary 2
strengthens analogous result by Bläser and Lysikov [3, Corollary 25]).
###### Example 4.
We consider $\mathbb{K}^{4}$, where we treat elements as column vectors. Let
us also consider matrices:
$A_{1}=\begin{bmatrix}0&0&1&0\\\ 0&0&0&0\\\ 0&0&0&0\\\ 0&0&0&0\\\
\end{bmatrix},\qquad A_{2}=\begin{bmatrix}0&0&0&1\\\ 0&0&0&0\\\ 0&0&0&0\\\
0&0&0&0\\\ \end{bmatrix},\qquad A_{3}=\begin{bmatrix}0&0&0&0\\\ 0&0&1&0\\\
0&0&0&0\\\ 0&0&0&0\\\ \end{bmatrix}.$
The matrices clearly pairwise commute, so we can introduce a structure of a
$\mathbb{K}[x_{1},\>x_{2},\>x_{3}]$-module on $\mathbb{K}^{4}$ where
multiplication of a vector $v$ by $x_{i}$ is simply the left multiplication of
$v$ by $A_{i}$ (so we identify $x_{i}$ with $A_{i}$ and take as a
multiplication in the module the left multiplication of a vector by a matrix).
Let us take the standard basis $(e_{1},\>e_{2},\>e_{3},\>e_{4})$ of
$\mathbb{K}^{4}$. We also have the basis of
$\mathbb{K}[x_{1},\>x_{2},\>x_{3}]_{\leq 1}$: $(A_{0}\coloneqq
Id,\>A_{1},\>A_{2},\>A_{3})$ (since we identified $x_{i}$ with $A_{i}$ for
$i=1,\>2,\>3$). To obtain the structure tensor we need to verify the results
of pairwise multiplications.
We have
$A_{0}\cdot e_{i}=e_{i},\qquad A_{1}\cdot e_{i}=\begin{bmatrix}\delta_{i=3}\\\
0\\\ 0\\\ 0\\\ \end{bmatrix},\qquad A_{2}\cdot
e_{i}=\begin{bmatrix}\delta_{i=4}\\\ 0\\\ 0\\\ 0\\\ \end{bmatrix},\qquad
A_{3}\cdot e_{i}=\begin{bmatrix}0\\\ \delta_{i=3}\\\ 0\\\ 0\\\ \end{bmatrix}.$
So the structure tensor of this module is
$A_{0}^{*}\otimes e_{1}^{*}\otimes e_{1}+A_{0}^{*}\otimes e_{2}^{*}\otimes
e_{2}+A_{0}^{*}\otimes e_{3}^{*}\otimes e_{3}+A_{0}^{*}\otimes
e_{4}^{*}\otimes e_{4}+A_{1}^{*}\otimes e_{3}^{*}\otimes
e_{1}+A_{2}^{*}\otimes e_{4}^{*}\otimes e_{1}+A_{3}^{*}\otimes
e_{3}^{*}\otimes e_{2}.$
Above we explained that every structure tensor of a commutative unital algebra
is the structure tensor of a module. Here we show that for an algebra
$A=\nicefrac{{S}}{{I}}$ the structure tensor of $A$ as an algebra and as an
$S$-module may differ. The difference comes from the fact that a given algebra
can have many $S$-module structures: the structure coming from a span
$(1,\>g_{1},\>g_{2},\ldots,\>g_{n-1})$ is in general different from the
structure coming from $\nicefrac{{S}}{{I}}$.
###### Example 5.
Let us consider the algebra
$\nicefrac{{\mathbb{K}[x_{1},\>x_{2}]}}{{(x_{1}^{3},\>x_{2})}}$. We have
standard basis of $\mathbb{K}[x_{1},\>x_{2}]_{\leq 1}$: $(A_{0}\coloneqq
1,\>A_{1}\coloneqq x_{1},\>A_{2}\coloneqq x_{2})$ and standard basis of the
algebra: $(A_{0},\>A_{1},\>A_{3}\coloneqq x_{1}^{2})$. To obtain the structure
tensors we need to verify results of pairwise multiplications, however for
structure tensor of module we multiply elements from the basis of
$\mathbb{K}[x_{1},\>x_{2}]_{\leq 1}$ by the basis of the algebra and for the
structure tensor of algebra - elements from the basis of the algebra by
themselves.
So the structure tensor of the module and the algebra are respectively
$\displaystyle A_{0}^{*}\otimes A_{0}^{*}\otimes A_{0}+A_{0}^{*}\otimes
A_{1}^{*}\otimes A_{1}+A_{0}^{*}\otimes A_{3}^{*}\otimes
A_{3}+A_{1}^{*}\otimes A_{0}^{*}\otimes A_{1}+A_{1}^{*}\otimes
A_{1}^{*}\otimes A_{3},$ $\displaystyle A_{0}^{*}\otimes A_{0}^{*}\otimes
A_{0}+A_{0}^{*}\otimes A_{1}^{*}\otimes A_{1}+A_{0}^{*}\otimes
A_{3}^{*}\otimes A_{3}+A_{1}^{*}\otimes A_{0}^{*}\otimes
A_{1}+A_{1}^{*}\otimes A_{1}^{*}\otimes A_{3}+A_{3}^{*}\otimes
A_{0}^{*}\otimes A_{3}.$
Note that both structure tensors come from the
$\mathbb{K}[x_{1},\>x_{2}]$-module structure on $A$: the structure tensor of
the module encodes the results of multiplication by $1,\>x_{1},\>x_{2}$, while
the structure tensor of the algebra encodes the results of multiplication by
$1,\>x_{1},\>x_{1}^{2}$.
A module is simple if it is non-zero and has no non-zero proper submodules. A
module is semi-simple if it is a direct sum of simple modules.
An arbitrary restriction of a tensor $t\in V_{1}\otimes V_{2}\otimes V_{3}$ is
$\phi(t)$, where $\phi\colon V_{1}\otimes V_{2}\otimes V_{3}\to
V_{1}^{\prime}\otimes V_{2}^{\prime}\otimes V_{3}^{\prime}$ is a linear map
induced by a triple of linear maps $\phi_{i}\colon V_{i}\to V_{i}^{\prime}$.
For a tensor $t\in V_{1}\otimes V_{2}\otimes V_{3}$ and a linear form $x\in
V_{1}^{*}$, the contraction $t\cdot x$ is defined as $(v_{1}\otimes
v_{2}\otimes v_{3})\cdot x=x(v_{1})(v_{2}\otimes v_{3})$ for rank one tensors
and extended to arbitrary tensors by linearity. Thus, a tensor $t\in
V_{1}\otimes V_{2}\otimes V_{3}$ defines a map $V_{1}^{*}\rightarrow
V_{2}\otimes V_{3}$ sending $x$ to $t\cdot x$. The two other maps
$V_{2}^{*}\rightarrow V_{1}\otimes V_{3}$ and $V_{3}^{*}\rightarrow
V_{1}\otimes V_{2}$ can be defined similarly. These maps are called
flattenings of the tensor $t$. A tensor is called concise if all its
flattenings are injective. Such a tensor does not lie in any non-trivial
subspace $V_{1}^{\prime}\otimes V_{2}^{\prime}\otimes V_{3}^{\prime}$ with
$V_{k}^{\prime}\subset V_{k}$. We denote the maximum of the three ranks of the
flattenings by $N(t)$. For a concise tensor, the ranks of the flattenings are
the dimensions of $V_{k}$, and
$N(t)=\max\left\\{\dim_{\mathbb{K}}V_{1},\>\dim_{\mathbb{K}}V_{2},\>\dim_{\mathbb{K}}V_{3}\right\\}$.
A tensor $t\in V_{1}\otimes V_{2}\otimes V_{3}$ is called $1_{V_{1}}$-generic
if $\dim_{\mathbb{K}}V_{2}=\dim_{\mathbb{K}}V_{3}$ and there exists $x\in
V_{1}^{*}$ such that the matrix $t\cdot x\in V_{2}\otimes V_{3}$ has full
rank. The notions of $1_{V_{2}}$-genericity and $1_{V_{3}}$-genericity are
defined analogously [3].
A block tensor is a tensor $t\in V_{1}\otimes V_{2}\otimes V_{3}$ with a
triple of direct sum decompositions $V_{1}=\bigoplus_{i\in
I_{1}}V_{1,\>i},\>V_{2}=\bigoplus_{i\in
I_{2}}V_{2,\>i},\>V_{3}=\bigoplus_{i\in I_{3}}V_{3,\>i}$. The decompositions
of $V_{k}$ induce the decomposition of the tensor space $V_{1}\otimes
V_{2}\otimes V_{3}=\bigoplus_{(i_{1},\>i_{2},\>i_{3})\in I_{1}\times
I_{2}\times I_{3}}V_{1,\>i_{1}}\otimes V_{2,\>i_{2}}\otimes V_{3,\>i_{3}}$.
For a block tensor $t$, we denote by $t_{i_{1}i_{2}i_{3}}$ its projection onto
$V_{1,\>i_{1}}\otimes V_{2,\>i_{2}}\otimes V_{3,\>i_{3}}$ [3, Definition 8].
The support of a block tensor $t$ is defined as
$supp\;t=\\{(i_{1},\>i_{2},\>i_{3})\in I_{1}\times I_{2}\times I_{3}\mid
t_{i_{1}i_{2}i_{3}}\neq 0\\}$ [3, Definition 9].
The block format of a block tensor is a triple $(n_{1},\>n_{2},\>n_{3})$ of
maps $n_{k}\colon I_{k}\rightarrow\mathbb{N}$ defined as
$n_{k}(i)=\dim_{\mathbb{K}}V_{k,\>i},\>k=1,\>2,\>3$. The relative block format
is a triple $(f_{1},\>f_{2},\>f_{3})$ defined as
$f_{k}(i)=\frac{n_{k}(i)}{N_{k}}$ where $N_{k}=\dim_{\mathbb{K}}V_{k}$. [3,
Definition 10]. A subset $S\subset I_{1}\times I_{2}\times I_{3}$ is called
s-subtight with numbering given by three maps $a_{k}\colon
I_{k}\rightarrow\mathbb{Z}$ if for each $(i_{1},\>i_{2},\>i_{3})\in S$ we have
$a_{1}(i_{1})+a_{2}(i_{2})+a_{3}(i_{3})\leq s$ [3, Definition 13].
A tensor $t\in V_{1}\otimes V_{2}\otimes V_{3}$ is unstable, if $0$ is
contained in the Zariski closure of $SL(V_{1})\times SL(V_{2})\times
SL(V_{3})$ orbit of $t$.
A set $S\in I_{1}\times I_{2}\times I_{3}$ is a combinatorially unstable
support in block format $(n_{1},\>n_{2},\>n_{3})$ if there exist exponents
$u_{k}\colon I_{k}\rightarrow\mathbb{Q}$ such that $\sum_{i\in
I_{k}}n_{k}(i)u_{k}(i)=0$ for each $k$ and
$u_{1}(i_{1})+u_{2}(i_{2})+u_{3}(i_{3})>0$ for each
$(i_{1},\>i_{2},\>i_{3})\in S$ [3, Definition 19]. A tensor is combinatorially
unstable if its support is a combinatorially unstable support.
It turns out that combinatorially unstable tensors are unstable [3,
Proposition 20].
Let a tensor $t$ have $s$-subtight support with numbering
$(a_{1},\>a_{2},\>a_{3})$. Let $(f_{1},\>f_{2},\>f_{3})$ be a relative block
format of $t$ and $\overline{a}_{k}=\sum_{i\in I_{k}}f_{k}(i)a_{k}(i)$ for
$k=1,\>2,\>3$. If the inequality
$\overline{a}_{1}+\overline{a}_{2}+\overline{a}_{3}>s$ holds, then we say then
that the tensor $t$ is $s$-subtight-unstable. It turns out that $s$-subtight-
instability implies combinatorial instability [3, Theorem 21].
By $SR(t)$ we denote slice rank of tensor $t$. By $\widetilde{SR}(t)$ we
denote the asymptotic slice rank of tensor $t$, i.e.
$\widetilde{SR}(t)=\limsup_{m\in\mathbb{N}}{SR(t^{\otimes m})^{\frac{1}{m}}}$.
For a ring $R$ we denote by $rad(R)$ the nilradical of $R$, i. e. the ideal
consisting of elements which raised to some power give zero.
## 3 Irreversibility of Structure Tensors of Modules
Let $\mathbb{K}$ be an algebraically closed field. Let
$S=\mathbb{K}[x_{1},\>x_{2},\>\ldots,\>x_{n-1}]$. For $S$-module $M$ of rank
$n=\dim_{\mathbb{K}}S_{\leq 1}$ let $\sigma_{M}\colon S_{\leq 1}\otimes
M\rightarrow M$ be its structure tensor, so $\sigma_{M}\in S_{\leq
1}^{*}\otimes M^{*}\otimes
M\simeq\mathbb{K}^{n}\otimes\mathbb{K}^{n}\otimes\mathbb{K}^{n}$.
Let $F_{n}$ be the set of tensors
$t\in\mathbb{K}^{n}\otimes\mathbb{K}^{n}\otimes\mathbb{K}^{n}$ such that there
exist a non-semisimple $S$-module $M$ such that $t\simeq\sigma_{M}$ (so
$F_{n}$ is the set of tensors considered in Theorem 1).
Let us distinguish the coordinates by taking
$\mathbb{K}^{n}\otimes\mathbb{K}^{n}\otimes\mathbb{K}^{n}=A\otimes B\otimes
C$.
Before proving this corollary we introduce a useful lemma which classifies
simple modules. It is quite known but we present it for self-containment.
###### Lemma 6.
An $S$-module $M$ is simple $\iff$ there exists a maximal ideal $\mathfrak{m}$
of $S$ such that $M=\nicefrac{{S}}{{\mathfrak{m}}}$.
###### Proof.
"$\impliedby$": $\nicefrac{{S}}{{\mathfrak{m}}}$ is a field which is clearly
simple.
"$\implies$": Let $N$ be a simple $S$-module. Let $n_{0}$ be a non-zero
element of $N$. Then $Sn_{0}$ is a non-trivial submodule and thus $Sn_{0}=N$.
Let us define $\pi_{N}\colon S\rightarrow N$, $\pi_{N}(s)=sn_{0}$. By
$Sn_{0}=N$, we have that $\pi_{N}$ is surjective and thus
$N\simeq\nicefrac{{S}}{{\ker\pi_{N}}}$. It now suffices to show that
$\ker\pi_{N}$ is a maximal ideal of $S$.
Since $N\neq 0$, $\ker\pi_{N}$ is a subideal of some maximal ideal
$\mathfrak{m}$ of $S$. Let us suppose that
$\ker\pi_{N}\subsetneq\mathfrak{m}$. Then
$\nicefrac{{\mathfrak{m}}}{{\ker\pi_{N}}}\neq 0$ and clearly
$\nicefrac{{\mathfrak{m}}}{{\ker\pi_{N}}}\subseteq\nicefrac{{S}}{{\ker\pi_{N}}}=N$.
However $N$ is simple, so it must hold
$\nicefrac{{\mathfrak{m}}}{{\ker\pi_{N}}}=\nicefrac{{S}}{{\ker\pi_{N}}}$.
However it implies that
$\nicefrac{{S}}{{\mathfrak{m}}}=\nicefrac{{(\nicefrac{{S}}{{\ker\pi_{N}}})}}{{\nicefrac{{(\mathfrak{m}}}{{\ker\pi_{N}}})}}=0$,
which is a contradiction. ∎
###### Proof of Corollary 2.
Let $T_{n}$ be the set of tensors
$t\in\mathbb{K}^{n}\otimes\mathbb{K}^{n}\otimes\mathbb{K}^{n}$ such that $t$
is $1_{A}$-generic, the rank of $t$ is larger than $n$ and the border rank of
$t$ is $n$. Let $t\in T_{n}$. We need to show that $t\in F_{n}$.
By the assumption that $t$ is $1_{A}$-generic we know that there exist
$\alpha$ such that $t(\alpha):C\rightarrow B^{*}$ has maximal rank. Thus
$a\otimes b\otimes c\mapsto a\otimes b\otimes t(\alpha)(c)$ is a tensor
isomorphism between $t$ and a tensor $\widetilde{t}\in A\otimes B\otimes
B^{*}$. Let us denote $V=\widetilde{t}(A^{*})\subseteq B\otimes B^{*}=End(B)$.
By [12, Lemma 2.6] the linear subspace $V$ consists of commutative matrices.
Since $\widetilde{t}(\alpha)$ is an identity matrix, we have $Id\in V$. We
also clearly have $\dim_{\mathbb{K}}V\leq n$.
We can now choose a linear span $\\{1,\>v_{1},\>v_{2},\>\ldots,\>v_{n-1}\\}$
of $V$ (here $1$ is $Id$). We define a
$\mathbb{K}[x_{1},\>x_{2},\>\ldots,\>x_{n-1}]$-module structure on $V$ by
$x_{i}\cdot b=v_{i}(b)$ for $i=1,\>2,\>\ldots,\>n-1$ and for all $b\in B$.
By commutativity it is indeed a module structure, since $x_{i}\cdot(x_{j}\cdot
b)=v_{i}(v_{j}(b))=v_{j}(v_{i}(b))=x_{j}\cdot(x_{i}\cdot b)$. One can read
more about such a construction in [11]. Let us denote the module as
$M_{\widetilde{t}}$. By construction its structure tensor
$\sigma_{M_{\widetilde{t}}}$ is isomorphic to $\widetilde{t}$. Since
$\widetilde{t}\simeq t$, we obtain that $t\simeq\sigma_{M_{\widetilde{t}}}$,
so $t$ is isomorphic to $S_{n}$-module structure tensor.
We now need to argue that $M_{\widetilde{t}}$ is non-semisimple. However, if
the module $M_{\widetilde{t}}$ were semisimple, then by Lemma 6 it would be a
direct sum of modules of rank $1$, so its structure tensor would be diagonal
in some basis and thus its rank would be $n$. Since $t$ is isomorphic to this
structure tensor, it would imply that $rank(t)=n$, which is contradiction with
the assumption $rank(t)>n$. ∎
Let us now fix a degree $n$ and a module $M$.
To prove 1, we will use an approach that directly generalizes the approach
from [3] \- we will show that if $M$ is not semisimple, then $\sigma_{M}$ is
$0$-subtight-unstable.
###### Theorem 7.
The tensor $\sigma_{M}$ is $0$-subtight-unstable or $M$ is semisimple.
The proof is given later, after handful lemmas.
By $Ann(M)$ we denote the annihilator of module $M$ over $S$. Let
$A=\nicefrac{{S}}{{Ann(M)}}$ and let $\pi$ be a surjection from $S$ to $A$.
We use the quotient ring $A$ because it has much more convenient structure: in
fact, as we show in Lemma 9, it is an Artinian ring, which will let us
conclude a lot about its spectrum. By Hilbert’s Nullstellensatz it will allow
us to analyze affine functions over $\mathbb{K}$, which correspond to $S_{\leq
1}$. Artinian structure will also allow us to use the nilradical construction
to obtain filtrations induced by powers of nilradical. However, we will have
to struggle with one fundamental issue - we want the preimage of $rad(A)$ in
$S$ to be non-trivial. In fact, we want more - preimage of $rad(A)$ has to
have a non-trivial intersection with $S_{\leq 1}$.
The next three lemmas are quite classical commutative algebra arguments, but
we present them to assure self-containment of the paper.
###### Lemma 8.
There exist a natural number $r$ such that $rad(A)^{r}=0$.
###### Proof.
By Hilbert’s basis theorem $S$ is Noetherian ring and since $A$ is its
quotient, it is Noetherian too. Hence $rad(A)$ is finitely generated. Let us
denote its generators as $g_{1},\>g_{2}\>,\ldots,\>g_{k}$. Let $p_{i}$ be such
positive integers that $g_{i}^{p_{i}}=0$ for $i=1,\>2,\>\ldots\>k$. Let
$r=k\max(p_{i})$. Then $rad(A)^{r}=0$.
Indeed, if $a\in rad(A)$, then $a=\sum_{i=1}^{k}{c_{i}g_{i}}$ for some
$c_{i}\in A$ and
$a^{r}=(\sum_{i=1}^{k}{c_{i}g_{i}})^{r}=\sum_{a_{1}+a_{2}+\ldots+a_{k}=r}{b_{(a_{1},\>a_{2},\>\ldots,\>a_{k})}g_{1}^{a_{1}}g_{2}^{a_{2}}\ldots
g_{k}^{a_{k}}}$. For every component of the sum by the pigeonhole principle
for some $i$ we have $a_{i}\geq\max(p_{i})$, thus $g_{i}^{a_{i}}=0$ and so
$b_{(a_{1},\>a_{2},\>\ldots,\>a_{k})}g_{1}^{a_{1}}g_{2}^{a_{2}}\ldots
g_{k}^{a_{k}}=0$. ∎
###### Lemma 9.
$A$ is an Artinian ring.
###### Proof.
Since $A$ is an $\mathbb{K}$-algebra, it suffices to show that $A$ has finite
dimension over $\mathbb{K}$ (every descending sequence of ideals is a
descending sequence of $\mathbb{K}$-linear subspaces so if $A$ has finite
dimension over $\mathbb{K}$ then such a sequence clearly stabilizes).
Let us define $\psi\colon A\rightarrow Hom_{\mathbb{K}}(M,\>M)$ by
$\psi(s+Ann(M))=(m\rightarrow sm)$. By definition of $Ann(M)$ the map $\psi$
is well-defined. Clearly $\psi$ is $\mathbb{K}$-linear.
Let us now observe that $\psi$ is injective. Indeed, if $\psi(s+Ann(M))=0$,
then for all $m\in M$ $sm=0$ and thus $s\in Ann(M)$. Thus $A$ is isomorphic as
a $\mathbb{K}$-linear subspace with $\operatorname{im}(\psi)$, which has
finite dimension over $\mathbb{K}$ as a linear subspace of
$Hom_{\mathbb{K}}(M,\>M)$ which has finite dimension. ∎
###### Lemma 10.
The spectrum of an Artinian ring is finite and equals its maximal spectrum.
This lemma is well-known, but we add a proof for completeness.
###### Proof.
Let $R$ be an Artinian ring and let $p$ be its prime ideal. We will first
argue that $p$ is maximal. Let $\pi_{p}$ be a projection from $R$ to
$\nicefrac{{R}}{{p}}$. Let us observe that $\nicefrac{{R}}{{p}}$ is also
Artinian - for every descending sequence of ideals $I_{1},\>I_{2},\>\ldots$ in
$\nicefrac{{R}}{{p}}$ the sequence of preimages
$\pi_{p}^{-1}(I_{1}),\>\pi_{p}^{-1}(I_{2}),\>\ldots$ in $R$ stabilizes and
thus the sequence $I_{1},\>I_{2},\>\ldots$ also stabilizes. The ring
$\nicefrac{{R}}{{p}}$ is a domain and it suffices to show that it is a field.
Let $x\in\nicefrac{{R}}{{p}},\>x\neq 0$. Let us consider the descending
sequence $(x)\supseteq(x^{2})\supseteq\ldots$ Since it stabilizes, for some
$k$ holds equality $(x^{k})=(x^{k+1})$ with $k\geq 1$. Thus there exist
$a\in\nicefrac{{R}}{{p}}$ such that $x^{k}=x^{k+1}a$. So $x^{k}(1-xa)=0$ and
since $\nicefrac{{R}}{{p}}$ is a domain and $x\neq 0$ we obtain $xa=1$, so $x$
is invertible. So $\nicefrac{{R}}{{p}}$ is a field and thus $p$ is a maximal
ideal of $R$.
We will now show that maximal spectrum of $R$ is finite. Let us suppose
otherwise. Then there exists an infinite sequence of maximal ideals
$\mathfrak{m_{i}}$. Let us consider the descending sequence
$\mathfrak{m_{1}},\>\mathfrak{m_{1}}\cap\mathfrak{m_{2}},\>\mathfrak{m_{1}}\cap\mathfrak{m_{2}}\cap\mathfrak{m_{3}},\>\ldots$
Since it stabilizes there exists such $l$ that
$\mathfrak{m_{1}}\cap\mathfrak{m_{2}}\cap\ldots\cap\mathfrak{m_{l}}=\mathfrak{m_{1}}\cap\mathfrak{m_{2}}\cap\ldots\cap\mathfrak{m_{l}}\cap\mathfrak{m_{l+1}}$.
Clearly $\mathfrak{m_{i}}\not\subseteq\mathfrak{m_{j}}$ for $i\neq j$. Let
$a_{i}\in\mathfrak{m_{i}}\setminus\mathfrak{m_{l+1}}$ for
$i=1,\>2,\>\ldots,\>l$. Then $a_{1}a_{2}\ldots
a_{l}\in\mathfrak{m_{1}}\cap\mathfrak{m_{2}}\cap\ldots\cap\mathfrak{m_{l}}$,
so by our assumption $a_{1}a_{2}\ldots
a_{l}\in\mathfrak{m_{1}}\cap\mathfrak{m_{2}}\cap\ldots\cap\mathfrak{m_{l}}\cap\mathfrak{m_{l+1}}$,
so $a_{1}a_{2}\ldots a_{l}\in\mathfrak{m_{l+1}}$. Thus by primeness of
$\mathfrak{m_{l+1}}$ there exists $c$ such that $a_{c}\in\mathfrak{m_{l+1}}$.
It is a contradiction with the definition with $a_{i}$. ∎
We now will be using definitions referring to block tensors, which are
introduced in Section 2.
###### Proposition 11.
If $\pi^{-1}(rad(A))\cap S_{\leq 1}\neq 0$, then $\sigma_{M}$ is $0$-subtight-
unstable.
###### Proof.
We will use the same approach as in [3, Example 14].
By $r$ we denote the minimal natural number such that $rad(A)^{r}=0$; by Lemma
8 such a natural number exists. Let us take the sequence
$M\supseteq rad(A)M\supseteq rad(A)^{2}M\supseteq\ldots\supseteq
rad(A)^{r}M=0$
. Let us take the sequence $(M_{k})$ of linear subspaces satisfying:
* •
$rad(A)^{k-1}M=M_{k-1}\oplus rad(A)^{k}M$ for $r>k>1$,
* •
$M_{r}=0$.
By definition $(M_{i})_{i=0}^{r}$ is a decomposition of $M$.
Let $(R_{i})_{i=0}^{r}$ be a decomposition of $S_{\leq 1}$ induced by
$\pi^{-1}({rad(A)}^{k})\cap S_{\leq 1}$, i.e. $\pi^{-1}({rad(A)}^{k})\cap
S_{\leq 1}=R_{k}\oplus\pi^{-1}({rad(A)}^{k+1})\cap S_{\leq 1}$, $R_{r}=0$.
We have $R_{1}\oplus R_{2}\oplus\ldots\oplus R_{r}=\pi^{-1}(rad(A))\cap
S_{\leq 1}$, so the assumption $R_{1}\oplus R_{2}\oplus\ldots\oplus R_{r}\neq
0$.
For all $i,\>j\geq 0$ we have
$R_{i}\>M_{j}\subseteq(rad(A)^{i}+Ann(M))\>rad(A)^{j}M=rad(A)^{i+j}M=\bigoplus_{k\geq
i+j}M_{k}$.
Let us now consider the numbering $a_{1}(i)=a_{2}(i)=i.\>a_{3}(i)=-i$. The
structure tensor $\sigma_{M}$ is a block tensor with decompositions obtained
from decompositions of $R_{i}$ and $M_{i}$. Let $I_{1},\>I_{2},\>I_{3}$ be its
indexing sets. As we observed above, for all $(i_{1},\>i_{2},\>i_{3})$ such
that $a_{1}(i_{1})+a_{2}(i_{2})+a_{3}(i_{3})>0$ it holds that
$\sigma_{M}(R_{i_{1}},\>M_{i_{2}},\>M_{i_{3}})=0$. It means that $\sigma_{M}$
has $0$-subtight support. Let $(f_{1},\>f_{2},\>f_{3})$ be a relative block
format of $\sigma_{M}$. Let $\overline{a}_{k}=\sum_{i\in
I_{k}}f_{k}(i)a_{k}(i)$ for $k=1,\>2,\>3$. We need to to show that
$\overline{a}_{1}+\overline{a}_{2}+\overline{a}_{3}>0$. Since decompositions
on the second and the third coordinate are dual, we have
$I_{2}=I_{3},\>f_{2}=f_{3}$ and thus
$\overline{a}_{2}+\overline{a}_{3}=\sum_{i\in
I_{2}}f_{2}(i)\>a_{2}(i)+\sum_{i\in I_{3}}f_{3}(i)\>a_{3}(i)=\sum_{i\in
I_{2}}f_{2}(i)\>i+\sum_{i\in I_{3}}f_{3}(i)\>(-i)=0$. So we just need to show
$\overline{a}_{1}>0$. We have $\overline{a}_{1}=\sum_{i\in
I_{1}}f_{1}(i)\>a_{1}(i)=\sum_{i\in I_{1}}f_{1}(i)\>i$. In our case
$I_{1}=\\{0,\>1,\>\ldots,\>r\\}$ and as we observed before $R_{1}\oplus
R_{2}\oplus\ldots\oplus R_{r}\neq 0$, so there exist
$s\in\\{1,\>2,\>\ldots,\>r\\}$ such that $f_{1}(s)>0$. Thus we have
$\sum_{i\in I_{1}}f_{1}(i)\>i=\sum_{i=0}^{r}f_{1}(i)\>i\geq f_{1}(s)\>s>0$,
which ends the proof. ∎
Now we will argue that $\pi^{-1}(rad(A))\cap S_{\leq 1}$ is zero only for
semisimple modules. To prove it, we will analyze the spectrum of $A$ and the
support of $M$.
Let us define $supp\>M$ as a set of such maximal ideals $\mathfrak{m}$ of $S$
that $\mathfrak{m}M\neq M$. Let us also denote by $V(I)$ the Zariski closure
of ideal $I$, i.e. the set of ideals which include $I$.
###### Lemma 12.
It holds that $supp\>M=V(Ann(M))$.
###### Proof.
First we argue that $V(Ann(M))$ contains only maximal ideals. Since by Lemma 9
we have that $A=\nicefrac{{S}}{{Ann(M)}}$ is Artinian, by Lemma 10 all its
prime ideals are maximal. Taking preimage induces bijection between
$\operatorname{Spec}(\nicefrac{{S}}{{Ann(M)}})$ and $V(Ann(M))$ and this
bijection preserves maximality of an ideal. Thus all elements of $V(Ann(M))$
are maximal ideals.
Let now $\mathfrak{m}$ be a maximal ideal of $S$. We need to show that
$Ann(M)\not\subseteq\mathfrak{m}\iff\mathfrak{m}M=M$.
On the one hand, if $Ann(M)\not\subseteq\mathfrak{m}$, then
$\mathfrak{m}+Ann(M)=(1)$ and thus
$\mathfrak{m}M=\mathfrak{m}M+Ann(M)M=(\mathfrak{m}+Ann(M))M=(1)M=M.$
On the other hand, if $\mathfrak{m}M=M$, then by Nakayama’s lemma there exist
an element $m_{0}\in\mathfrak{m}$ such that $(1-m_{0})M=0$ and thus
$1-m_{0}\in Ann(M)$. Since $1-m_{0}\not\in\mathfrak{m}$, it holds that
$Ann(M)\not\subseteq\mathfrak{m}$. ∎
###### Lemma 13.
We have the inequality $|supp\>M|\leq\deg M$ and if the equality holds, then
$M$ is semisimple.
###### Proof.
By Lemma 12 the set $supp\>M$ can be treated as a subset of $SpecMax(A)$. By
Lemma 9 we have that $A$ is Artinian and thus by Lemma 10 we have that
$SpecMax(A)$ is finite, so $supp\>M$ is also finite.
Let
$supp\>M=\\{\mathfrak{m}_{1},\>\mathfrak{m}_{2},\>\ldots,\>\mathfrak{m}_{k}\\}$.
We have $\deg M=dim_{\mathbb{K}}M\geq
dim_{\mathbb{K}}\nicefrac{{M}}{{\mathfrak{m}_{1}M\cap\mathfrak{m}_{2}M\cap\ldots\cap\mathfrak{m}_{k}M}}$.
By the Chinese Remainder Theorem we have
$\nicefrac{{M}}{{\mathfrak{m}_{1}M\cap\mathfrak{m}_{2}M\cap\ldots\cap\mathfrak{m}_{k}M}}\simeq\nicefrac{{M}}{{\mathfrak{m}_{1}M}}\times\nicefrac{{M}}{{\mathfrak{m}_{2}M}}\times\ldots\times\nicefrac{{M}}{{\mathfrak{m}_{k}M}}.$
By definition of $\mathfrak{m}_{i}$ we have
$\nicefrac{{M}}{{\mathfrak{m}_{i}M}}\neq 0$ and thus
$dim_{\mathbb{K}}(\nicefrac{{M}}{{\mathfrak{m}_{i}M}})\geq 1$ and so
$dim_{\mathbb{K}}(\nicefrac{{M}}{{\mathfrak{m}_{1}M}}\times\nicefrac{{M}}{{\mathfrak{m}_{2}M}}\times\ldots\times\nicefrac{{M}}{{\mathfrak{m}_{k}M}})\geq
k.$
Thus $\deg
M\geq\deg\nicefrac{{M}}{{\mathfrak{m}_{1}M\cap\mathfrak{m}_{2}M\cap\ldots\cap\mathfrak{m}_{k}M}}\geq
k=|supp\>M|$, which proves the first part of the statement.
Let us now suppose that equality holds, i.e. $\deg M=k$. Thus all inequalities
from the first part must be equalities and hence
$\mathfrak{m}_{1}M\cap\mathfrak{m}_{2}M\cap\ldots\cap\mathfrak{m}_{k}M=0$ and
for all $i$ we have $dim_{\mathbb{K}}\nicefrac{{M}}{{\mathfrak{m}_{i}M}}=1$.
First equality and Chinese Remainder Theorem give us
$M\simeq\nicefrac{{M}}{{\mathfrak{m}_{1}M}}\times\nicefrac{{M}}{{\mathfrak{m}_{2}M}}\times\ldots\times\nicefrac{{M}}{{\mathfrak{m}_{k}M}}.$
Now by Lemma 6 it suffices to show that for all $i$ we have
$\nicefrac{{M}}{{\mathfrak{m}_{i}M}}\simeq\nicefrac{{S}}{{\mathfrak{m}_{i}}}$
as $S$-modules.
Since $\nicefrac{{M}}{{\mathfrak{m}_{i}M}}$ is a linear space of dimension
one, there exist $m\in\nicefrac{{M}}{{\mathfrak{m}_{i}M}}$ such that
$\mathbb{K}m=\nicefrac{{M}}{{\mathfrak{m}_{i}M}}$.
Let us define $S$-module homomorphism $\varphi\colon
S\rightarrow\nicefrac{{M}}{{\mathfrak{m}_{i}M}}$ by $\varphi(1)=m$,
$\varphi(s)=sm$. Since $\mathbb{K}m=\nicefrac{{M}}{{\mathfrak{m}_{i}M}}$, the
map $\varphi$ is surjective. Thus
$\nicefrac{{M}}{{\mathfrak{m}_{i}M}}\simeq\nicefrac{{S}}{{\ker(\varphi)}}$.
Clearly $\mathfrak{m}_{i}\subseteq\ker(\varphi)$ and thus
$\nicefrac{{S}}{{\ker(\varphi)}}\subseteq\nicefrac{{S}}{{\mathfrak{m}_{i}}}$.
However
$dim_{\mathbb{K}}\nicefrac{{M}}{{\mathfrak{m}_{i}M}}=1=dim_{\mathbb{K}}\nicefrac{{S}}{{\mathfrak{m}_{i}}}$
and so the inclusion must be an equality and thus indeed
$\nicefrac{{M}}{{\mathfrak{m}_{i}M}}\simeq\nicefrac{{S}}{{\mathfrak{m}_{i}}}$
as $S$-modules. ∎
###### Proposition 14.
If it holds that $\pi^{-1}(rad(A))\cap S_{\leq 1}=0$, then we have an equality
$|supp\>M|=\deg M$.
###### Proof.
By definition $\pi^{-1}(rad(A))=\sqrt{Ann(M)}$. The ideal $rad(A)$ is a
nilradical of $A$ so it is an intersection of its all prime ideals. Since by
Lemma 9 we have that $A$ is Artinian, by Lemma 10 all its prime ideals are
maximal and the number of maximal ideals is finite. Thus $rad(A)$ is an
intersection of finite number of maximal ideals and so $\pi^{-1}(rad(A))$ is
also intersection of finite number of maximal ideals.
By Lemma 13 we have that $|supp\>M|\leq\deg M$. Let us suppose by
contradiction that $|supp\>M|<\deg M$. By Lemma 12 we have that
$|supp\>M|=|V(Ann(M))|$. Let
$V(Ann(M))=\\{\mathfrak{m}_{1},\>\mathfrak{m}_{2},\ldots,\>\mathfrak{m}_{r}\\}$,
let us remark that by Lemma 12 ideals $\mathfrak{m}_{i}$ are maximal for all
$i$. Moreover
$\pi^{-1}(rad(A))=\sqrt{Ann(M)}=\bigcap_{p_{i}\in
V(Ann(M))}{p_{i}}=\mathfrak{m}_{1}\cap\mathfrak{m}_{2}\cap\ldots\cap\mathfrak{m}_{r}.$
By Hilbert’s Nullstellensatz maximal ideals in polynomial ring over
algebraically closed field can be treated as points, we will be using this
equivalence. In it, the elements of $\pi^{-1}(rad(A))\cap S_{\leq 1}$ are
exactly affine functions $f$ such that $f(\mathfrak{m}_{i})=0$ for all $i$.
We have $r=|V(Ann(M))|=|supp\>M|<\deg M=\deg S_{\leq 1}$. So the space of
affine functions has the dimension at least $r+1$. Thus the dimension of a
subspace of affine functions such that $f(\mathfrak{m}_{i})=0$ for all $i$ is
at least one. Thus there exist a non-zero function $f$ satisfying these
conditions. Since $f$ is an element of $\pi^{-1}(rad(A))\cap S_{\leq 1}$, we
obtain $\pi^{-1}(rad(A))\cap S_{\leq 1}\neq 0$, which gives us expected
contradiction. ∎
###### Proof of Theorem 7.
Let us suppose that $M$ is not semisimple. We need to prove that $\sigma_{M}$
is $0$-subtight-unstable.
By Lemma 13 we have that $|supp\>M|<\deg M$. Thus by Proposition 14 we have
$\pi^{-1}(rad(A))\cap S_{\leq 1}\neq 0$. Proposition 11 implies that
$\sigma_{M}$ is $0$-subtight-unstable. ∎
###### Theorem 15.
Let $t\in\mathbb{K}^{n}\otimes\mathbb{K}^{n}\otimes\mathbb{K}^{n}$ be a non-
zero $s$-subtight-unstable tensor with numbering $(a_{1},\>a_{2},\>a_{3})$ and
relative block format $(f_{1},\>f_{2},\>f_{3})$. Let also
$\overline{a}_{k}=\sum_{i\in I_{k}}f_{k}(i)a_{k}(i)$ for $k=1,\>2,\>3$. Then
it holds that
$\widetilde{SR}(t)\leq
n\exp\left(-\frac{(\overline{a}_{1}+\overline{a}_{2}+\overline{a}_{3}-s)^{2}}{6\sum_{k=1}^{3}\sum_{i_{k}\in
I_{k}}a_{k}^{2}}\right).$
###### Proof.
The proof is given in the proof of [3, Theorem 22]. Although in a statement
there is an assumption that tensors are over $\mathbb{C}$, there proof does
not use this assumption (it operates on supports of the tensors), which is
also noted in [3, Remark 26].
Note that the fraction is well defined, since if it held that $a_{k}\equiv 0$
for $k=1,\>2,\>3$, then $\overline{a}_{k}=0$. We have that
$\overline{a}_{1}+\overline{a}_{2}+\overline{a}_{3}>s$, so $s<0$. However
since $t$ is non-zero, its support is non-empty, and by definition for some
$(i_{1},\>i_{2},\>i_{3})$ it holds that
$0=a_{1}(i_{1})+a_{2}(i_{2})+a_{3}(i_{3})\leq s$. So we would have $s\geq
0>s$, which is a contradiction. ∎
Having Theorem 7 and Theorem 15 proved, we are almost ready to prove the main
theorem, using similar tools as Bläser and Lysikov. However we have to
slightly change the assumptions, so we will need to prove one more lemma.
Let us remind that for tensor $t\in V_{1}\otimes V_{2}\otimes V_{3}$ we denote
by $N(t)$ the maximal rank of flattening of $t$ (see Section 2).
###### Lemma 16.
We have an equality $N(\sigma_{M})=n$.
###### Proof.
Clearly $N(\sigma_{M})\leq n$, since $n=\dim_{\mathbb{K}}S_{\leq
1}=\dim_{\mathbb{K}}M$ and $\sigma_{M}\in S_{\leq 1}^{*}\otimes M^{*}\otimes
M$. Since $1\in S_{\leq 1}$, the matrix of $\sigma_{M}(1,-)$ is an identity
matrix and thus the ranks of flattenings of $\sigma_{M}$ over the second and
the third coordinate are at least $\dim_{\mathbb{K}}M=n$, so
$N(\sigma_{M})\geq n$, and thus the equality must hold. ∎
We now present more general result which implies Theorem 1. In particular, to
prove Theorem 1 we will use $0$-subtight-instability.
###### Proposition 17.
In $\mathbb{K}^{n\times n\times n}$ with $n$ fixed, it is impossible to prove
$\omega=2$ using arbitrary restrictions from powers of elements of tensor
family $G_{n}$ such that for every $t\in G_{n}$ tensor $t$ is $s$-subtight-
unstable and satisfies $N(t)=n$.
###### Proof.
We use the approach as in [3, Theorem 16], but we have slightly weaker
assumptions - we do not require conciseness.
Note that zero tensor does not belong to $G_{n}$, since $N(\mathbf{0})=0$.
If we let
$B(n)\coloneqq\inf\left\\{\frac{\log N(t)}{\log\widetilde{SR}(t)}\mid t\in
G_{n}\right\\},$
then by [3, Proposition 15] the irreversibility of any tensor with $N(t)\leq
n$ is bounded from below by $B(n)$, and by [8, Theorem 9] [3, Theorem 7] the
best bound on $\omega$ we can get is at least $2B(n)$.
By the assumption we have for all $t\in G_{n}$ that $N(t)=n$. Also by the
assumption all tensors from family $G_{n}$ are $s$-subtight-unstable, so by
Theorem 15 we have for all $t\in G_{n}$ the bound
$\widetilde{SR}(t)\leq
n\exp\left(-\frac{(\overline{a}_{1}(t)+\overline{a}_{2}(t)+\overline{a}_{3}(t)-s)^{2}}{6\sum_{k=1}^{3}\sum_{i_{k}\in
I_{k}}a_{k}(t)^{2}}\right).$
Let us denote
$\overline{B}(n)(t)=\frac{\log n}{\log
n-\frac{(\overline{a}_{1}(t)+\overline{a}_{2}(t)+\overline{a}_{3}(t)-s)^{2}}{6\sum_{k=1}^{3}\sum_{i_{k}\in
I_{k}}a_{k}(t)^{2}}}$
and
$\overline{B}(n)=\inf\left\\{\overline{B}(n)(t)\mid t\in G_{n}\right\\}.$
We can bound $B(n)$ from below by $\overline{B}(n)$. So it is sufficient to
show that $\overline{B}(n)>1$.
Clearly the term
$\frac{(\overline{a}_{1}(t)+\overline{a}_{2}(t)+\overline{a}_{3}(t)-s)^{2}}{6\sum_{k=1}^{3}\sum_{i_{k}\in
I_{k}}a_{k}(t)^{2}}$ is non-negative. Moreover, all tensors from the family
$G_{n}$ are $s$-subtight-unstable, so by definition the term
$\overline{a}_{1}(t)+\overline{a}_{2}(t)+\overline{a}_{3}(t)-s$ is strictly
positive. It implies that for $t\in G_{n}$ the value of $\overline{B}(n)(t)$
is strictly greater than one. Moreover, the value of $\overline{B}(n)(t)$
depends only on the support of the tensor $t$. Since for fixed $n$ the number
of possible supports of tensors is finite, the set of possible values of
$\overline{B}(n)(t)$ is also finite. Thus we have
$\overline{B}(n)=\inf\left\\{\overline{B}(n)(t)\mid t\in
G_{n}\right\\}=\min\left\\{\overline{B}(n)(t)\mid t\in G_{n}\right\\}>1,$
which ends the proof. ∎
###### Proof of Theorem 1.
By Theorem 7 all tensors $t\in F_{n}$ are $0$-subtight-unstable. By Lemma 16
it also holds that $N(t)=n$. Applying result of Proposition 17 for
$G_{n}=F_{n}$, we obtain the claim. ∎
Now we present the exact bound of irreversibility of the structure tensor of a
non-semisimple module. Note that we have already proved that this
irreversibility is strictly greater than one.
###### Corollary 18.
If the module $M$ is non-semisimple, then the irreversibility of its structure
tensor $\sigma_{M}$ is at least
$\left(1-\left(\frac{\left(\sum_{i=0}^{r-1}is_{i}\right)^{2}}{6n^{2}\log
n\left(2\sum_{i=0}^{r-1}m_{i}^{2}+\sum_{i=0}^{r-1}s_{i}\right)}\right)\right)^{-1},$
where $r$ is minimal such that $rad(A)^{r}=0$ ($r$ exists by Lemma 8),
$t_{i}=\dim_{\mathbb{K}}\pi^{-1}({rad(A)}^{i})\cap S_{\leq 1}$
$s_{i}=t_{i}-t_{i+1}$,
$m_{i}=\dim_{\mathbb{K}}\left({rad(A)}^{i}M/{rad(A)}^{i+1}M\right)$.
###### Proof.
As we observed, if $M$ is non-semisimple, then its structure tensor is
$0$-subtight unstable and satisfies $N(\sigma_{M})=n$, so it satisfies the
assumptions from Proposition 17, so we can use the bound
$i(\sigma_{M})\geq\frac{\log n}{\log
n-\frac{(\overline{a}_{1}(t)+\overline{a}_{2}(t)+\overline{a}_{3}(t))^{2}}{6\sum_{k=1}^{3}\sum_{i_{k}\in
I_{k}}a_{k}(t)^{2}}}.$
Now it is sufficient to use definitions of $a_{i}$ and $\overline{a}_{i}$ from
the proof of Proposition 11 and we obtain the desired bound. Analogical bound
for the structure tensors of non-semisimple algebras is given in [3, Corollary
23]. ∎
## 4 Characteristic zero
We will now show that in Proposition 17 we can weaken the assumption of
$s$-subtight-instability for fields of characteristic zero.
###### Proposition 19.
In $\mathbb{C}^{n\times n\times n}$ with $n$ fixed, it is impossible to prove
$\omega=2$ using arbitrary restrictions from powers of elements of tensor
family $G_{n}$ such that for every $t\in G_{n}$ tensor $t$ is unstable and
satisfies $N(t)=n$.
###### Proof.
From [3, Theorem 5] (originally in [6]) and [3, Theorem 4] (originally in
[14]) it follows that the set of possible values of $\widetilde{SR}(t)$ for
tensors in $\mathbb{C}^{n\times n\times n}$ is finite. Therefore, the set of
possible ratios $\frac{\log N(t)}{\log\widetilde{SR}(t)}$ is also finite. So
we can let
$B(n)\coloneqq\min\left\\{\frac{\log N(t)}{\log\widetilde{SR}(t)}\mid t\in
G_{n}\right\\},$
and by [3, Proposition 15] the irreversibility of any tensor with $N(t)\leq n$
is bounded from below by $B(n)$, and also by [8, Theorem 9] [3, Theorem 7] the
best bound on $\omega$ we can get is at least $2B(n)$.
Since all tensors form $G_{n}$ are unstable, we have by [3, Theorem 3] that
for all $t\in G_{n}$ we have $\widetilde{SR}(t)<n$. Since we assumed that for
all $t\in G_{n}$ we have $N(t)=n$, it holds that $2B(n)>2$, so it is
impossible to show $\omega=2$. ∎
###### Proposition 20.
In $\mathbb{K}^{n\times n\times n}$ with $n$ fixed with characteristic of
$\mathbb{K}$ equal to zero, it is impossible to prove $\omega=2$ using
arbitrary restrictions from powers of elements of tensor family $G_{n}$ such
that for every $t\in G_{n}$ tensor $t$ is combinatorially unstable and
satisfies $N(t)=n$.
###### Proof.
Let us suppose that it is possible to prove $\omega=2$ over $\mathbb{K}$ using
arbitrary restrictions. We will prove that then it is possible to prove
$\omega=2$ over $\mathbb{C}$ using arbitrary restrictions and obtain
contradiction with Proposition 19.
Possibility to prove $\omega=2$ over $\mathbb{K}$ using arbitrary restrictions
means that there exists a sequence of pairs $(t_{i},\>N_{i})$ such that for
all $i$ it holds that $t_{i}\in G_{n}$ and using arbitrary restrictions (which
are basically some linear equations) we can obtain from $t_{i}^{\otimes
N_{i}}$ large matrix multiplication, i.e. such that for a limit at infinity we
obtain $\omega=2$.
For a tensor $t_{i}$ let $(x_{i_{j}})$ be a (finite) tuple of vectors
corresponding to the flatenning of $t_{i}$ with maximal rank (each of three
flattenings is a linear map, we choose basis of tensors $x_{i_{j}}$ for which
contractions with $t_{i}$ give linear space of maximal dimension). Let also
$Z_{i}$ be a set of coefficients of arbitrary restrictions (which are
basically linear equations) used with $t_{i}$. All sets $Z_{i}$ are finite.
Now we want to transfer this sequence to $\mathbb{C}$. Firstly let us observe
that we only need to consider powers of tensors from sequence $t_{i}$, which
means that if $\mathbb{K}$ is large we can drop some part of it. More
precisely, let $\mathbb{\widetilde{L}}$ be a smallest subfield of $\mathbb{K}$
such that $\mathbb{\widetilde{L}}$ contains all elements of sets $Z_{i}$ and
tensors $t_{i}$ and $x_{i_{j}}$ (which means that we treat tensors as tuples
and want $\mathbb{\widetilde{L}}$ to contain all elements of corresponding
tuples). Let the field $\mathbb{L}$ be an algebraic closure of the field
$\mathbb{\widetilde{L}}$. Since $\mathbb{K}$ is algebraically closed,
$\mathbb{L}$ is a subfield of $\mathbb{K}$. The field $\mathbb{\widetilde{L}}$
is a countable field and thus $\mathbb{L}$ is also a countable field as an
algebraic closure of a countable field, so $\mathbb{L}$ is a countable
extension of $\mathbb{Q}$. By construction of $\mathbb{L}$ there exists an
injection from $\mathbb{L}$ to $\mathbb{K}$. This injection is a bijection on
tensors $t_{i}$, so it preserves the whole construction of obtaining large
matrix multiplications using arbitrary restrictions. Since the field
$\mathbb{L}$ is a countable extension of $\mathbb{Q}$, there exists an
injection from $\mathbb{L}$ to $\mathbb{C}$. Let us denote by $t_{i}^{\prime}$
the preimages of $t_{i}$ in injection from $\mathbb{L}$ to $\mathbb{K}$ and by
$\widetilde{t_{i}}$ the images of $t_{i}^{\prime}$ in injection from
$\mathbb{L}$ to $\mathbb{C}$. Now we can induce a construction of large matrix
multiplications in $\mathbb{C}$ using tensors $\widetilde{t_{i}}$ and
arbitrary restrictions.
We now want to obtain a contradiction with Proposition 19. To do so, we need
to prove that tensors $\widetilde{t_{i}}$ are unstable and
$N(\widetilde{t_{i}})=n$. By [3, Theorem 20] combinatorial instability implies
instability, so it is sufficient to show that tensors $\widetilde{t_{i}}$ are
unstable and $N(\widetilde{t_{i}})=n$. First we show that $t_{i}^{\prime}$ are
combinatorially unstable and $N(t_{i}^{\prime})=n$. The latter statement is
quite clear, since we have that $N(t_{i})=n$, which is obtained by
contractions with tensors $x_{i_{j}}$ (by definition) and because tensors
$x_{i_{j}}$ can be injected to $\mathbb{L}$, we have also
$N(t_{i}^{\prime})=n$. Being combinatorially unstable is also quite clear -
because we added to $\mathbb{L}$ tensors $t_{i}$, supports of these tensors
over $\mathbb{K}$ and $\mathbb{L}$ are the same sets. Since being
combinatorially unstable is a property of tensor support, and supports over
both fields are equal, being combinatorially unstable over $\mathbb{K}$
implies being combinatorially unstable over $\mathbb{L}$. Because $\mathbb{L}$
can be injected to $\mathbb{C}$, we can inject $x_{i_{j}}$ to $\mathbb{C}$, so
analogically $\widetilde{t_{i}}$ are combinatorially unstable. Also $t_{i}$
can be injected to $\mathbb{C}$, so the support of $\widetilde{t_{i}}$ is
equal to supports of $t_{i}$ and $t_{i}^{\prime}$, so we also have
$N(\widetilde{t_{i}})=n$. ∎
###### Remark 21.
It is well-known (a result due to Schönhage) that $\omega$ depends only on the
characteristic of the field, not on the field itself [7, Corollary 15.18].
However, in Proposition 19 and Proposition 20 we do not show that $\omega$ in
characteristic zero has some specific value, but that it cannot be proved that
$\omega=2$ using arbitrary restrictions from the class of starting tensors.
Thus, the generalisation from the case of $\mathbb{C}$ to the case of any
field of characteristic zero needed justification.
## References
* [1] Josh Alman and Virginia Vassilevska Williams “Further Limitations of the Known Approaches for Matrix Multiplication” In _9th Innovations in Theoretical Computer Science Conference (ITCS 2018)_ 94, Leibniz International Proceedings in Informatics (LIPIcs) Dagstuhl, Germany: Schloss Dagstuhl–Leibniz-Zentrum fuer Informatik, 2018, pp. 25:1–25:15 DOI: 10.4230/LIPIcs.ITCS.2018.25
* [2] Josh Alman and Virginia Vassilevska Williams “Limits on all known (and some unknown) approaches to matrix multiplication” In _59th Annual IEEE Symposium on Foundations of Computer Science—FOCS 2018_ IEEE Computer Soc., Los Alamitos, CA, 2018, pp. 580–591 DOI: 10.1109/FOCS.2018.00061
* [3] Markus Bläser and Vladimir Lysikov “Slice Rank of Block Tensors and Irreversibility of Structure Tensors of Algebras” In _45th International Symposium on Mathematical Foundations of Computer Science (MFCS 2020)_ 170, Leibniz International Proceedings in Informatics (LIPIcs) Dagstuhl, Germany: Schloss Dagstuhl–Leibniz-Zentrum für Informatik, 2020, pp. 17:1–17:15 DOI: 10.4230/LIPIcs.MFCS.2020.17
* [4] Jonah Blasiak et al. “On cap sets and the group-theoretic approach to matrix multiplication” In _Discrete Analysis_ Alliance of Diamond Open Access Journals, 2017 DOI: 10.19086/da.1245
* [5] Jonah Blasiak et al. “Which groups are amenable to proving exponent two for matrix multiplication?”, 2017 arXiv:1712.02302 [math.GR]
* [6] Michel Brion “Sur l’image de l’application moment” In _Séminaire d’Algèbre Paul Dubreil et Marie-Paule Malliavin: Proceedings, Paris 1986, volume 1296 of Lecture Notes in Mathematics_ Springer, Berlin, 1987, pp. pages 177–192 DOI: 10.1007/BFb0078520
* [7] Peter Bürgisser, Michael Clausen and Mohammad A Shokrollahi “Algebraic complexity theory” Springer Science & Business Media, 2013
* [8] Matthias Christandl, Péter Vrana and Jeroen Zuiddam “Barriers for fast matrix multiplication from irreversibility” In _34th Computational Complexity Conference_ 137, LIPIcs. Leibniz Int. Proc. Inform. Schloss Dagstuhl. Leibniz-Zent. Inform., Wadern, 2019, pp. Art. No. 2617
* [9] Don Coppersmith and Shmuel Winograd “Matrix multiplication via arithmetic progressions” In _J. Symbolic Comput._ 9.3, 1990, pp. 251–280 DOI: 10.1016/S0747-7171(08)80013-2
* [10] A.. Davie and A.. Stothers “Improved bound for complexity of matrix multiplication” In _Proceedings of the Royal Society of Edinburgh: Section A Mathematics_ 143.2 Royal Society of Edinburgh Scotland Foundation, 2013, pp. 351–369 DOI: 10.1017/S0308210511001648
* [11] Joachim Jelisiejew and Klemen Šivic “Components and singularities of Quot schemes and varieties of commuting matrices”, 2021 arXiv:2106.13137 [math.AG]
* [12] Joseph M Landsberg and Mateusz Michałek “Abelian tensors” In _Journal de Mathématiques Pures et Appliquées_ 108.3 Elsevier, 2017, pp. 333–371
* [13] François Le Gall “Powers of tensors and fast matrix multiplication” In _Proceedings of the 39th international symposium on symbolic and algebraic computation_ , 2014, pp. 296–303
* [14] Jeroen Zuiddam Matthias Christandl “Universal points in the asymptotic spectrum of tensors” In _arXiv preprint:1709.07851_ , 2017
* [15] Volker Strassen “Gaussian elimination is not optimal” In _Numerische mathematik_ 13.4 Springer, 1969, pp. 354–356
* [16] V.. Williams “Multiplying matrices faster than Coppersmith-Winograd” In _STOC ’12_ , 2012
|
* [129] David Barr “RFC 1912: Common DNS operational and configuration errors” In _The Pennsylvania State University, Pennsylvania_ , 1996
* [130] Jawad Ahmed, Hassan Habibi Gharakheili, Craig Russell and Vijay Sivaraman “Automatic Detection of DGA-Enabled Malware Using SDN and Traffic Behavioral Modeling” In _IEEE Transactions on Network Science and Engineering_ , 2022, pp. 1–1 DOI: 10.1109/TNSE.2022.3173591
* [131] Cybercrime Magazine “Global Cybercrime Damages Predicted To Reach $6 Trillion Annually By 2021” Accessed on 28.04.2020, 2020 URL: https://bit.ly/2YpyEZZ
* [132] Accenture Security “The Cost of Cybercrime” Accessed on 28.04.2020, 2019 URL: https://accntu.re/2WgX3hx
* [133] Security Magazine “As Cyber Attacks Become More Prevalent, Here?s Why Your Small Business is at Risk” Accessed on 28.04.2020, 2020 URL: https://bit.ly/2YpyEZZ
* [134] Accenture Security “Enhancing the Resilience of the Internet and Communications Ecosystem Against Botnets and Other Automated, Distributed Threats” Accessed on 20.04.2020, 2018 URL: https://bit.ly/2VTd925
* [135] Ars Technica “Failure to patch two-month-old bug led to massive Equifax breach” Accessed on 10.04.2020, 2017 URL: https://bit.ly/3aSJgmO
* [136] Sophos “Troj/Menti-Fam” Accessed on 25.02.2020, 2020 URL: https://bit.ly/38ttVrJ
* [137] Microsoft “TrojanDownloader:Win32/Murlo.S” Accessed on 25.02.2020, 2020 URL: https://bit.ly/2xeYZyt
* [138] Trend Micro “WORM:NERIS.A” Accessed on 25.02.2020, 2020 URL: https://bit.ly/2wxU8YK
* [139] Fortinet “NSIS.botnet” Accessed on 25.02.2020, 2020 URL: https://bit.ly/39uofPh
* [140] F Secure “Backdoor:W32/RBot” Accessed on 25.02.2020, 2020 URL: https://bit.ly/2VKtoPr
* [141] Microsoft “Program:W32/Sogou” Accessed on 25.02.2020, 2020 URL: https://bit.ly/38spEEE
* [142] F Secure “Virus:Win32/Virut” Accessed on 25.02.2020, 2020 URL: https://bit.ly/2TtmRqw
* [143] Sebastian Garcia “The CTU-13 Dataset” Accessed on 20.01.2020, 2020 URL: https://www.stratosphereips.org/datasets-ctu13
* [144] Roberto Perdisci “Behavioral Clustering of HTTP-Based Malware and Signature Generation Using Malicious Network Traces” In _Proc. USENIX NSDI_ , 2010
* [145] Battista Biggio “Poisoning Behavioral Malware Clustering” In _Proc. ACM AISec_ , 2014
* [146] Government Canada “Canadian Centre for Cyber Security” Accessed on 21.01.2020, 2020 URL: https://cyber.gc.ca/en/
* [147] Threat Intelligence Team “Andromeda under the microscope” Accessed on 21.01.2020, 2020 URL: https://bit.ly/32VPcJp
* [148] Andrew Griffin “Andromeda taken down” Accessed on 25.01.2020, 2020 URL: https://bit.ly/32SGITa
* [149] blueliv “Inside Tinba- DGA infection” Accessed on 9.02.2020, 2020 URL: https://bit.ly/39uDHLg
* [150] Heimdal Security “Everything You Need to Know about the Notorious Zeus Gameover Malware” Accessed on 9.02.2020, 2020 URL: https://bit.ly/39tmX7o
* [151] Douglas Bonderud “Combating High-Risk, Low-Noise Threats” Accessed on 29.01.2020, 2020 URL: https://ibm.co/38u8FBW
* [152] Basil AsSadhan “Periodic behavior in botnet command and control channels traffic” In _Proc. IEEE GLOBECOM_ , 2009, pp. 1–6
* [153] Virus Total “Virus Total” Accessed on 27.02.2020, 2020 URL: https://bit.ly/32SskdI
* [154] urlscan “X.co urlscan.io” Accessed on 27.02.2020, 2020 URL: https://bit.ly/38ocE3b
* [155] David A. Cieslak “Learning Decision Trees for Unbalanced Data” In _Proc. MLKDD_ , 2008, pp. 241–256
* [156] Arunan Sivanathan “Detecting Behavioral Change of IoT Devices using Clustering-Based Network Traffic Modeling” In _IEEE Internet of Things Journal_ 7.8, 2020, pp. 7295–7309
* [157] S. Hariri “Extended Isolation Forest” In _IEEE Transactions on Knowledge and Data Engineering_ , 2019, pp. 1–1
* [158] Chris Ding “K-means clustering via principal component analysis” In _Proc. ICML_ , 2004
* [159] Bernhard Scholkopf “Estimating the support of a high-dimensional distribution” In _Neural computation_ 13.7 MIT Press, 2001, pp. 1443–1471
* [160] David J Ketchen “The application of cluster analysis in strategic management research: an analysis and critique” In _Strategic management journal_ 17.6 Wiley Online Library, 1996, pp. 441–458
* [161] NoviFlow “NoviSwitch 2122 High Performance OpenFlow Switch” Accessed on 01.02.2020, 2020 URL: https://bit.ly/3cyhvSx
* [162] RYU Dev “Ryu SDN Framework” Accessed on 27.02.2020, 2020 URL: https://osrg.github.io/ryu/
* [163] Zeek “The Zeek Network Security Monitor” Accessed on 19.02.2020, 2020 URL: https://zeek.org
* [164] “Major cyber attack disrupts internet service across Europe and US” Accessed on 28.03.2021, 2016 URL: https://bit.ly/3fP7LHN
* [165] “FBI Outlines Technique Behind DDoS Attacks on US Voter Registration Website” Accessed on 31.03.2021, 2020 URL: https://bit.ly/3wC0II7
* [166] Xi Luo et al. “A large scale analysis of dns water torture attack” In _Proceedings of the 2018 2nd International Conference on Computer Science and Artificial Intelligence_ , 2018, pp. 168–173
|
# From Hyperbolic Geometry Back to Word Embeddings
Sultan Nurmukhamedov
Yandex School of Data Analysis
<EMAIL_ADDRESS>
&Thomas Mach
University of Potsdam
<EMAIL_ADDRESS>Arsen Sheverdin
University of Amsterdam
<EMAIL_ADDRESS>
&Zhenisbek Assylbekov
Nazarbayev University
<EMAIL_ADDRESS>
###### Abstract
We choose random points in the hyperbolic disc and claim that these points are
already word representations. However, it is yet to be uncovered which point
corresponds to which word of the human language of interest. This
correspondence can be approximately established using a pointwise mutual
information between words and recent alignment techniques.
From Hyperbolic Geometry Back to Word Embeddings
Sultan Nurmukhamedov Yandex School of Data Analysis<EMAIL_ADDRESS>Thomas
Mach University of Potsdam<EMAIL_ADDRESS>
Arsen Sheverdin University of Amsterdam<EMAIL_ADDRESS>Zhenisbek Assylbekov Nazarbayev University<EMAIL_ADDRESS>
## 1 Introduction
Vector representations of words are ubiquitous in modern natural language
processing (NLP). There are currently two large classes of word embedding
models: they build (1) static and (2) contextualized word vectors
correspondingly.
Static embeddings map each word type into a vector of real numbers, regardless
of the context in which the word type is used. The most prominent
representatives of this class of models are word2vec Mikolov et al. (2013b, a)
and GloVe Pennington et al. (2014). The obvious problem with this approach is
the representation of polysemous words, such as bank—it becomes unclear
whether we are talking about a financial institution, or we are talking about
the river bank.
Contextualized word embeddings, such as ELMo Peters et al. (2018) and BERT
Devlin et al. (2019), solve this problem by mapping each word token into a
vector space depending on the context in which the given word token is used,
i.e. the same word will have different vector representations when used in
different contexts. The second approach can nowadays be considered mainstream,
despite relatively few papers offering theoretical justifications for
contextualized word embeddings.
For static embeddings, on the contrary, there is a number of theoretical
works, each of which offers its own version of what is happening when word
vectors are trained. An incomplete list of such works includes those of Levy
and Goldberg (2014), Arora et al. (2016), Hashimoto et al. (2016), Gittens et
al. (2017), Tian et al. (2017), Ethayarajh et al. (2019), Allen et al. (2019),
Allen and Hospedales (2019), Assylbekov and Takhanov (2019), Zobnin and
Elistratova (2019). Other advantages of static embeddings over contextualized
ones include faster training (few hours instead of few days) and lower
computing requirements (1 consumer-level GPU instead of 8–16 non-consumer
GPUs). Morevoer, static embeddings are still an integral part of deep neural
network models that produce contextualized word vectors, because embedding
lookup matrices are used at the input and output (softmax) layers of such
models. Therefore, we consider it necessary to further study static
embeddings.
Several recent works Nickel and Kiela (2017); Tifrea et al. (2019) argue that
static word embeddings should be better trained in hyperbolic spaces than in
Euclidean spaces, and provide empirical evidence that word embeddings trained
in hyperbolic spaces need less dimensions to achieve the same quality as
state-of-the-art Euclidean vectors.111The quality of word vectors is usually
measured by the performance of downstream tasks, such as similarity,
analogies, part-of-speech tagging, etc. Usually such works motivate the
hyperbolicity of word embeddings by the fact that hyperbolic spaces are better
suited for embedding hierarchical structures. Words themselves often denote
concepts with an underlying hierarchy. An example of such a hierarchy is the
WordNet database, an excerpt of which is shown in Fig. 1.
carnivorefelinebig catliontigercatcaninedogwolffox Figure 1: An excerpt from
the WordNet database.
In the present paper we will investigate where the hyperbolicity originates
from. If we take the state-of-the-art Euclidean embeddings, is it possible to
establish a direct connection between them and their counterparts from a
hyperbolic word embedding? This was answered positively by Assylbekov and
Jangeldin (2020) who established a chain of connections: from word embeddings
to co-occurrence matrices, then to complex networks, and, finally, to
hyperbolic spaces. In this paper, to provide an additional justification for
the constructed chain, we propose a way to move from the final point,
hyperbolic spaces, to the initial one, word embeddings. We show that drawing
random points from the hyperbolic plane results in a set of points that
reasonably well resembles word embeddings. In fact, we can match these points
to word embeddings. Contrary, the same trick does not work with points drawn
at random in the Euclidean space. Thus, one can argue that the hyperbolic
space provides the underlying structure for word embeddings, while in the
Euclidean space this structure has to be superimposed.
### Notation
We denote with $\mathbb{R}$ the real numbers. Bold-faced lowercase letters
($\mathbf{x}$) denote vectors, plain-faced lowercase letters ($x$) denote
scalars, bold-faced uppercase letters ($\mathbf{A}$) denote matrices,
$\langle\mathbf{x},\mathbf{y}\rangle$ is the Euclidean inner product. We use
$\mathbf{A}_{a:b,c:d}$ to denote a submatrix located at the intersection of
rows $a,a+1,\ldots,b$ and columns $c,c+1,\ldots,d$ of $\mathbf{A}$. ‘i.i.d.’
stands for ‘independent and identically distributed’, ‘p.d.f’ stands for
‘probability distribution function’. We use the sign $\propto$ to abbreviate
‘proportional to’, and the sign $\sim$ to abbreviate ‘distributed as’.
Assuming that words have already been converted into indices, let
$\mathcal{W}:=\\{1,\ldots,n\\}$ be a finite vocabulary of words. Following the
setup of the widely used word2vec model Mikolov et al. (2013a, b), we use two
vectors per each word $i$: (1) $\mathbf{w}_{i}\in\mathbb{R}^{d}$ when
$i\in\mathcal{W}$ is a center word, (2) $\mathbf{c}_{i}\in\mathbb{R}^{d}$ when
$i\in\mathcal{W}$ is a context word; and we assume that $d\ll n$.
In what follows we assume that our dataset consists of co-occurence pairs
$(i,j)$. We say that “the words $i$ and $j$ co-occur” when they co-occur in a
fixed-size window of words. Let $\\#(i,j)$ be the number of times the words
$i$ and $j$ co-occur.
Figure 2: Random hyperbolic graph.
Figure 3: Distribution of PMI values (top) and of $R-X$.
## 2 Background: From Word Embeddings to Hyperbolic Space
Our departure point is the skip-gram with negative sampling (SGNS) word
embedding model of Mikolov et al. (2013b) that maximizes the following
objective function
$\sum_{i\in\mathcal{W}}\sum_{j\in\mathcal{W}}\\#(i,j)\log\sigma(\langle\mathbf{w}_{i},\mathbf{c}_{j}\rangle)\\\
+k\cdot\mathbb{E}_{j^{\prime}\sim
p}[\log\sigma(-\langle\mathbf{w}_{i},\mathbf{c}_{j^{\prime}}\rangle)],$ (1)
where $\sigma(x)=\frac{1}{1+e^{-x}}$ is the logistic sigmoid function, $p$ is
a smoothed unigram probability distribution for words,222The authors of SGNS
suggest $p(i)\propto\\#(i)^{3/4}$. and $k$ is the number of negative samples
to be drawn. Interestingly, training SGNS is approximately equivalent to
finding a low-rank approximation of a shifted pointwise mutual information
(PMI) matrix Levy and Goldberg (2014) in the form
$\log\frac{p(i,j)}{p(i)p(j)}-\log
k\approx\langle\mathbf{w}_{i},\mathbf{c}_{j}\rangle,$ (2)
where the left-hand side is the shifted PMI between $i$ and $j$, and the
right-hand side is an $ij$-th element of a matrix with rank $\leq d$ since
$\mathbf{w}_{i},\mathbf{c}_{j}\in\mathbb{R}^{d}$. This approximation was later
re-derived by Arora et al. (2016), Zobnin and Elistratova (2019), Assylbekov
and Takhanov (2019), and Allen et al. (2019) under different sets of
assuptions. In a recent paper, Assylbekov and Jangeldin (2020) showed that the
removal of the sigmoid transformation in the SGNS objective (1) gives word
embeddings comparable in quality with the original SGNS embeddings. A
maximization of such modified objective results in a low-rank approximation of
a squashed shifted PMI ($\sigma$SPMI) matrix, defined as
$\mathbf{A}_{ij}:=\sigma\left(\log\frac{p(i,j)}{p(i)p(j)}-\log k\right).$ (3)
Moreover, treating the $\sigma$SPMI matrix as a connection probabilities
matrix of a random graph, the authors show that such graph is a complex
network, that is it has strong clustering and scale-free degree distribution,
and according to Krioukov et al. (2010), such graph possesses an effective
hyperbolic geometry underneath. The following chain summarizes this argument:
$\boxed{\text{Word
Embeddings}}\quad\longrightarrow\quad\boxed{\sigma\text{SPMI}}\quad\longrightarrow\\\
\boxed{\text{Complex Network}}\quad\longrightarrow\quad\boxed{\text{Hyperbolic
Space}}$ (4)
In our work, we go from the final point (hyperbolic space) to the starting one
(word embeddings), and the next section provides the details of our method.
Method | Word Similarity | POS Tagging
---|---|---
WS353 | Men | M. Turk | CoNLL-2000 | Brown
SGNS | .678 | .656 | .690 | 90.77 | 92.60
PMI + SVD | .669 | .674 | .666 | 92.25 | 93.76
$\sigma$SPMI + SVD | .648 | .622 | .666 | 92.76 | 93.78
RHG + SVD + Align | .406 | .399 | .509 | 92.23 | 93.19
Random + Align | .165 | .117 | .111 | 81.89 | 89.39
Table 1: Evaluation of word embeddings on the similarity and POS tagging
tasks. For the similarity tasks the evaluation metric is the Spearman’s
correlation with human ratings, for the POS tagging tasks it is accuracy.
Random stands for random vectors that were obtained as i.i.d. draws from
$\mathcal{N}(\mathbf{0},\mathbf{I})$.
## 3 Method: From Hyperbolic Geometry to Word Embeddings
It is difficult to visualize hyperbolic spaces because they cannot be
isometrically embedded into any Euclidean space.333This means that we cannot
map points of a hyperbolic space into points of a Euclidean space in such way
that the distances between points are preserved. However, there exist models
of hyperbolic spaces: each model emphasizes different aspects of hyperbolic
geometry, but no model simultaneously represents all of its properties. We
will consider here the so-called native model Krioukov et al. (2010), in which
the hyperbolic plane $\mathbb{H}^{2}$ is represented by a disk of radius $R$,
and we use polar coordinates $(r,\theta)$ to specify the position of any point
$v\in\mathbb{H}^{2}$, where the radial coordinate $r$ equals the hyperbolic
distance of $v$ from the origin. Given this notation, the distance $x$ between
two points with coordinates $(r,\theta)$ and $(r^{\prime},\theta^{\prime})$
satisfies the hyperbolic law of cosines
$\textstyle\cosh x=\cosh r\cosh r^{\prime}\\\ -\sinh r\sinh
r^{\prime}\cos(\theta-\theta^{\prime}),$ (5)
for the hyperbolic space of constant curvature $-1$.444Defining constant
curvature is beyond the scope of our paper. We just mention here that there
are only three types of isotropic spaces: Euclidean (zero curvature),
spherical (positively curved), and hyperbolic (negatively curved). A key
property of hyperbolic spaces is that they expand faster than Euclidean
spaces. E.g., a circle with radius $r$ has in the Euclidean plane a length of
$2\pi r=\Theta(r)$ and an area of $\pi r^{2}=\Theta(r^{2})$, while its length
and area in the hyperbolic plane are $2\pi\sinh(r)=\Theta(e^{r})$ and
$2\pi(\cosh r-1)=\Theta(e^{r})$ correspondingly. It is noteworthy that in a
balanced tree with branching factor $b$, the number of nodes that are $r$
edges from the root grows as $\Theta(b^{r})$, i.e. exponentially with $r$,
leading to the suggestion that hierarchical complex networks with tree-like
structures might be easily embeddable in hyperbolic space.
Based on the above facts, we construct a random hyperbolic (RHG) graph as in
the work of Krioukov et al. (2010): we place randomly $n$ points (nodes) into
a hyperbolic disk of radius $R$, and each pair of nodes $(i,j)$ is connected
with probability $\sigma(R-x_{ij})$, where $x_{ij}$ is the hyperbolic distance
(5) between points $i$ and $j$. Angular coordinates of the nodes are sampled
from the uniform distribution: $\theta\sim\mathcal{U}[0,2\pi]$, while the
radial coordinates are sampled from the exponential p.d.f.
$\rho(r)=\frac{\alpha\sinh\alpha r}{\cosh\alpha R-1}=\Theta(e^{\alpha r}).$
The hyperparameters $R$ and $\alpha$ are chosen based on the total number of
nodes $n$, the desired average degree $\bar{k}$ and the power-law exponent
$\gamma$ according to the equations (22) and (29) of Krioukov et al. (2010).
An example of such RHG is shown in Figure 3. Notice, that the connection
probabilities matrix of our graph is
$\mathbf{B}_{ij}:=\sigma(R-x_{ij}),$
Comparing this to (3), we see that if $\mathbf{A}$ and $\mathbf{B}$ induce
structurally similar graphs then the distribution of the PMI values
$\log\frac{p(i,j)}{p(i)p(j)}$ should be similar to the distribution of
$R-x_{ij}$ values (up to a constant shift). To test this empirically, we
compute a PMI matrix of a well-known corpus, text8, and compare the
distribution of the PMI values with the p.d.f. of $R-X$, where $X$ is a
distance between two random points of a hyperbolic disk (the exact form of
this p.d.f. is given in Proposition A.1). The results are shown in Figure 3.
As we can see, the two distributions are similar in the sense that both are
unimodal and right-skewed. The main difference is in the shift—distribution of
$R-X$ is shifted to the left compared to the distribution of the PMI values.
We hypothesize that the nodes of the RHG treated as points of the hyperbolic
space are already reasonable word embeddings for the words of our vocabulary
$\mathcal{W}$. The only thing that we do not know is the correspondence
between words $i\in\mathcal{W}$ and nodes of the RHG. Instead of aligning
words with nodes, we can align their vector representations. For this, we take
singular value decompositions (SVD) of $\mathbf{A}$ and $\mathbf{B}$:
$\mathbf{A}=\mathbf{U}_{A}\boldsymbol{\Sigma}_{A}\mathbf{V}_{A}^{\top},\quad\mathbf{B}=\mathbf{U}_{B}\boldsymbol{\Sigma}_{B}\mathbf{V}_{B}^{\top},$
and then obtain embedding matrices by
$\displaystyle\mathbf{W}_{A}$
$\displaystyle:=\mathbf{U}_{A,1:n,1:d}\boldsymbol{\Sigma}^{1/2}_{A,1:d,1:d}\in\mathbb{R}^{n\times
d}$ $\displaystyle\mathbf{W}_{B}$
$\displaystyle:=\mathbf{U}_{B,1:n,1:d}\boldsymbol{\Sigma}^{1/2}_{B,1:d,1:d}\in\mathbb{R}^{n\times
d}$
as in the work of Levy and Goldberg (2014). An $i^{\text{th}}$ row in
$\mathbf{W}_{A}$ is an embedding of the word $i\in\mathcal{W}$, while an
$i^{\text{th}}$ row in $\mathbf{W}_{B}$ is an embedding of the RHG’s node $i$.
To align these two sets of embeddings we apply a recent stochastic
optimization method of Grave et al. (2019) that solves
$\min_{\mathbf{Q}\in\mathcal{O}_{d}}\min_{\mathbf{P}\in\mathcal{P}_{n}}\|\mathbf{W}_{A}\mathbf{Q}-\mathbf{P}\mathbf{W}_{B}\|^{2}_{2},$
where $\mathcal{O}_{d}$ is the set of $d\times d$ orthogonal matrices and
$\mathcal{P}_{d}$ is the set of $n\times n$ permutation matrices. As one can
see, this method assumes that alignment between two sets of embeddings is not
only a permutation from one set to the other, but also an orthogonal
transformation between the two. Once the alignment is done, we treat
$\mathbf{PW}_{B}$ as an embedding matrix for the words in $\mathcal{W}$.
## 4 Evaluation
In this section we evaluate the quality of word vectors resulting from a
RHG555Our code is available at https://github.com/soltustik/RHG against those
from the SGNS, PMI, and $\sigma$SPMI. We use the text8 corpus mentioned in the
previous section. We were ignoring words that appeared less than 5 times
(resulting in a vocabulary of 71,290 tokens). We set window size to 2,
subsampling threshold to $10^{-5}$, and dimensionality of word vectors to 200.
The SGNS embeddings were trained using our custom
implementation.666https://github.com/zh3nis/SGNS The PMI and BPMI matrices
were extracted using the hyperwords tool of Levy et al. (2015) and SVD was
performed using the PyTorch library of Paszke et al. (2019).
The embeddings were evaluated on word similarity and POS tagging tasks. For
word similarity we used WordSim (Finkelstein et al., 2002), MEN (Bruni et al.,
2012), and M.Turk (Radinsky et al., 2011) datasets. For POS tagging we trained
a simple classifier777feedforward neural network with one hidden layer and
softmax output layer by feeding in the embedding of a current word and its
nearby context to predict its part-of-speech (POS) tag:
$\widehat{\mathrm{POS}}_{t}=\operatorname{softmax}(\sigma(\mathbf{A}[\mathbf{w}_{t-2};\ldots;\mathbf{w}_{t+2}]+\mathbf{b}))$
where $[\mathbf{x};\mathbf{y}]$ is concatenation of $\mathbf{x}$ and
$\mathbf{y}$. The classifier was trained on CoNLL-2000 Tjong Kim Sang and
Buchholz (2000) and Brown Kucera et al. (1967) datasets.
The results of evaluation are provided in Table 1. As we see, vector
representations of words generated from a RHG lag behind in word similarity
tasks from word vectors obtained by other standard methods. Note, however,
that the similarity task was designed with Euclidean geometry in mind. Even
though our RHG-based vectors are also ultimately placed in the Euclidean space
(otherwise the alignment step would not have been possible), their nature is
inherently non-Euclidean. Therefore, the similarity scores for them may not be
indicative. So, for example, when RHG vectors are fed into a nonlinear model
for POS tagging, they are comparable with other types of vectors.
We notice that random vectors—generated as i.i.d. draws from
$\mathcal{N}(\mathbf{0},\mathbf{I})$ and then aligned to the embeddings from
$\sigma$SPMI—show poor results in the similarity tasks and underperform all
other word embedding methods in the POS tagging tasks. This calls into
question whether multivariate Gaussian is a reasonable (prior) distribution
for word vectors as was suggested by Arora et al. (2016), Assylbekov and
Takhanov (2019).
## 5 Conclusion and Future Work
In this work we show that word vectors can be obtained from hyperbolic
geometry without explicit training. We obtain the embeddings by randomly
drawing points in the hyperbolic plane and by finding correspondence between
these points and the words of the human language. This correspondence is
determined by the relation (hyperbolic distance) to other words. This method
avoids the, often expensive, training of word vectors in hyperbolic spaces as
in Tifrea et al. (2019). A direct comparison is not what this paper
attempts—our method is cheaper but produces word vectors of lower quality. Our
method simply shows that word vectors do fit better into hyperbolic space than
into Euclidean space.
Finally, we want to sketch a possible direction for future work. The
hyperbolic space is a special case of a Riemannian manifold. Are Riemannian
manifolds better suited for word vectors? In particular which manifolds should
one use? At the moment, there is only limited empirical knowledge to address
these questions. For instance, Gu et al. (2019) obtained word vectors of
better quality, according to the similarity score, in the product of
hyperbolic spaces, which is still a Riemannian manifold but not a hyperbolic
space anymore. We are hopeful that future work may provide an explanation for
this empirical fact.
## Acknowledgements
Zhenisbek Assylbekov was supported by the Program of Targeted Funding “Economy
of the Future” #0054/ ПЦФ-НС-19. The work of Sultan Nurmukhamedov was
supported by the Nazarbayev University Faculty-Development Competitive
Research Grants Program, grant number 240919FD3921. The authors would like to
thank anonymous reviewers for their feedback.
## References
* Allen et al. (2019) Carl Allen, Ivana Balazevic, and Timothy Hospedales. 2019. What the vec? towards probabilistically grounded embeddings. In _Proceedings of NeurIPS_.
* Allen and Hospedales (2019) Carl Allen and Timothy Hospedales. 2019. Analogies explained: Towards understanding word embeddings. In _Proceedings of ICML_.
* Arora et al. (2016) Sanjeev Arora, Yuanzhi Li, Yingyu Liang, Tengyu Ma, and Andrej Risteski. 2016. A latent variable model approach to pmi-based word embeddings. _Transactions of the Association for Computational Linguistics_ , 4:385–399.
* Assylbekov and Jangeldin (2020) Zhenisbek Assylbekov and Alibi Jangeldin. 2020. Squashed shifted pmi matrix: bridging word embeddings and hyperbolic spaces. In _Proceedings of AJCAI_.
* Assylbekov and Takhanov (2019) Zhenisbek Assylbekov and Rustem Takhanov. 2019. Context vectors are reflections of word vectors in half the dimensions. _Journal of Artificial Intelligence Research_ , 66:225–242.
* Bruni et al. (2012) Elia Bruni, Gemma Boleda, Marco Baroni, and Nam-Khanh Tran. 2012. Distributional semantics in technicolor. In _Proceedings of ACL_.
* Devlin et al. (2019) Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. Bert: Pre-training of deep bidirectional transformers for language understanding. In _Proceedings of NAACL-HLT_.
* Ethayarajh et al. (2019) Kawin Ethayarajh, David Duvenaud, and Graeme Hirst. 2019. Towards understanding linear word analogies. In _Proceedings of ACL_.
* Finkelstein et al. (2002) Lev Finkelstein, Evgeniy Gabrilovich, Yossi Matias, Ehud Rivlin, Zach Solan, Gadi Wolfman, and Eytan Ruppin. 2002. Placing search in context: The concept revisited. _ACM Transactions on information systems_ , 20(1):116–131.
* Gittens et al. (2017) Alex Gittens, Dimitris Achlioptas, and Michael W Mahoney. 2017. Skip-gram- zipf+ uniform= vector additivity. In _Proceedings of ACL_ , pages 69–76.
* Grave et al. (2019) Edouard Grave, Armand Joulin, and Quentin Berthet. 2019. Unsupervised alignment of embeddings with wasserstein procrustes. In _Proceedings of AISTATS_.
* Gu et al. (2019) Albert Gu, Frederic Sala, Beliz Gunel, and Christopher Ré. 2019. Learning mixed-curvature representations in product spaces. In _7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, May 6-9, 2019_. OpenReview.net.
* Hashimoto et al. (2016) Tatsunori B Hashimoto, David Alvarez-Melis, and Tommi S Jaakkola. 2016. Word embeddings as metric recovery in semantic spaces. _Transactions of the Association for Computational Linguistics_ , 4:273–286.
* Krioukov et al. (2010) Dmitri Krioukov, Fragkiskos Papadopoulos, Maksim Kitsak, Amin Vahdat, and Marián Boguná. 2010. Hyperbolic geometry of complex networks. _Physical Review E_ , 82(3):036106.
* Kucera et al. (1967) Henry Kucera, Henry Kučera, and Winthrop Nelson Francis. 1967. _Computational analysis of present-day American English_. Brown university press.
* Levy and Goldberg (2014) Omer Levy and Yoav Goldberg. 2014. Neural word embedding as implicit matrix factorization. In _Proceedings of NeurIPS_.
* Levy et al. (2015) Omer Levy, Yoav Goldberg, and Ido Dagan. 2015. Improving distributional similarity with lessons learned from word embeddings. _Transactions of the Association for Computational Linguistics_ , 3:211–225.
* Mikolov et al. (2013a) Tomas Mikolov, Kai Chen, Greg Corrado, and Jeffrey Dean. 2013a. Efficient estimation of word representations in vector space. _arXiv preprint arXiv:1301.3781_.
* Mikolov et al. (2013b) Tomas Mikolov, Ilya Sutskever, Kai Chen, Greg S Corrado, and Jeff Dean. 2013b. Distributed representations of words and phrases and their compositionality. In _Proceedings of NeurIPS_.
* Nickel and Kiela (2017) Maximillian Nickel and Douwe Kiela. 2017. Poincaré embeddings for learning hierarchical representations. In _Advances in neural information processing systems_ , pages 6338–6347.
* Paszke et al. (2019) Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, et al. 2019\. Pytorch: An imperative style, high-performance deep learning library. In _Proceedings of NeurIPS_.
* Pennington et al. (2014) Jeffrey Pennington, Richard Socher, and Christopher Manning. 2014. Glove: Global vectors for word representation. In _Proceedings of EMNLP_.
* Peters et al. (2018) Matthew E Peters, Mark Neumann, Mohit Iyyer, Matt Gardner, Christopher Clark, Kenton Lee, and Luke Zettlemoyer. 2018. Deep contextualized word representations. In _Proceedings of NAACL-HLT_.
* Radinsky et al. (2011) Kira Radinsky, Eugene Agichtein, Evgeniy Gabrilovich, and Shaul Markovitch. 2011\. A word at a time: computing word relatedness using temporal semantic analysis. In _Proceedings of the 20th international conference on World wide web_ , pages 337–346. ACM.
* Tian et al. (2017) Ran Tian, Naoaki Okazaki, and Kentaro Inui. 2017. The mechanism of additive composition. _Machine Learning_ , 106(7):1083–1130.
* Tifrea et al. (2019) Alexandru Tifrea, Gary Bécigneul, and Octavian-Eugen Ganea. 2019. Poincaré glove: Hyperbolic word embeddings. In _Proceedings of ICLR_.
* Tjong Kim Sang and Buchholz (2000) Erik F. Tjong Kim Sang and Sabine Buchholz. 2000. Introduction to the CoNLL-2000 shared task chunking. In _Fourth Conference on Computational Natural Language Learning and the Second Learning Language in Logic Workshop_.
* Zobnin and Elistratova (2019) Alexey Zobnin and Evgenia Elistratova. 2019. Learning word embeddings without context vectors. In _Proceedings of the 4th Workshop on Representation Learning for NLP (RepL4NLP-2019)_ , pages 244–249.
## Appendix A Auxiliary Results
###### Proposition A.1.
Let $X$ be a distance between two points that were randomly uniformly placed
in the hyperbolic disk of radius $R$. The probability distribution function of
$X$ is given by
$f_{X}(x)=\int_{0}^{R}\int_{0}^{R}\\\
\frac{\sinh(x)\rho(r_{1})\rho(r_{2})dr_{1}dr_{2}}{\pi\sqrt{1-A(r_{1},r_{2},x)}\sinh(r_{1})\sinh(r_{2})},$
(6)
where
$A(r_{1},r_{2},x)=\frac{\cosh(r_{1})\cosh(r_{2})-\cosh(x)}{\sinh(r_{1})\sinh(r_{2})}$,
and $\rho(r)=\frac{\alpha\sinh\alpha r}{\cosh\alpha R-1}$.
###### Proof.
Let us throw randomly and uniformly two points $(r_{1},\theta_{1})$ and
$(r_{2},\theta_{2})$ into the hyperbolic disk of radius $R$, i.e.
$r_{1},r_{2}\,\,{\stackrel{{\scriptstyle\text{i.i.d.}}}{{\sim}}}\,\,\rho(r)$,
$\theta_{1},\theta_{2}\,\,{\stackrel{{\scriptstyle\text{i.i.d.}}}{{\sim}}}\,\,\text{Uniform}[0,2\pi)$.
Let $X$ be the distance between these points ($X$ is a random variable). Let
$\gamma$ be the angle between these points, then
$\gamma:=\pi-|\pi-|\theta_{1}-\theta_{2}||\sim\text{Uniform}[0,\pi)$ and thus
$f_{\cos\gamma}(t)=\frac{1}{\pi\sqrt{1-t^{2}}},\quad t\in[-1,1].$
Since the distance in our model of hyperbolic plane is given by
$X=\cosh^{-1}[\cosh r_{1}\cosh r_{2}-\sinh r_{1}\sinh r_{2}\cos\gamma]$
we have
$\displaystyle\Pr(X\leq x)$
$\displaystyle=\Pr\left(\cos\gamma\geq\underbrace{\frac{\cosh r_{1}\cosh
r_{2}-\cosh x}{\sinh r_{1}\sinh r_{2}}}_{A(r_{1},r_{2},x)}\right)$
$\displaystyle=\Pr(\cos\gamma\geq A(r_{1},r_{2},x))$
$\displaystyle=\int_{A(r_{1},r_{2},x)}^{+\infty}\frac{1}{\pi\sqrt{1-t^{2}}}$
$\displaystyle=\frac{1}{2}-\frac{\sin^{-1}A(r_{1},r_{2},x)}{\pi},$
and therefore
$f_{X\mid
r_{1},r_{2}}(x)=\frac{d}{dx}\left[\frac{1}{2}-\frac{\sin^{-1}A(r_{1},r_{2},x)}{\pi}\right]\\\
=\frac{\sinh x}{\pi\sqrt{1-A(r_{1},r_{2},x)}\sinh(r_{1})\sinh r_{2}}$
for $x\in(|r_{1}-r_{2}|,r_{1}+r_{2})$. Integrating $f_{X\mid
r_{1},r_{2}}(x)\rho(r_{1})\rho(r_{2})$ with respect to $r_{1}$ and $r_{2}$ we
get (6). ∎
|
# Unravelling the structure of magnetised molecular clouds with SILCC-Zoom:
sheets, filaments and fragmentation
Shashwata Ganguly,1 S. Walch,1,2 D. Seifried,1,2 S. D. Clarke1,3 and M. Weis1
1I. Physikalisches Insitut, Universität zu Köln, Zülpicher Str. 77, 50937
Köln, Germany
2Cologne Centre for Data and Simulation Science, University of Cologne,
Cologne, Germany
3Academia Sinica, Institute of Astronomy and Astrophysics, Taipei, Taiwan
E-mail<EMAIL_ADDRESS>
(Accepted XXX. Received YYY; in original form ZZZ)
###### Abstract
To what extent magnetic fields affect how molecular clouds (MCs) fragment and
create dense structures is an open question. We present a numerical study of
cloud fragmentation using the SILCC-Zoom simulations. These simulations follow
the self-consistent formation of MCs in a few hundred parsec sized region of a
stratified galactic disc; and include magnetic fields, self-gravity,
supernova-driven turbulence, as well as a non-equilibrium chemical network. To
discern the role of magnetic fields in the evolution of MCs, we study seven
simulated clouds, five with magnetic fields, and two without, with a maximum
resolution of 0.1 parsec. Using a dendrogram we identify hierarchical
structures which form within the clouds. Overall, the magnetised clouds have
more mass in a diffuse envelope with a number density between 1-100 cm-3. We
find that six out of seven clouds are sheet-like on the largest scales, as
also found in recent observations, and with filamentary structures embedded
within, consistent with the bubble-driven MC formation mechanism. Hydrodynamic
simulations tend to produce more sheet-like structures also on smaller scales,
while the presence of magnetic fields promotes filament formation. Analysing
cloud energetics, we find that magnetic fields are dynamically important for
less dense, mostly but not exclusively atomic structures (typically up to
$\sim 100-1000$ cm-3), while the denser, potentially star-forming structures
are energetically dominated by self-gravity and turbulence. In addition, we
compute the magnetic surface term and demonstrate that it is generally
confining, and some atomic structures are even magnetically held together. In
general, magnetic fields delay the cloud evolution and fragmentation by $\sim$
1 Myr.
###### keywords:
MHD – methods: numerical – stars: formation – ISM: clouds – ISM: kinematics
and dynamics
††pubyear: 2022††pagerange: Unravelling the structure of magnetised molecular
clouds with SILCC-Zoom: sheets, filaments and fragmentation–F
## 1 Introduction
Magnetic fields are ubiquitous in the interstellar medium (ISM, Crutcher et
al., 2003; Heiles & Troland, 2005; Fletcher et al., 2011; Beck, 2015). Since
the discovery of interstellar magnetic fields by Hiltner (1951) and Hall
(1951), they have been known to be integral to the dynamical evolution of the
ISM. Magnetic fields, however, are also notoriously difficult to measure
accurately and model theoretically. Decades of dedicated observations have
resulted in a good understanding of the morphology and strength of the
magnetic field in different ISM phases (Crutcher, 1999; Bourke et al., 2001;
Heiles & Crutcher, 2005; Troland & Crutcher, 2008; Crutcher, 2012; Beck, 2015;
Planck Collaboration et al., 2020; Lopez-Rodriguez et al., 2023).
However, the exact nature of how magnetic fields affect molecular cloud (MC)
formation and evolution is an open question and subject of intense scrutiny
(see e.g. reviews by Crutcher, 2012; Hennebelle & Inutsuka, 2019; Girichidis
et al., 2020; Pattle et al., 2022). Various numerical studies have performed
detailed analysis on the interplay of magnetic fields with other physical
processes (e.g. turbulence, thermal pressure) in order to determine how MCs
are shaped, formed, and how they evolve (e.g. Heitsch et al., 2001; Federrath
& Klessen, 2012; Walch et al., 2015; Körtgen & Banerjee, 2015; Girichidis et
al., 2016b; Körtgen et al., 2018; Seifried et al., 2019; Ibáñez-Mejía et al.,
2022).
On galactic scales, ordered magnetic fields have been observed, with a
correlation between the direction of the spiral arms and the magnetic field
(Beck, 2009; Fletcher et al., 2011; Li & Henning, 2011). In the diffuse ISM,
the magnetic field strength, $B$, does not show any correlation with the
density for number densities of up to roughly 300 cm-3 (Crutcher et al.,
2010). Above these densities, Crutcher et al. (2010) find
$B\propto\rho^{\kappa}$, with $\kappa\approx 2/3$, consistent with sub-
dominant magnetic field strengths, although there remains considerable scatter
in the observations.
The lack of correlation between the strength of the magnetic field and the
density of the ambient medium implies that in the diffuse ISM, magnetic fields
can channelise gas flows along the field lines and therefore influence the
environment in which MCs form. Pardi et al. (2017) show that magnetic fields
are more likely to cause a smoother gas distribution, while Molina et al.
(2012) find that they are more likely to affect the dynamics of lower-density
gas. Magnetic fields can add to the thermal pressure exerted by the gas and
slow down the formation of dense gas (Hill et al., 2012), as well as molecular
gas (Girichidis et al., 2018; Seifried et al., 2020a). A sufficiently strong
magnetic field can prevent the collapse of a MC altogether (Mouschovias, 1991;
Spitzer, 1978) or slow down cloud evolution (Heitsch et al., 2001; Padoan &
Nordlund, 2011; Federrath & Klessen, 2012; Ibáñez-Mejía et al., 2022).
In terms of morphology, they can facilitate the formation of elongated
filamentary structures (Hacar et al., 2022; Pineda et al., 2022) and are
essential in understanding the filamentary nature of the ISM (see e.g. Bally
et al., 1987; André et al., 2014). The direction of such elongation relative
to the direction of the magnetic field is a matter of great active research
(e.g. Soler & Hennebelle, 2017; Seifried et al., 2020b). In the lower density
range, for sub-Alfvénic gas, anisotropic turbulence can lead to structures
elongated parallel to field lines. In contrast, at higher densities, magnetic
fields can channelise flows along field lines and therefore facilitate
structures perpendicular to the field direction.
Magnetic fields are likely to also affect the fragmentation of clouds and
cloud cores. Commerçon et al. (2011) find that fragments in magnetized cloud
cores are more massive compared to those formed without magnetic fields.
Although the probability density function (PDF) of lower density gas is found
to be different in the presence of magnetic fields (Molina et al., 2012), the
high density, potentially star-forming part does not seem to significantly
affected (Klessen & Burkert, 2001; Slyz et al., 2005; Girichidis et al., 2014;
Schneider et al., 2015).
In this work, we perform a numerical investigation of the role that magnetic
fields play in the formation and shaping of density structures within MCs. We
do a detailed analysis of realistic MC simulations based on the SILCC-Zoom
simulations (Seifried et al., 2017) by comparing the morphological, dynamical,
and fragmentation properties in seven simulated clouds, five with magnetic
fields (magnetohydrodynamic or MHD clouds) and two without (hydrodynamic or HD
clouds).
The paper is structured as follows: In Section 2, we outline the numerical
setup of the simulation. Section 3 discusses the procedure for identifying and
classifying structures (Ganguly et al., 2022). We highlight the differences
density PDFs between HD and MHD clouds in Section 4. The morphological
properties of the obtained structures are presented in Section 5. We find all
the MCs to be sheet-like on the largest scales (tens of parsecs). On smaller
scales, we see that the presence of magnetic fields enhances the formation of
filamentary over sheet-like sub-structures. In Section 6, we analyse the
dynamics and energetic balance of magnetized structures and relate them to the
fragmentation of cloud sub-structures. We find that the presence of magnetic
fields slows down cloud evolution and, in particular, leads to more massive
fragments at low to intermediate densities (<100 cm-3). We attempt to make an
order of magnitude estimate of this slow-down effect in Section 6.5. Finally,
we present the summary of our findings in Section 7.
## 2 Numerical methods and simulation
We present here results based on the SILCC-Zoom simulations (Seifried et al.,
2017; Seifried et al., 2019). The SILCC-Zoom simulations are MCs with
realistic boundary conditions, generated by embedding the clouds within the
SILCC simulations of multi-phase interstellar gas, thus having realistic
initial conditions (Walch et al., 2015; Girichidis et al., 2016a). In this
section, we highlight some key features of the simulations. More details on
the simulations can be found in Seifried et al. (2017) and Seifried et al.
(2019).
All simulations were executed using the adaptive mesh refinement code FLASH,
version 4 (Fryxell et al., 2000; Dubey et al., 2008), which solves the ideal
MHD equations for an ideal fluid. If we consider a fluid parcel of density
$\rho$, velocity $\mathbf{v}$, total energy $e_{\mathrm{tot}}$, and magnetic
field vector $\mathbf{B}$ (zero if pure hydrodynamics), these are given as
follows:
$\displaystyle\frac{\partial\rho}{\partial t}+\nabla\cdot(\rho\mathbf{v})=0,$
(1) $\displaystyle\frac{\partial\rho\mathbf{v}}{\partial
t}+\nabla\cdot\left[\rho\mathbf{v}\otimes\mathbf{v}+\left(P+\frac{B^{2}}{8\pi}\right)\mathbf{I}-\frac{\mathbf{B}\otimes\mathbf{B}}{4\pi}\right]=\rho\mathbf{g},$
(2) $\displaystyle\frac{\partial e_{\mathrm{tot}}}{\partial
t}+\nabla\cdot\left[\left(e_{\mathrm{tot}}+P\right)\mathbf{v}-\frac{(\mathbf{B}\cdot\mathbf{v})\mathbf{B}}{4\pi}\right]=\rho\mathbf{v}\cdot\mathbf{g}+\dot{u}_{\mathrm{heat}},$
(3) $\displaystyle\frac{\partial\mathbf{B}}{\partial
t}-\nabla\times(\mathbf{v}\times\mathbf{B})=0.$ (4)
Here, Eqs. 1 to 4 represent the conservation of mass, momentum, energy, and
magnetic flux, respectively. $P$ represents the thermal pressure, $\mathbf{g}$
is the local gravitational acceleration obtained from solving the Poisson
equation, $u$ is the internal energy, and $\dot{u}_{\mathrm{heat}}$ is the
internal energy input rate due to the combination of heating and cooling
processes. The $\otimes$ is the outer product (i.e.
$(\mathbf{a}\otimes\mathbf{b})_{ij}=a_{i}b_{j}$).
The total energy and the pressure are computed as follows:
$\displaystyle e_{\mathrm{tot}}=u+\frac{1}{2}\rho v^{2}+\frac{1}{8\pi}B^{2},$
(5) $\displaystyle P=(\gamma-1)u,$ (6)
with $\gamma$ being the adiabatic index.
Here, we present results from runs with and without magnetic fields. The MHD
simulations shown are performed using an entropy-stable solver that guarantees
minimum possible dissipation (Derigs et al., 2016; Derigs et al., 2018). The
hydrodynamic simulations have been performed using the MHD Bouchut 5-wave
solver (Bouchut et al., 2007; Waagan, 2009) that guarantees positive entropy
and density. The magnetic field strength has been set to zero for these runs.
All simulations include self-gravity as well as an external galactic potential
due to the presence of old stars. This external potential is calculated
assuming a stellar population density of $\Sigma_{\rm
star}=30~{}\mathrm{M}_{\odot}~{}\mathrm{pc}^{-2}$, a sech2 vertical profile
and a scale height of 100 pc, according to Spitzer (1942). The self-gravity of
the gas is calculated using a tree-based algorithm (Wünsch et al., 2018).
The entire simulation domain consists of a box of 500 pc $\times$ 500 pc
$\times$ $\pm$ 5 kpc size, with the long axis representing the vertical
$z-$direction of a galactic disc. The box is set with periodic boundary
conditions in the $x-$ and $y-$ direction, and outflow boundary condition in
the $z-$direction. The initial gas surface density is set to $\Sigma_{\rm
gas}=10\ \mathrm{M}_{\odot}~{}\mathrm{pc}^{-2}$ which corresponds to solar
neighbourhood conditions. The vertical distribution of the gas is modelled as
a Gaussian, i.e. $\rho=\rho_{0}\ \mathrm{exp}(-z^{2}/2h_{z}^{2})$, where
$h_{z}$=30 pc is the scale height and $\rho_{0}=9\times 10^{-24}$ g cm-3. The
initial gas temperature is set to 4500 K. For runs with magnetic fields, the
magnetic field is initialized along the $x-$direction, i.e.
$\mathbf{B}=(B_{x},0,0)$ with $B_{x}=B_{x,0}\sqrt{\rho(z)/\rho_{0}}$ and the
magnetic field strength at the midplane $B_{x,0}=3\ \mu$G. The field strength
is chosen to be in accordance with recent observations (e.g. Beck &
Wielebinski, 2013).
The turbulence in the simulations is generated by supernova explosions. The
explosion rate is set to 15 SNe Myr-1, which is consistent with the Kennicutt-
Schmidt relation, which observationally determines the star formation rate
surface density for a given gas surface density (Schmidt, 1959; Kennicutt,
1998). 50% of the supernovae are placed following a Gaussian random
distribution along the $z-$direction up to a height of 50 pc, while the other
50% are placed at density peaks of the gas. This prescription of supernova
driving creates a multi-phase turbulent ISM which can be used as initial
conditions for the zoom-in simulations (Walch et al., 2015; Girichidis et al.,
2016a).
Apart from the dynamics of the gas, we also model its chemical evolution using
a simplified non-equilibrium chemical network based on hydrogen and carbon
chemistry (Nelson & Langer, 1997; Glover & Mac Low, 2007; Glover et al.,
2010). For this purpose, we follow the abundance of H+, H, H2, CO, C+, e-, and
O. At the beginning of the simulation, all hydrogen in the disc midplane is
neutral and carbon is in its ionized form (i.e. H and C+, respectively).
To correctly model the chemistry of the gas, we include an interstellar
radiation field (ISRF) of strength $G_{0}=1.7$ in Habing units (Habing, 1968;
Draine, 1978). The attenuation of this radiation field is taken into
consideration by computing the true optical depth inside any given point in
the simulation domain. This is computed as follows:
$\mathrm{A_{V,3D}=-\frac{1}{2.5}ln\left[\frac{1}{{\it
N}_{PIX}}\sum_{i=1}^{{\it N}_{PIX}}exp\left(-2.5\frac{N_{H,tot,i}}{1.87\times
10^{21}\ \mathrm{cm}^{-2}}\right)\right]},$ (7)
where the sum is carried over each Healpix pixel, with $N_{\rm PIX}$ being the
total number of such pixels (here 48), and $N_{\mathrm{H,tot},i}$ is the
column density computed for the $i-$th pixel. In essence, for any given point,
we compute the column density along various lines of sight and use that for an
effective $\mathrm{A_{V,3D}}$. The averaging is performed in an exponential
manner because the intensity of radiation decreases in an exponential manner
due to extinction caused by the gas column density along the line of sight.
The calculation for this is performed by the TreeRay Optical Depth module
developed by Wünsch et al. (2018).
To study the formation of MCs, all supernova explosions are stopped at a
certain time $t_{0}$. Up to this point, the maximum grid resolution is 3.9 pc.
At time $t_{0}$, different regions are identified for the zoom-in process,
primarily by determining which regions form molecular gas when the simulations
are run further at the original SILCC resolution of 3.9 pc. The time $t=t_{0}$
refers to the start of the evolution of the different clouds and is set as an
evolutionary time $t_{\rm evol}=0$. The total simulation time $t$ is related
to the evolution time as
$t=t_{0}+t_{\rm evol}.$ (8)
From $t_{\rm evol}=0$ on, in the selected regions, the AMR grid is allowed to
be refined to a higher resolution to capture structures that form as MCs.
These regions are called zoom-in regions and are of primary importance to us
as sites of MCs. Each SILCC simulation we run contains two such "zoom-in"
boxes simultaneously. All runs present here have a maximum resolution of 0.125
pc. For details of how the zoom-in process is achieved, see Seifried et al.
(2017).
## 3 Classification of structures
For the analysis presented in this work, we look at eight different cubic
boxes of 62.5 pc in size, each from a different SILCC zoom-in region. These
boxes are chosen by visual inspection, in order to capture the most
interesting features contained in each zoom-in region. For the purpose of this
work, we will refer to these cubic regions as MCs. They are named MC1-HD and
MC2-HD for the two hydrodynamic clouds, and MC$x$-MHD for the MHD clouds,
where $x$ is between one and six. We present some basic details of the
different MCs in Table 1. A projected view of all the different MCs is added
in the Appendix A. For more information on the presented clouds, we refer the
reader to Seifried et al. (2017) for the HD clouds and Seifried et al. (2019)
for the MHD clouds.
Run name | MHD | $t_{0}$ | Total mass | H2 mass | $\langle\mathrm{B}\rangle$
---|---|---|---|---|---
| | [Myr] | $[10^{4}\,{\rm M}_{\odot}]$ | $[10^{4}\,{\rm M}_{\odot}]$ | [$\mu$G]
MC1-HD | no | 12 | 7.3 | 2.1 | 0
MC2-HD | no | 12 | 5.4 | 1.6 | 0
MC1-MHD | yes | 16 | 7.8 | 1.3 | 4.8
MC2-MHD | yes | 16 | 6.2 | 0.86 | 3.9
(MC3-MHDa | yes | 16 | 2.0 | 0.19 | 2.0)
MC4-MHD | yes | 11.5 | 6.8 | 1.2 | 6.4
MC5-MHD | yes | 11.5 | 10.1 | 1.6 | 6.8
MC6-MHD | yes | 16 | 6.6 | 1.4 | 4.3
Table 1: Basic information on the eight analysed simulations. From left to
right we list the run name, whether magnetic fields are present or not, the
time when the AMR "zoom-in" starts, as well as the total mass, molecular
hydrogen mass and the average magnetic field strength at $t_{\rm evol}=2$ Myr.
aWe discard MC3-MHD from our further analysis because of its low molecular gas
content and lack of interesting density features (see also Fig. 14).
We perform a detailed analysis of the different clouds, following their
evolution from $t_{\rm evol}=2$ Myr to $t_{\rm evol}=3.5$ Myr, primarily
focusing on the latter time. The beginning and the end time are chosen to look
at relatively early stages of structure formation in the MCs. We do not look
at times earlier than 2 Myr primarily because the clouds undergo the
refinement process and are not fully resolved until $t_{\rm evol}\sim 1.5$
Myr.
### 3.1 Structure identification
To identify structures in our MCs, we use a dendrogram algorithm (Rosolowsky
et al., 2008). Dendrogram is a model-independent method to determine
hierarchical structures in two and three dimensions. Since we are interested
in 3-dimensional structures, we perform the dendrogram analysis on
3-dimensional density cubes. We do not use the 3D AMR grid structure inherent
in the data, but rather convert it into a uniform mesh at 0.125 and 0.25 pc
resolution (see also Table 2).
Given an initial density field, $\rho$, the dendrogram essentially depends on
three free parameters: the initial starting threshold, $\rho_{0}$, the density
jump, $\Delta\rho$, and the minimum number of cells that need to be included
in any structure, $N_{\rm cells}$. Due to high density contrasts, we build the
dendrogram tree on the logarithmic density profile of the gas, and therefore
have used density bins of $\Delta\mathrm{log}_{10}\ \rho$, rather than
$\Delta\rho$. In addition to the three parameters mentioned, we can choose a
pruning peak, $\rho_{\rm prune}$, to allow the dendrogram to create new
structures only when such a structure will have peak density $\rho_{\rm
peak}>\rho_{\rm prune}$, although this has not been used in the present work.
Using these parameters, the dendrogram algorithm allows us to define volumes
of gas as structures in a hierarchical tree, primarily defined by their
threshold density $\rho_{\rm thr}$, which is the minimum density value inside
a given structure. This can be thought of as equivalent to contour values for
two dimensional maps. The hierarchy is characterised by different dendrogram
branches, where a branch is a given dendrogram structure and all its parent
structures, up to the largest and most diffuse ancestor in the dendrogram
tree.
For probing both the higher and lower density ends of the data, we perform two
dendrgram analyses on the same regions: a higher density dendrogram analysis
performed at a resolution of 0.125 pc for probing gas above densities of 10-22
g cm-3 (referred to as high-den), and a lower density analysis performed at
0.25 pc for gas between the densities of 10-24 and 10-22 g cm-3 (referred to
as low-den). The low-den values are computed as volume averaged values from
the higher resolution grid. We present the dendrogram parameters used for both
analyses in Table 2.
dendrogram | Resolution | $\rho_{0}$ | $\Delta$ log${}_{10}\ \rho$ | $N_{\rm cells}$ | $\rho_{\rm prune}$ | additional
---|---|---|---|---|---|---
type | [pc] | [g cm-3] | | | [g cm-3] | criteria
high-den | 0.125 | $10^{-22}$ | 0.1 | 100 | None | None
low-den | 0.25 | $10^{-24}$ | 0.2 | 100 | None | $\rho_{\rm thr}<10^{-22}$ g cm-3
Table 2: Information on the parameters used for the two different kinds of
dendrogram analyses. From left to right are: the type of dendrogram, the grid
resolution at which it is performed, the starting density, the logarithmic
density jump, the minimum number of cells in structures, the density of the
pruning peak used, and if any additional criteria were used to select
structures.
In addition to the difference in the basic parameters between the two
dendrogram analyses, we remove all structures with $\rho_{\rm thr}>10^{-22}$ g
cm-3 for the low-den analysis. This is done in order to avoid double counting
of structures.
The parameter values mentioned in Table 2 have been chosen from a mixture of
practical considerations, such as CPU memory, computation time, and through
trial and error. We note that in principle the same analysis could be
performed by a single dendrogram analysis at $\rho_{\rm thr}=10^{-24}$ g cm-3
at the highest resolution of 0.125 pc. However, the computation cost of such
an analysis was prohibitive in our case. Combining the high-den and low-den
dendrogram analyses allows us to probe a much higher density range than would
be otherwise possible.
In terms of the parameters used, we have seen no unexpected change in the
results by changing the free parameters within a reasonable range. We refer
the reader to our companion paper (Ganguly et al., 2022) for a more thorough
discussion of the effect of altering the parameter values on the analysis.
Overall, we find that changing the parameters, while resulting in a varying
number of obtained structures, leaves the statistical properties of the
structures virtually unaffected.
An example of the leaf density structures (structures that contain no further
sub-structures) from the dendrogram analysis can be seen in Fig. 1 for MC1-MHD
at $t_{\rm evol}=3.5$ Myr, as contours over column density maps. The three
panels show, from left to right, the cloud projected along the $x-$, $y-$ and
$z-$direction. The contours are drawn as projections of the 3D dendrogram
structure outlines in the projected direction. We distinguish between
structures depending on their molecular H2 content, by plotting structures
with over 50% of their total hydrogen mass in molecular form (referred to as
molecular structures) in solid lines and otherwise in dashed lines (referred
to as atomic structures).
Figure 1: Left to right: Projections of MC1-MHD at $t_{\rm evol}=$3.5 Myr
along the $x$-, $y$-, and $z$-axis, respectively. The contours show the
projections of the leaf dendrogram structures along the same axis. Molecular
structures ($>50\%\;\mathrm{H}_{2}$ mass fraction) are plotted with solid, and
atomic structures ($<50\%\;\mathrm{H}_{2}$ mass fraction) are plotted with
dashed lines. The molecular structures nicely trace the dense spine of the two
main filaments, while the atomic structures mostly represent the envelope.
Due to the nature of the dendrogram algorithm, there are some structures which
touch the edge of the box. This can lead to structures whose morphology is
determined by their proximity to the edge. To avoid this, we do not classify
the morphology of any structures which have more than 5% of their surface
cells touching any edge. This is relevant especially for the large-scale
structures from the low-den dendrogram analysis. However, they can still be of
interest while considering cloud dynamics, and in such a case we add them as
an additional category of "unclassified". While in a different context, Alves
et al. (2017) have shown the importance of having closed contours while
studying 2D maps. We have attempted to follow the same principle here as much
as possible.
### 3.2 Structure classification
Once we obtain the tree of dendrogram density sub-structures, we aim to
classify their morphology. For each structure, we compute an equivalent
ellipsoid that has the same mass and the same moments of inertia (MOI) as the
original structure. We then use the axes lengths of this equivalent ellipsoid
to classify the shape of the different structures.
Let us consider a uniform density ellipsoid of mass $M$ and semi-axes lengths
$a,\ b,\ c$ with $a\geq b\geq c$. The moments of inertia along the three
principal axes will be given as follows:
$\begin{split}I_{a}&=\frac{1}{5}M(b^{2}+c^{2}),\\\
I_{b}&=\frac{1}{5}M(c^{2}+a^{2}),\\\
I_{c}&=\frac{1}{5}M(a^{2}+b^{2}),\end{split}$ (9)
where $I_{c}\geq I_{b}\geq I_{a}$. If we now compute the principal moments of
inertia of our given dendrogram structure to be $A$, $B$ and $C$,
respectively, then the ellipsoid has an equivalent moment of inertia if
$A=I_{a},\ B=I_{b},\ C=I_{c}.$ (10)
This leads to the following equation for computing the axis lengths of the
equivalent ellipsoids:
$\begin{split}a&=\sqrt{\frac{5}{2M}(B+C-A)}\ ,\\\
b&=\sqrt{\frac{5}{2M}(C+A-B)}\ ,\\\ c&=\sqrt{\frac{5}{2M}(A+B-C)}\
.\end{split}$ (11)
We then use the aspect ratio of the semi-axes of the corresponding ellipsoid
and the position of the center of mass (COM) of the structure relative to its
boundary (i.e. whether the COM is contained by the structure itself) to
categorise the different structures into four categories - sheets, curved
sheets (referred to as sheet_c in this paper), filaments, and spheroids:
$\begin{split}&\textbf{sheet: }\frac{a}{b}\leq f_{\rm asp},\frac{a}{c}>f_{\rm
asp}\\\ &\textbf{filament: }\frac{a}{b}>f_{\rm asp}\\\ &\textbf{spheroidal:
}\frac{a}{c}\leq f_{\rm asp},\text{ contains its own COM}\\\
&\textbf{sheet\\_c: }\frac{a}{c}\leq f_{\rm asp},\text{ does not contain its
own COM}\end{split}$ (12)
where we set the aspect ratio factor $f_{\rm asp}=3$.
The inclusion of the COM criterion in addition to the ratio of the ellipsoid
axes help us deal with especially the larger-scale structures which can be
highly curved. A highly curved sheet could have comparable MOI eigenvalues
along the different eigen-directions, but would not contain its own COM. We
highlight some visual examples of such highly curved sheet-like structures
when we discuss the large scale morphology of our clouds in Section 5. In
contrast to curved sheets, a spheroidal structure would contain its own COM.
Apart from using the normal moment of inertia, we also perform the
classification by computing a volume-weighted moment of inertia, where we
compute the moment of inertia of the structures (the quantities $A$, $B$ and
$C$) by assuming the structure is of the same mass but with uniform density,
but find statistically little to no difference in the resulting morphologies.
The discussion above highlights some possible caveats of our method. If we
have a situation of multiple crossing filaments (hub-like structure), or
parallel filaments joined by a more diffuse intermediate medium - the method
will identify it as a sheet-like structure splitting into filaments in the
dendrogram tree hierarchy. We must therefore emphasise that our definition of
a sheet in this context is more general and contains also situations where
multiple filamentary structures are connected by a more diffuse medium.
Further, for highly curved structures, it is possible that the simple fit
ellipsoid method may not result in a good description of the ellipsoid axis
lengths.
## 4 Density distribution and magnetic fields
We first consider the bulk properties of the different MCs to quantify the
differences between the hydrodynamic and MHD clouds. From Table 1, we see that
the volume-weighted root-mean-square average magnetic field strength for all
MHD clouds is comparable and varies between 3.9-6.8 $\mu$G. These values are
slightly higher than the initial magnetic field strength $B_{x,0}=3\,\mu{\rm
G}$. The cloud masses and their H2 masses are also within a factor of roughly
2 to each other (with the exception of MC3-MHD, see below). For a view of the
time evolution of the total and H2 masses, as well as the H2 mass fraction, we
refer the reader to Appendix A.
MC3-MHD stands out as it has a much lower H2 mass and H2 mass fraction
compared to the other clouds (Table 1). Visual inspection of this cloud shows
that its structures are still diffuse and not as prominent, suggesting that it
perhaps needs much longer to collapse, or may not collapse at all (see Fig.
14, bottom row left). Its molecular content remains at a roughly constant
level of 10% throughout. Since we are interested primarily in the problem of
density structures that eventually form stars, we exclude MC3-MHD from further
analysis considering its unevolved state and low molecular content.
It is of interest to examine whether the mass distribution in different clouds
is affected by the presence of magnetic fields. This can be seen in Fig. 2,
which shows the volume-weighted density PDF of all different clouds at $t_{\rm
evol}=2$ Myr (top) and $t_{\rm evol}=3.5$ Myr (bottom) in the density range
probed by the dendrogram analysis ($>10^{-24}$ g cm-3). The respective density
PDFs for the full density range can be found in Appendix B. The two
hydrodynamic clouds are plotted using reddish lines (red and salmon), while
the magnetised clouds are shown using darker colours. For all clouds, the
shown density range contains more than 99% of their total mass.
Figure 2: Volume-weighted density PDF for different HD and MHD clouds $t_{\rm
evol}=2$ Myr (top) and 3.5 Myr (bottom). The density range shown is used for a
dendrogram analysis, and contains more than 99% of the total mass of the
clouds. The two hydrodynamic clouds are plotted in reddish lines. The vertical
line demarcate the boundaries of the high-den ($>10^{-22}$ g cm-3) and the
low-den (between $10^{-24}-10^{-22}$ g cm-3) dendrogram analyses (see also
Table 2). The MHD clouds have more fraction of gas in the density range
between roughly $10^{-24}$ and $10^{-22}$ g cm-3, or between approximately 1
and 100 cm-3.
From Fig. 2, we see that between $10^{-24}$ and $10^{-22}$ $\mathrm{g\
cm^{-3}}$, corresponding to the rough number densities between 1 and 100
$\mathrm{cm^{-3}}$, the MHD clouds contain much more gas. This is more
prominent at $t_{\rm evol}=2$ Myr, but remains also clearly visible at $t_{\rm
evol}=3.5$ Myr. This effect can also be visually seen in the column density
plots of Fig. 14, where the denser parts of the hydrodynamic clouds seem to be
embedded in a more rarefied medium compared to their MHD counterparts. We
calculate the mass percentage at 2 Myr in different density regimes in Table
3, which shows that, at this time, the MHD clouds contain almost 50% of their
mass between $10^{-24}$ and $10^{-22}$ $\mathrm{g\ cm^{-3}}$, in contrast to
only around 26% for the hydrodynamic MCs.
Cloud | Mass percentage at 2 Myr
---|---
sample | $\frac{\rho}{{\rm g}\,{\rm cm}^{-3}}<10^{-24}$ | $10^{-24}\leq\frac{\rho}{{\rm g}\,{\rm cm}^{-3}}<10^{-22}$ | $\frac{\rho}{{\rm g}\,{\rm cm}^{-3}}\geq 10^{-22}$
HD | 0.9 | 25.9 | 73.2
MHD | 0.5 | 50.0 | 49.5
Table 3: The average mass percentage in different density regimes for the HD
and MHD clouds at $t_{\rm evol}=2$ Myr. The MHD clouds have twice the amount
of mass in the intermediate density range between $10^{-24}$ g cm-3 and
$10^{-22}$ g cm-3 compared to their HD counterparts.
Magnetic fields in our simulations therefore play an important role in shaping
the environment inside which denser, molecular, and potentially star-forming
structures live. This is consistent with the picture that magnetic fields have
a noticeable effect on the dynamics of low (here, $\lesssim 10^{-22}$ g cm-3)
density gas (Molina et al., 2012). Similar conclusions have been reached by
Seifried et al. (2020b) using the technique of relative orientation of
magnetic fields with respect to filaments who find that this change in the
relative impact of magnetic fields occurs around $\sim 100$ cm-3. We explore
the effects of magnetic fields in more detail by looking at the clouds’
fragmentation properties in Section 6.4.
For $\rho>10^{-22}$ g cm-3 we see no clear trend in the slope of the density
PDF between the hydrodynamic and MHD clouds. This is consistent with
simulations and observations showing that column density PDFs are not
sensitive to the presence of a magnetic fields in the high column density
regime (Klessen & Burkert, 2001; Slyz et al., 2005; Girichidis et al., 2014;
Schneider et al., 2015).
However, at $t_{\rm evol}=2$ Myr, the two hydrodynamic clouds seem to have a
bit more dense gas mass (see also Table 3), although the effect is visually
far from clear. If there were a "delay" in the formation of denser gas when
magnetic fields are present, this would be extremely relevant for the
formation of well-shielded, molecular gas. In Table 4, we show the mass above
an $\mathrm{A_{V,3D}}$ (Eq. 7) of 1 and 10 for one magnetised and one non-
magnetised cloud of comparable mass (MC1-MHD and MC1-HD, see Table 1).
Additionally, the mass-weighted PDF of $\mathrm{A_{V,3D}}$ for these two
clouds is shown in Appendix C, Fig. 18. From Table 4, as well as from Fig. 18,
we find that the amount of gas above $\mathrm{A_{V,3D}}>1$ and
$\mathrm{A_{V,3D}}>10$ in MC1-MHD is consistently lower compared to MC1-HD. In
Section 6.5, we attempt to quantify such a delay timescale due to magnetic
fields. For a more detailed analysis on the connection between magnetic fields
and $\mathrm{A_{V,3D}}$, we refer the reader to Seifried et al. (2020a).
cloud, time | mass above | mass above
---|---|---
| $\mathrm{A_{V,3D}}>1$ [%] | $\mathrm{A_{V,3D}}>10$ [%]
MC1-HD, 2 Myr | 41.1 | 2.3
MC1-HD, 3.5 Myr | 44.0 | 9.8
MC1-MHD, 2 Myr | 26.6 | 0
MC1-MHD, 3.5 Myr | 31.7 | 0.8
Table 4: The percentage of mass above values of $\mathrm{A_{V,3D}}=1,\ 10$ for
two similar mass clouds, MC1-HD and MC1-MHD.
## 5 Morphology
We perform a morphological classification of all simulated cloud structures
using the method described in Section 3.2. As an intuitive visual aid, we
first present 3D surfaces of three large-scale cloud dendrogram
structures111We show the largest structures from the high-den dendrogram
analysis ($\rho>10^{-22}$ g cm-3) as they are on the maximum resolution and
therefore capture the finer complexities of the cloud better. The large scale
structures for the low-den dendrogram analysis follow the same trend. (from
top to bottom: MC1-MHD, MC5-MHD, and MC6-MHD) seen from three different
viewing angles (different columns) in Fig. 3. The lighter blue colour shows
the large-scale structure (identified at $\rho_{\rm thr}\approx 10^{-22}$ g
cm-3) and in red, we show one of the primary embedded filamentary structures
(identified using values of $\rho_{\rm thr}$ between $10^{-20}-10^{-21}$ g
cm-3). Visual inspection seems to suggest that the large-scale, lighter blue
structures are rather thin and sheet-like, and indeed all three clouds shown
in Fig. 3 are identified as sheets or curved sheets according to the
classification algorithm of Section 3.2. This is even clearer in a video view,
which can be found here222https://hera.ph1.uni-koeln.de/~ganguly/silcc_zoom/.
The visual suggestion of the clouds being sheet-like on the largest scales is
also confirmed for all clouds in a quantitative analysis, presented below.
Figure 3: 3D surface rendering of example large-scale dendrogram structures
from the high-den dendrogram analysis for MC1-MHD (top row), MC5-MHD (middle
row), and MC6-MHD (bottom row), from different viewing angles (left to right).
The blue structures represent the large-scale sheets or curved sheets at
$\rho_{\rm thr}\approx 10^{-22}$ g cm-3, while the embedded red structures
show one of the more prominent embedded filaments ($\rho_{\rm thr}$ between
$10^{-20}-10^{-21}$ g cm-3). The units in the axes are in parsec. A video link
for the various structures can be found in https://hera.ph1.uni-
koeln.de/~ganguly/silcc_zoom/morphology_3d/.
We estimate the size of the structures simply from the volume $V$ as:
$R=V^{1/3}.$ (13)
We define $N_{\rm tot}$ as the total number of morphologically classified
structures, i.e. $N_{\rm tot}$ is
$N_{\rm tot}=N_{\rm sheet}+N_{\rm sheet\\_c}+N_{\rm filament}+N_{\rm
spheroid},$ (14)
with $N_{x}$ being the total number of structures (i.e. both parents and
leaves) of morphological class $x$ (where $x\in$ [sheet, sheet_c, filament,
spheroid]). We express the number of structures of type $x$ at a given size
$R$ by $N_{x}(R)$.
In Fig. 4 we plot the cumulative fraction (i.e. $N_{x}(R)/N_{\rm tot}$) of
sheets, curved sheets, filaments, and spheroidal structures against $R$ for
all structures (i.e. both parents and leaves) in the two hydrodynamic clouds
(left panel) and the five MHD clouds (right panel) at $t_{\rm evol}=3.5$ Myr.
The numerical values of the overall fractions across all scales, $\int
N_{x}(R)/N_{\rm tot}\ \mathrm{d}R$, for both HD and MHD clouds at two
different times can be found in Table 5.
Figure 4: Cumulative histogram of different morphologies (sheets, curved sheets, filaments, or spheroids) for all HD (left) and MHD (right) clouds at $t_{\rm evol}=$ 3.5 Myr. 6 out of the 7 analysed clouds are sheet-like on large scales, with filamentary networks embedded inside. Spheroidal structures are rarer in the presence of magnetic fields. Both HD and MHD clouds produce more sheets than filaments, but the MHD runs tend to have a relative increase in the fraction of filaments. cloud, time | $\frac{N_{\rm sheet}}{N_{\rm tot}}$ | $\frac{N_{\rm sheet\\_c}}{N_{\rm tot}}$ | $\frac{N_{\rm filament}}{N_{\rm tot}}$ | $\frac{N_{\rm spheroid}}{N_{\rm tot}}$ | $N_{\rm tot}$
---|---|---|---|---|---
HD, 2 Myr | 0.58 | 0.12 | 0.22 | 0.08 | 910
HD, 3.5 Myr | 0.63 | 0.07 | 0.19 | 0.11 | 1167
MHD, 2 Myr | 0.57 | 0.03 | 0.31 | 0.09 | 487
MHD, 3.5 Myr | 0.56 | 0.04 | 0.33 | 0.07 | 2087
Table 5: Fraction of sheets, curved sheets, filaments, and spheroids among all
morphologically classified structures, for both HD and MHD clouds at $t_{\rm
evol}=2,\ 3.5$ Myr. While all clouds are dominated by sheet-like structures,
the MHD clouds have a higher fraction of filaments compared to their
hydrodynamic counterparts.
We find that spheroidal structures, shown in green, are generally less
numerous compared to sheet-like or filamentary structures ($\sim$10% of
$N_{\rm tot}$ are spheroidal, Table 5). Sheets (including curved sheets)
appear to be the most abundant structures within all clouds (summing up to
$\sim$70% for the HD case and $\sim 60$% for the MHD case). However, filaments
are considerably more abundant in the MHD clouds compared to their HD
counterparts ($>30$% for MHD as opposed to $\sim$20% for HD clouds). In terms
of size, we find that at the largest $R$ values, indeed almost all clouds (six
out of seven) are either sheets or curved sheets, confirming the visual trend
we found in Fig. 3.
We highlight the morphological trends as a function of the molecular fraction
in Fig. 5. Similar to Fig. 4, we plot here the cumulative fraction of (curved)
sheets, filaments, and spheroids, but this time as a function of the molecular
mass fraction $f_{\rm H_{2}}$, which is the H2 mass in a structure divided by
the total hydrogen mass in the structure. Note that structures with high
$f_{\rm H_{2}}$ are usually small (located mostly at small $R$ in Fig. 4). We
see that around $f_{\rm H_{2}}>0.7$, there are more filaments than sheet-like
structures in the MHD case (right panel). This trend is absent for the HD
clouds (left panel). This implies that magnetic fields particularly enhance
the formation of filaments on the small scales, shaping the morphology of the
denser, well-shielded, molecular gas. This is in line with the fact that
magnetic fields can, in general, aid the formation of filamentary sub-
structures (Hacar et al., 2022; Pineda et al., 2022).
Figure 5: Cumulative histogram of different morphologies (sheets, curved
sheets, filaments, or spheroids) against H2 mass fraction for all HD (left)
and MHD (right) clouds at $t_{\rm evol}=$ 3.5 Myr. The most molecular
structures are more filamentary in presence of magnetic fields.
Gravitational collapse naturally proceeds anisotropically and tends to create
elongated structures (e.g. Burkert & Hartmann, 2004). However, we show in
Ganguly et al. (2022) that most of our cloud structures are unbound or only
marginally bound. This being the case, gravity cannot be the principal
contributor to forming elongated structures, and we must therefore identify
other possible sources of the lack of spheroidal structures. Shock compression
and turbulence are two such methods for producing elongated structures (see,
e.g., Inoue & Inutsuka 2016 for shock compression; Federrath 2016 for
turbulence; and Hacar et al. 2022 for a general overview). Sheets and
filaments are both elongated structures. However, it is interesting that for
the hydrodynamic clouds, sheets are by far the most numerous, whereas for the
MHD clouds filaments and sheets are more comparable in total number. This is
consistent with the results of Hennebelle (2013), who investigate setups of
both decaying supersonic turbulence and colliding flows, and find that their
simulations tend to produce more sheet-like structures for hydrodynamical
simulations, and more filamentary structures for MHD simulations.
Overall, we see primarily sheet-like MCs with an abundance of elongated
structures (filamentary or sheet-like), irrespective of whether the simulation
contains magnetic fields or not. Sheets are generally more numerous, probably
representing the fact that we trace a large number of structures belonging to
the sheet-like atomic envelope of the MCs. This is supported by the fact that,
in Fig. 5, both HD and MHD clouds show an abundance of sheet-like structures
below $f_{\rm H_{2}}\approx 0.5$. The presence of magnetic fields, however,
tends to somewhat increase the fraction of filamentary over sheet-like
structures.
The sheet-like nature of our clouds is consistent with a number of recent
observations. Kalberla et al. (2016) have argued that the cold, neutral
hydrogen in the ISM is organised in sheet-like structures. Investigating the
L1495 region of the Taurus molecular cloud, Arzoumanian et al. (2018) report
evidence of extended sheet-like structures too. Using the recent GAIA data,
Rezaei Kh. & Kainulainen (2022) have concluded that the California molecular
cloud is sheet-like in nature. Tritsis et al. (2022) have reached a similar
conclusion regarding the Musca molecular cloud using 3D dust extinction maps.
Based on a Herschel study of the giant molecular filament G214.5, Clarke et
al. (2023) have also posited that the filament is a result of the HI shell of
an expanding superbubble interacting with the local medium. Our findings here
are thus perfectly in line with these observations.
The morphology of MCs at larger (tens of parsecs) scales is of paramount
importance in relation to how the MCs themselves form. Our analysis shows that
the clouds are preferentially sheet-like, with and without magnetic fields.
The ISM in the SILCC simulations (and therefore also in the SILCC-Zoom
simulations) has a multi-phase structure (Walch et al., 2015; Girichidis et
al., 2016a). The MCs in these simulations form primarily at the shells or
intersections of expanding supernova bubbles. The large-scale sheets we see,
can therefore be interpreted as tracing these supernova-driven shells, with a
complex network of different morphological sub-structure contained within.
This picture is consistent with the bubble-driven structure formation scenario
(Koyama & Inutsuka, 2000; Inoue & Inutsuka, 2009; Inutsuka et al., 2015;
Pineda et al., 2022).
## 6 Dynamics and Fragmentation
### 6.1 The magnetic field - density scaling
The impact of magnetic fields on the MCs is naturally correlated to the field
strength. The initial 3 $\mu$G seed field in the original simulations is
expected to be enhanced when we look at denser structures inside the MCs. The
scaling behaviour of the magnetic field $B$ with $\rho$ is integral to
understanding the importance of magnetic fields at different scales.
If contraction of gas occurs exclusively along the magnetic field lines, this
should lead to no dependence of the magnetic field strength on the density,
i.e. $B\propto\rho^{0}$. If magnetic field lines do contract with the
enhancement of gas density, then one expects a scaling similar to
$B\propto\rho^{\kappa}$, with $\kappa=0.5,0.67$ for the strong and weak field
limits, respectively (see e.g. the review by Hennebelle & Inutsuka, 2019).
In the ISM, indeed the $\kappa=0$ relation is observed up to number densities
of $\sim$300 cm-3 (Troland & Heiles, 1986; Crutcher et al., 2010). This
corresponds to densities of roughly 1.1$\times 10^{-21}$ g cm-3, using a mean
molecular weight of 2.35. Crutcher et al. (2010) find that above these
densities, the data is consistent with $\kappa=2/3$, with considerable
scatter. The transition in power law is usually associated with the magnetic
fields becoming dynamically sub-dominant (Seifried et al., 2020b; Pattle et
al., 2022) and roughly matches with our observation that below $\sim 100$ cm-3
the mass in the MHD clouds is enhanced.
We can attempt to capture whether this transition in the importance of the
magnetic field is seen in the Alfvénic Mach number, $\mathcal{M}_{\rm A}$. For
a given sub-structure, we can compute $\mathcal{M}_{\rm A}$ as
$\mathcal{M}_{\rm A}=\sigma_{\rm 1D}/v_{\rm A}.$ (15)
Here $\sigma_{\rm 1D}$ is the one-dimensional velocity dispersion and $v_{\rm
A}$ is an estimate of the average Alfvén wave group velocity. For a structure
of mass $M$, we compute $\sigma_{\rm 1D}$ from
$\sigma_{\rm
1D}^{2}=\frac{1}{3M}\int_{V}\rho(\mathbf{v}-\mathbf{v}_{0})^{2}\mathrm{d}^{3}r,$
(16)
with $\mathbf{v}_{0}$ being the centre of mass velocity computed as
$\mathbf{v}_{0}=\frac{1}{M}\int_{V}\rho\mathbf{v}\mathrm{d}^{3}r.$ (17)
The integration is performed over the entire volume $V$ of the given
structure.
The Alfvén velocity can be computed as
$v_{\rm A}=\sqrt{\frac{\langle|\mathbf{B}|^{2}\rangle}{4\pi\rho_{\rm avg}}}.$
(18)
The density $\rho_{\rm avg}$ here is the volume-averaged density, i.e.
$\rho_{\rm avg}=M/V,$ (19)
and $\langle|\mathbf{B}|^{2}\rangle$ is the volume-averaged square of the
magnetic field $\mathbf{B}$,
$\langle|\mathbf{B}|^{2}\rangle=\frac{1}{V}\int_{V}|\mathbf{B}|^{2}\mathrm{d}^{3}r.$
(20)
The behaviour of the magnetic field strength with density for the MHD clouds
can be seen in Fig. 6, where we plot the root-mean-square magnetic field
strength against the threshold (minimum) density $\rho_{\rm thr}$ for all
dendrogram structures at $t_{\rm evol}=3.5$ Myr. The different dendrogram
structures are marked with filled/empty symbols depending on whether their H2
mass fraction (with respect to their total hydrogen mass) is greater/less than
50%. The colour bar shows $\mathcal{M}_{\rm A}$, as computed from Eq. 15. The
reddish points represent super-Alfvénic ($\mathcal{M}_{\rm A}>1$) structures,
while the blueish points are sub-Alfvénic ($\mathcal{M}_{\rm A}<1$). In the
sub-Alfvénic case, the fluid speed is smaller than the magnetic wave speed,
meaning that the magnetic field is dynamically important and guides the flow.
The vertical dotted line at $10^{-22}$ g cm-3 represents the boundary between
the points obtained from the low-den (left half) and high-den (right half)
dendrograms, respectively. The dash-dotted black line and the dotted power-law
represent the Crutcher et al. (2010) relation discussed previously and
$B\propto\rho^{0.5}$, respectively.
Figure 6: Relation between the root-mean-square magnetic field and $\rho_{\rm
thr}$ for all MHD clouds at $t_{\rm evol}$=3.5 Myr. The colour bar shows the
Alfvénic Mach number $\mathcal{M}_{\rm A}$. The dash-dotted line represents
the B$-\rho$ relation from Crutcher et al. (2010), while the dotted line
represents a $B\propto\rho^{0.5}$ power law. The cyan dashed line represents
the best fit power law for all points with $\rho_{\rm thr}>1.1\times 10^{-21}$
g cm-3.
The cyan dashed line represents the linear least-squares best fit performed on
the logarithm of the points for high densities ($\rho_{\rm thr}>1.1\times
10^{-21}$ g cm-3). The best fit of $\kappa=0.47\pm 0.03$ is consistent with
the strong-field limit of $B\propto\rho^{0.5}$. We have already shown in the
previous section (Section 5) that our structures are on average highly
elongated, and magnetic fields clearly help to deform the shape of the forming
structures. It is therefore not unexpected that we find a shallower scaling
compared to the weak field limit ($\kappa=0.67$).
We see that, while there is no clear transition from the sub- to the super-
Alfvénic regime, there is clearly a trend that higher Alfvénic Mach numbers
are preferentially obtained at the higher density end. This is confirmed by a
Kolmogorov-Smirnov (KS) two-sample test, which compares if two distributions
belong to the same population. In this case, we compare the $\rho_{\rm
thr}$-distributions of structures with $\mathcal{M}_{A}>1$ and
$\mathcal{M}_{A}\leq 1$. We find the $p$-values333If the $p$-value is larger
than a certain value (typically 0.05), this means that we cannot reject the
null hypothesis that the sub-Alfvénic and super-Alfvénic structures have the
same underlying density distribution. to be very low: $6\times 10^{-4}$ at 2
Myr and $5.2\times 10^{-15}$ at 3.5 Myr (see Table 6).
Crutcher et al. (2010) found that the observed magnetic field distribution is
rather flat at low density, in agreement with the idea that denser clouds are
swept up along the magnetic field lines on large scales, while at higher
density there is a power-law increase of the magnetic field strength. If
spherical clouds start to collapse and the magnetic field is not strong enough
to stop the collapse, one expects a power-law slope of $\kappa=0.5-0.67$ (see
above).
In the case of our clouds, we find that the high-density end is well
consistent with $\kappa=0.5$, and the lower-density end clearly shows a much
shallower slope. Nonetheless, there does not seem to be a clear single density
at which there is a sharp change in slope. Simulations by Li et al. (2015),
Mocz et al. (2017), Girichidis et al. (2018), Zhang et al. (2019) find
similarly the lack of a sharp transition density. Auddy et al. (2022) predict
that the transition density depends on the fourth power of $\mathcal{M}_{A}$.
While of potential interest, this is unfortunately not demonstrable from the
present analysis.
variable 1 | variable 2 | time [Myr] | p-value
---|---|---|---
$\rho_{\rm thr}(\mathcal{M}_{A}>1)$ | $\rho_{\rm thr}(\mathcal{M}_{A}\leq 1)$ | 2 | $6\times 10^{-4}$
| | 3.5 | $5.2\times 10^{-15}$
Table 6: The $p$-values of the 2-sample KS test for the density distribution
of sub-Alfvenic and super-Alfvénic structures. We can see that the $p$-value
is low for both 2 and 3.5 Myr, suggesting that sub-Alfvénic and super-Alfvénic
structures (corresponding to bluish and reddish points in Fig. 6,
respectively) have statistically significant differences in their density
distributions.
### 6.2 Impact of magnetic fields on the energetics of sub-structures
We are also interested in assessing the energetic relevance of magnetic fields
over different length scales in the MCs, especially with respect to
potentially star-forming structures. For this purpose, we compute the volume
term of the magnetic energy and compare it with the kinetic and potential
energies. Similar work for the same simulations has been performed by Ganguly
et al. (2022), who assess the virial balance of the cloud sub-structures.
Here, we extend the range of our analysis to include the dynamics of lower-
density gas (between 10-24 and 10-22 g cm-3; low-den dendrogram analysis, see
Table 2).
The magnetic energy of a given structure is computed as
$E_{\rm B}=\int_{V}\frac{1}{8\pi}|\mathbf{B}|^{2}\mathrm{d}^{3}r,$ (21)
where the integration is computed over the entire volume $V$ of the structure.
The kinetic energy is computed using the following relation:
$E_{\mathrm{KE}}=\frac{1}{2}\int_{V}\rho(\mathbf{v}-\mathbf{v}_{0})^{2}\mathrm{d}^{3}r.$
(22)
Here, $\mathbf{v}_{0}$ is the centre of mass velocity computed from Eq. 17.
The self-gravitating potential energy of a given structure is obtained using
the following relation:
$E_{\rm
PE}=-\frac{1}{2}G\int_{V}\int_{V}\frac{\rho(\mathbf{r})\rho(\mathbf{r^{\prime}})}{|\mathbf{r}-\mathbf{r^{\prime}}|}\mathrm{d}^{3}r\mathrm{d}^{3}r^{\prime},$
(23)
where $G$ is the gravitational constant. We compute the self-gravity of each
dendrogram structure using a KD-tree algorithm (Bentley, 1975) instead of an
$\mathcal{O}(N^{2})$ direct computation.
We show the relative importance of magnetic fields with respect to potential
and kinetic energy in the left and right panel of Fig. 7, respectively, for
all MHD cloud structures at $t_{\rm evol}=3.5$ Myr. For both plots, the
$x$-axis represents the density threshold $\rho_{\rm thr}$, and the $y$-axis
represents $E_{\rm B}/|E_{\rm PE}|$ (left) and $E_{\rm B}/|E_{\rm KE}|$
(right), respectively. The colours of the points represent their morphologies.
Here, for the purpose of understanding the dynamics of low-density gas, we
also include the "unclassified" structures (i.e. structures with >5% of their
surface cells touching the edge of the analysis box, see Section 3). The side
panels to the right and top of each plot show the marginal distributions of
$N_{x}/N_{\rm tot}$ for each morphology. Note that, since the definition of
$N_{\rm tot}$ (Eq. 14) does not contain unclassified structures, the fractions
in the two side panels add up to greater than unity. The filled symbols are
molecular structures, while the open symbols are atomic.
Figure 7: Ratio of magnetic energy to self-gravitating potential energy (left)
and to kinetic energy (right), respectively, plotted against the density
threshold for all dendrogram structures of all MHD clouds at time $t_{\rm
evol}=3.5$ Myr. The colours represent different morphologies. The dash-dotted
lines indicate a $\rho^{-1/2}$ relation. The top and the right panels show the
marginalised distributions (separated by morphology) over the density and the
corresponding energy ratio.
Typically, for low-density structures, which mostly consist of atomic gas, the
magnetic energy is either comparable to or much larger than the potential
energy (left panel of Fig. 7). The magnetic energy is also comparable to or
larger than the kinetic energy (right panel), but the spread in this energy
ratio is much smaller compared to the $E_{\rm B}/E_{\rm PE}$ ratio. For some
branches (a dendrogram branch is defined as a given structure and all its
parent structures, see Section 3.1), the energy ratio seems to roughly follow
a $\rho^{-1/2}$ power law. These branches represent the evolution from
diffuse, large-scale structures to denser, embedded structures. Camacho et al.
(2022) also find a tight power-law scaling between the potential and magnetic
energies. While not exactly the same, both scaling behaviours seem to imply
that magnetic fields become less important as we go deeper into the MCs
themselves. This is also in accordance with the findings of Seifried et al.
(2020b), Ibáñez-Mejía et al. (2022), as well as Ganguly et al. (2022), as
discussed previously.
From the marginal distributions, we find a weak trend that the high-density
end is dominated by filaments. Curved sheets and unclassified structures only
appear at lower densities because they are usually larger-scale structures.
There is no obvious correlation between the morphology of the structures and
the energy ratios. This suggests that the different morphological
configurations are created by the same formation mechanism, most likely
turbulent compression.
There also seems to be a difference in the energy ratios between atomic and
molecular structures. This can be clearly seen in the average behaviour of
these ratios over time. Fig. 8 plots the time evolution of the average value
of $E_{\rm B}/|E_{\rm PE}|$ (left) and $E_{\rm B}/E_{\rm KE}$ (right) for all
atomic (red), molecular (blue), and dense molecular (yellow) structures from
the MHD clouds, where we define dense molecular structures to be structures
that are both molecular and have $\rho_{\rm thr}>10^{-20}$ g cm-3. The error
bars here represent the standard error on the mean. From Fig. 8, we see that
magnetic energy dominates over potential and kinetic energies for atomic
structures, while it plays a subordinate role in molecular structures. There
is no clear trend indicating that this behaviour changes as a function of
time.
Figure 8: Time evolution of the average ratio of magnetic to potential energy
(left) and kinetic energy (right). The different colours represent atomic,
molecular, and dense molecular (molecular and $\rho_{\rm thr}>10^{-20}$ g
cm-3) structures in red, blue, and yellow, respectively. The errors bars are
the standard errors on the mean. For denser and molecular structures, magnetic
energy is less important compared to potential or kinetic energies. The atomic
structures, representing more the envelope of the molecular gas, have high
magnetic energies, especially compared to self-gravity.
The subservient role of magnetic energy for dense structures compared to
potential or kinetic energy suggests that while magnetic fields help to shape
the cloud structures across different scales, the dynamics of the denser, and
potentially star-forming structures, is determined by the interaction between
gravity and turbulence 444We explore the interplay between turbulence and
gravity in much greater detail in our companion paper by means of a virial
analysis (Ganguly et al., 2022). This explains why there is no discernible
difference in the power-law tail of the density PDFs between hydrodynamic and
MHD clouds (Fig. 2), confirming that the star-forming gas (see e.g. Klessen &
Burkert, 2001; Girichidis et al., 2014; Schneider et al., 2015) is virtually
unaffected by the presence of magnetic fields. However, magnetic fields change
the gas properties of the environment from which denser structures form,
accrete, and sit in (i.e. by making the surrounding envelope "fluffier"),
thereby also influencing the shape of these structures.
### 6.3 Magnetic surface energy
In the previous section, we have discussed the magnetic pressure term in
comparison to self-gravity and kinetic energy. The magnetic pressure relates
to the stretching and compression of magnetic field lines, and does not take
into account the effect of curvature in the field.
The magnetic surface term can be computed as an integral over the surface of a
given structure, $S$, as follows:
$E_{\rm B}^{\rm
surface}=\oint_{S}(\mathbf{r}-\mathbf{r}_{0})\mathbf{T}\hat{\mathbf{n}}\
\mathrm{d}S.$ (24)
Here $\mathbf{r}_{0}$ is the centre of mass, $\hat{\mathbf{n}}$ is the surface
normal vector that points outwards, and $\mathbf{T}$ is the Maxwell stress
tensor, which can be written as follows for ideal MHD:
$\mathbf{T}=\frac{1}{4\pi}\left(\mathbf{B}\otimes\mathbf{B}-\frac{1}{2}|\mathbf{B}|^{2}\hat{\mathbf{I}}\right).$
(25)
$\hat{\mathbf{I}}$ is here an identity matrix of rank two.
We evaluate Eq. 24 as a volume integral using the Gauss’ divergence theorem
for convenience. This gives us the following relation:
$E_{\rm B}^{\rm surface}=-E_{\rm
B}+\int_{V}(\mathbf{r}-\mathbf{r}_{0})\cdot\nabla\mathbf{T}\ \mathrm{d}V$ (26)
From Eq. 26, we can see that $E_{\rm B}^{\rm surface}$ can be both positive or
negative. When it is is positive, it adds to the magnetic pressure term and
acts as a dispersive term. In contrast, when $E_{\rm B}^{\rm surface}<0$, it
acts as a confining term.
The importance of $E_{\rm B}^{\rm surface}$ with respect to the volume term,
$E_{\rm B}$, can be seen in Fig. 9, left panel, which plots the magnitude of
the ratio of $E_{\rm B}^{\rm surface}/E_{\rm B}$ to the density threshold of
the cloud sub-structures for all MHD clouds at $t_{\rm evol}=3.5$ Myr.
Structures where $E_{\rm B}^{\rm surface}$ helps to disperse them ($E_{\rm
B}^{\rm surface}>0$) are marked in red, while structures where $E_{\rm B}^{\rm
surface}$ acts as a confining term ($E_{\rm B}^{\rm surface}<0$) are marked in
cyan. The vertical dotted line marks the difference between the results of the
low-den and high-den dendrogram runs at $\rho=10^{-22}$ g cm-3, as in the
previous plots. The horizontal dotted line represents a value of one, where
the volume and surface terms are equally important magnitude-wise. The top and
side panels show the marginal distributions.
Figure 9: Left: Ratio of the absolute value of the magnetic surface to volume
energy, plotted against the density threshold. The different colours represent
whether the magnetic surface term is positive and resists collapse or negative
and promotes collapse. The magnetic surface energy seems to be as relevant as
the volume energy, and for more than half of the structures acts as a
confining term. Right: The ratio of the total magnetic energy (surface plus
volume) to the self-gravitating potential energy, plotted against the magnetic
volume energy over the self-gravitating potential energy. The dashed line
represents a 1:1 ratio, and the shaded region represents a factor of 2. For
many small-scale atomic structures, the magnetic surface term seems to be
important as a confining force.
From the marginal distributions, we see that $E_{\rm B}^{\rm surface}$ acts as
a confining term for somewhat more number of structures compared to where the
surface term is dispersive. The magnetic surface term seems to be comparable
to and in some cases, even exceeding the volume term $E_{\rm B}$. This implies
that for diffuse and mostly atomic structures, where magnetic energy is
comparable or dominant, the surface term is important. This is especially
relevant when $E_{\rm B}^{\rm surface}$ acts as a confining term. However, for
dense structures, where $E_{\rm B}$ is one to two orders of magnitude smaller
than the potential and kinetic energies, the surface term is unlikely to
significantly affect the dynamics.
In the right panel of Fig. 9 we plot the magnitude of $(E_{\rm B}^{\rm
surface}+E_{\rm B})/E_{\rm PE}$ against $E_{\rm B}/E_{\rm PE}$ for all MHD
cloud sub-structures at 3.5 Myr. The colour bar here represents the size of
the structures. The horizontal and vertical dotted lines both represent a
value of unity along the $y-$ and $x-$ axes, respectively. The dashed line
represents a 1:1 line, and the shaded region around it represents a factor of
2 in each direction. The magnetic surface energy is not significant compared
to the volume energy for structures on or close to the 1:1 line. Structures
with strong dispersive $E_{\rm B}^{\rm surface}$ terms lie above the 1:1 line,
while points that lie below the 1:1 line represent structures where $E_{\rm
B}^{\rm surface}$ is confining in nature. Most interesting here are the points
that lie in the bottom right quadrant of the plot. They represent structures
where the magnetic pressure $E_{\rm B}$ is higher compared to the self-
gravity, and would be completely unbound in a traditional virial analysis.
However, the confining $E_{\rm B}^{\rm surface}$ term is strong enough that
the overall magnetic contribution becomes far less, thus allowing for a sort
of "magnetic confinement". These structures are mostly atomic and typically
seem to be $\lessapprox 1$ pc.
Two examples of structures belonging to MC2-MHD that exhibit such magnetic
confinement are plotted in Fig. 10 as black contour lines over a density slice
in the $y-z$ plane. The background colour here represents the density, while
the planar magnetic field is shown using the line integral convolution (LIC)
technique555The package used can be found in
https://github.com/alexus37/licplot.. For both structures, we mention the
magnitude of the $E_{\rm B}/E_{\rm PE}$ and $(E_{\rm B}^{\rm surface}+E_{\rm
B})/E_{\rm PE}$ ratios in the figure title. As can be clearly seen, the
magnetic surface term reduces the $|(E_{\rm B}^{\rm surface}+E_{\rm B})/E_{\rm
PE}|$ ratio to less than one. However, this naturally does not take into
account other energy terms, i.e. kinetic and thermal energy, and hence it is
not fully clear whether these structures are overall confined. Interestingly,
the structures for which the magnetic surface energy is important and of
confining nature (see right panel of Fig. 9) are usually located at the
"kinks" of magnetic field lines.
Figure 10: Two examples of structures confined by $E_{\rm B}^{\rm surface}$
from MC2-MHD, plotted as black contours over density slices in the $y-z$
plane, at $t_{\rm evol}=2$ Myr. The colour map is the logarithmic density, and
the direction of the planar magnetic field is plotted as line integral
convolution. The relevant energy ratios of the indicated structures are
denoted in the title. Structures for which the magnetic surface energy is
important and of confining nature, are usually located at the "kinks" of
magnetic field lines.
### 6.4 Fragmentation
In this section, we attempt to quantify to what extent magnetic fields affect
the fragmentation properties of molecular clouds. For this purpose, we study
the numbers and masses of different fragments, represented by leaf structures
(i.e. structures containing no further sub-structures) found in our dendrogram
analysis, and in addition perform a magnetic Jeans analysis on these
fragments.
#### 6.4.1 Number and mass distribution of fragments
Representing fragments by the leaves in the dendrogram analysis suffers from
the caveat of depending on the dendrogram parameters. Increasing the minimum
number of cells required in a dendrogram structure, for example, would
naturally reduce the number of fragments and increase their masses. The
absolute values of the masses and numbers we find, therefore, are sensitive to
the parameter values we have used. However, since we used the exact same
parameters for each HD and MHD run, and because all molecular clouds have
similar masses and identical environmental parameters (solar neighbourhood
parameters), the relative difference between the average behaviour of the HD
and MHD clouds is meaningful. With this caveat in mind, let us look at the
fragmentation properties of our dendrogram structures.
Figure 11: Top row: Cumulative distribution of the average number of leaf
structures against $\rho_{\rm thr}$ for HD and MHD clouds at $t_{\rm evol}=$
2, 2.5, 3.5 Myr, respectively (from left to right). The hydrodynamic clouds
have on average more new structures forming at earlier times, but this
distinction slowly disappears later on. Bottom row: Distribution of average
mass of leaf structures for both HD and MHD clouds at $t_{\rm evol}=$ 2, 2.5,
3.5 Myr, respectively (from left to right). The leaf structures, representing
fragments, are more massive for MHD clouds at earlier times, while this
distinction mostly disappears later on as gravity takes over.
We study the numbers and masses of leaf fragments in Fig. 11. The top row
plots the cumulative distribution of the average number of leaf structures,
$\langle N_{\rm structure}^{\rm leaf}\rangle$, as a function of $\rho_{\rm
thr}$ for both HD (blue) and MHD (red) clouds. The average here simply means
that we divide the total number of obtained structures by the number of
clouds, i.e. 5 for MHD and 2 for HD. The three panels (left to right) show
three different times, $t_{\rm evol}=2,\ 2.5$ and 3.5 Myr, respectively. The
vertical line at $10^{-22}$ g cm-3 marks the difference between the low-den
and the high-den dendrogram analysis. We see that at $t_{\rm evol}=2$ and 2.5
Myr, up to densities between $10^{-23}-10^{-22}$ g cm-3, the HD and MHD clouds
form roughly similar numbers of leaf fragments. However, at higher densities,
$\langle N_{\rm structure}^{\rm leaf}\rangle$ is much higher for the HD
clouds. This difference largely disappears at 3.5 Myr. This suggests that the
formation of structures is somewhat slowed down in the presence of magnetic
fields in the beginning, but at later stages, as gravity becomes dynamically
more and more important, this difference diminishes.
In the bottom row of Fig. 11 we plot the average mass of the leaf structures,
$\langle M_{\rm structure}^{\rm leaf}\rangle$, as a function $\rho_{\rm thr}$
for HD and MHD structures for the three different times. The shaded regions
represent the standard deviation of the average mass at a given $\rho_{\rm
thr}$. We see that at $t_{\rm evol}=2$ Myr, the MHD fragments are slightly
more massive compared to their hydrodynamic counterparts, in particular for
$\rho_{\rm thr}\lesssim 10^{-21}$ g cm-3. This difference disappears later.
For the densest structures, we do not seem to see a systematic difference in
$\langle M_{\rm structure}^{\rm leaf}\rangle$. This is in line with Fig. 2,
which shows that the difference in the density PDFs between the HD and MHD
clouds in the density range that corresponds primarily to the cloud envelope
(i.e. $\lessapprox 10^{-21}$ g cm-3) is most striking at $t_{\rm evol}=2$ Myr,
and less so later on.
Overall, the results shown in Fig. 11 indicate that the MHD clouds fragment
more slowly than the HD clouds but therefore have slightly more massive
fragments at early times. This is consistent with the result that magnetic
fields affect the dynamics of lower density gas more (Molina et al., 2012;
Seifried et al., 2020a, b; Ibáñez-Mejía et al., 2022). We also see that the
number and mass of the leaf structures are comparable at later times. This
suggests that the magnetic fields "slow down" the evolution of the cloud but
are less relevant once the cloud is more evolved, and gravity becomes
energetically more and more important, as shown in the previous energetic
analysis. This effect could be related to the overall strength of the magnetic
field. We investigate this further in the next Section.
#### 6.4.2 Magnetic Jeans analysis
The classic thermal Jeans analysis (Jeans, 1902) is a useful tool to
investigate the stability of MCs and their substructures (clumps and cores)
under thermal perturbations. Here, we perform its magnetic equivalent. The
thermal Jeans length, $\lambda_{\rm T}$, defines the largest length-scale
stable to thermal perturbations. For a given structure, this is defined as
$\lambda_{\rm T}=c_{s}\sqrt{\frac{\pi}{G\rho_{\rm avg}}},$ (27)
where $c_{s}$ is the average sound speed given by
$\displaystyle c_{s}=\frac{1}{V}\int_{V}\sqrt{\frac{P}{\rho}}\mathrm{d}^{3}r.$
(28)
Here, $P$ is the thermal pressure and the sound speed is calculated assuming
an isothermal equation of state due to the densities under consideration. We
remind the reader that $\rho_{\rm avg}$ is the volume-averaged density
computed in Eq. 19. From the Jeans length, a maximum mass stable under thermal
perturbations can be calculated. This mass is referred to as the thermal Jeans
mass, $M_{\rm T}$, and is given by
$M_{\rm T}=\frac{4}{3}\pi\rho_{\rm avg}\left(\frac{\lambda_{\rm
T}}{2}\right)^{3}.$ (29)
Similar to the thermal analysis, we can perform a magnetic Jeans analysis and
a Jeans analysis combining both magnetic and thermal support. For the magnetic
Jeans analysis, the relevant length ($\lambda_{\rm B}$) and mass ($M_{\rm B}$)
scales are given by,
$\displaystyle\lambda_{\rm B}=c_{\rm B}\sqrt{\frac{\pi}{G\rho_{\rm avg}}},$
(30) $\displaystyle M_{\rm B}=\frac{4}{3}\pi\rho_{\rm
avg}\left(\frac{\lambda_{\rm B}}{2}\right)^{3}.$ (31)
For a combination of thermal and magnetic effects, the relevant magneto-
thermal Jeans length ($\lambda_{\rm B,T}$) and Jeans mass ($M_{\rm B,T}$) are
$\displaystyle\lambda_{\rm B,T}=c_{\rm B,T}\sqrt{\frac{\pi}{G\rho_{\rm
avg}}},$ (32) $\displaystyle M_{\rm B,T}=\frac{4}{3}\pi\rho_{\rm
avg}\left(\frac{\lambda_{\rm B,T}}{2}\right)^{3}.$ (33)
The characteristic speeds are given by,
$\displaystyle c_{\rm B}=v_{\rm A},$ (34) $\displaystyle c_{\rm
B,T}=\sqrt{c_{s}^{2}+v_{\rm A}^{2}},$ (35)
where $v_{\rm A}$ is the Alfvén speed (Eq. 18).
In Fig. 12, we show the ratio of a structure’s mass to its magneto-thermal
Jeans mass, $M/M_{\rm B,T}$, as a function of $\rho_{\rm thr}$ for all MHD
cloud branch structures (top) and leaves (bottom) at $t_{\rm evol}=3.5$ Myr.
We remind the reader that branch structures contain sub-structures and leaves
do not. The Jeans mass can only be properly used when the corresponding length
is resolved. This is shown in Appendix E, Fig. 20, which depicts that some
structures with $\rho_{\rm thr}\gtrapprox 10^{-20}$ g cm-3 seem to be not
properly Jeans resolved. These are marked with black outlines in Fig. 12. Note
that the structures are (un-)resolved in the context of our dendrogram
analysis, which requires a minimum number of 100 cells per structure, and
therefore at least 200 cells to resolve fragmentation (as in the case of
fragmentation, each fragment would need to contain at least 100 cells). The
colour-bar denotes the ratio of $c_{s}$ to $v_{\rm A}$. Most of the structures
have $v_{\rm A}>c_{s}$ (blue points), suggesting support by magnetic fields
rather than by thermal pressure. This is confirmed by our purely magnetic
Jeans analysis (Fig. 21), which shows an almost identical distribution to the
magneto-thermal Jeans analysis of Fig. 12 (as well as Fig. 20). From Fig. 12,
top panel, we find that roughly below $10^{-22}$ g cm-3, all structures are
Jeans stable ($M/M_{\rm B,T}<1$). At higher densities, we have both, Jeans
stable and unstable structures. Some prominent branches clearly have $M/M_{\rm
B,T}>1$ above $10^{-22}$ g cm-3, indicating the growing importance of gravity
for fragmentation at higher densities. For leaves, this transition density
seems to occur at higher densities.
Figure 12: The ratio of the mass of a given structure to its magneto-thermal
Jeans mass ($M_{\rm B,T}$, Eq. 33) as a function of $\rho_{\rm thr}$ for all
MHD branch (top) and leaf (bottom) sub-structures at $t_{\rm evol}=3.5$ Myr. A
branch sub-structure has further sub-structures, while a leaf does not. The
horizontal dotted line represents a ratio of unity. The vertical line
separates the points obtained from the high-den and the low-den dendrogram
analysis. The colour-bar shows the ratio of the sound speed to the Alfvén wave
speed. For blue points, $v_{\rm A}>c_{s}$. A power-law is plotted in each
panel for rough guidance. Structures whose fragmentation is not well-resolved
(see Fig. 20) are marked with an additional black outline and mostly populate
the right-hand top corner of the plot. The magneto-thermal forces seem unable
to keep all the structures Jeans stable beyond $\sim 10^{-22}$ g cm-3,
suggesting the growing importance of gravity.
Interestingly, the leaves seem to have an overall sharper scaling behaviour
compared to the branches. However, this separation cannot be seen in the Jeans
length, where all structures show a consistent scaling of roughly
$\lambda_{\rm B,T}\propto\rho^{-2/3}$ (Fig. 20). This can be understood as
follows:
The mass of a structure is dependent on the density and size, i.e.
$M\propto\rho R^{3}.$ (36)
Combining Eq. 36 with Eq. 33, we obtain
$\frac{M}{M_{\rm B,T}}\propto R^{3}\lambda_{\rm B,T}^{-3}.$ (37)
As the size of the leaf structures is more determined by the choice of $N_{\rm
cells}$, they typically show very weak or no scaling between density and size,
and we can therefore approximate $R\propto\rho^{0}$. For the leaves, this
leads to $M/M_{\rm B,T}\propto\lambda_{\rm B,T}^{-3}$. As $\lambda_{\rm
B,T}\propto\rho^{-2/3}$ approximately (Fig. 20), this leads to $M/M_{\rm
B,T}\propto\rho^{2}$. For the branches, we find a shallower slope. In Ganguly
et al. (2022), we find many branches to follow $M\propto R$, which would lead
to $M/M_{\rm B,T}\propto\rho^{1/2}$, roughly consistent with the trend seen
for the branches here. The relation of the scaling between $\lambda_{\rm B,T}$
and $\rho$ is in itself interesting and we discuss it in Appendix F.
Overall, the Jeans analysis seems to show the emergence of potentially Jeans-
unstable structures at slightly lower densities ($\sim 30$ cm-3) compared to
that found in the previous energetic analysis. This could reflect the fact
that the Jeans analysis performed here does not include the kinetic energy,
which is often larger compared to $E_{\rm B}$ and the thermal energy (see
Section 4 in Ganguly et al., 2022). Turbulent motions can act as an effective
kinetic pressure term. Although the kinetic energy is often treated as an
effective pressure in the literature (see e.g. Chandrasekhar, 1951; Bonazzola
et al., 1987; Federrath & Klessen, 2012), we show in Ganguly et al. (2022)
that the volume and surface terms of the kinetic energy combine in a highly
non-trivial manner, with structures often being confined or even compressed
under ram pressure. This suggests that including a kinetic pressure in the
Jeans analysis would be too simplistic and not meaningful.
Overall, most leaf fragments in the Jeans analysis have $M/M_{\rm B,T}<1$,
suggesting that their fragmentation is unlikely to be primarily Jeans-like.
However, above 10-20 g cm-3, we begin to obtain Jeans unstable fragments which
are mostly unresolved and will likely undergo further fragmentation, possibly
ending up as the precursors of protostars.
### 6.5 Delay introduced by magnetic fields
The fragmentation analysis performed in 6.4 seems to suggest that magnetic
fields at least delay fragmentation in many cases. To estimate how much the
evolution of the cloud is slowed down by the effect of magnetic fields, we
define a delay timescale $\Delta t_{\rm B}$. Consider a structure of size $S$
which is compressed by an external flow with velocity $v$. In the absence of
either gravity or magnetic fields, as well as neglecting internal thermal and
kinetic pressure, the structure would be compressed on a crossing time,
$t_{\rm v}=S/v.$ (38)
For simplicity, we estimate the size of a structure using the shortest axis of
the equivalent ellipsoid, i.e. $S=2c$ (Section 3.2). We approximate the sweep-
up velocity $v$ to be equal to the bulk velocity of the structure, i.e.
$v=|\mathbf{v}_{0}|$, where $\mathbf{v}_{0}$ is obtained from Eq. 17. So
overall we have
$t_{\rm v}=2c/|\mathbf{v}_{0}|.$ (39)
Next, we consider an additional gravitational acceleration $a_{\rm g}$
assisting the sweeping up, where
$a_{\rm
g}=-\frac{1}{V}\int_{V}\mathbf{g}\cdot\frac{\mathbf{r}-\mathbf{r}_{0}}{|\mathbf{r}-\mathbf{r}_{0}|}d^{3}r$
(40)
is the average acceleration towards the centre of mass, $\mathbf{r}_{0}$. We
can then estimate (to first order) the gravitationally assisted sweep-up
timescale, $t_{\rm v,\ g}$, from
$S=vt_{\rm v,\ g}+\frac{1}{2}a_{\rm g}t_{\rm v,\ g}^{2}.$ (41)
For non-gravitating structures, this reduces to $t_{\rm v}$. For non-zero
gravitational field, taking the real root, we get
$t_{\rm v,g}=\frac{(-v+\sqrt{v^{2}+2Sa_{\rm g}})}{a_{\rm g}}.$ (42)
In the presence of magnetic fields, we can represent the combined acceleration
by gravity and magnetic fields as $a_{\rm g,B}$, where
$a_{\rm
g,B}=-\frac{1}{V}\int_{V}\left(\mathbf{g}-\frac{\nabla|\mathbf{B}|^{2}}{8\pi\rho}\right)\cdot\frac{\mathbf{r}-\mathbf{r}_{0}}{|\mathbf{r}-\mathbf{r}_{0}|}d^{3}r.$
(43)
We can then rewrite Eq. 42 to estimate a combined timescale
$t_{\rm v,g,B}=\frac{(-v+\sqrt{v^{2}+2Sa_{\rm g,B}})}{a_{\rm g,B}}.$ (44)
The time delay due to the presence of magnetic fields, $\Delta t_{\rm B}$, can
then be estimated as
$\Delta t_{\rm B}=t_{\rm v,g,B}-t_{\rm v,g}.$ (45)
Figure 13: The estimated delay timescale, $\Delta t_{\rm B}$ (Eq. 45), for
various MHD cloud structures from the high-den analysis at $t_{\rm evol}=2$
Myr. A power-law proportional to $R^{3}$ is plotted to show the rough scaling.
In Fig. 13, we plot $\Delta t_{\rm B}$ for various structures from the high-
den dendrogram analysis. We see that at the largest cloud scales at $t_{\rm
evol}=2$ Myr, $\Delta t_{\rm B}$ is of the order of $\sim 1$ Myr, and then
steadily decreases as a power-law roughly consistent with $\Delta t_{\rm
B}\propto R^{3}$. This timescale of $\sim 1$ Myr seems to be consistent with
the results of the fragmentation analysis in Section 6.4, where we found that
the significant differences in the cloud fragmentation properties at $t_{\rm
evol}=2$ Myr seem to have completely disappeared at $t_{\rm evol}=3.5$ Myr.
The general power-law trend We emphasise, however, that the calculation of
$\Delta t_{\rm B}$ should only be considered a first-order approximation. Note
that $\Delta t_{\rm B}$ does not directly depend on the magnetic field
strength but rather on its gradient. Hence, it is difficult to predict how
$\Delta t_{\rm B}$ would scale with different strengths of the background
field. This means that molecular clouds that form in a more magnetised medium
do not necessarily form structures more slowly.
### 6.6 Densities at which magnetic fields become dynamically sub-dominant
From the results presented in the previous sections, we can attempt to answer
the question of at what densities magnetic fields become dynamically sub-
dominant. From the density PDF of different clouds (Fig. 2), we find that the
density distribution is significantly different in the presence of magnetic
fields only below $\sim$ 100 cm-3. This is in accordance with previous
simulations and observations (Klessen & Burkert, 2001; Slyz et al., 2005;
Girichidis et al., 2014; Schneider et al., 2015), as well as the conclusions
drawn by Seifried et al. (2020a) using distributions of the three-dimensional
true optical depth, $A_{\rm V,3D}$. From the energetic analysis (Fig. 7), we
find that, magnitude-wise, gravity and kinetic energy supersede magnetic
fields above a few $\sim 100$ cm-3, consistent with the results of Ibáñez-
Mejía et al. (2022). Moreover, this density range is also in accordance with
the results of Seifried et al. (2020b), who find that relative orientation of
magnetic fields with respect to elongated filamentary structures changes at a
few $\sim 100$ cm-3 due to the occurrence of gravity-driven converging flows
(Soler & Hennebelle, 2017), suggesting energetic sub-dominance of magnetic
fields at higher densities. Lastly, also the fragmentation analysis presented
in this work (Fig. 11) shows differences in fragmentation patterns below a
similar density regime of $\sim 100$ cm-3. A Jeans fragmentation analysis
yields roughly consistent limits as well.
In summary, for clouds born from an ISM with typical magnetic field strengths
as in our Milky Way (Beck & Wielebinski, 2013), the density PDFs, the
energetic analysis, the histogram of relative orientation technique applied by
Seifried et al. (2020b), and the fragmentation analysis in this work - all
seem to point to the fact that the magnetic field becomes sub-dominant above
densities of around $100-1000$ cm-3. This overall trend is also fully
consistent with the $B-\rho$ relation obtained by Crutcher et al. (2010), who
conclude a transition density of $\sim 300$ cm-3.
## 7 Conclusions
We investigate the role magnetic fields play in determining the morphology,
energetics, and fragmentation properties of young molecular clouds by
analysing seven different simulated clouds (five with magnetic fields and two
without) from the SILCC-Zoom simulations. These simulations are geared to
study the evolution of the multi-phase interstellar medium in a supernova-
driven, turbulent, stratified galactic disc environment. To identify forming
structures, we use a dendrogram algorithm, and trace the statistical
properties of the identified structures. We include a simple chemical network
which allows us to follow the formation of H2 as the cloud assembles and
thereby distinguish between mostly atomic (H2 mass fraction < 50%) and mostly
molecular (H2 mass fraction > 50%) structures.
* •
We observe that the MHD clouds are fluffier, meaning that they have more
intermediate density gas between the number densities of roughly $1-100$ cm-3,
compared to their hydrodynamic counterparts. In the hydrodynamic clouds, the
lack of magnetic fields results in the denser structures being surrounded by a
comparatively more rarefied envelope.
* •
In terms of morphology, we find that almost all clouds are sheet-like, which
is consistent with recent observations of sheet-like envelopes around denser
filamentary cloud structures (Kalberla et al., 2016; Arzoumanian et al., 2018;
Rezaei Kh. & Kainulainen, 2022; Tritsis et al., 2022; Pineda et al., 2022;
Clarke et al., 2023). In our case, the MCs form due to compressions caused by
expanding supernova shells, consistent with the bubble-driven MC formation
scenario (Koyama & Inutsuka, 2000; Inoue & Inutsuka, 2009; Inutsuka et al.,
2015).
* •
We find that spheroidal structures within the clouds are rare on all spatial
scales, with $\sim 90$% of the structures being elongated. We further see that
the runs with magnetic fields have a roughly comparable fraction of filaments
and sheets, whereas the hydrodynamic runs overall produce more sheet-like
structures compared to filaments.
* •
Energetically, magnetic fields in our simulations are important for less dense
(up to $\sim$1000 cm-3) and mostly, but not exclusively, atomic structures.
The dynamics for denser and potentially star-forming structures is dominated
by the interplay of turbulence and gravity. This density threshold, above
which the magnetic fields seems to become sub-dominant, is supported by the
previous works of Seifried et al. (2020b), Ibáñez-Mejía et al. (2022) and is
consistent with the observed transition in the $B-\rho$ relation (Crutcher et
al., 2010).
* •
By investigating the magnetic surface energy term, we find that for most
structures it acts in a confining manner, and, for some low-density
structures, it even leads to overall magnetic confinement.
* •
By studying the numbers and masses of cloud fragments that form, we find that
at densities below roughly $\sim 100$ cm-3, the presence of magnetic fields
helps to create more massive fragments, but generally does not result in an
increase in the number of such structures. A stability analysis suggests that
in the resolved range, leaf fragments are mostly Jeans stable and the
fragmentation is not primarily governed by magnetic Jeans instabilities.
Instead of significantly altering the nature of fragmentation, magnetic fields
seem to rather slow down the fragmentation process. Using a simple order-of-
magnitude estimate, we find that this delay timescale is $\sim 1$ Myr.
Overall, using density PDFs, and an energetic as well as a fragmentation
analysis, we find a scenario where magnetic fields significantly affect the
flows and fragmentation in the lower density gas (below $\sim 100$ cm-3),
channelling flows and thereby affecting both, the morphology of the forming
structures as well as the formation timescale of the dense gas. Once the dense
structures (typically above $\sim 1000$ cm-3) form, however, the further
evolution and fragmentation of the dense gas seems to be mostly unaffected by
the magnetic field.
## Acknowledgements
We would like to thank the referee, Prof. Dr. Robi Banerjee, for their helpful
comments, suggestions, and overall discussion, which have increased the
quality of the paper. SG, SW, DS and MW would like to acknowledge the support
of Bonn-Cologne Graduate School (BCGS), which is funded through the German
Excellence Initiative, as well as the DFG for funding through SFB 956
’Conditions and Impact of Star Formation’ (subprojects C5 and C6). SDC is
supported by the Ministry of Science and Technology (MoST) in Taiwan through
grant MoST 108-2112-M-001-004-MY2. This research made use of astrodendro, a
Python package to compute dendrograms of Astronomical data
(http://www.dendrograms.org/); as well as yt, an open-source, permissively-
licensed python package for analyzing and visualizing volumetric data
(https://yt-project.org/). The 3D renderings in Fig. 3 were computed using
paraview. The FLASH code used in this work was partly developed by the Flash
Center for Computational Science at the University of Rochester.
## Data Availability
The data underlying this article can be shared for selected scientific
purposes after request to the corresponding author.
## References
* Alves et al. (2017) Alves J., Lombardi M., Lada C. J., 2017, A&A, 606, L2
* André et al. (2014) André P., Di Francesco J., Ward-Thompson D., Inutsuka S. I., Pudritz R. E., Pineda J. E., 2014, in Beuther H., Klessen R. S., Dullemond C. P., Henning T., eds, Protostars and Planets VI. p. 27 (arXiv:1312.6232), doi:10.2458/azu_uapress_9780816531240-ch002
* Arzoumanian et al. (2018) Arzoumanian D., Shimajiri Y., Inutsuka S.-i., Inoue T., Tachihara K., 2018, Publications of the Astronomical Society of Japan, 70, 96
* Auddy et al. (2022) Auddy S., Basu S., Kudoh T., 2022, The Astrophysical Journal, 928, L2
* Bally et al. (1987) Bally J., Stark A. A., Wilson R. W., Henkel C., 1987, The Astrophysical Journal Supplement Series, 65, 13
* Beck (2009) Beck R., 2009, Astrophysics and Space Sciences Transactions, 5, 43
* Beck (2015) Beck R., 2015, A&ARv, 24, 4
* Beck & Wielebinski (2013) Beck R., Wielebinski R., 2013, Planets, Stars and Stellar Systems. Volume 5: Galactic Structure and Stellar Populations, p. 641
* Bentley (1975) Bentley J. L., 1975, Communications of the ACM, 18, 509
* Bonazzola et al. (1987) Bonazzola S., Heyvaerts J., Falgarone E., Perault M., Puget J. L., 1987, A&A, 172, 293
* Bouchut et al. (2007) Bouchut F., Klingenberg C., Waagan K., 2007, Numerische Mathematik, 108, 7
* Bourke et al. (2001) Bourke T. L., Myers P. C., Robinson G., Hyland A. R., 2001, The Astrophysical Journal, 554, 916
* Burkert & Hartmann (2004) Burkert A., Hartmann L., 2004, The Astrophysical Journal, 616, 288
* Camacho et al. (2022) Camacho V., Vázquez-Semadeni E., Zamora-Avilés M., Palau A., 2022, arXiv e-prints, p. arXiv:2212.10610
* Chandrasekhar (1951) Chandrasekhar S., 1951, Proceedings of the Royal Society of London Series A, 210, 26
* Clarke et al. (2023) Clarke S. D., Sanchez-Monge A., Williams G. M., Howard A. D. P., Walch S., Schneider N., 2023, Monthly Notices of the Royal Astronomical Society, 519, 3098
* Commerçon et al. (2011) Commerçon B., Hennebelle P., Henning T., 2011, The Astrophysical Journal, 742, L9
* Crutcher (1999) Crutcher R. M., 1999, The Astrophysical Journal, 520, 706
* Crutcher (2012) Crutcher R. M., 2012, Annual Review of Astronomy and Astrophysics, vol. 50, p.29-63, 50, 29
* Crutcher et al. (2003) Crutcher R., Heiles C., Troland T., 2003, Turbulence and Magnetic Fields in Astrophysics, 614, 155
* Crutcher et al. (2010) Crutcher R. M., Wandelt B., Heiles C., Falgarone E., Troland T. H., 2010, The Astrophysical Journal, 725, 466
* Derigs et al. (2016) Derigs D., Winters A. R., Gassner G. J., Walch S., 2016, Journal of Computational Physics, 317, 223
* Derigs et al. (2018) Derigs D., Winters A. R., Gassner G. J., Walch S., Bohm M., 2018, Journal of Computational Physics, 364, 420
* Draine (1978) Draine B. T., 1978, ApJS, 36, 595
* Dubey et al. (2008) Dubey A., et al., 2008, in Pogorelov N. V., Audit E., Zank G. P., eds, Astronomical Society of the Pacific Conference Series Vol. 385, Numerical Modeling of Space Plasma Flows. p. 145
* Federrath (2016) Federrath C., 2016, Monthly Notices of the Royal Astronomical Society, 457, 375
* Federrath & Klessen (2012) Federrath C., Klessen R. S., 2012, The Astrophysical Journal, 761, 156
* Fletcher et al. (2011) Fletcher A., Beck R., Shukurov A., Berkhuijsen E. M., Horellou C., 2011, Monthly Notices of the Royal Astronomical Society, 412, 2396
* Fryxell et al. (2000) Fryxell B., et al., 2000, The Astrophysical Journal Supplement Series, 131, 273
* Ganguly et al. (2022) Ganguly S., Walch S., Clarke S. D., Seifried D., 2022, arXiv e-prints, p. arXiv:2204.02511
* Girichidis et al. (2014) Girichidis P., Konstandin L., Whitworth A. P., Klessen R. S., 2014, The Astrophysical Journal, 781, 91
* Girichidis et al. (2016a) Girichidis P., et al., 2016a, Monthly Notices of the Royal Astronomical Society, 456, 3432
* Girichidis et al. (2016b) Girichidis P., et al., 2016b, ApJ, 816, L19
* Girichidis et al. (2018) Girichidis P., Seifried D., Naab T., Peters T., Walch S., Wünsch R., Glover S. C. O., Klessen R. S., 2018, Monthly Notices of the Royal Astronomical Society, 480, 3511
* Girichidis et al. (2020) Girichidis P., et al., 2020, Space Science Reviews, 216, 68
* Glover & Mac Low (2007) Glover S. C. O., Mac Low M.-M., 2007, The Astrophysical Journal Supplement Series, 169, 239
* Glover et al. (2010) Glover S. C. O., Federrath C., Mac Low M.-M., Klessen R. S., 2010, Monthly Notices of the Royal Astronomical Society, 404, 2
* Habing (1968) Habing H. J., 1968, Bulletin of the Astronomical Institutes of the Netherlands, 19, 421
* Hacar et al. (2022) Hacar A., Clark S., Heitsch F., Kainulainen J., Panopoulou G., Seifried D., Smith R., 2022, arXiv e-prints, p. arXiv:2203.09562
* Hall (1951) Hall J. S., 1951, The Astronomical Journal, 56, 40
* Heiles & Crutcher (2005) Heiles C., Crutcher R., 2005, Cosmic Magnetic Fields, 664, 137
* Heiles & Troland (2005) Heiles C., Troland T. H., 2005, The Astrophysical Journal, 624, 773
* Heitsch et al. (2001) Heitsch F., Zweibel E. G., Mac Low M.-M., Li P., Norman M. L., 2001, The Astrophysical Journal, 561, 800
* Hennebelle (2013) Hennebelle P., 2013, Astronomy and Astrophysics, 556, A153
* Hennebelle & Inutsuka (2019) Hennebelle P., Inutsuka S.-i., 2019, Frontiers in Astronomy and Space Sciences, 6, 5
* Hill et al. (2012) Hill A. S., Joung M. R., Mac Low M.-M., Benjamin R. A., Haffner L. M., Klingenberg C., Waagan K., 2012, The Astrophysical Journal, 750, 104
* Hiltner (1951) Hiltner W. A., 1951, The Astrophysical Journal, 114, 241
* Ibáñez-Mejía et al. (2022) Ibáñez-Mejía J. C., Mac Low M.-M., Klessen R. S., 2022, The Astrophysical Journal, 925, 196
* Inoue & Inutsuka (2009) Inoue T., Inutsuka S.-i., 2009, The Astrophysical Journal, 704, 161
* Inoue & Inutsuka (2016) Inoue T., Inutsuka S.-i., 2016, The Astrophysical Journal, 833, 10
* Inutsuka et al. (2015) Inutsuka S.-i., Inoue T., Iwasaki K., Hosokawa T., 2015, Astronomy and Astrophysics, 580, A49
* Jeans (1902) Jeans J. H., 1902, Philosophical Transactions of the Royal Society of London Series A, 199, 1
* Kalberla et al. (2016) Kalberla P. M. W., Kerp J., Haud U., Winkel B., Ben Bekhti N., Flöer L., Lenz D., 2016, The Astrophysical Journal, 821, 117
* Kennicutt (1998) Kennicutt Jr. R. C., 1998, The Astrophysical Journal, 498, 541
* Klessen & Burkert (2001) Klessen R. S., Burkert A., 2001, The Astrophysical Journal, 549, 386
* Koyama & Inutsuka (2000) Koyama H., Inutsuka S.-I., 2000, The Astrophysical Journal, 532, 980
* Körtgen & Banerjee (2015) Körtgen B., Banerjee R., 2015, Monthly Notices of the Royal Astronomical Society, 451, 3340
* Körtgen et al. (2018) Körtgen B., Banerjee R., Pudritz R. E., Schmidt W., 2018, Monthly Notices of the Royal Astronomical Society, 479, L40
* Li & Henning (2011) Li H.-B., Henning T., 2011, Nature, 479, 499
* Li et al. (2015) Li P. S., McKee C. F., Klein R. I., 2015, Monthly Notices of the Royal Astronomical Society, 452, 2500
* Lopez-Rodriguez et al. (2023) Lopez-Rodriguez E., et al., 2023, The Astrophysical Journal, 942, L13
* Mocz et al. (2017) Mocz P., Burkhart B., Hernquist L., McKee C. F., Springel V., 2017, The Astrophysical Journal, 838, 40
* Molina et al. (2012) Molina F. Z., Glover S. C. O., Federrath C., Klessen R. S., 2012, Monthly Notices of the Royal Astronomical Society, 423, 2680
* Mouschovias (1991) Mouschovias T. C., 1991, The Astrophysical Journal, 373, 169
* Nelson & Langer (1997) Nelson R. P., Langer W. D., 1997, The Astrophysical Journal, 482, 796
* Padoan & Nordlund (2011) Padoan P., Nordlund A., 2011, The Astrophysical Journal, 730, 40
* Pardi et al. (2017) Pardi A., et al., 2017, Monthly Notices of the Royal Astronomical Society, 465, 4611
* Pattle et al. (2022) Pattle K., Fissel L., Tahani M., Liu T., Ntormousi E., 2022, arXiv e-prints, p. arXiv:2203.11179
* Pineda et al. (2022) Pineda J. E., et al., 2022, arXiv e-prints, p. arXiv:2205.03935
* Planck Collaboration et al. (2020) Planck Collaboration P., et al., 2020, Astronomy and Astrophysics, 641, A12
* Rezaei Kh. & Kainulainen (2022) Rezaei Kh. S., Kainulainen J., 2022, The Astrophysical Journal, 930, L22
* Rosolowsky et al. (2008) Rosolowsky E. W., Pineda J. E., Kauffmann J., Goodman A. A., 2008, The Astrophysical Journal, 679, 1338
* Schmidt (1959) Schmidt M., 1959, The Astrophysical Journal, 129, 243
* Schneider et al. (2015) Schneider N., et al., 2015, A&A, 578, A29
* Seifried et al. (2017) Seifried D., et al., 2017, Monthly Notices of the Royal Astronomical Society, 472, 4797
* Seifried et al. (2019) Seifried D., Walch S., Reissl S., Ibáñez-Mejía J. C., 2019, Monthly Notices of the Royal Astronomical Society, 482, 2697
* Seifried et al. (2020a) Seifried D., Haid S., Walch S., Borchert E. M. A., Bisbas T. G., 2020a, Monthly Notices of the Royal Astronomical Society, 492, 1465
* Seifried et al. (2020b) Seifried D., Walch S., Weis M., Reissl S., Soler J. D., Klessen R. S., Joshi P. R., 2020b, Monthly Notices of the Royal Astronomical Society, 497, 4196
* Slyz et al. (2005) Slyz A. D., Devriendt J. E. G., Bryan G., Silk J., 2005, Monthly Notices of the Royal Astronomical Society, 356, 737
* Soler & Hennebelle (2017) Soler J. D., Hennebelle P., 2017, Astronomy and Astrophysics, 607, A2
* Spitzer (1942) Spitzer Jr. L., 1942, The Astrophysical Journal, 95, 329
* Spitzer (1978) Spitzer L., 1978, Physical processes in the interstellar medium. John Wiley & Sons, Ltd, doi:10.1002/9783527617722
* Tritsis et al. (2022) Tritsis A., Bouzelou F., Skalidis R., Tassis K., Enßlin T., Edenhofer G., 2022, Monthly Notices of the Royal Astronomical Society
* Troland & Crutcher (2008) Troland T. H., Crutcher R. M., 2008, The Astrophysical Journal, 680, 457
* Troland & Heiles (1986) Troland T. H., Heiles C., 1986, The Astrophysical Journal, 301, 339
* Waagan (2009) Waagan K., 2009, Journal of Computational Physics, 228, 8609
* Walch et al. (2015) Walch S., et al., 2015, Monthly Notices of the Royal Astronomical Society, 454, 238
* Wünsch et al. (2018) Wünsch R., Walch S., Dinnbier F., Whitworth A., 2018, Monthly Notices of the Royal Astronomical Society, 475, 3393
* Zhang et al. (2019) Zhang Y., Guo Z., Wang H. H., Li H. b., 2019, The Astrophysical Journal, 871, 98
## Appendix A Basic information of clouds
We present here some basic properties of the different analysed molecular
clouds. Fig. 14 plots the column density projections of all different clouds,
both HD and MHD. Fig. 15 plots the total and H2 mass of the different MCs in
the left panel, and the H2 mass fraction in the right panel. We see that there
is no difference in the overall mass of the clouds depending on the magnetic
field, but that the H2 mass fraction in the HD clouds is higher. The cloud
MC3-MHD (cyan line), which has been excluded from this analysis, has the
lowest total H2 mass, as well as the lowest H2 mass fraction.
Figure 14: Column density projection along the x axis for different molecular
clouds at t${}_{\rm evol}=3.5$ Myr. The MHD clouds have typically more diffuse
emission. Note that we have excluded MC3-MHD from further analysis due to its
low molecular content (see Fig. 15).
Figure 15: Left: Time evolution of total mass and total H2 mass in the
different molecular clouds, both HD and MHD, from $t_{\rm evol}=2$ to 3.5 Myr.
The solid lines represent the total mass, and the dashed lines represent the
H2 mass. Middle: H2 mass fraction for the same clouds, both HD and MHD. The
two HD clouds are plotted in reddish lines. Apart from MC3-MHD, which we
discard due to its low molecular gas mass, the other MHD and HD and clouds
have comparable masses. The two HD clouds, however, have a much higher H2 mass
fraction. Right: The mass in dendrogram for each cloud. Note that the
discarded MC3-MHD is missing as we did not perform a dendrogram on it. The
dendrogram mass has similar trends to the total mass.
## Appendix B Alternative PDF views
We present additional views of density PDFs (both with a linear scale and
mass-weighted) in Figs. 16 and 17 as a complementary addition to Fig. 2.
Figure 16: Density PDF with a linear y-axis for all HD and MHD clouds at
tevol=3.5 Myr. Figure 17: Mass-weighted PDF for all clouds, both HD and MHD,
at t${}_{\rm evol}=3.5$ Myr. The average mass percentage in the different
regimes (shown by the vertical dotted bars), for both HD and MHD clouds, is
shown as text. The mass contained at $\rho<10^{-24}$ g cm-3 is $<1\%$ for all
clouds. The mass difference in the intermediate regime, even at 3.5 Myr, is
clearly seen.
## Appendix C Distribution of visual extinction for magnetized and non-
magnetized clouds
The gas mass distribution at different $\mathrm{A_{V,3D}}$ values (Eq. 7) for
one example HD and MHD cloud of comparable mass is presented in Fig. 18, for 2
Myr (top) and 3.5 Myr (bottom). The vertical dashed line represents
$\mathrm{A_{V,3D}}=1$. The HD cloud has consistently higher mass at high
$\mathrm{A_{V,3D}}$ values.
Figure 18: Mass weighted $\mathrm{A_{v,3D}}$ PDF for different HD and MHD
clouds at $t_{\rm evol}$=3.5 Myr. MC2-MHD stands out as having much less
shielded gas mass compared to the other clouds. The other HD and MHD clouds
have similar behaviour.
## Appendix D Alternative view of the magnetic field - density relation
As a companion view to Fig.6, we show here the same relation between the
magnetic field and density, but this time using the average density,
$\rho_{\rm avg}$ instead of $\rho_{\rm thr}$. Since for any given structure,
$\rho_{\rm avg}\geq\rho_{\rm thr}$, this results in a shallower fit at the
high density end using $\rho_{\rm avg}$.
Figure 19: Similar to Fig. 6, but using $\rho_{\rm avg}$ instead of $\rho_{\rm
thr}$. This creates a shallower slope at particularly the high density end, as
$\rho_{\rm avg}\geq\rho_{\rm thr}$. However, the overall statistical trend is
similar.
## Appendix E Supplement to the magnetic Jeans analysis
The Jeans mass analysis is only conclusive provided the Jeans length is
resolved. In Fig. 20, we plot the ratio of the magneto-thermal Jeans length,
$\lambda_{\rm B,T}$, to the maximum resolution, $\Delta x$ ($\sim 0.125$ pc
for $\rho_{\rm thr}>10^{-22}$ g cm-3 and $\sim 0.25$ pc for $\rho_{\rm
thr}<10^{-22}$ g cm-3, see also Table 2), as a function of $\rho_{\rm thr}$
for all sub-structures at 3.5 Myr. The colour-bar, similar to Fig. 12, denotes
$c_{s}/v_{\rm A}$. The horizontal dotted line denotes $(2N_{\rm
cells})^{1/3}$. As $N_{\rm cells}$ denotes the minimum number of cells
required in the dendrogram analysis for any structure, $2N_{\rm cells}$ is the
minimum number of cells a structure must have in order to fragment. Therefore,
$(2N_{\rm cells})^{1/3}$ represents the minimum number of cells required in
one direction by which the Jeans length should be resolved. We find that this
seems to not be the case only for some structures with $\rho_{\rm thr}\gtrsim
10^{-20}$ g cm-3. When we fit $\lambda_{\rm B,T}$ against $R$ using a linear
least-squares fit on the logarithm of the data, we obtain an exponent of
$-0.70\pm 0.01$, roughly consistent with an exponent of $-2/3$.
Figure 20: The ratio of the magneto-thermal Jeans length, $\lambda_{\rm B,T}$
to the maximum resolution $\Delta x$ ($\sim 0.125$ pc for $\rho_{\rm
thr}>10^{-22}$ g cm-3 and $\sim 0.25$ pc for $\rho_{\rm thr}<10^{-22}$ g cm-3,
see also Table 2), as a function of $\rho_{\rm thr}$ for all MHD sub-
structures at 3.5 Myr. The horizontal dotted line denotes the resolution limit
for the present dendrogram analysis ($2N_{\rm cells}^{1/3}$, with $N_{\rm
cells}=100$). The red dashed line denotes the best-fit exponent for a linear
least-squares fit on the logarithm of the data. Structures above $\rho_{\rm
thr}\approx 10^{-20}$ g cm-3 seem to be not well resolved enough to be
conclusive regarding the fragmentation analysis.
The analysis performed in Fig. 12 considers the combined contribution of
magnetic and thermal perturbations. It might be interesting to note their
relative contributions. For this purpose, we show a purely magnetic Jeans
analysis in Fig. 21. Comparing $M/M_{\rm B}$ to $M/M_{\rm B,T}$ (Fig. 12), we
find little to no difference, suggesting that the magnetic contribution is in
this density range more important than the thermal contribution. This can also
be seen in the fact that most of the points have larger $v_{\rm A}$ compared
to $c_{s}$ (bluish in the colour-bar). For completeness, we explicitly include
the thermal Jeans mass and length plot in Fig. 22.
Figure 21: Top: same as the combined panels of Fig. 12, but for a purely
magnetic Jeans mass. Bottom: same as Fig. 20, but for a purely magnetic Jeans
length.
Figure 22: Top: same as Fig. 21, but for a purely thermal Jeans mass. Bottom:
same as Fig. 20, but for a purely thermal Jeans length. Note that the
$y$-range in the top panel is different to the previous similar plots.
## Appendix F The scaling-relation between Jeans length and density
The Jeans length, $\lambda$, depends on the characteristic wave speed, $c$
($c_{s}$, $v_{\rm A}$, or a combination of the two), and the density, i.e.
$\lambda\propto c\rho^{-1/2}.$ (46)
In our case, we are dominated by the magnetic over kinetic pressure, i.e.
$\lambda_{\rm B,T}\approx\lambda_{\rm B}$. The Alfvén wave speed scales as
$v_{\rm A}\propto\frac{B}{\rho^{1/2}}.$ (47)
This leads to
$\lambda_{\rm B,T}\propto\frac{B}{\rho}.$ (48)
For a scaling of $B\propto\rho^{1/2}$, this leads to $\lambda_{\rm
B,T}\propto\rho^{-1}$. A scaling of $B\propto\rho^{1/3}$ leads to
$\lambda_{\rm B,T}\propto\rho^{-2/3}$. The fitted value seems to be somewhere
in-between, closer to $\rho^{-2/3}$, and is also roughly consistent with the
overall $B-\rho$ scaling in Fig. 6 and Fig. 19.
|
11institutetext: Université Paris-Saclay, Université Paris Cité, CEA, CNRS,
AIM, F-91191, Gif-sur-Yvette, France 22institutetext: INAF - Osservatorio di
Astrofisica e Scienza dello Spazio di Bologna, via Piero Gobetti 93/3, I-40129
Bologna, Italy 33institutetext: INFN - Sezione di Bologna, viale Berti Pichat
6/2, I-40127 Bologna, Italy
33email<EMAIL_ADDRESS>
# Fast multi-scale galaxy cluster detection with weak lensing:
Towards a mass-selected sample
G. Leroy 11 S. Pires 11 G.W. Pratt 11 C. Giocoli 2233
(Received XXXX ; accepted XXXX)
The sensitivity and wide area reached by ongoing and future wide-field optical
surveys allows for the detection of an increasing number of galaxy clusters
uniquely through their weak lensing (WL) signal. This motivates the
development of new methods to analyse the unprecedented volume of data faster
and more efficiently. Here we introduce a new multi-scale WL detection method
based on application of wavelet filters to the convergence maps. We compare
our results to those obtained from four commonly-used single scale approaches
based on the application of aperture mass filters to the shear in real and
Fourier space. The method is validated on Euclid-like mocks from the
DUSTGRAIN-pathfinder simulations. We introduce a new matching procedure that
takes into account the theoretical signal-to-noise of detection by WL and the
filter size. We perform an analysis of the filters, and a comparison of the
purity and the completeness of the resulting detected catalogues. We show that
equivalent results are obtained when the detection is undertaken in real and
Fourier space, and when the algorithms are applied to the shear and the
convergence. We show that the multiscale method applied to the convergence is
faster and more efficient at detecting clusters than single scale methods
applied to the shear. We obtained an increase of 25$\%$ in the number of
detections while maintaining the same purity compared to the most up-to-date
aperture mass filter. We analyse the detected catalogues and quantify the
efficiency of the matching procedure, showing that less than $5\%$ of the
detections from the multiscale method can be ascribed to line-of-sight
alignments. The method is well-adapted to the more sensitive, wider-area,
optical surveys that will be available in the future, and paves the way to
cluster samples that are as near as possible to being selected by total matter
content.
###### Key Words.:
Gravitational lensing : weak ; \- Galaxies: clusters : general ; \- methods :
data analysis ; \- Cosmology : dark matter, large-scale structure of Universe
\-
## 1 Introduction
Modern cosmological models show that gravitational collapse drives cosmic
structure formation through a hierarchical assembly process in which objects
merge into larger and larger structures. Clusters of galaxies sit at the
endpoint of this process, so the formation and evolution of the cluster
population directly traces the growth of cosmic structure over time. This
information can be used to constrain cosmological models. For example, the
number of clusters as a function of mass and redshift is highly sensitive to
the underlying cosmological parameters (e.g. White & Rees, 1978; Perrenod,
1980; Voit, 2005; Allen et al., 2011).
The matter content of clusters – composed of dark matter (DM; 85%), ionised
hot gas in the intracluster medium (12%), and stars ($\sim 3$%) – reflects
that of the Universe. Clusters are typically detected through their baryonic
components. Studies based on samples from X-ray and Sunyaev-Zeldovich (SZ)
surveys have been very successful at providing cosmological constraints (e.g.
Pacaud et al., 2006; Vikhlinin et al., 2009; Hasselfield et al., 2013; Bocquet
et al., 2019; Salvati et al., 2021), and increasingly-wide-field optical
surveys have become competitive in recent years (e.g. Rozo et al., 2010;
Hamana et al., 2015; Lesci et al., 2022). However, leveraging the data from
these surveys requires linking the baryonic observables (optical richness,
X-ray luminosity, and SZ flux) to the underlying mass. Today, survey sample
sizes are sufficiently large that the dominant uncertainties in cosmological
parameter estimation with clusters lie in systematic effects in the mass
estimates and the selection function (e.g. Pratt et al., 2019, and references
therein).
The vast amount of matter contained in clusters bends the light of background
galaxies. This coherent deflection of the path of light from background
sources by an intervening mass is termed gravitational lensing (e.g. Schneider
et al., 1992; Pyne & Birkinshaw, 1993). At large scales, the distortion in the
shapes of the background galaxies is detectable only statistically, and is
consequently termed weak gravitational lensing. The image distortions due to
the weak lensing effect can be characterised by the shear (a warping of the
background source images) and the convergence (a magnification effect on the
same). Both shear and convergence can provide insights into the statistical
properties of the weak lensing field, and has been shown that they contain
precisely the same information (e.g. Schneider et al., 2002; Pires et al.,
2020). The weak lensing effect is directly sensitive to the total projected
mass along the line of sight (LOS). As such, it is an attractive method for
cluster detection, potentially paving the way towards true mass-selected
samples. However, the sensitivity of cluster detection through the weak
lensing signal depends critically on the number of background sources that are
available to be lensed by the intervening matter. Only recently have deep,
wide-field surveys yielded the background source densities needed to detect
substantial numbers of objects.
A number of methods have been developed to detect galaxy clusters through
their weak lensing shear or convergence signal in optical imaging data.
Application of a simple Gaussian filter to the convergence, combined with
thresholding, was widely used in early optical surveys (e.g. White et al.,
2002; Miyazaki et al., 2002; Hamana et al., 2004; Tang & Fan, 2005; Gavazzi &
Soucail, 2006; Miyazaki et al., 2007; Fan et al., 2010; Shan et al., 2012,
2018). Such methods demonstrated that it was possible to detect clusters
through their weak lensing signal, while motivating the development of new,
more efficient, approaches. In the subsequent development of optimal filtering
techniques, the filter kernel was adapted to include information on the shape
of the expected halo profile, while excluding the shape noise and the
contribution of large-scale structure (LSS) to the noise budget (e.g. Hennawi
& Spergel, 2005; Maturi et al., 2005; Wittman et al., 2006).
The widely used aperture mass (AM) technique, introduced by Schneider (1996),
consists of convolving the lensing signal with a filter function of a specific
scale. A number of new filters have since been developed to maximise the
effectiveness of the AM method; they have been tested on simulations and
applied to various optical surveys (e.g. Schneider, 1996; Schneider et al.,
1998; Jarvis et al., 2004; Schirmer et al., 2004; Hetterscheidt et al., 2005;
Hennawi & Spergel, 2005; Maturi et al., 2005, 2007; Pace et al., 2007;
Schirmer et al., 2007; Dietrich & Hartlap, 2010; Hamana et al., 2012; Lin et
al., 2016; Miyazaki et al., 2018; Hamana et al., 2020; Oguri et al., 2021).
The filter function is a key component of the AM method and must be designed
to obtain the optimal signal-to-noise ratio (S/N) at a given scale. A defining
characteristic of the above approaches is that they all operate on a single
scale. As such, their detection efficiency is highly dependent on the filter
design and on its relation to the size of the structures we want to detect.
While some studies have proposed using multi-scale wavelet filters to de-noise
the convergence map (Starck et al., 2006; Lanusse et al., 2016), they were not
optimised for cluster detection because they do not take the LSS contribution
into account. Nevertheless, such multi-scale de-noising techniques have been
shown to be a promising approach for cluster detection (Leonard et al., 2015).
Motivated by the potential of current and upcoming deep wide-field optical
surveys with sufficient background source densities, such as the Hyper
Supreme-Cam (HSC) survey (Aihara et al. 2018), the Legacy Survey of Space and
Time (LSST111https://www.lsst.org/; Closson Ferguson et al. 2009; Ivezić et
al. 2019), Euclid222https://www.euclid-ec.org/ (Laureijs et al. 2011), and the
Roman Space Telescope333https://roman.gsfc.nasa.gov/ (formerly WFIRST; Spergel
et al. 2015), we revisit here the question of cluster detection through the
weak lensing effect. We perform a quantitative comparison of existing single-
scale detection methods, including a complete analysis of the filters. We
introduce a new multi-scale detection approach based on the wavelet transform
applied to the convergence. We chose to focus on the convergence because this
quantity explicitly traces the total matter distribution integrated along the
LOS and is computationally less expensive to analyse, making it ideal for
application to upcoming large-scale survey data. We quantify the performance
of our new multi-scale approach by applying it to the DUSTGRAIN-pathfinder
simulations detailed in Giocoli et al. (2018), which feature source densities
similar to those expected from the Euclid survey. We find that the new multi-
scale method operating on the convergence is faster and more efficient at
detecting clusters than currently used single-scale methods operating on the
shear.
This paper is organised as follows. In Sect. 2 we summarise the key aspects of
gravitational lensing. The mock dataset DUSTGRAIN-pathfinder simulations are
described in Sect. 3. Section 4 introduces the AM formalism and several
commonly used AM filters. A description of the wavelet formalism is also
provided. Section 5 details the detection procedure that we use to compare the
different filters. An analysis of the different options in the implementation
of the detection algorithm is provided in Sect. 6. In Sect. 7 we provide the
details of the matching procedure that we developed to allow for a fair
comparison of the methods. Finally, the performance of the detection methods
is evaluated in Sect. 8, and we conclude in Sect. 9.
Figure 1: Simulated shear maps covering a field of $5^{\circ}\times 5^{\circ}$
(left and middle panels) and the corresponding E-mode convergence map (right).
The E-mode convergence map directly traces the projected matter distribution,
the overdensities appearing as bright structures.
## 2 Weak gravitational lensing theory
### 2.1 Weak lensing theory
The gravitational field of massive objects affects the path of light in their
vicinity. Thus, the light from background galaxies is deflected as it travels
towards us and their images appear distorted. These distortions, or (reduced)
shear, are a direct observable and are an imprint of the intervening large-
scale matter distribution. Different structures, such as clusters of galaxies,
filaments, or even individual galaxies, can act as lenses and create this
warping effect.
We summarise here the gravitational lens theory that is sufficient for the
treatment of lensing by galaxy clusters (see e.g. Bartelmann & Schneider,
2001). We consider a lens at angular position $\boldsymbol{\theta}$ and at
distance $D_{\rm l}$ from the observer. Its surface mass density,
$\Sigma(\theta)$, is integrated from its 3D mass density,
$\rho(\boldsymbol{\theta},z),$ along the LOS:
$\Sigma(\boldsymbol{\theta})=\int_{0}^{+\infty}\rho(\boldsymbol{\theta},z)dz.$
(1)
From the surface mass density, we can define the lensing potential,
$\psi(\boldsymbol{\theta}),$ according to the position of the background
galaxy sources with distance $D_{\rm s}$ from the observer and distance
$D_{\rm ls}$ from the lens:
$\psi(\boldsymbol{\theta})=\frac{4G}{c^{2}}\frac{D_{\rm l}D_{\rm ls}}{D_{\rm
s}}\int_{\rm\mathbb{R}^{2}}\Sigma(\boldsymbol{\theta}^{\prime})\ln{(\lvert\boldsymbol{\theta}^{\prime}-\boldsymbol{\theta}\rvert)}\,d^{2}\boldsymbol{\theta}^{\prime}.$
(2)
When light rays travel close to a lens, they are bent, leading to image
distortions and, potentially, to multiple images. These images appear with a
deflection angle, $\alpha(\boldsymbol{\theta}),$ induced by the lensing
potential, $\psi(\boldsymbol{\theta})$:
$\alpha(\boldsymbol{\theta})=\nabla\psi(\boldsymbol{\theta}).$ (3)
Here $\alpha(\boldsymbol{\theta})$ is the difference between the angular
position, $\boldsymbol{\beta}$, where the images would be without the lens and
the observed position, $\boldsymbol{\theta}$. This is summarised by the lens
equation:
$\alpha(\boldsymbol{\theta})=\boldsymbol{\theta}-\boldsymbol{\beta}.$ (4)
From Eqs. 2, 3, and 4, the deviation, $\boldsymbol{\beta}$, of the observed
image with respect to the undistorted image can be derived:
$A\equiv\frac{\partial\boldsymbol{\beta}}{\partial\boldsymbol{\theta}}=\delta_{ij}-\frac{\partial^{2}\psi(\boldsymbol{\theta})}{\partial\theta_{i}\partial\theta_{j}}=\pmatrix{1}-\kappa-\gamma_{1}&-\gamma_{2}\\\
-\gamma_{2}1-\kappa+\gamma_{1}.$ (5)
Here
$\displaystyle\gamma_{1}=\frac{1}{2}(\partial_{1}^{2}-\partial_{2}^{2})\psi\
\rm{and}$ (6) $\displaystyle\gamma_{2}=\partial_{1}\partial_{2}\psi$ (7)
correspond to the two components of the shear $\gamma$, and
$\kappa=\frac{1}{2}(\partial_{1}^{2}+\partial_{2}^{2})\psi$ (8)
corresponds to the convergence that is the shape contraction or dilation of
the image.
From the above, we can express the convergence as a function of the critical
value of the surface mass density, $\Sigma_{\rm crit}$:
$\kappa(\boldsymbol{\theta})=\frac{\Sigma(\boldsymbol{\theta})}{\Sigma_{\rm
crit}},$ (9)
with
$\Sigma_{\rm crit}=\frac{c^{2}}{4\pi G}\frac{D_{\rm s}}{D_{\rm l}D_{\rm ls}}.$
(10)
Thus, the convergence is a tracer of the total matter distribution integrated
along the LOS, weighted by the redshift-dependent factor $1/\Sigma_{\rm
crit}$.
The direct observable is the reduced shear, $g$, that is derived from the
ellipticities of the observed background galaxies. The reduced shear, $g$, is
defined as follows:
$g=\frac{\gamma}{1-\kappa}.$ (11)
In the weak lensing regime ($\kappa\ll 1$), it approximates the shear,
$\gamma$.
### 2.2 Application
In practice, we observe a distorted image, from which we measure the galaxy
ellipticities and estimate the shear components, $\gamma_{1}$ and
$\gamma_{2}$. Kaiser & Squires (1993) introduced the mass inversion technique,
which involves the computation of the convergence map $\kappa$ from the
measured shear field.
We can consider a complex notation to represent the shear field,
$\gamma=\gamma_{1}+\rm{i}\gamma_{2}$, and the convergence field,
$\kappa=\kappa_{E}+\rm{i}\kappa_{B}$, with $\kappa_{E}$ and $\kappa_{B}$,
called E- and B-modes. Then, taking the Fourier transform of Eqs. 6, 7, and 8,
we obtain
$\displaystyle\hat{\gamma}=\hat{P}\,\hat{\kappa},$ (12) $\displaystyle{{\rm
with:}\,\,}\hat{P}=\hat{P}_{1}+i\,\hat{P}_{2}\rm{\,\,and\,\,},$ (13)
$\displaystyle\hat{P}_{1}=\frac{k_{1}^{2}-k_{2}^{2}}{k_{1}^{2}+k_{2}^{2}},$
(14)
$\displaystyle\hat{P}_{2}=\frac{2\,k_{1}^{2}k_{2}^{2}}{k_{1}^{2}+k_{2}^{2}},$
(15)
where the hat symbol refers to the Fourier transform, and $k_{i}$ is the wave
number at the angular position, $\theta_{i}$.
Considering the conjugate $\hat{P}*=\hat{P}_{1}-i\hat{P}_{2}$, we can
reconstruct $\kappa_{E}$ and $\kappa_{B}$ from the complex shear $\gamma$,
obtaining
$\displaystyle\hat{\kappa}_{E}=\hat{P}_{1}\hat{\gamma}_{1}+\hat{P}_{2}\hat{\gamma}_{2}$
(16)
$\displaystyle\hat{\kappa}_{B}=-\hat{P}_{2}\hat{\gamma}_{1}+\hat{P}_{1}\hat{\gamma}_{2}.$
(17)
Given that $\psi$ is a scalar potential, it can be shown that weak lensing
does not in principle produce B-modes. Thus, B-modes can be used to estimate
the level of the noise in the data.
Figure 1 shows an example of simulated shear and convergence maps without
shape noise derived from the DUSTGRAIN-pathfinder simulation (described in
Sect. 3) covering a field of $5^{\circ}\times 5^{\circ}$. The left and middle
panels correspond to the two components of the shear, $\gamma_{1}$ and
$\gamma_{2}$ and the right panel to the corresponding E-mode convergence map,
$\kappa_{E}$. Although both the shear and the convergence can be used for
cluster detection, the convergence is easier and computationally less
expensive to analyse because it is a scalar field that is proportional to the
projected matter distribution.
## 3 Weak lensing simulations
### 3.1 The DUSTGRAIN N-body simulations
We make use of a suite of cosmological N-body simulations called The
DUSTGRAIN-pathfinder (see Giocoli et al., 2018, for a detailed description).
This DM-only simulation traces the collisionless evolution of $783^{3}$ DM
particles with a mass $m_{\rm CDM}=8.1\times 10^{10}$ $h^{-1}$ M⊙ , contained
within a periodic cosmological box of side 750 $h^{-1}$ Mpc. In the present
work, we use a subset of the full DUSTGRAIN-pathfinder runs consisting of 256
realisations sharing the same standard cosmological parameters in agreement
with Planck Collaboration XIII 2016: $\Omega_{M}=\Omega_{\rm CDM}+\Omega_{\rm
b}+\Omega_{\nu}=0.31345$, $\Omega_{\rm b}=0.0481$, $\Omega_{\Lambda}=0.68655$,
$h=0.6731,$ and $\sigma_{8}=0.847$. The different LOS realisations were
obtained by randomising the stacked comoving cosmological boxes through
combinations of the following procedures: (i) changing the sign of the
Cartesian coordinates, (ii) redefining the position of the observer, and (iii)
modifying the order of the axes in the coordinate system. By construction,
these variations preserve the clustering properties of the particle
distribution at the scale of the comoving simulation snapshot.
### 3.2 Mock dark matter halo catalogues
The mock DM halo catalogues were generated by identifying the DM haloes in the
DUSTGRAIN simulation through a friends-of-friends algorithm (Davis et al.
1985) with linking distance $\lambda=0.16\times d$, where $d$ is the mean
separation distance between particles. The `SUBFIND` algorithm (Springel et
al. 2001) was then used to evaluate the standard parameters of each friends-
of-friends-identified halo: the redshift, $z$, and the virial mass, $\rm
M_{200c}$, and radius $\rm R_{200c}$, corresponding to the mass and radius of
a spherical region around the fiducial centre of each halo enclosing 200 times
the critical density of the Universe. Thus, for each of the 256 realisations,
the catalogue is composed of the positions (i.e. the right ascension and the
declination) of the identified DM haloes, their estimated redshift, $z_{l}$,
and their virial mass, $\rm M_{200c}$, and radius, $\rm R_{200c}$. This DM
halo catalogue is used in our study for the matching procedure to quantify the
purity and completeness of each selected sample.
### 3.3 Mock galaxy catalogues
The particles stored in 21 different snapshots were used to construct
continuous past light cones from $z=0$ to $z=4$ using the `MAPsim` pipeline
(Giocoli et al., 2015, 2017). The routine extracts the positions of each
particle to recreate the past light cone and at the same time, the particles
are binned to build $27$ different lens planes to recompose the projected
matter density distribution. Then, the shear of each galaxy is computed by
projection, using the Born approximation, which assumes unperturbed light
paths to integrate the lensing distortions (Bartelmann & Schneider, 2001). The
redshift distribution of the galaxies, $n(z)$, was built to follow a realistic
distribution approaching the one expected for the Euclid survey (e.g. Cropper
et al. 2013). Each realisation contains around 3 million galaxies in a field
of view of $5^{\circ}\times 5^{\circ}$, with a galaxy distribution extending
up to $z=4$. This leads to a galaxy density of about $n_{\rm{g}}=30$ gal.
arcmin-2.
Noise was added to the shear to mimic realistic surveys. Uncertainties in the
shear (referred to as ‘shape noise’) arise from a combination of the
unavoidable intrinsic shape of the galaxies (referred to as the ‘intrinsic
shape noise’), and measurement errors that include, among other factors,
uncertainties in the galaxy shape measurement and the point spread function
correction. The intrinsic shape of the galaxies being the dominant component
and the galaxies being randomly distributed, the shape noise can be modelled
as an additive noise. This can be well approximated by a Gaussian distribution
with a mean of $\mu=0$ and a standard deviation of
$\sigma_{\rm{\epsilon}}=0.26$ (e.g. Leauthaud et al., 2007; Schrabback et al.,
2015, 2018). For each galaxy, the shape noise was therefore included in the
catalogue by adding Gaussian noise to the two components of the shear. For
each of the 256 realisations, the galaxy catalogue is composed of: (i) the
positions (i.e. the right ascension and the declination) of the galaxies, (ii)
the two components of the shear, and (iii) the redshift. These noisy galaxy
catalogues are then used as the inputs to evaluate the detection algorithms
described below.
## 4 Filters
In this section we present several different filters that have been used for
weak lensing cluster detection in previous studies, and we compare them to the
wavelet filters that we use in our new multi-scale detection approach. A
complete analysis and comparison of the different filters both in real and
Fourier space is also provided.
Although used since the very first weak lensing detection algorithms (e.g.
White et al., 2002; Miyazaki et al., 2002; Hamana et al., 2004; Tang & Fan,
2005; Gavazzi & Soucail, 2006; Miyazaki et al., 2007; Fan et al., 2010; Shan
et al., 2012, 2018), we excluded the Gaussian filter from this analysis
because it behaves as a low-pass filter and is thus not optimal for cluster
detection. Although low-pass filters appear to be natural tools for reducing
the shape noise that dominates at high-frequencies, they are not adapted to
diminish the low-frequency signal coming from LSS, which causes many spurious
peaks (see e.g. Maturi et al., 2005). Since the use of such low-pass filters
is sub-optimal for the current application, we focus our analysis on pass-band
filters that are better able to target the signal around a given scale.
### 4.1 The aperture mass formalism
The AM method (Schneider, 1996) is commonly used to reconstruct maps of the
projected matter distribution at a given scale. The AM map can be evaluated at
a position $\boldsymbol{\theta_{\rm o}}$ by convolving the convergence field,
$\kappa$, with a filter, $U$, such that
$M_{\rm ap}(\boldsymbol{\theta_{\rm
o}})=\int_{\mathbb{R}^{2}}\kappa(\boldsymbol{\theta})\,U(\lvert\,\boldsymbol{\theta}\,-\,\boldsymbol{\theta_{\rm
o}}\,\rvert)\,d^{2}\boldsymbol{\theta},$ (18)
where $U$ must be compensated; that is,
$\int_{0}^{\rm\theta_{\rm
ap}}\rm\theta^{\prime}\,U(\rm\theta^{\prime})\,d\theta^{\prime}=0$ (19)
must be fulfilled within the aperture of radius $\theta_{\rm ap}$.
Alternatively, the AM map can also be computed from the tangential component
of the shear, $\gamma_{\rm t}$, which is a quantity that can directly be
computed from $\gamma_{1}$ and $\gamma_{2}$:
$\gamma_{\rm t}=\gamma_{1}\cos(\rm\phi)+\gamma_{2}\sin(\rm\phi),$ (20)
where $\phi$ is the polar angle
$\phi(\boldsymbol{\theta},\boldsymbol{\theta_{0})}$ relative to the centre of
the aperture, $\boldsymbol{\theta_{0}}$. The AM map at position
$\boldsymbol{\theta}_{\rm o}$ can then be obtained by convolving the
tangential shear with a filter, $Q$, such that
$M_{\rm
ap}(\boldsymbol{\theta_{o}})=\int_{\mathbb{R}^{2}}\gamma_{t}(\boldsymbol{\rm\theta})\,Q(\lvert\,\boldsymbol{\rm\theta}\,-\,\boldsymbol{\rm\theta_{\rm
o}}\,\rvert)\,d^{2}\boldsymbol{\rm\theta},$ (21)
the filters $Q$ and $U$ being linked by the relations (Schneider & Bartelmann,
1997)
$Q(\rm\theta)=\frac{2}{\rm\theta^{2}}\int_{0}^{\rm\theta}\rm\theta^{\prime}U(\rm\theta^{\prime})\,d\theta^{\prime}-U(\rm\theta)$
(22)
and
$U(\rm\theta)=2\int_{\theta}^{+\infty}\rm\frac{Q(\rm\theta^{\prime})}{\theta^{\prime}}\,d\theta^{\prime}-Q(\rm\theta).$
(23)
The performance of the AM technique at finding clusters on a map depends on
the exact choice of filter, and on its capacity to isolate the lensing signal
from the shape noise and the LSS component. The optimal filters $U$ or $Q$
should be designed to fulfil two properties. In real-space, the filters should
be local, that is, decrease smoothly to zero within a finite radius. Ideally,
they should fit the shape of the structures of interest. Moreover, they should
be local in Fourier space, focusing on a particular angular scale. Ideally,
their size should match as closely as possible the size of the structures of
interest.
A number of functions or families of functions have been proposed (e.g.
Schneider, 1996; Schirmer, 2004; Schirmer et al., 2007; Jarvis et al., 2004;
Hamana et al., 2012; Miyazaki et al., 2018). In the present study, we consider
the following four well-known AM filters.
The first is the original AM family of filters defined in the paper by
Schneider (1996, hereafter $\rm S96$):
$U_{\rm S96}(\theta)=\left\\{\begin{array}[]{ll}1&\mbox{x $\in$ [0,
$\nu_{1}R$],}\\\
\frac{1}{1-c}\big{[}\frac{\nu_{1}R}{\sqrt{(x-\nu_{1}R)^{2}+(\nu_{1}R)^{2}}}\big{]}&\mbox{x
$\in$ [$\nu_{1}R$, $\nu_{2}R$],}\\\ \frac{b}{R^{3}}(R-x)^{2}\,(x-\alpha
R)&\mbox{x $\in$ [$\nu_{2}R$, $R$],}\\\ 0&\mbox{x > $R$},\end{array}\right.$
(24)
where parameters $\alpha$, $b,$ and $c$ ensure that the filter is compensated
and continuous in real space. The parameters $\nu_{1}$, $\nu_{2}$, and $R$
define the angular extent of the filter and must be adapted to the size and
shape of the clusters we want to detect. We implemented the different
parameterisations presented in Schneider (1996) and we use $\nu_{1}=0.1$,
$\nu_{2}=0.9$, $R=9\arcmin$, $\alpha=0.8531$, $b=-329.8,$ and $c=0.2415$,
which we found to maximise the purity and completeness. The corresponding
$Q_{\rm S96}$ can be calculated analytically with Eq. 22.
The second is the filter developed by Schirmer (2004, hereafter $\rm TANH$):
$Q_{{\rm TANH}}(\theta)=\frac{1}{1+{\rm e}^{6-150\,\frac{\theta}{R}}+{\rm
e}^{-47+50\,\frac{\theta}{R}}}\frac{\tanh\left(\frac{\theta}{x_{c}R}\right)}{\frac{\theta}{x_{c}R}},$
(25)
where $R$ is the truncation radius and $x_{c}$ (dimensionless) defines the
width of the filter. This filter was directly derived from the shape of the
tangential shear of the Navarro-Frenk-White (NFW) profile (Navarro et al.,
1996). Following Schirmer et al. (2007), we used $x_{c}=0.1$. We fixed the
truncation radius $R=7\arcmin$ as the optimal choice to maximise the number of
detections, following the procedure described by Hetterscheidt et al. (2005).
The corresponding $U_{\rm TANH}$ can be computed numerically using Eq. 23.
The third is the filter proposed by Jarvis et al. (2004, hereafter J04):
$U_{\rm J04}(\rm\theta)=\frac{1}{\rm
2\pi\sigma^{2}}\,\left(1-\frac{\rm\theta^{2}}{\rm
2\sigma^{2}}\right)\,\exp{\left(-\frac{\rm\theta^{2}}{2\sigma^{2}}\right)},$
(26) $Q_{\rm J04}(\rm\theta)=\frac{1}{\rm
2\pi\sigma^{2}}\frac{\rm\theta^{2}}{\rm
2\sigma^{2}}\exp{\left(-\frac{\rm\theta^{2}}{\rm 2\sigma^{2}}\right)},$ (27)
where $U_{\rm J04}$ was defined following Van Waerbeke (1998) as the second
derivative of a Gaussian function with standard deviation $\sigma$. The
$U_{\rm J04}$ is also known as the Mexican hat wavelet filter. In the
following mass aperture computation, we use $\sigma=4\arcmin$, corresponding
to an apparent angular radius of about $4^{\prime}$, and a truncation radius
$R=20\arcmin$, as suggested by Leonard et al. (2012) to minimise oscillations
and high frequency mode contamination.
The fourth is the filter introduced by Hamana et al. (2012) and used by
Miyazaki et al. (2018, hereafter M18), which is defined as a truncated
Gaussian filter:
$U_{\rm
M18}(\rm\theta)=\frac{1}{\rm\pi\,\theta_{s}^{2}}\,\exp{\left(\frac{\rm\theta^{2}}{\rm\theta_{s}^{2}}\right)}\,-\,U_{\rm
o},$ (28)
where $U_{0}$ is a parameter to ensure that the filter $U$ is compensated and
$Q_{\rm
M18}(\theta)=\frac{1}{\rm\pi\,\theta^{2}}\bigg{[}1\,+\,\left(1-\frac{\rm\theta}{\rm\theta_{s}}\right)^{2}\bigg{]}\,\exp{\left(\frac{\rm\theta^{2}}{\rm\theta_{s}^{2}}\right)},$
(29)
with $\theta_{s}$ being defined as the angular scale of the aperture. In
practice, the choice of $\theta_{s}$ influences the angular radius of the
cluster the filter targets, but the relation is not direct. Following Miyazaki
et al. (2018), we use $\theta_{s}=1.5\arcmin$, and $U_{\rm M18}(\theta)=0$ if
$\theta>R$, with a truncation radius $R=15\arcmin$. We fixed this value of $R$
to reduce the impact of the truncation on the filter behaviour as it derives
from a Gaussian filter. Indeed, the filter $U_{\rm M18}$ reaches negative
values at $U_{\rm M18}(R)$ and is subsequently forced to $0$ by the
truncation. This step in the function results in contamination in the signal.
As the choice of truncation radius strongly impacts the computation time,
there is a clear trade off between the computation time and the filter
behaviour.
### 4.2 The wavelet formalism
The wavelet formalism can also be used to reconstruct maps of the projected
matter distribution at a given scale (see e.g. Leonard et al., 2012; Pires et
al., 2012). In Leonard et al. (2012), the authors showed that wavelet filter
functions at a given scale are formally identical to AM filter functions at
that scale. Similar to the AM functions, many different wavelet functions
exist, including starlet, Mexican hat, Morlet, and biorthogonal (see e.g.
Starck et al., 1998, 2006, for reviews). Many wavelet functions have been
specifically designed to fulfil the properties of localisation in real and
Fourier spaces, which is a distinct advantage compared to the AM approach.
A wavelet map of the convergence $\kappa$ at position $\theta$ and scale $a$
can be computed as
$W_{a}(\theta)=\frac{1}{\sqrt{a}}\int\kappa(t)\,\Psi_{a}(\,t-\theta\,)\,dt,$
(30)
where $\Psi_{a}$ is the wavelet filter of scale $a$. All wavelet filters are
defined such that they respect the condition of zero mean. Therefore, they are
compensated by definition.
Figure 2: Comparison of different filter functions, $U$ (left) and $Q$
(right), for AM map reconstruction. The figure shows in blue (dotted lines)
the AM filter functions defined by Schneider (1996, S96) with $R=9\arcmin$; in
red (dash-dotted lines) the AM filter functions defined in Schirmer (2004,
TANH) with $R=7\arcmin$; in purple (dash-dotted lines) the AM filter functions
defined in Jarvis et al. (2004, J04) with $\sigma=4\arcmin$; and in green
(dashed lines) the AM filter functions defined by Miyazaki et al. (2018, M18).
The filter $U_{\rm M18}$ is also a compensated filter thanks to the $U_{0}$
term, but due to the high truncation radius we used, this is not distinctly
visible. Finally, the AM filter functions corresponding to the wavelet filter
functions at scale $i=4$ are displayed in orange (solid lines). All the
filters are normalised to have a maximum amplitude of 1.
Another advantage of the wavelet formalism is that it can decompose a given
image into several complementary scale components. This is possible using the
wavelet transform, for which fast algorithms exist. For our study, we use the
isotropic un-decimated wavelet transform, also called the starlet transform
(e.g. Starck et al., 1998, 2006), which is able to simultaneously compute
several wavelet maps on dyadic scales. The starlet functions are isotropic,
which makes them well suited to extracting galaxy clusters, which appear
roughly circular in the convergence maps. As such functions decompose an image
into several complementary scales (see Sect. 4.3), they are ideal for
undertaking a multi-scale analysis and for targeting structures of different
angular sizes. The starlet transform decomposes the convergence as follows:
$\kappa(\theta)=C_{J}(\theta)+\sum_{i=1}^{J}W_{i}(\theta),$ (31)
where $J$ is the number of scales of the decomposition, $C_{J}$ is the
corresponding smoothed version of the convergence $\kappa$, and $W_{i}$ are
the wavelet maps targeting clusters with an apparent angular radius of
$2^{i-1}$ times the size of the pixel. As a consequence, the wavelet scales
are fixed by the pixel size, and, by definition, increase as a power of two.
In the starlet transform, the wavelet functions are defined as the difference
between two $B^{3}$-spline functions at different resolutions. Their
application to the convergence map was shown to be equivalent to applying the
following AM filter (Leonard et al., 2012):
$U_{W_{i}}(u)=\frac{1}{9}\,\bigg{[}\,93\,\left\lvert
u\right\rvert^{3}\,-\,64\,\left(\,\left\lvert\frac{1}{2}\,-\,u\,\right\rvert^{3}\,+\,\left(\,\frac{1}{2}\,+\,u\,\right)^{3}\right)\\\
+18\,\left(\,\left\lvert\,1\,-\,u\,\right\rvert^{3}\,+\,\left\lvert\,1+\,u\,\right\rvert^{3}\,\right)-\frac{1}{2}\,\left(\,\left\lvert\,2-\,u\,\right\rvert^{3}+\left\lvert\,2+\,u\,\right\rvert^{3}\right)\bigg{]},$
(32)
where
$u=\frac{\theta}{2^{i}x},$ (33)
and $x$ is the pixel size. The new multi-scale detection method that we
introduce in Sect. 5.3 takes advantage of this multi-scale decomposition, the
wavelet transform being able rapidly to decompose the convergence into
different scales.
### 4.3 Filter analysis
The performance of each detection method is closely linked to the capacity of
the filter functions to reduce the shape noise and the contribution from LSS,
while simultaneously minimising the signal loss. In the following, we analyse
the properties of the filters described above both in real and Fourier spaces
to help in the interpretation of the results presented in Sect. 8.
Figure 3: Fourier space comparison of several AM filters and wavelet filters,
expressed as a function of wave number $k$ in arcmin-1 and filter radius in
arcmin. The dotted blue line (upper panel) represents the S96 filter, the
dash-dotted red and purple lines (middle panels) represent the TANH and J04
filters, and the dashed green line (lower panel) corresponds to the M18
filter. The solid lines show the different wavelet filters: the teal, green,
orange, and red light lines correspond to the W2, W3, W4, and W5 wavelet
filters with radii of $1\aas@@fstack{\prime}17$, $2\aas@@fstack{\prime}32$,
$4\aas@@fstack{\prime}68$, and $9\aas@@fstack{\prime}37,$ respectively. For
comparison purposes, the J04 and M18 filters are forced to the same amplitude
as the W4 function to illustrate their behaviour in terms of filter radius.
In Fig. 2 we compare the different filter functions $Q$ and $U$, defined in
Sects. 4.1 and 4.2, in real space. This comparison underlines the differences
in filter shape. In particular, we focus our attention on their compensated
and local properties. The dotted blue curves correspond to the AM functions
defined in Schneider (1996). The filter $U_{\rm S96}$ is compensated and local
by design. The dash-dotted red curves correspond to the AM functions defined
in Schirmer (2004). The filter $U_{\rm TANH}$ drops rapidly to zero, but it is
not local and is only compensated at infinity. The localisation is imposed by
a truncation radius at $R=7^{\prime}$. The dash-dotted purple curves
correspond to the AM functions defined in Jarvis et al. (2004). This filter
$U_{\rm J04}$ is also not local, and is also only compensated at infinity
since it tends to zero when $\theta$ tends to $+\infty$. In practice, the
filter is truncated at R=$20\arcmin$ to have a local support. The dashed green
curves correspond to the AM functions defined in Miyazaki et al. (2018). In
the definition of $U_{\rm M18}$, $U_{0}$ ensures that the filter is
compensated within an aperture of radius $R$. However, the $Q_{\rm M18}$
function is non-local, which can introduce errors when truncating the
aperture. The solid orange curves correspond to the wavelet filter function
$U_{W_{4}}$ defined in Eq. 32 for $i=4$ (hereafter W4). The wavelet filter is
shown to be compensated within an aperture radius equal to two times the
filter radius.
It is also instructive to study the representation of the filter functions in
Fourier space as this allows us to fully understand how the filters are
designed to locate structures. In Fourier space, an optimal band-pass filter
should be local, in order to select only a specific range of frequencies
around the scale of interest, and thus reduce the noise and the contribution
from LSS. The design of AM filters can be delicate, and some such filters
suffer from oscillations in Fourier space owing to truncations applied in real
space that deteriorate the band-pass ability of the filter (see e.g. Leonard
et al., 2012, for more details). In contrast, many wavelet functions have been
specifically designed to fulfil the properties of localisation in real and
Fourier space.
To analyse the filters precisely in Fourier space, we studied their impulse
response. For this purpose, we simulated a null convergence map with a single
peak at the centre. Then we passed this convergence map through the four AM
filters and each of the wavelet functions, and computed the corresponding
power spectra in Fourier space. The resulting power spectra are compared in
Fig. 3. This representation highlights each characteristic filter radius,
corresponding to the characteristic size of the structures the filter targets.
The 3′, radius of the S96 filter falls between that of the W3 and W4 wavelet
filters. The TANH filter, with a radius of about 2′, matches that of the W3
wavelet filter. The J04 filter, which has a radius of about 4′, matches that
of the W4 wavelet filter. Finally, the M18 filter can be compared to a
combination of the W3 and W4 wavelet filters. In the Fourier domain, we can
clearly see the band-pass behaviour of the different functions. In particular,
the S96 and TANH filters show oscillatory behaviour, which can be explained by
their design.
Figure 4: Wavelet decomposition of an E-mode convergence map (including shape
noise), covering a simulated field of $5^{\circ}\times 5^{\circ}$. The upper-
left panel corresponds to the E-mode convergence map to be decomposed. The
other panels correspond to the wavelet maps at scale W1 =
$1\aas@@fstack{\prime}17$ (upper-middle panel), scale W2 =
$2\aas@@fstack{\prime}32$ (upper-right panel), scale W3 =
$4\aas@@fstack{\prime}68$ (lower-left panel), scale W4 =
$9\aas@@fstack{\prime}37$ (lower-middle panel), and scale W5 =
$18\aas@@fstack{\prime}75$ (lower-right panel). The wavelet maps are sensitive
to structures of different apparent angular sizes. The W1 map (upper-middle
panel) appears to be dominated by the shape noise. The wavelet decomposition
also includes the smoothed version, which we excluded from this
representation.
This Fourier space representation further shows that the selected scales of
the wavelet filter appear to be well defined. In the wavelet formalism, the
wavelet filter functions have complementary dyadic scales, defined by the
pixel resolution as explained in Sect. 4.2. Their comparison with the four AM
filters highlights that a wavelet filter at a given scale can directly be
compared to an AM filter. However, unlike AM filters, which focus essentially
on one scale, the wavelet formalism allows us to extract signal at more than
one complementary scale. Generally, the standard AM methods target clusters
located at relatively low redshift with a large apparent angular size. Thus,
typical AM filter radii are taken between $1\arcmin$ and $5\arcmin$ (e.g.
Hamana et al., 2012) to maximise the S/N for clusters with a similar angular
extent. In contrast, the lensing signal from clusters with different apparent
angular sizes can easily be targeted using the wavelet formalism.
## 5 Detection algorithms
In this section we discuss the implementation of the detection approach we
used to compare the filters. The detection algorithms are all based on the
same basic principles. The lensing signal is first convolved with a filter
function of a specific scale (described in the previous section) to
reconstruct E- and B-mode AM maps. Thresholding is then applied to the E-mode
AM map to locate the overdensity peaks corresponding to clusters.
### 5.1 Binning and filtering
In the following, we describe the steps used to produce the E- and B-mode AM
maps on which the detection is performed. As explained in Sect. 3.3, the noisy
galaxy catalogues are the basic inputs for the detection algorithms. These
contain the shear at the discrete positions of each of the galaxies.
At this stage the convolution by the filter function can be undertaken in real
space, on the shear directly at the position of the galaxies, or in Fourier
space, by binning the shear on a regular grid to build a shear map. A third
possibility is to reconstruct the convergence from the shear maps and then to
perform the filter convolution on the resulting convergence map. A comparison
of these approaches is provided in Sects. 6.1 and 6.2, where we show that we
obtain statistically similar results irrespective of approach. For the
following, we use the third approach because it is computationally faster, in
particular for wavelet filters, with the use of the wavelet transform.
We first binned the observed galaxy ellipticities on a regular grid to create
what we refer to as the noisy shear maps. In our study, the simulated field
being $5^{\circ}\times 5^{\circ}$, we decided to bin the galaxies in a grid of
$512\times 512$ pixels, yielding a pixel size of $0\aas@@fstack{\prime}58$. On
average, with a galaxy density of 30 galaxies per arcmin2, about 10 galaxies
fall into each pixel. The standard method for binning the shear on a regular
grid consists of simply calculating the average shear per pixel (see e.g.
Kaiser & Squires, 1993; van Waerbeke, 2000; Pires et al., 2020). In this
paper, we introduce a new binning strategy to be closer to the real space
approach. This consists of summing the shear of all the galaxies that fall
into each pixel. Thus, the shear in each pixel is effectively weighted by the
number of galaxies in that pixel. This binning strategy is more adapted to
weak lensing detection because additional information on the galaxy density is
included in each pixel. The resulting shear map can then be normalised by the
mean number of galaxies per pixel to facilitate the comparison with other
approaches. However, this global normalisation has no impact on the detection
because it affects the E- and B-modes in a similar way. The impact of these
two different binning approaches is discussed in more detail in Sect. 6.1. The
disadvantage of this approach is that the weak lensing signal depends
significantly on the number of galaxies per pixel, complicating its
interpretation in terms of mass. However, such calculations can be undertaken
in a second step, after detection.
We then applied the Kaiser & Squires inversion described in Sect. 2.2 to the
noisy shear maps $\gamma_{1}$ and $\gamma_{2}$, to reconstruct the E- and
B-mode convergence maps. Convolution by the different filters was then
performed both on the E- and B-mode convergence maps, to produce the
corresponding AM maps.
In the following, we refer to maps of the projected matter distribution at a
given scale as AM maps. In practice, these can be obtained by applying AM
filters or wavelet filters to the shear or to the convergence. The procedure
to detect the haloes on these E- and B-mode AM maps using a single scale is
identical, regardless of the method that has been used to produce them.
### 5.2 Single-scale detection procedure
Once the E- and B-mode AM maps at a specific scale are produced, the detection
procedure starts by applying a thresholding step. Since the weak lensing
effect produces only E-modes, the B-modes are simply due to the noise and can
therefore be used to evaluate the level of the noise in the E-mode AM map. As
proposed by Miyazaki et al. (2018), we used the maximum pixel value in the
B-modes to define the threshold that is then applied to the E-mode AM map. We
then only keep the pixels in the E-mode AM map that are greater than this
threshold, setting the other pixels to zero.
In their study of the HSC data, Miyazaki et al. (2018) find that this
threshold corresponds to $4.7$ times the standard deviation of the noise. The
standard deviation depends on the filter under consideration and is computed
from the B-mode AM map. For the DUSTGRAIN-pathfinder simulation we use here,
the equivalent average thresholds for the 256 realisations correspond to
$4.09$, $4.37$, $4.01$, $4.35$, $4.57$, $4.35,$ and $4.04$ times the standard
deviation for the S96, TANH, J04, M18, W2, W3, and W4 filters, respectively.
The final detection step is to identify the peaks in each of the thresholded
E-mode AM maps. A peak is defined as a pixel whose value is greater than its
eight nearest neighbours. The peaks at the border of the map over a width
equal to the filter radius are discarded to avoid contamination due to
boundary effects introduced by the convolution step. Following this procedure,
we obtain a list of detections and associated peak coordinates. While this
procedure is sufficient for single-scale methods, it needs to be further
developed in the multi-scale case to deal with multiple detections, as we
describe below.
### 5.3 Multi-scale detection procedure
In the detection method based on the wavelet transform, we chose to decompose
the lensing signal into $J=5$ scales (see Eq. 31), as this will encompass a
maximum of possible galaxy cluster apparent angular sizes. Figure 4 shows the
result of such a decomposition applied to the noisy convergence map of a
$5^{\circ}\times 5^{\circ}$ simulated field. The top left panel shows a noisy
version of the E-mode convergence map displayed on the right panel of Fig. 1,
and the other panels show the wavelet maps corresponding to scales $i=1$ to 5
(W1-W5). Each wavelet map gives details of the original convergence map at
different scales. The first scale (W1) is mostly dominated by the noise. In
contrast, the scales from W2 to W5 shed light on different signals within the
input E-mode convergence map. In our analysis, we decide to only keep the
scales W2, W3 and W4. We remove scale W1 because it is dominated by noise.
Scales W5 and above have also been removed because these filter sizes greatly
exceed the expected angular size of the clusters we want to detect.
On the remaining scales W2, W3, and W4, we apply the single-scale detection
procedure described above to extract the local maxima at each individual scale
(hereafter W234). We now have to deal with an issue that is specific to the
multi-scale approach. A signal from a given object can be detected at several
scales, and with a position that varies slightly from one scale to another. We
refer to these henceforth as multiple detections. Once we have obtained a peak
candidate list for each scale, it is important to identify such multiple
detections and to recombine them. This is achieved by computing the separation
distance between all the pairs of detections on consecutive scales. If two
peaks on different scales have a separation distance smaller than the larger
of the two filter radii, we consider them to be a multiple detection. These
are then recombined, and the position of the recombined peak is set by the
position of the detection at the finest scale. If there are more than two
peaks, we repeat this procedure for all the detection peaks and all the scales
consecutively. We keep track of the index $i$ of the finest scale where the
recombined detection appears, as this will be considered as the scale at which
the signal was detected for the matching procedure described below in Sect. 7.
We also saved the detections at each individual wavelet scale to study their
complementarity before combination and to allow a comparison of their
performance with respect to the corresponding AM filters.
Figure 5: Comparison of the real and Fourier space implementations. Left:
Power spectra of a $5^{\circ}\times 5^{\circ}$ AM map with $512\times 512$
pixels obtained using the real space implementation (blue) and using the
Fourier space implementations (orange and green lines for summed and averaged
pixel binning, respectively). Compared to the real space implementation, there
is a $1.2\%$ and $1.5\%$ loss of power for the Fourier space implementation
with summed and averaged pixel binning, respectively. Right: Detections
obtained on the same field when the computation is undertaken in real space
(blue circles) or in Fourier space using either the summed pixel binning
(golden circles) or the averaged pixel binning (green circles). The Fourier
space implementation with summed pixel binning is very close to the real space
implementation in terms of detections. When the binning is performed by
averaging the shear values, there are fewer detections, and about 30% of these
detections are different from those obtained with the real space
implementation.
Figure 6: Comparison of the shear and convergence approaches. Left: Power
spectra of a $5^{\circ}\times 5^{\circ}$ AM map with $512\times 512$ pixels
obtained by applying the AM J04 filter to the shear maps (blue line) and to
the E-mode convergence map (orange dots) computed from the same data as in
Fig. 6. The power spectra agree perfectly. Right: Detections obtained on the
same field when the convolution by the J04 AM filter is performed on the shear
map (blue circles) and on the convergence map (golden circles). There is a
one-to-one correspondence between the detections.
## 6 Analysis of implementation options for the detection algorithms
In practice, the detection is undertaken on a catalogue of galaxies, for each
of which a noisy shear measurement is available. As described above, this is
typically undertaken through the convolution of the shear with a compensated
filter. This convolution can be performed in real space, directly at the
position of the galaxies (e.g. Maturi et al., 2005; Wittman et al., 2006;
Miyazaki et al., 2018; Hamana et al., 2020), or in Fourier space, binning the
shear on a regular grid (e.g. Hennawi & Spergel, 2005; Leonard et al., 2015;
Oguri et al., 2021). Moreover, some cluster weak lensing detection algorithms
are based on the convolution of the convergence (derived from the shear and
also defined on a regular grid) by a compensated filter (e.g. White et al.,
2002; Hamana et al., 2004; Tang & Fan, 2005; Miyazaki et al., 2007). In this
section we explore how binning impacts the signal, and whether the choice to
use the shear or the convergence has any effect on the detection efficiency.
### 6.1 Detection in real space versus Fourier space
In the limit of a perfectly homogeneous distribution of background sources,
the detection of convergence peaks, whether in real space or in Fourier space,
depends only on the number density of these sources. However, as the
background sources are not homogeneously distributed, an additional
uncertainty is introduced into the method by the binning. This can impact the
position and the S/N of the detections. This point has not been fully studied
here because the distribution of the background sources is homogeneous in the
simulations. However, as the number of background sources per pixel varies
slightly for statistical reasons, our study already provides some insights on
the impact of the background source distribution.
#### 6.1.1 Binning strategy
The computation time for AM-based detection algorithms is considerable for the
real space approach. Computation time is significantly reduced for AM-based
detection algorithms when using the Fourier approach. However, there is a loss
of resolution compared to the real space approach because the galaxies must be
binned on a shear map before performing the Fourier transform, instead of
using their exact position. This loss of resolution introduces a loss of power
for the high S/N pixels in the shear map, affecting the efficiency of the
detection method. An additional disadvantage of the Fourier approach is that
there are artefacts at the border of the field due to the periodic assumption
of the discrete Fourier transform. Such effects can be significantly mitigated
by removing the image boundaries or by dealing correctly with borders during
the mass inversion, as proposed in Pires et al. (2020).
The impact of the loss of resolution when working in Fourier space depends on
the map resolution (i.e. the number of pixels of the map), and can be
mitigated by using an appropriate value. Indeed, there is a trade-off to be
made on the map resolution. Decreasing the resolution too much leads to power
leakage, which can affect the detection method. On the other hand, finer
binning can oversample the underlying data, and can introduce missing data
into the map.
Both real space and Fourier space approaches can be found in the literature
(e.g. Hennawi & Spergel, 2005; Gavazzi & Soucail, 2006; Miyazaki et al., 2018;
Hamana et al., 2020; Oguri et al., 2021). In this paper we have introduced an
additional Fourier space approach with a different binning strategy as
explained in Sect. 5.1.
#### 6.1.2 Impact on the signal
We first assessed the impact of the choice of a real or Fourier space approach
on the signal conservation, when using an appropriate map resolution. We
considered three different implementations: (i) application of an AM filter to
the shear at the positions of the galaxies, directly in real space; (ii)
application of an AM filter in Fourier space to the binned shear obtained by
averaging the signal from the galaxies that fall in each pixel (averaged pixel
binning); and (iii) application of an AM filter in Fourier space to the binned
shear obtained by summing the signal from the galaxies that fall in each pixel
(summed pixel binning).
First, we studied the differences between the three implementations through
their spectra in the Fourier domain. For this test, we used the filter
proposed by Jarvis et al. (2004) because it can be easily applied to the
shear, either at the position of the galaxies or binned on a regular grid.
From the shear catalogue of a $5^{\circ}\times 5^{\circ}$ field without noise,
we computed the E-mode AM maps for the three different implementations and
derived the corresponding power spectrum.
The results are presented in the left-hand panel of Fig. 6, with the power
spectra shown for implementations 1, 2, and 3. The power spectra are very
similar, showing that on average the signal is equally well preserved by all
three approaches. To quantitatively compare the implementations, we computed
the integral of the power spectra. Compared to the real space approach, we
find that the power spectrum integral of the Fourier space implementation
suffers from a $1.2\%$ loss for summed pixel binning, and a $1.5\%$ loss for
averaged pixel binning.
#### 6.1.3 Impact on the detections
We now analyse the impact of the three implementation choices discussed above
on the detection efficiency. For each of the three implementations, we built
the E- and B-mode AM map and applied the detection procedure described in
Sect. 5.2 using the Jarvis et al. (2004) AM filter. This was undertaken on all
256 realisations. The right-hand panel of Fig. 6 compares the detections
obtained with the three different implementations for one of the realisations.
When averaging these results over the 256 realisations, we found that the mean
number of detections using the real space approach ($49.2\pm 0.6$) is very
close to that from the Fourier space approach with summed pixel binning
($47.9\pm 0.5$). In contrast, the Fourier space approach with averaged pixel
binning gives a smaller mean number of detections ($43.4\pm 0.5$). Moreover, a
closer look on the detections made by the different approaches reveals
noticeable differences. We found that $28.2\%\pm 2.1\%$ of the detections
obtained from the averaged pixel binning approach differ from those made by
the real space approach. In contrast, with the summed pixel binning approach
only $7.6\%\pm 1.3\%$ of the detections are different.
These results show that the binning strategy has a non-negligible impact on
the detection efficiency, both in terms of number and distribution of the
detections. In light of these results, we conclude that the Fourier space
approach with summed pixel binning strategy agrees very well with the real
space approach, producing results that are statistically similar. In the
following, we decided to perform all tests and comparisons using the third
implementation (i.e. the Fourier space approach using a summed pixel binning
strategy) because it is computationally less demanding. It has the additional
advantage of providing a substantial speed increase with respect to detection
in real space: for our applications, we find a speed increase of two orders of
magnitude, in agreement with the study by Leonard et al. (2012).
### 6.2 Detection using shear map versus convergence map
Both shear and convergence can give insights into the statistical properties
of the weak lensing field, and indeed, it can be shown that they contain
precisely the same information (e.g. Schneider et al., 2002; Pires et al.,
2020). The (reduced) shear is a direct observable and is usually preferred for
reasons of simplicity. However, the convergence has the key advantage that
$\kappa_{E}$ encapsulates all the lensing signal, while it is inevitably
shared between $\gamma_{1}$ and $\gamma_{2}$ in the shear. In this connection,
the convergence is more adapted for galaxy cluster detection because it
explicitly traces the total matter distribution integrated along the LOS. A
further advantage of using the convergence is that it is computationally less
expensive to analyse.
Figure 7: Application of a theoretical S/N pre-selection to the halo
catalogue. Left: Weak-lensing selection function for clusters of galaxies, in
the redshift–mass plane.The selection function has been computed assuming a
Euclid-like redshift distribution with $\rm n_{\rm g}=30\,gal.{arcmin}^{-2}$
and $\sigma_{\epsilon}=0.26$. Contours denote the theoretical S/N for an NFW
profile. Right: Example of halo catalogue pre-selection using the theoretical
S/N, for one $5^{\circ}\times 5^{\circ}$ field. The light blue dots correspond
to all the haloes in the field, and the dark blue dots are the selected haloes
with S/N$>2$.
Although the reconstruction of the convergence field from survey data is a
difficult task, Pires et al. (2020) show that the lensing signal is preserved
in the convergence maps provided the mass inversion is performed without noise
regularisation, and systematic effects such as irregular sampling and complex
survey geometries are well controlled.
The AM formalism can be expressed either on the shear or on the convergence
maps. We now assess the difference between these two approaches. We first
compared them in Fourier space. For each realisation of the noisy shear
catalogue, we produced the E- and B-mode AM maps using the J04 AM filter
applied to either the shear or the convergence maps. Then, we computed and
compared the power spectra. The result of this comparison for one realisation
is shown in the left-hand panel of Fig. 6, where we see that the power spectra
obtained from the shear aperture map and the convergence aperture map are in
good agreement. Averaging over all 256 realisations, we measure a discrepancy
of $0.03\%$ in the power spectrum integral between the shear and the
convergence approaches. This difference can be explained by residual border
effects that are not totally suppressed when removing the image boundaries.
We then compared the two approaches in terms of detections, applying the
detection procedure described in Sect. 5.2. The results, shown in the right-
hand panel of Fig. 6, highlight the perfect agreement between the approaches.
This result remains valid for wavelet filters that are formally identical to
AM filters (see Leonard et al. 2012 for more details). However, while the AM
formalism can be applied to either the shear or the convergence, the wavelet
transform is not designed to process spin-2 fields (i.e. with two independent
components) such as the shear field. So, the shear and convergence approaches
being equivalent, we decided to perform all the tests and comparisons applying
the AM and Wavelet filters to the convergence maps, to be able to use the
wavelet transform.
## 7 Matching
Once the peaks are detected in the E-mode AM maps, they must be cross-matched
with the position of the haloes in the DM halo catalogue. This matching
procedure is essential to characterise the performance of the different
methods, and the results can be biased if not undertaken correctly. This is
particularly true when the filter characteristics differ significantly between
methods. We first identify two main points in the matching procedure that can
significantly increase the number of false detections.
The first concerns the characteristics of the halo catalogue used for
matching. The DM halo catalogue contains objects down to a mass of $\rm
M_{200c}=10^{12}\,h^{-1}\,M_{\odot}$ and up to a redshift of $\rm z=3.6$. For
the cosmology under consideration, a halo catalogue with these characteristics
contains about $15,000$ haloes in a field of $5^{\circ}\times 5^{\circ}$.
Obviously, only a small fraction of these haloes will be detectable through
the weak lensing effect, which is only sensitive to the most massive clusters
with sufficient galaxy sources behind them to trace their mass distribution.
Changing the characteristics of the halo catalogue, such as imposing limits in
mass and/or redshift, changes the performance of the methods with respect to a
number of key points (e.g. matching, false associations).
The second is the association distance. Typically, the matching is performed
within a fixed physical, comoving, or angular radius centred on the candidate
peak, the association distance of which is usually optimised for a given
filter (e.g. Hennawi & Spergel, 2005; Gavazzi & Soucail, 2006; Miyazaki et
al., 2018; Hamana et al., 2020; Oguri et al., 2021). However, when comparing
different filters, this association distance must be adapted for each
detection method to allow a fair comparison.
In the following, we address the first issue by applying a pre-selection to
the halo catalogue based on the theoretical S/N of the detection of an NFW
halo profile by weak lensing. Regarding the second issue, we developed a
method that allows us to adapt the matching distance to the filter
characteristics used in the detection method.
### 7.1 Halo catalogue pre-selection by S/N
Imposing a pre-selection on the DM halo catalogue, by removing those haloes
that are unlikely to be detected through weak lensing, reduces significantly
the number of false associations. We defined our pre-selection by deriving a
measure of the halo detectability in the form of the theoretical S/N of the
detection by weak lensing (e.g. Hamana et al., 2004; Hetterscheidt et al.,
2005; Bergé et al., 2010; Andreon & Bergé, 2012). We followed the approach
proposed by Bergé et al. (2010). Assuming the filter perfectly represents the
signal $\kappa$, the S/N of an halo of mass $\rm M_{200c}$ and redshift
$z_{l}$ can be expressed as
$\nu=\frac{\sqrt{\rm n_{\rm
g}}}{\sigma_{\epsilon}}\sqrt{\iint\kappa(\boldsymbol{\theta})^{2}\,\rm
d\boldsymbol{\theta}^{2}},$ (34)
where $\sigma_{\epsilon}$ is the mean intrinsic shape noise of the source
galaxies and $\rm n_{\rm g}$ the mean density of background sources. Assuming
the mass of the cluster follows an NFW distribution (Navarro et al., 1996) its
density profile is given by
$\rho(r)=\frac{\rho_{s}}{(r/r_{s})\,(1+r/r_{s})^{2}},$ (35)
where $\rho_{s}$ is the characteristic density and $r_{s}=R_{200c}/c$ is the
scale radius. The surface mass density $\Sigma$, projected along the LOS, can
be written as (see Bartelmann & Schneider, 2001)
$\Sigma(x)=2\,r_{\rm s}\,\rho_{\rm s}\,g(x),$ (36)
where $x=r/r_{\rm s}$ is a dimensionless radius. The function $g$ is defined
with respect to the distance to the halo centre and the concentration, $c$. If
$c>1$, it can be written as (Wright & Brainerd, 2000)
$g(x)=\frac{1}{1+c}\left\\{\begin{array}[]{ll}-\frac{\sqrt{c^{2}-x^{2}}}{(1-x^{2})(1+c)}+\frac{1}{(1-x^{2})^{3/2}}\rm
arccosh\left(\frac{x^{2}+c}{x(1+c)}\right)&\mbox{(x\,<\,1),}\\\
\frac{\sqrt{c^{2}+1}}{3}\left(1+\frac{1}{1+c}\right)&\mbox{ (x\,=\,1),}\\\
-\frac{\sqrt{c^{2}-x^{2}}}{(1-x^{2})(1+c)}-\frac{1}{(x^{2}-1)^{3/2}}\rm
arccos\left(\frac{x^{2}+c}{x(1+c)}\right)&\mbox{(1<x<c),}\\\
0&\mbox{(x>c).}\end{array}\right.$ (37)
The concentration parameter $c$ is derived using the semi-analytical model
introduced by Diemer & Joyce (2019), as implemented in COLOSSUS (Diemer,
2018). The S/N of the halo can then be written as (Bergé et al., 2010)
$\nu=\,<Z>\,2\sqrt{2\,\pi}\,\frac{\sqrt{\rm n_{\rm
g}}}{\sigma_{\epsilon}}\frac{\rm\,r_{\rm s}^{2}\,\rho_{\rm s}}{D_{\rm
d}\Sigma_{\infty}}\sqrt{\int_{0}^{\,c}xg(x)^{2}dx}.$ (38)
The weight parameter $Z$ is designed to take into account the impact of the
distribution of the sources (e.g. Seitz & Schneider, 1997). the angular-
diameter distance to the lens is described by $D_{\rm d}$, and
$\Sigma_{\infty}$ is the value of $\Sigma_{\rm crit}$ for a source at a
redshift of infinity. For each halo in the catalogue, we computed its
theoretical S/N ($\nu$) assuming a Euclid-like redshift source distribution
$\rm p_{\rm z_{s}}$ (derived from the simulation), a galaxy density of $\rm
n_{\rm g}=30\,gal.{arcmin}^{-2}$ and a shape noise of
$\sigma_{\rm\epsilon}=0.26$.
The left-hand panel of Fig. 7 shows the resulting ideal weak lensing detection
selection function in the mass-redshift plane for the characteristics of the
simulation. The figure is obtained by generating a grid in the mass-redshift
plane and computing the S/N value for each point of the grid using Eq. 38. The
colour map highlights the increase in S/N towards the upper-left corner,
corresponding to the detection of massive clusters at low redshift. The solid
white lines correspond to S/N contours of level S/N
=[0,1,2,3,4,5,6,7,10,12,15,20,25,30,35], which have been interpolated from the
$(z,M)$ grid points.
However, the above only yields an average value for the S/N of detection
assuming a mean source distribution and an optimal filter. In practice, the
filters are never perfectly adapted to the cluster shape as assumed in the
theoretical S/N estimation. Furthermore, the detection S/N depends on the
redshift distribution of the sources that are behind the cluster. The galaxies
behind the cluster being few in number, small variations in their redshift can
artificially boost or decrease the weak lensing signal.
Figure 8: Dispersion of the association distance (without maximum association
distance) for the W3 wavelet filter, considering all 256 realisations. In all
the panels, the blue dots correspond to an association between a peak
detection and its closest halo. Upper left: Absolute association distance,
$D$, as a function of its normalised expression with respect to the
characteristic halo radius, $\theta_{200c}$. Lower right: Zoomed-in view of
the central $20\times 20$ arcminute distribution of absolute association
distances expressed in RA and Dec, to highlight their dispersion. The two
remaining panels show the projection of the absolute association distance
distribution onto the RA and Dec axes: the projection onto the Dec axis
(lower-left panel) and the projection onto the RA axis (upper-right panel).
The RA and Dec axes are zoomed around the central area. On these two panels,
the dashed orange line shows the result of a fit by a Gaussian function whose
parameters ($\mu$, $\sigma$) are labelled ‘Mean’ and ‘Stddev’.
We applied a conservative threshold to the 256 DM halo catalogues by keeping
all the haloes with S/N$>2$. Although a halo with S/N = 2 is unlikely to be
detected, we applied this value to take into account the variance in the
detection due to the distribution of the sources, and the use of an ideal
filter. Below this value we do not expect to detect many clusters. We discuss
this choice of pre-selection threshold further in Sect 8.3.2. With this pre-
selection, in practice, from a typical $5^{\circ}\times 5^{\circ}$ field, only
2-3% of the $15,000$ haloes in the catalogue are kept. The remaining $300-400$
haloes are then the basis for the matching procedure. The right-hand panel of
Fig. 7 shows the selected haloes in the mass-redshift plane for one
$5^{\circ}\times 5^{\circ}$ field.
Figure 9: Matched detections for the W3 filter. The two panels are similar to
Fig. 8. The blue dots correspond to an association between a peak detection
and its closest halo. The orange dots correspond to the detections with an
absolute association distance lower than the defined MMD.
### 7.2 Matching procedure
The matching procedure between the catalogue haloes and the peak candidates is
a complex task. The typical approach consists of evaluating the distance
between the candidate peak positions and the closest halo, and to consider
these matched if the distance is below a given maximum matching distance
(MMD).
In practice, most matching methods simply draw a circle of a given apparent
angular radius around the candidate peak and match the closest halo that falls
in the circle (e.g. Hamana et al., 2004; Gavazzi & Soucail, 2006; Hamana et
al., 2015, 2020). However, in some procedures the matching is more refined,
drawing a circle of a given comoving distance (in Mpc) around the halo and
then matching the detection with highest S/N within this radius (e.g. Miyazaki
et al., 2018; Oguri et al., 2021). The latter is more precise, but requires
access to the redshift of the haloes in the catalogue.
Our study is further complicated by the characteristics of the different
filter functions, and the inherent multi-scale nature of the wavelet approach.
We therefore used a different method, drawing on SZ cluster survey matching
procedures (see e.g. Planck Collaboration et al., 2014). In our study, the
matching is undertaken by computing the apparent angular distance $D$
(hereafter the association distance) between each candidate peak and each halo
in the DM halo catalogue, after S/N pre-selection. This association distance
is then normalised by the apparent halo radius $\theta_{200c}$. This
normalised distance $D/\theta_{200c}$, together with the MMD defined below,
are our matching criteria. Together, they take into account the filter size
and have the effect of favouring the closest and most massive clusters in the
matching.
The matching procedure is implemented iteratively. For $N$ detections and $M$
haloes, we obtain an array of $N\times M$ measured distances. We start the
iterative process for a given candidate peak by selecting the minimum
normalised distance $D/\theta_{200c}$ among all the measured normalised
distances. The detection and halo corresponding to this minimum normalised
distance are considered as matched and the corresponding row and column are
removed from the array. The procedure is repeated until we reach the defined
MMD. This procedure is perfectly suited to cross-match detections with
simulated halo catalogues, yielding excellent results compared to typical
methods. However, for application to cross-matching with real cluster
catalogues (e.g. optical, X-ray, etc.) access to an estimate of the $\rm
R_{200c}$ (or $\rm M_{200c}$) of the haloes is required.
### 7.3 Maximum matching distance
The choice of the MMD is critical, and must take the positional precision of
the detection method into account, as larger filter sizes will lead to a loss
of precision on the coordinates of the peak centre. At the same time, if the
chosen MMD is too large, false associations may be introduced even after pre-
selecting the halo catalogue, while associations may be missed if the MMD is
too small. Therefore, the MMD must be optimised and adapted to the detection
method in question.
For this study, we developed an empirical method that is able to adapt the MMD
to the detection method. Our approach derives the MMD from the uncertainty in
the position of the detection introduced by the size of the filter. In
practice, this is undertaken by performing an association procedure on all the
256 fields, which corresponds to applying the matching procedure without any
limit in the matching distance.
Figure 8 shows an example of the distribution of the corresponding association
distances for all the 256 realisations (about $75,000$ haloes after catalogue
pre-selection) for the filter W3, for which the filter radius equals
$2\aas@@fstack{\prime}$32\. Each blue dot corresponds to a pairing between a
peak detection and its closest halo. The upper-left panel shows the absolute
association distance $D$, expressed in arcminutes, as a function of the
association distance $D$ normalised by the angular halo radius
$\theta_{200c}$. We clearly distinguish two distinct groups of associations.
In the upper group, the detections are outside the characteristic halo radius.
Therefore, we can consider this group to be false associations. The lower
group is composed of associations for which the peak position is within the
$\theta_{200c}$ radius of the halo and can thus be considered as true
associations. To isolate the correct associations and define a correct MMD, we
use a different visualisation of the results. In the lower-right panel of Fig.
8, the blue dots represent a unique peak detection-halo pair whose association
distance $D$ has been decomposed in terms of RA and Dec (we show only the
central $20\times 20$ arcminutes). The high-density central region corresponds
to peak detections that have been correctly matched to an underlying halo. The
dispersion in this high-density central region in the RA-Dec plane highlights
the uncertainties in the position of the peak detection introduced by the size
of the filter.
To highlight the properties of the dispersion, we projected the distribution
on both RA and Dec axes, displayed for the W3 filter in the upper-right and
lower-left panels of Fig. 8, respectively. The positional uncertainty is well
approximated by a Gaussian distribution. In practice, we obtain the same
parameters if we fit a 2D Gaussian distribution to the RA-Dec plane or a 1D
Gaussian distribution separately to the RA and Dec components. From the fit we
extracted the standard deviation $\sigma$, which corresponds to the dispersion
of the positional uncertainty of the detection with respect to the true halo
position. The Gaussian fit shows a y-offset that can be explained by randomly
matched haloes. The Gaussian fit also shows an x-offset corresponding, to the
mean of the Gaussian distribution $\mu$, which is due to the pixelisation.
From these parameters, we defined the MMD to be $5\sigma\,+\,\lvert\mu\rvert$.
We then applied this procedure to all the filters, and the resulting MMDs for
each detection method are shown in Table 1. Comparing these values with the
values of the filter radii shown in Fig. 3, we see that the MMD is directly
proportional to the filter radius, as expected.
Table 1: Estimated maximum matching distance (MMD) for different filter functions, evaluated using the procedure described in Sect. 7.3. Filter function | $\sigma$ | $\lvert\mu\rvert$ | MMD [arcmin]
---|---|---|---
S96 | 0.83 | 0.26 | 4.4
TANH | 0.47 | 0.28 | 2.6
J04 | 0.87 | 0.31 | 4.7
M18 | 0.60 | 0.28 | 3.3
W2 | 0.23 | 0.27 | 1.7
W3 | 0.45 | 0.30 | 2.5
W4 | 0.89 | 0.29 | 4.7
Table 2: Completeness of the different detection methods for different S/N
halo pre-selections, in percent.
Filter function | Completeness (S/N¿2) | Completeness (S/N¿5) | Completeness (S/N¿7) | Completeness (S/N¿10)
---|---|---|---|---
S96 | $13.3\pm 0.3$ | $65.6\pm 0.1$ | $89.3\pm 0.1$ | $98.1\pm 0.1$
TANH | $11.9\pm 0.2$ | $62.3\pm 0.1$ | $85.3\pm 0.1$ | $95.8\pm 0.1$
J04 | $12.9\pm 0.2$ | $59.7\pm 0.2$ | $87.7\pm 0.1$ | $97.4\pm 0.1$
M18 | $17.1\pm 0.2$ | $71.5\pm 0.1$ | $93.9\pm 0.1$ | $98.1\pm 0.1$
W4 | $12.7\pm 0.2$ | $60.0\pm 0.2$ | $89.1\pm 0.1$ | $98.2\pm 0.1$
W234 | $21.0\pm 0.2$ | $75.3\pm 0.1$ | $95.7\pm 0.1$ | $99.2\pm 0.1$
444The completeness is computed on each field of 5∘ $\times$ 5∘ using
different S/N halo pre-selection in the matching procedure, with S/N ¿ 2, 5, 7
and 10. The results give the mean and associated uncertainty in the
completeness (in percent), estimated from the 256 realisations.
Once the MMDs are defined, we can apply the matching procedure defined in
Sect. 7.2. Figure 9 shows the results of the matching for the W3 filter. The
blue dots still refer to the associations from all the detections and their
closest halo. The orange dots correspond to the detections that have been
successfully matched. Considering only the pairs for which the association
distance is lower than the MMD allows us to separate the two groups of
associations distinctly. This further allows us to build two catalogues: one
of matched detections, and another of unmatched detections.
### 7.4 Multi-scale matching procedure
In the multi-scale approach, a same cluster can be detected on several
successive scales, complicating the matching analysis. To assess the
performance of the multi-scale approach, it is important to recombine the
multiple detections and to associate them with one single scale (the finest
scale), as explained in Sect. 5.3.
In practice, the multi-scale matching procedure works as follows. Starting
with the finest scale, the matching procedure is performed using the
appropriate matching distance defined in Table 1. Associations at that scale
are considered to be unique, and these haloes are removed from the catalogue
that is used on subsequent scales. This procedure is repeated for each
subsequent scale.
The final result is a catalogue of associated and non-associated detections
for each wavelet scale. These are important to measure the individual
performance of each wavelet scale, and to compare their overall contribution
in terms of detections and associations. The final catalogues of associated
and non-associated detections are then obtained by concatenation of the
individual single-scale catalogues.
## 8 Results and performance
We now assess and compare the performance of the detection methods. We start
by quantifying the completeness and purity of each sample, as commonly done by
other studies (e.g. Miyazaki et al., 2018; Euclid Collaboration et al., 2019;
Hamana et al., 2020). Then we analyse in more detail our detections by
performing an analysis of the distribution of the matched detections in terms
of redshift, mass and scale. We then quantify the number of false detections
due to the shape noise for each method. Finally, we undertake a
characterisation and a classification of the matched and unmatched detections.
For the wavelet filters, we undertake the above both on one single scale (W4),
and using the full multi-scale approach (W234).
### 8.1 Detection method performance
#### 8.1.1 Completeness
To quantify the performance of the different methods and to compare them, we
used the completeness, C, defined as follows:
$\rm
C=\frac{Number\,\,of\,\,matched\,\,detections}{Number\,\,of\,\,clusters\,\,in\,\,the\,\,halo\,\,catalogue}.$
(39)
We note that the completeness depends on the characteristics of the halo
catalogue used for matching. Changing the pre-selection S/N in the halo
catalogue will change the values of the completeness for each method. However,
this does not affect the comparison between the methods.
To further develop our analysis, we computed the completeness for each
detection method using different S/N thresholds in the halo catalogue. For
each method, we computed the completeness of the 256 realisations considering
only the haloes in the catalogue whose theoretical S/N is above a given value.
We repeated this operation for different S/N limits. The results, averaged
over the 256 realisations, are shown in Table 2. We see that considering the
haloes with S/N$>2$ results in a completeness below 25% for all the detection
methods. As expected, the completeness increases with the halo pre-selection
S/N threshold, and almost all of the highest S/N clusters are detected by the
different detection methods. Even if the overall behaviour is comparable, this
quantification shows that the detection methods based on the W234 and M18
filters outperform those based on the S96, TANH, and J04 filters in terms of
completeness.
#### 8.1.2 Purity
Although the completeness for a given S/N halo selection is a good indicator
of the performance of the different detection methods, it does not provide any
information on the proportion of unmatched detections. That is why a measure
of the purity of the sample is also needed to refine the comparison. The
purity, P, is defined as
$\rm
P=\frac{Number\,\,of\,\,matched\,\,detections}{Total\,\,number\,\,of\,\,detections}.$
(40)
This quantity measures the proportion of matched detections in the sample
(e.g. Euclid Collaboration et al., 2019) and can be used to evaluate the false
detection rate. We computed the purity of the samples obtained from each
detection method for a fixed catalogue pre-selection threshold of S/N$>2$.
This threshold was chosen so as to not decrease the purity artificially as
explained in Sect. 7.1. The results for the different detection methods,
averaged over the 256 realisations, are summarised in Table 3. The estimated
purity for all detection methods is comparable, at around 85%. This is
somewhat lower than the expected number from the false detections estimated in
Sect. 8.2.1. This led us to perform a case-by-case analysis of the unmatched
detections, as detailed in Sect. 8.3.2. In particular, we observed that for
some detections the matching procedure has failed or is too conservative.
Although this obviously has an impact on the estimated purity, it does not
affect the comparison between the methods because it affects all of them to
the same extent.
#### 8.1.3 Purity versus completeness
Table 3: Detection method performance.
Filter function | Number of detections | Purity P [%] | Completeness C [%]
---|---|---|---
S96 | $57.6\pm 1.0$ | $83.7\pm 0.3$ | $13.3\pm 0.3$
TANH | $50.9\pm 0.9$ | $84.5\pm 0.4$ | $11.9\pm 0.2$
J04 | $47.9\pm 0.5$ | $88.3\pm 0.3$ | $12.9\pm 0.2$
M18 | $66.6\pm 0.7$ | $83.1\pm 0.3$ | $17.1\pm 0.2$
W4 | $48.0\pm 0.6$ | $88.3\pm 0.3$ | $12.7\pm 0.2$
W234 | $83.2\pm 0.9$ | $82.7\pm 0.3$ | $21.0\pm 0.2$
555 For each detection method, the detections are obtained applying the
detection procedure described in Sect. 5. The purity and the completeness are
measured using the matching described in Sect. 7 using a S/N$>2$ pre-selection
in the halo catalogue. The results correspond to the mean and its
uncertainties (in percent) estimated from the 256 realisations.
Comparing the results in Table 3, we see that the detection methods based on
the S96, TANH, J04, and W4 filters lead to higher purity and lower
completeness. Although the completeness is not directly linked to the number
of detections, but rather to the number of matched detections, the lower
completeness for these filters is consistent with their lower mean number of
detections compared to the two other methods, and also with their Fourier
space representation, whose band-pass width is narrower. We note that the
wavelet filter W4 alone attains similar performance to the single-scale
filters S96, TANH and J04. Conversely, the completeness is higher and the
purity is lower for the M18 and W234 filters, which use larger band-pass
filters. However, the multi-scale approach spans a wider range of possible
cluster scales, and therefore allows us to reach a higher completeness. As a
consequence, the multi-scale approach results in a higher mean number of
detections for the same purity. Compared to the M18 AM filter, for instance,
the multi-scale approach yields 25% more detections at a purity of $\sim
85\%$.
Table 3 also shows that some detection methods are more efficient in terms of
completeness, while others are more efficient in terms of purity. Since the
purity and completeness cannot be compared separately, a fair comparison
between the detection methods requires an analysis of the evolution of the
completeness as a function of the purity. We therefore computed this for each
detection method using the pre-selected S/N$>2$ catalogue for the matching. In
practice, we varied the threshold in the detection procedure described in
Sect. 5.2 in steps of $5\%$ and obtained new values of the completeness and
purity for each new detection threshold value.
Figure 10: Completeness as a function of purity for the detection methods
based on the S96 (blue), TANH (red), J04 (purple), and M18 (green) AM filters,
the W4 wavelet filter (light blue), and the W234 wavelet filter (orange). The
cross markers represent precisely where the methods are evaluated in terms of
purity and completeness. The purity and the completeness are computed using a
S/N$>2$ halo pre-selection and averaging over the 256 realisations.
The results, averaged over the 256 realisations, are shown in Fig. 10. We see
that the multi-scale approach outperforms the other detection methods at all
values of purity. Compared to the detection method based on the TANH filter,
there is an increase in the completeness of up to 65% at low purity. The
increase in completeness is less for the detection method based on the M18
filter, but it remains significant in the range of purities studied. We also
note that the performance of the wavelet filter W4 alone and the single-scale
filter J04 are comparable.
### 8.2 Analysis of the detections
Here we estimate the number of false detections due to the shape noise, and
characterise the matched detections in terms of mass, redshift and scale.
Figure 11: Distribution of the matched clusters in the $z-M$ plane for one
$5^{\circ}\times 5^{\circ}$ patch, using the S96 AM filter in the upper-left
panel, the TANH AM filter in the upper-right panel, the J04 AM filter in the
middle-left panel, the M18 filter in the middle-right panel, the W4 wavelet
filter in the bottom-left panel, and the W234 wavelet filters in the bottom-
right panel. The light blue dots correspond to all the haloes within the halo
catalogue, and the dark blue dots to the haloes we have pre-selected to have
S/N$>2$. The overplotted orange crosses correspond to the haloes that have
been matched with a detection, and in parentheses we note the number of
matches for each method on one patch. This patch was chosen to be
representative of the performance of the different methods in terms of
detection number.
#### 8.2.1 False detections
In the following we define false detections as those detections not related to
any lensing signal (i.e. overdensity), and therefore due only to the noise.
Then, we quantify the number of false detections due to the shape noise for
each detection method. For this purpose, we randomised the orientation of the
sources in the noisy galaxy catalogues to suppress the lensing signal and to
generate pure noise galaxy catalogues. We then derived the associated
convergence maps as described in Sect. 5 and applied the various detection
methods. For each method, we then computed the mean false detection rate
averaged over the 256 realisations. This was obtained from the ratio of the
number of detections due purely to the shape noise to the number of detections
obtained when the signal is not suppressed. The resulting false detection rate
due to the shape noise is given, for each detection method, in the second
column of Table 4.
The false detection rate due to shape noise is $1.1\%$ for the J04 detection
method, $1.8\%$ for the detection methods based on the S96, TANH and W234
filters, and 3% for the M18 detection method. The impact of the shape noise
depends on the filter. The noise level is lower at large scales compared to
smaller scales, which explains why the J$04$ detection method is less
sensitive to the noise compared to the other methods. In contrast, the M$18$
detection method is the most sensitive because it includes small scales.
Similarly, the oscillation behaviour of the S96 and TANH filters increases
their sensitivity to the noise, by integrating smaller scales. The W$234$
detection method also includes small scales, but the impact of the shape noise
is reduced owing to its multi-scale detection procedure. Despite these
differences, the number of false detections due to the shape noise is small,
irrespective of detection method.
#### 8.2.2 Characterisation of the matched detections
Once the matching is completed, we characterise the matched detections by
plotting their distribution in the redshift-mass plane. Figure 11 shows the
distribution of the clusters detected by the five different detection methods
in the redshift-mass plane. The figure is obtained from one single
realisation, but is representative of the full 256 realisation sample. The
matching is performed using only the haloes with S/N$>2$.
As expected, all the detection methods are most sensitive to the signal from
high mass clusters lying at low redshift. Comparison shows that the multi-
scale detection method W234 has more matched detections than the others, in
agreement with the completeness results presented in Table 2. In addition,
almost all the clusters detected by the S96, TANH, J04, and M18 detection
methods are also detected by the multi-scale detection procedure. Compared to
the four other detection methods, the multi-scale approach generally detects
more clusters at lower masses, and up to higher redshifts. In Appendix A we
further analyse the multi-scale approach by investigating the contribution of
each wavelet scale to the total number of matched detections.
A striking aspect of Fig. 11 is that some haloes with a very high predicted
S/N are not detected, regardless of the method used. We performed a case-by-
case analysis of these haloes and we found several explanations for this. In
some cases, the detection is missed because of the shape noise. In other
cases, the detection is missed because the lensing signal is decreased by the
redshift distribution of the sources behind the cluster (through Eqs. 9 and
10). In some other cases, the cluster is detected but the matching procedure
fails because the normalised distance is not always the best criterion.
### 8.3 Classification of the detections
We now undertake a further analysis of the matched and unmatched detections by
classifying them into different categories. In particular, Fig. 12 summarises
these categories and the results for the W234 method, after detection and
matching of all 256 realisations.
#### 8.3.1 Classification of the matched detections
The matched detections can be classified into two categories.
Table 4: Classification of the detections.
Filter function | Matched [%] | False detections [%] | Boosted objects ($\rm p_{z}$) [%] | Blended haloes [%] | Others [%]
---|---|---|---|---|---
S96 | $83.7\pm 0.3$ (48) | $2.0\pm 0.2$ (2) | $3.5\pm 0.2$ (2) | $5.8\pm 0.3$ (3) | $\sim 5.0(3)$
TANH | $84.5\pm 0.3$ (43) | $1.7\pm 0.3$ (1) | $4.3\pm 0.3$ (2) | $6.2\pm 0.3$ (3) | $\sim 3.3(2)$
J04 | $88.3\pm 0.3$ (42) | $1.1\pm 0.2$ (1) | $3.6\pm 0.2$ (2) | $5.9\pm 0.4$ (2) | $\sim 1.1$ (1)
M18 | $83.1\pm 0.3$ (56) | $3.0\pm 0.3$ (2) | $4.0\pm 0.1$ (3) | $5.6\pm 0.3$ (3) | $\sim 4.3$ (3)
W234 | $82.7\pm 0.3$ (67) | $1.8\pm 0.3$ (1) | $6.4\pm 0.2$ (6) | $5.5\pm 0.2$ (5) | $\sim 3.6$ (4)
666The detections have been classified in the categories described in Sect.
8.3.1 and 8.3.2, for each detection method and averaged on each field of size
5∘ $\times$ 5∘. The first column gives the proportion of matched detection
using a S/N¿2 pre-selected halo catalogue in the matching procedure described
in Sect. 7. Then, the unmatched detections have been classified in the
categories described in Sect. 8.3.2. The false detections due to shape noise
are in the second column. The unmatched detections due to haloes whose lensing
signal has been boosted are in the third column. The unmatched detections due
to blended lensing signal from LOS alignments or the merging of two haloes of
S/N¿1.5 are in the fourth column. The last column gives the proportion of
unmatched detections that remain unclassified. These results correspond to the
mean and associated uncertainty of the detections that fall into these
categories. The numbers in parentheses correspond to the average number of
detections in each category in a 5∘ $\times$ 5∘ realisation (rounded up to an
integer number).
True detections correctly matched. Almost all the matched detections
correspond to correct associations to a single halo of the catalogue (‘Single
matched’ in Fig. 12). However, it is inevitable that some true detections can
be correctly matched but should in fact be associated with two (or possibly
more) haloes whose lensing signal is merged. This may occur when two haloes
are aligned in the LOS or when their apparent angular distance is small
compared to the size of the filter. In this case, the matching procedure does
not differentiate and the association is undertaken simply with the closest
halo. We estimate that about 5% of the matched detections fall in this
category by searching for multiple haloes in the vicinity of the matched
detections (’Multiple Matched (LOS)’ in Fig. 12). These imperfect associations
can slightly reduce the estimated completeness, but all the methods are
impacted in the same way by this effect.
False detections incorrectly matched. Among the matched detections, a small
fraction are in fact randomly-matched false detections. These are detections
due to the shape noise that are matched owing to being close to a halo in the
catalogue. Their number depends both on the number of false detections (see
Sect. 8.2.1) and on the false association rate (given in Table 5). In this
study, we intentionally keep this number very small for all the detection
methods (less than 0.3 %). However, one must be careful because both the
detection and matching procedures can make this number become large. Thus, it
is essential to always compute this number to fully assess the performance of
the detection methods.
#### 8.3.2 Classification of the unmatched detections
Here we focus on the detections that are not matched with a halo, to fully
complete the analysis of the different detection methods. These unmatched
detections can be classified into several categories.
False detections unmatched correctly. The most intuitive class for of
unmatched detections is that composed of those due to the shape noise (see
Sect. 8.2.1.) Beyond the small fraction that is incorrectly matched owing to
being located close to a halo in the catalogue (see above), almost all of
these false detections are not matched. Their number thus corresponds
approximately to the false detection rate, and is given in the second column
of Table 4. This number is very small compared to the estimated purity in
Sect. 8.1, and other explanations need to be found to explain the remaining
$\sim 15\%$ of the unmatched detections.
Table 5: False association rate for different filter functions.
Filter function | False association rate [%] |
---|---|---
S96 | $12.8\pm 0.3$ |
TANH | $9.0\pm 0.3$ |
J04 | $14.2\pm 0.4$ |
M18 | $10.6\pm 0.4$ |
W234 | $10.8\pm 0.4$ |
W2 | $2.8\pm 0.3$ |
W3 | $8.7\pm 0.4$ |
W4 | $13.6\pm 0.4$ |
777The false association rate is computed for each detection method including
each individual wavelet filter W2, W3 and W4. The fraction of false
associations is estimated by matching random detection positions with the halo
catalogues with a S/N$>2$ halo pre-selection. The errors are computed using
the 256 realisations.
Boosted lensing signal unmatched incorrectly. One possible explanation for the
remaining unmatched detections is that some of these detections should be
associated with haloes that have been removed from the halo catalogue after
the S/N pre-selection. This could happen because the S/N estimated for each
halo in the catalogue to perform the selection is theoretical. It is computed
assuming an average theoretical source density and a theoretical distribution,
$\rm p_{\rm z_{s}}$, for the redshift of the sources. In practice the sources
behind a cluster are few in number. As a consequence their number and their
redshift distribution can deviate significantly from the theoretical values,
and can in some cases boost the lensing signal of a cluster, regardless of its
mass and redshift. At the same time, selecting haloes with a much lower S/N in
the halo catalogue serves to increase the false association rate, and so the
number of false detections matched incorrectly (see Sect 8.3.1). Table 5 gives
the false association rate when using a S/N$>2$ pre-selection in the halo
catalogue. This remains below 15% for all detection methods. Thus, the choice
of imposing a S/N$>2$ pre-selection in the halo catalogue is a trade-off
between the false association rate and the missed association rate.
We estimated the fraction of unmatched detections due to a boosted lensing
signal by undertaking the matching procedure a second time for all 256
realisations, but with a halo pre-selection of S/N$>1.5$. The results for each
method is given in the third column of Table 4.
Blended lensing signal unmatched incorrectly. A further explanation for the
remaining unmatched detections is the case of two haloes contributing to the
same detection:
1. 1.
Blended LOS: The blended lensing signal can be caused by LOS alignments (e.g.
filamentary structures). This means that the detection is coming from two
haloes with S/N$<2$ that are aligned along the LOS and whose lensing signals
cannot be separated without considering the redshift distribution of the
background galaxies in the detection step.
2. 2.
Blended haloes: The blended lensing signal can also be linked to the size of
the filter. Indeed, some unmatched detections can come from two haloes whose
apparent distance is small compared to the filter radius. The lensing signal
of these two haloes being merged within the AM map, there is a single
detection associated with the two haloes, the position of which depends on the
mass and redshift of the two haloes. In this case, the detection is unmatched
when the theoretical S/N of the two haloes is less than 2, or when the
position of the detection is too far away from the centre of the two haloes.
We estimated the proportion of blended lensing signal falling in the two
categories by searching for multiple haloes with S/N $>1.5$ in the vicinity of
the unmatched detections. The proportion of unmatched detections due to the
blending of the signal coming from two (or possibly more) haloes is given in
the fourth column of Table 4. In this class of unmatched detections, we found
that about a third of them are due to intrinsic LOS alignments.
A small fraction of unmatched detections do not fall into the above three
categories. The fraction for each method is listed as ‘others’ in the last
column of Table 4. Most of these remaining detections can be matched with
haloes with S/N$<1.5$. But further decreasing the halo pre-selection threshold
increases the false association rate, making this classification very
uncertain.
Figure 12: Characterisation of the detections from the W234 method. The
matched detections (blue segments) are categorised as correctly matched with
single or multiple haloes, or incorrectly matched as defined in Sect. 8.3.1.
The unmatched detections are sub-categorised as blended signal, boosted signal
from the source distribution, false detections, or others, as defined in Sect.
8.3.2. The results are averaged over the 256 realisations.
## 9 Summary and conclusions
The sensitivity of wide-field optical surveys now allows for the blind
detection of galaxy clusters through their weak lensing signal. This situation
will vastly improve with the launch of Euclid and the Roman Space Telescope.
However, the construction of weak-lensing-selected galaxy cluster catalogues
requires the improvement or development of new galaxy cluster detection
algorithms able to cope with the increasingly large volume of data.
In this paper we have introduced a new, fast, multi-scale approach, based on
wavelet filters operating on complementary scales, for detecting galaxy
clusters through their weak lensing signal. This new method, W234, was
compared to four commonly used approaches (S96, TANH, J04, and M18) based on
AM filters. The comparison was performed on the same set of Euclid-like mocks
obtained from the DUSTGRAIN-pathfinder simulations, composed of 256
realisations of a field of 5∘ $\times$ 5∘, corresponding to a total area of 6
400 deg2.
We first undertook a thorough examination of the filter characteristics in
real and Fourier space. This allowed us to demonstrate the equivalence between
some of the commonly used AM filters (S96, TANH, J04, and M18) and wavelet
filters at individual scales; moreover, we show that the M18 AM filter is
comparable to a combination of two wavelet filters. We then investigated
different options for implementing the detection methods. Applying the
detection methods to the 256 realisations, we find:
* •
Similar results are obtained when applying the detection algorithms to the
shear directly at the galaxy positions in real space or to the binned shear in
Fourier space. Binning a 5∘ $\times$ 5∘ field into $512\times 512$ bins, the
power spectrum integral of the AM map suffers a maximum 1.5% loss when passing
from real to Fourier space.
* •
The binning scheme that is adopted can have an impact on the detection
efficiency in terms of the number and distribution of detections. In
particular, binning by averaging the shear in each pixel yields a smaller
number of detections than binning by summing the shear, and a significant
fraction of these detections ($\sim 28\%$) are different from those detected
in the real space approach. In contrast, the summed pixel binning yields very
similar results, in terms of both number and distribution, to application to
the shear in real space.
* •
Application of the detection methods to the shear or the convergence yields
similar results. In particular, the power spectrum loss on the AM map is
negligible when passing from the shear to the convergence, and the number and
distribution of detections agree exactly.
To quantitatively compare the results obtained from the different filters,
each having different characteristic radii, we developed a new adaptive
matching procedure to match the detections to the haloes in the simulations.
We find:
* •
The characteristics of the halo catalogue used for the matching can have a
dramatic effect on the number of false associations. We thus applied a pre-
selection to the halo catalogue based on the theoretical weak lensing S/N
expected for an NFW halo, and kept only haloes with S/N$>2$ for the subsequent
matching step.
* •
The association distance is a critical parameter when dealing with filters
that have different characteristic radii. It must take both the spatial
resolution of the filter and the angular size of the haloes into account. We
obtained an optimal matching distance for the detections from each filter by
considering the absolute association distance, $D$, of all haloes as a
function of the association distance normalised by the characteristic radius
of each halo, $D/\theta_{200c}$. Then, decomposing the distribution of the
absolute association distance in terms of RA and Dec and fitting with a
Gaussian, we obtained an optimal MMD for each filter.
* •
For the multi-scale wavelet detections, a halo might be detected on more than
one successive scale. We thus carried out a special matching procedure for the
detections from the wavelet filters, building a cumulative catalogue from the
finest to the coarsest scales.
We then compared the performance of the detection methods, after running them
on the 256 simulations and using the matching procedure developed in this
paper. We find:
* •
The completeness is a strong function of the halo catalogue pre-selection
threshold. However, irrespective of the exact S/N pre-selection threshold, the
S96, TANH, and J04 filters exhibit a lower completeness than the M18 and W234
filters. In contrast, the J04 filter exhibits slightly higher purity at the
expense of completeness.
* •
The evolution of the completeness as a function of purity shows that the
multi-scale approach outperforms the methods based on AM filters at all values
of purity.
* •
The multi-scale method detects practically all the detections from the
individual AM approaches, with the addition of supplemental detections,
particularly at lower masses and higher redshifts.
Finally, we performed an exhaustive analysis of the matched and unmatched
detections for each method, using a S/N$>2$ pre-selection on the halo
catalogue. We found similar results for all methods. For the W234 method, we
find:
* •
More than $82\%$ of the detections are correctly matched, and of these, the
overwhelming majority (95%) are due to a single halo. For our matching
procedure, correctly matched detections due to multiple haloes along the LOS
account for less than 5% of the total. Furthermore, because our detection and
matching methods aim at high purity, the number of incorrectly matched haloes
is negligible ($<1\%$).
* •
About $17\%$ of the detections are unmatched. Only a small fraction of these
($10\%$) are due to false detections. The remainder can be explained mostly by
haloes below the S/N threshold whose signal has been boosted because of a
fortuitous redshift distribution of background sources ($37\%$), or a blended
lensing signal from LOS alignments, or angular separation distances smaller
than the filter size ($32\%$).The final 20% of the unmatched detections can
likely be associated with haloes with S/N$<1.5$ that do not appear in the
catalogue due to the pre-selection.
In this work we have shown that a multi-scale approach is faster and gives
better results, in terms of purity and completeness, than the currently used
AM methods. We have also introduced a new adaptative matching procedure that
allows a fair comparison between detection methods that operate on different
scales. A significant advantage of our method is its computational speed.
Compared to methods that apply single-scale AM filters to the shear directly
at the galaxy positions, our multi-scale method applied to the convergence
yields a gain of two to three orders of magnitude in speed.
Based on these results, we aim to apply our multi-scale approach to current
and future large-area survey data, which will allow the construction of large
catalogues of clusters selected through their weak lensing signal. This next
step also involves new challenges. As detailed in Hamana et al. (2020), the
presence of foreground galaxies that are not part of the cluster can dilute
the lensing signal. This dilution effect contaminates the weak lensing dataset
and analyses. One way to tackle this issue is to use photometric redshifts to
select sources and minimise contamination. By taking the redshift distribution
within the dataset into account, the tomographic approach reduces the impact
of the dilution effect. This solution has already shown promising results for
the detection of galaxy clusters in the HSC survey (e.g. Hamana et al., 2020;
Oguri et al., 2021). The tomographic approach may also yield coarse
information on the redshift of the detected sources. We are confident that a
multi-scale approach combined with tomography will improve on the results
shown here. Application to upcoming deep, large-area surveys will allow us to
develop our method further.
###### Acknowledgements.
The authors would like to thank Joel Bergé for providing his code to estimate
the theoretical S/N of detection of a halo by weak lensing, Jean-Baptiste
Melin for his advices and Peter Schneider for his helpful comments on the
filters. GL acknowledges funding from CNES, the French space agency. GC
acknowledges support from INAF theory Grant 2022: Illuminating Dark Matter
using Weak Lensing by Cluster Satellites.
## References
* Aihara et al. (2018) Aihara, H., Armstrong, R., Bickerton, S., et al. 2018, PASJ, 70, S8
* Allen et al. (2011) Allen, S. W., Evrard, A. E., & Mantz, A. B. 2011, Annual Review of Astronomy and Astrophysics, 49, 409–470
* Andreon & Bergé (2012) Andreon, S. & Bergé , J. 2012, Astronomy & Astrophysics, 547, A117
* Bartelmann & Schneider (2001) Bartelmann, M. & Schneider, P. 2001, Physics Reports, 340, 291–472
* Bergé et al. (2010) Bergé, J., Amara, A., & Réfrégier, A. 2010, ApJ, 712, 992
* Bocquet et al. (2019) Bocquet, S., Dietrich, J. P., Schrabback, T., et al. 2019, ApJ, 878, 55
* Closson Ferguson et al. (2009) Closson Ferguson, H., Armus, L., Borne, K., et al. 2009, in American Astronomical Society Meeting Abstracts, Vol. 213, American Astronomical Society Meeting Abstracts #213, 460.07
* Cropper et al. (2013) Cropper, M., Hoekstra, H., Kitching, T., et al. 2013, MNRAS, 431, 3103
* Davis et al. (1985) Davis, M., Efstathiou, G., Frenk, C. S., & White, S. D. M. 1985, ApJ, 292, 371
* Diemer (2018) Diemer, B. 2018, ApJS, 239, 35
* Diemer & Joyce (2019) Diemer, B. & Joyce, M. 2019, ApJ, 871, 168
* Dietrich & Hartlap (2010) Dietrich, J. P. & Hartlap, J. 2010, MNRAS, 402, 1049
* Euclid Collaboration et al. (2019) Euclid Collaboration, Adam, R., Vannier, M., et al. 2019, A&A, 627, A23
* Fan et al. (2010) Fan, Z., Shan, H., & Liu, J. 2010, ApJ, 719, 1408
* Gavazzi & Soucail (2006) Gavazzi, R. & Soucail, G. 2006, Astronomy & Astrophysics, 462, 459–471
* Giocoli et al. (2018) Giocoli, C., Baldi, M., & Moscardini, L. 2018, Monthly Notices of the Royal Astronomical Society, 481, 2813–2828
* Giocoli et al. (2017) Giocoli, C., Di Meo, S., Meneghetti, M., et al. 2017, MNRAS, 470, 3574
* Giocoli et al. (2015) Giocoli, C., Metcalf, R. B., Baldi, M., et al. 2015, MNRAS, 452, 2757
* Hamana et al. (2012) Hamana, T., Oguri, M., Shirasaki, M., & Sato, M. 2012, MNRAS, 425, 2287
* Hamana et al. (2015) Hamana, T., Sakurai, J., Koike, M., & Miller, L. 2015, PASJ, 67, 34
* Hamana et al. (2020) Hamana, T., Shirasaki, M., & Lin, Y.-T. 2020, Publications of the Astronomical Society of Japan, 72
* Hamana et al. (2004) Hamana, T., Takada, M., & Yoshida, N. 2004, MNRAS, 350, 893
* Hasselfield et al. (2013) Hasselfield, M., Hilton, M., Marriage, T. A., et al. 2013, J. Cosmology Astropart. Phys., 2013, 008
* Hennawi & Spergel (2005) Hennawi, J. & Spergel, D. 2005, The Astrophysical Journal, 624
* Hetterscheidt et al. (2005) Hetterscheidt, M., Erben, T., Schneider, P., et al. 2005, A&A, 442, 43
* Ivezić et al. (2019) Ivezić, Ž., Kahn, S. M., Tyson, J. A., et al. 2019, ApJ, 873, 111
* Jarvis et al. (2004) Jarvis, M., Bernstein, G., & Jain, B. 2004, Monthly Notices of the Royal Astronomical Society, 352, 338–352
* Kaiser & Squires (1993) Kaiser, N. & Squires, G. 1993, ApJ, 404, 441
* Lanusse et al. (2016) Lanusse, F., Starck, J. L., Leonard, A., & Pires, S. 2016, A&A, 591, A2
* Laureijs et al. (2011) Laureijs, R., Amiaux, J., Arduini, S., et al. 2011, Euclid Definition Study Report
* Leauthaud et al. (2007) Leauthaud, A., Massey, R., Kneib, J.-P., et al. 2007, ApJS, 172, 219
* Leonard et al. (2015) Leonard, A., Lanusse, F., & Starck, J.-L. 2015, MNRAS, 449, 1146
* Leonard et al. (2012) Leonard, A., Pires, S., & Starck, J.-L. 2012, Monthly Notices of the Royal Astronomical Society, 423, 3405–3412
* Lesci et al. (2022) Lesci, G. F., Marulli, F., Moscardini, L., et al. 2022, A&A, 659, A88
* Lin et al. (2016) Lin, C.-A., Kilbinger, M., & Pires, S. 2016, A&A, 593, A88
* Maturi et al. (2005) Maturi, M., Meneghetti, M., Bartelmann, M., Dolag, K., & Moscardini, L. 2005, Astronomy & Astrophysics, 442, 851–860
* Maturi et al. (2007) Maturi, M., Schirmer, M., Meneghetti, M., Bartelmann, M., & Moscardini, L. 2007, A&A, 462, 473
* Miyazaki et al. (2007) Miyazaki, S., Hamana, T., Ellis, R. S., et al. 2007, ApJ, 669, 714
* Miyazaki et al. (2002) Miyazaki, S., Hamana, T., Shimasaku, K., et al. 2002, ApJ, 580, L97
* Miyazaki et al. (2018) Miyazaki, S., Oguri, M., Hamana, T., et al. 2018, Publications of the Astronomical Society of Japan, 70
* Navarro et al. (1996) Navarro, J. F., Frenk, C. S., & White, S. D. M. 1996, ApJ, 462, 563
* Oguri et al. (2021) Oguri, M., Miyazaki, S., Li, X., et al. 2021, Publications of the Astronomical Society of Japan, 73, 817–829
* Pacaud et al. (2006) Pacaud, F., Pierre, M., Refregier, A., et al. 2006, MNRAS, 372, 578
* Pace et al. (2007) Pace, F., Maturi, M., Meneghetti, M., et al. 2007, A&A, 471, 731
* Perrenod (1980) Perrenod, S. C. 1980, ApJ, 236, 373
* Pires et al. (2012) Pires, S., Leonard, A., & Starck, J.-L. 2012, MNRAS, 423, 983
* Pires et al. (2020) Pires, S., Vandenbussche, V., Kansal, V., et al. 2020, Astronomy & Astrophysics, 638, A141
* Planck Collaboration et al. (2014) Planck Collaboration, Ade, P. A. R., Aghanim, N., et al. 2014, A&A, 571, A29
* Planck Collaboration XIII (2016) Planck Collaboration XIII. 2016, A&A, 594, A13
* Pratt et al. (2019) Pratt, G. W., Arnaud, M., Biviano, A., et al. 2019, Space Science Reviews, 215
* Pyne & Birkinshaw (1993) Pyne, T. & Birkinshaw, M. 1993, ApJ, 415, 459
* Rozo et al. (2010) Rozo, E., Wechsler, R. H., Rykoff, E. S., et al. 2010, ApJ, 708, 645
* Salvati et al. (2021) Salvati, L., Saro, A., Bocquet, S., et al. 2021, arXiv e-prints, arXiv:2112.03606
* Schirmer (2004) Schirmer. 2004, PhD thesis, Rheinische Friedrich-Wilhelms-Universität Bonn
* Schirmer et al. (2007) Schirmer, M., Erben, T., Hetterscheidt, M., & Schneider, P. 2007, A&A, 462, 875
* Schirmer et al. (2004) Schirmer, M., Erben, T., Schneider, P., Wolf, C., & Meisenheimer, K. 2004, A&A, 420, 75
* Schneider (1996) Schneider, P. 1996, Monthly Notices of the Royal Astronomical Society, 283, 837–853
* Schneider & Bartelmann (1997) Schneider, P. & Bartelmann, M. 1997, MNRAS, 286, 696
* Schneider et al. (1992) Schneider, P., Ehlers, J., & Falco, E. E. 1992, Gravitational Lenses
* Schneider et al. (1998) Schneider, P., van Waerbeke, L., Jain, B., & Kruse, G. 1998, MNRAS, 296, 873
* Schneider et al. (2002) Schneider, P., van Waerbeke, L., Kilbinger, M., & Mellier, Y. 2002, A&A, 396, 1
* Schrabback et al. (2015) Schrabback, T., Hilbert, S., Hoekstra, H., et al. 2015, MNRAS, 454, 1432
* Schrabback et al. (2018) Schrabback, T., Schirmer, M., van der Burg, R. F. J., et al. 2018, A&A, 610, A85
* Seitz & Schneider (1997) Seitz, C. & Schneider, P. 1997, A&A, 318, 687
* Shan et al. (2012) Shan, H., Kneib, J.-P., Tao, C., et al. 2012, ApJ, 748, 56
* Shan et al. (2018) Shan, H., Liu, X., Hildebrandt, H., et al. 2018, MNRAS, 474, 1116
* Spergel et al. (2015) Spergel, D., Gehrels, N., Baltay, C., et al. 2015, Wide-Field InfrarRed Survey Telescope-Astrophysics Focused Telescope Assets WFIRST-AFTA 2015 Report
* Springel et al. (2001) Springel, V., White, S. D. M., Tormen, G., & Kauffmann, G. 2001, MNRAS, 328, 726
* Starck et al. (1998) Starck, J., Murtagh, F., & Bijaoui, A. 1998, Image Processing and Data Analysis: The Multiscale Approach (Cambridge University Press)
* Starck et al. (2006) Starck, J. L., Pires, S., & Réfrégier, A. 2006, A&A, 451, 1139
* Tang & Fan (2005) Tang, J. Y. & Fan, Z. H. 2005, ApJ, 635, 60
* Van Waerbeke (1998) Van Waerbeke, L. 1998, A&A, 334, 1
* van Waerbeke (2000) van Waerbeke, L. 2000, MNRAS, 313, 524
* Vikhlinin et al. (2009) Vikhlinin, A., Kravtsov, A. V., Burenin, R. A., et al. 2009, ApJ, 692, 1060
* Voit (2005) Voit, G. M. 2005, Reviews of Modern Physics, 77, 207–258
* White et al. (2002) White, M., van Waerbeke, L., & Mackey, J. 2002, ApJ, 575, 640
* White & Rees (1978) White, S. D. M. & Rees, M. J. 1978, MNRAS, 183, 341
* Wittman et al. (2006) Wittman, D., Dell’Antonio, I. P., Hughes, J. P., et al. 2006, The Astrophysical Journal, 643, 128–143
* Wright & Brainerd (2000) Wright, C. O. & Brainerd, T. G. 2000, ApJ, 534, 34
## Appendix A Contribution of each wavelet scale
In this section we study the contribution of each wavelet scale to the matched
detections made by the multi-scale approach.
Figure 13 shows the distribution of the matched clusters in the $z-M$ plane
for the W2 wavelet filter (upper-left panel), the W3 wavelet filter (upper-
right panel) and the W4 wavelet filter (lower-left panel). The lower-right
panel shows the distribution of the matched clusters in the $z-M$ plane for
the multi-scale detection method. The colours allow us to see the contribution
of each scale after recombination of the multiple detections. The different
panels also highlight the differences in terms of targeted clusters by the
different wavelet scales. In particular, we see the complementarity of the
different wavelet filters.
If we look at the number of matched detections for each wavelet scale after
recombination, we see that the proportion of detections made at each scale
does not vary much from one realisation to another. On average, the W2 wavelet
filter provides up to $29\%\pm 0.39$, W3 around $33\%\pm 0.57,$ and W4 around
$38\%\pm 0.42$ of the matched detections. Hence, the three wavelet scales are
useful and contribute similarly to the total number of matched detections.
Figure 13: Distribution of the matched clusters in the z-M plane for one
$5^{\circ}\times 5^{\circ}$ patch, obtained using the W2 wavelet filter
(upper-left panel), W3 wavelet filter (upper-right panel), W4 wavelet filter
(lower-left panel), and the multi-scale approach (lower-right panel).
|
# Non-imaging metasurface design for collimated beam shaping
Kirstine E. S. Nielsen Department of Physics, Technical University of
Denmark, Fysikvej, DK-2800 Kongens Lyngby, Denmark<EMAIL_ADDRESS>Mads A. Carlsen Department of Physics, Technical University of Denmark,
Fysikvej, DK-2800 Kongens Lyngby, Denmark Søren Raza Department of Physics,
Technical University of Denmark, Fysikvej, DK-2800 Kongens Lyngby, Denmark
<EMAIL_ADDRESS>
## Abstract
Metasurfaces provide a versatile platform for realizing ultrathin flat optics
for use in a wide variety of optical applications. The design process involves
defining or calculating the phase profile of the metasurface that will yield
the desired optical output. Here, we present an inverse design method for
determining the phase profile for shaping the intensity profile of a
collimated incident beam. The model is based on the concept of optimal
transport from non-imaging optics and enables a collimated beam with an
arbitrary intensity profile to be redistributed to a desired output intensity
profile. We derive the model from the generalized law of refraction and
numerically solve the resulting differential equation using a finite-
difference scheme. Through a variety of examples, we show that our approach
accommodates a range of different input and output intensity profiles, and
discuss its feasibility as a design platform for non-imaging optics.
## Introduction
Metasurfaces are gaining a lot of attention in recent years due to their
highly engineerable properties, excellent optical performance and resulting
applications as ultrathin flat optical devices [1, 2, 3, 4, 5, 6, 7, 8, 9,
10]. Their functionality is based on imparting a spatially-dependent phase
delay using judiciously designed nanoantennas [11, 12] to shape the
transmitted [13, 14, 15, 16] or reflected wavefront [17]. The freedom offered
by metasurfaces to implement any phase profile enables the design of flat
free-form optics [18, 19] with new opportunities for non-imaging applications
[20]. Non-imaging free-form optics focuses on the optimal transfer of light
energy from a light source onto a target without requiring the formation of an
image of the source. These optics have important applications in light
concentration and illumination [21].
While determining the phase profiles for imaging optics is well established,
either by analytical expressions [22] or numerical calculations using
commercial ray-tracing software [23] and Fourier methods [24], these methods
are generally not applicable for non-imaging optics. Non-imaging free-form
lenses are formulated as an inverse problem with the aim to redistribute a
light source with a known intensity profile into a desired output intensity
[21]. A variety of different strategies have been implemented to tackle this
inverse problem [25, 26, 27, 28, 29, 21], but a particularly elegant and
general way to do it is through an optimal transport formulation [30, 31].
This method makes no a priori symmetry assumptions on the lens shape and has
been demonstrated to work for a variety of incident light distributions and
output illumination patterns [32].
Here, we demonstrate a metasurface formulation of the optimal transport
approach to determine the phase profile for arbitrary one-dimensional
intensity shaping of a collimated beam. Using the generalized law of
refraction [33], we derive a nonlinear differential equation for the
metasurface phase profile, which depends on the intensity profile of the
incident beam and the desired output intensity profile. The differential
equation is transformed into a system of nonlinear algebraic equations using a
finite difference scheme and numerically solved with a root-finding algorithm.
The results from four different combinations of incident and output
illuminations are shown to demonstrate the versatility of the method in the
design of non-imaging metasurfaces.
## Theoretical framework
Figure 1: Schematic representation of the functionality of the non-imaging
metasurface. A normally-incident collimated beam $\boldsymbol{\mathrm{I}}$
with intensity profile $I(x)$ is incident on a metasurface with phase profile
$\phi(x)$. The metasurface, positioned at $z=0$, refracts the beam onto the
target plane positioned a distance $L$ from the metasurface. The refracted ray
$\boldsymbol{\mathrm{O}}$ hits the target plane at a specific coordinate
$t_{x}$ to produce the desired intensity profile $E(t_{x})$. The boundary rays
of the beam are marked with blue.
We consider a normally-incident collimated beam of light with a one-
dimensional intensity profile $I(x)$, which is redistributed by a refracting
metasurface with phase profile $\phi(x)$ to a desired target illumination with
intensity profile $E(t_{x})$ (Fig. 1). The coordinate $t_{x}$ at the target is
located a distance $L$ away from the metasurface and is connected to the
$x$-coordinate at the metasurface through the refraction caused by the
metasurface phase profile. Our aim is to determine the metasurface phase
profile that maps the incident intensity profile to the desired target
intensity profile. This constitutes an inverse problem, which we formulate by
drawing inspiration from the concept of optimal transport in non-imaging free-
form optics [30].
We assume that the metasurface is a refractor, which transmits all incident
light and is lossless. Metasurfaces with near unity transmission and $2\pi$
phase coverage can be realized with high-refractive-index dielectrics using
Huygens’-type [14, 34] and nanopost metaatoms [15]. Energy conservation then
dictates that the power incident on the metasurface is equal to the power
received at the target [35]
$\displaystyle\int_{S_{1}}I(x)\textrm{d}x=\int_{S_{2}}E(t_{x})\textrm{d}t_{x}=\int_{S_{1}}E(t_{x})\left|\frac{\partial
t_{x}}{\partial x}\right|\textrm{d}x,$ (1)
where $S_{1}$ and $S_{2}$ denote the metasurface and target widths,
respectively, and $\left|\frac{\partial t_{x}}{\partial x}\right|$ is the
Jacobian describing the change of variable from the target coordinate $t_{x}$
to the metasurface coordinate $x$. Physically, the Jacobian describes the ray
expansion or contraction due to the refraction by the metasurface. Equation
(1) requires that the power is locally conserved
$\displaystyle E(t_{x})\left|\frac{\partial t_{x}}{\partial x}\right|=I(x).$
(2)
The aim is to establish a coordinate relationship between the metasurface and
target planes. From geometrical considerations (Fig. 1), we obtain
$t_{x}=x+L\frac{O_{x}}{O_{z}},$ (3)
where $\boldsymbol{\mathrm{O}}=(O_{x},O_{z})$ is the unit vector describing
the direction of the refracted ray. The direction of the refracted ray is
determined by its wave vector $\boldsymbol{\mathrm{k_{r}}}$, which, in turn,
is found from the generalized law of refraction [33, 36]. Since the incident
beam impinges normally and the metasurface is flat, the direction of the
refracted ray is solely determined by the metasurface phase gradient as
$\boldsymbol{\mathrm{O}}=\frac{\boldsymbol{\mathrm{k_{r}}}}{|\boldsymbol{\mathrm{k_{r}}}|}=\frac{1}{k}(k_{x},k_{z})=\left(\frac{\partial\tilde{\phi}}{\partial
x},\sqrt{1-\left(\frac{\partial\tilde{\phi}}{\partial x}\right)^{2}}\right).$
(4)
Here, we have introduced a scaled wavelength-independent phase
$\tilde{\phi}=\phi/(n_{0}k_{0})$, where $k_{0}$ is the vacuum wave number and
$n_{0}$ is the refractive index of the medium between the metasurface and the
target plane. Combining Eqs. (3-4), we evaluate the Jacobian in Eq. (2) to
arrive at
$\displaystyle E(t_{x})\left|1+L\frac{\partial^{2}\tilde{\phi}}{\partial
x^{2}}\left[1-\left(\frac{\partial\tilde{\phi}}{\partial
x}\right)^{2}\right]^{-\frac{3}{2}}\right|-I(x)=0.$ (5)
Equation (5) is our main result and is a nonlinear differential equation for
the metasurface phase profile, given a known input intensity profile $I(x)$
and a desired target intensity profile $E(t_{x})$. We note that only the phase
gradient $\partial\tilde{\phi}/\partial x$ impacts the beam shaping, as the
phase itself does not explicitly appear in Eq. (5). We exploit this in the
numerical implementation described in Sec. Numerical implementation.
As the final step we treat the boundaries of the system. The boundary
condition is defined by the edge-ray principle, which is the basis for most
non-imaging design problems. The edge-ray principle states that a mapping of
all rays from the edge of the source distribution should be directed to the
edge of the target illumination distribution, in order to ensure that all rays
fall within the target [37]. In our case this condition is represented by the
coordinate mapping in Eq. (3), which at $x=x_{\text{min}}$ (Fig. 1) requires
the phase gradient to satisfy the boundary condition
$\displaystyle\left.\frac{\partial\tilde{\phi}}{\partial
x}\right|_{x=x_{\text{min}}}$
$\displaystyle=A_{\text{min}}\left(1+A_{\text{min}}^{2}\right)^{-\frac{1}{2}},\,\,\,\,\,\,\,\,A_{\text{min}}=\frac{t_{x,\text{min}}-x_{\text{min}}}{L},$
(6)
and similarly at $x=x_{\text{max}}$ for $A_{\text{max}}$ and
$t_{x,\text{max}}$.
The theoretical framework given by Eqs. (5-6) describes the metasurface phase
profile needed to transport the incident intensity to the target intensity.
The nonlinear nature of Eq. (5) renders a general analytical solution
unfeasible. Instead, we transform the nonlinear differential equation into a
system of nonlinear algebraic equations using a finite difference scheme, and
solve the resulting system of equations numerically.
## Numerical implementation
We noted that the phase does not explicitly appear in Eqs. (5-6), and thus the
problem can be simplified by solving for the phase gradient directly. We
discretize the metasurface into $N$ grid points in the interval
$\left[-\tfrac{X}{2};\tfrac{X}{2}\right]$ with a spacing $h=X/(N-1)$, where
$X$ is the width of the metasurface. Denoting the scaled phase gradient as
$y\equiv\partial\tilde{\phi}/\partial x$ and employing a centered finite
difference scheme [38], Eq. (5) turns into the following system of algebraic
equations for the interior grid points
$\displaystyle F_{i}^{\text{int}}$
$\displaystyle=E(t_{x}(y_{i}))\left|1+L\frac{y_{i+1}-y_{i-1}}{2h(1-y_{i}^{2})^{\frac{3}{2}}}\right|-I(x_{i})=0,$
(7)
where $t_{x}(y_{i})=x_{i}+Ly_{i}/\sqrt{1-y_{i}^{2}}$. The boundary points
$i=1,N$ described by Eq. (6) are given by
$\displaystyle F_{i}^{\text{BC}}$
$\displaystyle=y_{i}-A_{i}\left(1+A_{i}^{2}\right)^{-\frac{1}{2}}=0.$ (8)
These discretized equations describe a system of $N$ equations with $N$
variables, which we solve numerically using a root-finding algorithm
implemented in the commercial software MATLAB. The implementation is based on
the Levenberg–Marquardt algorithm, which employs the Jacobian matrix for fast
and robust convergence. The elements $J_{i,j}=\partial F_{i}/\partial y_{j}$
in the Jacobian matrix for our system of equations [Eqs. (7-8)] are only non-
zero for $j=i+1,\,i,\,i-1$ for the interior points and for $j=i$ on the
boundary, making the Jacobian matrix sparse. For the interior points we find
the Jacobian elements
$\displaystyle J_{i,i}$
$\displaystyle=E^{\prime}(t_{x}(y_{i}))\frac{L|K_{i}|}{\left(1-y_{i}^{2}\right)^{\frac{3}{2}}}+E(t_{x}(y_{i}))\frac{3L(y_{i+1}-y_{i-1})y_{i}}{2h\left(1-y_{i}^{2}\right)^{\frac{5}{2}}}\frac{K_{i}}{|K_{i}|},$
(9) $\displaystyle J_{i,i+1}$
$\displaystyle=-J_{i,i-1}=\frac{E(t_{x}(y_{i}))L}{2h\left(1-y_{i}^{2}\right)^{\frac{3}{2}}}\frac{K_{i}}{|K_{i}|},$
(10)
where we have defined
$K_{i}=1+L\frac{y_{i+1}-y_{i-1}}{2h(1-y_{i}^{2})^{\frac{3}{2}}}$. In Eq. (9),
prime denotes the differentiation with respect to the argument. For the
boundary points described by Eq. (8), the Jacobian elements are given by
$J_{i,i}=1$.
The iterative root-finding algorithm based on Eqs. (7-10) requires an initial
guess for the phase gradient. For complex beam shaping, the phase gradient may
vary in a manner that is a priori difficult to guess. To alleviate this
problem, we use numerical continuation where we iteratively increase the
complexity of the beam shaping, using in each iteration the obtained phase
gradient from the previous iteration as an initial guess. For example, in the
Gaussian beam shaping presented in Sec. Design examples, our starting point is
a broad Gaussian profile (close to constant) which we gradually sharpen by
decreasing the linewidth to obtain the desired beam shaping funcitonality. We
find that this remedies the dependence on providing an accurate initial guess
and ensures robust convergence. With this numerical implementation we obtain
the phase gradient to which we fit a function using linear interpolation and
perform integration to extract the phase profile.
Figure 2: (a) Scaled phase profiles and phase gradients, and (b) resulting ray
plots for a focusing non-imaging metasurface. (c-d) Similar to (a-b) for a
defocusing metasurface. In (a,c) the dashed black lines indicate the
analytical phase profile [Eq. 11], while in (b,d) they indicate the boundary
rays. The ray plots (b,d) show that the calculated phase profile indeed
achieves the desired focus (b) and defocus (d) effect.
## Design examples
We verify the theoretical model by testing four beam shaping functionalities.
In the following examples, the metasurface width is $X=1$ mm and the distance
between the metasurface and the target is $L=10$ cm. First we demonstrate
focusing and defocusing functionalities which can be verified analytically
(Fig. 2). For focusing, the target width $T$ is set to 1000 times smaller than
the metasurface width [Fig. 2(a-b)], while defocusing is modelled by setting
the target width to $T=15X$, i.e., 15 times larger than the metasurface width
[Fig. 2(c-d)]. The incident beam and target illumination both have constant
intensity profiles, and we ensure energy conservation by scaling the target
intensity profile using Eq. (1). In these cases the scaled phase profile of
the metasurface is given by [22]
$\tilde{\phi}(x)=\pm\left(\sqrt{x^{2}+f^{2}}-f\right).$ (11)
The $-$ ($+$) solution in Eq. (11) gives a focusing (defocusing) phase profile
with the focal length given by $f=L$ [$f=LX/(T-X)$]. Consequently, we expect
convex and concave phase distributions for the metasurface in the focus and
defocus cases, respectively. For the focusing case [Fig. 2(a)], we indeed
obtain a convex phase profile in excellent agreement with Eq. (11). The ray
plot in Fig. 2(b) shows to which point $t_{x}$ on the target plane a ray
incident on the metasurface in point $x$ will be refracted. The target points
$t_{x}$ are evaluated using Eq. (3) with the numerically obtained phase
gradients as input. The ray plots visualize the redistribution of the incident
beam onto the target plane and show that the boundary rays are indeed mapped
to the boundary of the target in accordance with the edge-ray principle. For
the defocus case [Fig. 2(c-d)], the numerically obtained concave phase profile
is again in agreement with the analytical expression. These intuitive cases
demonstrate that our model produces results in agreement with existing theory.
Figure 3: (a) Incident Gaussian intensity profile $I(x)$ and constant target
intensity $E(t_{x})$. (b) Numerically calculated scaled phase profile and
scaled phase gradient. (c) Comparison between the intended target intensity
$E(t_{x})$ (black dashed line) and the resulting target intensity profile
based on the calculated scaled phase gradient (green). (d) Ray distribution
calculated from the obtained phase gradient, which illustrates the refraction
of the incident beam by the metasurface phase profile to obtain the desired
constant intensity.
From here we move to more complex beam shaping, where the input has a Gaussian
intensity profile which is mapped to a constant target intensity profile [Fig.
3(a)]. The metasurface and target widths are kept the same. We choose a
configuration where the input intensity is not centered with respect to the
metasurface, in order to stress the model. In Fig. 3b we show the numerically
obtained phase and phase gradient profiles. We find a convex-type phase
profile, which is reasonable as the beam shaping taking place is similar to
the defocus case [Fig. 2(a)]. The rays are directed towards the perimeter of
the target plane, which is consistent with the flattening of the Gaussian beam
shape [Fig. 3(d)]. We also check that the calculated phase gradient produces
the expected target intensity profile by using Eq. (2). The calculated target
intensity profile is in excellent agreement with the desired target intensity
[Fig. 3(c)].
Finally, we investigate the inverse case where the target is a Gaussian
intensity profile $E(t_{x})=E_{0}e^{-t_{x}^{2}/2b^{2}}$ with $b=0.2$ mm, while
the input intensity profile remains constant [Fig. 4(a)]. In Fig. 4(b), we see
that the phase profile has a concave shape similar to the focus case [Fig.
2(a)], which is reasonable since the constant incident beam is focused to a
Gaussian target illumination [Fig. 4(c-d)]. However, the phase gradient
differs from the focusing case as the edge rays at the metasurface must be
mapped to the edge of the target. This forces the phase gradient to go to zero
at the edges of the metasurface, since the target and metasurface widths are
identical. This example serves to show that rather complex beam shaping can be
handled by our theoretical framework.
Figure 4: (a) Incident constant intensity $I(x)$ and Gaussian target intensity
profile $E(t_{x})$. (b) Numerically calculated scaled phase profile and scaled
phase gradient. (c) The intended target intensity $E(t_{x})$ (black dashed
line) and the resulting target intensity profile based on the calculated phase
gradient (green). (d) Ray distribution calculated from the obtained phase
gradient.
## Metasurface design and efficiency
The theoretical model and its numerical implementation determines a
wavelength-independent scaled phase and phase gradient profile for the desired
beam shaping. We can retrieve the actual phase profile of the metasurface at
any desired operating wavelength $\lambda$ by the operation
$\phi(x)=2\pi/(n_{0}\lambda)\tilde{\phi}(x)$. Fig. 5 shows half of the
symmetric phase profile for the defocus case [Fig. 2(c)] calculated for
$\lambda=660$ nm and $\lambda=1550$ nm and wrapped on the interval $[0,2\pi]$.
Here, we have assumed that the rays travel through air ($n_{0}=1$). Since a
metasurface is made up of a periodic array of nanostructures, which each
provide a specific phase shift to incident light, we need to discretize and
wrap the phase profile in modulo of $2\pi$. This gets increasingly tougher for
shorter wavelengths, as they yield larger phase gradients (Fig. 5). It is
advantageous to have as many phase levels in the discretization as possible in
order to best approximate the continuous phase profile [16, 39]. However, for
simple metalenses 4 phase levels have proven to give sufficient resolution and
efficiency [40, 41]. For more advanced lens designs as many as 8 phase levels
are required [42] , and for production of detailed holograms 8-32 phase levels
have been proven necessary to achieve the required resolution [43, 44, 45]. In
our case the steepest phase gradient is found for $\lambda=660$ nm at the very
edge of the metasurface where the largest refraction occurs (Fig. 5). Here we
need to cover the full $2\pi$ phase range over 9 µm. The metasurface period
needs to be subwavelength in order to limit diffraction, meaning that 16 phase
levels are reasonable to achieve with a theoretical efficiency of 99% [16].
For more complex beam shaping with larger phase gradients the phase
discretization impacts the achievable efficiency of the metasurface.
Figure 5: Phase profiles for the defocus lens described in Fig. 2 at the
wavelengths $\lambda=660$ nm and $\lambda=1550$ nm. Since the phase profile is
symmetric, only one half is plotted. The phase is wrapped on the interval
$[0,2\pi]$ to illustrate the phase changes. The shortest distance over which
the phase goes from $0$ to $2\pi$ is 9 µm (23 µm) for $\lambda=660$ nm
($\lambda=1550$ nm).
## Conclusion
In summary we have demonstrated a theoretical optimal transport framework for
obtaining the phase profile of a metasurface for arbitrary one-dimensional
beam shaping with collimated incident light. We show how the resulting
nonlinear differential equation can be solved numerically using a finite
difference scheme and a root-finding algorithm. Four specific cases are
investigated to demonstrate the versatility of the method within the design of
non-imaging metasurfaces. The resulting phase profiles are smooth and our
results are visualized through ray tracing. We find that the phase profiles
can be realized with a reasonable phase resolution.
### Funding
S. R. acknowledges support by the Independent Research Funding Denmark
(7026-00117B).
### Acknowledgments
We thank Hugh Simons and Gor Nahapetyan for assistance with the numerical
implementation.
### Disclosures
The authors declare no conflicts of interest.
### Data availability
Data underlying the results presented in this paper are not publicly available
at this time but may be obtained from the authors upon reasonable request.
## References
* [1] R. Paniagua-Domínguez, Y. F. Yu, E. Khaidarov, S. Choi, V. Leong, R. M. Bakker, X. Liang, Y. H. Fu, V. Valuckas, L. A. Krivitsky, and A. I. Kuznetsov, “A metalens with a near-unity numerical aperture,” Nano Letters, vol. 18, pp. 2124–2132, Mar 2014.
* [2] Z. Zhou, J. Li, R. Su, B. Yao, H. Fang, K. Li, L. Zhou, J. Liu, D. Stellinga, C. P. Reardon, T. F. Krauss, and X. Wang, “Efficient silicon metasurfaces for visible light,” ACS Photonics, vol. 4, pp. 544–551, Jan 2017.
* [3] J. van de Groep, J.-H. Song, U. Celano, Q. Li, P. G. Kik, and M. L. Brongersma, “Exciton resonance tuning of an atomically thin lens,” Nature Photonics, vol. 14, pp. 426–430, Jul 2020.
* [4] M. Lawrence, D. R. Barton, J. Dixon, J.-H. Song, J. van de Groep, M. L. Brongersma, and J. A. Dionne, “High quality factor phase gradient metasurfaces,” Nature Nanotechnology, vol. 15, pp. 956–961, Nov 2020.
* [5] Z. Cai, Y. Deng, C. Wu, C. Meng, Y. Ding, S. I. Bozhevolnyi, and F. Ding, “Dual-functional optical waveplates based on gap-surface plasmon metasurfaces,” Advanced Optical Materials, vol. 9, no. 11, p. 2002253, 2021\.
* [6] Y. Zhou, I. I. Kravchenko, H. Wang, J. R. Nolen, G. Gu, and J. Valentine, “Multilayer noninteracting dielectric metasurfaces for multiwavelength metaoptics,” Nano Letters, vol. 18, no. 12, pp. 7529–7537, 2018.
* [7] A. Weiss, C. Frydendahl, J. Bar-David, R. Zektzer, E. Edrei, J. Engelberg, N. Mazurski, B. Desiatov, and U. Levy, “Tunable metasurface using thin-film lithium niobate in the telecom regime,” ACS Photonics, vol. 9, no. 2, pp. 605–612, 2022.
* [8] I. Kim, R. J. Martins, J. Jang, T. Badloe, S. Khadir, H.-Y. Jung, H. Kim, J. Kim, P. Genevet, and J. Rho, “Nanophotonics for light detection and ranging technology,” Nature Nanotechnology, vol. 16, pp. 508–524, May 2021\.
* [9] J. Engelberg and U. Levy, “The advantages of metalenses over diffractive lenses,” Nature Communications, vol. 11, p. 1991, Apr 2020.
* [10] S. Wang, P. C. Wu, V.-C. Su, Y.-C. Lai, C. Hung Chu, J.-W. Chen, S.-H. Lu, J. Chen, B. Xu, C.-H. Kuan, T. Li, S. Zhu, and D. P. Tsai, “Broadband achromatic optical metasurface devices,” Nature Communications, vol. 8, p. 187, Aug 2017.
* [11] S. M. Kamali, E. Arbabi, A. Arbabi, and A. Faraon, “A review of dielectric optical metasurfaces for wavefront control,” Nanophotonics, vol. 7, no. 6, pp. 1041–1068, 2018.
* [12] W. T. Chen, A. Y. Zhu, and F. Capasso, “Flat optics with dispersion-engineered metasurfaces,” Nat. Rev. Mater., vol. 5, no. 8, pp. 604–620, 2020.
* [13] N. Yu, P. Genevet, M. a. Kats, F. Aieta, J.-P. Tetienne, F. Capasso, and Z. Gaburro, “Light Propagation with Phase Reflection and Refraction,” Science, vol. 334, pp. 333–337, 2011.
* [14] M. Decker, I. Staude, M. Falkner, J. Dominguez, D. N. Neshev, I. Brener, T. Pertsch, and Y. S. Kivshar, “High-efficiency dielectric Huygens’ surfaces,” Advanced Optical Materials, vol. 3, pp. 813–820, Feb 2015.
* [15] A. Arbabi, Y. Horie, M. Bagheri, and A. Faraon, “Dielectric metasurfaces for complete control of phase and polarization with subwavelength spatial resolution and high transmission,” Nat. Nanotechnol., vol. 10, no. 11, pp. 937–943, 2015.
* [16] P. Lalanne and P. Chavel, “Metalenses at visible wavelengths: past, present, perspectives,” Laser Photon. Rev., vol. 11, p. 1600295, May 2017.
* [17] A. Pors, O. Albrektsen, I. P. Radko, and S. I. Bozhevolnyi, “Gap plasmon-based metasurfaces for total control of reflected light,” Sci. Rep., vol. 3, no. 2155, p. 2155, 2013.
* [18] A. Zhan, S. Colburn, C. M. Dodson, and A. Majumdar, “Metasurface Freeform Nanophotonics,” Sci. Rep., vol. 7, p. 1673, May 2017.
* [19] D. K. Nikolov, A. Bauer, F. Cheng, H. Kato, A. Nick Vamivakas, and J. P. Rolland, “Metaform optics: Bridging nanophotonics and freeform optics,” Sci. Adv., vol. 7, no. eabe5112, p. eabe5112, 2021.
* [20] I. Moreno, M. Avendaño-Alejo, and C. P. Castañeda-Almanza, “Nonimaging metaoptics,” Opt. Lett., vol. 45, no. 10, p. 2744, 2020\.
* [21] R. Wu, Z. Feng, Z. Zheng, R. Liang, P. Benítez, J. C. Miñano, and F. Duerr, “Design of Freeform Illumination Optics,” Laser Photon. Rev., vol. 12, no. 7, p. 1700310, 2018.
* [22] F. Aieta, P. Genevet, M. A. Kats, N. Yu, R. Blanchard, Z. Gaburro, and F. Capasso, “Aberration-free ultrathin flat lenses and axicons at telecom wavelengths based on plasmonic metasurfaces,” Nano Letters, vol. 12, pp. 4932–4936, Aug 2012.
* [23] A. Arbabi, E. Arbabi, S. M. Kamali, Y. Horie, S. Han, and A. Faraon, “Miniature optical planar camera based on a wide-angle metasurface doublet corrected for monochromatic aberrations,” Nat. Commun., vol. 7, p. 13682, 2016.
* [24] B. Desiatov, N. Mazurski, Y. Fainman, and U. Levy, “Polarization selective beam shaping using nanoscale dielectric metasurfaces,” Opt. Express, vol. 23, no. 17, pp. 22611–22618, 2015.
* [25] H. Ries and J. Muschaweck, “Tailored freeform optical surfaces,” J. Opt. Soc. Am. A, vol. 19, pp. 590–595, Mar 2002.
* [26] Y. Ding, X. Liu, Z. Zheng, and P. Gu, “Freeform LED lens for uniform illumination,” Opt. Express, vol. 16, pp. 12958–12966, Aug 2008.
* [27] L. Sun, S. Jin, and S. Cen, “Free-form microlens for illumination applications,” Appl. Opt., vol. 48, pp. 5520–5527, Oct 2009.
* [28] F. R. Fournier, W. J. Cassarly, and J. P. Rolland, “Fast freeform reflector generation using source-target maps,” Opt. Express, vol. 18, pp. 5295–5304, Mar 2010.
* [29] J.-J. Chen, T.-Y. Wang, K.-L. Huang, T.-S. Liu, M.-D. Tsai, and C.-T. Lin, “Freeform lens design for LED collimating illumination,” Opt. Express, vol. 20, pp. 10984–10995, May 2012.
* [30] R. Wu, L. Xu, P. Liu, Y. Zhang, Z. Zheng, H. Li, and X. Liu, “Freeform illumination design: a nonlinear boundary problem for the elliptic Monge-Ampere equation,” Opt. Lett., vol. 38, pp. 229–231, Jan 2013\.
* [31] C. E. Gutiérrez and L. Pallucchini, “Reflection and refraction problems for metasurfaces related to Monge-Ampere equations,” J. Opt. Soc. Am. A, vol. 35, pp. 1523–1531, Sep 2018.
* [32] R. Wu, P. Benítez, Y. Zhang, and J. C. Miñano, “Influence of the characteristics of a light source and target on the Monge-Ampere equation method in freeform optics design,” Opt. Lett., vol. 39, pp. 634–637, Feb 2014.
* [33] F. Aieta, P. Genevet, N. Yu, M. A. Kats, Z. Gaburro, and F. Capasso, “Out-of-plane reflection and refraction of light by anisotropic optical antenna metasurfaces with phase discontinuities,” Nano Letters, vol. 12, pp. 1702–1706, Feb 2012.
* [34] S. Raza, “Slow light using magnetic and electric Mie resonances,” Opt. Lett., vol. 45, no. 5, pp. 1260–1263, 2020.
* [35] R. Wu, P. Liu, Y. Zhang, Z. Zheng, H. Li, and X. Liu, “A mathematical model of the single freeform surface design for collimated beam shaping,” Opt. Express, vol. 21, pp. 20974–20989, Sep 2013.
* [36] C. P. Castañeda-Almanza and I. Moreno, “Ray tracing in metasurfaces,” Opt. Continuum, vol. 1, pp. 958–964, May 2022.
* [37] H. Ries and A. Rabl, “Edge-ray principle of nonimaging optics,” J. Opt. Soc. Am. A, vol. 11, pp. 2627–2632, Oct 1994.
* [38] S. Chakraverty, N. R. Mahato, P. Karunakar, and T. D. Rao, Advanced numerical and semi-analytical methods for differential equations. Hoboken, New Jersey: John Wiley & Sons. Inc., 1 ed., 2019.
* [39] G. J. Swanson, “Binary optics technology: The theory and design of multi-level diffractive optical elements,” Tech. Rep. Technical Report 854, Massachusetts Institute of Technology, Lexington, MA, Aug 1989.
* [40] F. Aieta, P. Genevet, M. Kats, and F. Capasso, “Aberrations of flat lenses and aplanatic metasurfaces,” Opt. Express, vol. 21, pp. 31530–31539, Dec 2013\.
* [41] M. Y. Shalaginov, S. An, Y. Zhang, F. Yang, P. Su, V. Liberman, J. B. Chou, C. M. Roberts, M. Kang, C. Rios, Q. Du, C. Fowler, A. Agarwal, K. A. Richardson, C. Rivero-Baleine, H. Zhang, J. Hu, and T. Gu, “Reconfigurable all-dielectric metalens with diffraction-limited performance,” Nat. Comm., vol. 12, p. 1225, Feb 2021.
* [42] S. Gao, C.-S. Park, C. Zhou, S.-S. Lee, and D.-Y. Choi, “Twofold polarization-selective all-dielectric trifoci metalens for linearly polarized visible light,” Advanced Optical Materials, vol. 7, p. 1900883, Nov 2019\.
* [43] J. Scheuer, “Metasurfaces-based holography and beam shaping: engineering the phase profile of light,” Nanophotonics, vol. 6, pp. 137–152, Jan 2017\.
* [44] X. Ni, A. V. Kildishev, and V. M. Shalaev, “Metasurface holograms for visible light,” Nat. Comm., vol. 4, p. 2807, Nov 2013.
* [45] W. Zhao, H. Jiang, B. Liu, J. Song, Y. Jiang, C. Tang, and J. Li, “Dielectric Huygens’ metasurface for high-efficiency hologram operating in transmission mode,” Sci. Rep., vol. 6, p. 30613, Jul 2016.
|
11institutetext: State University of Campinas (UNICAMP) 22institutetext:
Maritaca AI 33institutetext: NeuralMind AI
# BLUEX: A benchmark based on Brazilian Leading Universities Entrance eXams
Thales Sales Almeida 1122 Thiago Laitz 1133 Giovana K. Bonás 11 Rodrigo
Nogueira 1122
###### Abstract
One common trend in recent studies of language models (LMs) is the use of
standardized tests for evaluation. However, despite being the fifth most
spoken language worldwide, few such evaluations have been conducted in
Portuguese. This is mainly due to the lack of high-quality datasets available
to the community for carrying out evaluations in Portuguese. To address this
gap, we introduce the Brazilian Leading Universities Entrance eXams (BLUEX), a
dataset of entrance exams from the two leading universities in Brazil: UNICAMP
and USP. The dataset includes annotated metadata for evaluating the
performance of NLP models on a variety of subjects. Furthermore, BLUEX
includes a collection of recently administered exams that are unlikely to be
included in the training data of many popular LMs as of 2023. The dataset is
also annotated to indicate the position of images in each question, providing
a valuable resource for advancing the state-of-the-art in multimodal language
understanding and reasoning. We describe the creation and characteristics of
BLUEX and establish a benchmark through experiments with state-of-the-art LMs,
demonstrating its potential for advancing the state-of-the-art in natural
language understanding and reasoning in Portuguese. The data and relevant code
can be found at https://github.com/Portuguese-Benchmark-Datasets/BLUEX
## 1 Introduction
Recent advances in Language Models (LMs) have generated significant interest
due to their demonstrated capabilities on a wide range of language tasks,
including text classification, language translation, and text generation [3,
7]. LM performance has been particularly impressive on standardized tests,
which present challenging questions requiring high levels of domain-specific
knowledge and reasoning. For instance, recent benchmarks on GPT-4 [16] showed
that it can achieve human-level performance on a variety of graduate-level
benchmarks.
Despite the impressive performance of LMs on standardized tests, few
evaluations have been performed in Portuguese [15], partially due to the lack
of available datasets in the language. This lack of high-quality, standardized
datasets presents a significant challenge for researchers interested in
developing and evaluating LMs in Portuguese. To address this gap for Brazilian
Portuguese, we introduce BLUEX, a dataset consisting of entrance exams for the
two leading universities in Brazil. Our dataset offers a rich source of high-
quality high school-level questions annotated with their respective subjects,
as well as flags indicating the required capabilities necessary to respond
accurately to the questions, such as knowledge of Brazilian culture and the
application of mathematical reasoning. These annotations can be used to
evaluate the performance of LMs on a variety of subjects and capabilities such
as domain-specific knowledge and reasoning. Additionally, BLUEX includes a
collection of recently administered entrance exams that are unlikely to be
included in the training data of many currently popular LMs.
In anticipation of the emergence of multimodal models that combine text and
image understanding, we have annotated BLUEX to indicate the position of
images in each question. Additionally, we have included all necessary images
with the dataset to facilitate research on multimodal language tasks. We
believe that this resource will be essential in evaluating the performance of
models that reason with both text and image inputs to solve complex problems.
In this paper, we describe the creation and characteristics of BLUEX and
establish a benchmark through experiments with state-of-the-art LMs. Our
findings suggest that BLUEX provides a valuable resource for benchmarking and
advancing the state-of-the-art in natural language understanding and reasoning
in Portuguese. This is particularly relevant since even the current state-of-
the-art models, such as GPT-4, still have considerable room for improvement
and do not achieve the highest cutoff grades for both universities.
## 2 Related Work
In the realm of Portuguese Natural Language Processing (NLP) datasets, there
appears to be a limited availability.
For question-answering tasks, Faquad [21] is available, which exhibits an
extractive style akin to SQuAD [18]. It features questions concerning
Brazilian higher education institutions, with documents sourced from a federal
university and supplemented by Wikipedia articles. Another option is the
Multilingual Knowledge Questions and Answers (MKQA) dataset, which covers 26
languages [12]. This dataset was generated by selecting 10,000 queries from
the Natural Questions dataset [10] and acquiring new passage-independent
answers for each question. Subsequently, human translators translated the
questions and answers into 25 non-English, typologically diverse languages,
including Portuguese.
Regarding sentence entailment tasks, ASSIN 1 and 2 [5, 19] are available.
These datasets encompass Recognizing Textual Entailment (RTE), also referred
to as Natural Language Inference (NLI), and Semantic Textual Similarity (STS)
tasks. The former involves predicting if a given text (premise) implies
another text (hypothesis), while the latter quantifies the semantic
equivalence between two sentences.
The Portuguese Language Understanding Evaluation (PLUE) benchmark [6] provides
Portuguese translations of the GLUE [26], SNLI [1], and SciTAIL [8] datasets.
These translations have been generated using automatic translation tools
including Google Translate and OpusMT [24].
The Winograd Schema Challenge (WSC) dataset [9] contains pairs of sentences
with minimal differences, featuring an ambiguous pronoun that is resolved
divergently between the two sentences. Melo et al. [13] manually translated
and adapted this dataset to Portuguese.
For sentiment analysis tasks, the TweetsentBr dataset [2] consists of 15,000
tweets related to the TV show domain, collected between January and July 2017.
The tweets were manually annotated by seven annotators into three classes:
positive, neutral, and negative.
The Multilingual Amazon Slu resource package (SLURP) for Slot-filling, Intent
classification, and Virtual assistant Evaluation (MASSIVE) [4] is a 1M-example
dataset containing realistic virtual utterances in 51 languages, including
Portuguese. Professional translators translated the dataset from English, and
it is annotated for slot (55 classes) and intent (60 classes) prediction
tasks.
A dataset more closely related to BLUEX is the ENEM-challenge dataset [22],
which includes the editions of the Brazilian national exam, Exame Nacional do
Ensino Medio (ENEM), from 2009 to 2017. Additionally, Nunes et al. [15]
introduced a dataset containing the ENEM exam of 2022, the same paper
evaluated the performance of LMs such as GPT-3.5-Turbo and GPT-4 on both the
ENEM-challenge and the ENEM 2022 datasets.
## 3 The BLUEX Dataset
### 3.1 Dataset Creation
BLUEX is a dataset comprising more than 1,000 multiple choice questions from
the entrance exams of the two leading universities in Brazil, Unicamp and USP,
administered between 2018 and 2023. The dataset was created by automatically
extracting each question text, alternatives, and related images using scripts,
and subsequently each example was manually annotated to correct extraction
errors and provide additional metadata such as image positioning.
### 3.2 Annotated Question Metadata
The annotated metadata is described below.
* •
Prior Knowledge (PRK) \- Indicates whether the question requires knowledge
from outside of what has been provided in the question, such as familiarity
with a particular author’s work or a specific mathematical formula.
* •
Text Understanding (TU) \- Indicates whether the question requires
understanding of a particular text.
* •
Image Understanding (IU) \- Indicates whether the question requires
understanding of an image. It should be noted that not all questions with
images require their understanding to answer the question.
* •
Mathematical Reasoning (MR) \- Indicates whether the question requires
mathematical reasoning, such as the ability to perform calculations and
symbolic manipulations.
* •
Multilingual (ML) \- Indicates whether the question requires knowledge of two
or more languages, such as questions designed to test English skills of
Portuguese speakers.
* •
Brazilian Knowledge (BK) \- Indicates whether the question involves knowledge
specific to Brazil, such as Brazilian history, literature, geography, or
culture.
* •
Subjects \- A list of subjects related to the question, such as geography,
physics, etc.
* •
Related Images \- A list of all the related images for the question.
* •
Alternative Type \- Indicates whether the answer choices are presented as text
or as images. This is important because some questions may use images as
answer choices, which requires different processing techniques than questions
with only textual answers.
By providing such annotations along with the questions we aim to facilitate
research into language understanding and reasoning in Portuguese for both pure
language models and multimodal models. We believe that BLUEX will be a
valuable resource for researchers to evaluate and improve the performance of
future language models in the context of Portuguese-language standardized
tests.
### 3.3 Image Positioning
Many of the questions in the exams require a contextual or informational
understanding of images. Despite active research in the field of multimodal
models, models that can adeptly process both text and image data and yield
satisfactory results remain scarce in the public domain. We believe that BLUEX
can serve as an essential evaluation tool for such models. Anticipating the
use of models that will process images and text in an interleaved manner, we
also provide precise information regarding the placement of images within the
question, as illustrated in Figure 1.
Figure 1: Example of image annotation in BLUEX.
### 3.4 Dataset Distribution
The BLUEX dataset covers a wide range of high school subjects, including
Mathematics, Physics, Chemistry, Biology, History, Geography, English,
Philosophy and Portuguese, as well as multidisciplinary questions that involve
two or more subjects. The distribution of questions is shown in Table 1, where
we also provide the distribution for the subset of questions without images,
which accounts for approximately 58% of the total dataset.
Furthermore, Table 2 shows the distribution of the dataset across annotated
categories, as explained in Section 3.2. We observe that the majority of
questions require specific knowledge and the ability to comprehend text, two
expected capabilities in students taking these exams. Note that any given
question can be part of multiple categories.
biology chemistry english geography history mathematics philosophy physics
portuguese multidisciplinary Total UNICAMP 60 45 57 51 60 89 1 61 86 46 556
USP 50 63 41 55 63 69 4 63 88 43 539 BLUEX 110 108 98 106 123 158 5 124 174 89
1095 No images UNICAMP 35 15 20 22 49 64 1 36 65 31 338 USP 23 15 25 16 52 36
4 26 80 23 300 BLUEX 58 30 45 38 101 100 5 62 145 54 638
Table 1: Distribution over subjects.
DS TU IU MR ML BK UNICAMP 431 440 160 209 60 69 USP 446 442 203 174 43 63
BLUEX 877 882 363 383 103 132 No Images UNICAMP 273 282 0 118 23 46 USP 237
269 0 70 25 43 BLUEX 510 551 0 188 48 89
Table 2: Distribution over categories.
## 4 Results
Model BLUEX UNICAMP USP MR BK Highest Cutoff Score 0.863 0.855 0.872 - -
Average Human Score 0.521 0.530 0.511 - - Random 0.220 0.250 0.200 0.223 0.228
GPT-4 [16] 0.748 0.749 0.747 0.447 0.854 Sabiá 65B [17] 0.632 0.615 0.650
0.239 0.775 GPT-3.5-Turbo 0.582 0.580 0.583 0.277 0.764 LLaMA 65B [25] 0.542
0.530 0.557 0.271 0.652 OPT 66B [30] 0.223 0.246 0.197 0.186 0.258 Sabiá 7B
[17] 0.466 0.494 0.433 0.25 0.551 Alpaca 7B [23] 0.284 0.308 0.257 0.261 0.258
BloomZ 7B [14] 0.284 0.275 0.293 0.17 0.326 LLaMA 7B [25] 0.255 0.275 0.233
0.255 0.247 Bertin 6B [20] 0.241 0.293 0.183 0.261 0.315 Bloom 7B [29] 0.238
0.302 0.167 0.255 0.281 XGLM 7.5B [11] 0.205 0.219 0.19 0.213 0.202 OPT 6.7B
[30] 0.205 0.240 0.167 0.207 0.281 GPT-J 6B [27] 0.197 0.222 0.17 0.186 0.236
Table 3: Accuracy in the BLUEX dataset.
To enable future comparisons, we evaluated our dataset using several language
models, ranging from 6B to 66B parameters, including OpenAI’s GPT-4 and
GPT-3.5-Turbo models. Our experiments were conducted using large language
models with no specific training for this task. Each model was provided with
one example in the input and then asked to answer a question from the test
set. The example was randomly selected from an exam of the same university as
the current question, but from a different year. For example, if the current
question is from UNICAMP 2019, the example provided in the prompt would be a
question from a UNICAMP exam, but not from 2019. We excluded all questions
containing images from our experiments since the language models we used can
only process text. This resulted in a total of 638 questions being used, which
corresponds to approximately 60% of the dataset
Table 3 summarizes our experimental findings, including the mean score
achieved by exam-taking students, as well as the mean cutoff score of the most
competitive major, which is medicine in both universities.111The average and
cutoff scores are reported by the entities responsible for administering the
exams. The results presented in Table 3 are the average of all the exams
contained in the BLUEX dataset. The BLUEX column shows the accuracy of the
whole subset used in the evaluation, while the UNICAMP and USP columns account
for only the questions from the respective universities. The MR and BK columns
account only for questions that include those categories.
Among the language models tested in the 7B-parameter range, Sabiá [17], a
model further pre-trained in Portuguese, consistently outperformed all other
models, coming close to matching the average human score. Among the open-
source models in the 60B-parameter range, LLaMA 65B [25] significantly
outperformed OPT 66B [30] and achieved similar performance to GPT-3.5-Turbo.
Sabiá 65B achieved better performance than GPT-3.5-Turbo but still lagged
behind GPT-4 by ten points. GPT-4 was by far the best model in our evaluations
but did not achieve an average score high enough to pass in medicine, the most
competitive major. It is worth noting that the average and cutoff scores
provided in Table 3 are computed taking into account the whole exam, including
questions with images, while the scores obtained by the language models
utilize only the subset of questions with no images.
We also conducted a more detailed analysis of the models’ performance by
examining their ability to handle specific question types. Table 3 presents
the findings for questions that required Mathematical Reasoning (MR) and
Brazilian Knowledge (BK). We observe that, with the exception of GPT-4, all
models struggled to perform significantly better than random chance in
questions that required Mathematical Reasoning. Even GPT-4 only achieved an
accuracy of 44% in MR questions. On the other hand, when considering questions
that require brazilian knowledge, Sabiá greatly outperformed all the other
models in the 7B-parameter range, indicating that the extra pretraining in
Portuguese provided the model with additional regional knowledge. In the
60B-parameter range, Sabiá also showed improvement over LLaMA, increasing the
accuracy in these questions by 10 points and slightly outperforming
GPT-3.5-Turbo. Nevertheless, it could not match the remarkable performance of
GPT-4.
Figure 2: Accuracy of the best models over the years of the exams.
Moreover, Figure 2 displays the performance of the top four models on the
exams conducted each year. It can be observed that the models have a small
variance between the years, which is expected as the difficulty of each exam
and the number of questions in the subset vary across years. A surprising
result, however, is the increased performance that all models seem to exhibit
in 2023. The average and highest cutoff scores also increased slightly over
the years, indicating that the exams became slightly easier in recent years.
Since the 2023 exams were very recently administered, it is unlikely that they
are part of any of the studied models’ training data. Therefore, since the
models’ performance in the most recent years is comparable to that in older
exams, it is reasonable to assume that the models are not merely memorizing
the answers for the questions in the dataset.
## 5 Conclusion
This work introduced BLUEX, a new dataset that consists of 13 college entrance
exams applied between 2018 and 2023 from two of the leading Brazilian
universities, UNICAMP and USP. Each question of these exams was extensively
annotated to help measure different abilities across multiple subjects in
Portuguese. Beyond that, by providing images and their corresponding positions
within the text, BLUEX is one of the few Portuguese datasets that are ready to
evaluate multimodal models. We provide results from multiple LMs as baselines
and reference scores based on students performance to facilitate future
comparisons. We believe that BLUEX will be a important benchmark in the
evaluation of the Portuguese capabilities of future models.
## 6 Future Work
The models used in this study employed a single in-context example. However,
there’s room for further investigation, such as determining whether increasing
the number of few-shot examples could boost the performance of each model, as
well as assessing their zero-shot performance. Furthermore, Nunes et al. [15]
showed that GPT-4’s performance on ENEM questions was significantly boosted
when chain-of-thought prompts [28] were used. Adopting a similar approach here
could potentially lead to performance improvement.
Finally, regarding multimodal models, their performance can be assessed
utilizing the BLUEX dataset. This provides an opportunity for researchers to
investigate the models’ capabilities in integrating visual and textual
information to address high school level questions.
## References
* [1] Bowman, S., Angeli, G., Potts, C., Manning, C.D.: A large annotated corpus for learning natural language inference. In: Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing. pp. 632–642 (2015)
* [2] Brum, H.B., das Graças Volpe Nunes, M.: Building a sentiment corpus of tweets in brazilian portuguese (2017)
* [3] Chowdhery, A., Narang, S., Devlin, J., Bosma, M., Mishra, G., Roberts, A., Barham, P., Chung, H.W., Sutton, C., Gehrmann, S., Schuh, P., Shi, K., Tsvyashchenko, S., Maynez, J., Rao, A., Barnes, P., Tay, Y., Shazeer, N., Prabhakaran, V., Reif, E., Du, N., Hutchinson, B., Pope, R., Bradbury, J., Austin, J., Isard, M., Gur-Ari, G., Yin, P., Duke, T., Levskaya, A., Ghemawat, S., Dev, S., Michalewski, H., Garcia, X., Misra, V., Robinson, K., Fedus, L., Zhou, D., Ippolito, D., Luan, D., Lim, H., Zoph, B., Spiridonov, A., Sepassi, R., Dohan, D., Agrawal, S., Omernick, M., Dai, A.M., Pillai, T.S., Pellat, M., Lewkowycz, A., Moreira, E., Child, R., Polozov, O., Lee, K., Zhou, Z., Wang, X., Saeta, B., Diaz, M., Firat, O., Catasta, M., Wei, J., Meier-Hellstern, K., Eck, D., Dean, J., Petrov, S., Fiedel, N.: Palm: Scaling language modeling with pathways (2022)
* [4] FitzGerald, J., Hench, C., Peris, C., Mackie, S., Rottmann, K., Sanchez, A., Nash, A., Urbach, L., Kakarala, V., Singh, R., Ranganath, S., Crist, L., Britan, M., Leeuwis, W., Tur, G., Natarajan, P.: MASSIVE: A 1m-example multilingual natural language understanding dataset with 51 typologically-diverse languages (2022)
* [5] Fonseca, E., Santos, L., Criscuolo, M., Aluisio, S.: Assin: Avaliacao de similaridade semantica e inferencia textual. In: Computational Processing of the Portuguese Language-12th International Conference, Tomar, Portugal. pp. 13–15 (2016)
* [6] GOMES, J.R.S.: Plue: Portuguese language understanding evaluation. https://github.com/jubs12/PLUE (2020)
* [7] Hoffmann, J., Borgeaud, S., Mensch, A., Buchatskaya, E., Cai, T., Rutherford, E., de Las Casas, D., Hendricks, L.A., Welbl, J., Clark, A., Hennigan, T., Noland, E., Millican, K., van den Driessche, G., Damoc, B., Guy, A., Osindero, S., Simonyan, K., Elsen, E., Rae, J.W., Vinyals, O., Sifre, L.: Training compute-optimal large language models (2022)
* [8] Khot, T., Sabharwal, A., Clark, P.: Scitail: A textual entailment dataset from science question answering. In: Proceedings of the AAAI Conference on Artificial Intelligence. vol. 32 (2018)
* [9] Kocijan, V., Lukasiewicz, T., Davis, E., Marcus, G., Morgenstern, L.: A review of winograd schema challenge datasets and approaches. arXiv preprint arXiv:2004.13831 (2020)
* [10] Kwiatkowski, T., Palomaki, J., Redfield, O., Collins, M., Parikh, A., Alberti, C., Epstein, D., Polosukhin, I., Devlin, J., Lee, K., et al.: Natural questions: a benchmark for question answering research. Transactions of the Association for Computational Linguistics 7, 453–466 (2019)
* [11] Lin, X.V., Mihaylov, T., Artetxe, M., Wang, T., Chen, S., Simig, D., Ott, M., Goyal, N., Bhosale, S., Du, J., Pasunuru, R., Shleifer, S., Koura, P.S., Chaudhary, V., O’Horo, B., Wang, J., Zettlemoyer, L., Kozareva, Z., Diab, M., Stoyanov, V., Li, X.: Few-shot learning with multilingual language models (2022)
* [12] Longpre, S., Lu, Y., Daiber, J.: MKQA: A linguistically diverse benchmark for multilingual open domain question answering. Transactions of the Association for Computational Linguistics 9, 1389–1406 (2021)
* [13] de Melo, G., Imaizumi, V., Cozman, F.: Winograd schemas in portuguese. In: Anais do XVI Encontro Nacional de Inteligência Artificial e Computacional. pp. 787–798. SBC (2019)
* [14] Muennighoff, N., Wang, T., Sutawika, L., Roberts, A., Biderman, S., Scao, T.L., Bari, M.S., Shen, S., Yong, Z.X., Schoelkopf, H., Tang, X., Radev, D., Aji, A.F., Almubarak, K., Albanie, S., Alyafeai, Z., Webson, A., Raff, E., Raffel, C.: Crosslingual generalization through multitask finetuning (2022)
* [15] Nunes, D., Primi, R., Pires, R., Lotufo, R., Nogueira, R.: Evaluating gpt-3.5 and gpt-4 models on brazilian university admission exams (2023)
* [16] OpenAI: Gpt-4 technical report (2023)
* [17] Pires, R., Abonizio, H., Almeida, T.S., Nogueira, R.: Sabiá: Portuguese large language models (2023)
* [18] Rajpurkar, P., Zhang, J., Lopyrev, K., Liang, P.: Squad: 100,000+ questions for machine comprehension of text. In: Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing. pp. 2383–2392 (2016)
* [19] Real, L., Fonseca, E., Gonçalo Oliveira, H.: The ASSIN 2 shared task: a quick overview. In: Computational Processing of the Portuguese Language: 14th International Conference, PROPOR 2020, Evora, Portugal, March 2–4, 2020, Proceedings 14. pp. 406–412. Springer (2020)
* [20] de la Rosa, J., Ponferrada, E.G., Villegas, P., de Prado Salas, P.G., Romero, M., Grandury, M.: Bertin: Efficient pre-training of a spanish language model using perplexity sampling (2022)
* [21] Sayama, H.F., Araujo, A.V., Fernandes, E.R.: Faquad: Reading comprehension dataset in the domain of brazilian higher education. In: 2019 8th Brazilian Conference on Intelligent Systems (BRACIS). pp. 443–448. IEEE (2019)
* [22] Silveira, I.C., Mauá, D.D.: Advances in automatically solving the enem. In: 2018 7th Brazilian Conference on Intelligent Systems (BRACIS). pp. 43–48. IEEE (2018)
* [23] Taori, R., Gulrajani, I., Zhang, T., Dubois, Y., Li, X., Guestrin, C., Liang, P., Hashimoto, T.B.: Stanford alpaca: An instruction-following llama model. https://github.com/tatsu-lab/stanford_alpaca (2023)
* [24] Tiedemann, J., Thottingal, S.: OPUS-MT — Building open translation services for the World. In: Proceedings of the 22nd Annual Conferenec of the European Association for Machine Translation (EAMT). Lisbon, Portugal (2020)
* [25] Touvron, H., Lavril, T., Izacard, G., Martinet, X., Lachaux, M.A., Lacroix, T., Rozière, B., Goyal, N., Hambro, E., Azhar, F., Rodriguez, A., Joulin, A., Grave, E., Lample, G.: Llama: Open and efficient foundation language models (2023)
* [26] Wang, A., Singh, A., Michael, J., Hill, F., Levy, O., Bowman, S.R.: GLUE: A multi-task benchmark and analysis platform for natural language understanding. In: International Conference on Learning Representations (2019), https://openreview.net/forum?id=rJ4km2R5t7
* [27] Wang, B., Komatsuzaki, A.: GPT-J-6B: A 6 Billion Parameter Autoregressive Language Model. https://github.com/kingoflolz/mesh-transformer-jax (May 2021)
* [28] Wei, J., Wang, X., Schuurmans, D., Bosma, M., brian ichter, Xia, F., Chi, E.H., Le, Q.V., Zhou, D.: Chain of thought prompting elicits reasoning in large language models. In: Oh, A.H., Agarwal, A., Belgrave, D., Cho, K. (eds.) Advances in Neural Information Processing Systems (2022), https://openreview.net/forum?id=_VjQlMeSB_J
* [29] Workshop, B., :, Scao, T.L., Fan, A., Akiki, C., Pavlick, E., Ilić, S., Hesslow, D., Castagné, R., Luccioni, A.S., Yvon, F., Gallé, M., Tow, J., Rush, A.M., Biderman, S., Webson, A., Ammanamanchi, P.S., Wang, T., Sagot, B., Muennighoff, N., del Moral, A.V., Ruwase, O., Bawden, R., Bekman, S., McMillan-Major, A., Beltagy, I., Nguyen, H., Saulnier, L., Tan, S., Suarez, P.O., Sanh, V., Laurençon, H., Jernite, Y., Launay, J., Mitchell, M., Raffel, C., Gokaslan, A., Simhi, A., Soroa, A., Aji, A.F., Alfassy, A., Rogers, A., Nitzav, A.K., Xu, C., Mou, C., Emezue, C., Klamm, C., Leong, C., van Strien, D., Adelani, D.I., Radev, D., Ponferrada, E.G., Levkovizh, E., Kim, E., Natan, E.B., Toni, F.D., Dupont, G., Kruszewski, G., Pistilli, G., Elsahar, H., Benyamina, H., Tran, H., Yu, I., Abdulmumin, I., Johnson, I., Gonzalez-Dios, I., de la Rosa, J., Chim, J., Dodge, J., Zhu, J., Chang, J., Frohberg, J., Tobing, J., Bhattacharjee, J., Almubarak, K., Chen, K., Lo, K., Werra, L.V., Weber, L., Phan, L., allal, L.B., Tanguy, L., Dey, M., Muñoz, M.R., Masoud, M., Grandury, M., Šaško, M., Huang, M., Coavoux, M., Singh, M., Jiang, M.T.J., Vu, M.C., Jauhar, M.A., Ghaleb, M., Subramani, N., Kassner, N., Khamis, N., Nguyen, O., Espejel, O., de Gibert, O., Villegas, P., Henderson, P., Colombo, P., Amuok, P., Lhoest, Q., Harliman, R., Bommasani, R., López, R.L., Ribeiro, R., Osei, S., Pyysalo, S., Nagel, S., Bose, S., Muhammad, S.H., Sharma, S., Longpre, S., Nikpoor, S., Silberberg, S., Pai, S., Zink, S., Torrent, T.T., Schick, T., Thrush, T., Danchev, V., Nikoulina, V., Laippala, V., Lepercq, V., Prabhu, V., Alyafeai, Z., Talat, Z., Raja, A., Heinzerling, B., Si, C., Taşar, D.E., Salesky, E., Mielke, S.J., Lee, W.Y., Sharma, A., Santilli, A., Chaffin, A., Stiegler, A., Datta, D., Szczechla, E., Chhablani, G., Wang, H., Pandey, H., Strobelt, H., Fries, J.A., Rozen, J., Gao, L., Sutawika, L., Bari, M.S., Al-shaibani, M.S., Manica, M., Nayak, N., Teehan, R., Albanie, S., Shen, S., Ben-David, S., Bach, S.H., Kim, T., Bers, T., Fevry, T., Neeraj, T., Thakker, U., Raunak, V., Tang, X., Yong, Z.X., Sun, Z., Brody, S., Uri, Y., Tojarieh, H., Roberts, A., Chung, H.W., Tae, J., Phang, J., Press, O., Li, C., Narayanan, D., Bourfoune, H., Casper, J., Rasley, J., Ryabinin, M., Mishra, M., Zhang, M., Shoeybi, M., Peyrounette, M., Patry, N., Tazi, N., Sanseviero, O., von Platen, P., Cornette, P., Lavallée, P.F., Lacroix, R., Rajbhandari, S., Gandhi, S., Smith, S., Requena, S., Patil, S., Dettmers, T., Baruwa, A., Singh, A., Cheveleva, A., Ligozat, A.L., Subramonian, A., Névéol, A., Lovering, C., Garrette, D., Tunuguntla, D., Reiter, E., Taktasheva, E., Voloshina, E., Bogdanov, E., Winata, G.I., Schoelkopf, H., Kalo, J.C., Novikova, J., Forde, J.Z., Clive, J., Kasai, J., Kawamura, K., Hazan, L., Carpuat, M., Clinciu, M., Kim, N., Cheng, N., Serikov, O., Antverg, O., van der Wal, O., Zhang, R., Zhang, R., Gehrmann, S., Mirkin, S., Pais, S., Shavrina, T., Scialom, T., Yun, T., Limisiewicz, T., Rieser, V., Protasov, V., Mikhailov, V., Pruksachatkun, Y., Belinkov, Y., Bamberger, Z., Kasner, Z., Rueda, A., Pestana, A., Feizpour, A., Khan, A., Faranak, A., Santos, A., Hevia, A., Unldreaj, A., Aghagol, A., Abdollahi, A., Tammour, A., HajiHosseini, A., Behroozi, B., Ajibade, B., Saxena, B., Ferrandis, C.M., Contractor, D., Lansky, D., David, D., Kiela, D., Nguyen, D.A., Tan, E., Baylor, E., Ozoani, E., Mirza, F., Ononiwu, F., Rezanejad, H., Jones, H., Bhattacharya, I., Solaiman, I., Sedenko, I., Nejadgholi, I., Passmore, J., Seltzer, J., Sanz, J.B., Dutra, L., Samagaio, M., Elbadri, M., Mieskes, M., Gerchick, M., Akinlolu, M., McKenna, M., Qiu, M., Ghauri, M., Burynok, M., Abrar, N., Rajani, N., Elkott, N., Fahmy, N., Samuel, O., An, R., Kromann, R., Hao, R., Alizadeh, S., Shubber, S., Wang, S., Roy, S., Viguier, S., Le, T., Oyebade, T., Le, T., Yang, Y., Nguyen, Z., Kashyap, A.R., Palasciano, A., Callahan, A., Shukla, A., Miranda-Escalada, A., Singh, A., Beilharz, B., Wang, B., Brito, C., Zhou, C., Jain, C., Xu, C., Fourrier, C., Periñán, D.L., Molano, D., Yu, D., Manjavacas, E., Barth, F., Fuhrimann, F., Altay, G., Bayrak, G., Burns, G., Vrabec, H.U., Bello, I., Dash, I., Kang, J., Giorgi, J., Golde, J., Posada, J.D., Sivaraman, K.R., Bulchandani, L., Liu, L., Shinzato, L., de Bykhovetz, M.H., Takeuchi, M., Pàmies, M., Castillo, M.A., Nezhurina, M., Sänger, M., Samwald, M., Cullan, M., Weinberg, M., Wolf, M.D., Mihaljcic, M., Liu, M., Freidank, M., Kang, M., Seelam, N., Dahlberg, N., Broad, N.M., Muellner, N., Fung, P., Haller, P., Chandrasekhar, R., Eisenberg, R., Martin, R., Canalli, R., Su, R., Su, R., Cahyawijaya, S., Garda, S., Deshmukh, S.S., Mishra, S., Kiblawi, S., Ott, S., Sang-aroonsiri, S., Kumar, S., Schweter, S., Bharati, S., Laud, T., Gigant, T., Kainuma, T., Kusa, W., Labrak, Y., Bajaj, Y.S., Venkatraman, Y., Xu, Y., Xu, Y., Xu, Y., Tan, Z., Xie, Z., Ye, Z., Bras, M., Belkada, Y., Wolf, T.: Bloom: A 176b-parameter open-access multilingual language model (2023)
* [30] Zhang, S., Roller, S., Goyal, N., Artetxe, M., Chen, M., Chen, S., Dewan, C., Diab, M., Li, X., Lin, X.V., Mihaylov, T., Ott, M., Shleifer, S., Shuster, K., Simig, D., Koura, P.S., Sridhar, A., Wang, T., Zettlemoyer, L.: Opt: Open pre-trained transformer language models (2022)
## 7 Appendix
### 7.1 Prompt for evaluation
The prompt used for all the experiments in this paper is shown in the Figure
3.
| Select the correct alternative
---|---
… Few-shot examples go here, separated with ###
Nth example:
Question: | Times change, desires change, Beings change, trust changes: The whole world is composed of change, Always taking on new qualities. Continually we see novelties, Different in everything from hope: From evil, only the sorrows remain in memory, And from good (if any existed), the longing. Time covers the ground with a green cloak, Which was once covered in cold snow, And in me, it turns the sweet song into tears. And besides this daily change, Another change causes even more astonishment, That no longer changes as it used to. (Luís Vaz de Camões). (Luís de Camões, 20 sonnets. Campinas: Unicamp Publisher, p.91.) Indicate the statement that applies to the sonnet written by Camões.
Alternatives: | A. The poem takes up the Renaissance theme of the change of things, which the poet feels as a reason for hope and faith in life.
B. The idea of transformation refers to worldly things, but it does not affect
the poet’s state of mind due to his love belief.
C. Everything is always renewed, unlike the poet’s hopes, which harbor his
sorrows and longings.
D. Not only does the poet’s state of mind change, but also his experience of
change itself.
Answer: | D.
Figure 3: Example of prompt used in the experiments, the question was
translated into English for the convenience of readers. The text in red is the
expected output.
### 7.2 Benchmark per Subject
Table 4 provides a detailed report of each model achieved accuracy by subject.
Questions that were associated with more than one subject contributed to the
accuracy of both scores. For example, a question related to mathematics and
English will be taken into account when calculating the accuracy of both
mathematics and English subjects.
Model Biology Chemistry English Geography History Mathematics Philosophy
Physics Portuguese GPT-4 0.871 0.675 0.918 0.935 0.930 0.389 1.000 0.557 0.805
Sabiá 65B 0.771 0.350 0.837 0.774 0.883 0.278 1.000 0.257 0.755 GPT-3.5-Turbo
0.700 0.350 0.714 0.806 0.805 0.259 0.714 0.329 0.629 LLaMA 65B 0.657 0.350
0.816 0.677 0.719 0.306 0.429 0.286 0.572 OPT 66B 0.229 0.275 0.286 0.161
0.273 0.176 0.286 0.200 0.189 Sabiá 7B 0.514 0.350 0.592 0.565 0.672 0.241
0.571 0.271 0.509 Alpaca 7B 0.286 0.225 0.347 0.306 0.320 0.269 0.143 0.229
0.264 BloomZ 7B 0.243 0.075 0.551 0.371 0.336 0.185 0.143 0.171 0.308 LLaMA 7B
0.229 0.325 0.286 0.210 0.266 0.231 0.000 0.314 0.245 Bertin 6B 0.186 0.225
0.347 0.226 0.234 0.259 0.286 0.243 0.245 Bloom 7B 0.243 0.225 0.327 0.210
0.219 0.259 0.143 0.214 0.239 XGLM 7.5B 0.143 0.300 0.245 0.161 0.164 0.204
0.000 0.171 0.264 OPT 6.7B 0.186 0.250 0.143 0.145 0.234 0.185 0.000 0.257
0.214 GPTJ 6B 0.214 0.200 0.204 0.113 0.227 0.194 0.000 0.200 0.195
Table 4: Results for each model by subject in BLUEX.
|
problemgen Input: Output: problemdec Input: Question:
# On the enumeration of signatures of XOR-CNF’s
Nadia Creignou Aix-Marseille Université, CNRS, LIS, Marseille, France. Oscar
Defrain Aix-Marseille Université, CNRS, LIS, Marseille, France. Frédéric
Olive Aix-Marseille Université, CNRS, LIS, Marseille, France. Simon Vilmin
Aix-Marseille Université, CNRS, LIS, Marseille, France.
(February 28, 2024)
###### Abstract
Given a CNF formula $\varphi$ with clauses $C_{1},\dots,C_{m}$ over a set of
variables $V$, a truth assignment $\mathbf{a}:V\to\\{0,1\\}$ generates a
binary sequence
$\sigma_{\varphi}(\mathbf{a})=(C_{1}(\mathbf{a}),\ldots,C_{m}(\mathbf{a}))$,
called a signature of $\varphi$, where $C_{i}(\mathbf{a})=1$ if clause $C_{i}$
evaluates to 1 under assignment $\mathbf{a}$, and $C_{i}(\mathbf{a})=0$
otherwise. Signatures and their associated generation problems have given rise
to new yet promising research questions in algorithmic enumeration. In a
recent paper, Bérczi et al. interestingly proved that generating signatures of
a CNF is tractable despite the fact that verifying a solution is hard. They
also showed the hardness of finding maximal signatures of an arbitrary CNF due
to the intractability of satisfiability in general. Their contribution leaves
open the problem of efficiently generating maximal signatures for tractable
classes of CNFs, i.e., those for which satisfiability can be solved in
polynomial time. Stepping into that direction, we completely characterize the
complexity of generating all, minimal, and maximal signatures for XOR-CNFs.
Keywords: algorithmic enumeration, XOR-CNF, signatures, maximal bipartite
subgraphs enumeration, extension, proximity search.
## 1 Introduction
Propositional formulas are ubiquitous in computer science. The complexity of
the satisfiability problem has been extensively studied from the point of view
of several algorithmic tasks such as decidability, counting, or enumeration to
mention but a few. Given a CNF formula $\varphi$ with clauses
$C_{1},\dots,C_{m}$ over a set of variables $V=\\{v_{1},\dots,v_{n}\\}$, a
truth assignment $\mathbf{a}:V\to\\{0,1\\}$ leads to a binary sequence
$\sigma_{\varphi}(\mathbf{a})$, called a _signature_ of $\varphi$, defined by
$\sigma_{\varphi}(\mathbf{a})=(C_{1}(\mathbf{a}),\ldots,C_{m}(\mathbf{a}))$
where $C_{i}(\mathbf{a})=1$ if clause $C_{i}$ evaluates to 1 under assignment
$\mathbf{a}$, and $C_{i}(\mathbf{a})=0$ otherwise. Deciding the satisfiability
of $\varphi$ boils down to deciding whether the all-one sequence is a
signature of $\varphi$.
In this paper, we investigate the problems of listing all, minimal, and
maximal signatures of a CNF, where minimal and maximal are meant bitwise. The
task of finding all signatures of a CNF originates from well-design pattern
trees and has first been posed during the Dagstuhl seminar on enumeration in
data management [BKPS19]. However, enumerating the signatures of a given
CNF—may it be all, minimal, or maximal ones—will generally produce an
exponential number of solutions. Therefore, input sensitive polynomial time
complexity is not a suitable yardstick of efficiency when analyzing algorithms
performance. Instead, _output-sensitive complexity_ estimates the running time
of an enumeration algorithm using both input and output size. More precisely,
an enumeration algorithm runs in _total-polynomial time_ (or _output-
polynomial_ time) if its execution time is bounded by a polynomial in the
combined size of the input and the output. Still, the regularity of
enumeration algorithms is more relevant than their total running time. For
this reason, _polynomial delay_ and _incremental-polynomial time_ are
customarily regarded as better notions of tractability for enumeration
complexity. On the one hand, polynomial delay means that the delay between
consecutive outputs is polynomial in the input size. Incremental-polynomial
time, on the other hand, means that the time to produce the $i^{\text{th}}$
solution is bounded by a polynomial in the input size plus $i$ (see e.g.,
[JYP88, Str19]).
Generating all the signatures of a CNF turns out to be of great fundamental
interest, for it is an example of enumeration problem which can be solved in
total-polynomial time even though solutions cannot be recognized in polynomial
time.111These problems thus lie outside of the class EnumP which is commonly
studied in algorithmic enumeration; see e.g. [Str19]. In a recent contribution
[BBČ+21], Bérczi et al. address this problem. They give an incremental-
polynomial time algorithm for listing all the signatures of a CNF, and improve
their result to polynomial delay for tractable CNF’s. On the other hand, they
show that generating maximal signatures is hard, while minimal signatures can
be listed with polynomial delay, regardless of the CNF under consideration.
Their positive result relies on the equivalence with maximal independent sets
enumeration in graphs, known to be tractable [TIAS77]. As of their hardness
result, it relies on the fact that the existence of at least two solutions
would imply the non-satisfiability of the CNF at hand. Thus, the difficulty of
the enumeration heavily relies on the intractability of the satisfiability of
CNF’s in general. This observation naturally leads to the following open
question, where by tractable CNF’s we mean families of CNF’s for which the
satisfiability of a formula $\varphi$ in the family (as well as its sub-
formulas) can be decided in polynomial time.
###### Question 1.
Can the maximal signatures of tractable CNF’s be enumerated in total-
polynomial time?
Among tractable formulas, Horn and 2-CNF’s are two natural candidates for
approaching Question 1. These two cases however proved to be very challenging,
and the question of their tractability was explicitly stated as an open
problem in the 2022’s edition of the Workshop on Enumeration Problems and
Applications (WEPA’22).
Another natural step toward answering Question 1 is to consider XOR-CNF
formulas, being conjunctions of clauses using the “exclusive-or” connective
instead of the usual “or” connective. On top of being tractable for
satisfiability, XOR-CNF’s appear in many complexity classifications and enjoy
polynomial time algorithms for a broad variety of problems including counting
and enumerating their models (see e.g., [CH96, CH97, COS11]). Note that these
formulas do not fall in the framework of CNF’s though. Indeed, rewritting a
XOR-CNF as a CNF may result in a blowup on the number of clauses and
signatures, as discussed in Section 2. Hence, we remark that the positive
results from [BBČ+21] do not directly apply to XOR-CNF. Yet, we show that
signatures enumeration remains tractable using flashlight search as it is done
in [BBČ+21]. Namely, we prove the following.
###### Theorem 1.1.
The set of signatures of a XOR-CNF formula can be enumerated with polynomial
delay and polynomial space.
Concerning the enumeration of minimal signatures, we show that the problem is
in fact equivalent to the case of listing maximal signatures.
###### Proposition 1.2.
There is polynomial-delay and polynomial space algorithm listing all minimal
signatures of a XOR-CNF formula if and only if there is one listing all
maximal signatures of a XOR-CNF formula.
This leave us with the study of maximal signatures enumeration for XOR-CNF’s.
Quite surprisingly, we prove this problem to relate well with other problems
from graph theory and matroid theory. Specifically, we show the case of 2-XOR-
CNF to generalize maximal bipartite subgraphs enumeration, a problem which was
recently shown to admit a polynomial-delay algorithm [CGM+22]. As of the
general case of XOR-CNF’s, it can be seen as listing all maximal satisfiable
subsystems of a linear system of equations, which is related to the
enumeration of circuits passing through a given element in a matroid [BEGK03];
see also [KBE+05, KBB+09]. Relying on these results, we derive that maximal
signatures enumeration for XOR-CNF’s is tractable. Namely, we obtain the
following.
###### Theorem 1.3.
There is an incremental-polynomial time algorithm generating the maximal (or
minimal) signatures of a XOR-CNF.
Concerning the case of 2-XOR-CNF, we show it to be even more tractable, that
is, we show that it can be solved with polynomial delay using proximity
search.
###### Theorem 1.4.
There is a polynomial-delay algorithm generating the maximal (or minimal)
signatures of a 2-XOR-CNF.
To prove Theorem 1.4, we represent the input 2-XOR-CNF as an edge-bicolored
multigraph where each color codes the parity of the XOR clause. Then, the
enumeration amounts to list maximal bipartite subgraphs with additional
constraints on colored edges.
A caveat of Theorems 1.3 and 1.4 is that a queue of solutions has to be
maintained in the generation. More precisely, in [BEGK03] the queue is needed
to generate a new solution as the algorithm uses the saturation technique
where new solutions are derived from obtained ones. In [CGM+22], the queue has
the other purpose of preventing repetitions. In both cases, getting rid of a
potentially exponential space use is a natural enhancement one may seek. It
should however be noted that improving the algorithm from [BEGK03] to get
polynomial space is open for two decades now. As of improving Theorem 1.4 to
use polynomial space, it would imply such a result for maximal bipartite
subgraphs enumeration, another open question [CGM+22]. Toward such directions,
we prove that the folklore technique of flashlight search used to obtain
Theorem 1.1 may probably not be of great use as deciding whether a subsequence
of clauses can be extended into a signature is NP-complete.
### Organization of the paper.
In the next section we introduce the notions that will be used in this paper,
basic properties, and prove Theorem 1.1. The equivalence between maximal and
minimal signatures enumeration for XOR-CNF is detailed in Section 3. Theorems
1.3 and 1.4 are proved in Section 4. Future directions and the limitations of
flashlight search toward an improvement of Theorem 1.4 are finally discussed
in Section 5.
## 2 Preliminaries
All the objects considered in this paper are finite. If $V$ is a set,
$\mathbf{2}^{V}$ is its powerset. Let $m\in\mathbb{N}$ and let
$\sigma\in\\{0,1\\}^{m}$ be a binary sequence. We write $\sigma[i]$ to denote
the value of the $i^{\text{th}}$ element of $\sigma$. It will be furthermore
convenient to note $\mathbf{1}(\sigma)$ the set of indices
$I\subseteq\\{1,\dots,m\\}$ whose corresponding values in $\sigma$ is 1;
$\mathbf{0}(\sigma)$ is defined analogously. The _complementary_ of $\sigma$
is the binary sequence $\bar{\sigma}$ obtained from $\sigma$ by flipping all
the bits, i.e., $\bar{\sigma}[i]=1+\sigma[i]$ (mod $2$) for $1\leq i\leq m$.
If $\sigma,\tau\in\\{0,1\\}^{m}$, we write $\sigma\leq\tau$ if
$\sigma[i]\leq\tau[i]$ for all each $1\leq i\leq m$. In other words, $\leq$
corresponds to the bitwise order.
### Enumeration complexity.
An enumeration algorithm aims at enumerating a set of solutions of some
problem, one after the other, with no repetition. We already have defined
relevant time complexities in the introduction. However, space is also an
important consideration in the analysis of enumeration algorithms. We assume
that the solutions we produce are not stored in memory but rather flashed and
discarded. Thus, when measuring space we only consider the space needed by the
algorithm in order to conduct the enumeration, in terms of the input size. We
refer to [JYP88, Str19] for more details on the complexity of enumeration
algorithms.
### Graphs.
Given an undirected graph $G$, we write $V(G)$ its set of vertices and $E(G)$
its set of edges. For convenience, we will note $uv$ instead of $\\{u,v\\}$
for edges of $G$. A graph $G$ is bipartite if there exists a bipartition
$V_{1},V_{2}$ of $V(G)$ such that each edge of $E(G)$ has one endpoint in
$V_{1}$ and the other in $V_{2}$. Equivalently, a graph $G$ is bipartite if
and only if it does not contain any odd cycles. In this paper we will consider
_multigraphs_ as well, meaning graphs with potentially several edges sharing
two same endpoints, i.e., with $E(G)=\\{e_{1},\dots,e_{m}\\}$ a multiset. In
this case and to avoid any ambiguity we will refer to the edges of $G$ by
their label rather than their endpoints.
### XOR-formulas.
We assume basic familiarity with propositional logic. Let
$V=\\{x_{1},\dots,x_{n}\\}$ be a set of Boolean variables. A _literal_ is a
variable $x_{i}$ (positive literal) or its negation $\neg x_{i}$ (negative
literal). A _clause_ $C$ is a disjunction of literals. The size $|C|$ of $C$
is its number of literals. We say that $C$ is a $k$-clause if $|C|\leq k$. A
formula in conjunctive normal form (or a CNF for short) $\varphi$ is a
conjunction of clauses, i.e., $\varphi=C_{1}\land\dots\land C_{m}$ where
$C_{i}$, for $1\leq i\leq m$. It will be convenient to denote by $V(\varphi)$
the variables of $\varphi$. The (total) _size_ of $\varphi$ is defined as
$\sum_{1\leq i\leq m}|C_{i}|$. An assignment of the variables in $V$ is a
mapping $\mathbf{a}:V\to\\{0,1\\}$. Given an assignment $\mathbf{a}$, we write
$C_{i}(\mathbf{a})=1$ (resp. $C_{i}(\mathbf{a})=0$) if $C_{i}$ evaluates to
$1$ (resp. $0$) under assignment $\mathbf{a}$. The notations
$\varphi(\mathbf{a})=1$ and $\varphi(\mathbf{a})=0$ are defined analogously.
The formula $\varphi$ is _satisfiable_ if there exists an assignment
$\mathbf{a}$ of $V(\varphi)$ such that $\varphi(\mathbf{a})=1$; it is
unsatisfiable otherwise.
A _XOR-clause_ is a clause in which the usual connective “or” is replaced by
the “exclusive-or” connective. It is well-known that any XOR-clause can be
represented as a linear equation $x_{1}+\dots+x_{k}=\varepsilon$,
$\varepsilon\in\\{0,1\\}$ over the two-elements field $\mathbf{F}_{2}$. In
particular, the $2$-XOR-clause $x_{1}+x_{2}=0$ (resp. $x_{1}+x_{2}=1$) is
satisfied if and only if $x_{1}=x_{2}$ (resp. $x_{1}\neq x_{2}$). A XOR-clause
is _odd_ (resp. _even_) if $\varepsilon=1$ (resp. $0$). Given a XOR-clause
$C=(x_{1}+\dots+x_{k}=\varepsilon)$, with $\varepsilon\in\\{0,1\\}$, we put
$\bar{C}=(x_{1}+\dots+x_{k}=1-\varepsilon)$ and call $\smash{\bar{C}}$ the
_negation_ of $C$. In other words, the bar operator _changes the parity_ of
$C$. A XOR-CNF $\varphi$ is a conjunction of XOR-clauses. Equivalently,
$\varphi$ can be seen as a system of linear equations over $\mathbf{F}_{2}$.
For a XOR-CNF $\varphi=C_{1}\land\dots\land C_{m}$, we denote by
$\bar{\varphi}$ the XOR-CNF $\bar{C}_{1}\land\dots\land\bar{C}_{m}$ and call
it the _inverse_ of $\varphi$. We recall that the satisfiability of a XOR-CNF
can be tested in polynomial time; see e.g., [CH11, Theorem 2.18].
### Signatures.
The next notations and terminology are borrowed from [BBČ+21]. Let
$\varphi=C_{1}\land\dots\land C_{m}$ be a CNF. We refer to the introduction
for the definition of a signature, and shall say that an assignment
$\mathbf{a}$ of $\varphi$ _produces_ the signature
$\sigma_{\varphi}(\mathbf{a})=(C_{i}(\mathbf{a}),\ldots,C_{i}(\mathbf{a}))$.
We will further drop the subscript $\varphi$ from this notation when it is
clear from the context. Note that in particular, $\varphi$ is satisfiable if
and only if $(1,\dots,1)$ is a signature of $\varphi$. A signature $\sigma$ of
$\varphi$ is _minimal_ if it is minimal with respect to $\leq$ among all
signatures of $\varphi$. Similarly, we call _maximal_ a signature which is
maximal for $\leq$ among signatures of $\varphi$.
In this paper we are interested in the following problems.
Observe that in the above problems, and by the previous discussions, we may
indifferently consider $\varphi$ as a set of XOR clauses or as a set of binary
equations over $\mathbf{F}_{2}$. A subset of simultaneously satisfiable
clauses on one side corresponds to a feasible subset of equations on the other
side. In the rest of the paper, we will assume that $\varphi$ is in this
latter form. In particular, when $\varphi$ is seen as a system of linear
equations over $\mathbf{F}_{2}$, there is a one-to-one correspondence between
maximal signatures of $\varphi$ and maximal feasible subsets of equations.
Another remark is that any XOR-CNF may be rewritten as a CNF, i.e., the
$\oplus$ operator may be rewritten using $\land$ and $\lor$. For example, the
XOR-clause $x_{1}\neq x_{2}$ can be rewritten as $(x_{1}\vee
x_{2})\land(\bar{x_{1}}\vee\bar{x_{2}})$. We however note that we do not have
a bijection between the sets of signatures of the two formulas. Typically,
generalizing the above example to disjoint clauses will produce a XOR-CNF with
one minimal signature while the equivalent CNF has exponentially many such
signatures. Though this does not rule out the existence of a reduction between
(maximal) signatures of XOR-CNF’s to (maximal) signatures of CNF’s it should
be noted that the status of maximal signatures enumeration is still open for
2-CNF’s, a point that is discuss in Section 5. As of the case of all
signatures, it admits a simple and direct proof using flashlight search, as
explained in Section 3.
We finally argue that duplicate clauses in $\varphi$ may be ignored as far as
the enumeration of signatures is concerned. Indeed, having one of these
clauses to 1 implies that all other copies are to 1, and vice versa. Hence, we
will assume without loss of generality that all formulas are without
duplicated clauses in the rest of the paper, i.e., that they are defined as
pairwise distinct XOR-clauses.
## 3 Signatures of XOR-CNF’s and their properties
We prove that testing whether a binary sequence is a signature (resp. minimal
signature, maximal signature) of a XOR-CNF can be done in polynomial time in
the size of the CNF at hand, and use it to show that listing all signatures of
a XOR-CNF is tractable. For the rest of the section, let us fix some XOR-CNF
$\varphi=C_{1}\land\dots\land C_{m}$ on variable set
$V=\\{x_{1},\dots,x_{n}\\}$. A preliminary step is the following.
###### Proposition 3.1.
A sequence $\sigma\in\\{0,1\\}^{m}$ is a signature of $\varphi$ if and only if
$\overline{\sigma}$ is a signature of $\bar{\varphi}$. In particular, $\sigma$
is a minimal (resp. maximal) signature of $\varphi$ if and only if
$\overline{\sigma}$ is a maximal (resp. minimal) signature of $\bar{\varphi}$.
###### Proof.
Let $\mathbf{a}$ be an assignment producing the signature $\sigma$ of
$\varphi$. By definition we have $C_{i}(\mathbf{a})=1$ and
$C_{j}(\mathbf{a})=0$ for each $i,j\in\\{1,\dots,m\\}$ such that $\sigma[i]=1$
and $\sigma[j]=0$. This can be equivalently written as
$\smash{\bar{C}_{i}(\mathbf{a})=0}$ if $\overline{\sigma}[i]=0$ and
$\smash{\bar{C}_{i}(\mathbf{a})=1}$ when $\overline{\sigma}[i]=1$. This is in
turn equivalent to $\mathbf{a}$ producing the signature $\overline{\sigma}$ of
$\bar{\varphi}$. The second part of the statement follows from the fact that
for every sequences $\sigma,\tau\in\\{0,1\\}^{m}$ we have $\sigma\leq\tau$ if
and only if $\overline{\tau}\leq\overline{\sigma}$. ∎
From Proposition 3.1 we already derive Proposition 1.2.
Given two disjoint subsets $A,B\subseteq\\{1,\dots,m\\}$, we define
$\varphi(A,B):=\left(\bigwedge_{i\in A}C_{i}\right)\land\left(\bigwedge_{j\in
B}\bar{C}_{j}\right)$
to be the formula obtained from $\varphi$ by taking the clauses with index in
$A$ and the negation of clauses with index in $B$. We furthermore set this
formula to be defined on variable set $V(\varphi)$ even though some variables
in $V(\varphi)$ may not appear in any of the clauses of $\varphi(A,B)$—we want
this to ensure that assignments of $\varphi(A,B)$ are well-defined for
$\varphi$. Remark that the size of $\varphi(A,B)$ is at most that of
$\varphi$, and that it is defined on precisely $m$ clauses whenever $A,B$ is a
bipartition of $\\{1,\dots,m\\}$. In the next proposition, we show that
testing whether a binary sequence is a signature of a XOR-CNF can be reduced
to satisfiability testing.
###### Proposition 3.2.
A sequence $\sigma\in\\{0,1\\}^{m}$ is a signature of $\varphi$ if and only if
the formula $\varphi(\mathbf{1}(\sigma),\mathbf{0}(\sigma))$ is satisfiable.
###### Proof.
We start with the only if part. Assume that $\sigma$ is a signature of
$\varphi$. Then there is an assignment $\mathbf{a}$ that produces $\sigma$. By
definition $\mathbf{a}$ satisfies the clauses $C_{i}$ such that $\sigma[i]=1$,
and it does not satisfy the clauses $C_{j}$ such that $\sigma[j]=0$. This
latter case is equivalent to $\mathbf{a}$ satisfying the clauses
$\smash{\bar{C}_{j}}$. Hence $\mathbf{a}$ is an assignment which satisfies the
formula $\varphi(\mathbf{1}(\sigma),\mathbf{0}(\sigma))$ as desired.
We move to the if part. Assume that
$\varphi(\mathbf{1}(\sigma),\mathbf{0}(\sigma))$ has a satisfying assignment
$\mathbf{a}$ defined over $V$. Now since $\mathbf{a}$ satisfies the clauses
$C_{i}$ of $\varphi$ such that $\sigma[i]=1$, and not the clauses $C_{j}$ of
$\varphi$ such that $\sigma[j]=0$, we conclude that $\sigma$ is indeed a
signature of $\varphi$ as required. ∎
###### Proposition 3.3.
Deciding whether $\sigma$ is a minimal (resp. maximal) signature of $\varphi$
can be done in polynomial time in the size of $\varphi$.
###### Proof.
To check that $\sigma$ is a signature, we rely on Proposition 3.2: we build
the XOR-CNF $\varphi(\mathbf{1}(\sigma),\mathbf{0}(\sigma))$ and test its
satisfiability, all of which in polynomial time.
We prove the if part. Suppose that $\psi(j)$ is satisfiable for some
$j\in\mathbf{0}(\sigma)$, and consider an arbitrary assignment $\mathbf{a}$
over $V$ satisfying $\psi(j)$. Let $\tau$ be the signature of $\varphi$
produced by assignment $\mathbf{a}$. Since $\mathbf{a}$ satisfies $\psi(j)$,
every $1$ in $\sigma$ is a $1$ in $\tau$, so that $\sigma\leq\tau$. Moreover,
$\sigma[j]<\tau[j]$ holds, and $\sigma<\tau$ thus follows. Hence $\sigma$ is
not a maximal signature concluding the if implication. As for the only if
part, the existence of a signature $\tau>\sigma$ implies by definition that
$\psi(j)$ is satisfiable for some $j\in\mathbf{0}(\sigma)$ such that
$\tau[j]>\sigma[j]$.
Henceforth, deciding whether $\sigma$ is a maximal signature of $\varphi$ can
be done by $m$ calls to the satisfiability of a XOR-CNF whose size is no
greater than $\varphi$. We conclude that it can be done in polynomial time.
For minimality, we conclude using Proposition 3.1. ∎
To conclude this section, we argue that enumerating all the signatures of a
XOR-CNF is tractable using the above properties and flashlight search. We
recall that the main idea of this technique is to conduct binary search over
the space of solutions. In order for flashlight search to run with polynomial
delay and polynomial space, it is sufficient to solve the so-called “extension
problem” in polynomial time: given two disjoint subsets $A,B$ of the
groundset, decide whether there exists a solution including $A$ and avoiding
$B$. We refer to e.g. [BEG04, MS19] for a more detailed description of this
folklore technique.
In the case of XOR-CNF signatures, $A$ will consist of indices of clauses to
be satisfied in the signature—the 1’s in the sequence—while $B$ will consist
of indices of clauses not to be satisfied—the 0’s in the sequence. Then, it
follows from the discussion above that $\varphi$ admits a signature $\sigma$
satisfying $\sigma[i]=1$ and $\sigma[j]=0$ for all $i\in A$ and $j\in B$ if
and only if $\varphi(A,B)$ is satisfiable. Indeed, a satisfying assignment
$\mathbf{a}$ of $\varphi(A,B)$ satisfies the clauses of $\varphi$ with index
in $A$, and not those with index in $B$. Thus the signature of $\varphi$
produced by $\mathbf{a}$ meets the requirements. As for the converse
direction, it follows directly from the definition of a signature. We conclude
by Proposition 3.3 that the extension problem can be solved in polynomial time
for XOR-CNF signatures, hence proving Theorem 1.1 that we recall here.
See 1.1
We point that Theorem 1.1 is not a direct consequence of [BBČ+21, Theorem 1].
Indeed, in the later theorem the authors consider families of tractable CNF’s
which do not contain XOR-CNF’s in general. The arguments, however, follow the
same line.
## 4 Maximal signatures of XOR-CNF’s
We investigate the complexity of enumerating the maximal signatures of a given
XOR-CNF. More precisely, we prove the general case to be solvable in
incremental-polynomial time by a result of Boros et al. [BEGK03], and the
2-XOR-CNF case to be solvable with polynomial delay using proximity search.
In Section 2 we have mentioned that the clauses of a given XOR-CNF $\varphi$
may be seen as linear equations $x_{1}+\dots+x_{k}=\varepsilon$,
$\varepsilon\in\\{0,1\\}$ over the two-elements field $\mathbf{F}_{2}$. Hence
$\varphi$ can be seen as a system of linear equations whose maximal feasible
subsystems correspond to its maximal signatures. In [BEGK03], the authors
present an incremental polynomial-time algorithm for enumerating all circuits
of a matroid containing an element, and more generally, one for enumerating
all maximal subsets not spanning an element. Using Farkas’ lemma (more
precisely the Fredholm alternative) it can be derived an algorithm which,
given an infeasible system of linear equations, enumerates all its maximal
feasible (or minimal infeasible) subsystems within the same time; see also
[KBE+05, KBB+09] on the use of this characterization of infeasible systems.
This immediately yields Theorem 1.3 that we restate here.
See 1.3
We now focus on 2-XOR-CNF’s. Recall from Section 2 that we can consider these
formulas as a set of 2-clauses of the form $x=y$ or $x\neq y$ for positive
literals $x,y$ only, together with a set of 1-clauses $(z)$ with $z$ a
positive or negative literal. Furthermore we may assume that no clause is
repeated in the formula, as far as signatures enumeration is concerned. We
shall call _graphic_ a 2-XOR-CNF with clauses of the first type only—that is,
without 1-clauses—and will assume for now on that our XOR-CNF’s are graphic.
Later, we will show how to deal with 1-clauses with small modifications on the
instance.
We start by defining an underlying multigraph structure of graphic 2-XOR-CNF
that will be convenient for applying the framework of proximity search. Given
such a XOR-CNF $\varphi$ we define $G(\varphi)$ (or simply $G$ when it is
clear from the context) as the edge-bicolored multigraph defined on vertex set
$V(\varphi)$ and where there is a blue edge $xy$ whenever $x\neq y$ is a
clause of $\varphi$, and a red edge $xy$ whenever $x=y$ is a clause of
$\varphi$. A same pair of variables $x,y$ can produce both a red and a blue
edge: this corresponds to having clauses $x=y$ and $x\neq y$ in $\varphi$
which typically makes it unsatisfiable. However as no clause is repeated in
$\varphi$, no two edges of a same color share the same endpoints. In other
words, our multigraph has edge multiplicity at most two, and two edges between
a same pair of endpoints means that one is blue and the other is red; see
Figure 1 for an illustration of such multigraphs. In the following, we will
label edges of $G$ as $e_{1},\dots,e_{m}$ so that the edge $e_{j}$ corresponds
to the clause $C_{j}$ in $\varphi$. We note $R(G)$ the set of red edges of
$G$, and $B(G)$ the set of its blue edges. By definition, we have $R(G)\cap
B(G)=\emptyset$ and $R(G)\cup B(G)=E(G)$. A naive but crucial property we get
from that partition is that the set of red edges characterizes the set of blue
edge, and vice versa. By subgraph $H$ of an edge-bicolored multigraph $G$ we
mean the graph on same vertex set $V(H)=V(G)$ and $E(H)\subseteq E(G)$, that
is, we consider subgraphs as sets of edges. Finally, given two disjoint
subsets $X,Y\subseteq V(\varphi)$, by $\delta(X,Y)$ we mean the set of edges
having an endpoint in $X$ and the other in $Y$.
Figure 1: An edge-bicolored multigraph $G$ on nine vertices, with red edges
denoted by thin red lines and blue edges denoted by bold blue lines.
Note that an arbitrary (simple, uncolored) graph $H$ is bipartite whenever it
admits a bipartition $(X,Y)$ of its vertex set with $E(H)=\delta(X,Y)$. In the
next lemma, we relate the satisfiability of a graphic 2-XOR-CNF to a property
of its underlying edge-bicolored multigraph which generalizes bipartiteness.
We stress the fact that this property may only be stated on blue edges, which
is still reasonable as blue and red edges partition the edges of the graph.
###### Lemma 4.1.
Let $\varphi$ be a graphic 2-XOR-CNF. Then $\varphi$ is satisfiable if and
only if there is a bipartition $(X,Y)$ of the vertex set of $G(\varphi)$ such
that $B(G)=\delta(X,Y)$. Moreover, every such bipartition corresponds to a
model $\mathbf{a}$ of $\varphi$ with $X=\mathbf{a}^{-1}(1)$ and
$Y=\mathbf{a}^{-1}(0)$.
###### Proof.
We prove the first implication. If $\varphi$ is satisfiable then it has a
satisfying assignment $\mathbf{a}$ defining a bipartition $(X,Y)$ of
$V(\varphi)$ where $X=\mathbf{a}^{-1}(1)$ and $Y=\mathbf{a}^{-1}(0)$. Clearly
$B(G)\subseteq\delta(X,Y)$ as otherwise we have two variables $x,y$ with
$x\neq y$ yet $\mathbf{a}(x)=\mathbf{a}(y)$. On the converse, if there is a
red edge in $\delta(X,Y)$ we deduce the existence of two variables $x,y$ with
$x=y$ yet $\mathbf{a}(x)\neq\mathbf{a}(y)$. Hence, $\delta(X,Y)\subseteq
E(G)\setminus R(G)=B(G)$ and $B(G)=\delta(X,Y)$ follows. This concludes the
first implication.
We prove the other implication. Let $(X,Y)$ be a bipartition of $V(G)$ such
that $B(G)=\delta(X,Y)$. Consider the assignment $\mathbf{a}$ defined by
$\mathbf{a}(x)=1$ and $\mathbf{a}(y)=0$ for all $x\in X$ and $y\in Y$. By
construction to each clause of $\varphi$ of the form $x\neq y$ corresponds a
blue edge lying in $\delta(X,Y)$ and hence such that
$\mathbf{a}(x)\neq\mathbf{a}(y)$. As of the clauses of the form $x=y$ they
correspond to red edges not in $\delta(X,Y)$ and hence such that
$\mathbf{a}(x)=\mathbf{a}(y)$. All these clauses are satisfied by $\mathbf{a}$
and we conclude that $\varphi$ is indeed satisfiable. ∎
###### Lemma 4.2.
Let $\varphi$ be a graphic 2-XOR-CNF. Then the maximal signatures of $\varphi$
are exactly the maximal subgraphs $H$ of $G(\varphi)$ admitting a bipartition
$(X,Y)$ of their vertex set such that $\delta(X,Y)=B(H)$.
###### Proof.
Let $\sigma$ be a maximal signature of $\varphi$ and fix $H:=G(\psi)$ for
$\psi:=\varphi(\mathbf{1}(\sigma),\emptyset)$—recall that $\psi$ is defined on
the full variable set $V(\varphi)$ while some of its variables may not appear
in any clause. Since $\psi$ is satisfiable by definition of a signature, and
using Lemma 4.1, we derive that $B(H)=\delta(X,Y)$ for a bipartition $(X,Y)$
of $V(G)$ corresponding to a model $\mathbf{a}$ of $\psi$. Suppose that $H$ is
not maximal with this property. Then there exist an edge $e_{j}\in
E(G)\setminus E(H)$, $1\leq j\leq m$ and a bipartition
$(X^{\prime},Y^{\prime})$ with $H^{\prime}:=H+e_{j}$ satisfying
$B(H^{\prime})=\delta(X^{\prime},Y^{\prime})$. Then by Lemma 4.1 we derive
that $\psi:=\varphi(\mathbf{1}(\sigma)\cup\\{j\\},\emptyset)$ is satisfiable.
Let $\mathbf{a}$ be an assignment satisfying $\psi$. Since $j$ belongs to
$\mathbf{0}(\sigma)$, we obtain that $\mathbf{a}$ produces a signature $\tau$
with $\tau>\sigma$, a contradiction.
Let $H$ be a maximal subgraph of $G$ admitting a bipartition $(X,Y)$ of its
vertex set such that $B(G)=\delta(X,Y)$. By Lemma 4.1 the formula
$\psi:=\varphi(\\{j:e_{j}\in E(H)\\},\emptyset)$ corresponding to the edges of
$H$ is satisfiable. Let $\mathbf{a}$ be a satisfying assignment of $\psi$ and
consider the signature $\sigma$ it produces for $\varphi$. We argue that it is
a maximal signature of $\varphi$. Suppose that this is not the case and let
$\tau$ be a signature of $\varphi$ such that $\tau>\sigma$. Then
$\psi^{\prime}:=\varphi(\mathbf{1}(\sigma)\cup\\{j\\},\emptyset)$ is
satisfiable for some $1\leq j\leq m$ with $\sigma[j]<\tau[j]$. Hence by Lemma
4.1, the graph $H^{\prime}:=G(\psi^{\prime})$ satisfies
$B(H^{\prime})=\delta(X^{\prime},Y^{\prime})$. Moreover $E(H)\subset
E(H^{\prime})$ which contradicts the fact that $H$ is chosen maximal. We
conclude that $\sigma$ is a maximal signature. ∎
In the following, given an edge-bicolored multigraph, we call _red-blue
bipartite_ any of its subgraph with the property described in Lemma 4.2. In
Figure 2, we give an example of a maximal red-blue bipartite subgraph $H$ of
the multigraph $G$ of Figure 1. Note that red-blue bipartite subgraphs do not
contain both a red edge and a blue edge between a same pair of vertices, i.e.,
they define edge-bicolored graphs.
Figure 2: A maximal (connected) red-blue bipartite subgraph $H$ of the
multigraph $G$ defined in Figure 1. The bipartition $(H_{0},H_{1})$ of $H$ is
highlighted. The edges of $G$ not in $H$ are dotted.
###### Theorem 4.3.
There is a polynomial-delay and polynomial-space algorithm enumerating the
maximal signatures of a graphic 2-XOR-CNF $\varphi$ if and only if there is
one listing the maximal red-blue bipartite subgraphs of its underlying
multigraph $G(\varphi)$.
###### Proof.
This is a consequence of Lemma 4.2 observing that constructing either the
formula or the edge-bicolored multigraph can be done in polynomial time in the
size of the structure at hand. As there is a bijection between the two
solutions sets this reduction preserves polynomial delay and polynomial space.
∎
A consequence of Lemma 4.2 is that the maximal signatures of a 2-XOR-CNF only
consisting of disequalities $x\neq y$ is equivalent to the enumeration of
maximal (edge) bipartite subgraphs of a given graph. Indeed in that case,
$R(G)$ is empty—hence it is a graph—and we only require $E(G)=\delta(X,Y)$. We
derive the following using a recent result by Conte and Uno [CGM+22] on
maximal bipartite subgraphs enumeration using proximity search.
###### Corollary 4.4.
There is a polynomial-delay algorithm listing the maximal signatures of a
2-XOR-CNF consisting of disequalities $x\neq y$ only.
We end the section showing that the result from [CGM+22] can be extended into
our context using proximity search, and arguing how to capture clauses of size
one.
In the following, we say that an edge-bicolored multigraph $G$ is _connected_
if its underlying graph (where colors are ignored and multi-edges are
considered as a single edge) is connected. Note that we can assume without
loss of generality that $G(\varphi)$ is connected, as far as signatures
enumeration is concerned. Indeed, if it was not the case we could enumerate
the signature of each subformula of $\varphi$ corresponding to connected
components of $G(\varphi)$ and combine the results by doing their Cartesian
product. Hence, we may further assume that $G(\varphi)$ is connected, and
shall call _connected_ a formula such that its underlying multigraph is. We
first show that if $G$ is connected then its maximal red-blue bipartite
subgraphs are as well, yielding a canonical bipartition.
###### Lemma 4.5.
Let $G$ be a connected edge-bicolored multigraph. Then any of its maximal red-
blue bipartite subgraph is connected.
###### Proof.
Let $H$ be an arbitrary maximal red-blue bipartite subgraph of $G$ and suppose
toward a contradiction that it is not connected. Since $G$ is connected we can
find a shortest path $P$ connecting two distinct and connected components
$C_{1},C_{2}$ of $H$. For convenience, we consider two such components
minimizing the length of $P$ so that adding $P$ only connects $C_{1},C_{2}$.
We obtain a red-blue bipartite supergraph of $H$ by the following procedure:
we start from the endpoint of $P$ in $C_{1}$ and for each of the edges
$e_{1},\dots,e_{k}$ of $P$ in order, we either keep its other endpoint in the
same part if $e_{i}$ is red, or put its endpoint in the other part if $e_{i}$
is blue. When reaching $e_{k}$, we may need to swap the original bipartition
of $C_{2}$. No contradiction can be found during this process and the obtained
graph is indeed a supergraph of $H$. This contradicts the fact that $H$ is
maximal and we conclude that $H$ is connected as desired. ∎
###### Lemma 4.6.
If $G$ is a connected edge-bicolored bipartite graph then it admits a unique
bipartition $(X,Y)$ of its vertex set with $B(G)=\delta(X,Y)$ and $X$
containing the vertex of smallest label in $G$.
###### Proof.
We prove the statement using induction on the number of vertices of $G$. If
$G$ has 0, 1, or 2 vertices, the result is clear. Now, assume that the
statement holds for every graph with at most $k$ vertices. Let $G$ be a
connected edge-bicolored bipartite graph with vertices
$V(G)=\\{v_{1},\dots,v_{k+1}\\}$. We consider the edge-colored bipartite
induced subgraph $H:=G[\\{v_{1},\dots,v_{k}\\}]$ with $v_{k+1}$ chosen so that
$H$ is connected. Note that this can be done by considering a BFS of $G$
launched at its vertex $v_{1}$ of smallest label, and choosing $v_{k+1}$ as
the vertex obtained last in the traversal. By inductive hypothesis, there
exists a unique bipartition $(X^{\prime},Y^{\prime})$ of $V(H)$ such that
$\delta(X^{\prime},Y^{\prime})=B(H)$ and $v_{1}\in X^{\prime}$. Since $G$ is
bipartite, there exists at least one bipartition $(X,Y)$ of $V(G)$ such that
$\delta(X,Y)=B(G)$. Moreover, every such bipartition must comply with
$(X^{\prime},Y^{\prime})$, that is, without loss of generality,
$X^{\prime}\subseteq X$ and $Y^{\prime}\subseteq Y$. Henceforth, $G$ has at
most two possible valid bipartitions:
$(X^{\prime},Y^{\prime}\cup\\{v_{k+1}\\})$ and
$(X^{\prime}\cup\\{v_{k+1}\\},Y^{\prime})$. Now, $G$ is connected, so there
exists a vertex $v_{i}$ in $V(G)\setminus\\{v_{k+1}\\}$, say in $X^{\prime}$,
such that $v_{i}v_{k+1}$ is an edge of $G$. We have two disjoint cases:
* •
$v_{i}v_{k+1}$ is red in which case $(X^{\prime}\cup\\{v_{k+1}\\},Y^{\prime})$
is the unique correct bipartition; and
* •
$v_{i}v_{k+1}$ is blue in which case
$(X^{\prime},Y^{\prime}\cup\\{v_{k+1}\\})$ is the unique correct bipartition.
We deduce using induction that there exists exactly one bipartition $(X,Y)$ of
$G$ satisfying $\delta(X,Y)=B(G)$ with $X$ containing the vertex of smallest
label in $G$. ∎
Proximity search has been introduced in [CU19] as a special case of solution
graph traversal in which the proof that the solution graph is strongly
connected relies on a notion of proximity between solutions that is more
involved than their intersection. More precisely, the proximity between two
solutions $S,S^{*}$ is defined as the length of the longest prefix of $S^{*}$
(according to a carefully chosen ordering of its elements) that is subset of
$S$, and need not to be symmetric. We refer to [CGM+22] for more details on
the technique. In [CGM+22], the authors extract the necessary conditions that
make an enumeration problem “proximity searchable” and thus to be solvable
with polynomial delay. We restate these conditions here in the context of XOR-
CNF for self-containment and better readability.
In the reformulation below, let $G$ be an edge-bicolored multigraph and
$\mathcal{H}(G)$ denote the set of maximal (connected) red-blue bipartite
subgraphs of $G$ we shall call _maximal solutions_. We further call
_solutions_ the connected red-blue bipartite subgraphs of $G$.
###### Theorem 4.7 (Reformulation of [CGM+22, Definition 4.2 and Theorem
4.3]).
For arbitrary $H,H^{*}\in\mathcal{H}(G)$, let $\mu(H^{*})$ denote an ordering
of the edges in $H^{*}$, and $H\,\tilde{\cap}\,H^{*}$ denote the elements in
the longest prefix of $\mu(H^{*})$ that is a subset of $H$. Then, there is a
polynomial-delay algorithm enumerating $\mathcal{H}(G)$ whenever:
1. 1.
For $H\in\mathcal{H}(G)$ any prefix of $\mu(H)$ is a solution;
2. 2.
There is a polynomial-time computable function $\operatorname{{COMP}}$, which
given a solution $H^{\prime}$ produces $H=\operatorname{{COMP}}(H^{\prime})$ a
maximal solution such that $H\supseteq H^{\prime}$;
3. 3.
Given $H\in\mathcal{H}(G)$ and $e\in E(G)\setminus E(H)$ there is a family
$\mathcal{K}\subseteq 2^{E(H)}$ of sets called removables such that:
* •
the family $\mathcal{K}$ can be computed in polynomial time;
* •
for any $K\in\mathcal{K}$ the set $H\setminus K\cup\\{e\\}$ is a solution; and
* •
for any $H^{*}\in\mathcal{H}(G)$ and minimal element $e$ in $\mu(H^{*})$ that
is not in $H$, there exists $K\in\mathcal{K}$ with
$(H\,\tilde{\cap}\,H^{*})\cap K=\emptyset$.
The rest of the section is dedicated to proving that the conditions of Theorem
4.7 are fulfilled for well-chosen $\mu$ and set $\mathcal{K}$. The fact that
$\operatorname{{COMP}}$ can be computed in polynomial time follows from the
fact that the red-blue bipartiteness is a hereditary property which can be
tested in polynomial time, hence that we can obtain a maximal solution by a
greedy procedure.
We start with the definition of $\mu$. Given a maximal red-blue bipartite
graph $H$ of $G$ we denote by $(H_{0},H_{1})$ the unique bipartition given by
Lemma 4.6 where $H_{0}$ is the side containing the vertex $v_{0}$ of smallest
index in $G$. Consider a BFS ordering of the vertices of $G$ starting at
$v_{0}$: $v_{0}$ is first in the order and at each step, given the last vertex
in the ordering, we add its neighborhood in ascending order of their label.
Clearly, each prefix of such an ordering defines a connected induced subgraph.
We define $\mu(H)$ as the increasing ordering of the edges of $H$ with respect
to their endpoint occurring later in the BFS ordering launched at $v_{0}$, and
break the tie by increasing order of their earlier endpoint in the BFS; see
Figure 3 for an example of such an edge-ordering. This choice of an edge-
ordering of a maximal solution yields the following property which will be
crucial in the rest of the proof.
Figure 3: A BFS on the maximal red-blue bipartite subgraph $H$ of $G$ (Figure
2) will give the following order on the vertices:
$v_{0},v_{1},v_{3},v_{4},v_{2},v_{5},v_{8},v_{7},v_{6}$. The order of the
edges of $H$ it produces is
$\mu(H)=v_{0}v_{1},v_{0}v_{3},v_{0}v_{4},v_{3}v_{4},v_{1}v_{2},v_{1}v_{5},v_{2}v_{5},v_{1}v_{8},v_{3}v_{7},v_{8}v_{6},v_{7}v_{6}$.
###### Observation 4.8.
If $H\in\mathcal{H}(G)$ then any prefix of $\mu(H)$ is a connected red-blue
bipartite subgraph of $G$ and the first edge in this ordering is incident to
$v_{0}$.
Let us now consider arbitrary $H,H^{*}\in\mathcal{H}(G)$ and put
$H^{\prime}:=H\,\tilde{\cap}\,H^{*}$ as defined in Theorem 4.7. Note that if
$H^{\prime}$ is empty then the trivial removable $K=E(G)$ satisfies the above
conditions. Hence we may assume for convenience (in the rest of the analysis)
that $|H\,\tilde{\cap}\,H^{*}|\neq 0$, while it will appear clear later that
this assumption is in fact not needed. By Observation 4.8 since the first edge
in $\mu(H^{*})$ belongs to both $H^{\prime}$, $H$ and $H^{*}$ we derive that
$v_{0}$ belongs to all three graphs. Again by Observation 4.8, $H^{\prime}$ is
a connected red-blue bipartite subgraph of $G$. Thus by Lemma 4.6 it admits a
unique bipartition $(H^{\prime}_{0},H^{\prime}_{1})$ with $v_{0}\in
H^{\prime}_{0}$. Now since $H$ and $H^{*}$ contain $H^{\prime}$ they must
agree with that bipartition, meaning that $H^{\prime}_{0}\subseteq H_{0}\cap
H^{*}_{0}$ and $H^{\prime}_{1}\subseteq H_{1}\cap H^{*}_{1}$, which is a
crucial point. In the following, for each edge $e=ab$ in $E(G)\setminus E(H)$,
we define $K_{1}:=\\{av:av\in E(G)\\}$ and $K_{2}:=\\{bv:bv\in E(G)\\}$ to be
the two removables of $(H,e)$ and argue that $\mathcal{K}:=\\{K_{1},K_{2}\\}$
meets the requirements of the theorem.
Consider the first edge $e=ab$ in $\mu(H^{*})$ that is not an edge of $H$. We
distinguish two symmetric cases on whether $e$ is red or blue, and only detail
the blue case here. This situation is depicted in Figure 4.
We assume without loss of generality that the endpoint $a$ belongs to
$H^{*}_{0}$ and that the other endpoint $b$ lies in $H^{*}_{1}$. Recall that
$e\not\in H$. Hence since $e$ is blue, either the two endpoints $a,b$ belong
to $H_{0}$, or they both belong to $H_{1}$. Two symmetric cases arise. Let us
assume $\\{a,b\\}\subseteq H_{0}$. Note that in $H^{*}$ we have no blue edge
between $b$ and $H^{\prime}_{1}$, and no red edge between $b$ and
$H^{\prime}_{0}$. In $H$ this is the opposite: we have no blue edge between
$b$ and $H^{\prime}_{0}$, and no red edge between $b$ and $H^{\prime}_{1}$.
Now as $E(H^{\prime})\subseteq E(H^{*})\cap E(H)$ we conclude that $b$ is not
incident to blue or red edges in $H^{\prime}$, that is, $E(H^{\prime})\cap
K_{2}=\emptyset$. This concludes the case and the other situation of
$\\{a,b\\}\subseteq H_{1}$ yields the symmetric situation of
$E(H^{\prime})\cap K_{1}=\emptyset$ by the same arguments. The other situation
of $e$ being red follows the exact same line of arguments, swapping red for
blue.
Figure 4: The situation occurring when considering two solutions $H,H^{*}$ in
Theorem 4.7 with $H^{\prime}=H\,\tilde{\cap}\,H^{*}$: $H^{\prime}_{0}\subseteq
H_{0}\cap H^{*}_{0}$, $H^{\prime}_{1}\subseteq H_{1}\cap H^{*}_{1}$ and the
blue edge $ab$ of $H^{*}$ satisfies $E(H^{\prime})\cap K_{2}=\emptyset$ where
$K_{2}:=\\{bv:bv\in E(G)\\}$.
We finish the proof observing that the graphic conditions of 2-XOR-CNF’s can
be relaxed. As a preliminary step, note that 1-clauses containing variables
that do not appear in 2-clauses can be removed from the instance as they will
be set to true in any maximal signature. Thus we may assume that $\varphi$
does not contain such isolated 1-clauses. We code every other 1-clauses $(x)$
and $(\overline{y})$—which can also be seen as clauses $x=1$ and $y=0$—by a
blue edge $xu$ and a red edge $yu$ in the edge-bicolored multigraph $G$ we
have defined above, for some special vertex $u$ that will be connecting all
such 1-clauses. Note that $G$ stays connected by the above assumption that
$\varphi$ has no isolated 1-clause. Then we set $u$ to be the vertex of
smallest label in $G$, followed by $v_{0}$. Then by the definition of $\mu(H)$
the vertex $u$ will be placed in $H_{0}$ which can be seen as forcing the
gadget to false, satisfying the clause $y=0$ if the red edge $uy$ is selected,
and satisfying the clause $x=1$ if the blue edge $ux$ is selected. We conclude
with Theorem 1.4 that we restate here as a consequence of Theorem 4.7 and the
above discussion.
See 1.4
## 5 Discussions
In this work we have showed that enumerating all (resp. minimal, maximal)
signatures of a XOR-CNF formula is tractable, namely, Theorems 1.1, 1.3 and
1.4. Observe that the algorithm of Theorem 1.3 runs in incremental-polynomial
time while the other two run with polynomial delay. Hence it is natural to ask
whether Theorem 1.3 can be improved to polynomial delay.
###### Question 2.
Can the maximal signatures of a XOR-CNF be generated with polynomial delay?
The algorithm by Boros et al. [BEGK03] underlying Theorem 1.3 relies on the
enumeration of the circuits of a matroid, for which the same question seems
open since two decades. Hence, answering Question 2 would either require new
techniques, or would need to answer the open question from [BEGK03].
Furthermore, except for Theorems 1.1 and 1.3 that run with polynomial space,
the algorithm of Theorem 1.4 requires a space that is potentially exponential
as solutions must be stored into a queue. A natural question is the following.
###### Question 3.
Can the maximal signatures of a 2-XOR-CNF be generated with polynomial delay
and polynomial space?
We however stress the fact that a positive answer to Question 3 would improve
the algorithm by Conte et al. for maximal bipartite subgraphs enumeration for
which the questions of achieving polynomial delay and space is open [CGM+22].
Toward this direction, we now argue that flashlight search may not be adapted
in order to give a positive answer to Question 3, as the extension problem is
hard in that case. Moreover, this result even holds when restricted to 2-XOR-
CNF’s consisting of disequalities, that is, in the context of maximal
bipartite subgraph enumeration, which may be of independent interest.
###### Theorem 5.1.
The problem of deciding, given a graph $G$ and two disjoint edge sets
$A,B\subseteq E(G)$, whether there exists a maximal bipartite (edge) subgraph
$H$ of $G$ such that $A\subseteq E(H)$ and $B\cap E(H)=\emptyset$ is ${\sf
NP}$-complete.
###### Proof.
We first argue that the problem belongs to NP. Note that bipartiteness is
hereditary, meaning that every subgraph of a bipartite graph is bipartite.
Moreover since bipartiteness can be decided in polynomial time, we can check
maximality of $H$ by trying to add every edge in $E(G)\setminus E(H)$ and
checking whether the resulting graph is bipartite. Now since the two extra
conditions $A\subseteq E(H)$ and $B\cap E(H)=\emptyset$ can be checked in
polynomial time, we conclude that $H$ itself is a polynomial certificate.
We now prove NP-hardness by reduction from 3-SAT. Let
$\varphi=C_{1}\land\dots\land C_{m}$ be an instance of 3-SAT on variable set
$V(\varphi)=\\{v_{1},\dots,v_{n}\\}$. We construct $G$ by creating two
vertices $x_{i},y_{i}$ representing the literals of variable $v_{i}$ for every
$1\leq i\leq n$, a special vertex $u$, and a vertex $c_{j}$ representing the
clause $C_{j}$ for every $1\leq j\leq m$. We add the edges $x_{i}y_{i}$ for
every $1\leq i\leq n$, the edge $x_{i}c_{j}$ if $v_{i}$ appears positively in
$C_{j}$, and the edge $y_{i}c_{j}$ if it appears negated. Finally, we connect
$u$ to every other vertex in $G$ making it a universal vertex. This completes
the construction of the graph. Note that it is not bipartite as it contains
triangles $x_{i}y_{i}u$ for $1\leq i\leq n$. The two sets $A,B\subseteq E(G)$
are defined as $A:=\\{x_{i}y_{i}:1\leq i\leq n\\}$ and $B:=\\{uc_{j}:1\leq
j\leq m\\}$. We illustrate the reduction in Figure 5.
Figure 5: The reduction of Theorem 5.1 with the CNF $\varphi=(v_{1}\lor
v_{2}\lor v_{3})\land(\bar{v_{1}}\lor v_{2}\lor v_{3})\land(\bar{v_{1}}\lor
v_{2}\lor\bar{v_{3}})$. We have $A=\\{x_{1}y_{1},x_{2}y_{2},x_{3}y_{3}\\}$ and
$B=\\{c_{1}u,c_{2}u,c_{3}u\\}$. A maximal bipartite subgraph including $A$ and
avoiding $B$ is highlighted in grey. It corresponds to the model of $\varphi$
which assigns $1$ to $v_{1}$ and $v_{2}$, and $0$ to $v_{3}$.
Let us first show that there exists a maximal bipartite subgraph of $G$
containing $A$ and avoiding $B$ whenever $\varphi$ is satisfiable. Consider a
model $\mathbf{a}$ of $\varphi$ and consider the subgraph $H$ of $G$ defined
as
$\displaystyle E(H):=A$
$\displaystyle\cup\\{ux_{i}:\mathbf{a}(v_{i})=1\\}\cup\\{uy_{i}:\mathbf{a}(v_{i})=0\\}$
$\displaystyle\cup\\{c_{j}x_{i}:\mathbf{a}(v_{i})=1\ \text{and}\ v_{i}\in
C_{j}\\}$ $\displaystyle\cup\\{c_{j}y_{i}\hskip 0.85355pt:\mathbf{a}(v_{i})=0\
\text{and}\ \overline{v}_{i}\in C_{j}\\}.$
Clearly $H$ extends subset $A$. We first argue that it is bipartite. Consider
the bipartition $(V_{1},V_{2})$ where
$V_{1}=\\{x_{i}:\mathbf{a}(v_{i})=1\\}\cup\\{y_{i}:\mathbf{a}(v_{i})=0\\}$ and
$V_{2}=V\setminus V_{1}$. By construction $V_{1}$ is edgeless. The set $V_{2}$
is edgeless too as by the definition of $H$ vertices adjacent to $u$ or to
$c_{j}$, $1\leq j\leq m$ in $H$ are precisely those in $V_{1}$. Thus $H$ is
bipartite. We greedily extend $H$ into a maximal bipartite subgraph and note
$H^{\prime}$ the result. Suppose toward a contradiction that $H^{\prime}$ does
not satisfy the required properties, i.e., that it contains an edge $uc_{j}$
for some $1\leq j\leq m$. Then by definition of $V_{1}$ and the fact that
$\mathbf{a}$ is a model of $\varphi$, we get that $uc_{j}w$ would induce a
triangle for $w$ the neighbor of $c_{j}$ in $V_{1}$. This contradicts the fact
that $H^{\prime}$ is bipartite. Consequently $H^{\prime}$ extends $A$ and
avoids $B$ as required.
Suppose now that there exists a maximal bipartite subgraph $H$ of $G$ with
$A\subseteq E(H)$ and $B\cap E(H)=\emptyset$. Let $(V_{1},V_{2})$ be a
bipartition of $H$. Note that since $x_{1}y_{1},\dots,x_{n}y_{n}$ belong to
$A$ (hence to $H$), the sets $V_{1}$ and $V_{2}$ contain exactly one and
distinct elements from $\\{x_{i},y_{i}\\}$ for each $1\leq i\leq n$. Let us
assume without loss of generality that $u\in V_{2}$. Then since every edge
$uc_{j}$, $1\leq j\leq m$ belongs to $B$, it must be that every such $c_{j}$
lies in $V_{2}$ as well. Thus, except for the the choice of which of
$\\{x_{i},y_{i}\\}$ belongs to $V_{1}$, $1\leq i\leq n$, the bipartition
$(V_{1},V_{2})$ is now completely characterized. Consider a triangle
$ux_{i}y_{i}$, $1\leq i\leq n$. Then one of $x_{i},y_{i}$ belongs to $V_{1}$,
call it $\ell_{1}$, and the other belongs to $V_{2}$, call it $\ell_{2}$.
Clearly $u\ell_{2}$ is not an edge of $H$ as both $u$ and $\ell_{2}$ lie in
the same side of the bipartition. We argue that $u\ell_{1}$ is an edge of $H$.
Suppose that it is not the case. Then by maximality of $H$, adding $u\ell_{1}$
creates an odd cycle. Consequently there exists an even path $P$ going from
$u$ to $\ell_{1}$ in $H$, where by even we mean that $P$ contains an even
number of edges. Since $\ell_{1}\ell_{2}$ belongs to $H$ we deduce that the
path $P\ell_{2}$ obtained by adding $\ell_{2}$ to $P$ is odd, which
contradicts the fact that $u$ and $\ell_{2}$ both lie in $V_{2}$. A
consequence of these observations is that the set of edges $u\ell$ in $H$ for
$\ell\in V_{1}$ defines an assignment $\mathbf{a}$ of $\varphi$, where
$\mathbf{a}(v_{i})=1$ if $\ell=x_{i}$, and $\mathbf{a}(v_{i})=0$ if
$\ell=y_{i}$. Hence the edges of $H$ having some $c_{j}$ for endpoint are
precisely those connecting $c_{j}$ to $V_{1}$. We claim that every such
$c_{j}$ has at least one such incident edge, hence that $\mathbf{a}$ is a
satisfying truth assignement of $\varphi$. Indeed, if it was not the case then
$c_{j}$ would be disconnected from the rest of the graph. But in that case, we
could have added $uc_{j}$ to the graph still maintaining the bipartiteness
(changing $c_{j}$ to the other side of the bipartition) contradicting the
maximality of $H$. This concludes the proof. ∎
Concerning prospective research directions, let us restate here the implicit
question from [BBČ+21] that motivated this work, which additionally was posed
at the WEPA’19 open problem session, and that is still open to date. Recall
that a family of CNF’s is tractable if there exists a polynomial-time
algorithm to decide the satisfiability of any formula (and its subformulas) in
the family.
See 1
Natural examples of tractable CNF’s include 2-CNF’s and Horn CNF’s for which,
to the best of our knowledge, no progress has been made. Toward this
direction, the case of Horn 2-CNF also seems open.
Another question left open by the work of Bérczi et al. deals with the
_dimension_ of the formula, defined as the maximum size of a clause it
contains. Namely, the algorithm of [BBČ+21, Theorem 4] only performs in
incremental-polynomial time for fixed dimension, and it is open whether the
same result can be obtained for formulas of arbitrary dimension.
Finally, another promising direction concerns the parameterized study of
signatures enumeration. More particularly, it can be seen that the algorithm
of [BBČ+21, Theorem 4] is XP-incremental parameterized by the dimension $k$,
that is, it generates the $i^{\text{th}}$ solution in time
$(\|\varphi\|+i)^{f(k)}$ for some computable function $f$. It would be
interesting to know whether this algorithm can be improved to run with FPT-
delay, that is, to run with $f(k)\cdot\|\varphi\|^{O(1)}$ delay; see e.g.
[CMM+17, GKKL22] for more details on the parameterized aspects of enumeration
problems. A preliminary step in that direction would be to get such running
times for a double parameterization. Among parameters, the maximum number
$\omega$ of occurrence of a variable in the formula is natural. This parameter
is relevant as in [BBČ+21, Theorem 3] the authors give a simpler algorithm for
signatures enumeration when both $k$ and $\omega$ are bounded. Improving this
XP-incremental algorithm into one that is FPT-delay parameterized by $k$ plus
$\omega$ seems open.
### Acknowledgements.
The authors would like to thank Kazuhiro Kurita, Kazuhisa Makino, Kunihiro
Wasa, and Yasuaki Kobayashi for preliminary discussions on the topic of this
paper.
## References
* [BBČ+21] Kristóf Bérczi, Endre Boros, Ondřej Čepek, Khaled Elbassioni, Petr Kučera, and Kazuhisa Makino. Generating clause sequences of a CNF formula. Theoretical computer science, 856:68–74, 2021.
* [BEG04] Endre Boros, Khaled Elbassioni, and Vladimir Gurvich. Algorithms for generating minimal blockers of perfect matchings in bipartite graphs and related problems. In European Symposium on Algorithms, pages 122–133. Springer, 2004.
* [BEGK03] Endre Boros, Khaled Elbassioni, Vladimir Gurvich, and Leonid Khachiyan. Algorithms for enumerating circuits in matroids. In International Symposium on Algorithms and Computation, pages 485–494. Springer, 2003.
* [BKPS19] Endre Boros, Benny Kimelfeld, Reinhard Pichler, and Nicole Schweikardt. Enumeration in Data Management (Dagstuhl Seminar 19211). Dagstuhl Reports, 9(5):89–109, 2019.
* [CGM+22] Alessio Conte, Roberto Grossi, Andrea Marino, Takeaki Uno, and Luca Versari. Proximity search for maximal subgraph enumeration. SIAM Journal on Computing, 51(5):1580–1625, 2022.
* [CH96] Nadia Creignou and Miki Hermann. Complexity of generalized satisfiability counting problems. Inf. Comput., 125(1):1–12, 1996.
* [CH97] Nadia Creignou and Jean-Jacques Hébrard. On generating all solutions of generalized satisfiability problems. RAIRO Theor. Informatics Appl., 31(6):499–511, 1997.
* [CH11] Yves Crama and Peter L. Hammer. Boolean functions: Theory, algorithms, and applications. Cambridge University Press, 2011.
* [CMM+17] Nadia Creignou, Arne Meier, Julian-Steffen Müller, Johannes Schmidt, and Heribert Vollmer. Paradigms for parameterized enumeration. Theory of Computing Systems, 60:737–758, 2017.
* [COS11] Nadia Creignou, Frédéric Olive, and Johannes Schmidt. Enumerating all solutions of a boolean CSP by non-decreasing weight. In Karem A. Sakallah and Laurent Simon, editors, Theory and Applications of Satisfiability Testing - SAT 2011 - 14th International Conference, SAT 2011, Ann Arbor, MI, USA, June 19-22, 2011. Proceedings, volume 6695 of Lecture Notes in Computer Science, pages 120–133. Springer, 2011.
* [CU19] Alessio Conte and Takeaki Uno. New polynomial delay bounds for maximal subgraph enumeration by proximity search. In Proceedings of the 51st Annual ACM SIGACT Symposium on Theory of Computing, STOC 2019, pages 1179–1190, New York, NY, USA, 2019. Association for Computing Machinery.
* [GKKL22] Petr A. Golovach, Christian Komusiewicz, Dieter Kratsch, and Van Bang Le. Refined notions of parameterized enumeration kernels with applications to matching cut enumeration. Journal of Computer and System Sciences, 123:76–102, 2022.
* [JYP88] David S. Johnson, Mihalis Yannakakis, and Christos H. Papadimitriou. On generating all maximal independent sets. Information Processing Letters, 27(3):119–123, 1988.
* [KBB+09] Leonid Khachiyan, Endre Boros, Konrad Borys, Vladimir Gurvich, and Khaled Elbassioni. Generating all vertices of a polyhedron is hard. Twentieth Anniversary Volume: Discrete & Computational Geometry, pages 1–17, 2009.
* [KBE+05] Leonid Khachiyan, Endre Boros, Khaled Elbassioni, Vladimir Gurvich, and Kazuhisa Makino. On the complexity of some enumeration problems for matroids. SIAM Journal on Discrete Mathematics, 19(4):966–984, 2005.
* [MS19] Arnaud Mary and Yann Strozecki. Efficient enumeration of solutions produced by closure operations. Discret. Math. Theor. Comput. Sci., 21(3), 2019.
* [Str19] Yann Strozecki. Enumeration complexity. Bulletin of EATCS, 1(129), 2019.
* [TIAS77] Shuji Tsukiyama, Mikio Ide, Hiromu Ariyoshi, and Isao Shirakawa. A new algorithm for generating all the maximal independent sets. SIAM Journal on Computing, 6(3):505–517, 1977.
|
# Freeze-in bino dark matter in high scale supersymmetry
Chengcheng Han Peiwen Wu Jin Min Yang Mengchao Zhang
###### Abstract
We explore a scenario of high scale supersymmetry where all supersymmetric
particles except gauginos stay at a high energy scale $M_{\rm SUSY}$ which is
much larger than the reheating temperature $T_{\text{RH}}$. The dark matter is
dominated by bino component with mass around the electroweak scale and the
observed relic abundance is mainly generated by the freeze-in process during
the early universe. Considering the various constraints, we identify two
available scenarios in which the supersymmetric sector at an energy scale
below $T_{\text{RH}}$ consists of: a) bino; b) bino and wino. Typically, for a
bino mass around 0.1-1 TeV and a wino mass around 2 TeV, we find that $M_{\rm
SUSY}$ should be around $10^{12-14}$ GeV with $T_{\text{RH}}$ around
$10^{4-6}$ GeV.
## 1 Introduction
Supersymmetry (SUSY) [1, 2, 3, 4, 5, 6] is a significant theoretical framework
aiming at extending the Standard Model (SM), drawing inspiration from the
pursuit of a quantum gravity theory, particularly within the context of
superstring theory. In the field of phenomenology, SUSY not only provides a
viable candidate for dark matter (DM) which plays a crucial role in the
formation of large-scale structures in the universe, but also contributes to
the renormalization group running of gauge couplings through the inclusion of
additional particles near the electroweak scale. This property of SUSY
facilitates the potential unification of the three fundamental forces at high
energy scales. It has long been postulated that SUSY DM takes the form of
Weakly Interacting Massive Particles (WIMPs) that can be probed through
diverse experiments [7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17]. However, the
absence of confirmed DM signals poses significant challenges to the standing
of SUSY DM. The current LHC search results indicate that SUSY particles seem
to be heavier than the electroweak (EW) scale [18, 19], thus challenging the
WIMP paradigm of SUSY (for recent reviews on SUSY in light of current
experiments, see, e.g., [20, 21, 22]).
Given the current situation, in this study we consider an alternative scenario
of SUSY DM in which gauginos are located at a low energy scale while all other
SUSY partners exist at a significantly higher scale $M_{\rm SUSY}$. This
scenario is a special case of the Split SUSY [23, 24, 25, 26] where higgsinos
are also taken to be a similar scale as sfermions. One should note that the
Higgs sector in this scenario is fine-tuned [27, 28, 29, 30, 31] and it might
be a consequence of the anthropic principle. However, in this work we will
assume that SUSY still provides a candidate of DM and we will specifically
consider the Minimal Supersymmetric Standard Model (MSSM). Since the
measurement of gamma-ray from the MAGIC [32] has strongly constrained the
possibility of wino DM111There is still viable parameter space for wino dark
matter assuming core profile of the DM., the only viable DM candidate in the
MSSM is bino. However, it is widely known that pure bino DM is typically
overabundant from the freeze-out mechanism [33] due to its weak coupling with
the visible sector [34, 35]. Alternatively, a bino particle with a rather weak
coupling may serve as a suitable candidate for Feebly Interacting Massive
Particle (FIMP) DM with a correct relic abundance generated via the freeze-in
mechanism [36], with assumptions that the reheating process solely occurs in
the Standard Model (SM) sector and the reheating temperature $T_{\text{RH}}$
is lower than the SUSY scale $M_{\text{SUSY}}$ .
In this work we study the possibility that the bino DM in MSSM is generated
via the freeze-in process during the early universe. We assume that all MSSM
particles except gauginos share similar mass $M_{\rm SUSY}$ which is much
higher than the reheating temperature $T_{\text{RH}}$ of the universe. To
generate enough relic abundance of bino dark matter, we always require the
bino mass lower than the reheating temperature. While for the mass of wino or
gluino, they could be either higher or lower than the reheating temperate
$T_{\text{RH}}$ depending on the different scenarios we consider.
The paper is organized as follows. In Section 2 we present the model set up.
In Section 3 we first overview the physics related to dark matter and then
study the dominate channels for bino freeze-in production. In Section 4 we
give the numerical results and discuss the experimental limits on the model
parameter space relevant for our scenarios. We draw the conclusions in Section
5 and leave the calculation details in Appendices.
## 2 Model of heavy supersymemtry
Since we are considering a scenario of high scale supersymmetry in which only
gauginos are at low energy scale, the relevant Lagrangian terms are
$\displaystyle\mathcal{L}$ $\displaystyle\supset$
$\displaystyle-\sum_{f=q,l}M^{2}_{\tilde{f}}\tilde{f}^{\ast}\tilde{f}+\bigg{[}\bigg{(}\sum_{A=1,2,3}-\frac{1}{2}M_{A}\tilde{V}^{A,a}\tilde{V}^{A,a}\bigg{)}-\mu\tilde{H}_{u}\cdot\tilde{H}_{d}+b\mu
H_{u}\cdot H_{d}+h.c.\bigg{]}$ (2.1)
$\displaystyle-\sum_{A=1,2}\sqrt{2}g_{A}\bigg{[}{H}^{\ast}_{u}\bigg{(}T^{A,a}\tilde{V}^{A,a}\bigg{)}\tilde{H}_{u}+{H}^{\ast}_{d}\bigg{(}T^{A,a}\tilde{V}^{A,a}\bigg{)}\tilde{H}_{d}+h.c.\bigg{]}$
$\displaystyle-\sum_{A=1,2,3}\sqrt{2}g_{A}\bigg{[}\sum_{f=q,l}\tilde{f}^{\ast}\bigg{(}T^{A,a}\tilde{V}^{A,a}\bigg{)}f+h.c.\bigg{]}$
$\displaystyle-(M^{2}_{H_{u}}+|\mu|^{2})H^{\ast}_{u}H_{u}-(M^{2}_{H_{d}}+|\mu|^{2})H^{\ast}_{d}H_{d}~{},$
where $A=1,2,3$ correspond to the SM gauge group $\rm
U(1)_{Y},SU(2)_{L},SU(3)_{C}$, respectively, and $a$ denotes the corresponding
indices in adjoint representation of group $A$. Fields
$\tilde{V}^{A,a},\tilde{H}_{u},\tilde{H}_{d},\tilde{f}$ are the superpartners
of the SM vector gauge bosons $V^{A,a}=B,W^{1\sim 3},G^{1\sim 8}$, scalar
doublets $H_{u},H_{d}$ and fermions $f$. The fields $H_{u}$, $H_{d}$,
$\tilde{H}_{u}$, $\tilde{H}_{d}$ are defined as
$\displaystyle H_{u}$ $\displaystyle=$
$\displaystyle\left(\begin{array}[]{c}H_{u}^{+}\\\
H_{u}^{0}\end{array}\right),\quad\tilde{H}_{u}=\left(\begin{array}[]{c}\tilde{H}_{u}^{+}\\\
\tilde{H}_{u}^{0}\end{array}\right),\quad
H_{d}=\left(\begin{array}[]{c}H_{d}^{0}\\\
H_{d}^{-}\end{array}\right),\quad\tilde{H}_{d}=\left(\begin{array}[]{c}\tilde{H}_{d}^{0}\\\
\tilde{H}_{d}^{-}\end{array}\right).$ (2.10)
For the Higgs sector, we need a SM-like Higgs boson $H_{\text{SM}}$ near the
electroweak scale [37, 38]. This is obtained from the mixing between the two
Higgs doublets $H_{u}$ and $H_{d}$ in the MSSM:
$\displaystyle H_{u}=\left(\begin{array}[]{c}H_{u}^{+}\\\
H_{u}^{0}\end{array}\right)$ $\displaystyle=$ $\displaystyle\sin\beta\,H_{\rm
SM}+\cos\beta\,H_{\rm NP}=\sin\beta\left(\begin{array}[]{c}G_{\rm SM}^{+}\\\
H_{\rm SM}^{0}\end{array}\right)+\cos\beta\left(\begin{array}[]{c}H_{\rm
NP}^{+}\\\ H_{\rm NP}^{0}\end{array}\right)~{},$ (2.17)
$\displaystyle(-i\sigma^{2})H^{\ast}_{d}=\left(\begin{array}[]{c}-(H_{d}^{-})^{*}\\\
(H_{d}^{0})^{*}\end{array}\right)$ $\displaystyle=$
$\displaystyle\cos\beta\,H_{\rm SM}-\sin\beta\,H_{\rm
NP}=\cos\beta\left(\begin{array}[]{c}G_{\rm SM}^{+}\\\ H_{\rm
SM}^{0}\end{array}\right)-\sin\beta\left(\begin{array}[]{c}H_{\rm NP}^{+}\\\
H_{\rm NP}^{0}\end{array}\right)~{},$ (2.24)
where $\sigma^{2}$ is the second Pauli matrix, and $\tan\beta=\langle
H^{0}_{u}\rangle/\langle H^{0}_{d}\rangle$ with $\langle H^{0}_{u}\rangle$ and
$\langle H^{0}_{d}\rangle$ being the vacuum expectation values (VEVs). Such
mixings can be realized by properly choosing Higgs mass parameters
$\mu,M_{H_{u}},M_{H_{d}}$, and $b$. The subscription "NP" in $H_{\text{NP}}$
denotes the new physics (NP) Higgs doublet in the MSSM accompanying the SM
one222Note that in order not to increase the complexity of notation, we don’t
further perform the expansion of the complex but electrically neutral scalars
$H_{\rm SM}^{0},H_{\rm NP}^{0}$ into real and imaginary parts. However, one
needs to beware that $G_{\rm SM}^{\pm},H_{\rm SM}^{0}$ contain the Goldstone
boson modes to be absorbed into vector gauge bosons $W^{\pm},Z^{0}$ after the
electroweak symmetry breaking (EWSB).. Since the mass parameters $M_{H_{u}}$,
$M_{H_{d}}$, $b$, $\mu$ are all much larger than the electroweak scale, a
tuning of these parameters are needed to get a light Higgs at electroweak
scale [27, 28, 29, 30, 31]. We need also match the Higgs self-coupling to be
the SUSY value at the scale of $M_{\rm SUSY}$,
$\displaystyle\lambda(M_{\rm
SUSY})=\frac{{g^{\prime}_{1}}^{2}+g_{2}^{2}}{4}\cos^{2}2\beta~{}.$ (2.25)
Note that the Higgs self-coupling $\lambda$ becomes very small at high energy
scale due to the RGE running, and thus the $\beta$ value should get close to
$\pi/4$ and $\tan\beta\approx 1$. We will fix $\tan\beta=1$ as the benchmark
parameter throughout this work for simplicity.
Generally, when considering physical processes at temperature $T\ll
M_{\text{SUSY}}$, we can integrate out the heavy mediators with mass
$\mu,M_{\tilde{f}}\sim M_{\text{SUSY}}\gg T_{\text{RH}}$ and get the following
effective operators at the level of dimension 5 and 6, respectively,
$\displaystyle{\text{dimension-5:}}\quad\propto\quad\frac{1}{\mu}\,|H_{\rm
SM}|^{2}(\tilde{B}\tilde{B},\tilde{B}\tilde{W})~{},$ (2.26)
$\displaystyle{\text{dimension-6:}}\quad\propto\quad\frac{1}{M^{2}_{\tilde{f}}}(f^{\dagger}\tilde{B}^{\dagger})(f\tilde{B},f\tilde{W},f\tilde{G})~{}.$
(2.27)
Since we assume the mass parameters of higgsinos $\mu$ and sfermions
$M_{\tilde{f}}$ around $M_{\text{SUSY}}$, the dominant process would be from
the dimension-5 (dim-5) operators. Nevertheless we also present the processes
related to dim-6 operators for completeness.
We acknowledge that a majority of the significant processes are evaluated at
energy scales considerably beneath $M_{\text{SUSY}}$. The recommended approach
entails initiating the integration procedure for the massive particle to
derive the effective operators of dimension 5 and 6, along with their
corresponding Wilson coefficients, within the realm of $M_{\text{SUSY}}$.
Subsequently, the computation of these Wilson coefficients at the pertinent
scale is achieved by employing the Renormalization Group Equations to track
the evolution of the operators. Notably, there exists a potential correction
to the primary outcome, potentially on the order of $O(1)$, yet the
fundamental framework remains robust. We leave the investigation of this
effect for future study.
## 3 Freeze-in bino dark matter in MSSM
### 3.1 Particle spectrum
Despite the existence of new Higgs bosons and many supersymmetric partners of
the SM particles, the MSSM particle spectrum we consider in this work consist
of two sectors distinguished by their characteristic mass scales. Although not
making significant difference for the mass spectrum structure before and after
EWSB, we take the pre-EWSB case as an illustration.
* •
Heavy sector, inactive after cosmological reheating
Mass: $M\sim M_{\text{SUSY}}\gg T_{RH}$
* $\star$
Higgs bosons not in SM: $H^{0}_{\text{NP}}$, $A$, $H^{\pm}_{\text{NP}}$
* $\star$
Sfermions $\tilde{f}$
* $\star$
Higgsinos $\tilde{H}_{u},\tilde{H}_{d}$
* •
Light sector, active after cosmological reheating
Mass: $M\sim\mathcal{O}(1)\,\text{TeV}\ll M_{\text{SUSY}}$
* $\star$
SM particles
* $\star$
Bino $\tilde{B}$, consisting cosmological DM with mass $M_{1}<T_{RH}$
* $\star$
Winos $\tilde{W}$, with mass $M_{2}$
* $\star$
Gluinos $\tilde{G}$, with mass $M_{3}$
In the above we utilized gauge eigenstates for description, since
$\tilde{B},\tilde{W}$ do not mix with higgsinos $\tilde{H}_{u},\tilde{H}_{d}$
before EWSB when the SM Higgs $H_{\text{SM}}$ has not acquired the VEV.
### 3.2 Bino production from freeze-in mechanism
In the early stage of universe before EWSB when the gaugino states
$\tilde{B},\tilde{W}$ do not mix with higgsinos $\tilde{H}_{u},\tilde{H}_{d}$,
pure $\tilde{B}$ acting as DM can only interact with SM via mediators with
heavy mass near the scale $M_{\text{SUSY}}$, as shown in Fig. 1. Due to the
suppressed interacting strength, the cosmological production of bino DM in our
scenario proceed via the freeze-in mechanism. In the follows we consider the
contributions to bino DM production from several typical processes333After
electroweak phase transition occurs and $H_{\text{SM}}$ acquires VEV, the top
and bottom vertex in the left panel of Fig.1 imply the mixing between
$\tilde{B},\tilde{W}$ and $\tilde{H}_{u},\tilde{H}_{d}$, resulting in the mass
eigenstates of electrically neutral neutralinos $\tilde{\chi}^{0}_{1,2,3,4}$
and charged $\tilde{\chi}^{\pm}_{1,2}$ (see discussions in Section 4.1)..
Figure 1: Schematic plots for interactions of DM composed of pure $\tilde{B}$
with SM after cosmological reheating considered in this work, which would
induce dimension-5 (left) and dimension-6 (right) effective operators. The SM
Higgs $H_{\text{SM}}$ originates from the mixing between MSSM Higgs doublets
$H_{u},H_{d}$. Colored lines indicate the direction of freeze-in production
when applicable. Additional Hermitian conjugated processes also exist when the
amplitudes are complex. See more discussions in the main texts.
### 3.3 Case I: bino freeze-in from $HH^{\ast}\to\tilde{B}\tilde{B}$
This case corresponds to the left panel of Fig. 1 but without winos
$\tilde{W}$. After integrating out the heavy higgsinos, the relevant dim-5
effective interaction is given by (the details are given in Appendix A)
$\displaystyle\mathcal{L}^{\text{eff}}_{HH^{\ast}\to\tilde{B}\tilde{B}}=\frac{2g_{1}^{2}\,Y_{H}^{2}}{\mu}\sin\beta\cos\beta(|H_{\rm
SM}|^{2})(\tilde{B}\tilde{B}+\tilde{B}^{\dagger}\tilde{B}^{\dagger})~{},$
(3.1)
where $|H_{\rm SM}|^{2}=G_{\rm SM}^{+}(G_{\rm SM}^{+})^{\ast}+(H_{\rm
SM}^{0})(H_{\rm SM}^{0})^{*}$. In the subscription
$HH^{\ast}\to\tilde{B}\tilde{B}$ on the left side (and hereafter when not
causing any confusion), we denote $H_{\text{SM}}$ as $H$ to simplify the
notation, and all fields in the initial and final states of the process should
be understood in the sense of physical particles444Discussion on the naming
convention of particles, states and filed can be found in, e.g. [39].. With
more details given in Appendix B, Eq.(3.1) would induce the Boltzmann equation
of the bino number density:
$\displaystyle\frac{d}{dt}n_{\tilde{B}}+3\mathcal{H}n_{\tilde{B}}=\textbf{C}_{HH^{\ast}\to\tilde{B}\tilde{B}}\approx\frac{g_{1}^{4}}{4}\frac{1}{\pi^{5}}\frac{\,\sin^{2}\beta\cos^{2}\beta}{\mu^{2}}T^{6}~{}.$
(3.2)
The above equation can be modified to a differential equation about bino yield
$Y_{\tilde{B}}=n_{\tilde{B}}/S$ ($S$ is the entropy density) and temperature
$T$:
$\displaystyle\frac{dY_{HH^{\ast}\to\tilde{B}\tilde{B}}(T)}{dT}=-\frac{\textbf{C}_{HH^{\ast}\to\tilde{B}\tilde{B}}}{ST\mathcal{H}}$
$\displaystyle\approx-(1.25\times 10^{-3})\times
M_{\text{Pl}}\frac{\textbf{C}_{HH^{\ast}\to\tilde{B}\tilde{B}}}{T^{6}}\approx-(1\times
10^{-6})\times M_{\text{Pl}}\
g_{1}^{4}\frac{\,\sin^{2}\beta\cos^{2}\beta}{\mu^{2}}~{},$
where $M_{\text{Pl}}\approx 1.22\times 10^{19}$ GeV is the Planck mass,
$S=2\pi^{2}g_{\ast}T^{3}/45$, and Hubble expansion rate $\mathcal{H}\approx
1.66\sqrt{g_{\ast}}T^{2}/M_{\text{Pl}}$ with $g_{\ast}=106.75$ before EWSB.
Performing a simple integral from reheating temperature, it can be found that
the final yield of $\tilde{B}$ depends on the reheating temperature
$T_{\text{RH}}$ which corresponds to the Ultraviolet (UV) freeze-in scenario
[36, 40]:
$\displaystyle Y_{HH^{\ast}\to\tilde{B}\tilde{B}}(\infty)\approx(1\times
10^{-6})\times M_{\text{Pl}}\
g_{1}^{4}\frac{\,\sin^{2}\beta\cos^{2}\beta}{\mu^{2}}\ T_{\text{RH}}~{},$
(3.3)
and the corresponding current relic abundance is given by
$\displaystyle\left(\Omega_{\tilde{B}}h^{2}\right)_{HH^{\ast}\to\tilde{B}\tilde{B}}=M_{1}\
\frac{Y_{HH^{\ast}\to\tilde{B}\tilde{B}}(\infty)S_{0}}{\rho_{cr}}\approx
Y_{HH^{\ast}\to\tilde{B}\tilde{B}}(\infty)\left(\frac{M_{1}}{\text{TeV}}\right)\times(2.72\times
10^{11})~{}.$
### 3.4 Case II: fermion scattering process $f\bar{f}\to\tilde{B}\tilde{B}$
After integrating out sfermions with heavy mass $M_{\tilde{q},\tilde{l}}\sim
M_{\text{SUSY}}$ in the right panel of Fig.1, the effective interactions
between SM fermion pair and $\tilde{B}$ pair have the following form at
dimention 6 (for more details, see Appendix C):
$\displaystyle\mathcal{L}^{\text{eff}}_{f\bar{f}\to\tilde{B}\tilde{B}}=\sum_{f=q,l}\frac{(\sqrt{2}g_{1}Y_{f})(\sqrt{2}g_{1}Y_{f})}{M^{2}_{\tilde{f}}}(f^{\dagger}\tilde{B}^{\dagger})(f\tilde{B})~{},$
(3.5)
where for simplicity, we consider an universal mass for all the fermions, i.e.
$M_{\tilde{f}}\equiv M_{\tilde{q}}=M_{\tilde{l}}$.
Thus the Boltzmann equation is
$\displaystyle\frac{dY_{f\bar{f}\to\tilde{B}\tilde{B}}(T)}{dT}=-\frac{\textbf{C}_{f\bar{f}\to\tilde{B}\tilde{B}}}{ST\mathcal{H}}$
$\displaystyle\approx-(1.25\times 10^{-3})\times
M_{\text{Pl}}\frac{\textbf{C}_{f\bar{f}\to\tilde{B}\tilde{B}}}{T^{6}}\approx-(8.6\times
10^{-5})\times M_{\text{Pl}}\frac{g_{1}^{4}}{M^{4}_{\tilde{f}}}T^{2}~{},$
(3.6)
and correspondingly,
$\displaystyle Y_{f\bar{f}\to\tilde{B}\tilde{B}}(\infty)$
$\displaystyle\approx$ $\displaystyle(4.7\times
10^{-7})\times\frac{M_{\text{Pl}}}{M^{4}_{\tilde{f}}}\ T^{3}_{\text{RH}}~{},$
(3.7)
$\displaystyle\left(\Omega_{\tilde{B}}h^{2}\right)_{f\bar{f}\to\tilde{B}\tilde{B}}$
$\displaystyle=$ $\displaystyle M_{1}\
\frac{Y_{f\bar{f}\to\tilde{B}\tilde{B}}(\infty)S_{0}}{\rho_{cr}}\approx
Y_{f\bar{f}\to\tilde{B}\tilde{B}}(\infty)\left(\frac{M_{1}}{\text{TeV}}\right)\times(2.72\times
10^{11})~{}.$ (3.8)
### 3.5 Case III: gluino/wino scattering or decay processes
As indicated by blue colored arrows in Fig. 1, the $2\to 2$ scattering
processes consist of two ways of generating bino DM when combining $\rm
U(1)_{Y}$ with $\rm SU(2)_{L}$ or $\rm SU(3)_{C}$ interactions, related by the
cross symmetry. Moreover, we can also have the red colored arrow indicating
$1\to 3$ ($1\to 2$) decay processes generating binos before (after) EWSB when
the cosmological temperature drops below the scale of $M_{2}$ or $M_{3}$
(equivalently, when the age of the universe reach the lifetime of $\tilde{W}$
and $\tilde{G}$).
Similar to the previous two cases, integrating out heavy higgsino and
sfermions would generate the following dim-5 and dim-6 effective operators:
$\displaystyle\mathcal{L}^{\text{eff}}_{\text{case-III}}=$
$\displaystyle\bigg{\\{}-\sum_{b=1}^{3}\frac{(\sqrt{2}g_{1}Y_{H})(\sqrt{2}g_{2})}{\mu}\sin\beta\cos\beta(H^{\ast}\frac{1}{2}\sigma^{b}H)(\tilde{B}\tilde{W}^{b})$
(3.9)
$\displaystyle+\sum_{f=u_{L},d_{L},e_{L},\nu}\quad\sum_{b=1}^{3}\frac{(\sqrt{2}g_{1}Y_{f})(\sqrt{2}g_{2})}{M^{2}_{\tilde{f}}}(f^{\dagger}\tilde{B}^{\dagger})(\frac{1}{2}\sigma^{b}f\tilde{W}^{b})$
$\displaystyle+\sum_{f=u_{L},d_{L},{u^{\dagger}_{R}},{d^{\dagger}_{R}}}\quad\sum_{a=1}^{8}\frac{(\sqrt{2}g_{1}Y_{f})(\sqrt{2}g_{3})}{M^{2}_{\tilde{f}}}(f^{\dagger}\tilde{B}^{\dagger})(\frac{1}{2}\lambda^{a}f\tilde{G}^{a})\bigg{\\}}+h.c.~{}.$
Note that the index $f$ in the second line includes only $\rm SU(2)_{L}$
doublets, while the index $f$ in the third line includes only quarks. To
highlight the difference, we use index $a$ and $b$ to denote generators of
$\rm SU(3)_{C}$ and $\rm SU(2)_{L}$ interactions, respectively.
Correspondingly, $\lambda^{a}$ and $\sigma^{b}$ are Gell-Mann and Pauli
matries, respectively.
In the following, we consider the contributions to the bino DM production from
$2\to 2$ scattering and $1\to 3$ decay separately, while leaving the effects
of $1\to 2$ decay appearing after EWSB in Section 4.1.
#### 3.5.1 Case III A: $2\to 2$ scattering involving gluino/wino
With more detailed given in Appendix D, the collision terms in the Boltzmann
equation for dim-5 and dim-6 operators are approximated as (ignoring the
masses of all external particles)
$\displaystyle\textbf{C}_{\rm dim-5}$ $\displaystyle=$
$\displaystyle\frac{T}{2048\pi^{6}}\int^{\infty}_{4M_{1}^{2}}ds\
(s-4M_{1}^{2})^{1/2}K_{1}({\sqrt{s}}/{T})\sum_{\rm
internal\,d.o.f}\int\,d\Omega$ (3.10)
$\displaystyle\times\bigg{(}|\mathcal{M}|^{2}_{HH^{\ast}\to\tilde{B}\tilde{B}}+|\mathcal{M}|^{2}_{HH^{\ast}\to\tilde{B}\tilde{W}}+N_{\text{conj}}|\mathcal{M}|^{2}_{\tilde{W}H\to\tilde{B}H}\bigg{)}$
$\displaystyle=\bigg{(}\frac{1}{4}g_{1}^{4}+\frac{3}{2}g_{1}^{2}g_{2}^{2}\bigg{)}\frac{1}{\pi^{5}}\frac{\,\sin^{2}\beta\cos^{2}\beta}{\mu^{2}}T^{6}~{},$
$\displaystyle\textbf{C}_{\rm dim-6}$ $\displaystyle=$
$\displaystyle\frac{T}{2048\pi^{6}}\int^{\infty}_{4M_{1}^{2}}ds\
(s-4M_{1}^{2})^{1/2}K_{1}({\sqrt{s}}/{T})\sum_{\rm
internal\,d.o.f}\int\,d\Omega$ (3.11)
$\displaystyle\times\bigg{(}|\mathcal{M}|^{2}_{f\bar{f}\to\tilde{B}\tilde{B}}+|\mathcal{M}|^{2}_{f\bar{f}\to\tilde{B}\tilde{W}}+N_{\text{conj}}|\mathcal{M}|^{2}_{\tilde{W}f\to\tilde{B}f}+|\mathcal{M}|^{2}_{f\bar{f}\to\tilde{B}\tilde{G}}+N_{\text{conj}}|\mathcal{M}|^{2}_{\tilde{G}f\to\tilde{B}f}\bigg{)}$
$\displaystyle=$
$\displaystyle(\frac{190}{9}g^{4}_{1}+30g_{1}^{2}g_{2}^{2}+\frac{440}{3}g_{1}^{2}g_{3}^{2})\frac{1}{\pi^{5}}\frac{1}{M^{4}_{\tilde{f}}}T^{8}~{},$
where $N_{\text{conj}}=2$ denotes the effects of conjugated process.
#### 3.5.2 Case III B: decay of gluino/wino
Following the method in [36] with $f_{\tilde{G}}$ and $f_{\tilde{W}}$
approximated by $e^{-E_{\tilde{G}}/T}$ and $e^{-E_{\tilde{W}}/T}$, the
Boltzmann equation of freeze-in production for the $1\to 3$ decay processes is
$\displaystyle\frac{d}{dt}n_{\tilde{B}}+3\mathcal{H}n_{\tilde{B}}=\textbf{C}$
$\displaystyle\approx\frac{g_{\tilde{G}}M_{3}^{2}}{2\pi^{2}}TK_{1}(\frac{M_{3}}{T})\Gamma_{\tilde{G}\to
f\bar{f}\tilde{B}}+\frac{g_{\tilde{W}}M_{2}^{2}}{2\pi^{2}}TK_{1}(\frac{M_{2}}{T})(\Gamma_{\tilde{W}\to
f\bar{f}\tilde{B}}+\Gamma_{\tilde{W}\to HH^{\ast}\tilde{B}})~{},$ (3.12)
where $g_{\tilde{G}}=16$ and $g_{\tilde{W}}=6$ are the internal d.o.f. of
$\tilde{G}$ and $\tilde{W}$, respectively. The expressions of decay width
involved in the above results are listed in Appendix E. Changing variables to
yield $Y_{\tilde{B}}$ and temperature $T$, we then integrate over temperature
evolution to obtain the final yield. If reheating temperature $T_{\text{RH}}$
is much larger than $M_{2}$ and $M_{3}$, then the final yield from $1\to 3$
decay can be approximated by
$\displaystyle Y^{\rm 1\to
3}_{\tilde{B}}(\infty)\approx\int^{T_{\text{RH}}}_{T_{min}}\frac{\textbf{C}}{ST\mathcal{H}}dT$
$\displaystyle\approx(3\times 10^{-4})\times
M_{\text{Pl}}\left(\frac{1}{M^{2}_{3}}g_{\tilde{G}}\Gamma_{\tilde{G}\to
f\bar{f}\tilde{B}}+\frac{1}{M^{2}_{2}}g_{\tilde{W}}\Gamma_{\tilde{W}\to
f\bar{f}\tilde{B}}+\frac{1}{M^{2}_{2}}g_{\tilde{W}}\Gamma_{\tilde{W}\to
HH^{\ast}\tilde{B}}\right)~{}.$ (3.13)
It is worth pointing out that the above result is not sensitive to
$T_{\text{RH}}$. Taking a low reheating temperature $T_{\text{RH}}=1.1\,M_{3}$
as an example, increasing the value of $T_{\text{RH}}$ does no modify the
result significantly.
In addition to the $1\to 3$ decay, we should also note that wino $\tilde{W}$
with mass $M_{2}<T_{\text{RH}}$ keeps staying in the thermal bath until
reaching its freeze-out moment yielding a relic wino number density, which
would later convert to the equal amount of bino number density $n_{\tilde{B}}$
via $1\to 2$ decay $\tilde{W}\to\tilde{B}+h$ after EWSB occurs. Depending on
the bino mass $M_{1}$, this freeze-out component would also contribute to the
total bino DM abundance in today’s epoch. We checked that with wino mass
$M_{2}=2$ TeV, the $1\to 2$ decay contribution of $Y^{1\to 2}_{\tilde{B}}$ to
final bino yield is around $25\%$ ($1\%$) on the percentage level for $M_{1}=$
1(0.1) TeV [41], thus not affecting the freeze-in domination scenario of this
work. We properly include the wino freeze-out contribution in our results.
There is also contribution from gluino late time decay. However, to avoid the
constraints from BBN, we have to set the gluino mass higher than the
$T_{\text{RH}}$, thus we do not include its contribution here.
## 4 Numerical results and discussion
In Fig. 2 we show the required scales of $\mu$ ($M_{\tilde{f}}$) for dim-5(6)
operators with various $T_{\text{RH}}$ to produce the observed bino DM relic
abundance. The upper (lower) two lines correspond to dim-5 (6) operators. We
can see that due to the more suppression of dim-6 operators, the needed
$M_{\tilde{f}}$ are generally $\mathcal{O}(10^{-4})$ smaller than $\mu$ in the
dim-5 case. If we assume $\mathcal{O}(\mu)\approx\mathcal{O}(M_{\tilde{f}})$,
in order not to overclose the Universe, the dim-6 contributions would be
completely negligible.
From Fig. 2, we can see that for the case $M_{\tilde{B}}<T_{RH}\ll
M_{\tilde{W}}$, the dominant production of bino dark matter is from the
process $HH^{*}\rightarrow\tilde{B}\tilde{B}$ from the dim-5 operator.
Generally, $M_{\rm SUSY}$ should be around $10^{13-14}$ GeV for
$T_{RH}<10^{6}$ GeV. Since the final relic abundance is proportional to
${T_{RH}}/{\mu^{2}}$, the $M_{\rm SUSY}$ could continue increasing if the
reheating temperature $T_{\text{RH}}$ becomes higher. Note that this is
similar to the model of Higgs portal to fermion dark matter which are studied
in [42], with which we find our result are consistent. We emphasize that our
model is motivated by a more complete framework and [42] falls into one of
cases we consider. Moreover, For the case
$M_{\tilde{B}},M_{\tilde{W}}<T_{RH}$, we find the wino-included process can
largely enhance the annihilate rate and a higher scale is needed to satisfy
the relic abundance. In this case, $M_{\rm SUSY}$ should be around
$10^{14-15}$ GeV for $T_{RH}<10^{6}$ GeV.
Notice that if the gluino is in the thermal equilibrium with SM in the early
universe and the sfermions mediating the gluino decay are heavier than
$10^{9}$ GeV, gluino’s lifetime could be longer than the age of Universe when
the big bang nucleosynthesis (BBN) happens, leading to energy injection into
the cosmic plasma and altering the BBN profile. In all cases considered in
this work we find $M_{\rm SUSY}$ is much larger than $10^{9}$ GeV, therefore
we always need $M_{\tilde{G}}\gg T_{RH}$ to avoid the limit from BBN [43].
More discussions on BBN limits are given in 4.1.
Figure 2: Values of $\mu$ and $M_{\tilde{f}}$ to produce the observed DM
abundance via the UV freeze-in processes. See more discussions in the main
texts.
In Fig. 3 we show the comparison of final contributions and intermediate
profile of UV and IR freeze-in processes to the bino DM relic abundance. It
can be clearly seen that the IR freeze-in final yields from wino 3-body decays
are negligible compared to that of UV freeze-in processes generated by $2\to
2$ annihilation. Moreover, the critical production moment determining the
final yield of UV freeze-in locates in a much smaller $x$ (and thus much
higher temperature) than the IR freeze-in case.
Figure 3: Comparison between UV freeze-in and IR freeze-in. Note the
difference between temperatures indicated by $x=M_{2}/T$ producing the correct
relic density of bino DM.
### 4.1 Limits from BBN
After EWSB, the SM-like Higgs doublet needs to be replaced by:
$\displaystyle H=\left(\begin{array}[]{c}G^{+}\\\
\frac{1}{\sqrt{2}}(v+h+iG^{0})\end{array}\right)~{},$ (4.3)
where $v=246$ GeV is the VEV of SM Higgs 555If wino decays much later than
electroweak phase transition, then $v=246$ GeV is a good approximation. and
$h$ is the observed SM-like Higgs scalar. $G^{\pm}$ ($G^{-}=(G^{+})^{\ast}$)
and $G^{0}$ are Goldstone bosons that form the longitudinal modes of SM gauge
bosons $W^{\pm}$ and $Z$. As mentioned earlier, the SM-like Higgs VEV will
generate mixings among the gauge states
$\tilde{B},\tilde{W},\tilde{H}_{u},\tilde{H}_{d}$ and form mass eigenstates of
charge-neutral neutralinos $\tilde{\chi}_{1,2,3,4}$ and charged charginos
$\tilde{\chi}^{\pm}_{1,2}$ (with ascending mass order inside sectors of
neutralinos and charginos, respectively). For the scenario considered in this
work, the component of neutralino $\tilde{\chi}^{0}_{1}$
($\tilde{\chi}^{0}_{2}$) is dominated by bino $\tilde{B}$ (wino
$\tilde{W}^{3}$), and component of chargino $\tilde{\chi}^{\pm}_{1}$ is
dominated by winos $\frac{1}{\sqrt{2}}(\tilde{W}^{1}\mp i\tilde{W}^{2})$. More
details of the approximated masses and couplings can be found in [44, 45, 46].
In the following, we would utilize the language of gauge states (bino
$\tilde{B}$, wino $\tilde{W}$, higgsinos $\tilde{H}_{u},\tilde{H}_{d}$ ) and
mass eigenstates (neutralino $\tilde{\chi}^{0}$, chargino
$\tilde{\chi}^{\pm}$) interchangeably before and after EWSB.
Now we study the limit of BBN on our scenario from lifetimes of neutralinos,
charginos. In our scenario, only neutralino
$\tilde{\chi}^{0}_{1}\approx\tilde{B},\
\tilde{\chi}^{0}_{2}\approx\tilde{W}^{3}$ and chargino
$\tilde{\chi}^{\pm}_{1}\approx\frac{1}{\sqrt{2}}(\tilde{W}^{1}\mp
i\tilde{W}^{2})$ existed in the primordial thermal bath. Due to the loop
induced mass-splitting between $\tilde{\chi}^{\pm}_{1}$ and
$\tilde{\chi}^{0}_{2}$, chargino $\tilde{\chi}^{\pm}_{1}$ can have the 2-body
decay $\tilde{\chi}^{\pm}_{1}\to\tilde{\chi}^{0}_{2}\pi^{\pm}$ [47, 48, 49,
50]. It makes the lifetime of $\tilde{\chi}^{\pm}_{1}$ much shorter than 1
sec, and thus not affecting the BBN profile. However, we need to scrutinize
the lifetime of $\tilde{\chi}^{0}_{2}$ more carefully. If
$\tilde{\chi}^{0}_{2}$ decays after the onset of BBN, then the highly
energetic decay products will cause the photodissociation or hadrodissociation
and thus change the final abundances of light elements. So a bound from BBN
can be put on the model parameters, especially on the SUSY scale
$M_{\text{SUSY}}$ [51, 52].
It is easy to see that Fig. 1 implies the 2-body decay mode of
$\tilde{\chi}^{0}_{2}\to\tilde{\chi}^{0}_{1}h$ at the level of dim-5 after
EWSB, in which case we will have:
$\displaystyle\mathcal{L}_{\text{eff}}$ $\displaystyle=$
$\displaystyle-\sum_{b=1}^{3}\frac{(\sqrt{2}g_{1}Y_{H})(\sqrt{2}g_{2})}{\mu}\sin\beta\cos\beta(H^{\ast}\frac{1}{2}\sigma^{b}H)(\tilde{B}\tilde{W}^{b})+h.c.$
(4.4) $\displaystyle=$
$\displaystyle-\frac{g_{1}g_{2}v}{2\mu}\sin\beta\cos\beta(G^{\mp}\tilde{W}^{\pm}\tilde{B}-h\tilde{W}^{0}\tilde{B}+h.c.)$
$\displaystyle\approx$
$\displaystyle-\frac{g_{1}g_{2}v}{2\mu}\sin\beta\cos\beta(G^{\mp}\tilde{\chi}^{\pm}_{1}\tilde{\chi}^{0}_{1}-h\tilde{\chi}^{0}_{2}\tilde{\chi}^{0}_{1}+h.c.)~{},$
where the first term containing Goldstone boson $G^{\mp}$ can be understood in
the context of Goldstone equivalence theorem (GET) for
$\tilde{\chi}^{\pm}_{1}\to\tilde{\chi}^{0}_{1}W^{\pm}$. It should be noticed
that Eq.(4.4) does not contain the three-particle coupling
$G^{0}\tilde{W}\tilde{B}$ and thus would not provide a way of inferring the
2-body decay mode $\tilde{\chi}^{0}_{2}\to\tilde{\chi}^{0}_{1}Z$ via the GET.
In fact, $\tilde{\chi}^{0}_{2}\to\tilde{\chi}^{0}_{1}\ Z$ comes from the gauge
covariant kinetic terms of gauginos and higgsinos combined with gaugino
mixings after EWSB. However, the decay width of
$\tilde{\chi}^{0}_{2}\to\tilde{\chi}^{0}_{1}Z$ suffers from an extra
suppression of $\frac{1}{\mu^{2}}$ embedded in the mass mixings compared to
$\tilde{\chi}^{0}_{2}\to\tilde{\chi}^{0}_{1}h$ and thus can be ignored [53] .
Therefore, we have the following dominant 2-body decay (see Appendix F for
more details):
$\displaystyle\Gamma_{\tilde{\chi}^{0}_{2}\to\tilde{\chi}^{0}_{1}h}\approx
M_{2}\frac{1}{16\pi}\left(\frac{v}{\mu}g_{1}g_{2}\sin\beta\cos\beta\right)^{2}\left(1-\frac{M_{1}^{2}}{M_{2}^{2}}\right)\left(1+\frac{M_{1}}{M_{2}}\right)^{2}~{}.$
(4.5)
Using the GET we would obtain the same results for
$\Gamma_{\tilde{\chi}^{\pm}_{1}\to\tilde{\chi}^{0}_{1}W^{\pm}}$ when
neglecting the gauge boson masses.
In this work, we apply the limit of BBN to the requirement that lifetime of
$\tilde{\chi}^{0}_{2}$ must be less than 0.3 second [36]. In Fig. 4 , we show
the interplay between BBN constraints and freeze-in production, where regions
below black lines are allowed while region above blue lines are allowed. We
can see that for bino mass around 0.1-1 TeV, an upper bound of
$M_{\text{SUSY}}\sim 10^{14}$ TeV is needed to satisfy both phenomenological
requirements.
Figure 4: Interplay between BBN constraints and freeze-in production, where
regions below black lines are allowed while region above blue lines are
allowed.
### 4.2 Limits from direct/indirect detection
Our scenario can easily escape from the current limits from the direct and
indirect detection. In the case of direct detection, Eq.(3.1) after EWSB would
generate the $t$-channel scattering of $\tilde{\chi}^{0}_{1}$ with quarks and
gluons in SM neucleons mediated by SM Higgs, of which the event rate is
suppressed by $1/\mu^{2}$ and thus negligibly small. In the case of indirect
detection, which is basically the inversed process of the freeze-in DM
production, would generate cosmic rays via DM pair annihilations
$\tilde{\chi}^{0}_{1}\tilde{\chi}^{0}_{1}\to h^{\ast}\to\text{SM}$ and
$\tilde{\chi}^{0}_{1}\tilde{\chi}^{0}_{1}\to hh\to\text{SM}$, of which the
flux is again suppressed by $1/\mu^{2}$ and thus not violating the current
experimental bounds.
### 4.3 Limits from the LHC
The collider signals of our scenario mainly come from
$pp\to\tilde{\chi}^{\pm}_{1}\tilde{\chi}^{\mp}_{1},\tilde{\chi}^{\pm}_{1}\tilde{\chi}^{0}_{2}$
followed by $\tilde{\chi}^{\pm}_{1}\to\tilde{\chi}^{0}_{2}\pi^{\pm}$ and
$\tilde{\chi}^{0}_{2}\to\tilde{\chi}^{0}_{1}h$ which both generate the long-
lived particle (LLP) signals. The LLP signatures manifest as disappearing
track for $\tilde{\chi}^{\pm}_{1}\to\tilde{\chi}^{0}_{2}\pi^{\pm}$ and
displaced vertices for $\tilde{\chi}^{0}_{2}\to\tilde{\chi}^{0}_{1}h$,
respectively. However,
$\tau_{\tilde{\chi}^{0}_{2}\to\tilde{\chi}^{0}_{1}h}>\mathcal{O}(10^{-2})\,\text{s}$
would make $\tilde{\chi}^{0}_{2}$ traverse through the whole detector before
decaying without leaving any energy deposit in the calorimeters, thus can
easily evade the current ATLAS [54] and CMS [55] searches for displaced vertex
signals at $\sqrt{s}=13\,\text{TeV}$. As for the disappearing track signature
of $\tilde{\chi}^{\pm}_{1}\to\tilde{\chi}^{0}_{2}\pi^{\pm}$, ATLAS [56] and
CMS [57] also performed dedicated searches using dataset at
$\sqrt{s}=13\,\text{TeV}$ and imply that
$\tilde{\chi}^{\pm}_{1},\tilde{\chi}^{0}_{2}$ should be heavier than 500-600
GeV, therefore our benchmark points with $M_{2}=2\,\text{TeV}$ are still
available.
## 5 Conclusion
We studied a scenario of dark matter generated from UV freeze-in mechanism,
realized in the framework of high scale MSSM. The bino is the dark matter
candidate and its relic abundance is generated by the freeze-in processes via
the dim-5 or dim-6 operators. We found that the SUSY scale $M_{\text{SUSY}}$
should be around $10^{13-15}~{}\text{GeV}$ for reheating temperature in the
range of $10^{4-6}~{}\text{GeV}$. We also illustrated the interplay between
BBN constraints from neutral wino decay and the experimentally observed dark
matter relic abundance, implying an upper bound of $M_{\text{SUSY}}$ around
$10^{14}~{}\text{GeV}$ for wino mass around 2 TeV and bino mass of $0.1\sim 1$
TeV.
## Acknowledgments
This work was supported by the Natural Science Foundation of China (NSFC)
under grant numbers 12105118, 11947118, 12075300, 11821505 and 12335005, the
Peng-Huan-Wu Theoretical Physics Innovation Center (12047503), the CAS Center
for Excellence in Particle Physics (CCEPP), and the Key Research Program of
the Chinese Academy of Sciences under grant No. XDPB15. CH acknowledges
support from the Sun Yat-Sen University Science Foundation and the Fundamental
Research Funds for the Central Universities, Sun Yat-sen University under
Grant No. 23qnpy58. PW acknowledges support from Natural Science Foundation of
Jiangsu Province (Grant No. BK20210201), Fundamental Research Funds for the
Central Universities, Excellent Scholar Project of Southeast University (Class
A), and the Big Data Computing Center of Southeast University. PW also
acknowledges his wife for being tolerant and supportive (freezing in or
squeezing in time to give birth to a son the same day the work was finished).
## Appendix A Notation conventions and dim-5 operator in Case I
In Eq.(2.1), the dot product means
$\tilde{H}_{u}\cdot\tilde{H}_{d}=\tilde{H}_{u,i}(i\sigma^{2})^{ij}\tilde{H}_{d,j}=\tilde{H}^{+}_{u}\tilde{H}^{-}_{d}-\tilde{H}^{0}_{u}\tilde{H}^{0}_{d}$
to realize the isospin symmetry ${\rm SU(2)_{L}}$ where $\sigma^{2}$ is the
second Pauli matrix. The Kronecker delta function $\delta_{i}^{\,\,\,j}$
manifests the ${\rm SU(2)_{L}}$-blindness of the ${\rm U(1)_{Y}}$ interactions
under consideration for binos production and $Y_{H_{u}}=+1/2,Y_{H_{d}}=-1/2$
are the hypercharges of doublets $H_{u},H_{d}$, respectively. We follow the
convention of [39] and impose the left-chiral two-component spinor formalism
for higgisnos
$\tilde{H}^{+}_{u},\tilde{H}^{0}_{u},\tilde{H}^{0}_{d},\tilde{H}^{-}_{d}$ and
bino $\tilde{B}$ (as well as winos $\tilde{W}$ and gluinos $\tilde{g}$ in
later discussion). For the Case I in Section 3.2, the relevant Lagrangian
terms are
$\displaystyle\mathcal{L}$ $\displaystyle\supset$
$\displaystyle-\frac{1}{2}M_{1}\tilde{B}\tilde{B}-\mu\left(\tilde{H}^{+}_{u}\tilde{H}^{-}_{d}-\tilde{H}^{0}_{u}\tilde{H}^{0}_{d}\right)+h.c.$
(A.1)
$\displaystyle-\frac{g_{1}}{\sqrt{2}}({H}^{+}_{u})^{\ast}\tilde{H}^{+}_{u}\tilde{B}-\frac{g_{1}}{\sqrt{2}}({H}^{0}_{u})^{\ast}\tilde{H}^{0}_{u}\tilde{B}+\frac{g_{1}}{\sqrt{2}}({H}^{-}_{d})^{\ast}\tilde{H}^{-}_{d}\tilde{B}+\frac{g_{1}}{\sqrt{2}}({H}^{0}_{d})^{\ast}\tilde{H}^{0}_{d}\tilde{B}+h.c.~{},$
After integrating out higgsinos with mass $\mu$, we obtain dim-5 operator
between SM Higgs $H_{\text{SM}}$ and $\tilde{B}$ DM:
$\displaystyle\mathcal{L}^{\text{eff}}_{HH^{\ast}\to\tilde{B}\tilde{B}}$
$\displaystyle=$
$\displaystyle-\frac{(\sqrt{2}g_{1}Y_{H})(\sqrt{2}g_{1}Y_{H})}{\mu}(H_{u}^{\ast}\cdot
H_{d}^{\ast})\tilde{B}\tilde{B}+h.c.$ (A.2) $\displaystyle=$
$\displaystyle-\frac{2g_{1}^{2}\,Y_{H}^{2}}{\mu}\sin\beta\cos\beta(H^{\ast}_{\rm
SM}\cdot i\sigma^{2}H_{\rm
SM})(\tilde{B}\tilde{B}+\tilde{B}^{\dagger}\tilde{B}^{\dagger})$
$\displaystyle=$
$\displaystyle\frac{2g_{1}^{2}\,Y_{H}^{2}}{\mu}\sin\beta\cos\beta(|H_{\rm
SM}|^{2})(\tilde{B}\tilde{B}+\tilde{B}^{\dagger}\tilde{B}^{\dagger})~{},$
where $Y_{H}=|Y_{H_{u}}|=|Y_{H_{d}}|=1/2$ and the dot products are
$H_{u}^{\ast}\cdot
H_{d}^{\ast}=({H}^{+}_{u})^{\ast}({H}^{-}_{d})^{\ast}-({H}^{0}_{u})^{\ast}({H}^{0}_{d})^{\ast}$.
## Appendix B Boltzmann equation and calculation details of freeze-in DM in
Case I
In the homogeneous and isotropic universe, the production of bino is described
by following Boltzmann equation [33]:
$\displaystyle\frac{d}{dt}n_{\tilde{B}}+3\mathcal{H}n_{\tilde{B}}=\textbf{C}~{},$
(B.1)
with $n_{\tilde{B}}$ denoting the number density of bino particle, and
$\mathcal{H}$ is the Hubble expansion rate. Taking
$HH^{\ast}\to{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\tilde{B}\tilde{B}}$
($\tilde{B}$ means the physical bino particle) in Case I of Section 3.2 as an
example, we have [58]
$\displaystyle\textbf{C}_{ij\to kl}$ $\displaystyle=$ $\displaystyle
N\times\frac{1}{S}\times\bigg{\\{}\int\frac{d^{3}p_{i}}{(2\pi)^{3}2E_{i}}\frac{d^{3}p_{j}}{(2\pi)^{3}2E_{j}}\frac{d^{3}p_{k}}{(2\pi)^{3}2E_{k}}\frac{d^{3}p_{l}}{(2\pi)^{3}2E_{l}}$
(B.2) $\displaystyle\times(2\pi)^{4}\delta^{4}(p_{i}+p_{j}-p_{k}-p_{l})\
\bigg{[}f_{i}f_{j}(1-f_{k})(1-f_{l})-f_{k}f_{l}(1+f_{i})(1+f_{j})\bigg{]}$
$\displaystyle\times\sum_{\rm internal\,d.o.f}|\mathcal{M}|^{2}_{ij\to kl}\
\bigg{\\}}~{},$
where $f_{i,j,k,l}$ are the phase space distribution functions. The number
density, taking $f_{i}$ as example, is defined as
$\displaystyle n_{i}\equiv g_{i}\int\frac{d^{3}p}{(2\pi)^{3}}f_{i}(p)~{},$
(B.3)
in which $g_{i}$ is the internal degree of freedom (d.o.f.) of particle $i$.
The factor $N$ denotes the number of particles under consideration produced in
the final state and the factor $1/S$ originates from the phase space
suppression due to the identical particles in the initial and final states.
For
$HH^{\ast}\to{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\tilde{B}\tilde{B}}$
we have $N=2$ and $1/S=1/(N!)=1/2$. After some manipulations and neglecting
the negligible backward process, we have [58]
$\displaystyle\textbf{C}_{ij\to kl}$ $\displaystyle\approx$
$\displaystyle\frac{T}{32\pi^{4}}\int^{\infty}_{(m_{k}+m_{l})^{2}}ds\,p_{ij}\,W_{ij\to
kl}\,K_{1}({\sqrt{s}}/{T})$ (B.4) $\displaystyle W_{ij\to kl}$
$\displaystyle=$ $\displaystyle\frac{p_{kl}}{16\pi^{2}\sqrt{s}}\,\sum_{\rm
internal\,d.o.f}\int\,d\Omega\,|\mathcal{M}|^{2}_{ij\to kl}$ (B.5)
$\displaystyle p_{ij}$ $\displaystyle=$
$\displaystyle\frac{\sqrt{s-(m_{i}+m_{j})^{2}}\sqrt{s-(m_{i}-m_{j})^{2}}}{2\sqrt{s}}~{},$
(B.6)
where $p_{kl}$ is similar to $p_{ij}$. After summing over all bino spin states
$s_{1},s_{2}$ and isospin states of the SM-like Higgs , we have the amplitude
square ($s$ is the square of the central energy):
$\displaystyle\sum_{\rm internal\,d.o.f}\int d\Omega\
|\mathcal{M}|^{2}_{HH^{\ast}\to\tilde{B}\tilde{B}}~{},$
$\displaystyle\approx(2\pi)\times\bigg{[}\sum_{i,j=1}^{2}(\delta_{i}^{\,\,\,j})^{2}\bigg{]}\bigg{[}Y^{4}_{H}\bigg{]}\bigg{(}\frac{g_{1}g_{2}\sin\beta\cos\beta}{\mu}\bigg{)}^{2}\bigg{[}64\
s\bigg{(}1-\frac{4M_{1}^{2}}{s}\bigg{)}^{\frac{3}{2}}\bigg{]}$
$\displaystyle\approx(16\pi)\times\frac{g_{1}^{4}}{\mu^{2}}\sin^{2}\beta\cos^{2}\beta\
s~{}.$ (B.7)
We modify the MSSM model file available in FeynRules [59, 60] to highlight the
gauge state interactions and then export to FeynArts [61] augmented with
FeynCalc [62] to perform the calculation.
Since we are considering freeze-in production of $\tilde{B}$, $f_{1,2}$ in Eq.
(B.2) can be ignored. We can further approximate $f_{3,4}$ by Maxwell-
Boltzmann distribution, i.e. $f_{3,4}\approx e^{-E_{3,4}/T}$. Then the
collision term can be rewritten as [58, 63, 40]
$\displaystyle\textbf{C}_{HH^{\ast}\to\tilde{B}\tilde{B}}$
$\displaystyle\approx$
$\displaystyle\frac{T}{2048\pi^{6}}\int^{\infty}_{4M_{1}^{2}}ds\
(s-4M_{1}^{2})^{1/2}K_{1}({\sqrt{s}}/{T})\sum_{\rm
internal\,d.o.f}\int\,d\Omega\,|\mathcal{M}|^{2}_{HH^{\ast}\to\tilde{B}\tilde{B}}~{},$
(B.8) $\displaystyle\approx$
$\displaystyle\frac{T}{128\pi^{5}}\frac{g_{1}^{4}\,\sin^{2}\beta\cos^{2}\beta}{\mu^{2}}\int^{\infty}_{4M_{1}^{2}}ds\
s^{3/2}K_{1}({\sqrt{s}}/{T})~{}.$
Here $K_{1}$ is the Bessel function of the second kind, and we treat the SM-
like Higgs in the initial state as being massless. In the case where $M_{1}\ll
T$, the collision term can be approximated as (using
$\int^{\infty}_{0}dxx^{4}K_{1}(x)=16$)
$\displaystyle\int^{\infty}_{4M_{1}^{2}}ds\ s^{3/2}K_{1}({\sqrt{s}}/{T})$
$\displaystyle\approx$ $\displaystyle\int^{\infty}_{0}(dx\,T)\
(2xT)(xT)^{3}K_{1}(x)$ (B.9) $\displaystyle=$ $\displaystyle
2\,T^{5}\int^{\infty}_{0}dx\,x^{4}K_{1}(x)=32\ T^{5}~{}.$
## Appendix C The calculation details in Case II
We use $f=q,l$ with $q=u_{L},d_{L},u^{\dagger}_{R},d^{\dagger}_{R}$ and
$l=\nu,e_{L},e^{\dagger}_{R}$ to denote the left-handed two-component Weyl
spinor of SM quarks and leptons, where the bars are simply notations and do
not mean the Dirac conjugation. Hypercharges are given by
$\\{Y_{Q_{L}}=Y_{u_{L}}=Y_{d_{L}},\ Y_{{u^{\dagger}_{R}}},\
Y_{{d^{\dagger}_{R}}},\ Y_{L_{L}}=Y_{e_{L}}=Y_{\nu},\
Y_{{e^{\dagger}_{R}}}\\}=\\{1/6,\ -2/3,\ 1/3,\ -1/2,\ 1\\}$. After integrating
out sfermions with mass $M_{\tilde{f}}$ in the right panel of Fig.1, we obtain
dim-6 operators between SM fermion pair and $\tilde{B}$ pair:
$\displaystyle\mathcal{L}_{\text{eff}}=\sum_{f=q,l}\frac{(\sqrt{2}g_{1}Y_{f})(\sqrt{2}g_{1}Y_{f})}{M^{2}_{\tilde{f}}}(f^{\dagger}\tilde{B}^{\dagger})(f\tilde{B})~{},$
(C.1)
where for simplicity we consider an universal mass for all the fermions, i.e.
$M_{\tilde{f}}\equiv M_{\tilde{q}}=M_{\tilde{l}}$.
The amplitude squared terms in the collision term for
$f\bar{f}\to\tilde{B}\tilde{B}$ scattering process is given by666 Again,
fields in the initial and final states in the process should be understood in
the sense of physical particles, where $\bar{f}$ denotes the physical anti-
particle. Discussion on the naming convention of particles, states and filed
can be found in, e.g. [39].
$\displaystyle\sum_{\rm internal\,d.o.f}\int d\Omega\
|\mathcal{M}|^{2}_{f\bar{f}\to\tilde{B}\tilde{B}}$ $\displaystyle\approx$
$\displaystyle 2\pi N_{\rm flavor}\bigg{[}N_{\rm
color}\bigg{(}\sum_{i,j=1}^{2}(\delta_{i}^{\,\,\,j})^{2}Y^{4}_{Q_{L}}+Y_{{u^{\dagger}_{R}}}^{4}+Y_{{d^{\dagger}_{R}}}^{4}\bigg{)}$
(C.2)
$\displaystyle+\bigg{(}\sum_{i,j=1}^{2}(\delta_{i}^{\,\,\,j})^{2}Y^{4}_{L_{L}}+Y_{{e^{\dagger}_{R}}}^{4}\bigg{)}\bigg{]}\bigg{(}\frac{g_{1}^{2}}{M^{2}_{\tilde{f}}}\bigg{)}^{2}\bigg{[}\frac{16}{3}\
s^{2}\bigg{]}$ $\displaystyle=$
$\displaystyle\frac{1520\pi}{27}\frac{g_{1}^{4}}{M^{4}_{\tilde{f}}}\ s^{2}$
where $N_{\rm flavor}=N_{\rm color}=3$. As in Eq. (B.2), if we neglect bino
mass, then the collision term can be approximately given by (using
$\int^{\infty}_{0}dxx^{6}K_{1}(x)=384$)
$\displaystyle\textbf{C}_{f\bar{f}\to\tilde{B}\tilde{B}}$
$\displaystyle\approx$
$\displaystyle\frac{T}{2048\pi^{6}}\int^{\infty}_{4M_{1}^{2}}ds\
(s-4M_{1}^{2})^{1/2}K_{1}({\sqrt{s}}/{T})\sum_{\rm
internal\,d.o.f}\int\,d\Omega\,|\mathcal{M}|^{2}_{ff^{\dagger}\to\tilde{B}\tilde{B}^{\dagger}}$
(C.3) $\displaystyle\approx$
$\displaystyle\frac{T}{2048\pi^{6}}\bigg{(}\frac{1520\pi}{27}\frac{g_{1}^{4}}{M^{4}_{\tilde{f}}}\bigg{)}\int^{\infty}_{4M_{1}^{2}}ds\
s^{5/2}K_{1}({\sqrt{s}}/{T})$ $\displaystyle\approx$
$\displaystyle\frac{T}{2048\pi^{6}}\bigg{(}\frac{1520\pi}{27}\frac{g_{1}^{4}}{M^{4}_{\tilde{f}}}\bigg{)}\int^{\infty}_{0}(Tdx)(2Tx)\
(xT)^{5}K_{1}({x})$ $\displaystyle=$
$\displaystyle\frac{190}{9}g_{1}^{4}\frac{1}{\pi^{5}}\frac{1}{M^{4}_{\tilde{f}}}T^{8}~{}.$
## Appendix D The calculation details in Case III A
When neglecting all particle masses in the final state, we have
$\displaystyle\sum_{\rm internal\,d.o.f}\int d\Omega\
|\mathcal{M}|^{2}_{HH^{\ast}\to\tilde{B}\tilde{W}}$
$\displaystyle\approx(2\pi)\times\bigg{[}\sum_{b=1}^{3}\text{tr}\bigg{(}\frac{1}{2}\sigma^{b}\frac{1}{2}\sigma^{b}\bigg{)}\bigg{]}\bigg{[}Y^{2}_{H}\bigg{]}\bigg{(}\frac{g_{1}g_{2}\sin\beta\cos\beta}{\mu}\bigg{)}^{2}\bigg{[}64\
s\bigg{]}$
$\displaystyle=(48\pi)\times\frac{g_{1}^{2}g_{2}^{2}}{\mu^{2}}\sin^{2}\beta\cos^{2}\beta\
s~{},$ (D.1) $\displaystyle\sum_{\rm internal\,d.o.f}\int d\Omega\
|\mathcal{M}|^{2}_{\tilde{W}H\to\tilde{B}H}=\sum_{\rm internal\,d.o.f}\int
d\Omega\ |\mathcal{M}|^{2}_{\tilde{W}H^{\ast}\to\tilde{B}H^{\ast}}$
$\displaystyle\approx(2\pi)\times\bigg{[}\sum_{b=1}^{3}\text{tr}\bigg{(}\frac{1}{2}\sigma^{b}\frac{1}{2}\sigma^{b}\bigg{)}\bigg{]}\bigg{[}Y^{2}_{H}\bigg{]}\bigg{(}\frac{g_{1}g_{2}\sin\beta\cos\beta}{\mu}\bigg{)}^{2}\bigg{[}32\
s\bigg{]}$
$\displaystyle=(24\pi)\times\frac{g_{1}^{2}g_{2}^{2}}{\mu^{2}}\sin^{2}\beta\cos^{2}\beta\
s~{},$ (D.2) $\displaystyle\sum_{\rm internal\,d.o.f}\int d\Omega\
|\mathcal{M}|^{2}_{f\bar{f}\to\tilde{B}\tilde{W}}$ $\displaystyle\approx
2\pi\bigg{[}\sum_{b=1}^{3}\text{tr}\bigg{(}\frac{1}{2}\sigma^{b}\frac{1}{2}\sigma^{b}\bigg{)}\bigg{]}\bigg{[}N_{\rm
flavor}\bigg{(}Y^{2}_{L_{L}}+N_{\rm
color}Y^{2}_{Q_{L}}\bigg{)}\bigg{]}\bigg{(}\frac{g_{1}g_{2}}{M^{2}_{\tilde{f}}}\bigg{)}^{2}\bigg{[}\frac{16}{3}\
s^{2}\bigg{]}$
$\displaystyle=(16\pi)\times\frac{g_{1}^{2}g_{2}^{2}}{M^{4}_{\tilde{f}}}\
s^{2}~{},$ (D.3) $\displaystyle\sum_{\rm internal\,d.o.f}\int d\Omega\
|\mathcal{M}|^{2}_{\tilde{W}f\to\tilde{B}f}=\sum_{\rm internal\,d.o.f}\int
d\Omega\ |\mathcal{M}|^{2}_{\tilde{W}\bar{f}\to\tilde{B}\bar{f}}$
$\displaystyle\approx
2\pi\bigg{[}\sum_{b=1}^{3}\text{tr}\bigg{(}\frac{1}{2}\sigma^{b}\frac{1}{2}\sigma^{b}\bigg{)}\bigg{]}\bigg{[}N_{\rm
flavor}\bigg{(}Y^{2}_{L_{L}}+N_{\rm
color}Y^{2}_{Q_{L}}\bigg{)}\bigg{]}\bigg{(}\frac{g_{1}g_{2}}{M^{2}_{\tilde{f}}}\bigg{)}^{2}\bigg{[}\frac{32}{3}\
s^{2}\bigg{]}$
$\displaystyle=(32\pi)\times\frac{g_{1}^{2}g_{2}^{2}}{M^{4}_{\tilde{f}}}\
s^{2}~{},$ (D.4) $\displaystyle\sum_{\rm internal\,d.o.f}\int d\Omega\
|\mathcal{M}|^{2}_{f\bar{f}\to\tilde{B}\tilde{G}}$ $\displaystyle\approx
2\pi\bigg{[}\sum_{a=1}^{8}\text{tr}\bigg{(}\frac{1}{2}\lambda^{a}\frac{1}{2}\lambda^{a}\bigg{)}\bigg{]}\bigg{[}N_{\rm
flavor}\bigg{(}\sum_{i,j=1}^{2}(\delta_{i}^{\,\,\,j})^{2}Y^{2}_{Q_{L}}+Y_{{u^{\dagger}_{R}}}^{2}+Y_{{d^{\dagger}_{R}}}^{2}\bigg{)}\bigg{]}\bigg{(}\frac{g_{1}g_{3}}{M^{2}_{\tilde{f}}}\bigg{)}^{2}\bigg{[}\frac{16}{3}\
s^{2}\bigg{]}$
$\displaystyle=(\frac{704\pi}{9})\times\frac{g_{1}^{2}g_{3}^{2}}{M^{4}_{\tilde{f}}}\
s^{2}~{},$ (D.5) $\displaystyle\sum_{\rm internal\,d.o.f}\int d\Omega\
|\mathcal{M}|^{2}_{\tilde{G}f\to\tilde{B}f}=\sum_{\rm internal\,d.o.f}\int
d\Omega\ |\mathcal{M}|^{2}_{\tilde{G}\bar{f}\to\tilde{B}\bar{f}}$
$\displaystyle\approx
2\pi\bigg{[}\sum_{a=1}^{8}\text{tr}\bigg{(}\frac{1}{2}\lambda^{a}\frac{1}{2}\lambda^{a}\bigg{)}\bigg{]}\bigg{[}N_{\rm
flavor}\bigg{(}\sum_{i,j=1}^{2}(\delta_{i}^{\,\,\,j})^{2}Y^{2}_{Q_{L}}+Y_{{u^{\dagger}_{R}}}^{2}+Y_{{d^{\dagger}_{R}}}^{2}\bigg{)}\bigg{]}\bigg{(}\frac{g_{1}g_{3}}{M^{2}_{\tilde{f}}}\bigg{)}^{2}\bigg{[}\frac{32}{3}\
s^{2}\bigg{]}$
$\displaystyle=(\frac{1408\pi}{9})\times\frac{g_{1}^{2}g_{3}^{2}}{M^{4}_{\tilde{f}}}\
s^{2}~{}.$ (D.6)
## Appendix E The calculation details in Case III B
The $1\to 3$ decay processes are indicated by the red colored arrow in Fig.1.
When neglecting all particle masses in the final state, we have
$\displaystyle\Gamma_{\tilde{W}\to\tilde{B}HH^{\ast}}$ $\displaystyle=$
$\displaystyle\frac{1}{(2\pi)^{3}}\frac{1}{32M_{2}^{3}}\
\frac{1}{g_{\tilde{W}}}\sum_{\text{internal d.o.f.}}\int dm^{2}_{12}\
dm^{2}_{23}\ |\mathcal{M}|^{2}_{\tilde{W}\to\tilde{B}HH^{\ast}}$ (E.1)
$\displaystyle=$ $\displaystyle\frac{1}{(2\pi)^{3}}\frac{1}{32M_{2}^{3}}\
\frac{1}{\sum_{b=1}^{3}(2s_{\tilde{W}}+1)}\bigg{[}\sum_{b=1}^{3}\text{tr}\bigg{(}\frac{1}{2}\sigma^{b}\frac{1}{2}\sigma^{b}\bigg{)}\bigg{]}$
$\displaystyle\times\bigg{[}Y^{2}_{H}\bigg{]}\bigg{[}\frac{32}{3}M_{2}^{6}\left(\frac{g_{1}g_{2}\sin\beta\cos\beta}{\mu}\right)^{2}\bigg{]}$
$\displaystyle=$
$\displaystyle\frac{1}{384\pi^{3}}\left(\frac{g_{1}g_{2}\sin\beta\cos\beta}{\mu}\right)^{2}M_{2}^{3}~{}.$
$\displaystyle\Gamma_{\tilde{W}\to\tilde{B}f\bar{f}}$ $\displaystyle=$
$\displaystyle\frac{1}{(2\pi)^{3}}\frac{1}{32M_{2}^{3}}\
\frac{1}{g_{\tilde{W}}}\sum_{\text{internal d.o.f.}}\int dm^{2}_{12}\
dm^{2}_{23}\ |\mathcal{M}|^{2}_{\tilde{W}\to\tilde{B}f\bar{f}}$ (E.2)
$\displaystyle=$ $\displaystyle\frac{1}{(2\pi)^{3}}\frac{1}{32M_{2}^{3}}\
\frac{1}{\sum_{b=1}^{3}(2s_{\tilde{W}}+1)}\bigg{[}\sum_{b=1}^{3}\text{tr}\bigg{(}\frac{1}{2}\sigma^{b}\frac{1}{2}\sigma^{b}\bigg{)}\bigg{]}$
$\displaystyle\times\bigg{[}N_{\rm flavor}\bigg{(}Y^{2}_{L_{L}}+N_{\rm
color}Y^{2}_{Q_{L}}\bigg{)}\bigg{]}\bigg{[}\frac{2}{3}M_{2}^{8}\left(\frac{g_{1}g_{2}}{M^{2}_{\tilde{f}}}\right)^{2}\bigg{]}$
$\displaystyle=$
$\displaystyle\frac{1}{1536\pi^{3}}\left(\frac{g_{1}g_{2}}{M^{2}_{\tilde{f}}}\right)^{2}M_{2}^{5}~{},$
$\displaystyle\Gamma_{\tilde{G}\to\tilde{B}f\bar{f}}$ $\displaystyle=$
$\displaystyle\frac{1}{(2\pi)^{3}}\frac{1}{32M_{3}^{3}}\
\frac{1}{g_{\tilde{G}}}\sum_{\text{internal d.o.f.}}\int dm^{2}_{12}\
dm^{2}_{23}\ |\mathcal{M}|^{2}_{\tilde{G}\to\tilde{B}f\bar{f}}$ (E.3)
$\displaystyle=$ $\displaystyle\frac{1}{(2\pi)^{3}}\frac{1}{32M_{3}^{3}}\
\frac{1}{\sum_{a=1}^{8}(2s_{\tilde{G}}+1)}\bigg{[}\sum_{a=1}^{8}\text{tr}\bigg{(}\frac{1}{2}\lambda^{a}\frac{1}{2}\lambda^{a}\bigg{)}\bigg{]}$
$\displaystyle\times\bigg{[}N_{\rm
flavor}\bigg{(}N_{\text{iso},Q_{L}}Y^{2}_{Q_{L}}+Y_{{u^{\dagger}_{R}}}^{2}+Y_{{d^{\dagger}_{R}}}^{2}\bigg{)}\bigg{]}\bigg{[}\frac{2}{3}M_{2}^{8}\left(\frac{g_{1}g_{3}}{M^{2}_{\tilde{f}}}\right)^{2}\bigg{]}$
$\displaystyle=$
$\displaystyle\frac{11}{9216\pi^{3}}\left(\frac{g_{1}g_{3}}{M^{2}_{\tilde{f}}}\right)^{2}M_{2}^{5}~{}.$
where $dm^{2}_{12},dm^{2}_{23}$ are defined in [64].
## Appendix F The calculation details of 2-body decay after EWSB
As discussed in Section 4.1, we have the following $1\to 2$ decay possibly
affecting the cosmological BBN:
$\displaystyle\Gamma_{\tilde{\chi}^{0}_{2}\to\tilde{\chi}^{0}_{1}h}$
$\displaystyle=$
$\displaystyle\frac{1}{2s_{\tilde{\chi}^{0}_{2}}+1}\frac{1}{2M_{2}}\sum_{\text{spin
d.o.f.}}\int d\Pi_{2}\ |M|^{2}_{\tilde{\chi}^{0}_{2}\to\tilde{\chi}^{0}_{1}h}$
$\displaystyle=$ $\displaystyle\frac{1}{2}\frac{1}{2M_{2}}\ \int d\Pi_{2}\
\left(\frac{v}{\mu}g_{1}g_{2}\sin\beta\cos\beta\right)^{2}\bigg{[}4(p_{\tilde{\chi}^{0}_{2}}\cdot
p_{\tilde{\chi}^{0}_{1}}+M_{\tilde{\chi}^{0}_{2}}M_{\tilde{\chi}^{0}_{1}})\bigg{]}$
$\displaystyle\approx$ $\displaystyle\frac{1}{2}\frac{1}{2M_{2}}\ \bigg{[}\int
d\Omega\frac{1}{16\pi^{2}}\frac{|\vec{p}_{\tilde{\chi}^{0}_{1}}|}{M_{2}}\bigg{]}\
\left(\frac{v}{\mu}g_{1}g_{2}\sin\beta\cos\beta\right)^{2}\bigg{[}4(M_{2}E_{\tilde{\chi}^{0}_{1}}+M_{2}M_{1})\bigg{]}~{},$
where
$\displaystyle E_{\tilde{\chi}^{0}_{1}}$ $\displaystyle=$
$\displaystyle\frac{M_{2}^{2}+M_{1}^{2}-M_{h}^{2}}{2M_{2}}\approx\frac{M_{2}^{2}+M_{1}^{2}}{2M_{2}}~{},$
(F.2) $\displaystyle|\vec{p}_{\tilde{\chi}^{0}_{1}}|$ $\displaystyle=$
$\displaystyle\sqrt{E_{\tilde{\chi}^{0}_{1}}^{2}-M_{1}^{2}}=\frac{\left(M_{2}^{4}+M_{1}^{4}+M_{h}^{4}-2M_{2}^{2}M_{1}^{2}-2M_{2}^{2}M_{h}^{2}-2M_{1}^{2}M_{h}^{2}\right)^{\frac{1}{2}}}{2M_{2}}$
(F.3) $\displaystyle\approx$
$\displaystyle\frac{M_{2}^{2}-M_{1}^{2}}{2M_{2}}~{}.$
Finally, we have [44]
$\displaystyle\Gamma_{\tilde{\chi}^{0}_{2}\to\tilde{\chi}^{0}_{1}h}$
$\displaystyle\approx$
$\displaystyle\frac{1}{2}\frac{1}{2M_{2}}\Bigg{[}4\pi\frac{1}{16\pi^{2}}\frac{1}{2}\left(1-\frac{M_{1}^{2}}{M_{2}^{2}}\right)\bigg{]}\left(\frac{v}{\mu}g_{1}g_{2}\sin\beta\cos\beta\right)^{2}\bigg{[}4M_{2}\frac{(M_{2}+M_{1})^{2}}{2M_{2}}\bigg{]}$
(F.4) $\displaystyle\approx$ $\displaystyle
M_{2}\frac{1}{16\pi}\left(\frac{v}{\mu}g_{1}g_{2}\sin\beta\cos\beta\right)^{2}\left(1-\frac{M_{1}^{2}}{M_{2}^{2}}\right)\left(1+\frac{M_{1}}{M_{2}}\right)^{2}~{}.$
Using the GET we would obtain the same results in the high energy limit for
$\Gamma_{\tilde{\chi}^{\pm}_{1}\to\tilde{\chi}^{0}_{1}W^{\pm}}$.
## References
* [1] Y. A. Golfand and E. P. Likhtman, _Extension of the Algebra of Poincare Group Generators and Violation of p Invariance_ , _JETP Lett._ 13 (1971) 323.
* [2] D. V. Volkov and V. P. Akulov, _Is the neutrino a goldstone particle?_ , _Physics Letters B_ 46 (1973) 109.
* [3] J. Wess and B. Zumino, _Supergauge transformations in four dimensions_ , _Nuclear Physics B_ 70 (1974) 39.
* [4] A. Salam and J. Strathdee, _Super-symmetry and non-Abelian gauges_ , _Physics Letters B_ 51 (1974) 353.
* [5] J. Wess and B. Zumino, _Supergauge invariant extension of quantum electrodynamics_ , _Nuclear Physics, Section B_ 78 (1974) 1.
* [6] S. Ferrara and B. Zumino, _Supergauge invariant Yang-Mills theories_ , _Nuclear Physics, Section B_ 79 (1974) 413.
* [7] G. Steigman and M. S. Turner, _Cosmological constraints on the properties of weakly interacting massive particles_ , _Nuclear Physics B_ 253 (1985) 375.
* [8] G. Jungman, M. Kamionkowski and K. Griest, _Supersymmetric Dark Matter_ , _Physics Reports_ 267 (1995) 195 [9506380].
* [9] S. P. MARTIN, _A SUPERSYMMETRY PRIMER_ , pp. 1–98. jul, 1998. 9709356. DOI.
* [10] J. L. Feng, _Dark Matter Candidates from Particle Physics and Methods of Detection_ , _Annual Review of Astronomy and Astrophysics_ 48 (2010) 495 [1003.0904].
* [11] J. Cao, C. Han, L. Wu, J. M. Yang and Y. Zhang, _Probing Natural SUSY from Stop Pair Production at the LHC_ , _JHEP_ 11 (2012) 039 [1206.3865].
* [12] J. Cao, F. Ding, C. Han, J. M. Yang and J. Zhu, _A light Higgs scalar in the NMSSM confronted with the latest LHC Higgs data_ , _JHEP_ 11 (2013) 018 [1309.4939].
* [13] C. Han, A. Kobakhidze, N. Liu, A. Saavedra, L. Wu and J. M. Yang, _Probing Light Higgsinos in Natural SUSY from Monojet Signals at the LHC_ , _JHEP_ 02 (2014) 049 [1310.4274].
* [14] C. Han, K.-i. Hikasa, L. Wu, J. M. Yang and Y. Zhang, _Current experimental bounds on stop mass in natural SUSY_ , _JHEP_ 10 (2013) 216 [1308.5307].
* [15] C. Han, D. Kim, S. Munir and M. Park, _Accessing the core of naturalness, nearly degenerate higgsinos, at the LHC_ , _JHEP_ 04 (2015) 132 [1502.03734].
* [16] C. Han, J. Ren, L. Wu, J. M. Yang and M. Zhang, _Top-squark in natural SUSY under current LHC run-2 data_ , _Eur. Phys. J. C_ 77 (2017) 93 [1609.02361].
* [17] C. Han, K.-i. Hikasa, L. Wu, J. M. Yang and Y. Zhang, _Status of CMSSM in light of current LHC Run-2 and LUX data_ , _Phys. Lett. B_ 769 (2017) 470 [1612.02296].
* [18] ATLAS collaboration, _SUSY March 2023 Summary Plot Update,https://cds.cern.ch/record/2852738_, .
* [19] CMS collaboration, _CMS SUS Physics Results,https://twiki.cern.ch/twiki/bin/view/CMSPublic/PhysicsResultsSUS_, .
* [20] H. Baer, V. Barger, S. Salam, D. Sengupta and K. Sinha, _Status of weak scale supersymmetry after LHC Run 2 and ton-scale noble liquid WIMP searches_ , _Eur. Phys. J. ST_ 229 (2020) 3085 [2002.03013].
* [21] F. Wang, W. Wang, J. Yang, Y. Zhang and B. Zhu, _Low Energy Supersymmetry Confronted with Current Experiments: An Overview_ , _Universe_ 8 (2022) 178 [2201.00156].
* [22] J. M. Yang, P. Zhu and R. Zhu, _A brief survey of low energy supersymmetry under current experiments_ , _PoS_ LHCP2022 (2022) 069 [2211.06686].
* [23] G. Giudice and A. Romanino, _Split supersymmetry_ , _Nuclear Physics B_ 699 (2004) 65 [0406088].
* [24] N. Arkani-Hamed, S. Dimopoulos, G. Giudice and A. Romanino, _Aspects of Split Supersymmetry_ , _Nuclear Physics B_ 709 (2005) 3 [0409232].
* [25] N. Arkani-Hamed and S. Dimopoulos, _Supersymmetric unification without low energy supersymmetry and signatures for fine-tuning at the LHC_ , _Journal of High Energy Physics_ 2005 (2005) 073 [0405159].
* [26] J. D. Wells, _PeV-scale supersymmetry_ , _Physical Review D_ 71 (2005) 015013 [0411041].
* [27] L. J. Hall and Y. Nomura, _A Finely-Predicted Higgs Boson Mass from A Finely-Tuned Weak Scale_ , _JHEP_ 03 (2010) 076 [0910.2235].
* [28] G. F. Giudice and A. Strumia, _Probing High-Scale and Split Supersymmetry with Higgs Mass Measurements_ , _Nucl. Phys. B_ 858 (2012) 63 [1108.6077].
* [29] L. E. Ibanez and I. Valenzuela, _The Higgs Mass as a Signature of Heavy SUSY_ , _JHEP_ 05 (2013) 064 [1301.5167].
* [30] L. J. Hall, Y. Nomura and S. Shirai, _Grand Unification, Axion, and Inflation in Intermediate Scale Supersymmetry_ , _JHEP_ 06 (2014) 137 [1403.8138].
* [31] S. A. R. Ellis and J. D. Wells, _High-scale supersymmetry, the Higgs boson mass, and gauge unification_ , _Phys. Rev. D_ 96 (2017) 055024 [1706.00013].
* [32] MAGIC collaboration, H. Abe et al., _Search for Gamma-Ray Spectral Lines from Dark Matter Annihilation up to 100 TeV toward the Galactic Center with MAGIC_ , _Phys. Rev. Lett._ 130 (2023) 061002 [2212.10527].
* [33] E. W. Kolb and M. S. Turner, _The Early Universe_ , vol. 69. 1990, 10.1201/9780429492860.
* [34] K. A. Olive and M. Srednicki, _New Limits on Parameters of the Supersymmetric Standard Model from Cosmology_ , _Phys. Lett. B_ 230 (1989) 78.
* [35] K. Griest, M. Kamionkowski and M. S. Turner, _Supersymmetric Dark Matter Above the W Mass_ , _Phys. Rev. D_ 41 (1990) 3565.
* [36] L. J. Hall, K. Jedamzik, J. March-Russell and S. M. West, _Freeze-In Production of FIMP Dark Matter_ , _JHEP_ 03 (2010) 080 [0911.1120].
* [37] ATLAS collaboration, G. Aad et al., _Observation of a new particle in the search for the Standard Model Higgs boson with the ATLAS detector at the LHC_ , _Phys. Lett. B_ 716 (2012) 1 [1207.7214].
* [38] CMS collaboration, S. Chatrchyan et al., _Observation of a New Boson at a Mass of 125 GeV with the CMS Experiment at the LHC_ , _Phys. Lett. B_ 716 (2012) 30 [1207.7235].
* [39] H. K. Dreiner, H. E. Haber and S. P. Martin, _Two-component spinor techniques and Feynman rules for quantum field theory and supersymmetry_ , _Phys. Rept._ 494 (2010) 1 [0812.1594].
* [40] F. Elahi, C. Kolda and J. Unwin, _UltraViolet Freeze-in_ , _JHEP_ 03 (2015) 048 [1410.6157].
* [41] M. Beneke, R. Szafron and K. Urban, _Sommerfeld-corrected relic abundance of wino dark matter with NLO electroweak potentials_ , _JHEP_ 02 (2021) 020 [2009.00640].
* [42] J. Ikemoto, N. Haba, S. Yasuhiro and T. Yamada, _Higgs Portal Majorana Fermionic Dark Matter with the Freeze-in Mechanism_ , 2212.14660.
* [43] A. Arvanitaki, C. Davis, P. W. Graham, A. Pierce and J. G. Wacker, _Limits on split supersymmetry from gluino cosmology_ , _Phys. Rev. D_ 72 (2005) 075011 [hep-ph/0504210].
* [44] J. F. Gunion and H. E. Haber, _Two-body Decays of Neutralinos and Charginos_ , _Phys. Rev. D_ 37 (1988) 2515.
* [45] J. F. Gunion and H. E. Haber, _Errata for Higgs bosons in supersymmetric models: 1, 2 and 3_ , hep-ph/9301205.
* [46] A. Djouadi, Y. Mambrini and M. Muhlleitner, _Chargino and neutralino decays revisited_ , _Eur. Phys. J. C_ 20 (2001) 563 [hep-ph/0104115].
* [47] Y. Yamada, _Electroweak two-loop contribution to the mass splitting within a new heavy SU(2)(L) fermion multiplet_ , _Phys. Lett. B_ 682 (2010) 435 [0906.5207].
* [48] M. Ibe, S. Matsumoto and R. Sato, _Mass Splitting between Charged and Neutral Winos at Two-Loop Level_ , _Phys. Lett. B_ 721 (2013) 252 [1212.5989].
* [49] J. McKay and P. Scott, _Two-loop mass splittings in electroweak multiplets: winos and minimal dark matter_ , _Phys. Rev. D_ 97 (2018) 055049 [1712.00968].
* [50] M. Ibe, M. Mishima, Y. Nakayama and S. Shirai, _Precise estimate of charged Wino decay rate_ , _JHEP_ 01 (2023) 017 [2210.16035].
* [51] M. Kawasaki, K. Kohri, T. Moroi and A. Yotsuyanagi, _Big-Bang Nucleosynthesis and Gravitino_ , _Phys. Rev. D_ 78 (2008) 065011 [0804.3745].
* [52] M. Kawasaki, K. Kohri, T. Moroi and Y. Takaesu, _Revisiting Big-Bang Nucleosynthesis Constraints on Long-Lived Decaying Particles_ , _Phys. Rev. D_ 97 (2018) 023502 [1709.01211].
* [53] K. Rolbiecki and K. Sakurai, _Long-lived bino and wino in supersymmetry with heavy scalars and higgsinos_ , _JHEP_ 11 (2015) 091 [1506.08799].
* [54] ATLAS collaboration, G. Aad et al., _Search for neutral long-lived particles in $pp$ collisions at $\sqrt{s}$ = 13 TeV that decay into displaced hadronic jets in the ATLAS calorimeter_, _JHEP_ 06 (2022) 005 [2203.01009].
* [55] CMS collaboration, A. Tumasyan et al., _Search for long-lived particles decaying to a pair of muons in proton-proton collisions at $\sqrt{s}$ = 13 TeV_, _JHEP_ 05 (2023) 228 [2205.08582].
* [56] ATLAS collaboration, G. Aad et al., _Search for long-lived charginos based on a disappearing-track signature using 136 fb -1 of pp collisions at $\sqrt{s}$ = 13 TeV with the ATLAS detector_, _Eur. Phys. J. C_ 82 (2022) 606 [2201.02472].
* [57] CMS collaboration, A. M. Sirunyan et al., _Search for disappearing tracks in proton-proton collisions at $\sqrt{s}=$ 13 TeV_, _Phys. Lett. B_ 806 (2020) 135502 [2004.05153].
* [58] J. Edsjo and P. Gondolo, _Neutralino relic density including coannihilations_ , _Phys. Rev. D_ 56 (1997) 1879 [hep-ph/9704361].
* [59] N. D. Christensen and C. Duhr, _FeynRules - Feynman rules made easy_ , _Comput. Phys. Commun._ 180 (2009) 1614 [0806.4194].
* [60] A. Alloul, N. D. Christensen, C. Degrande, C. Duhr and B. Fuks, _FeynRules 2.0 - A complete toolbox for tree-level phenomenology_ , _Comput. Phys. Commun._ 185 (2014) 2250 [1310.1921].
* [61] T. Hahn, _Generating Feynman diagrams and amplitudes with FeynArts 3_ , _Comput. Phys. Commun._ 140 (2001) 418 [hep-ph/0012260].
* [62] V. Shtabovenko, R. Mertig and F. Orellana, _New Developments in FeynCalc 9.0_ , _Comput. Phys. Commun._ 207 (2016) 432 [1601.01167].
* [63] P. Gondolo and G. Gelmini, _Cosmic abundances of stable particles: Improved analysis_ , _Nucl. Phys. B_ 360 (1991) 145.
* [64] Particle Data Group collaboration, R. L. Workman et al., _Review of Particle Physics_ , _PTEP_ 2022 (2022) 083C01.
|
# A Framework using Contrastive Learning for Classification with Noisy Labels
Madalina Ciortan§
EURA NOVA BE
Mont-Saint-Guibert, Belgium
<EMAIL_ADDRESS>
&Romain Dupuis§
EURA NOVA BE
Mont-Saint-Guibert, Belgium
<EMAIL_ADDRESS>
&Thomas Peel
EURA NOVA BE
Mont-Saint-Guibert, Belgium
<EMAIL_ADDRESS>
###### Abstract
We propose a framework using contrastive learning as a pre-training task to
perform image classification in the presence of noisy labels. Recent
strategies such as pseudo-labelling, sample selection with Gaussian Mixture
models, weighted supervised contrastive learning have been combined into a
fine-tuning phase following the pre-training. This paper provides an extensive
empirical study showing that a preliminary contrastive learning step brings a
significant gain in performance when using different loss functions: non
robust, robust, and early-learning regularized. Our experiments performed on
standard benchmarks and real-world datasets demonstrate that: i) the
contrastive pre-training increases the robustness of any loss function to
noisy labels and ii) the additional fine-tuning phase can further improve
accuracy, but at the cost of additional complexity.
§§footnotetext: Equal contribution
## 1 Introduction
Collecting large and well-annotated datasets for image classification tasks
represents a challenge as human quality annotations are expensive and time-
consuming. Alternative methods exist, such as web crawlers [27]. Nevertheless,
these methods generate noisy labels decreasing the performance of deep neural
networks. They tend to overfit to noisy labels due to their high capacity
[44]. That is why developing efficient noisy-label learning (NLL) techniques
is of great importance.
Various strategies have been proposed to deal with NLL: i) Noise transition
matrix [33, 9, 41] estimates the noise probability and corrects the loss
function, ii) a small and clean subset can help to avoid overfitting [14],
iii) samples selection identifies true-labeled samples [15, 10, 22], and iv)
robust loss functions solve the classification problem only by adapting the
loss function to be less sensitive to noisy labels [47, 37, 26]. Methods also
combine other strategies (eg. ELR+ [22], DivideMix [25]): two networks, semi-
supervised learning, label correction, or mixup. They show the most promising
results but lead to a large number of hyperparameters. That is why we explore
improvement strategies for robust loss functions. They are simpler to
integrate and faster to train, but as illustrated in Figure 1, they tend to
overfit and have lower performance for high noise ratios.
Figure 1: Top-1 test accuracy for a ResNet18 trained on the CIFAR-100 dataset
with a symmetric noise of 80% for three losses: Cross Entropy (CE), Normalized
Focal Loss + Reverse Cross Entropy (NFL+RCE), and Early Learning
Regularization (ELR).
Meanwhile, new self-supervised learning algorithms for image representations
have been recently developed [5, 12]. Such algorithms extract representation
(or features) in unsupervised settings. These representations can then be used
for downstream tasks such as classification. Methods based on contrastive
learning compete with fully supervised learning while fine-tuning only on a
small fraction of all available labels. Therefore, using contrastive learning
for NLL appears as promising. In this work, contrastive learning aims to pre-
train the classifier to improve its robustness.
The key contributions of this work are:
* •
A framework increasing robustness of any loss function to noisy labels by
adding a contrastive pre-training task.
* •
The adaptation of the supervised contrastive loss to use sample weight values,
representing the probability of correctness for each sample in the training
set
* •
An extensive empirical study identifying and benchmarking additional state of
the art strategies to boost the performance of pre-trained models: pseudo-
labeling, sample selection with GMM, weighted supervised contrastive learning,
and mixup with bootstrapping.
## 2 Related works
Existing approaches dealing with NLL and contrastive learning in computer
vision are briefly reviewed. Extra details can be found in Song et al. [36],
Le-Khac et al. [21].
### 2.1 Noise tolerant classification
Sample Selection: This method identifies noisy and clean samples within the
training data. Several strategies leverage the interactions between multiple
networks to identify the probably correct labels [10, 15, 22]. Recent works
[1, 35] exploit the small loss trick to identify clean and noisy samples by
considering a certain number of small-loss training samples as true-labeled
samples. This approach can be justified by the memorization effect: deep
neural networks first fit the training data with clean labels during a so-
called early learning phase, before overfitting the noisy samples during the
memorization phase [2, 25].
Robust Loss Function: Commonly used loss functions, such as Cross Entropy (CE)
or Focal Loss, are not robust to noisy labels. Therefore, new loss functions
have been designed. Such robust loss functions can be easily incorporated into
existing pipelines to improve performance regarding noisy labels. The
symmetric cross entropy [37] has been proposed by adding a reverse CE loss to
the initial CE. This combination improves the accuracy of the model compared
to classical loss functions. Ma et al. [26] show theoretically that
normalization can convert classical loss functions into loss functions robust
to noise labels. The combination of two robust loss functions can also improve
robustness. However, the performance of normalized loss functions remains
quite low for high noise rates as illustrated in Figure 1.
Semi-supervised: Semi-supervised approaches deal with both labeled and
unlabeled data. Recent works [30, 22, 38] combine sample selection with semi-
supervised methods: the possibly noisy samples are treated as unlabeled and
the possibly clean samples are treated as labeled. Such approaches leverage
information contained in noisy data, for instance by using MixMatch [3]. Semi-
supervised approaches show competitive results. However, they use several
hyperparameters that can be sensitive to changes in data or noise type [36,
31].
Contrastive learning: recent developments in self-supervised and contrastive
learning [46, 31, 23] inspire new approaches in NLL. Li et al. [23] employed
features learned by contrastive learning to detect out-of-distribution
samples.
### 2.2 Contrastive learning for vision data
Contrastive learning extracts features by comparing each data sample with
different samples. The central idea is to bring different instances of the
same input image closer and spread instances from different images apart. The
inputs are usually divided into positive (similar inputs) and negative pairs
(dissimilar inputs). Frameworks have been recently developed, such as CPCv2
[13], SimCLR [5], Moco [12]. Once the self-supervised model is trained, the
extracted representations can be used for downstream tasks.In this work, the
representations are used for noisy label classification.
Chen et al. [5] demonstrate that large sets of negatives (and large batches)
are crucial in learning good representations. However, large batches are
limited by GPU memory. Maintaining a memory bank accumulating a large number
of negative representations is an elegant solution decoupling the batch size
from the number of negatives [28]. Nevertheless, the representations get
outdated in a few iterations. The Momentum Encoder [12] addresses the issues
by generating a dynamic memory queue of representations. Other strategies aim
at getting more meaningful negative samples to reduce the memory/batch size
[16].
## 3 Preliminaries
Let
$D=\\{(\bm{x_{i}},\overline{y_{i}})\\}_{i=1..n},\bm{x_{i}}\in\mathbb{R}^{d},\overline{y_{i}}\in\\{1,\cdots,K\\}$
denote a noisy input dataset with an unknown number of samples incorrectly
labelled. The associated true and unobservable labels are written $y_{i}$. The
images $\bm{x_{i}}$ are of size $d$ and the classification problem has $K$
classes. The goal is to train a deep neural network (DNN) $f$. Using a robust
loss function for training consists of minimizing the empirical risk defined
by robust loss functions in order to find the set of optimal parameters
$\theta$. The one-hot encoding of the label is denoted by the distribution
$q(k|\bm{x})$ for a sample $\bm{x}$ and a class $k$, such as
$q(y_{i}|\bm{x_{i}})=1$ and $q(k\neq y_{i}|\bm{x_{i}})=0,\>\forall
i\in\\{1,\cdots,n\\}$. The probability vector of $f$ is defined by the softmax
function $p(k|\bm{x})=\frac{e^{z_{k}}}{\sum_{j=1}^{K}e^{z_{j}}}$ where $z_{k}$
denotes the logits output with respect to class $k$.
### 3.1 Classification with robust loss functions
The method employs noise-robust losses to train the classifier in the presence
of noisy labels. Such losses improve the classification accuracy compared to
the commonly used Cross Entropy (CE), as illustrated in Figure 1. In this
section, the general empirical risk for a given mini-batch is defined by
$L=\sum_{i=1}^{N}\mathcal{L}(f(x_{i}),\overline{y_{i}})=\sum_{i=1}^{N}l_{i}$ .
The term $l_{i}$ is modified by each loss function.
The classical CE is used as a baseline loss function not robust to noisy
labels [8] and is defined as:
$l_{ce}=-\sum_{k=1}^{K}q(k|\bm{x_{i}})log(p(k|\bm{x_{i}})).$ (1)
As presented in section 2, Ma et al. [26] introduce robust loss functions
called Active Passive Losses that do not suffer from underfitting. We
investigate the combination between the Normalized Focal Loss (NFL) and the
Reversed Cross Entropy (RCE) called NFL+RCE. It shows promising results on
various benchmarks. The NFL is defined as:
$l_{nfl}=\frac{-\sum\limits_{k=1}^{K}q(k|\bm{x_{i}})(1-p(k|\bm{x_{i}}))^{\gamma}log(p(k|\bm{x_{i}}))}{-\sum\limits_{j=1}^{K}\sum\limits_{k=1}^{K}q(y=j|\bm{x_{i}})(1-p(k|\bm{x_{i}}))^{\gamma}log(p(k|\bm{x_{i}}))},$
(2)
where $\gamma\geq 0$ is an hyperparameter. The RCE loss is:
$l_{rce}=-\sum_{k=1}^{K}p(k|\bm{x_{i}})log\left(q(k|\bm{x_{i}})\right).$ (3)
The final combination following the framework simply gives a different
$\alpha$ and $\beta$ to each loss:
$l_{nfl+rce}=\alpha.l_{nfl}+\beta.l_{rce}.$ (4)
The two hyperparameters $\alpha$ and $\beta$ control the balancing between
more active learning and less passive learning. For simplicity, $\alpha$ and
$\beta$ are set to 1.0 without any tuning.
Liu et al. [25] propose another framework to deal with noisy annotations based
on the “early learning” phase. The loss, called Early Learning Regularization
(ELR), adds a regularization term to capitalize on early learning. ELR is not
strictly speaking a robust loss but belongs to robust penalization and label
correction methods. The penalization term corrects the CE based on estimated
soft labels identified with semi-supervised learning techniques. It prevents
memorization of false labels by steering the model towards these targets. The
regularization term maximizes the inner product between model outputs and
targets:
$l_{elr}=l_{ce}+\frac{\lambda_{elr}}{N}log\left(1-\sum_{k=1}^{K}p(k|\bm{x_{i}})t(k|\bm{x_{i}})\right).$
(5)
The target is not set equal to the model output but is estimated with a
temporal ensembling from semi-supervised methods. Let $t(k|\bm{x_{i}})^{(l)}$
denote the target for example $\bm{x_{i}}$ at iteration $l$ of training with a
momentum $\beta$:
$t(k|\bm{x_{i}})^{(l)}=\beta
t(k|\bm{x_{i}})^{(l-1)}+(1-\beta)p(k|\bm{x_{i}})^{(l)}.$ (6)
### 3.2 Contrastive learning
Contrastive learning methods learn representations by contrasting positive and
negative examples. A typical framework is composed of several blocks [7]:
* •
Data augmentation: Data augmentation is used to decouple the pretext tasks
from the network architecture. Chen et al. [5] study broadly the impact of
data augmentation. We follow their suggestion combining random crop (and
flip), color distortion, Gaussian blur, and gray-scaling.
* •
Encoding: The encoder extracts features (or representation) from augmented
data samples. A classical choice for the encoder is the ResNet model [11] for
image data. The final goal of the contrastive approach is to find correct
weights for the encoder.
* •
Loss function: The loss function usually combines positive and negative pairs.
The Noise Contrastive Estimation (NCE) and its variants are popular choices.
The general formulation for such loss function is defined for the i-th pair as
[40]:
$L_{i}=-log\frac{exp(\bm{z_{i}}^{T}\bm{z_{j(i)}}/\tau)}{\sum_{a\in
A(i)}exp(\bm{z_{i}}^{T}\bm{z_{a}}/\tau)},\;\text{with}\;i\in I,$ (7)
where $\bm{z}$ is a feature vector, $I$ is the set of indexes in the mini-
batch, $i$ is the index of the anchor, $j(i)$ is the index of an augmented
version of the anchor source image, $A(i)=I\setminus\\{i\\}$, and $\tau$ is a
temperature controlling the dot product. The denominator includes one positive
and $K$ negative pairs.
* •
Projection head: That step is not used in all frameworks. The projection head
maps the representation to a lower-dimensional space and acts as an
intermediate layer between the representation and the embedding pairs. Chen et
al. [5, 6] show that the projection head helps to improve the representation
quality.
## 4 A framework coupling contrastive learning and noisy labels
As illustrated in Figure 2, our method classifies noisy samples in a two
phased process. First, a classifier pre-trained with contrastive learning
produces train-set pseudo-labels (pre-training phase, in panel a), used during
the training of a subsequent fine-tuning phase (panel b). The underlying
intuition is that the predicted pseudo-lables are more accurate than the
original noisy labels. The contrastive learning performed in the first phase
(panel a1) improves the performance the classifier (panel a2), sensitive to
noisy labels; the resulting model can be also used in a standalone way with a
reduced number of hyperparameters, without the underlying fine-tuning phase.
The second phase leverages the pseudo-labels predicted by the pre-training in
all underlying steps (b1-b3). To mitigate the effect of potentially
incorrectly predicted pseudo-labels, a Gaussian Mixture Model (GMM, panel b1)
with 2 components follows the small loss-trick to predict for each sample the
probability of correctness. This value is used as a weight in a supervised
contrastive step (panel b2), performed to improve the learned representations
by taking advantage of the label information. A classification head is added
to the contrastive model in order to produce the final predictions (panel b3).
The fine-tuning phase can be seen as an adaptation of the pre-training phase
to handle pseudo-labels.
Figure 2: Overview of the framework consisting of two phases: pre-training
(panel a) and fine-tuning (panel b). After a contrastive learning phase (a1) a
classifier (a2) is trained to predict train-set pseudo-labels $\widehat{y}$.
The fine-tuning phase uses $\widehat{y}$ as a new ground truth. First, a GMM
model (b1) predicts the probability of correctness for each sample, used as a
corrective weight factor in a supervised contrastive training (panel b2). The
final predictions $\widehat{y}_{final}$ are produced by the (b3) classifier.
To maximize the impact of the contrastive learning on the underlying
classification, the supervised training is performed in 2 steps: a warm-up
step, updating only the classifier layer (while keeping the encoder frozen) is
followed by the full model training. We compared three different loss
functions for the supervised classification: the classical CE, the robust
NFL+RCE, and the ELR loss.
### 4.1 Sample selection and correction with pseudo-labels
Pseudo labels represent one hot encoded model’s predictions on the training
set. Pseudo-labels were initially used in semi-supervised learning to produce
annotations for unlabelled data; in the noisy label setting, various
techniques (e.g. DivideMix, etc) identify a subset with a high likelihood of
correctness and treat the remaining samples as the unlabeled counterpart in
semi-supervised learning. In this work, we elaborate on the observation that
the training set labels, predicted after training the model with a noise-
robust loss function (i.e. the pseudo labels), are more accurate than the
ground truth. This observation is supported by the results in Figure 3,
depicting the accuracy of pseudo labels predicted on CIFAR100, contaminated
with various levels of asymmetric (panel a) and symmetric (panel b) noise. The
pseudo labels are more accurate than the corrupted ground truth in both
settings and bring a higher gain in performance as the noise ratio increases.
Figure 3: Accuracy of pseudo labels on all simulated settings with asymmetric
(a) and symmetric (b) noise, evaluated on CIFAR100. The correctness of the
ground truth is represented on the x axis, while the accuracy of predicted
pseudo labels on the y axis. In all experiments, the pseudo labels have a
higher accuracy than the corrupted ground truth and this gain increases with
the noise ratio
As proposed by other approaches [1], the loss value on train samples can be
used to discriminate between clean and mislabeled samples. The sample
correctness probability is computed by fitting a 2 components GMM on the
distribution of losses [22]. The underlying probability is used as a sample
weight:
$w_{i}=p(k=0|l_{i}),$ (8)
where $l_{i}$ is the loss for sample $i$ and $k=0$ is the GMM component
associated to the clean samples (lowest loss). Figure 4 depicts the evolution
of the clean training set identified by GMM on an example: its accuracy grows
from 0.6 to 0.93 while the size stabilizes at 60% of the training set.
Figure 4: Accuracy of the entire training set (in blue) compared to the clean
train subset (in red); The clean subset’s percentual size is depicted in
green. The example is performed on CIFAR100, with 40% symmetric noise.
### 4.2 Weighted supervised contrastive learning
A modification to the contrastive loss defined in Equation 7 has been proposed
to leverage label information [18]:
$L_{i}=-log\frac{1}{|P(i)|}\sum_{p\in
P(i)}\frac{exp(\bm{z_{i}}^{T}\bm{z_{p}}/\tau)}{\sum_{a\in
A(i)}exp(\bm{z_{i}}^{T}\bm{z_{a}}/\tau)},$ (9)
where $P(i)=\\{j\in I\setminus\\{i\\},y_{j}=\widetilde{y_{i}}\\}$ with
$\widetilde{y_{i}}$ the prediction of the model for input $\bm{x_{i}}$.
As explained in the previous section, the loss value for the training set
samples is used to fit a GMM with 2 components, corresponding to correctly and
incorrectly labeled samples. We adapted the supervised representation loss to
employ $w$, a weighting factor representing the sample probability of
membership to the correctly labeled component. Thus, likely mislabeled samples
having large loss values would contribute only marginally to the supervised
representations:
$L_{i}=-log\frac{1}{|P(i)|}\sum_{p\in
P(i)}\frac{\widetilde{w_{p,i}}exp(\bm{z_{i}}^{T}\bm{z_{p}}/\tau)}{\sum_{a\in
A(i)}exp(\bm{z_{i}}^{T}\bm{z_{p}}/\tau)},$ (10)
where $\widetilde{w_{p,i}}$ is a modified version of $w_{p}$ such as
$\widetilde{w_{p,i}}=1$ if $p=j(i)$ else $\widetilde{w_{p,i}}=w_{i}$. If all
samples are considered as noisy, Equation 10 is simplified into the classical
unsupervised contrastive loss in Equation 7.
## 5 Experiments
The framework is assessed on three benchmarks and the contribution of each
block identified in Figure 2 is analyzed.
### 5.1 Datasets
CIFAR10 and CIFAR100 [20]. These experiments assess the accuracy of the method
against synthetic label noise. The two datasets are contaminated with
simulated symmetric or asymmetric label noise reproducing the heuristic in Ma
et al. [26]. The symmetric noise consists in corrupting an equal arbitrary
ratio of labels for each class. The noise level varies from $0.2$ to $0.8$.
For asymmetric noise [33, 25], sample labels have been flipped within a
specific set of classes, thus providing confusion between predetermined pairs
of labels. For CIFAR100, 20 groups of super-classes have been created, each
consisting of 5 sub-classes. The label flipping is performed only within each
super-class circularly. The asymmetric noise ratio is explored between $0.2$
and $0.4$.
Webvision [24]. This is a real-world dataset with noisy labels. It contains
2.4 million images crawled from the web (Google and Flickr) that share the
same 1,000 classes from the ImageNet dataset. The noise ratio varies from 0.5%
to 88%, depending on the class. In order to speed-up the training time, we
used mini Webvision [15], consisting of only top 50 classes in the Google
subset (66,000 images).
Clothing1M [42]. Clothing 1M is a large real-world dataset consisting of 1
million images on 14 classes of clothing articles. Being gathered from
e-commerce websites, Clothing1M embeds an unknown ratio of label noise.
Additional validation and test sets, consisting of 14k and 10k clean labeled
samples have been made available. In order to speed-up the training time, we
selected a subset of 56,000 images keeping the initial class distribution.
Both Webvision and Clothing1M images were resized to $128\times 128$.
Therefore, the reported results may differ from other papers cropping the
images to a $224\times 224$ resolution.
### 5.2 Settings
We use the contrastive SimCLR framework [5] with a ResNet18 [11] (without
ImageNet pre-training) as encoder. A projection head was added after the
encoder for the contrastive learning with the following architecture: a multi-
layer perceptron with one hidden layer and a ReLu non-linearity. The
classifier following the contrastive learning step has a simple multilayer
architecture: a single hidden layer with batch normalization and a ReLU
activation function. A comparison with a linear classifier is provided in the
supplementary materials.
For all supervised classification, we use SGD optimizer with momentum 0.9 and
cosine learning rate annealing. The NFL hyperparameter $\gamma$ is set to
$0.5$. Unlike the original paper, the ELR hyperparameters do no depend on the
noise type: the regularization coefficient $\lambda_{elr}$ and the momentum
$\beta$ are set to $3.0$ and $0.7$. Details on the experiment setting can be
found in the supplementary materials.
All codes are implemented in the PyTorch framework [32]. The experiments for
CIFAR are performed with a single Nvidia TITAN V-12GB and the experiments for
Webvision and Clothing1M are performed with a single Nvidia Tesla V100-32GB,
demonstrating the accessibility of the method. Our implementation has been
made available along with the supplementary materials.
## 6 Results
All experiments presented in this secion evaluate our method’s performance
with the top-1 accuracy score.
### 6.1 Impact of contrastive pre-training
To evaluate the impact of the contrastive pre-training on the classification
model, the proposed method (pre-training phase) is compared with a baseline
classifier, trained for 200 epochs without contrastive learning. For each
simulated dataset, we compare robust losses (e.g. NLF+RCE and ELR) and cross
entropy. Results for CIFAR10 and CIFAR100 are depicted in Table 1 for
different levels of symmetric and asymmetric noise. The pre-training improves
the accuracy of the three different baselines for both datasets with different
types and ratios of label noise. The largest differences are observed for the
noisiest case with $80\%$ noise. The pre-training outperforms the baselines by
large margins between $10$ and $75$ for CIFAR10 and between $5$ and $30$ for
CIFAR100.
Table 1: Results on both CIFAR10 and CIFAR100 using symmetric noise (0.2 -
0.8) and asymmetric noise (0.2 - 0.4). We compare training from scratch or
from pre-trained representation. Best scores are in bold for each noise
scenario and each loss.
| | | CIFAR10 | CIFAR100
---|---|---|---|---
Type | $\eta$ | Loss | Base | Pre-t. | Base | Pre-t.
Sym | 0.2 | ce | 77.2 | 87.7 | 55.6 | 56.5
elr | 90.3 | 93.0 | 64.1 | 67.4
nfl+rce | 91.0 | 92.7 | 66.6 | 68.8
0.4 | ce | 58.2 | 78.0 | 39.9 | 41.9
elr | 82.3 | 92.0 | 56.9 | 62.0
nfl+rce | 87.0 | 91.4 | 60.2 | 66.3
0.6 | ce | 35.2 | 59.2 | 21.8 | 26.8
elr | 64.2 | 90.4 | 40.6 | 55.7
nfl+rce | 80.2 | 88.1 | 47.0 | 61.8
0.8 | ce | 17.0 | 27.3 | 7.80 | 12.4
elr | 18.3 | 84.8 | 16.2 | 45.3
nfl+rce | 42.8 | 59.9 | 20.1 | 50.2
Asym | 0.2 | ce | 84.0 | 87.9 | 59.0 | 57.8
elr | 91.8 | 92.4 | 70.3 | 70.2
nfl+rce | 90.2 | 91.5 | 63.9 | 68.4
0.3 | ce | 79.2 | 83.9 | 50.6 | 50.4
elr | 89.6 | 91.7 | 69.8 | 69.3
nfl+rce | 86.7 | 89.9 | 53.5 | 63.5
0.4 | ce | 75.3 | 77.8 | 41.8 | 42.4
elr | 72.3 | 89.5 | 67.6 | 67.6
nfl+rce | 80.0 | 82.4 | 40.6 | 47.8
In addition to the comparisons with ELR and NFL+RCE, performed using our
implementations (column Base in Table 1), we present the results reported by
other recent competing methods. As shown in the introduction, numerous
contributions have been made to the field in the last years. Six recent
representative methods are selected for comparison: Taks [34], Co-teaching+
[43], ELR [25], DivideMix [22], SELF [30], and JoCoR [39]. The results are
presented in Table 2. The difference between the scores reported by ELR and
those obtained with our run (using the same implementation, but slightly
different hyper-parameters and a ResNet18 instead of a ResNet34) suggests that
the method is less stable on data contaminated with asymmetric noise and
sensitive to small changes hyperparameters. Moreover, ELR proposes
hyperparameters having different values depending on the type of dataset (i.e.
CIFAR10/CIFAR100) and underlying noise (i.e. symmetric/asymmetric), identified
after a hyperparameter search exercise. The best scores are reported by
DivideMix and they surpass all other techniques. One can note DivideMix uses a
PreAct ResNet18 while we use a classical ResNet18. Moreover, a recent study
[31] attempted to replicate these values and reported significantly lower
results on CIFAR100 (i.e. $49.5\%$ instead of $59.6\%$ on symmetric data and
$50.9\%$ instead of $72.1\%$ on asymmetric data). Our framework compares
favourably with the other competing methods, both on symmetric and asymmetric
noise.
Table 2: Accuracy scores compared with 6 methods (Taks, Co-teaching+, ELR,
DivideMix, SELF, and JoCoR) on CIFAR10 (C10) and CIFAR100 (C100). The cases
most affected by dropout are presented, with symmetric (S) and asymmetric (A)
noise. Top-2 scores are in bold
| | C10
---
80% S
| C10
---
40% A
| C100
---
80% S
| C100
---
40% A
| Ours
---
(ELR)
84.8 | 89.5 | 45.3 | 67.6
| ELR [25]
---
73.9 | 91.1 | 29.7 | 73.2
| Taks [34]
---
40.2 | 73.4 | 16.0 | 35.2
| Co-teach+ [43]
---
23.5 | 68.5 | 14.0 | 34.3
| DivideMix [22]
---
92.9 | 93.4 | 59.6 | 72.1
| SELF [30]
---
69.9 | 89.1 | 42.1 | 53.8
| JoCoR [39]
---
25.5 | 76.1 | 12.9 | 32.3
Webvision and Clothing1M results are presented in Table 3. The contrastive
framework outperforms the respective baselines for the three loss functions.
Because the images have a reduced size, and for Clothing1M, we use a smaller
training set, the direct comparison with competing methods is less relevant.
However, the observed gap in performance is significant and promising for
training images with higher resolution. Moreover, a ResNet50 model has been
trained with our framework on the Webvision dataset with a higher resolution
($224\times 224$). The accuracy reaches respectively $75.7\%$ and $76.2\%$ for
CE and ELR. These results are very close to the values reported with DivideMix
($77.3\%$) and ELR+ ($77.8\%$) using a larger model, Inception-ResNet-v2 (the
difference is more than $4\%$ on the ImageNet benchmark [4]).
Table 3: Top-1 accuracy for mini-Webvision and Clothing1M. Best scores are in
bold for each dataset and each loss. Pre-t represents the pre-training phase
while Fine-tune refers to the results after the fine-tuning step.
| Webvision | Clothing1M
---|---|---
Loss | Base. | Pre-t. | Fine-tune | Base. | Pre-t. | Fine-tune
ce | 51.8 | 57.1 | 58.4 | 54.8 | 59.1 | 61.5
elr | 53.0 | 58.1 | 59.0 | 57.4 | 60.8 | 60.4
nfl+rce | 49.9 | 54.8 | 58.2 | 57.4 | 59.4 | 60.1
Supported by this first set of experiments, the preliminary pre-training with
contrastive learning shows great performances. The accuracy of both
traditional and robust-loss classification models is significantly improved.
### 6.2 Sensitivity to the hyperparameters
Estimating the best hyperparameters is complex for datasets with noisy labels
as clean validation sets are not available. For instance, Ortego et al. [31]
show that two efficient methods (eg. ELR and DivideMix) could be sensitive to
specific hyperparameters. Therefore a hyperparameter sensitivity study has
been carried out to estimate the stability of the framework for the learning
rate. Figure 5 depicts the sensitivity on CIFAR100 with $80\%$ noise. CE and
NFL+RCE seem to have opposite behaviors. The CE reaches competitive results
with small learning rates but is prompt to overfitting for higher learning
rates. The NFL+RCE loss tends to underfitting for the lowest learning rates
but is quite robust for higher values. The ELR loss has the smallest
sensitivity to the learning for the investigated range but does not reach the
best values obtained with CE or NFL+RCE. We can assume that the regularization
term coupled with pre-training is very efficient. It prevents memorization of
the false labels as observed with CE. Results for other noise ratios have been
documented in the supplementary materials.
Figure 5: Learning rate sensitivity for CIFAR100 with 80% noise. The explored
learning rate values are $\\{0.001,0.01,0.1,1.0\\}$. The baseline (dashed
line) is compared with our framework (solid line).
This sensitivity analysis is limited to the learning rate. Investigating the
impact of other hyperparameters, such as the momentum $\beta$ or the
regularization factor $\lambda_{elr}$, could be interesting. In their original
papers, ELR and NFL+RCE reach respectively $25.2\%$ and $30.3\%$ with other
hyperparameters. These values are still far from the improvements brought by
the contrastive pre-training but it suggests that the results could be
improved with different hyperparameters.
Our empirical results indicate that the analyzed methods may be sensitive to
hyperparameters. Despite the promised robustness to label noise, the analyzed
robust losses are also affected by overfitting or underfitting. Our
experiments have been built upon the parameters recommended in each issuing
paper (e.g. ELR, SIMCLR) but, since the individual building blocks can be
affected by small variations in input parameters, the performance of our
method may also be impacted. Finding a relevant method to estimate proper
hyperparameters in NLL remains a challenge. In the absence of a clean
validation set, identifying when overfitting starts also remains an open
challenge. This is demonstrated by our studies on the behaviour of the (also
noise-corrupted) validation set and another two recently proposed methods,
analyzing the stability of the loss function on the train set and the changes
in the upstream layers. These experiments are detailed in Supplementary
Materials.
### 6.3 Impact of the fine-tuning phase
Experimental results on synthetic label noise, depicted in Figure 6, show that
continuing the presented pre-training block (Figure 2) with the fine-tuning
phase increases the accuracy in over 65% of cases on CIFAR10 and over 80% of
cases on CIFAR100. For both datasets, asymmetric noise data benefit more from
this approach than symmetric noise. All experiments only use the input
parameters proposed in the loss-issuing papers.
Figure 6: Accuracy gain when performing the fine-tuning phase after the pre-
training block (computed as the difference between fine-tuning accuracy and
pre-training accuracy). The plot gathers the results for all noise ratios on
CIFAR10 (panels a, b) and CIFAR100 (c, d) with symmetric (first column) and
asymmetric (second column) noise.
The sample selection has also got a positive impact on the two real-world
datasets, as shown in Table 3 by the ”Fine-tune” columns. The average accuracy
improvement is about $1.8\%$. Only the ELR loss function slightly decreases
the performance on Clothing1M.
Enriching pretrained models with sample weighting and selection, pseudo labels
instead of corrupted targets, and supervised contrastive pre-training can
improve the classification accuracy. However, such an approach raises the
question of a trade-off between complexity, accuracy improvement, and
computation time.
## 7 Discussion and limits of the framework
In addition to the presented fine-tuning phase, we evaluated the performance
of other promising techniques, such as the dynamic bootstrapping with mixup
[1]. This strategy has been developed to help convergence under extreme label
noise conditions. Details can be found in the supplementary materials. The
improvement that dynamic bootstraping can bring when used after pre-training
is depicted in Figure 7. In most of the cases, this technique improves the
accuracy, as indicated by the positive accuracy gain scores, measuring the
difference between the accuracy after dynamic bootstraping and the accuracy of
the pre-training phase. ELR and CE benefit most from this addition for
CIFAR100. The impact of the dynamic boostrapping should also be analyzed for
the fine-tuning phase and for larger datasets, such as Webvision or
Clothing1M.
Figure 7: Top-1 accuracy gain for the dynamic bootstrapping on CIFAR100 with
asymmetric (a) and symmetric noise (b). Dynamic bootstrapping is an
alternative to the proposed fine-tuning phase.
One of the major drawbacks of our method is the extra computational time
needed to learn representations with contrastive learning. A detailed study,
comparing the execution time of our framework with 6 other competing methods
has been provided in supplementary materials. The pre-training phase doubles
the execution time of a reference baseline, consisting of performing only a
single classification step, while the entire framework increases the execution
time 3 to 4 times the baseline value. However, the constrastive learning does
not increase the need for GPU memory if the batch size is limited for the
contrastive learning [29, 12]. The computational time could be reduced by
initializing the contrastive step with the pretrained weights from ImageNet.
Most state-of-the-art approaches also leverage computationally expensive
settings, consisting of larger models (e.g. ResNet50), dual model training, or
data augmentation such as mixup. In this work, we explored the limits of a
restricted computational setting, consisting of a single GPU and 8GB RAM. All
experiments use a ResNet18 model, batch sizes of $256$, and for real-world
datasets, the images have been rescaled (e.g. $128\times 128$ instead of
$224\times 224$). We also foresee that the constrastive learning step could be
improved by images with higher resolutions as smaller details could be
identified in the representation embedding.
There remain multiple open problems for future research, such as: i)
identifying the start of the memorization phase in the absence of a clean
dataset, ii) studying the impact of contrastive learning on other models for
noisy labels such as DivideMix, iii) comparing SimCLR approach in the context
of noisy labels with other contrastive frameworks (the impact of Moco is
studied in the supplementary materials) and other self-supervised approaches,
and iv) having a better theoretical understanding of the interaction between
the initial state precomputed with contrastive learning and the classifier in
presence of noisy labels. Moreover, the analysis carried out in this work
should be validated on larger settings, in particular on Clothing1M with a
ResNet50, higher resolutions, and the full dataset.
## 8 Conclusions
In this work, we presented a contrastive learning framework optimized with
several adaptations for noisy label classification. Supported by an extensive
range of experiments, we conclude that a preliminary representation pre-
training improves the performance of both traditional and robust-loss
classification models. Additionally, multiple techniques can be used to fine-
tune and further optimize these results; however, no approach provides a
significant improvement systematically on all types of datasets and label
noise. The cross-entropy penalized by Early-Learning Regularization (ELR)
shows the best overall results for synthetic noise but also real-world
datasets.
However, the training phases remain sensitive to input configuration.
Overfitting is the common weakness of all studied models. When trained with
tuned parameters, even traditional (cross-entropy) models provide competitive
results, while robust-losses are less sensitive. The typical noisy label
adaptations, such as sample selection or weighting, the usage of pseudo
labels, or supervised contrastive losses, improve the performance to a lesser
extent but increase the framework’s complexity. We hope that this work will
promote the use of contrastive learning to improve the robustness of the
classification process with noisy labels.
## References
* Arazo et al. [2019] Eric Arazo, Diego Ortego, Paul Albert, Noel O’Connor, and Kevin McGuinness. Unsupervised label noise modeling and loss correction. In _International Conference on Machine Learning_ , pages 312–321. PMLR, 2019.
* Arpit et al. [2017] Devansh Arpit, Stanisław Jastrzebski, Nicolas Ballas, David Krueger, Emmanuel Bengio, Maxinder S Kanwal, Tegan Maharaj, Asja Fischer, Aaron Courville, Yoshua Bengio, et al. A closer look at memorization in deep networks. In _International Conference on Machine Learning_ , pages 233–242. PMLR, 2017.
* Berthelot et al. [2019] David Berthelot, Nicholas Carlini, Ian Goodfellow, Nicolas Papernot, Avital Oliver, and Colin A Raffel. Mixmatch: A holistic approach to semi-supervised learning. In _Advances in Neural Information Processing Systems_ , pages 5049–5059, 2019.
* Bianco et al. [2018] Simone Bianco, Remi Cadene, Luigi Celona, and Paolo Napoletano. Benchmark analysis of representative deep neural network architectures. _IEEE Access_ , 6:64270–64277, 2018.
* Chen et al. [2020a] Ting Chen, Simon Kornblith, Mohammad Norouzi, and Geoffrey Hinton. A simple framework for contrastive learning of visual representations. _arXiv preprint arXiv:2002.05709_ , 2020a.
* Chen et al. [2020b] Xinlei Chen, Haoqi Fan, Ross Girshick, and Kaiming He. Improved baselines with momentum contrastive learning. _arXiv preprint arXiv:2003.04297_ , 2020b.
* Falcon and Cho [2020] William Falcon and Kyunghyun Cho. A framework for contrastive self-supervised learning and designing a new approach. _arXiv preprint arXiv:2009.00104_ , 2020.
* Ghosh et al. [2017] Aritra Ghosh, Himanshu Kumar, and PS Sastry. Robust loss functions under label noise for deep neural networks. In _Proceedings of the AAAI Conference on Artificial Intelligence_ , volume 31, 2017.
* Goldberger and Ben-Reuven [2017] J. Goldberger and E. Ben-Reuven. Training deep neural-networks using a noise adaptation layer. In _ICLR_ , 2017.
* Han et al. [2018] Bo Han, Quanming Yao, Xingrui Yu, Gang Niu, Miao Xu, Weihua Hu, Ivor Tsang, and Masashi Sugiyama. Co-teaching: Robust training of deep neural networks with extremely noisy labels. In _Advances in neural information processing systems_ , pages 8527–8537, 2018.
* He et al. [2016] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In _Proceedings of the IEEE conference on computer vision and pattern recognition_ , pages 770–778, 2016.
* He et al. [2020] Kaiming He, Haoqi Fan, Yuxin Wu, Saining Xie, and Ross Girshick. Momentum contrast for unsupervised visual representation learning. In _Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition_ , pages 9729–9738, 2020.
* Henaff [2020] Olivier Henaff. Data-efficient image recognition with contrastive predictive coding. In _International Conference on Machine Learning_ , pages 4182–4192. PMLR, 2020.
* Hendrycks et al. [2018] Dan Hendrycks, Mantas Mazeika, Duncan Wilson, and Kevin Gimpel. Using trusted data to train deep networks on labels corrupted by severe noise. In _Advances in neural information processing systems_ , pages 10456–10465, 2018.
* Jiang et al. [2018] Lu Jiang, Zhengyuan Zhou, Thomas Leung, Li-Jia Li, and Li Fei-Fei. Mentornet: Learning data-driven curriculum for very deep neural networks on corrupted labels. In _International Conference on Machine Learning_ , pages 2304–2313. PMLR, 2018.
* Kalantidis et al. [2020] Yannis Kalantidis, Mert Bulent Sariyildiz, Noe Pion, Philippe Weinzaepfel, and Diane Larlus. Hard negative mixing for contrastive learning. _arXiv preprint arXiv:2010.01028_ , 2020.
* Kamabattula et al. [2020] Sree Ram Kamabattula, Venkat Devarajan, Babak Namazi, and Ganesh Sankaranarayanan. Identifying training stop point with noisy labeled data, 2020.
* Khosla et al. [2020] Prannay Khosla, Piotr Teterwak, Chen Wang, Aaron Sarna, Yonglong Tian, Phillip Isola, Aaron Maschinot, Ce Liu, and Dilip Krishnan. Supervised contrastive learning. _Advances in Neural Information Processing Systems_ , 33, 2020.
* Kornblith et al. [2019] Simon Kornblith, Mohammad Norouzi, Honglak Lee, and Geoffrey Hinton. Similarity of neural network representations revisited. In _International Conference on Machine Learning_ , pages 3519–3529. PMLR, 2019.
* Krizhevsky [2009] A Krizhevsky. Learning multiple layers of features from tiny images. _Master’s thesis, University of Tront_ , 2009.
* Le-Khac et al. [2020] Phuc H Le-Khac, Graham Healy, and Alan F Smeaton. Contrastive representation learning: A framework and review. _IEEE Access_ , 2020.
* Li et al. [2020a] Junnan Li, Richard Socher, and Steven C.H. Hoi. Dividemix: Learning with noisy labels as semi-supervised learning. In _International Conference on Learning Representations_ , 2020a.
* Li et al. [2020b] Junnan Li, Caiming Xiong, and Steven CH Hoi. Mopro: Webly supervised learning with momentum prototypes. _arXiv preprint arXiv:2009.07995_ , 2020b.
* Li et al. [2017] Wen Li, Limin Wang, Wei Li, Eirikur Agustsson, and Luc Van Gool. Webvision database: Visual learning and understanding from web data. _arXiv preprint arXiv:1708.02862_ , 2017.
* Liu et al. [2020] Sheng Liu, Jonathan Niles-Weed, Narges Razavian, and Carlos Fernandez-Granda. Early-learning regularization prevents memorization of noisy labels. _Advances in Neural Information Processing Systems_ , 33, 2020.
* Ma et al. [2020] Xingjun Ma, Hanxun Huang, Yisen Wang, Simone Romano, Sarah Erfani, and James Bailey. Normalized loss functions for deep learning with noisy labels. In _International Conference on Machine Learning_ , pages 6543–6553. PMLR, 2020.
* Mahajan et al. [2018] Dhruv Mahajan, Ross Girshick, Vignesh Ramanathan, Kaiming He, Manohar Paluri, Yixuan Li, Ashwin Bharambe, and Laurens van der Maaten. Exploring the limits of weakly supervised pretraining. In _Proceedings of the European Conference on Computer Vision (ECCV)_ , pages 181–196, 2018.
* Misra and Maaten [2020] Ishan Misra and Laurens van der Maaten. Self-supervised learning of pretext-invariant representations. In _Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition_ , pages 6707–6717, 2020.
* Mitrovic et al. [2020] Jovana Mitrovic, Brian McWilliams, and Melanie Rey. Less can be more in contrastive learning. In _”I Can’t Believe It’s Not Better!” NeurIPS 2020 workshop_ , 2020\.
* Nguyen et al. [2019] Duc Tam Nguyen, Chaithanya Kumar Mummadi, Thi Phuong Nhung Ngo, Thi Hoai Phuong Nguyen, Laura Beggel, and Thomas Brox. Self: Learning to filter noisy labels with self-ensembling. In _International Conference on Learning Representations_ , 2019.
* Ortego et al. [2020] Diego Ortego, Eric Arazo, Paul Albert, Noel E. O’Connor, and Kevin McGuinness. Multi-objective interpolation training for robustness to label noise, 2020\.
* Paszke et al. [2019] Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, Alban Desmaison, Andreas Kopf, Edward Yang, Zachary DeVito, Martin Raison, Alykhan Tejani, Sasank Chilamkurthy, Benoit Steiner, Lu Fang, Junjie Bai, and Soumith Chintala. Pytorch: An imperative style, high-performance deep learning library. In H. Wallach, H. Larochelle, A. Beygelzimer, F. d'Alché-Buc, E. Fox, and R. Garnett, editors, _Advances in Neural Information Processing Systems 32_ , pages 8024–8035. Curran Associates, Inc., 2019.
* Patrini et al. [2017] Giorgio Patrini, Alessandro Rozza, Aditya Krishna Menon, Richard Nock, and Lizhen Qu. Making deep neural networks robust to label noise: A loss correction approach. In _Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition_ , pages 1944–1952, 2017.
* Song et al. [2020a] Heon Song, Nariaki Mitsuo, Seiichi Uchida, and Daiki Suehiro. No regret sample selection with noisy labels. _arXiv preprint arXiv:2003.03179_ , 2020a.
* Song et al. [2019] Hwanjun Song, Minseok Kim, and Jae-Gil Lee. Selfie: Refurbishing unclean samples for robust deep learning. In _International Conference on Machine Learning_ , pages 5907–5915. PMLR, 2019.
* Song et al. [2020b] Hwanjun Song, Minseok Kim, Dongmin Park, and Jae-Gil Lee. Learning from noisy labels with deep neural networks: A survey. _arXiv preprint arXiv:2007.08199_ , 2020b.
* Wang et al. [2019] Yisen Wang, Xingjun Ma, Zaiyi Chen, Yuan Luo, Jinfeng Yi, and James Bailey. Symmetric cross entropy for robust learning with noisy labels. In _Proceedings of the IEEE International Conference on Computer Vision_ , pages 322–330, 2019.
* Wang et al. [2020] Zhuowei Wang, Jing Jiang, Bo Han, Lei Feng, Bo An, Gang Niu, and Guodong Long. Seminll: A framework of noisy-label learning by semi-supervised learning. _CoRR_ , abs/2012.00925, 2020. URL https://arxiv.org/abs/2012.00925.
* Wei et al. [2020] Hongxin Wei, Lei Feng, Xiangyu Chen, and Bo An. Combating noisy labels by agreement: A joint training method with co-regularization, 2020.
* Wu et al. [2018] Zhirong Wu, Yuanjun Xiong, Stella X Yu, and Dahua Lin. Unsupervised feature learning via non-parametric instance discrimination. In _Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition_ , pages 3733–3742, 2018.
* Xia et al. [2019] Xiaobo Xia, Tongliang Liu, Nannan Wang, Bo Han, Chen Gong, Gang Niu, and Masashi Sugiyama. Are anchor points really indispensable in label-noise learning? In _Advances in Neural Information Processing Systems_ , pages 6838–6849, 2019.
* Xiao et al. [2015] Tong Xiao, Tian Xia, Yi Yang, Chang Huang, and Xiaogang Wang. Learning from massive noisy labeled data for image classification. In _Proceedings of the IEEE conference on computer vision and pattern recognition_ , pages 2691–2699, 2015.
* Yu et al. [2019] Xingrui Yu, Bo Han, Jiangchao Yao, Gang Niu, Ivor Tsang, and Masashi Sugiyama. How does disagreement help generalization against label corruption? In _International Conference on Machine Learning_ , pages 7164–7173. PMLR, 2019.
* Zhang et al. [2016] Chiyuan Zhang, Samy Bengio, Moritz Hardt, Benjamin Recht, and Oriol Vinyals. Understanding deep learning requires rethinking generalization. _arXiv preprint arXiv:1611.03530_ , 2016.
* Zhang et al. [2018] Hongyi Zhang, M. Cissé, Yann Dauphin, and David Lopez-Paz. mixup: Beyond empirical risk minimization. In _International Conference on Learning Representations (ICLR)_ , 2018.
* Zhang and Yao [2020] Hui Zhang and Quanming Yao. Decoupling representation and classifier for noisy label learning, 2020\.
* Zhang and Sabuncu [2018] Zhilu Zhang and Mert Sabuncu. Generalized cross entropy loss for training deep neural networks with noisy labels. _Advances in neural information processing systems_ , 31:8778–8788, 2018.
## Supplementary Materials
## Appendix A Description of the datasets
Table 4 gives a detailed description of datasets, including size of the
training and test sets, the image resolution, and the number of classes.
Table 4: Description of the datasets used in the experiments. Data set | Train | Test | Size | # classes
---|---|---|---|---
CIFAR10 | 50K | 10K | 32x32 | 10
CIFAR100 | 50K | 10K | 32x32 | 100
Clothing1M | 56K | 5K | 128x128 | 14
Mini-Webvision | 66K | 2.5K | 128x128 | 50
## Appendix B Detailed settings of the experiments
All experiments use the ResNet18 as encoder. The classification steps are
combined with data augmentation: a random crop with a padding of $4$, an
horizontal flip with a probability of $50\%$, and a random rotation of
$20^{\circ}$. All other hyperparameters are resumed in Table 5.
Table 5: Training parameters. Symbols: l.r means learning rate, w.d means weight decay, opti. means optimizer, Repre. means representation step, Classi. means the supervised classification step. | | C10/C100 | Webvision | Clothing1M
---|---|---|---|---
Repre. | Batch | 512 | 512 | 512
Opti. | Adam | Adam | Adam
l.r. | $10^{-3}$ | $10^{-3}$ | $10^{-3}$
w.d. | $10^{-6}$ | $10^{-6}$ | $10^{-6}$
epochs | 500 | 500 | 500
Classi. | Batch | 256 | 256 | 256
Opti. | SGD | SGD | SGD
l.r. | 0.01/0.1 | 0.4 | 0.01
w.d. | $10^{-5}$ | $3.10^{-5}$ | $10^{-4}$
epochs | 200 | 200 | 200
## Appendix C Ablation study
### C.1 Contrastive learning with a momentum encoder
The momentum encoder from the Moco framework [12] maintains a dynamic memory
queue of representations. The current mini-batch is added to the memory queue
while the oldest mini-batch is dequeued. The offline momentum encoder is a
copy of the online encoder by taking an exponentially-weighted average of the
parameter of the online encoder. The main advantage of Moco is to be able to
reduce the batch size (and the GPU memory) while keeping a very large number
of negative pairs for the contrastive learning.
Table 6: Top-1 accuracy on CIFAR100 with $80\%$ noise. Two different contrastive learning frameworks are evaluated for the pre-training: SimCLR and Moco. The third column gives the accuracy for a classifier with a smaller learning rate. | SimCLR | Moco | | Moco - Fine
---
tune
CE | 12.4 | 12.0 | 49.0
ELR | 45.3 | 38.8 | 42.3
NFL+RCE | 50.2 | 26.3 | 47.0
The different representations computed by SimCLR and Moco are compared on
CIFAR100. Both approaches are trained for 500 epochs following the usual
hyperparameter parameters from the initial papers. As the two methods use
different strategies to compute the representations, their quality is assessed
by learning a linear classifier on top of the frozen encoder network. It can
be seen as a proxy for representation quality. The SimCLR framework reaches
$55.3\%$ of accuracy while Moco gets $55.0\%$ of accuracy. However, the two
encoders do not behave in a similar way with regard to noisy labels. The same
classifier (multi-layer, same learning rate and weight decay) is trained
starting from the representation computed by SimCLR and Moco. As depicted in
Table 6, the representations computed by Moco are more sensitive to the noisy
labels. However, reducing the learning rate of the optimizer by a factor $10$
(column Moco - Fine Tune) significantly increases the accuracy.
Even if pretraining the encoder increases the accuracy for both contrastive
methods, the two approaches do not have the same behavior. In particular, the
best parameters for the classifier optimizer seem to be different. This raises
several questions about the difference between the two representations and
what properties of these representations improve the robustness of the
classifier.
### C.2 Sensitivity to the learning rate
We perform an hyperparameter search on the CIFAR100 datasets. The learning
rate is chosen in $\\{10^{-3},10^{-2},10^{-1},10^{0}\\}$. Results are
presented in Figure 8. The configuration with $80\%$ noise is clearly the most
sensitive case, in particular for the NFL+RCE loss and the CE. The ELR method
is quiet robust over the investigated range.
(a) CIFAR100 with $80\%$ noise. (b) CIFAR100 with $60\%$ noise.
(c) CIFAR100 with $40\%$ noise. (d) CIFAR100 with $20\%$ noise.
Figure 8: Hyperparameter sensitivity for CIFAR100.
### C.3 Impact of the classifier architecture
The impact of the 2 classifier architectures is detailed in Table 7. The
multilayer architecture performs better on datasets contaminated with a
significant amount of asymmetric noise.
Table 7: Results on both CIFAR10 and CIFAR100 using symmetric noise (0.2 -
0.8) and asymmetric noise (0.2 - 0.4). We compare a single linear layer (L) to
multiple layers (M) final classification head, for three losses: CE, ELR, and
NFL+RCE.
| | | CIFAR10 | CIFAR100
---|---|---|---|---
Type | $\eta$ | Loss | L | M | L | M
Sym | 0.2 | ce | 91.7 | 87.7 | 58.6 | 56.5
elr | 92.9 | 93.0 | 66.4 | 67.4
nfl_rce | 93.2 | 92.7 | 69.7 | 68.8
0.4 | ce | 90.6 | 78.0 | 44.2 | 41.9
elr | 92.1 | 92.0 | 60.8 | 62.0
nfl_rce | 92.1 | 91.4 | 67.0 | 66.3
0.6 | ce | 88.1 | 59.2 | 28.9 | 26.8
elr | 89.7 | 90.4 | 54.0 | 55.7
nfl_rce | 90.2 | 88.1 | 63.7 | 61.8
0.8 | ce | 72.6 | 27.3 | 14.1 | 12.4
elr | 82.0 | 84.8 | 41.6 | 45.3
nfl_rce | 78.9 | 59.9 | 54.2 | 50.2
Asym | 0.2 | ce | 91.6 | 87.9 | 60.1 | 57.8
elr | 92.7 | 92.4 | 69.3 | 70.2
nfl_rce | 92.5 | 91.5 | 69.1 | 68.4
0.3 | ce | 90.2 | 83.9 | 52.3 | 50.4
elr | 90.6 | 91.7 | 68.5 | 69.3
nfl_rce | 91.2 | 89.9 | 68.0 | 63.5
0.4 | ce | 84.7 | 77.8 | 43.7 | 42.4
elr | 68.4 | 89.5 | 65.5 | 67.6
nfl_rce | 62.6 | 82.4 | 63.0 | 47.8
## Appendix D Dynamic bootstrapping with mixup
In addition to the presented fine-tuning phase, we also evaluated the
performance of other techniques recently proposed for noisy label
classification. The weights $w$ computed by the sample selection phase can
also be combined with a mixup data augmentation strategy [45]. A specific
strategy for noisy labels, called dynamic bootstrapping with mixup [1], has
been developed to help convergence under extreme label noise conditions. The
convex combinations of sample pairs $\bm{x_{p}}$ (loss $l_{p}$) and
$\bm{x_{q}}$ (loss $l_{q}$) is weighted by the probability $w$ to belong to
the clean dataset:
$\bm{x}=\frac{w_{p}}{w_{p}+w_{q}}\bm{x_{p}}+\frac{w_{q}}{w_{p}+w_{q}}\bm{x_{q}}.$
(11) $l=\frac{w_{p}}{w_{p}+w_{q}}l_{p}+\frac{w_{q}}{w_{p}+w_{q}}l_{q}.$ (12)
The associated CE is corrected according to the weights:
$l_{ce}=-\sum_{k=1}^{K}\left(w_{i}q(k|\bm{x_{i}})+(1-w_{i})z_{i}\right)log(p(k|\bm{x_{i}})),$
(13)
where $z(k|\bm{x_{i}})=1$ if $k=\operatorname*{\arg\\!\max}p(k|\bm{x_{i}})$ or
zero for all the other cases. If the GMM probability are well estimated,
combining one noisy sample with one clean sample leads to a large weight for
the clean sample and a small weight for the noisy sample. Clean-clean and
noisy-noisy cases remain similar to a classical mixup with weights around
$0.5$.
The dynamic bootstrapping for ELR is derived by replacing the CE term by the
corrected version:
$l_{elr}(\theta)=l_{ceb}(\theta)+\frac{\lambda_{elr}}{N}log\left(1-\sum_{k=1}^{K}p(k|\bm{x_{i}}).t(k|\bm{x_{i}})\right).$
(14)
Regarding the robust loss function NFL+RCE, the two losses have to be
modified:
$\displaystyle
l_{nfl}=w_{i}\frac{-\sum\limits_{k=1}^{K}q(k|\bm{x_{i}})(1-p(k|\bm{x_{i}}))^{\gamma}log(p(k|\bm{x_{i}}))}{-\sum\limits_{j=1}^{K}\sum\limits_{k=1}^{K}q(y=j|\bm{x_{i}})(1-p(k|\bm{x_{i}}))^{\gamma}log(p(k|\bm{x_{i}}))}$
(15)
$\displaystyle+(1-w_{i})\frac{-\sum\limits_{k=1}^{K}z(k|\bm{x_{i}})(1-p(k|\bm{x_{i}}))^{\gamma}log(p(k|\bm{x_{i}}))}{-\sum\limits_{j=1}^{K}\sum\limits_{k=1}^{K}z(y=j|\bm{x_{i}})(1-p(k|\bm{x_{i}}))^{\gamma}log(p(k|\bm{x_{i}}))}$
where $q$ is the one-hot encoding of the label (the zero value is fixed to a
low value to avoid $log(0)$).
$l_{rce}=-\sum_{k=1}^{K}p(k|\bm{x_{i}})log\left(w_{i}.q(k|\bm{x_{i}})+(1-w_{i})z_{i}\right)$
(16)
## Appendix E Classification warmup
This section compares the classification accuracy of models trained with and
without a warm-up phase after the representation learning. The warm-up phase
consists of freezing the entire model except for the classification head.
Figure 9 depicts the gain in performance brought by the warm-up phase. When
using the default values, its inclusion is beneficial only for significant
amounts of symmetric noise. Our experiments have been performed using only the
recommended classifier learning rates, detailed in the experimental setup.
Having different learning rates for the warm-up phase and the classification
optimizing all weights (encoder and classifier) could have a different impact
on the warmup phase.
Figure 9: Gain in performance when using a supplementary classifier warm-up
phase before training the entire model on CIFAR 100 with symmetric (panel a)
and asymmetric noise (panel b).
## Appendix F Execution time analysis
In order to estimate our method’s computational cost, we compared the
execution time of both approaches, consisting of performing only the pre-
training phase and the pre-training followed by fine-tuning with the execution
time of performing only one supervised classification phase (i.e. the
baseline). The number of times our methods were slower than the baseline has
been depicted in Table 8. We provided similar metrics for the methods making
available this informations (i.e. Taks, Co-teaching+, JoCoR). As expected, the
pre-training doubles the execution time of the baseline as, in addition to
training the classifier, a contrastive learning phase has to be performed
beforehand. The entire framework introduces a computational cost 3 to 4.5
times higher. However, all methods leveraging pre-trained models (using for
instance supervised pre-training) also hide a similar computational cost.
Table 8: Comparison of execution time results reported as a factor with respect to the training time of the baseline, representing the supervised training of the model with the CE loss. The abbreviations Ours (Pre-t) indicate the pre-training phase while Ours (Fine-tune) the pre-training phase followed by fine-tuning | | C10
---
80% S
| C10
---
40% A
| C100
---
80% S
| C100
---
40% A
| Ours
---
(Pre-t)
2.36 | 2.53 | 2.40 | 2.32
| Ours
---
(Fine-tune)
3.42 | 3.63 | 4.31 | 4.36
Taks | 0.53 | 1.04 | 0.52 | 0.98
Co-teach+ | 2.00 | 2.00 | 2.00 | 2.01
JoCoR | 1.73 | 1.74 | 1.72 | 1.74
## Appendix G An attempt to prevent overfitting with early stopping
Overfitting is the common weakness of all studied models. Several strategies
understanding and preventing overfitting have been explored: i) analysing the
model behaviour on a validation set, ii) identifying the start of the
memorization phase using Training Stop Point [17], and iii) characterizing
changes in the model using Centered Kernel Alignment [19]. A clean validation
set is generally used to find the best moment for early stopping and to
estimate the hyperparameter sets. However, we assume that clean validation
samples are not available. Therefore, the methods must be robust to
overfitting and to a wide range of hyperparameter values.
As typical noisy label settings lack a clean reference set, we contrasted the
behavior of the model on a corrupted validation set with that on a clean test
set, where overfitting can be easily identified. Train/validation sets have
been generated using 5 cross validation folds. In the figure below, panel (a)
depicts the evolution of accuracy scores on the corrupted train/validation
sets as well as on the test set. After the first 50 epochs, the model starts
overfitting as the test accuracy drops by 10% ( Figure 10 panel a). The
accuracy on the corrupted train continues to increase as the model memorizes
the input labels. However, on the corrupted validation set a plateau followed
by a loss of performance is indicative of the same phenomena, but without
being always aligned with the overfitting phase observed on the test-set. The
memorization phenomena of the train-set labels incapacitates the model to
generalize on the corrupted validation set and explains the significant
difference in scores between the train and validation accuracies.
A second perspective on the analysis of overfitting explores the stability of
the network’s predictions on the validation set. Panel (b) depicts the number
of samples predicted in different classes across consecutive epochs. As the
model starts overfitting, the prediction stability also increases. After 200
epochs, only 500 from 10000 samples on the validation set change class from
one epoch to another. As expected, the network stability is correlated with
model overfitting on severe label noise.
Figure 10: Evolution of accuracy across train/validation/test sets (a)
Prediction stability on the validation set computed as the number of samples
changing class across consecutive epochs. The rolling mean average of the
number of predictions has been depicted in black. The experiments have been
performed on CIFAR 100, with 80% symmetric noise during the first
classification phase and used NFL + RCE loss.
Figure 11: Evolution of accuracy across train/validation/test sets (a)
Prediction stability on the validation set computed as the number of samples
changing class across consecutive epochs. The rolling mean average of the
number of predictions has been depicted in black. The experiments have been
performed on CIFAR 100, with 40% asymmetric noise during the first
classification phase and used NFL + RCE loss.
Several recent contributions studied the overfitting phenomena of neural
networks in an attempt to identify an early stopping point corresponding to
the maximum obtainable test accuracy. Traditional approaches leverage a clean
test set which is often unavailable when confronted with noisy labelled data.
Kamabattula et al. [17] proposed to find a Training Stop Point (TSP), a
heuristic analyzing the rate of change in the training accuracy and correlated
its transition towards the memorization phase with a transition towards a
smoother (smaller variance) regime, as depicted below. Our experimental
results showed that the theoretical conditions to identify the early stopping
point are not always met as suggested by TSP. Figure 12 indicates that the
overfitting phase, starting after the first 5 epochs, does not change the
variance of the train loss.
Figure 12: Evolution of train loss and test accuracy on CIFAR, 60% symmetric
noise. The theoretical conditions of higher variance on the train loss,
associated with the start of the memorization phase, as suggested by TSP, are
not fulfilled.
Centered Kernel Alignment (CKA) [19] provides a similarity index comparing
representations between layers of different trained models. In particular, CKA
shows interesting properties as CKA can consistently identify correspondences
between layers trained from different initializations.
The objective is twofold: i) observing if a specific behavior can be
identified for the overfitting and ii) comparing the CKA values with and
without contrastive pre-training. The CKA index is computed at three different
locations in the network: the input layer, the middle of the network, and the
final layer. Figure 13 shows the CKA similarity computed between the
initialization/pre-trained model and the same layer at different epochs during
the training process. It is interesting to note that the first layer of the
pre-trained model remains very similar to the same layer computed by
contrastive learning. Such behavior was expected in order to improve the
robustness against noisy labels. Indeed, if contrastive learning can extract
good representations for semi-supervised or transfer learning, being close to
such representations can also help to avoid learning noisy labels. As
expected, all layers of the model trained from a random initialization vary
much more during the training.
The training phase of the pre-trained model reaches its maximum accuracy
around 50 epochs but the CKA values of the middle and last layers continue to
drop until 130 epochs. On the other hand, the CKA values of the initialized
model remain stable after $150$ epochs when the test accuracy reaches almost
its maximum value. At first glance, the CKA behavior cannot be related to
overfitting.
(a) CKA from a pre-trained encoder with contrastive learning. (b) CKA from a
random initialization.
Figure 13: CKA similarity for a model trained with NFL+RCE loss function on
CIFAR100 with $80\%$ noise.
None of the studied approaches provides a solution preventing overfitting
across all our experiments and this problem remains an open question.
|
order estimate in perturbation theory gives for the nucleon ground state with
$n=0$
$\displaystyle\Delta
M_{\pm}=\mp\frac{g_{X}}{4}\int\frac{dz}{z}X_{0}(z)\left(\left|\tilde{f}_{L}^{0}(z)\right|^{2}-\left|\tilde{f}_{R}^{0}(z)\right|^{2}\right)=\mp\frac{g_{X}\sigma}{8\tilde{\kappa}^{2}}$
(B.181)
with $X_{0}(z)$ given in (II.5.1). (B.180) through the expansion around the
vev, $X(x,z)\approx X_{0}(z)e^{i\Pi(x,z)}$, would also generate a contribution
to the pion-nucleon coupling and also the axial-charge of the direct and
transition axial form factors YEE . Since our central interest is neutrino DIS
scattering we can neglect this coupling and its effects on our results, as
most of our analysis involves the behavior near the UV boundary where the
effects of $\sigma$ is negligible both in the nucleonic wavefunctions and the
spectrum.
## Appendix C Details of the Reggeon exchange
The bulk gauge field $L^{0}_{\mu}(k,z)$ exchange contribution to the
diffractive Compton scattering amplitude in the t-channel is given by
$\displaystyle i{\cal A}^{L}_{Lp\rightarrow Lp}(s,t)=\sum_{n}i\tilde{{\cal
A}}^{L}_{Lp\rightarrow Lp}(m_{n},s,t)$ $\displaystyle i\tilde{{\cal
A}}^{L}_{Lp\rightarrow
Lp}(m_{n},s,t)=(-i)V_{LLL}^{\mu}(q,q^{\prime},k,m_{n})\times\tilde{G}_{\mu\nu}(m_{n},t)\times(-i)V_{L\bar{\Psi}\Psi}^{\nu}(p_{1},p_{2},k,m_{n})\,,$
with the bulk vertices ($k=p_{2}-p_{1}=q-q^{\prime}$)
$\displaystyle V_{LLL}^{\mu}(q,q^{\prime},k,m_{n})\equiv$
$\displaystyle\left(\frac{\delta
S_{LLL}^{k}}{\delta(\epsilon^{0}_{\mu}\partial_{z}L^{0}(k,z))}\right)\,J_{L}(m_{n},z)+\left(\frac{\delta
S_{LLL}^{k}}{\delta(\epsilon^{0}_{\mu}L^{0}(k,z))}\right)\,J_{L}(m_{n},z)$
$\displaystyle=$ $\displaystyle
g_{5}^{3}\kappa_{CS}B^{\mu}(q,q^{\prime},\epsilon^{\pm})\int
dz\,\Big{(}\mathcal{V}(Q,z)\mathcal{V}(Q^{\prime},z)\partial_{z}J_{L}(m_{n},z)-\partial_{z}\mathcal{V}(Q,z)\mathcal{V}(Q^{\prime},z)J_{L}(m_{n},z)\Big{)}\,,$
$\displaystyle V_{L\bar{\Psi}\Psi}^{\nu}(p_{1},p_{2},k,m_{n})\equiv$
$\displaystyle\left(\frac{\delta
S_{L\bar{\Psi}\Psi}^{k}}{\delta(\epsilon^{0}_{\nu}L^{0}(k,z))}\right)\,J_{L}(m_{n},z)=g_{5}\int
dz\sqrt{g}\,e^{-\phi}z\bar{\Psi}(p_{2},z)\gamma^{\nu}\Psi(p_{1},z)J_{L}(m_{n},z)\,,$
We have defined $p=({p_{1}+p_{2}})/{2}$, $t=-K^{2}$, $\mathcal{V}(Q,z)\equiv
L(q=\sqrt{-Q^{2}},z)$ as given in (A.1.155), and used the bulk-to-bulk gauge
field propagator (A.1.2) with the substitutions $q\rightarrow k$,
$Q\rightarrow K$, and $V\rightarrow L$. We have also used the vertices in
(LABEL:vertices33), and defined
$\displaystyle
B^{\mu}(q,q^{\prime},\epsilon^{\pm})\equiv(-i)\epsilon^{\rho\sigma\nu\mu}\epsilon_{\rho}^{+}(q)\epsilon_{\sigma}^{-}(q^{\prime})(q^{\prime}_{\nu}+q_{\nu})\,.$
For $z^{\prime}\rightarrow 0$, we can use (A.1.159) and simplify (C) as
$\displaystyle i{\cal A}^{L}_{Lp\rightarrow
Lp}(s,t)\approx(-i)\mathcal{V}^{\mu}_{LLL}(q_{1},q_{2},k_{z})\times\big{(}-i\eta_{\mu\nu}\big{)}\times(-i)\mathcal{V}^{\nu}_{L\bar{\Psi}\Psi}(p_{1},p_{2},k_{z})\,,$
with
$\displaystyle\mathcal{V}^{\mu}_{LLL}(q_{1},q_{2},k_{z})=g_{5}^{3}\kappa_{CS}B^{\mu}(q,q^{\prime},\epsilon^{\pm})\int
dz\,\Big{(}\mathcal{V}(Q,z)\mathcal{V}(Q^{\prime},z)z-\partial_{z}\mathcal{V}(Q,z)\mathcal{V}(Q^{\prime},z)\frac{z^{2}}{2}\Big{)}\,,$
$\displaystyle\mathcal{V}^{\nu}_{L\bar{\Psi}\Psi}(p_{1},p_{2},k_{z})=g_{5}\times\frac{3}{2}\times\int
dz\,\sqrt{g}\,e^{-\phi}z\,\bar{\Psi}(p_{2},z)\gamma^{\nu}\Psi(p_{1},z)\mathcal{V}(K,z)=g_{5}F_{1}^{(LN)}(K)\,,$
where $\mathcal{V}(K,z)\equiv L^{0}(k=\sqrt{-K^{2}},z)$, and $F_{1}^{(LN)}(K)$
is the form factor of the the nucleon due to $L_{\mu}^{0}$.
The Reggeization of the bulk spin-1 gauge field $L_{\mu}^{0}(k,z)$ exchange
can be obtained, in a similar way to the Reggezation of the spin-2 graviton
exchange, through the substitution
$\displaystyle J_{L}(m_{n}(j),z)\rightarrow
z^{-(j-1)}\phi_{n}(j,z)=z^{-(j-1)}\frac{\tilde{\phi}_{n}(j,z)}{z}$ (C.187)
followed by the summation over all spin-j meson exchanges using the
Sommerfeld-Watson formula
$\displaystyle\frac{1}{2}\sum_{j\geq
1}(s^{j-1}+(-s)^{j-1})\rightarrow-\frac{\pi}{2}\int_{\mathbb{C}}\frac{dj}{2\pi
i}\left(\frac{s^{j-1}+(-s)^{j-1}}{{\rm sin}\,\pi j}\right)$
The contour ${\mathbb{C}}$ is to the left of all odd poles $j=1,3,...$ (in
contrast to the Reggeized graviton where the contour is chosen to the left of
the even poles), and requires the analytical continuation of the exchanged
amplitudes to the complex j-plane.
The spin-j normalized meson wavefunctions $J_{L}(m_{n}(j),z)$ (C.187) are
expressed in terms of the wavefunctions of massive scalar fields
$\tilde{\phi}_{n}(j,z)$ which are given, for the soft wall model, in terms of
the generalized Laguerre polynomials as
$\displaystyle\tilde{\phi}_{n}(j,z)=c_{n}(j)\,z^{\Delta(j)}L_{n}^{\Delta(j)-2}(w)\,,$
(C.189)
with $w=\tilde{\kappa}^{2}z^{2}$. The normalization coefficients are
$\displaystyle
c_{n}(j)=\Big{(}\frac{2\tilde{\kappa}^{2(\Delta(j)-1)}\Gamma(n+1)}{\Gamma(n+\Delta(j)-1)}\Big{)}^{\frac{1}{2}}\,,$
(C.190)
and the dimension of the massive scalar fields (with an additional mass coming
from the massive open string states attached to the D9 or D7-branes)
$\Delta(j)$ is given by
$\displaystyle\Delta(j)$ $\displaystyle=$ $\displaystyle
2+\sqrt{4+m^{2}R^{2}+\frac{R^{2}}{\alpha^{\prime}}(j-1)}$ (C.191)
$\displaystyle=$ $\displaystyle 2+\sqrt{\sqrt{\lambda}(j-j_{0})}\,,$
where, in the last line, we have used the fact that $m^{2}R^{2}=-3$. The
spin-1 transverse bulk gauge field defined as $zL_{\mu}^{0}(m_{n},z)$ obeys
the same bulk equation of motion as a bulk massive scalar field
$\tilde{\phi}_{n}(j=1,z)$ with $m^{2}R^{2}=-3$ which is manifest in (E.265).
We have also used the open string quantized mass spectrum
$m_{j}^{2}R^{2}=(j-1)({R^{2}}/{\alpha^{\prime}})=\sqrt{\lambda}(j-1)$ for open
strings attached to the D9 or D7-branes in bulk, and we have defined
$j_{0}=1-{1}/{\sqrt{\lambda}}$.
We now recall that the non-normalized bulk-to-boundary propagators of massive
scalar fields are given in terms of Kummer’s (confluent hypergeometric)
function of the second kind, and their integral representations are (for
space-like momenta $k^{2}=-K^{2}$)
$\displaystyle\mathcal{\tilde{V}}(j,K,z)=$ $\displaystyle
z^{\Delta(j)}U\Big{(}a_{K}+\frac{\Delta(j)}{2},\Delta(j)-1;w\Big{)}=z^{\Delta(j)}w^{2-\Delta(j)}U\Big{(}\tilde{a}(j),\tilde{b}(j);w\Big{)}$
$\displaystyle=$ $\displaystyle
z^{\Delta(j)}w^{2-\Delta(j)}\frac{1}{\Gamma(\tilde{a}(j))}\int_{0}^{1}dx\,x^{\tilde{a}(j)-1}(1-x)^{-\tilde{b}(j)}{\rm
exp}\Big{(}-\frac{x}{1-x}w\Big{)}\,,$ (C.192)
with $w=\tilde{\kappa}^{2}z^{2}$
$\displaystyle
a_{K}=a=\frac{K^{2}}{4\tilde{\kappa}^{2}}\qquad\tilde{a}(j)=a_{K}+2-\frac{\Delta(j)}{2}\qquad\tilde{b}(j)=3-\Delta(j)$
(C.193)
after using the identity $U(m,n;y)=y^{1-n}U(1+m-n,2-n,y)$. Therefore, the
bulk-to-bulk propagator of spin-j mesons
$\displaystyle
J_{L}(m_{n}(j),z)\,G(j,z,z^{\prime})\,J_{L}(m_{n}(j),z^{\prime})=z^{-(j-1)}\,G(j,z,z^{\prime})\,z^{\prime-(j-1)}$
can be approximated at the boundary as (for space-like momenta $k^{2}=-K^{2}$)
$\displaystyle G(j,z\rightarrow 0,z^{\prime})\approx$
$\displaystyle-\bigg{[}\frac{\phi_{n}(j,z\rightarrow
0)}{-g_{5}\mathcal{F}_{n}(j)}\bigg{]}\times\sum_{n}\frac{-g_{5}\mathcal{F}_{n}(j)\phi_{n}(j,z^{\prime})}{K^{2}+m_{n}^{2}(j)}$
$\displaystyle=$
$\displaystyle(\tilde{\kappa}^{2})^{\Delta(j)-2}\frac{z^{\Delta(j)-1}}{\Delta(j)-1}\frac{\Gamma(\Delta(j)-2+a)}{\Gamma(\Delta(j)-2)}\,\mathcal{V}(j,K,z^{\prime})$
where $\phi_{n}(j,z\rightarrow 0)=\frac{1}{z}\tilde{\phi}_{n}(j,z\rightarrow
0)$. We have defined the non-normalized bulk-to-boundary propagator of spin-j
mesons
$\displaystyle\mathcal{V}(j,K,z^{\prime})=\sum_{n}\frac{-g_{5}\mathcal{F}_{n}(j)\phi_{n}(j,z^{\prime})}{K^{2}+m_{n}^{2}(j)}=\frac{1}{z}\mathcal{\tilde{V}}(j,K,z^{\prime})|_{a+\frac{\Delta(j)}{2}\rightarrow
a+1+(\Delta(j)-3)}\,,$
with the shift $a+\frac{\Delta(j)}{2}\rightarrow a+1+(\Delta(j)-3)$ defined in
such a way that the mass spectrum of massive scalar fields
$m_{n}^{2}=4\tilde{\kappa}^{2}(n+\frac{\Delta(j=1)}{2})$ and the mass spectrum
of spin-1 gauge fields $m_{n}^{2}=4\tilde{\kappa}^{2}(n+1)$ match, i.e., we
shift $n+\frac{\Delta(j=1)}{2}\rightarrow n+1+(\Delta(j=1)-3)$, giving the
mass spectrum of spin-j mesons
$\displaystyle m_{n}^{2}(j)=4\tilde{\kappa}^{2}(n+1+(\Delta(j)-3))\,.$ (C.197)
We have also used
$\displaystyle\mathcal{F}_{n}(j)=$
$\displaystyle\frac{\mathcal{C}(j,K,\epsilon)}{g_{5}}\bigg{(}-\sqrt{g}\,e^{-\phi}\,\big{(}g^{xx}\big{)}^{2}\,\partial_{z^{\prime}}\phi_{n}(j,z^{\prime})\bigg{)}_{z^{\prime}=\epsilon}\,,$
$\displaystyle\mathcal{C}(j,K,\epsilon)=$
$\displaystyle\mathcal{V}(j,K,\epsilon)$ (C.198)
and the substitution $\phi_{n}(j,z\rightarrow
0)=\frac{1}{z}\tilde{\phi}_{n}(j,z\rightarrow 0)\approx
c_{n}(j)\,z^{\Delta(j)-1}L_{n}^{\Delta(j)-2}(0)$ for the soft wall model.
After the Reggeization, the scattering amplitude for the spin-j meson exchange
becomes
$\displaystyle i{\cal A}^{L}_{Lp\rightarrow
Lp}(j,s,t)\approx(-i)\mathcal{V}^{\mu}_{LLL}(j,q_{1},q_{2},k_{z})\times\big{(}-i\eta_{\mu\nu}\big{)}\times(-i)\mathcal{V}^{\nu}_{L\bar{\Psi}\Psi}(j,p_{1},p_{2},k_{z})\,,$
with
$\displaystyle\mathcal{V}^{\mu}_{LLL}(j,q_{1},q_{2},k_{z})=$
$\displaystyle\frac{1}{g_{5}^{2}}\times
g_{5}^{3}\kappa_{CS}B^{\mu}(q,q^{\prime},\epsilon^{\pm})\int dz\,z^{2(j-1)}$
$\displaystyle\times\bigg{(}\mathcal{V}(Q,z)\mathcal{V}(Q^{\prime},z)\times\frac{(\Delta(j)-1-(j-1))z^{\Delta(j)-1-(j-1)-1}}{\Delta(j)-1}$
$\displaystyle-\partial_{z}\mathcal{V}(Q,z)\mathcal{V}(Q^{\prime},z)\times\frac{z^{\Delta(j)-1-(j-1)}}{\Delta(j)-1}\bigg{)}\times(\tilde{\kappa}^{2})^{\Delta(j)-2}\frac{\Gamma(\Delta(j)-2+a)}{\Gamma(\Delta(j)-2)}$
$\displaystyle=$ $\displaystyle\mathcal{V}_{LLL}(j,Q,Q^{\prime})\times
B^{\mu}(q,q^{\prime},\epsilon^{\pm})\,,$
$\displaystyle\mathcal{V}^{\nu}_{L\bar{\Psi}\Psi}(p_{1},p_{2},k_{z})=$
$\displaystyle g_{5}\times\frac{3}{2}\times\int
dz\,\sqrt{g}\,e^{-\phi}z^{1+2(j-1)}\,\bar{\Psi}(p_{2},z)\gamma^{\nu}\Psi(p_{1},z)z^{-(j-1)}\mathcal{V}(j,K,z)$
$\displaystyle=$ $\displaystyle
g_{5}F_{1}^{(LN)}(j,K)\times\bar{u}(p_{2})\gamma^{\nu}u(p_{1})\,,$
We have defined
$\displaystyle\mathcal{V}_{LLL}(j,Q,Q^{\prime})=$
$\displaystyle\frac{1}{g_{5}^{2}}\times g_{5}^{3}\kappa_{CS}\int
dz\,z^{2(j-1)}$
$\displaystyle\times\bigg{(}\mathcal{V}(Q,z)\mathcal{V}(Q^{\prime},z)\times\frac{(\Delta(j)-1-(j-1))z^{\Delta(j)-1-(j-1)-1}}{\Delta(j)-1}$
$\displaystyle-\partial_{z}\mathcal{V}(Q,z)\mathcal{V}(Q^{\prime},z)\times\frac{z^{\Delta(j)-1-(j-1)}}{\Delta(j)-1}\bigg{)}(\tilde{\kappa}^{2})^{\Delta(j)-2}\frac{\Gamma(\Delta(j)-2+a)}{\Gamma(\Delta(j)-2)}$
$\displaystyle=$ $\displaystyle\frac{1}{g_{5}^{2}}\times
g_{5}^{3}\kappa_{CS}\times Q^{2-j-\Delta(j)}\times
I_{\xi}(j,Q,Q^{\prime})\times\tilde{\kappa}^{2\Delta(j)-4}\times\frac{\Gamma(\Delta(j)-2+a)}{\Gamma(\Delta(j)-2)}\times\frac{1}{\Delta(j)-1}$
with
$\displaystyle I_{\xi}(j,Q,Q^{\prime})$ $\displaystyle=$ $\displaystyle\int
d\xi\,\xi^{j-2+\Delta(j)}\bigg{(}\mathcal{V}(\xi)\mathcal{V}(\xi
Q^{\prime}/Q)\times(\Delta(j)-1-(j-1))\xi^{-1}-\partial_{\xi}\mathcal{V}(\xi)\mathcal{V}(\xi
Q^{\prime}/Q)\bigg{)}$ (C.204) $\displaystyle\approx$
$\displaystyle(\Delta(j)-1-(j-1))\times
2^{\Delta(j)+j-3}\,\frac{Q^{\prime}}{Q}\,G_{2,2}^{2,2}\left(\frac{Q^{2}}{Q^{\prime
2}}\bigg{|}\begin{array}[]{c}\frac{1}{2},\frac{3}{2}\\\
\frac{1}{2}(j+\Delta(j)-1),\frac{1}{2}(j+\Delta(j)+1)\\\ \end{array}\right)$
$\displaystyle+$ $\displaystyle
2^{\Delta(j)+j-2}\,\frac{Q^{\prime}}{Q}\,G_{2,2}^{2,2}\left(\frac{Q^{2}}{Q^{\prime
2}}\bigg{|}\begin{array}[]{c}\frac{1}{2},\frac{3}{2}\\\
\frac{1}{2}(j+\Delta(j)+1),\frac{1}{2}(j+\Delta(j)+1)\\\
\end{array}\right)\,,$ (C.207)
where $G_{p,q}^{m,n}\left(z\bigg{|}\begin{array}[]{c}a_{1},...,a_{p}\\\
b_{1},...,b_{q}\\\ \end{array}\right)$ is the Meijer G-function. We have used
the identities
$\displaystyle\lim_{\frac{Q}{\tilde{\kappa}}\rightarrow\infty}\mathcal{V}(\xi)=\xi
K_{1}(\xi)\qquad{\rm
and}\qquad\partial_{\xi}(\xi^{\nu}K_{\nu}(\xi))=-\xi^{\nu}K_{\nu-1}(\xi)$
(C.209)
to evaluate the integrals with $\xi=Qz$. The function $F_{1}^{(LN)}(j,K)$ in
(C) admits the integral representation
$\displaystyle F_{1}^{(LN)}(j,K)$ $\displaystyle=$
$\displaystyle\frac{3}{2}\times\frac{1}{2}\frac{\tilde{\kappa}^{-(j-1)-\Delta(j)-1}}{\Gamma(a)}\int_{0}^{1}dx\,x^{a-1}(1-x)^{-\tilde{b}(j)}$
$\displaystyle\times\bigg{(}\bigg{(}\frac{\tilde{n}_{R}}{\tilde{\kappa}^{\tau-1}}\bigg{)}^{2}\times\Gamma(c(j))\bigg{(}\frac{1}{1-x}\bigg{)}^{-c(j)}+\bigg{(}\frac{\tilde{n}_{L}}{\tilde{\kappa}^{\tau}}\bigg{)}^{2}\times\Gamma(c(j)+1)\bigg{(}\frac{1}{1-x}\bigg{)}^{-(c(j)+1)}\bigg{)}\,,$
$\displaystyle=$
$\displaystyle\frac{3}{2}\times\frac{1}{2}\tilde{\kappa}^{-(j-1)-\Delta(j)-1}$
$\displaystyle\times\bigg{(}\bigg{(}\frac{\tilde{n}_{R}}{\tilde{\kappa}^{\tau-1}}\bigg{)}^{2}\times\frac{\Gamma(c(j))\Gamma(1-\tilde{b}(j)+c(j))}{\Gamma(1-\tilde{b}(j)+c(j)+a)}+\bigg{(}\frac{\tilde{n}_{L}}{\tilde{\kappa}^{\tau}}\bigg{)}^{2}\times\frac{\Gamma(c(j)+1)\Gamma(2-\tilde{b}(j)+c(j))}{\Gamma(2-\tilde{b}(j)+c(j)+a)}\bigg{)}$
where
$\displaystyle\tilde{b}(j)=3-\Delta(j)\,,\qquad\qquad
c(j)=(\tau+1)+\frac{j-1}{2}-\frac{\Delta(j)}{2}-\frac{1}{2}\,.$ (C.211)
After summing over all contributions from the spin-j mesons, the total
amplitude ${\cal A}^{tot}_{Lp\rightarrow Lp}(s,t)$ is given by
$\displaystyle{\cal A}^{tot}_{Lp\rightarrow Lp}(s,t)=$
$\displaystyle-\int_{\mathbb{C}}\frac{dj}{2\pi
i}\left(\frac{s^{j-1}+(-s)^{j-1}}{{\rm sin}\,\pi j}\right){\cal
A}^{L}_{Lp\rightarrow Lp}(j,s,t)$ $\displaystyle{\cal A}^{L}_{Lp\rightarrow
Lp}(j,s,t)=$ $\displaystyle\mathcal{V}_{LLL}(j,Q,Q^{\prime})\times
B^{\mu}(q,q^{\prime},\epsilon^{\pm})\times
g_{5}F_{1}^{(LN)}(j,K)\times\bar{u}(p_{2})\gamma_{\mu}u(p_{1})\,,$ (C.212)
The contour $\mathbb{C}$ is at the rightmost of the branch-point of
$F_{1}^{LN}(j,K)$ and the leftmost of $j=1,3,...$. From (C), we determine the
single Reggeon amplitude (total amplitude) in momentum space, after wrapping
the j-plane contour ${\mathbb{C}}$ to the left,
$\displaystyle{\cal A}^{tot}_{Lp\rightarrow
Lp}(s,t)=-s^{j_{0}-1}\int_{-\infty}^{j_{0}}\frac{dj}{\pi}\left(\frac{1+e^{-i\pi}}{{\rm
sin}\,\pi j}\right)s^{j-j_{0}}\,\text{Im}[{\cal A}^{L}_{Lp\rightarrow
Lp}(j,s,t)]$ (C.213)
The imaginary part follows from the discontinuity of the $\Gamma$-function
$\displaystyle\text{Im}[{\cal A}^{L}_{Lp\rightarrow Lp}(j,s,t)]$
$\displaystyle\approx$
$\displaystyle\bigg{(}\Gamma(\Delta(j)-2)\mathcal{V}_{LLL}(j,Q,Q^{\prime})\times
B^{\mu}(q,q^{\prime},\epsilon^{\pm})\times
g_{5}F_{1}^{(LN)}(j,K)\times\bar{u}(p_{2})\gamma_{\mu}u(p_{1})\bigg{)}\bigg{|}_{j\rightarrow
j_{0},\Delta(j)\rightarrow 2}$ $\displaystyle\times$
$\displaystyle\text{Im}\bigg{[}\frac{1}{\Gamma(\tilde{\Delta}(j))}\bigg{]}$
with the complex argument
$\displaystyle\tilde{\Delta}(j)=\Delta(j)-2=i\sqrt{\sqrt{\lambda}(j_{0}-j)}\equiv
iy$
and $j_{0}=1-{1}/{\sqrt{\lambda}}$. For $y\rightarrow 0$, we may approximate
$1/\Gamma(iy)\approx iy\,e^{i\gamma y}$, with the Euler-Mascheroni constant
$\gamma=0.55772...$. The single Reggeon amplitude (total amplitude) in
momentum space (C) can now be cast in block form
$\displaystyle{\cal A}^{tot}_{Lp\rightarrow Lp}(s,t)=I(j_{0},s)\times
G_{5}(j_{0},s,t)$ (C.216)
with
$\displaystyle
I(j_{0},s)=-\tilde{s}^{j_{0}}\int_{-\infty}^{j_{0}}\frac{dj}{\pi}\left(\frac{1+e^{-i\pi}}{{\rm
sin}\,\pi
j}\right)\tilde{s}^{j-j_{0}}\,\sin\left[\tilde{\xi}\sqrt{\sqrt{\lambda}(j_{0}-j)}\right]$
$\displaystyle
G_{5}(j_{0},s,t)=\frac{1}{\tilde{s}}\bigg{(}\tilde{\kappa}^{2(j-1)}\Gamma(\Delta(j)-2)\mathcal{V}_{LLL}(j,Q,Q^{\prime})\times
B^{\mu}(q,q^{\prime},\epsilon^{\pm})\times
g_{5}F_{1}^{(LN)}(j,K)\times\bar{u}(p_{2})\gamma_{\mu}u(p_{1})\bigg{)}\bigg{|}_{j\rightarrow
j_{0},\,\Delta(j)\rightarrow 2}$
We have set $\tilde{s}\equiv{s}/{\tilde{\kappa}^{2}}$, and
$\tilde{\xi}-\pi/2=\gamma=0.55772.....$ is Euler-Mascheroni constant. We note
that the apparent pole in the Gamma-function at the Reggeon intercept, cancels
out in the combination $\Gamma(\Delta(j_{0})-2){\cal
V}_{LLL}(j_{0},Q,Q^{\prime})$.
In the block form (C.216), the spin-j integral $I(j_{0},s)$ is similar to the
spin-j integral in POLX (see Eq. 4.19), with the identifications
$\mathcal{K}(s,b^{\perp},z,z^{\prime})\leftrightarrow{\cal
A}^{tot}_{Lp\rightarrow Lp}(s,t)$,
$(zz^{\prime}/R^{4})G_{3}(j_{0},v)\leftrightarrow G_{5}(j_{0},s,t)$,
$\xi(v)\leftrightarrow\tilde{\xi}$, and $\widehat{s}\leftrightarrow\tilde{s}$.
We then follow POLX to evaluate the spin-j integral by closing the j-contour
appropriately. In the high energy limit
$\sqrt{\lambda}/\tilde{\tau}\rightarrow 0$
($\tilde{\tau}\equiv\log\tilde{s}$), the single Reggeon contribution to the
amplitude is
$\displaystyle{\cal A}^{tot}_{Lp\rightarrow Lp}(s,t)\simeq
e^{j_{0}\tilde{\tau}}\left[(\sqrt{\lambda}/\pi)+i\right](\sqrt{\lambda}/2\pi)^{1/2}\;\tilde{\xi}\;\frac{e^{-\sqrt{\lambda}\tilde{\xi}^{2}/2\tilde{\tau}}}{\tilde{\tau}^{3/2}}\left(1+{\cal
O}\bigg{(}\frac{\sqrt{\lambda}}{\tilde{\tau}}\bigg{)}\right)\times
G_{5}(j_{0},s,t)\,.$ (C.218)
We can rewrite the amplitude (C.218) as
$\displaystyle{\cal A}^{tot}_{Lp\rightarrow Lp}(s,t)$ $\displaystyle\simeq$
$\displaystyle 4\times 4\times g_{5}\times\frac{1}{g_{5}^{2}}\times
g_{5}^{3}\kappa_{CS}\times\Big{(}\frac{Q}{\tilde{\kappa}}\Big{)}^{2-j_{0}-\Delta(j_{0})}\times\Big{(}\frac{s}{\tilde{\kappa}^{2}}\Big{)}^{j_{0}}\times\frac{e^{-\sqrt{\lambda}\tilde{\xi}^{2}/2\log[s/\tilde{\kappa}^{2}]}}{(\log[s/\tilde{\kappa}^{2}])^{3/2}}$
$\displaystyle\times$
$\displaystyle\left[(\sqrt{\lambda}/\pi)+i\right]\times(\sqrt{\lambda}/2\pi)^{1/2}\;\tilde{\xi}\;\left(1+{\cal
O}\bigg{(}\frac{\sqrt{\lambda}}{\log[s/\tilde{\kappa}^{2}]}\bigg{)}\right)\times\tilde{G}_{5}(j_{0},t,Q,Q^{\prime})$
where
$\displaystyle\tilde{G}_{5}(j_{0},s,t)$ $\displaystyle\equiv$
$\displaystyle\frac{1}{4}\times\frac{1}{4}\times\frac{1}{g_{5}}\times\frac{1}{\frac{1}{g_{5}^{2}}\times
g_{5}^{3}\kappa_{CS}}\times\frac{1}{\tilde{\kappa}^{2}}\times\frac{1}{Q^{2-j-\Delta(j)}}\times\frac{1}{\kappa^{2\Delta(j)-4}}\times\frac{1}{\tilde{\kappa}^{-(j-1)-\Delta(j)-1}}\times\frac{1}{\tilde{\kappa}^{2(j-1)}}\times
G_{5}(j_{0},s,t)$ (C.220) $\displaystyle=$ $\displaystyle
I_{\xi}(j_{0},Q,Q^{\prime})\times\mathcal{F}_{1}^{(LN)}(j_{0},K)\times
s^{-1}\times\frac{1}{4}\times
B^{\mu}(q,q^{\prime},\epsilon^{\pm})\times\frac{1}{4}\times\bar{u}(p_{2})\gamma_{\mu}u(p_{1})$
with
$\displaystyle\mathcal{F}_{1}^{(LN)}(j_{0},K)\equiv\frac{\Gamma(\Delta(j_{0})-2+a)}{\Delta(j_{0})-1}\times\frac{1}{\tilde{\kappa}^{-(j_{0}-1)-\Delta(j_{0})-1}}\times
F_{1}^{(LN)}(j_{0},K)\,.$ (C.221)
## Appendix D Details of the Pomeron exchange
The transverse and traceless part of the graviton
($\eta_{\mu\nu}\rightarrow\eta_{\mu\nu}+h_{\mu\nu}$) follows from the
quadratic part of the Einstein-Hilbert action in de-Donder gauge,
$\displaystyle S=$ $\displaystyle\int
d^{5}x\sqrt{g}\,e^{-2\phi}\,\mathcal{L}_{h}\,,$
$\displaystyle\mathcal{L}_{h}=$
$\displaystyle-\frac{1}{4\tilde{g}_{5}^{2}}\,g^{\mu\nu}\,\eta^{\lambda\rho}\eta^{\sigma\tau}\partial_{\mu}h_{\lambda\sigma}\partial_{\nu}h_{\rho\tau}\,,$
(D.222)
with Newton constant $16\pi
G_{N}={8\pi^{2}}/{N_{c}^{2}}=\tilde{g}_{5}^{2}=2\kappa^{2}$. The massive
glueball spectrum is determined by solving the equation of motion for
$h_{\mu\nu}$ following from (D), with for spin-2 glueballs
$\displaystyle
m_{n}^{2}=8\tilde{\kappa}^{2}_{N}(n+1)\qquad\tilde{g}_{5}f_{n}=2\tilde{\kappa}_{N}$
(D.223)
### D.1 Graviton coupling in bulk
For the graviton in the axial gauge $h_{\mu z}=h_{zz}=0$. Using
$\eta_{\mu\nu}\rightarrow\eta_{\mu\nu}+h_{\mu\nu}$ in the linearized bulk
action gives
$\displaystyle h\overline{\Psi}\Psi:\quad$
$\displaystyle-\frac{\sqrt{2\kappa^{2}}}{2}\int
d^{5}x\,\sqrt{g}\,h_{\mu\nu}T_{F}^{\mu\nu}$ $\displaystyle hLL:\quad$
$\displaystyle-\frac{\sqrt{2\kappa^{2}}}{2}\int
d^{5}x\,\sqrt{g}\,h_{\mu\nu}T_{L}^{\mu\nu}$
with the energy-momentum tensors for the fermions and left gauge fields
$\displaystyle T_{F}^{\mu\nu}$ $\displaystyle=$ $\displaystyle
e^{-\phi}\frac{i}{2}\,z\,\overline{\Psi}\gamma^{\mu}\overset{\leftrightarrow}{\partial^{\nu}}\Psi-\eta^{\mu\nu}\mathcal{L}_{F}\,,$
$\displaystyle T_{L}^{\mu\nu}$ $\displaystyle=$
$\displaystyle-e^{-\phi}\Big{(}z^{4}\eta^{\rho\sigma}\eta^{\mu\beta}\eta^{\nu\gamma}\,F^{L}_{\beta\rho}F^{L}_{\gamma\sigma}-z^{4}\,\eta^{\mu\beta}\eta^{\nu\gamma}\,F^{L}_{\beta
z}F^{L}_{\gamma z}\Big{)}-\eta^{\mu\nu}\mathcal{L}_{L}\,.$ (D.225)
and the rescaling
$\displaystyle\Psi\rightarrow\sqrt{2g_{5}^{2}}\Psi\qquad L_{N}\rightarrow
g_{5}L_{N}\qquad h_{\mu\nu}\rightarrow\sqrt{2\kappa^{2}}\,h_{\mu\nu}\,.$
(D.226)
Evaluating the couplings or the vertices (LABEL:vertices1) on the solutions,
Fourier transforming the fields to momentum space, and integrating by part the
trace-full part for the fermions, we find for the couplings to the fermions
($h\overline{\Psi}\Psi$) to the left gauge fields ($hLL$)
$\displaystyle h\overline{\Psi}\Psi:\quad$
$\displaystyle\int\frac{d^{4}p_{2}d^{4}p_{1}d^{4}k}{(2\pi)^{12}}(2\pi)^{4}\delta^{4}(p_{2}-k-p_{1})\big{(}S^{k}_{h\bar{\Psi}\Psi}\big{)}$
$\displaystyle hLL:\quad$
$\displaystyle\int\frac{d^{4}q^{\prime}d^{4}qd^{4}k}{(2\pi)^{12}}(2\pi)^{4}\delta^{4}(q-k-q^{\prime})\big{(}S^{k}_{hLL}\big{)}$
with
$\displaystyle S^{k}_{h\bar{\Psi}\Psi}$ $\displaystyle=$
$\displaystyle-\frac{\sqrt{2\kappa^{2}}}{2}\int
dz\sqrt{g}\,e^{-\phi}z\,\epsilon^{TT}_{\mu\nu}h(k,z)\bar{\Psi}(p_{2},z)\gamma^{\mu}p^{\nu}\Psi(p_{1},z)\,,$
$\displaystyle S^{k}_{hLL}$ $\displaystyle=$
$\displaystyle\sqrt{2\kappa^{2}}\int
dz\sqrt{g}\,e^{-\phi}z^{4}\,\epsilon^{TT}_{\mu\nu}h(k,z)K^{\mu\nu}(q,q^{\prime},\epsilon,\epsilon^{\prime},z)\,.$
We have set $h_{\mu\nu}=\epsilon^{TT}_{\mu\nu}h(k,z)$ (where
$\epsilon^{TT}_{\mu\nu}$ is transverse and traceless polarization tensor),
$q^{2}=-Q^{2}$, $q^{\prime 2}=-Q^{\prime 2}$ for space-like momenta, and
defined
$\displaystyle K^{\mu\nu}(q,q^{\prime},\epsilon,\epsilon^{\prime},z)\equiv
B_{1}^{\mu\nu}\mathcal{V}(Q,z)\mathcal{V}(Q^{\prime},z)-B_{0}^{\mu\nu}\partial_{z}\mathcal{V}(Q,z)\partial_{z}\mathcal{V}(Q^{\prime},z)\,,$
$\displaystyle
B_{0}^{\mu\nu}(\epsilon,\epsilon^{\prime})\equiv\epsilon^{\mu}\epsilon^{\prime\nu}\,,$
$\displaystyle
B_{1}^{\mu\nu}(q,q^{\prime},\epsilon,\epsilon^{\prime})\equiv\epsilon\cdot\epsilon^{\prime}\,q^{\mu}q^{\prime\nu}-q\cdot\epsilon^{\prime}\,\epsilon^{\mu}q^{\prime\nu}-q^{\prime}\cdot\epsilon\,q^{\mu}\epsilon^{\prime\nu}+q\cdot
q^{\prime}\,\epsilon^{\mu}\epsilon^{\prime\nu}\,.$ (D.229)
with $B_{1,0}=\eta_{\mu\nu}B_{1,0}^{\mu\nu}$, $K=\eta_{\mu\nu}K^{\mu\nu}$, and
the non-normalizable wave function for the virtual photon $\mathcal{V}(Q,z)$
given in (LABEL:Gauge).
### D.2 Scattering amplitude
The t-channel Compton exchange of a spin-2 glueball of mass $m_{n}$ in AdS
reads
$\displaystyle i{\cal A}^{h}_{Lp\rightarrow Lp}(s,t)=\sum_{n}i\tilde{{\cal
A}}^{h}_{Lp\rightarrow Lp}(m_{n},s,t)$ $\displaystyle i\tilde{{\cal
A}}^{h}_{Lp\rightarrow
Lp}(m_{n},s,t)=(-i)V_{hLL}^{\mu\nu(TT)}(q,q^{\prime},k,m_{n})\times\tilde{G}^{TT}_{\mu\nu\alpha\beta}(m_{n},t)\times(-i)V_{h\bar{\Psi}\Psi}^{\alpha\beta(TT)}(p_{1},p_{2},k,m_{n})\,,$
with the bulk vertices ($k=p_{2}-p_{1}=q-q^{\prime}$)
$\displaystyle V_{hLL}^{\mu\nu(TT)}(q,q^{\prime},k,m_{n})\equiv$
$\displaystyle\left(\frac{\delta
S_{hLL}^{k}}{\delta(\epsilon^{TT}_{\mu\nu}h(k,z))}\right)\,J_{h}(m_{n},z)=\sqrt{2\kappa^{2}}\times\frac{1}{2}\int
dz\sqrt{g}\,e^{-\phi}z^{4}K^{\mu\nu}(q,q^{\prime},\epsilon,\epsilon^{\prime},z)J_{h}(m_{n},z)\,,$
$\displaystyle
V_{h\bar{\Psi}\Psi}^{\alpha\beta(TT)}(p_{1},p_{2},k,m_{n})\equiv$
$\displaystyle\left(\frac{\delta
S_{h\bar{\Psi}\Psi}^{k}}{\delta(\epsilon^{TT}_{\alpha\beta}h(k,z))}\right)\,J_{h}(m_{n},z)=-\sqrt{2\kappa^{2}}\times\int
dz\sqrt{g}\,e^{-\phi}z\bar{\Psi}(p_{2},z)\gamma^{\alpha}p^{\beta}\Psi(p_{1},z)J_{h}(m_{n},z)\,,$
with $p=({p_{1}+p_{2}})/{2}$. The bulk-to-bulk transverse and traceless
graviton propagator $G_{\mu\nu\alpha\beta}=G_{\mu\nu\alpha\beta}^{TT}$ for the
$2^{++}$ glueball is Raju:2011mp ; DHoker:1999bve
$\displaystyle
G_{\mu\nu\alpha\beta}^{TT}(m_{n},t,z,z^{\prime})=J_{h}(m_{n},z)\tilde{G}_{\mu\nu\alpha\beta}^{TT}(m_{n},t)J_{h}(m_{n},z^{\prime})\,,$
$\displaystyle\tilde{G}_{\mu\nu\alpha\beta}^{TT}(m_{n},t)={1\over
2}\left({\cal T}_{\mu\alpha}{\cal T}_{\nu\beta}+{\cal T}_{\mu\beta}{\cal
T}_{\nu\alpha}-\frac{2}{3}{\cal T}_{\mu\nu}{\cal
T}_{\alpha\beta}\right)\frac{i}{t-m_{n}^{2}+i\epsilon}\,,$
with
$\displaystyle{\cal T}_{\mu\nu}=$
$\displaystyle-\eta_{\mu\nu}+k_{\mu}k_{\nu}/m_{n}^{2}\,,$ $\displaystyle
J_{h}(m_{n},z)\equiv$
$\displaystyle\psi_{n}(z)=c_{n}\,z^{4}L_{n}^{\Delta(j)-2}(2\xi)$
and
$\displaystyle
c_{n}=\Bigg{(}\frac{2^{4}\tilde{\kappa}_{N}^{6}\Gamma(n+1)}{\Gamma(n+3)}\Bigg{)}^{\frac{1}{2}}\,,$
(D.233)
normalized according to
$\displaystyle\int
dz\,\sqrt{g}e^{-\phi}\,{\left|g^{xx}\right|}\,\psi_{n}(z)\psi_{m}(z)=\delta_{nm}\,.$
(D.234)
For $z^{\prime}\rightarrow 0$, we can simplify (D.2) as ($t=-K^{2}$),
$\displaystyle i{\cal A}^{h}_{Lp\rightarrow
Lp}(s,t)\approx(-i)\mathcal{V}^{\mu\nu(TT)}_{hLL}(q_{1},q_{2},k_{z})\times\bigg{(}\frac{i}{2}\eta_{\mu\alpha}\eta_{\nu\beta}\bigg{)}\times(-i)\mathcal{V}^{\alpha\beta(TT)}_{h\bar{\Psi}\Psi}(p_{1},p_{2},k_{z})\,,$
with
$\displaystyle\mathcal{V}^{\mu\nu(TT)}_{hLL}(q_{1},q_{2},k_{z})=\sqrt{2\kappa^{2}}\times\frac{1}{2}\int
dz\sqrt{g}\,e^{-\phi}z^{4}K^{\mu\nu}(q,q^{\prime},\epsilon,\epsilon^{\prime},z)\frac{z^{4}}{4}\,,$
$\displaystyle\mathcal{V}^{\alpha\beta(TT)}_{h\bar{\Psi}\Psi}(p_{1},p_{2},k_{z})=-\sqrt{2\kappa^{2}}\times\int
dz\,\sqrt{g}\,e^{-\phi}z\,\bar{\Psi}(p_{2},z)\gamma^{\mu}p^{\nu}\,\Psi(p_{1},z)\mathcal{H}(K,z)\,,$
where
$\displaystyle\mathcal{H}(K,z)$ $\displaystyle=\sum_{n}\frac{\sqrt{2}\kappa
F_{n}\psi_{n}(z^{\prime})}{K^{2}+m_{n}^{2}}$ (D.237)
$\displaystyle=4z^{4}\Gamma(a_{K}+2)U\Big{(}a_{K}+2,3;2\xi\Big{)}=\Gamma(a_{K}+2)U\Big{(}a_{K},-1;2\xi\Big{)}$
$\displaystyle=\frac{\Gamma(a_{K}+2)}{\Gamma(a_{K})}\int_{0}^{1}dx\,x^{a_{K}-1}(1-x){\rm
exp}\Big{(}-\frac{x}{1-x}(2\xi)\Big{)}\,,$
with $a_{K}={a}/{2}={K^{2}}/{8\tilde{\kappa}_{N}^{2}}$,
$\displaystyle F_{n}=\frac{1}{\sqrt{2}\kappa}\bigg{(}-\frac{1}{z^{\prime
3}}\partial_{z^{\prime}}\psi_{n}(z^{\prime})\bigg{)}_{z^{\prime}=\epsilon}=-\frac{4}{\sqrt{2}\kappa}c_{n}L_{n}^{2}(0)\,,$
We have used the transformation $U(m,n;y)=y^{1-n}U(1+m-n,2-n,y)$ in the second
line of (D.237).
### D.3 High energy limit
In the high energy limit $\sqrt{\lambda}/\tilde{\tau}\rightarrow 0$ with
$\tilde{\tau}\equiv\log\tilde{s}=\log[s/\tilde{\kappa}_{N}^{2}]$, the single
Pomeron (or spin-j gluballs) contribution to the Compton Scattering amplitude
has been evaluated in Mamo:2019mka , with the result
$\displaystyle{\cal A}^{tot}_{Lp\rightarrow Lp}(s,t)\simeq
e^{j_{0}\tilde{\tau}}\left[(\sqrt{\lambda}/\pi)+i\right](\sqrt{\lambda}/2\pi)^{1/2}\;\tilde{\xi}\;\frac{e^{-\sqrt{\lambda}\tilde{\xi}^{2}/2\tilde{\tau}}}{\tilde{\tau}^{3/2}}\left(1+{\cal
O}\bigg{(}\frac{\sqrt{\lambda}}{\tilde{\tau}}\bigg{)}\right)\times
G_{5}(j_{0},s,t,Q)$ (D.239)
with $\tilde{\xi}-\pi/2=\gamma=0.55772.....$ is Euler-Mascheroni constant, and
$\displaystyle G_{5}(j_{0},s,t,Q)=$
$\displaystyle\Big{(}\frac{\tilde{\kappa}_{N}}{\tilde{\kappa}_{V}}\Big{)}^{4-\Delta(j)+j-2}$
(D.240)
$\displaystyle\times\frac{1}{s^{2}}\bigg{(}\frac{1}{2}\tilde{\kappa}_{V}^{4-\Delta(j)+j-2}\Gamma(\Delta(j)-2)\Big{(}\mathcal{V}^{T}_{hLL}(j,Q,Q^{\prime})\times
B_{1}^{\alpha\beta}-\mathcal{V}^{L}_{hLL}(j,Q,Q^{\prime})\times
B_{0}^{\alpha\beta}\Big{)}$
$\displaystyle\times\frac{\sqrt{2\kappa^{2}}}{g_{5}^{2}}\times\tilde{\kappa}_{N}^{j-2+\Delta(j)}A(j,K)\bar{u}(p_{2})\gamma_{\alpha}p_{\beta}u(p_{1})\bigg{)}\bigg{|}_{j\rightarrow
j_{0},\,\Delta(j)\rightarrow 2}$
with, $Qz=\xi$,
$\displaystyle\mathcal{V}^{T}_{hLL}(j,Q,Q^{\prime})=$
$\displaystyle\frac{\sqrt{2\kappa^{2}}}{2}\int_{0}^{\infty}dz\sqrt{g}e^{-z^{2}\tilde{\kappa}_{V}^{2}}\,z^{4+2(j-2)}\times\mathcal{V}(Q,z)\times\mathcal{V}(Q^{\prime},z)\times
C(j)\times z^{\Delta(j)-(j-2)}$ $\displaystyle=$ $\displaystyle
Q^{4-(j+\Delta(j))}\times\frac{\sqrt{2\kappa^{2}}}{2}\int_{0}^{\infty}d\xi\,e^{-\xi^{2}\frac{\tilde{\kappa}_{V}^{2}}{Q^{2}}}\,\xi^{4+2(j-2)}\times\mathcal{V}(\xi)\times\mathcal{V}(\xi
Q^{\prime}/Q)\times C(j)\times\xi^{\Delta(j)-(j-2)}\,,$
$\displaystyle\mathcal{V}^{L}_{hLL}(j,Q,Q^{\prime})=$
$\displaystyle\frac{\sqrt{2\kappa^{2}}}{2}\int_{0}^{\infty}dz\sqrt{g}e^{-z^{2}\tilde{\kappa}_{V}^{2}}\,z^{4+2(j-2)}\times\partial_{z}\mathcal{V}(Q,z)\times\partial_{z}\mathcal{V}(Q^{\prime},z)\times
C(j)\times z^{\Delta(j)-(j-2)}$ $\displaystyle=$ $\displaystyle
Q^{6-(j+\Delta(j))}\times\frac{\sqrt{2\kappa^{2}}}{2}\int_{0}^{\infty}d\xi\,e^{-\xi^{2}\frac{\tilde{\kappa}_{V}^{2}}{Q^{2}}}\,\xi^{4+2(j-2)}\times\partial_{\xi}\mathcal{V}(\xi)\times\partial_{\xi}\mathcal{V}(\xi
Q^{\prime}/Q)\times C(j)\times\xi^{\Delta(j)-(j-2)}\,,$
and
$\displaystyle
A(j,K)=\frac{\tilde{\kappa}_{N}^{-(j-2)-\Delta(j)}}{2}\,\frac{\Gamma(c)\Gamma(1-\tilde{b}+c)}{\Gamma(1-\tilde{b}+c+\tilde{a})}$
$\displaystyle\times\bigg{(}\bigg{(}\frac{\tilde{n}_{R}}{\tilde{\kappa}_{N}^{\tau-1}}\bigg{)}^{2}\,_{2}F_{1}(\tilde{a},c+1,1-\tilde{b}+c+\tilde{a},-1)+\bigg{(}\frac{\tilde{n}_{L}}{\tilde{\kappa}_{N}^{\tau}}\bigg{)}^{2}\frac{c(1-\tilde{b}+c)}{1-\tilde{b}+c+\tilde{a}}\,_{2}F_{1}(\tilde{a}+1,c+1,2-\tilde{b}+c+\tilde{a},-1)\bigg{)}\,.$
The parameters are fixed as
$\displaystyle 1-\tilde{b}+c=(\tau-1)+\frac{j-2}{2}+\frac{\Delta(j)}{2}$
$\displaystyle 1-\tilde{b}+c+\tilde{a}=(\tau+1)+\frac{j-2}{2}+a_{K}$
$\displaystyle c=(\tau+1)+\frac{j-2}{2}-\frac{\Delta(j)}{2}$
$\displaystyle\tilde{n}_{R}=\tilde{n}_{L}\tilde{\kappa}_{N}^{-1}\sqrt{\tau-1}\qquad\tilde{n}_{L}=\tilde{\kappa}_{N}^{\tau}\sqrt{{2}/{\Gamma(\tau)}}$
and
$\displaystyle
C(j)=\tilde{\kappa}_{V}^{2\Delta(j)-4}\times\frac{1}{\Delta(j)}\frac{2^{\Delta(j)-2}\Gamma(a_{K}+\frac{\Delta(j)}{2})}{\Gamma(\Delta(j)-2)}\qquad$
$\displaystyle\Delta(j)=2+\sqrt{2\sqrt{\lambda}(j-j_{0})}\qquad{\rm and}\qquad
a_{K}=\frac{a}{2}=\frac{K^{2}}{8{\tilde{\kappa}}^{2}}\qquad{\rm and}\qquad
j_{0}=2-\frac{2}{\sqrt{\lambda}}\,.$ (D.244)
We can rewrite $G_{5}(j_{0},s,t,Q,Q^{\prime})$ of (D.240) more compactly as
$\displaystyle
G_{5}(j_{0},s,t,Q,Q^{\prime})=\frac{2\kappa^{2}}{g_{5}^{2}}\times\Big{(}\frac{Q}{\tilde{\kappa}}\Big{)}^{2-(j+\Delta(j))}$
$\displaystyle\times\frac{1}{s^{2}}\mathcal{F}(j,K)\bigg{(}I_{\xi}^{T}(j,Q,Q^{\prime})\times
B_{1}^{\alpha\beta}p_{\alpha}p_{\beta}-I_{\xi}^{L}(j,Q,Q^{\prime})\times
B_{0}^{\alpha\beta}p_{\alpha}p_{\beta}Q^{2}\bigg{)}\bigg{|}_{j\rightarrow
j_{0},\,\Delta(j)\rightarrow 2}$
where we have set $\tilde{\kappa}_{V}=\tilde{\kappa}_{N}=\tilde{\kappa}$, and
defined the dimensionless functions
$\displaystyle\mathcal{F}(j,K)$ $\displaystyle\equiv$
$\displaystyle\tilde{\kappa}^{j+2-\Delta(j)}\times\Gamma(\Delta(j)-2)\times
C(j,K)\times A(j,K)\,,$ $\displaystyle I_{\xi}^{T}(j,Q,Q^{\prime})$
$\displaystyle\equiv$
$\displaystyle\frac{1}{2}\int_{0}^{\infty}d\xi\,e^{-\xi^{2}\frac{\tilde{\kappa}_{V}^{2}}{Q^{2}}}\,\xi^{\Delta(j)+j+2}\times\mathcal{V}(\xi)\times\mathcal{V}(\xi
Q^{\prime}/Q)\,,$ (D.248) $\displaystyle\approx$
$\displaystyle\frac{1}{2}\times
2^{\Delta(j)+j+2}\,\frac{Q^{\prime}}{Q}\,G_{2,2}^{2,2}\left(\frac{Q^{2}}{Q^{\prime
2}}\bigg{|}\begin{array}[]{c}\frac{1}{2},\frac{3}{2}\\\
\frac{1}{2}(j+\Delta(j)+4),\frac{1}{2}(j+\Delta(j)+6)\\\
\end{array}\right)\,,$ $\displaystyle I_{\xi}^{L}(j,Q,Q^{\prime})$
$\displaystyle\equiv$
$\displaystyle\frac{1}{2}\int_{0}^{\infty}d\xi\,e^{-\xi^{2}\frac{\tilde{\kappa}_{V}^{2}}{Q^{2}}}\,\xi^{\Delta(j)+j+2}\times\partial_{\xi}\mathcal{V}(\xi)\times\partial_{\xi}\mathcal{V}(\xi
Q^{\prime}/Q)$ (D.251) $\displaystyle\approx$ $\displaystyle\frac{1}{2}\times
2^{\Delta(j)+j+2}\,\frac{Q^{\prime
2}}{Q^{2}}\,G_{2,2}^{2,2}\left(\frac{Q^{2}}{Q^{\prime
2}}\bigg{|}\begin{array}[]{c}1,1\\\
\frac{1}{2}(j+\Delta(j)+5),\frac{1}{2}(j+\Delta(j)+5)\\\
\end{array}\right)\,,$
using the identities (C.209).
We can also rewrite the amplitude (D.239) as
$\displaystyle{\cal A}^{tot}_{Lp\rightarrow Lp}(s,t)$ $\displaystyle\simeq$
$\displaystyle\frac{2\kappa^{2}}{g_{5}^{2}}\times\Big{(}\frac{Q}{\tilde{\kappa}}\Big{)}^{2+2/\sqrt{\lambda}}\times\Big{(}\frac{s}{\tilde{\kappa}^{2}}\Big{)}^{-2/\sqrt{\lambda}}\times\frac{e^{-\sqrt{\lambda}\tilde{\xi}^{2}/2\log[s/\tilde{\kappa}^{2}]}}{(\log[s/\tilde{\kappa}^{2}])^{3/2}}$
$\displaystyle\times$
$\displaystyle\left[(\sqrt{\lambda}/\pi)+i\right]\times(\sqrt{\lambda}/2\pi)^{1/2}\;\tilde{\xi}\;\left(1+{\cal
O}\bigg{(}\frac{\sqrt{\lambda}}{\log[s/\tilde{\kappa}^{2}]}\bigg{)}\right)\times\tilde{G}_{5}(j_{0},t,Q,Q^{\prime})$
where we have explicitly used $\tilde{\tau}=\log[s/\tilde{\kappa}^{2}]$,
$j_{0}=2-\frac{2}{\sqrt{\lambda}}$, $\Delta(j_{0})=2$,
$\tilde{\xi}-\pi/2=\gamma=0.55772.....$ is Euler-Mascheroni constant, and
defined
$\displaystyle\tilde{G}_{5}(j_{0},s,t,Q,Q^{\prime})\equiv\frac{1}{Q^{4}}\times\frac{g_{5}^{2}}{2\kappa^{2}}\times\Big{(}\frac{\tilde{\kappa}}{Q}\Big{)}^{2-(j_{0}+\Delta(j_{0}))}\times
s^{2}\times G_{5}(j_{0},s,t,Q,Q^{\prime})$ (D.253)
For small-x, we have $s\simeq{Q^{2}}/{x}$, we can rewrite the amplitude (D.3)
in terms of $x$ as
$\displaystyle{\cal A}^{tot}_{Lp\rightarrow Lp}(x,Q,t)$ $\displaystyle\simeq$
$\displaystyle\frac{1}{2}\times\frac{2\kappa^{2}}{g_{5}^{2}}\times\Big{(}\frac{Q}{\tilde{\kappa}}\Big{)}^{2-2/\sqrt{\lambda}}\times\Big{(}\frac{1}{x}\Big{)}^{1-2/\sqrt{\lambda}}\times\frac{e^{-\sqrt{\lambda}\tilde{\xi}^{2}/2(\log[Q^{2}/\tilde{\kappa}^{2}]+\log[1/x])}}{(\log[Q^{2}/\tilde{\kappa}^{2}]+\log[1/x])^{3/2}}$
$\displaystyle\times$
$\displaystyle\left[(\sqrt{\lambda}/\pi)+i\right]\times(\sqrt{\lambda}/2\pi)^{1/2}\;\tilde{\xi}\;\left(1+{\cal
O}\bigg{(}\frac{\sqrt{\lambda}}{\log[Q^{2}/\tilde{\kappa}^{2}]+\log[1/x]}\bigg{)}\right)\times\tilde{\tilde{G}}_{5}(j_{0},x,t,Q,Q^{\prime})$
where we have defined $\tilde{\tilde{G}}_{5}(j_{0},x,t,Q,Q^{\prime})\equiv
2x\,\tilde{G}_{5}(j_{0},s,t,Q,Q^{\prime})$ with ($\epsilon\cdot q=0$)
$\displaystyle\tilde{\tilde{G}}_{5}(j_{0},x,t,Q,Q^{\prime})=\mathcal{F}(j,K)\bigg{(}I_{\xi}^{T}(j,Q,Q^{\prime})\times\Big{(}\frac{1}{2x}\epsilon^{2}-\frac{2x}{Q^{2}}(\epsilon\cdot
p)^{2}\Big{)}-I_{\xi}^{L}(j,Q,Q^{\prime})\times\frac{2x}{Q^{2}}(\epsilon\cdot
p)^{2}\bigg{)}\bigg{|}_{j\rightarrow j_{0},\,\Delta(j)\rightarrow 2}\,,$
## Appendix E Operator Product Expansion
The parton model emerges in QCD through a leading twist contribution to the
structure functions. The twist expansion follows from the operator product
expansion (OPE) of the JJ currents. We now illustrate this expansion to
leading order for the charged current contributions in T-ordered product in
(IV.92). Specifically, we have as $x\rightarrow 0$
$\displaystyle T^{*}\left(J^{W^{-}}_{\mu}(x)J^{W^{+}}_{\nu}(0)\right)\approx$
$\displaystyle
e_{W}^{2}\left(\overline{q}(x)T^{-}T^{+}\gamma_{\mu}\frac{1}{2}(1-\gamma_{5})\,S(x)\,\gamma_{\nu}\frac{1}{2}(1-\gamma_{5})q(0)+\overline{q}(0)T^{+}T^{-}\gamma_{\nu}\frac{1}{2}(1-\gamma_{5})\,S(-x)\,\gamma_{\mu}\frac{1}{2}(1-\gamma_{5})q(x)\right)$
with $S(x)=2i\gamma\cdot x/(2\pi x^{2})^{2}$. With the help of the identity
$\displaystyle\gamma_{\mu}\gamma\cdot
x\gamma_{\nu}=\left(S_{\mu\nu\alpha\beta}+i\epsilon_{\mu\nu\alpha\beta}\gamma_{5}\right)x^{\alpha}\gamma^{\beta}$
(E.257)
with the symmetric tensor
$\displaystyle
S_{\mu\nu\alpha\beta}=\eta_{\mu\alpha}\eta_{\nu\beta}+\eta_{\mu\beta}\eta{\nu\alpha}-\eta_{\mu\nu}\eta_{\alpha\beta}$
(E.258)
in (E), the short distance contribution to the T-ordered product in (IV.92) is
$\displaystyle T_{\mu\nu}^{-+}\approx
2e_{W}^{2}\frac{q^{\alpha}}{q^{2}}\left<P\right|\bigg{(}S_{\mu\nu\alpha\beta}\overline{q}(0)\tau^{3}\gamma^{\beta}(1-\gamma_{5})q(0)-i\epsilon_{\mu\nu\alpha\beta}\overline{q}(0)\gamma^{\beta}(1-\gamma_{5})q(0)\bigg{)}\left|P\right>$
(E.259)
For unpolarized scattering, a comparison of (E.259) to (IV.92) suggests that
the parity odd structure function $F_{3}$ can be identified with the
antisymmetric tensor contribution,
$\displaystyle F_{3}(x,Q^{2})P^{\beta}\approx(2e_{W})^{2}{\rm
Im}\bigg{(}\frac{1}{x}\left<P\right|\overline{q}\gamma^{\beta}(1-\gamma_{5})q\left|P\right>\bigg{)}$
(E.260)
(E.260) is suggestive of a $j=1$ exchange through the left singlet current,
since $s^{j=1}\leftrightarrow 1/x$.
Although the twist-2 contribution to the OPE is real, it provides the relevant
starting point for the Reggeization by summing over higher spin-j states in
holography HATTAX . For that, we first note that the holographic dual of the
singlet current form factor is
$\displaystyle\left<P_{X}\right|\overline{q}\gamma^{\beta}(1-\gamma_{5})q\left|P\right>=$
$\displaystyle\left[e_{N}\overline{u}_{N}(P_{X})\gamma^{\beta}(1-\gamma^{5})u_{N}(P)\right]$
(E.261) $\displaystyle\times\int dz\sqrt{g}e^{-{\phi}}\,{{\cal
V}(Q,z)}\,\frac{z}{R}\,\bigg{[}\frac{1}{2}\bigg{(}\frac{z^{2}}{R^{2}}\bigg{)}^{2}\,\tilde{f}_{L}^{n_{x}}(z)\tilde{f}_{L}^{0}(z)+\frac{1}{2}\bigg{(}\frac{z^{2}}{R^{2}}\bigg{)}^{2}\,\tilde{f}_{R}^{n_{x}}(z)\tilde{f}_{R}^{0}(z)\bigg{]}$
The bulk-to-boundary propagator ${\cal V}(Q,z^{\prime})$ relates to the bulk-
to-bulk propagator $G(Q,z,z^{\prime})$ through
$\displaystyle\lim_{z\to 0}\frac{2}{z^{\prime 2}}\,G_{1}(Q,z^{\prime},z)={\cal
V}(Q,z)$ (E.262)
Using (E.262) into (E.261) gives
$\displaystyle\left<P_{X}\right|\overline{q}\gamma^{\beta}(1-\gamma_{5})q\left|P\right>=$
$\displaystyle\left[e_{N}\overline{u}_{N}(P_{X})\gamma^{\beta}(1-\gamma^{5})u_{N}(P)\right]\,$
(E.263) $\displaystyle\times\lim_{z^{\prime}\to 0}\frac{2}{z^{\prime 2}}\int
dz\sqrt{g}e^{-{\phi}}\,G_{1}(Q,z^{\prime},z)\,$
$\displaystyle\times\frac{z}{R}\,\bigg{[}\frac{1}{2}\bigg{(}\frac{z^{2}}{R^{2}}\bigg{)}^{2}\,\tilde{f}_{L}^{n_{x}}(z)\tilde{f}_{L}^{0}(z)+\frac{1}{2}\bigg{(}\frac{z^{2}}{R^{2}}\bigg{)}^{2}\,\tilde{f}_{R}^{n_{x}}(z)\tilde{f}_{R}^{0}(z)\bigg{]}$
### E.1 Hard wall
For the hard wall model with $\phi=0$, the bulk-to-bulk propagator can be
readily constructed
$\displaystyle
G_{1}(Q,z^{\prime},z)=zz^{\prime}\bigg{(}I_{0}(Qz_{0})K_{1}(Qz_{>})+K_{0}(Qz_{0})I_{1}(Qz_{>})\bigg{)}\frac{I_{1}(Qz_{<})}{I_{0}(Qz_{0})}$
(E.264)
and (E.262) explicitly checked. However, for the Reggeization it is more
useful to recall that the bulk-to-bulk propagator for the $U_{L}(1)$ vector
field, obeys the Green′s equation in warped space ($Q^{2}=-q^{2}$)
$\displaystyle\bigg{(}\bigg{(}\Delta_{j=1}-z^{2}Q^{2}+m_{j=1}^{2}\bigg{)}=\bigg{(}-z^{2}(\partial_{z}^{2}+Q^{2})+z\partial_{z}\bigg{)}\bigg{)}G_{j=1}(Q,z^{\prime},z)=\frac{\delta(z-z^{\prime})}{\sqrt{g}}$
(E.265)
with $m^{2}_{j=1}=-3$. Using the open-string Regge trajectory
$\displaystyle j=1+\alpha^{\prime}(m^{2}_{j}-m_{1}^{2})\qquad{\rm
with}\qquad\alpha^{\prime}=l_{s}^{2}=1/\sqrt{\lambda}$ (E.266)
(E.265) generalizes to spin-j
$\displaystyle\bigg{(}\Delta_{j}-z^{2}Q^{2}+m_{j}^{2}\bigg{)}G_{j}(Q,z^{\prime},z)=\frac{\delta(z-z^{\prime})}{\sqrt{g}}$
(E.267)
with the recursive relation for the warped Laplacian-like
$\displaystyle\Delta_{j}=z^{1-j}\,\Delta_{1}\,z^{j-1}$ (E.268)
(E.267) can be formally inverted
$\displaystyle{\sqrt{g}}\,G_{j}(Q,z^{\prime},z)=\frac{1}{(\Delta_{j}-z^{2}Q^{2}+m_{j}^{2})}\,{\delta(z-z^{\prime})}=z^{2-j}\,\frac{1}{(\Delta_{2}-z^{2}Q^{2}+m_{j}^{2})}\,z^{j-2}\,{\delta(z-z^{\prime})}$
(E.269)
Changing to the conformal variable $z^{2}=e^{-\rho}$, noting that
$\Delta_{2}=-4\partial_{\rho}^{2}+4$ and using the plane-wave identity
$\displaystyle
z^{\prime}\delta(z-z^{\prime})=\bigg{(}\frac{z^{\prime}}{z}\bigg{)}^{j-2}\int\frac{d\nu}{\pi}\,e^{i\nu(\rho-\rho^{\prime})}$
(E.270)
we can recast (E.269) in the form
$\displaystyle{\sqrt{g}}\,G_{j}(0,z^{\prime},z)=\sqrt{g^{\prime}}(g^{\prime
xx})^{j}(zz^{\prime})^{2-j}\int\frac{d\nu}{\pi}\,\frac{1}{4\nu^{2}+4+m_{j}^{2}}\,e^{i\nu(\rho-\rho^{\prime})}$
(E.271)
for $Q=0$. The Reggeized form of the spin-j and twist-2 extension of (E.263)
is
$\displaystyle\sum_{j}\,\frac{1}{x^{j}}\,\left<P\right|\overline{q}\gamma^{\beta}\,\partial^{j-1}\,(1-\gamma_{5})q\left|P\right>=\int_{\mathbb{C}}\frac{dj}{4i}\frac{1-e^{-i\pi
j}}{{\rm sin}\pi
j}\frac{1}{x^{j}}\,\left[e_{N}\overline{u}_{N}(P)\gamma^{\beta}\,\partial^{j-1}\,(1-\gamma^{5})u_{N}(P)\right]$
$\displaystyle\times\lim_{z^{\prime}\to 0}\frac{2}{z^{\prime 2}}\int
dz\sqrt{g^{\prime}}(g^{\prime
xx})^{j}(zz^{\prime})^{2-j}\int\frac{d\nu}{\pi}\,\frac{1}{4\nu^{2}+1+\sqrt{\lambda}(j-1)}\,e^{i\nu(\rho-\rho^{\prime})}\,\frac{z}{R}\,\bigg{[}\frac{1}{2}\psi_{L}^{2}(z)+\frac{1}{2}\psi_{R}^{2}(z)\bigg{]}\,$
using the open string Regge trajectory (E.266) in the forward limit ($Q=0$).
Here $\psi_{L}(z)$ is the lowest left-chirality bulk fermionic wavefunction
for the hard wall. The contour $\mathbb{C}$ is to the left-most of the poles
$j=1,3,...$ and to the right of the pole $j_{0}=1-1/\sqrt{\lambda}$. Undoing
the contour integration-${\mathbb{C}}$ by closing to the left and picking the
single pole $j_{0}$, and then performing the $\nu$-integration yield
$\displaystyle{\rm Im}\,\sum^{\rm
odd}_{j}\,\frac{1}{x^{j}}\,\left<P\right|\overline{q}\gamma^{\beta}\,\partial^{j-1}\,(1-\gamma_{5})q\left|P\right>=\frac{1}{x^{j_{0}}}\,\left[e_{N}\overline{u}_{N}(P)\gamma^{\beta}\,\partial^{j_{0}-1}\,(1-\gamma^{5})u_{N}(P)\right]$
$\displaystyle\times\frac{\pi}{2\sqrt{\lambda}}\lim_{z^{\prime}\to
0}\frac{2}{z^{\prime 2}}\int dz\sqrt{g^{\prime}}(g^{\prime
xx})^{j_{L}}(zz^{\prime})^{2-j_{0}}\,\frac{e^{-(\rho-\rho^{\prime})/4D\chi}}{\sqrt{\pi
D\chi}}\,\frac{z}{R}\,\bigg{[}\frac{1}{2}\psi_{L}^{2}(z)+\frac{1}{2}\psi_{R}^{2}(z)\bigg{]}\,$
The Gribov time is $\chi={\rm ln}(1/x)$ and the diffusion constant of the
Reggeon is $D=4/\sqrt{\lambda}$. (E.1) fixes the odd structure function in
(E.260) in the forward direction using this semi-quantitative OPE argument,
$\displaystyle
F_{3}(0,x)\approx\frac{1}{x^{j_{0}}}\approx\frac{1}{x^{1-1/\sqrt{\lambda}}}$
(E.274)
### E.2 Soft wall
For the soft-wall model, the Reggeized current form factor is given by
$\displaystyle\sum^{\rm
odd}_{j}\,\frac{1}{x^{j}}\,\left<P\right|\overline{q}\gamma^{\beta}\,\partial^{j-1}\,(1-\gamma_{5})q\left|P\right>=$
$\displaystyle-x^{j_{0}}\int_{-\infty}^{j_{0}}\frac{dj}{\pi}\left(\frac{1+e^{-i\pi}}{{\rm
sin}\,\pi
j}\right)x^{j-j_{0}}\,\text{Im}\left[\,2\times\frac{2}{3}\times\tilde{\kappa}^{(j-1)+\Delta(j)+1}\times\frac{\Gamma(\Delta(j)-2+a)}{\Gamma(\Delta(j)-2)}\times\frac{1}{g_{5}}\times\mathcal{V}^{\beta}_{L\bar{\Psi}\Psi}(p_{1}=p_{2}=p,k_{z}=0)\right]$
$\displaystyle=$
$\displaystyle-x^{j_{0}}\int_{-\infty}^{j_{0}}\frac{dj}{\pi}\left(\frac{1+e^{-i\pi}}{{\rm
sin}\,\pi
j}\right)x^{j-j_{0}}\,\text{Im}\left[\,2\times\frac{2}{3}\times\tilde{\kappa}^{(j-1)+\Delta(j)+1}\times\frac{\Gamma(\Delta(j)-2+a)}{\Gamma(\Delta(j)-2)}\times\frac{1}{g_{5}}\times
g_{5}F_{1}^{(LN)}(j,K=0)\times\bar{u}(p)\gamma^{\beta}u(p)\right]$
where $\mathcal{V}^{\beta}_{L\bar{\Psi}\Psi}(p_{1},p_{2},k_{z})$ is given by
(C), and $F_{1}^{(LN)}(j,K)$ is given by (C). Note that the bracket
$\displaystyle\bigg{[}2\times\frac{2}{3}\times\tilde{\kappa}^{(j-1)+\Delta(j)+1}\times\frac{\Gamma(\Delta(j)-2+a)}{\Gamma(\Delta(j)-2)}\times\frac{1}{g_{5}}\times
g_{5}F_{1}^{(LN)}(j,K)\bigg{]}$ (E.276)
is the spin-j form factor which reduces to the spin-1 form factor for $j=1$,
for the current operator
$2\times\tilde{J}_{L}^{\beta}(0)=\overline{q}\gamma^{\beta}(1-\gamma_{5})q$
sourced by $\frac{1}{2}\times\frac{3}{2}\times L_{\beta}^{0}(K,z\rightarrow
0)$ at the boundary. Also note that the momentum transfer is $k_{z}\equiv
q_{z}$ and that $-k^{2}=K^{2}\equiv Q^{2}$ with
$a={K^{2}}/{4\tilde{\kappa}^{2}}\equiv{Q^{2}}/{4\tilde{\kappa}^{2}}$. The
momentum of the in-coming nucleon is $p_{1}=p$, and the momentum of the out-
going nucleon is $p_{2}=p$, with $k=p_{2}-p_{1}\equiv q$.
Following the reasoning in Appendix C, we can evaluate the integral in (E.2)
with the result
$\displaystyle{\rm Im}\sum^{\rm
odd}_{j}\,\frac{1}{x^{j}}\,\left<P\right|\overline{q}\gamma^{\beta}\,\partial^{j-1}\,(1-\gamma_{5})q\left|P\right>\simeq$
$\displaystyle
e^{j_{0}\tau_{x}}\left[0\times(\sqrt{\lambda}/\pi)+i/i\right](\sqrt{\lambda}/2\pi)^{1/2}\;\tilde{\xi}\;\frac{e^{-\sqrt{\lambda}\tilde{\xi}^{2}/2\tau_{x}}}{\tau_{x}^{3/2}}\left(1+{\cal
O}\bigg{(}\frac{\sqrt{\lambda}}{\tau_{x}}\bigg{)}\right)\times\mathcal{G}_{5}(j_{0},x,Q=0)$
with
$\displaystyle\mathcal{G}_{5}(j_{0},x,Q=0)=$
$\displaystyle\bigg{(}\Gamma(\Delta(j)-2)\times
2\times\frac{2}{3}\times\tilde{\kappa}^{(j-1)+\Delta(j)+1}\times\frac{\Gamma(\Delta(j)-2+a)}{\Gamma(\Delta(j)-2)}\times\frac{1}{g_{5}}\times
g_{5}F_{1}^{(LN)}(j,Q)\times 2P^{\beta}\bigg{)}\bigg{|}_{j\rightarrow
j_{0},\,\Delta(j)\rightarrow 2,Q\rightarrow 0}$
Again, $j_{0}=1-\frac{1}{\sqrt{\lambda}}$, $\tau_{x}=\log[1/x]$,
$\bar{u}(p)\gamma^{\beta}u(p)=2p^{\beta}=2P^{\beta}$, and
$\tilde{\xi}-\pi/2=\gamma=0.55772.....$ is the Euler-Mascheroni constant.
Finally, comparing (E.2) to (E.260), we find
$\displaystyle
F_{3}(0,x)\approx\frac{1}{x^{1-1/\sqrt{\lambda}}}\times(\sqrt{\lambda}/2\pi)^{1/2}\;\tilde{\xi}\;\frac{e^{-\sqrt{\lambda}\tilde{\xi}^{2}/2\tau_{x}}}{\tau_{x}^{3/2}}\left(1+{\cal
O}\bigg{(}\frac{\sqrt{\lambda}}{\tau_{x}}\bigg{)}\right)\times\tilde{\mathcal{G}}_{5}(j_{0},x,0)$
(E.279)
with
$\displaystyle\tilde{\mathcal{G}}_{5}(j_{0},x,0)=$
$\displaystyle\bigg{(}\Gamma(\Delta(j)-2)\times
2\times\frac{2}{3}\times\tilde{\kappa}^{(j-1)+\Delta(j)+1}\times\frac{\Gamma(\Delta(j)-2+a)}{\Gamma(\Delta(j)-2)}\times\frac{1}{g_{5}}\times
g_{5}F_{1}^{(LN)}(j,Q)\bigg{)}\bigg{|}_{j\rightarrow
j_{0},\,\Delta(j)\rightarrow 2,Q\rightarrow 0}\,.$
## Appendix F Trace of Gamma matrices
Note that the Dirac traces do not depend on the specific form of the
$\gamma^{0},\gamma^{1},\gamma^{2},\gamma^{3}$ matrices but are completely
determined by the Clifford algebra
$\displaystyle\\{\gamma^{\mu},\gamma^{\nu}\\}\ \equiv\
\gamma^{\mu}\gamma^{\nu}\,+\,\gamma^{\nu}\gamma^{\mu}\ =\ 2\eta^{\mu\nu}\,,$
(F.281)
and some useful identities for carrying some of the Dirac traces of gamma
matrices above, are given by (note that
$\gamma^{5}=i\gamma^{0}\gamma^{1}\gamma^{2}\gamma^{3}$ and it satisfies
$\gamma^{5}\gamma^{\mu}=-\gamma^{\mu}\gamma^{5}$)
$\displaystyle{\rm tr}(\gamma^{\mu}\gamma^{\nu})=4\eta^{\mu\nu}\,,$ (F.282)
$\displaystyle{\rm tr}(\gamma^{\mu}\gamma^{\nu}\gamma^{5})=0\,,$ (F.283)
$\displaystyle{\rm
tr}(\gamma^{\alpha}\gamma^{\mu}\gamma^{\beta}\gamma^{\nu})=4\eta^{\alpha\mu}\eta^{\beta\nu}\
-\ 4\eta^{\alpha\beta}\eta^{\mu\nu}\ +\ 4\eta^{\alpha\nu}\eta^{\mu\beta}\,,$
(F.284)
$\displaystyle{\rm
tr}(\gamma^{\alpha}\gamma^{\mu}\gamma^{\beta}\gamma^{\nu}\gamma^{5})=-4i\epsilon^{\alpha\mu\beta\nu}\,,$
(F.285)
$\displaystyle{\rm
tr}(\gamma^{\alpha}\gamma^{\mu}\gamma^{\tilde{\nu}}\gamma^{\beta}\gamma^{\tilde{\mu}}\gamma^{\nu})$
$\displaystyle=4\eta^{\alpha\mu}\times\Bigl{(}\eta^{{\tilde{\nu}}\beta}\eta^{{\tilde{\mu}}\nu}\,-\,\eta^{{\tilde{\nu}}{\tilde{\mu}}}\eta^{\beta\nu}\,+\,\eta^{{\tilde{\nu}}\nu}\eta^{\beta{\tilde{\mu}}}\Bigr{)}$
(F.286) $\displaystyle\qquad-\
4\eta^{\alpha{\tilde{\nu}}}\times\Bigl{(}\eta^{\mu\beta}\eta^{{\tilde{\mu}}\nu}\,-\,\eta^{\mu{\tilde{\mu}}}\eta^{\beta\nu}\,+\,\eta^{\mu\nu}\eta^{\beta{\tilde{\mu}}}\Bigr{)}$
(F.287) $\displaystyle\qquad+\
4\eta^{\alpha\beta}\times\Bigl{(}\eta^{\mu{\tilde{\nu}}}\eta^{{\tilde{\mu}}\nu}\,-\,\eta^{\mu{\tilde{\mu}}}\eta^{{\tilde{\nu}}\nu}\,+\,\eta^{\mu\nu}\eta^{{\tilde{\nu}}{\tilde{\mu}}}\Bigr{)}$
(F.288) $\displaystyle\qquad-\
4\eta^{\alpha{\tilde{\mu}}}\times\Bigl{(}\eta^{\mu{\tilde{\nu}}}\eta^{\beta\nu}\,-\,\eta^{\mu\beta}\eta^{{\tilde{\nu}}\nu}\,+\,\eta^{\mu\nu}\eta^{{\tilde{\nu}}\beta}\Bigr{)}$
$\displaystyle\qquad+\
4\eta^{\alpha\nu}\times\Bigl{(}\eta^{\mu{\tilde{\nu}}}\eta^{\beta{\tilde{\mu}}}\,-\,\eta^{\mu\beta}\eta^{{\tilde{\nu}}{\tilde{\mu}}}\,+\,\eta^{\mu{\tilde{\mu}}}\eta^{{\tilde{\nu}}\beta}\Bigr{)}\,,$
$\displaystyle{\rm
tr}(\gamma^{\alpha}\gamma^{\mu}\gamma^{{\tilde{\nu}}}\gamma^{\beta}\gamma^{{\tilde{\mu}}}\gamma^{\nu}\gamma^{5})=$
$\displaystyle-$ $\displaystyle
4i(\eta^{\alpha\mu}\epsilon^{{\tilde{\nu}}\beta{\tilde{\mu}}\nu}-\eta^{\alpha{\tilde{\nu}}}\epsilon^{\mu\beta{\tilde{\mu}}\nu}+\eta^{{\tilde{\nu}}\mu}\epsilon^{\alpha\beta{\tilde{\mu}}\nu}-\eta^{{\tilde{\mu}}\nu}\epsilon^{\beta\alpha\mu{\tilde{\nu}}}$
(F.291) $\displaystyle+$
$\displaystyle\eta^{\beta\nu}\epsilon^{{\tilde{\mu}}\alpha\mu{\tilde{\nu}}}-\eta^{\beta{\tilde{\mu}}}\epsilon^{\nu\alpha\mu{\tilde{\nu}}})\,,$
and
$\displaystyle{\rm tr}(\gamma^{\nu_{1}}\cdots\gamma^{\nu_{n}}\gamma^{5})\ $
$\displaystyle=$ $\displaystyle\ 0\quad\forall\ {\rm odd}\ n\,,$
$\displaystyle{\rm tr}(\gamma^{\nu_{1}}\cdots\gamma^{\nu_{n}})\ $
$\displaystyle=$ $\displaystyle\ 0\quad\forall\ {\rm odd}\ n\,.$ (F.292)
## References
* (1) C. Adloff et al. [H1 Collaboration], Eur. Phys. J. C 13, 609 (2000) [hep-ex/9908059]; C. Adloff et al. [H1 Collaboration], Phys. Lett. B 520, 183 (2001) [hep-ex/0108035]; C. Adloff et al. [H1 Collaboration], Eur. Phys. J. C 21, 33 (2001) [hep-ex/0012053].
* (2) S. Chekanov et al. [ZEUS Collaboration], Eur. Phys. J. C 21, 443 (2001) [hep-ex/0105090].
* (3) K. J. Golec-Biernat and M. Wusthoff, Phys. Rev. D 59, 014017 (1998) [hep-ph/9807513].
* (4) I. Balitsky, Nucl. Phys. B 463, 99 (1996) [hep-ph/9509348]; Y. V. Kovchegov, Phys. Rev. D 60, 034008 (1999) [hep-ph/9901281].
* (5) J. Polchinski and M. J. Strassler, JHEP 0305, 012 (2003) [hep-th/0209211].
* (6) S. J. Brodsky and G. R. Farrar, Phys. Rev. Lett. 31, 1153 (1973).
* (7) Y. Hatta, E. Iancu and A. H. Mueller, JHEP 0801, 063 (2008) [arXiv:0710.5297 [hep-th]].
* (8) K. A. Mamo and I. Zahed, Phys. Rev. D 100 (2019) no.4, 046015 [arXiv:1808.01952 [hep-ph]]; K. A. Mamo and I. Zahed, Phys. Rev. D 101 (2020) no.6, 066014 [arXiv:1905.07864 [hep-th]].
* (9) J. Hirn and V. Sanz, JHEP 0512, 030 (2005) [hep-ph/0507049];
* (10) S. K. Domokos, H. R. Grigoryan and J. A. Harvey, Phys. Rev. D 80, 115018 (2009) [arXiv:0905.1949 [hep-ph]].
* (11) D. K. Hong, T. Inami and H. U. Yee, Phys. Lett. B 646 (2007), 165-171 [arXiv:hep-ph/0609270 [hep-ph]].
* (12) T. Gutsche, V. E. Lyubovitskij, I. Schmidt and A. Vega, Phys. Rev. D 86, 036007 (2012) [arXiv:1204.6612 [hep-ph]].
* (13) A. Cherman, T. D. Cohen and E. S. Werbos, Phys. Rev. C 79 (2009), 045203 [arXiv:0804.1096 [hep-ph]].
* (14) J. Erlich, E. Katz, D. T. Son and M. A. Stephanov, Phys. Rev. Lett. 95 (2005), 261602 [arXiv:hep-ph/0501128 [hep-ph]].
* (15) T. Sakai and S. Sugimoto, Prog. Theor. Phys. 113 (2005), 843-882 [arXiv:hep-th/0412141 [hep-th]].
* (16) C. Alexandrou, M. Constantinou, K. Hadjiyiannakou, K. Jansen, C. Kallidonis, G. Koutsou and A. Vaquero Aviles-Casco, Phys. Rev. D 96, no.5, 054507 (2017) [arXiv:1705.03399 [hep-lat]].
* (17) N. Tsutsui et al. [CP-PACS and JLQCD], Phys. Rev. D 70, 111501 (2004) [arXiv:hep-lat/0402026 [hep-lat]].
* (18) N. R. F. Braga and A. Vega, Eur. Phys. J. C 72 (2012), 2236 [arXiv:1110.2548 [hep-ph]]; C. A. Ballon Bayona, H. Boschi-Filho and N. R. F. Braga, JHEP 03 (2008), 064 [arXiv:0711.0221 [hep-th]].
* (19) Z. Abidin and C. E. Carlson, Phys. Rev. D 79 (2009), 115003 [arXiv:0903.4818 [hep-ph]].
* (20) Z. Abidin and P. T. P. Hutauruk, Phys. Rev. D 100 (2019) no.5, 054026 [arXiv:1905.08953 [hep-ph]].
* (21) J. M. Conrad, M. H. Shaevitz and T. Bolton, Rev. Mod. Phys. 70 (1998), 1341-1392 [arXiv:hep-ex/9707015 [hep-ex]].
* (22) M. Rho, S. J. Sin and I. Zahed, Phys. Lett. B 466, 199 (1999) [hep-th/9907126]. G. Basar, D. E. Kharzeev, H. U. Yee and I. Zahed, Phys. Rev. D 85, 105005 (2012) [arXiv:1202.0831 [hep-th]]; A. Stoffers and I. Zahed, Phys. Rev. D 87, 075023 (2013) [arXiv:1205.3223 [hep-ph]]; A. Stoffers and I. Zahed, arXiv:1210.3724 [nucl-th]. A. Stoffers and I. Zahed, Phys. Rev. D 88, no. 2, 025038 (2013) [arXiv:1211.3077 [nucl-th]].
* (23) R. A. Janik and R. B. Peschanski, Nucl. Phys. B 586, 163 (2000) [hep-th/0003059].
* (24) R. C. Brower, J. Polchinski, M. J. Strassler and C. I. Tan, JHEP 0712, 005 (2007) [hep-th/0603115]; R. C. Brower, M. J. Strassler and C. I. Tan, JHEP 0903, 092 (2009) [arXiv:0710.4378 [hep-th]]; R. C. Brower, M. S. Costa, M. Djuric, T. Raben and C. I. Tan, JHEP 1502, 104 (2015) [arXiv:1409.2730 [hep-th]]; A. Ballon-Bayona, R. Carcasses Quevedo, M. S. Costa and M. Djuric, Phys. Rev. D 93, 035005 (2016) [arXiv:1508.00008 [hep-ph]].
* (25) Y. Hatta, E. Iancu and A. H. Mueller, JHEP 0801, 026 (2008) [arXiv:0710.2148 [hep-th]]. Y. Hatta, T. Ueda and B. W. Xiao, JHEP 0908, 007 (2009) [arXiv:0905.2493 [hep-ph]].
* (26) Y. Hatta, T. Ueda and B. W. Xiao, JHEP 0908, 007 (2009) [arXiv:0905.2493 [hep-ph]].
* (27) N. Kovensky, G. Michalski and M. Schvellinger, JHEP 1810, 084 (2018) [arXiv:1807.11540 [hep-th]].
* (28) S. Raju, Phys. Rev. D 83 (2011), 126002 [arXiv:1102.4724 [hep-th]].
* (29) E. D’Hoker, D. Z. Freedman, S. D. Mathur, A. Matusis and L. Rastelli, Nucl. Phys. B 562 (1999), 330-352 [arXiv:hep-th/9902042 [hep-th]].
* (30) K. A. Mamo and I. Zahed, Phys. Rev. D 101 (2020) no.8, 086003 [arXiv:1910.04707 [hep-ph]].
* (31) D. Jorrin, G. Michalski and M. Schvellinger, JHEP 06, 063 (2020) [arXiv:2004.02909 [hep-th]].
* (32) E. Folco Capossoli, M. A. Martin Contreras, D. Li, A. Vega and H. Boschi-Filho, [arXiv:2007.09283 [hep-ph]].
* (33) H. Abramowicz et al. [H1 and ZEUS], Eur. Phys. J. C 75, no.12, 580 (2015) [arXiv:1506.06042 [hep-ex]].
* (34) M. R. Adams et al. [E665], Phys. Rev. D 54, 3006-3056 (1996)
* (35) L. W. Whitlow, “Deep Inelastic Structure Functions From Electron Scattering on Hydrogen, Deuterium, and Iron at 0.6-GeV${}^{2}\leq Q^{2}\leq 30$-GeV2,” SLAC-0357.
* (36) S D. Drell and T Yan Phys. Rev. Lett. 24, 181 (1970)
* (37) S. P. Malace et al. [Jefferson Lab E00-115], Phys. Rev. C 80, 035207 (2009) [arXiv:0905.2374 [nucl-ex]].
* (38) D. B. Clark, E. Godat and F. I. Olness, Comput. Phys. Commun. 216, 126-137 (2017) [arXiv:1605.08012 [hep-ph]].
|
# Berezinskii-Kosterlitz-Thouless phase transitions with long-range couplings
Guido Giachetti<EMAIL_ADDRESS>SISSA and INFN Sezione di Trieste, Via
Bonomea 265, I-34136 Trieste, Italy Nicolò Defenu Institute for Theoretical
Physics, ETH Z$\ddot{u}$rich, Wolfgang-Pauli-Str. 27, 8093 Z$\ddot{u}$rich,
Switzerland Stefano Ruffo SISSA and INFN Sezione di Trieste, Via Bonomea
265, I-34136 Trieste, Italy Istituto dei Sistemi Complessi, Consiglio
Nazionale delle Ricerche, Via Madonna del Piano 10, I-50019 Sesto Fiorentino,
Italy Andrea Trombettoni Department of Physics, University of Trieste,
Strada Costiera 11, I-34151 Trieste, Italy SISSA and INFN Sezione di Trieste,
Via Bonomea 265, I-34136 Trieste, Italy CNR-IOM DEMOCRITOS Simulation Center,
Via Bonomea 265, I-34136 Trieste, Italy
###### Abstract
The Berezinskii-Kostelitz-Thouless (BKT) transition is the paradigmatic
example of a topological phase transition without symmetry-breaking, where a
quasi-ordered phase, characterized by a power law scaling of the correlation
functions at low temperature, is disrupted by the proliferation of topological
excitations above the critical temperature $T_{\rm BKT}$. In this letter, we
consider the effect of long-range decaying couplings $\sim r^{-2-\sigma}$ on
this phenomenon. After pointing out the relevance of this non trivial problem,
we discuss the phase diagram, which is far richer than the corresponding
short-range one. It features – for $7/4<\sigma<2$ – a quasi ordered phase in a
finite temperature range $T_{c}<T<T_{\rm BKT}$, which occurs between a
symmetry broken phase for $T<T_{c}$ and a disordered phase for $T>T_{\rm
BKT}$. The transition temperature $T_{c}$ displays unique universal features
quite different from those of the traditional, short-range XY model. Given the
universal nature of our findings, they may be observed in current experimental
realizations in $2D$ atomic, molecular and optical quantum systems.
## I Introduction
Two-dimensional interacting systems are well known not to display conventional
symmetry breaking transitions at finite temperature, due to the Hohenberg-
Mermin-Wagner theorem [1]. Yet, a phase transition may appear driven by
topological defects, according to the celebrated Berezinskii-Kosterlitz-
Thouless (BKT) mechanism [2]. In the presence of long-range interactions the
Hohenberg-Mermin-Wagner theorem no longer holds and local order parameters,
such as the magnetization [3], may have a non-zero expectation value. The
general question addressed by this Letter is the fate of the BKT transition
when the range of the interactions is increased. The Sak’s criterion [4]
provides an argument for understanding whether the long-range, power law
coupling $\sim 1/r^{d+\sigma}$ in the classical $O(N)$ model affects
criticality. It can be formulated as follows: at low momenta the short-range
(SR) and long-range (LR) critical two-points functions behave as
$p^{-2+\eta_{\rm sr}}\hskip 14.22636pt\text{vs}\hskip 14.22636ptp^{-\sigma}$
(1)
respectively, where $\eta_{\rm sr}$ is the anomalous dimension of the SR
limit. Therefore, one can define a critical value of the range of the
interactions, $\sigma_{*}=2-\eta_{\rm sr}$, such that, for
$\sigma>\sigma^{*}$, the critical behavior is not affected by LR. The validity
of Sak’s criterion for the classical $O(N)$ models has been the subject of
considerable scrutiny. Indeed, numerical evidences supporting (or rejecting)
the Sak’s result are notoriously hard to obtain [5, 6, 7]. Intense theoretical
investigations both via MC simulations [5, 8, 9], renormalization group (RG)
theory [10, 11, 12] and conformal bootstrap [13] appeared all to confirm the
validity of Sak’s conjecture for the LR-SR crossover so that it is fair to
conclude that the criterion has been a useful tool to understand the critical
behaviour of LR interacting systems [14, 15, 16, 17]. The criterion, is
believed to apply to all symmetry breaking transition in $d\geq 2$. The status
of the $d=2$ classical XY model is rather different, and only few results
(later commented) are known. The main reasons are
i) The Sak criterion cannot be straightforwardly applied, since in the SR
limit the critical behavior is not described by a single RG fixed point, but
rather by a whole line of fixed points with a temperature-dependent exponent
$\eta_{\rm sr}$.
ii) Numerically, the large number of non-vanishing couplings, coming form the
LR nature of the interaction, along with the logarithmic scaling typical of
$2D$ systems (the so-called “Texas state argument” [18]) make the study of the
$2D$ XY universality notoriously challenging.
iii) In the nearest-neighbours $2D$ XY model, the classical treatment takes
advantage of the duality construction [19], through which one can famously
relate the model to the Coulomb gas [20, 21] or the sine-Gordon model [22,
21]. However, this is no longer the case already for next-to-nearest-neighbors
couplings.
iv) It is known that in the SR limit, the physics of the $2D$ classical XY
model can be related to the one of the $1D$ quantum XXZ model via its transfer
matrix [23].
This approach is based on the mapping to hard-core bosons, and therefore to
the XXZ model, and cannot be straightforwardly applied to the the case of $XY$
LR interactions, as one should show the RG irrelevance of terms violating the
hard-core condition. Moreover, let us remark that $2D$ boson gas at finite
temperature with (isotropic) $1/r^{3}$ density-density interaction do exhibit
a BKT transition [24]; but this interaction corresponds to a quantum 1D XXZ in
which only in the $z-z$ interaction is long range.
v) Finally, we observe that the treatment of the SR XY model in $2D$ is very
much simplified by the introduction of the Villain model, [25, 26], which can
be mapped exactly onto the Coulomb gas, and shares the same universality class
of the SR $XY$ model. The physical reason of their connection in the SR
regime, is that the (gapped) amplitude fluctuations of the corresponding
$O(2)$ action are irrelevant [27]. Thus, once the periodic nature of the phase
is taken care of, all the relevant information is present in the theory.
However, in the LR regime the interplay between amplitude and phase
fluctuations cannot be neglected and it is not known whether they still share
the same universality class.
Despite these difficulties, the study of LR XY model is of great interest:
first, since its introduction, the BKT mechanism [28, 29, 30, 31] has been
found to quantitatively describe the universal scaling appearing in several
$2D$ systems with $U(1)$ symmetry, ranging from thin 4He films [32] to
quasi-2D layered superconductors [33, 34, 35, 36, 37], exciton-polariton
systems [38], cold atoms in $2D$ traps [39, 40] and $2D$ electron gases at the
interface between insulating oxides in artificial heterostructures [41, 42,
43]. Apart from these experimental realizations, topological defects are
expected to be relevant in several natural phenomena outside the condensed
matter realm, such as DNA tangling or stripe formation [44, 45, 46]. To
understand how $\sigma_{*}$ is modified, is then a crucial question in all the
cases in which a LR tail of the interaction can be added or tuned, especially
because the spin-wave interaction term, already present in the SR case, may
destroy, partially or totally, the topological nature of the phase transition.
Moreover, the physics of LR interacting systems has recently experienced a new
wave of interest, due to the current experimental realizations on atomic,
molecular and optical (AMO) systems. In particular, trapped ions [47, 48],
Rydberg gases [49] and optical cavities [50, 51] allowed the observation of
plenty of exotic equilibrium and dynamical phenomena induced by LR
interactions, including entanglement and correlations propagation [52, 53],
dynamical phase transitions [54, 55], time crystals [56, 57, 54] and defect
scaling [58, 59]. These experimental results stimulated an impressive
theoretical activity to characterize the equilibrium and dynamical critical
scaling induced by LR interactions in a wide variety of different systems [60,
61, 62, 63, 17, 64, 65].Despite this outpouring theoretical activity and the
long-standing relation between topological scaling and LR interactions, the
possible corrections induced by power-law decaying couplings to the
topological BKT scaling remain an open question, testable in experiments.
## II Model & Preliminaries
We consider a system of planar rotators on a $2D$ lattice of spacing $a$,
described by the Hamiltonian:
$\beta
H=\frac{1}{2}\sum_{\mathbf{i},\mathbf{j}}J_{|\mathbf{i}-\mathbf{j}|}\left[1-\cos(\theta_{\mathbf{j}}-\theta_{\mathbf{i}})\right]$
(2)
where $\mathbf{i},\mathbf{j}\in\mathbb{Z}^{2}$ and
$J_{|\mathbf{i}-\mathbf{j}|}$ has a power-law tail:
$J_{|\mathbf{i}-\mathbf{j}|}\sim\frac{g}{|\mathbf{i}-\mathbf{j}|^{2+\sigma}}$
for $|\mathbf{i}-\mathbf{j}|\gg 1$. The exponent $\sigma$ is assumed positive
in order to ensure additivity of the thermodynamic quantities [66]. For the
following arguments the specific form of the couplings is not important, as
long as that there are no frustration effects nor competing interactions.
Let us now summarize what we do know for sure about the LR XY model (2):
a) For $\sigma<2$, at low enough temperatures, the system magnetizes, as
rigorously proven in [3]. MC simulations at $\sigma=1$ indicate an order-
disorder transition and no BKT phase at finite temperature [67]. Moreover,one
could expect that For $\sigma\leq 1$ the critical exponents of the ferro-
paramagnetic transition are expected to be mean-field [11].
b) In agreement with a), the spin-wave theory in which the cosine is expanded
to the quadratic order, without imposing the periodicity, as in the original
Berezinskii calculation [28], does also magnetize for $\sigma<2$, since the
contribution of the spin fluctuations is of the form $\int d^{2}q/q^{\sigma}$
and thus infrared finite.
c) An upper bound for $\sigma_{*}$ has to be $\sigma_{*}=2$, i.e. for sure
there is BKT for $\sigma>2$, as one can deduce even from the Sak’s argument
since $\eta$ is positive. This result is supported by the self-consistent
harmonic calculation recently presented in [68], which anyway is unable to
make even qualitative predictions for $\sigma<2$.
## III Effective model
We decompose the coupling as
$J_{|\mathbf{i}-\mathbf{j}|}=J^{S}_{|\mathbf{i}-\mathbf{j}|}+g|\mathbf{i}-\mathbf{j}|^{-(2+\sigma)}$
where $J^{S}$ is a SR term taking into account the small-distances behavior of
the coupling. At low temperatures, the spin direction varies smoothly from
site to site and, as a consequence, we can expand the SR term for small phase
differences as $\cos(\theta(\mathbf{x}+\mathbf{r})-\theta(\mathbf{x}))\sim
1-|\nabla\theta|^{2}/2$. The same, however, it is not automatically true for
the LR term, since far-away pairs, whose phase difference is not necessarily
small, give a significant contribution to the Hamiltonian.
These considerations allow us to write a continuous version of the Hamiltonian
in Eq. (2) in terms of the field $\theta(\mathbf{x})$, namely the Euclidean
action
$S[\theta]=\frac{J}{2}\int d^{2}x|\nabla\theta|^{2}+S_{\mathrm{L}R},$ (3)
where the LR part can be written as
$S_{\mathrm{L}R}=-\frac{g}{2\gamma_{2,\sigma}}\int
d^{2}x(\cos\theta\,\nabla^{\sigma}\cos\theta+\sin\theta\,\nabla^{\sigma}\sin\theta),$
(4)
with
$\gamma_{2,\sigma}=2^{\sigma}\Gamma(\scriptstyle\frac{1+\sigma}{2}\displaystyle)\pi^{-1}|\Gamma(\scriptstyle-\frac{\sigma}{2}\displaystyle)|^{-1}$,
by using the definition of (bulk) fractional derivative given in Appendix A.
The first and the second term in Eq. (3) account for the short- and long-range
contributions respectively, with $J\sim 1/T$ and $g\sim 1/T$ being the
temperature dependent couplings. Notice that the result would be different for
a quantum $1D$ chain with LR interactions, where interactions are still SR
along the imaginary time axis [69].
If $g=0$, by following the usual duality procedure [26], one can take into
account the periodic nature of the field $\theta$ in Eq. (3) by isolating the
topological configurations and introducing the vortex fugacity
$y=\exp(-\varepsilon_{c})$, being $\varepsilon_{c}$ the corresponding core
energy. This, in turn, leads to the Kosteritz-Thouless RG equations [29, 30,
26, 70] (see [71, 72] for textbook presentations) which feature a line of
stable Gaussian fixed points for $y=0$ and $J>\frac{2}{\pi}$, describing the
power-law scaling observed in the low-temperature BKT phase. For $g$ small
enough, we expect to have then a continuum theory described by the three
parameters $J$, $g$ and $y$.
In order to explore the effects of LR interactions, we deform the traditional
BKT fixed-points theory with the non-local operator in the second term of Eq.
(3). Since only those fixed-points which are stable under topological
perturbation correspond to an infra-red (IR) limit of the SR BKT theory, we
can restrict ourselves to the region in which the topological excitations are
irrelevant ($J>\frac{2}{\pi}$). The relevance of the LR perturbation depends
on the scaling dimension $\Delta_{g}$ of the coupling $g$, which is defined
according to the asymptotic behavior $g_{\ell}\approx\exp(\Delta_{g}\ell)$ for
$\ell\gg 1$, where as usual in the BKT literature, we put $\ell=\ln(r/a)$. On
the other hand, due to the Gaussian nature of the measure,
$\left\langle\cos\left(\theta(\mathbf{x})-\theta(\mathbf{x^{\prime}})\right)\right\rangle=e^{-\frac{1}{2}\left\langle\left(\theta(\mathbf{x})-\theta(\mathbf{x^{\prime}})\right)^{2}\right\rangle}=|\mathbf{x}-\mathbf{x^{\prime}}|^{-\eta_{\rm
sr}(J)},$ (5)
where $\eta_{\rm sr}(J)=\frac{1}{2\pi J}$ is the exponent of the SR two-point
function, [29, 30, 26]. Following Eq. (5), the scaling dimension of the LR
term reads
$\displaystyle\Delta_{g}=2-\sigma-\eta_{\mathrm{sr}}(J)$ (6)
so that the LR perturbation becomes relevant only if $\sigma<2-\eta_{\rm
sr}(J)$, similarly to the traditional spontaneous symmetry breaking (SSB) case
[11], but with a temperature-dependent anomalous dimension. This confirms that
for $\sigma>2$ the LR perturbation is always irrelevant, as expected.
Let us now consider the case $\sigma<2$. There, the LR perturbation becomes
relevant at small temperatures, since $\eta_{\rm sr}\simeq 0$ for $T\simeq 0$.
Since $\eta_{sr}$ in Eq. (6) is the one of the SR unperturbed theory, we can
apply the results of the traditional BKT theory [73] as long as the LR
perturbation is not relevant. In particular we know that topologicaly
excitations are irrelevant for $\eta_{sr}<1/4$, so that in the range
$7/4<\sigma<2$, a subset of the BKT fixed points remains stable and we have
quasi-long-range order (qlro) for a certain temperature window. This result is
rather non trivial, since in SSB transitions the traditional Sak’s result [4]
predicts the irrelevance of LR couplings at all temperatures for
$\sigma>2-\eta_{\rm sr}$.
## IV RG Flow
These results may be confirmed by deriving the flow equations for the LR term
at the leading order in $g$ for $y=0$, obtaining (see Appendix B):
$\begin{split}\frac{dg_{\ell}}{d\ell}&=\Big{(}2-\sigma-\eta_{\rm
sr}(J_{\ell})\Big{)}g_{\ell}\\\
\frac{dJ_{\ell}}{d\ell}&=c_{\sigma}\eta_{sr}(J_{\ell})g_{\ell}\end{split}$ (7)
where $c_{\sigma}=\frac{\pi}{2}a^{2-\sigma}\int^{\infty}_{1}du\
u^{1-\sigma}\mathcal{J}_{0}(2\pi u)$ and $\mathcal{J}_{0}(x)$ is the zeroth
order Bessel function of the first kind. As shown in Appendix B the above
result is reliable as long as $a^{2-\sigma}g_{\ell}\ll J_{\ell}$ or,
equivalently, as long as $\frac{dJ}{d\ell}\ll J_{\ell}$. As expected, we see
that the flow equations (29) support a line of SR fixed points $g=0$ which
becomes unstable for $\eta_{\rm sr}(J)<2-\sigma$. As long as our hypothesis of
small $g$ holds, we can explicitly identify the form of the flow trajectories
of Eqs. (29):
$g_{\ell}(J)=\frac{\pi(2-\sigma)}{c_{\sigma}}\left[\left(J_{\ell}-J_{\sigma}\right)^{2}+k\right],$
(8)
where $k$ is a real number and $J_{\sigma}=\frac{1}{2\pi(2-\sigma)}$. The sign
of $k$ divides the trajectories which met the fixed point $g=0$ and those
which do not, the first ones ending at (starting from) the fixed point line
for $J\leq J_{\sigma}$ ($J>J_{\sigma}$). The separatrix corresponds to the
semi-parabola with $k=0$, $J\leq J_{\sigma}$. For $k>0$ $g\to\infty$, showing
the existence of a new low-temperature phase, where LR interactions are
relevant. The critical temperature $T_{c}$, below which this new phase
appears, is such that $\eta_{\rm sr}(J_{c})>2-\sigma$ .
Since, as in the traditional BKT calculation [29], Eqs. (29) were derived for
small $g$ and $y$, its use for $T<T_{c}$ is in principle not justified, since
LR interactions are relevant and $g_{\ell}$ grows indefinitely. However, let
us notice that the scaling of $g_{\ell}$ with $T$ for $T\rightarrow T_{c}^{-}$
can be reliably predicted from Eqs. (29), since in this limit the flow spends
a divergent amount of RG time $\ell$ in the vicinity of the line of fixed
points $g=0$. This scaling is derived in Appendix C. Moreover, we can guess
the infrared form of the action in the low temperature phase by observing that
the rigorous result of Ref. [3] implies that for $T<T_{c}$ the system displays
finite magnetization and, then, phase fluctuations are limited even at large
distances. Therefore, the expansion of the trigonometric function in Eq. (3)
is justified leading to an action of the form
$S_{g}=-\frac{\bar{g}}{2}\int d^{2}x\ \theta\nabla^{\sigma}\theta,$ (9)
where $\bar{g}=g\gamma_{2,\sigma}^{-1}$. Being the above action quadratic, the
properties of this exotic low temperature phase can be worked out: in
particular the scaling of the magnetization for $T\rightarrow T_{c}^{-}$ is
found to be (see Appendix C for details)
$\ln m\sim-e^{B(T_{c}-T)^{-1/2}}$ (10)
where $B$ is a non universal constant. Since all the derivatives of $m$ with
respect to $T$ vanish at $T=T_{c}$ (essential singularity), and since $m$ is
linked to the derivative of the free energy with respect to the external
field, we have that the phase transition between the ordered and disordered
phase is actually of infinite order. Moreover, the connected correlation
functions have a power-law decay for $T<T_{c}$ given by
$\left\langle\mathbf{S}(\mathbf{r})\cdot\mathbf{S}(\mathbf{0})\right\rangle_{c}\sim\frac{1}{r^{2-\sigma}}$,
where
$\mathbf{S}(\mathbf{r})=(\cos{\theta}_{\mathbf{r}},\sin{\theta}_{\mathbf{r}})$.
Figure 1: Sketch of the RG flow lines for $\frac{7}{4}<\sigma<2$ in the $y=0$
plane. The dashed red line is a possible realization of the physical
parameters line, from which the flow starts, as the temperature is varied. On
the right/left of the gray dotted line the vortex fugacity $y$ is
irrelevant/relevant ($\dot{y}_{\ell}/y_{\ell}\gtrless 0$). The two
separatrices (bold black lines) divide the flow in three regions: a high-
temperature region (orange, the flow ends up in the disordered phase), an
intermediate one (blue, the flow reaches a $g=0$ fixed point) and the low-
temperature region (green, the LR perturbation brings the system away from the
critical line).
We have so far assumed $y=0$; let us now consider the effect of topological
excitations. At leading order in both $g$ and $y$ the two perturbations remain
independent and, since the vortices contribute to the $J_{\ell}$ flow only
beyond leading order in $y$, Eqs.(29) are unchanged. Moreover, one has
$\frac{dy_{\ell}}{d\ell}=(2-\pi J_{\ell})y_{\ell}$ valid up to second order
terms in $y_{\ell}$ and $g_{\ell}$. Then, in agreement with what we stated
above, as long as $\frac{7}{4}<\sigma<2$, the temperature range $T$ between
$T_{c}$ and $T_{\rm BKT}$ of the line of fixed points $g=y=0$ remains stable
under both topological and LR perturbations. In the low-temperature phase
instead, it is natural to suppose $y$ to be irrelevant, due to the fact that a
non-negligible LR coupling increases the cost of, highly non-local topological
excitations. This idea is made rigorous in Appendix D where the interaction
energy between vortices in the low temperature phase is computed, and it is
shown that they cannot proliferate.
Summarizing, for $\sigma\in(7/4,2)$ we find three phases: i) an ordered phase
for $T<T_{c}$ with finite magnetization and temperature independent power-law
correlation functions ii) an intermediate BKT phase for $T_{c}<T<T_{\rm BKT}$,
where the magnetization vanishes and the exponent of the two-point correlation
function depends on $T$ iii) a disordered phase for $T>T_{\rm BKT}$. Due to
the LR character of the interactions, also the high-temperature phase displays
power-law decaying two-point functions
$\left\langle\mathbf{S}(\mathbf{r})\cdot\mathbf{S}(0)\right\rangle\sim
r^{-2-\sigma}$,[74, 75, 76]. As $\sigma\rightarrow 7/4^{+}$ the critical
temperature $T_{c}$ reaches $T_{BKT}$ from below. Therefore, for $\sigma<7/4$,
the whole BKT line fixed points becomes unstable either with respect to
topological or LR perturbations and the intermediate phase vanishes, leaving
only a SSB phase transition. However, our approach cannot reliably be used to
fully characterize this transition: as $T$ approaches $T_{c}$ from below, the
RG flow slows down close to the $g=0$, $J=J_{\sigma}$ fixed point. Since for
$\sigma<\frac{7}{4}$ $J_{\sigma}<J_{\rm BKT}$, $y$ grows indefinitely, away
from the $y\ll 1$ regime. Our results are summarized in Fig. 2.
Figure 2: Sketch of the possible phases of the model: ordered with
magnetization (solid black), BKT qlro (dashed light gray), disordered (dashed
dark gray). If $\sigma>2$ we find the usual SR phenomenology with a BKT phase
transition. For $\sigma<2$ an ordered phase appears at low-temperatures, the
BKT qlro phase disappearing for $\sigma<\frac{7}{4}$.
## Conclusions
We have shown that the introduction of long-range (LR) power decaying
couplings in the $2D$ XY model Hamiltonian produces a rich phase diagram,
different from the short-range (SR) case [29] and from the one of $O(N)$ LR
systems [4]. Remarkably, for $7/4<\sigma<2$, the system displays both BKT qlro
in the temperature interval $T_{c}<T<T_{\rm BKT}$ and actual long-range order
for $T<T_{c}$.
The introduction of complex interaction patterns in systems with $U(1)$
symmetry is known to generate exotic critical features, as in the anisotropic
$3D$ XY model [77], coupled XY planes [78], 2D systems with anisotropic
dipolar interactions [79, 80] or four-body interactions [81], and high-
dimensional systems with Lifshitz criticality [82, 83] The present work
constitutes a further milestone along this path, as it identifies a peculiar
critical behavior, namely a non-analytic exponential vanishing of the order
parameter, that eludes the current classification of universal scaling
behaviors [84].
Our predictions may be tested in several low dimensional AMO systems. It would
be interesting to perform extensive numerical simulations in order to observe
the scaling of the critical quantities, and especially the magnetization,
close to the low-temperature endpoint of the BKT line in the regime
$7/4<\sigma<2$. These simulations will be useful to classify this
unprecedented transition and to investigate possible corrections near the
$\sigma=7/4$ endpoint due to higher-order effects caused by spin-wave
excitations [85]. Further investigation is also needed to compare our results
with the LR diluted model studied in [86, 87]. In this model, at
$\sigma=1.875$, the numerical simulations presented in [87] do not find any
intermediate BKT region, but the general question whether the $2D$ LR diluted
XY model and the $2D$ LR non-diluted one have the same phase diagram remains
open.
Our results have also implications for LR quantum XXZ chains [69, 88, 89]. One
would need to perform the exact mapping of the classical $2D$ LR XY model to
an effective $1D$ quantum model, following the calculation presented in [23]
and valid for the classical $2D$ SR XY model. If the non-local/LR terms
violating the hard-core boson condition can be shown to be irrelevant, then
one could put in correspondence our phase diagram with that of the LR quantum
XXZ chains having LR couplings both for $x-y$ and $z-z$ terms [69]. This seems
to be confirmed by the similar structure of the RG flow equations of [69] with
our Eqs.(29) taken at low temperatures. If this is the case, then the two
lines, black and white, of Fig.1 of [69] would merge in a point, with the XY
phase disappearing, corresponding to our $\sigma=7/4$ point. Finally, we
mention that it would be interesting to study in detail the phase diagram of
the $2D$ LR Villain model for $\sigma<2$.
## V Acknowledgment
Valuable discussions with G. Parisi, F. Ricci-Tersenghi and A. Scardicchio are
gratefully acknowledged. N.D. and A.T. also acknowledges useful discussion
with M. Ibáñez Berganza. This work is supported by the Deutsche
Forschungsgemeinschaft (DFG, German Research Foundation) under Germany’s
Excellence Strategy EXC2181/1-390900948 (the Heidelberg STRUCTURES Excellence
Cluster). This work is supported by the CNR/HAS (Italy-Hungary) project
“Strongly interacting systems in confined geometries”. This work is part of
MUR-PRIN2017 project “Coarse-grained description for non-equilibrium systems
and transport phenomena (CONEST)” No. 201798CZL whose partial financial
support is acknowledged.
## VI Appendix
### VI.1 A. Defintion of the fractional Laplacian
Given a real parameter $\sigma\in(0,2)$, one can define the fractional
Laplacian of order $\sigma$ of a function
$f(\mathbf{x}):\,\mathbb{R}^{d}\rightarrow\mathbb{R}$ as:
$\nabla^{\sigma}f(\mathbf{x})\equiv\gamma_{d,\sigma}\int
d^{d}r\frac{f(\mathbf{x}+\mathbf{r})-f(\mathbf{x})}{r^{d+\sigma}},$ (11)
where
$\gamma_{d,\sigma}=\frac{2^{\sigma}\Gamma(\frac{d+\sigma}{2})}{\pi^{d/2}|\Gamma(-\frac{\sigma}{2})|}$
and $r=\mid\mathbf{r}\mid$. Another expression for this quantity can be
derived in terms of the Fourier transform, $f(\mathbf{q})$, of
$f(\mathbf{x})$:
$\nabla^{\sigma}f(\mathbf{x})=-\gamma_{d,\sigma}\int d^{d}q\ f(\mathbf{q})\
e^{i\mathbf{q}\cdot\mathbf{x}}\int
d^{d}r\frac{1-e^{i\mathbf{q}\cdot\mathbf{r}}}{r^{d+\sigma}}.$ (12)
Since
$\int
d^{d}r\frac{1-e^{i\mathbf{q}\cdot\mathbf{r}}}{r^{d+\sigma}}=\gamma_{d,\sigma}^{-1}\
q^{\sigma},$ (13)
we find the alternative definition:
$\nabla^{\sigma}f(\mathbf{x})=-\int d^{d}q\
q^{\sigma}f(\mathbf{q})e^{i\mathbf{q}\cdot\mathbf{x}}.$ (14)
In our case $d=2$ and we have to evaluate the quantity:
$\int
d^{2}x\int_{r>a}\frac{d^{2}r}{r^{2+\sigma}}[1-\cos\left(\theta(\mathbf{x})-\theta(\mathbf{x}+\mathbf{r})\right)]$
(15)
For $\sigma<2$, one can actually disregard the contribution coming from the
lattice spacing $a$, since it would just result in a correction of the short-
range term. Then, through trivial trigonometric manipulations we can write the
above expression as:
$\begin{split}&\int
d^{2}x\cos\theta(\mathbf{x})\int\frac{d^{2}r}{r^{2+\sigma}}[\cos\theta(\mathbf{x})-\cos\theta(\mathbf{x}+\mathbf{r})]\\\
+&\int
d^{2}x\sin\theta(\mathbf{x})\int\frac{d^{2}r}{r^{2+\sigma}}[\sin\theta(\mathbf{x})-\sin\theta(\mathbf{x}+\mathbf{r})].\end{split}$
(16)
Finally, using the definition (11) of the fractional derivative, we get
$-\gamma_{2,\sigma}^{-1}\int
d^{2}x\left(\cos\theta\nabla^{\sigma}\cos\theta+\sin\theta\nabla^{\sigma}\sin\theta\right)$
(17)
which justifies the alternative form of the long-range term given in the main
text as Eq.(4).
### VI.2 B. Renormalization group for $y=0$
We will now derive the set of RG equations (7) given in the main text, valid
for $y=0$. We then start form the action written in the form
$S[\theta]=\int
d^{2}x\left(\frac{J_{\ell}}{2}|\nabla\theta|^{2}+\frac{g_{\ell}}{2}\int_{r>a}\frac{d^{2}r}{r^{2+\sigma}}\left[1-\cos\left(\Delta_{\mathbf{r}}\theta(\mathbf{x})\right)\right]\right)$
(18)
where, as in the main text,
$\Delta_{\mathbf{r}}\theta(\mathbf{x})=\theta(\mathbf{x}+\mathbf{r})-\theta(\mathbf{x})$,
and compute the flux perturbatively around $g=0$. The field is split into fast
and slow modes with respect to the momentum cutoff $\Lambda=\frac{2\pi}{a}$,
namely $\theta=\theta^{>}+\theta^{<}$ with
$\begin{split}\theta^{<}(\mathbf{x})&=\int_{q<\Lambda
e^{-d\ell}}\frac{d^{2}q}{(2\pi)^{2}}\theta(\mathbf{q})e^{i\mathbf{q}\cdot\mathbf{x}}\\\
\theta^{>}(\mathbf{x})&=\int_{\Lambda>q>\Lambda
e^{-d\ell}}\frac{d^{2}q}{(2\pi)^{2}}\theta(\mathbf{q})e^{i\mathbf{q}\cdot\mathbf{x}},\end{split}$
(19)
where $\ell=\ln(r/a)$. If we assume the interacting long-range term in Eq.
(18) to be small with respect to the quadratic one, we can perform the
integration perturbatively. It is easy to see that this is possible if
$ga^{2-\sigma}<<J$. Under this assumption then we integrate out the fast
modes, expanding the partition function in cumulants of the non-Gaussian part
$S_{g}$:
$S_{\rm eff}[\theta^{<}]=S_{0}[\theta^{<}]+\left\langle
S_{g}\right\rangle_{>}+O(g^{2}).$ (20)
Writing
$\cos(\Delta_{\mathbf{r}}\theta)=\cos(\Delta_{\mathbf{r}}\theta^{>})\cos(\Delta_{\mathbf{r}}\theta^{<})+\sin(\Delta_{\mathbf{r}}\theta^{>})\sin(\Delta_{\mathbf{r}}\theta^{<})$,
one sees that only the first term will give a contribution. Then, up to
additive constants we have:
$\left\langle S_{g}\right\rangle_{>}=\frac{g_{\ell}}{2}\int
d^{2}x\int\frac{d^{2}r}{r^{2+\sigma}}\left\langle\cos(\Delta_{\mathbf{r}}\theta^{>})\right\rangle_{>}\left[1-\cos\left(\Delta_{\mathbf{r}}\theta^{<}\right)\right]$
(21)
(from now on we omit the $r>a$ condition in the integral over $r$). On the
other hand,
$\left\langle\cos(\Delta_{\mathbf{r}}\theta^{>})\right\rangle_{>}=e^{-\frac{1}{2}\left\langle\left(\theta(\mathbf{r})-\theta(0)\right)^{2}\right\rangle_{>}}$
and
$\frac{1}{2}\left\langle\left(\theta(\mathbf{r})-\theta(0)\right)^{2}\right\rangle_{>}=\int_{\Lambda>q>\Lambda
e^{-d\ell}}\frac{d^{2}q}{(2\pi)^{2}}\frac{1-\cos(\mathbf{q}\cdot\mathbf{r})}{J_{\ell}q^{2}}=\frac{d\ell}{2\pi
J_{\ell}}\Big{(}1-\mathcal{J}_{0}(\Lambda r)\Big{)},$ (22)
where $\mathcal{J}_{0}(x)$ is the zeroth-order Bessel function of the first
kind. Then, introducing $\eta_{\rm sr}(J)=\frac{1}{2\pi J_{\ell}}$, the
exponent of the correlations at the cutoff scale $\ell$, we have:
$\begin{split}\left\langle\cos(\Delta_{\mathbf{r}}\theta^{>})\right\rangle_{>}&=e^{-\eta_{\rm
sr}(J_{\ell})d\ell\left(1-\mathcal{J}_{0}(\Lambda r)\right)}\\\ &=1-\eta_{\rm
sr}(J_{\ell})d\ell+\eta_{\rm sr}(J_{\ell})d\ell\mathcal{J}_{0}(\Lambda
r)\end{split}$ (23)
up to second order corrections. The first two terms provide an anomalous
dimension of the coupling $g_{\ell+d\ell}=g_{\ell}e^{-\eta_{\rm
sr}(J_{\ell})d\ell}$, as expected, while the last one modifies the power-law
dependence on $r$ of the long-range term:
$\begin{split}\left\langle S_{g}\right\rangle_{>}=&\frac{1}{2}\int
d^{2}x\Biggl{\\{}\int\frac{d^{2}r}{r^{2+\sigma}}ge^{-\eta_{\rm
sr}(J_{\ell})d\ell}\left[1-\cos\left(\Delta_{\mathbf{r}}\theta^{<}\right)\right]\\\
+&g\eta_{\rm
sr}(J_{\ell})d\ell\int\frac{d^{2}r}{r^{2+\sigma}}\mathcal{J}_{0}(\Lambda
r)\left[1-\cos\left(\Delta_{\mathbf{r}}\theta^{<}\right)\right]\Biggr{\\}}.\end{split}$
(24)
Let us now examine the last term of the above equation. This can be seen as an
interaction term of the original $XY$ form. Since $\mathcal{J}_{0}(x)\sim
x^{-1/2}\cos(x-\pi/4)$ for large $x$, the new coupling decays faster than the
original and has an oscillating behavior, which provides a natural cutoff for
$r\sim\Lambda^{-1}$. It is then reasonable to approximate this with a short-
range coupling of the form $|\nabla\theta|^{2}$. The simplest way is to
replace
$1-\cos(\Delta_{r}\theta)\approx\frac{1}{2}(\mathbf{r}\cdot\nabla_{\mathbf{x}}\theta)^{2}$
and observe that
$\int\frac{d^{2}r}{r^{2+\sigma}}\mathcal{J}_{0}(\Lambda
r)(\mathbf{r}\cdot\nabla_{\mathbf{x}}\theta^{<})^{2}=\pi|\nabla_{\mathbf{x}}\theta^{<}|^{2}\int^{\Lambda^{-1}}_{a}drr^{1-\sigma}\mathcal{J}_{0}(\Lambda
r).$ (25)
For $\sigma>\frac{1}{2}$, we can neglect the cutoff and, with the substitution
$r=au$, we can express the correction in the action as
$\frac{c_{\sigma}}{2}(g_{\ell}a^{2-\sigma})\eta_{\rm sr}(J_{\ell})d\ell\int
d^{2}x|\nabla_{\mathbf{x}}\theta^{<}|^{2},$ (26)
where
$c_{\sigma}=\frac{\pi}{2}\int_{1}^{\infty}duu^{1-\sigma}\mathcal{J}_{0}(2\pi
u)>0$. The integral is actually ill-defined for $\sigma<\frac{1}{2}$ signaling
that our approximation breaks down. Let us notice however that the precise
value of the coefficient is not important for our analysis. Moreover, it
should be noticed that this entire procedure is only reliable for
$\sigma>7/4$, where part of the BKT fixed points line remains stable and
furnishes a viable expansion point, see the discussion in the main text. Up to
the first order in $g$, then the integration of the fast modes gives the
corrections:
$\begin{split}dg&=-\eta_{\rm sr}(J_{\ell})g_{\ell}d\ell\\\
dJ&=c_{\sigma}\eta_{\rm sr}(J_{\ell})(g_{\ell}a^{2-\sigma})d\ell.\end{split}$
(27)
In order to obtain a theory with the same cutoff scale, we have to do the
replacement $\mathbf{x}\rightarrow\mathbf{x}e^{-d\ell}$ in the action. This
modifies the couplings $g$, $J$ by their own bare length dimension, i.e.
$2-\sigma$ and $0$ respectively:
$\begin{split}dg&=(2-\sigma-\eta_{\rm sr}(J_{\ell}))g_{\ell}d\ell\\\
dJ&=c_{\sigma}\eta_{\rm
sr}(J_{\ell})(g_{\ell}a_{0}^{2-\sigma})d\ell.\end{split}$ (28)
In turn, one finally obtains the RG equations:
$\begin{split}\frac{dg}{d\ell}&=\left(2-\sigma-\eta_{\rm
sr}(J_{\ell})\right)g_{\ell}\\\ \frac{dJ}{d\ell}&=c_{\sigma}\eta_{\rm
sr}(J_{\ell})g_{\ell},\end{split}$ (29)
i.e. Eqs. (7) of the main text (we absorbed the constant ultraviolet cutoff
$a^{2-\sigma}$ in the definition of $c_{\sigma}$).
### VI.3 C. Magnetization in the low-temperature phase
We will now derive the scaling behavior (10) given in the main text for the
magnetization near $T_{c}$, for $T\to T_{c}^{-}$. We start from the Gaussian
theory, Eq. (9) of the main text, describing the low temperature phase of the
theory in the infrared (IR). Being the theory Gaussian, it is
$m=\left\langle\cos\theta(\mathbf{x})\right\rangle=e^{-\frac{1}{2}\left\langle\theta^{2}(\mathbf{x)}\right\rangle}$.
Being
$\left\langle\theta^{2}(\mathbf{x})\right\rangle=\int_{q<2\pi/a}\frac{d^{2}q}{(2\pi^{2})}\frac{1}{\bar{g}q^{\sigma}}\sim\frac{1}{\bar{g}a^{\sigma-2}}.$
(30)
we find
$m=e^{-A/\bar{g}},$ (31)
where $A$ is a non-universal constant. Now, from the flow equations (29), we
find:
$g_{\ell}=ge^{(2-\sigma)\ell}e^{-\int\eta_{\rm sr}(J_{\ell})d\ell},$ (32)
which is reliable as long as $g_{\ell}$ is small. Let us consider a trajectory
which runs very close to the separatrix which, according to Eq. (8) of the
main text, is described by the trajectory
$g=\frac{\pi(2-\sigma)}{c_{\sigma}}\left[(J-J_{\sigma})^{2}+k\right]$ with
$k\rightarrow 0^{+}$. Let us consider a point in the flow $\ell^{*}$ such that
$g(\ell^{*})$ is small and $J(\ell^{*})>J_{\sigma}$. Then:
$\int^{\ell^{*}}_{0}\eta_{\rm
sr}(J_{\ell})d\ell=\int^{\ell^{*}}_{J_{0}}\eta_{\rm sr}(J)\
\frac{dJ}{\dot{J}}=c_{\sigma}^{-1}\int^{\ell^{*}}_{J_{0}}\frac{dJ}{g(J)}=\pi(2-\sigma)\int^{J(\ell^{*})}_{J_{0}}\frac{dJ}{\left(J-J_{\sigma}\right)^{2}+k}$
(33)
By changing the value of the temperature, we have that $J_{0}$ crosses the
separatrix ($k\rightarrow 0^{+}$) for some $J_{c}<J_{\sigma}$ that corresponds
to the critical temperature $T_{c}$, and consequently $k\sim T_{c}-T$. Since
in this case the integration interval on $J$ contains the second order
singularity $J_{\sigma}$, we have that the integral diverges as $k^{-1/2}$ as
$k\rightarrow 0^{+}$. Then we have
$g_{\ell^{*}}\sim e^{-B(T-T_{c})^{-1/2}}$ (34)
where $B$ is a non universal constant. Since, as $k\rightarrow 0^{+}$, the
trajectories corresponding to different values of $k$ run close in the
parameter space, for large $g$ as well, we do not expect this scaling to be
modified in the non-perturbative region. Finally, exploiting Eq. (31), one has
the scaling:
$\ln m\sim-Ae^{B(T-T_{c})^{-1/2}}$ (35)
### VI.4 D. Irrelevance of topological excitations in the low-temperature
phase
We start from the quadratic action of Eq. (9) of the main text, which
describes the low temperature phase, we express it in terms of the Fourier
transform of $\mathbf{v}(\mathbf{x})=\nabla\theta$
$S_{g}=\bar{g}\int\frac{d^{2}q}{(2\pi)^{2}}\ q^{\sigma}\
|\mathbf{\theta}(\mathbf{q})|^{2}=\bar{g}\int\frac{d^{2}q}{(2\pi)^{2}}\
q^{\sigma-2}\ |\mathbf{v}(\mathbf{q})|^{2}$ (36)
We notice that circling around a topologically non-trivial region we have
$\oint\nabla\theta\cdot d\mathbf{r}=\oint\mathbf{v}\cdot d\mathbf{r}=2\pi
m_{enc}$ (37)
were $m_{enc}$ is the sum of all the topological charges $m_{i}$ enclosed in
the integration contour. This can be rephrased by saying that
$\nabla\times\mathbf{v}(\mathbf{x})=2\pi n(\mathbf{x})$, where
$n(\mathbf{x})=\sum_{j}m_{j}\delta(x-x_{j})$ is the vortex-density and $x_{j}$
correspond to the positions of the vortices. This can be further simplified if
we introduce the dual $\mathbf{u}(\mathbf{x})$ of $\mathbf{v}(\mathbf{x})$,
defined as $u_{j}=\epsilon_{jk}v_{k}$ where $\epsilon_{jk}$ is the fully
antisymmetric tensor of rank $2$. We then find the condition
$\nabla\cdot\mathbf{u}(\mathbf{x})=2\pi n(\mathbf{x})$ (38)
In turn, this can be solved in the Fourier space:
$\mathbf{u}(\mathbf{q})=\frac{2\pi\mathbf{q}}{q^{2}}n(\mathbf{q)}+\mathbf{u}_{\perp}(\mathbf{q})$
(39)
where $\mathbf{u}_{\perp}(\mathbf{q})$ is a generic function such that
$\mathbf{q}\cdot\mathbf{u}_{\perp}(\mathbf{q})=0$ and which represent the
topologically-trivial component of the field $\theta$. Now, since
$|\mathbf{v}(\mathbf{q})|^{2}=|\mathbf{u}(\mathbf{q})|^{2}=\frac{(2\pi)^{2}}{q^{2}}|n(\mathbf{q})|^{2}+|\mathbf{u}_{\perp}(\mathbf{q})|^{2}$
(40)
we have that the action $S_{g}$ splits into the sum on the non-topological and
topological part, the latter being:
$S_{\rm top}=\bar{g}\int d^{2}q\ q^{\sigma-4}|n(\mathbf{q})|^{2}$ (41)
Coming back to the real space we have:
$S_{\rm top}=\bar{g}\sum_{ij}m_{i}m_{j}G(\mathbf{r}_{i}-\mathbf{r}_{j})$ (42)
with $G(\mathbf{x})=\int d^{2}q\
q^{\sigma-4}e^{i\mathbf{q}\cdot\mathbf{x}}\sim L^{2-\sigma}-x^{2-\sigma}$, $L$
being the system size. The first term in $G$ gives raise to a term
proportional to
$L^{2-\sigma}\sum_{i,j}m_{i}m_{j}=L^{2-\sigma}\left(\sum_{i}m_{i}\right)^{2}$
which, in the thermodynamic limit, ensures the neutrality of the gas of
charges. We find then
$S_{\rm
top}\sim-\bar{g}\sum_{ij}m_{i}m_{j}|\mathbf{x}_{i}-\mathbf{x}_{j}|^{2-\sigma}$
(43)
As expected, this interaction is more binding than the logarithmic one for the
short-range case. A simple entropy-energy argument shows that the charges will
never unbound at any temperature: indeed the energetic cost of creating two
far apart vortices grows like $\bar{g}L^{2-\sigma}$ while the entropy as $\ln
L$ so that the free energy
$F\sim\ln L-\bar{g}TL^{2-\sigma}$ (44)
is always dominated by the interaction term for large enough $L$.
## References
* Zinn-Justin [2002] J. Zinn-Justin, 4th ed. (Oxford University Press, Oxford, 2002).
* Kosterlitz [2017] J. M. Kosterlitz, Rev. Mod. Phys. 89, 040501 (2017).
* Kunz and Pfister [1976] H. Kunz and C. E. Pfister, Communications in Mathematical Physics 46, 245 (1976).
* Sak [1973] J. Sak, Phys. Rev. B 8, 281 (1973).
* Luijten and Blöte [2002] E. Luijten and H. W. J. Blöte, Phys. Rev. Lett. 89, 1 (2002).
* Blanchard _et al._ [2013] T. Blanchard, M. Picco, and M. A. Rajabpour, EPL (Europhysics Letters) 101, 56003 (2013), arXiv:1211.6758 [cond-mat.stat-mech] .
* Grassberger [2013] P. Grassberger, Journal of Statistical Physics 153, 289 (2013), arXiv:1305.5940 [cond-mat.stat-mech] .
* Angelini _et al._ [2014] M. C. Angelini, G. Parisi, and F. Ricci-Tersenghi, Phys. Rev. E 89, 062120 (2014).
* Horita _et al._ [2017] T. Horita, H. Suwa, and S. Todo, Phys. Rev. E 95, 012143 (2017).
* Brezin _et al._ [2014] E. Brezin, G. Parisi, and F. Ricci-Tersenghi, Journal of Statistical Physics 157, 855 (2014), arXiv:1407.3358 [cond-mat.stat-mech] .
* Defenu _et al._ [2015] N. Defenu, A. Trombettoni, and A. Codello, Phys. Rev. E 92, 052113 (2015).
* Defenu _et al._ [2016] N. Defenu, A. Trombettoni, and S. Ruffo, Phys. Rev. B 94, 224411 (2016).
* Behan _et al._ [2017] C. Behan, L. Rastelli, S. Rychkov, and B. Zan, Phys. Rev. Lett. 118, 241601 (2017).
* Luijten [1997] E. Luijten, Ph.D. thesis, Technische Universiteit Delft (1997).
* Defenu _et al._ [2020] N. Defenu, A. Codello, S. Ruffo, and A. Trombettoni, Journal of Physics A Mathematical General 53, 143001 (2020).
* Dutta and Bhattacharjee [2001] A. Dutta and J. K. Bhattacharjee, Phys. Rev. B 64, 184106 (2001).
* Defenu _et al._ [2017] N. Defenu, A. Trombettoni, and S. Ruffo, Phys. Rev. B 96, 104432 (2017).
* Bramwell and Holdsworth [1994] S. T. Bramwell and P. C. W. Holdsworth, Phys. Rev. B 49, 8811 (1994).
* Savit [1980] R. Savit, Rev. Mod. Phys. 52, 453 (1980).
* Minnhagen [1987] P. Minnhagen, Rev. Mod. Phys. 59, 1001 (1987).
* Gulácsi and Gulácsi [1998] Z. Gulácsi and M. Gulácsi, Adv. Phys. 47, 1 (1998).
* Amit _et al._ [1980] D. J. Amit, Y. Y. Goldschmidt, and S. Grinstein, J. Phys. A: Math. Gen. 13, 585 (1980).
* Mattis [1984] D. C. Mattis, Physics Letters A 104, 357 (1984).
* Filinov _et al._ [2010] A. Filinov, N. V. Prokof’ev, and M. Bonitz, Phys. Rev. Lett. 105, 070401 (2010).
* Villain [1975] J. Villain, J. Phys. France 36, 581 (1975).
* José _et al._ [1977] J. V. José, L. P. Kadanoff, S. Kirkpatrick, and D. R. Nelson, Phys. Rev. B 16, 1217 (1977).
* V. N. Popov [1983] J. N. V. N. Popov, 1st ed., Mathematical Physics and Applied Mathematics (Kluwer, 1983).
* Berezinskii [1972] V. L. Berezinskii, Sov. Phys. JETP 34, 610 (1972).
* Kosterlitz and Thouless [1973] J. M. Kosterlitz and D. J. Thouless, Journal of Physics C: Solid State Physics 6, 1181 (1973).
* Kosterlitz [1974] J. M. Kosterlitz, J. Phys. C 7, 1046 (1974).
* José [2013] J. V. José, ed., (World Scientific, Singapore, 2013).
* Bishop and Reppy [1978] D. J. Bishop and J. D. Reppy, Phys. Rev. Lett. 40, 1727 (1978).
* Corson _et al._ [1999] J. Corson, R. Mallozzi, J. Orenstein, J. N. Eckstein, and I. Bozovic, Nature (London) 398, 221 (1999).
* Hetel _et al._ [2007] I. Hetel, T. R. Lemberger, and M. Randeira, Nat. Phys. 3, 700 (2007).
* Bilbro _et al._ [2011] L. S. Bilbro, R. V. Aguilar, G. Logvenov, O. Pelleg, I. Boz̆ovic̀, and N. P. Armitage, Nat. Phys. 7, 298 (2011).
* Yong _et al._ [2012] J. Yong, M. J. Hinton, A. McCray, M. Randeria, M. Naamneh, A. Kanigel, and T. R. Lemberger, Phys. Rev. B 85, 180507 (2012).
* Baity _et al._ [2016] P. G. Baity, X. Shi, Z. Shi, L. Benfatto, and D. Popović, Phys. Rev. B 93, 024519 (2016).
* Nitsche _et al._ [2014] W. H. Nitsche, N. Y. Kim, G. Roumpos, C. Schneider, M. Kamp, S. Höfling, A. Forchel, and Y. Yamamoto, Phys. Rev. B 90, 205430 (2014).
* Hadzibabic _et al._ [2006] Z. Hadzibabic, P. Krüger, M. Cheneau, B. Battelier, and J. Dalibard, Nature (London) 441, 1118 (2006).
* Murthy _et al._ [2015] P. A. Murthy, I. Boettcher, L. Bayha, M. Holzmann, D. Kedar, M. Neidig, M. G. Ries, A. N. Wenz, G. Zürn, and S. Jochim, Phys. Rev. Lett. 115, 010401 (2015).
* Reyren _et al._ [2007] N. Reyren, S. Thiel, A. D. Caviglia, L. F. Kourkoutis, G. Hammerl, C. Richter, C. W. Schneider, T. Kopp, A.-S. Rüetschi, D. Jaccard, M. Gabay, D. A. Muller, J.-M. Triscone, and J. Mannhart, Science 317, 1196 (2007).
* Daptary _et al._ [2016] G. N. Daptary, S. Kumar, P. Kumar, A. Dogra, N. Mohanta, A. Taraphder, and A. Bid, Phys. Rev. B 94, 085104 (2016).
* Monteiro _et al._ [2017] A. M. R. V. L. Monteiro, D. J. Groenendijk, I. Groen, J. de Bruijckere, R. Gaudenzi, H. S. J. van der Zant, and A. D. Caviglia, Phys. Rev. B 96, 020504(R) (2017).
* Nisoli and Bishop [2014] C. Nisoli and A. R. Bishop, Phys. Rev. Lett. 112, 070401 (2014).
* Seul _et al._ [1991] M. Seul, L. R. Monar, L. O’Gorman, and R. Wolfe, Science 254, 1616 (1991).
* Mendoza-Coto and Stariolo [2012] A. Mendoza-Coto and D. A. Stariolo, Phys. Rev. E 86, 051130 (2012).
* Monroe _et al._ [2019] C. Monroe, W. C. Campbell, L. M. Duan, Z. X. Gong, A. V. Gorshkov, P. Hess, R. Islam, K. Kim, G. Pagano, P. Richerme, C. Senko, and N. Y. Yao, arXiv e-prints , arXiv:1912.07845 (2019), arXiv:1912.07845 [quant-ph] .
* Britton _et al._ [2012] J. W. Britton, B. C. Sawyer, A. C. Keith, C. C. J. Wang, J. K. Freericks, H. Uys, M. J. Biercuk, and J. J. Bollinger, Nature 484, 489 (2012).
* Baranov _et al._ [2012] M. A. Baranov, M. Dalmonte, G. Pupillo, and P. Zoller, arXiv e-prints , arXiv:1207.1914 (2012), arXiv:1207.1914 [cond-mat.quant-gas] .
* Landig _et al._ [2016] R. Landig, L. Hruby, N. Dogra, M. Land ini, R. Mottl, T. Donner, and T. Esslinger, Nature 532, 476 (2016).
* Botzung _et al._ [2021a] T. Botzung, D. Hagenmüller, G. Masella, J. Dubail, N. Defenu, A. Trombettoni, and G. Pupillo, Phys. Rev. B 103, 155139 (2021a).
* Jurcevic _et al._ [2014] P. Jurcevic, B. P. Lanyon, P. Hauke, C. Hempel, P. Zoller, R. Blatt, and C. F. Roos, Nature 511, 202 (2014).
* Richerme _et al._ [2014] P. Richerme, Z.-X. Gong, A. Lee, C. Senko, J. Smith, M. Foss-Feig, S. Michalakis, A. V. Gorshkov, and C. Monroe, Nature 511, 198 (2014).
* Zhang _et al._ [2017] J. Zhang, P. W. Hess, A. Kyprianidis, P. Becker, A. Lee, J. Smith, G. Pagano, I. D. Potirniche, A. C. Potter, A. Vishwanath, N. Y. Yao, and C. Monroe, Nature 543, 217 (2017).
* Baumann _et al._ [2010] K. Baumann, C. Guerlin, F. Brennecke, and T. Esslinger, Nature 464, 1301 (2010).
* Rovny _et al._ [2018] J. Rovny, R. L. Blum, and S. E. Barrett, Phys. Rev. Lett. 120, 180603 (2018).
* Choi _et al._ [2017] S. Choi, J. Choi, R. Landig, G. Kucsko, H. Zhou, J. Isoya, F. Jelezko, S. Onoda, H. Sumiya, V. Khemani, C. von Keyserlingk, N. Y. Yao, E. Demler, and M. D. Lukin, Nature 543, 221 (2017).
* Safavi-Naini _et al._ [2018] A. Safavi-Naini, R. J. Lewis-Swan, J. G. Bohnet, M. Gärttner, K. A. Gilmore, J. E. Jordan, J. Cohn, J. K. Freericks, A. M. Rey, and J. J. Bollinger, Phys. Rev. Lett. 121, 040503 (2018).
* Keesling _et al._ [2019] A. Keesling, A. Omran, H. Levine, H. Bernien, H. Pichler, S. Choi, R. Samajdar, S. Schwartz, P. Silvi, S. Sachdev, P. Zoller, M. Endres, M. Greiner, Vuletić, V. , and M. D. Lukin, Nature 568, 207 (2019).
* Hauke and Tagliacozzo [2013] P. Hauke and L. Tagliacozzo, Phys. Rev. Lett. 111, 207202 (2013), arXiv:1304.7725 [quant-ph] .
* Vodola _et al._ [2014] D. Vodola, L. Lepori, E. Ercolessi, A. V. Gorshkov, and G. Pupillo, Phys. Rev. Lett. 113, 156402 (2014).
* Maghrebi _et al._ [2016] M. F. Maghrebi, Z.-X. Gong, M. Foss-Feig, and A. V. Gorshkov, Phys. Rev. B 93, 125128 (2016).
* Gong _et al._ [2016] Z.-X. Gong, M. F. Maghrebi, A. Hu, M. Foss-Feig, P. Richerme, C. Monroe, and A. V. Gorshkov, Phys. Rev. B 93, 205115 (2016).
* Defenu _et al._ [2019] N. Defenu, T. Enss, and J. C. Halimeh, Phys. Rev. B 100, 014434 (2019).
* Uhrich _et al._ [2020] P. Uhrich, N. Defenu, R. Jafari, and J. C. Halimeh, Phys. Rev. B 101, 245148 (2020).
* Campa _et al._ [2014] A. Campa, T. Dauxois, D. Fanelli, and S. Ruffo, _Physics of Long-Range Interacting Systems_ (Oxford Univ. Press, 2014).
* Romano [1987] S. Romano, Nuovo Cimento B Serie 100, 447 (1987).
* Giachetti, Guido _et al._ [2021] Giachetti, Guido, Defenu, Nicolò, Ruffo, Stefano, and Trombettoni, Andrea, EPL 133, 57004 (2021).
* Maghrebi _et al._ [2017] M. F. Maghrebi, Z.-X. Gong, and A. V. Gorshkov, Phys. Rev. Lett. 119, 023001 (2017).
* Benfatto _et al._ [2013] L. Benfatto, C. Castellani, and T. Giamarchi, in _40 Years of Berezinskii-Kosterliz-Thouless Theory_, edited by J. V. José (World Scientific, Singapore, 2013).
* Itzykson and Drouffe [1989] C. Itzykson and J.-M. Drouffe, _Statistical field theory. Volume 1, From Brownian motion to renormalization and lattice gauge theory_ (Cambridge, Cambridge University Press, 1989).
* Le Bellac [1991] M. Le Bellac, _Quantum and statistical field theory_ (Oxford, Clarendon, 1991).
* José _et al._ [1977] J. V. José, L. P. Kadanoff, S. Kirkpatrick, and D. R. Nelson, Phys. Rev. B 16, 1217 (1977).
* Spohn and Zwerger [1999] H. Spohn and W. Zwerger, Journal of Statistical Physics 94, 1037 (1999), arXiv:cond-mat/9811219 [cond-mat.stat-mech] .
* Kargol [2005] A. Kargol, Reports on Mathematical Physics 56, 379 (2005).
* Kargol [2014] A. Kargol, Journal of Statistical Mechanics: Theory and Experiment 2014, 10006 (2014).
* Shenoy and Chattopadhyay [1995] S. R. Shenoy and B. Chattopadhyay, Phys. Rev. B 51, 9129 (1995).
* Bighin _et al._ [2019] G. Bighin, N. Defenu, I. Nándori, L. Salasnich, and A. Trombettoni, Phys. Rev. Lett. 123, 100601 (2019).
* Maier and Schwabl [2004] P. G. Maier and F. Schwabl, Phys. Rev. B 70, 134430 (2004).
* Vasilyev _et al._ [2014] A. Vasilyev, A. Tarkhov, L. Men’shikov, P. Fedichev, and U. Fischer, New Journal of Physics 16, 053011 (2014).
* Antenucci _et al._ [2015] F. Antenucci, M. Ibáñez Berganza, and L. Leuzzi, Phys. Rev. B 92, 014204 (2015).
* Jacobs and Savit [1983] L. Jacobs and R. Savit, Annals of the New York Academy of Sciences 410, 281 (1983).
* Defenu _et al._ [2021] N. Defenu, A. Trombettoni, and D. Zappalà, Nuclear Physics B 964, 115295 (2021).
* Raju _et al._ [2019] A. Raju, C. B. Clement, L. X. Hayden, J. P. Kent-Dobias, D. B. Liarte, D. Z. Rocklin, and J. P. Sethna, Phys. Rev. X 9, 021014 (2019).
* Maccari _et al._ [2020] I. Maccari, N. Defenu, L. Benfatto, C. Castellani, and T. Enss, Phys. Rev. B 102, 104505 (2020).
* Ibáñez-Berganza and Leuzzi [2013] M. Ibáñez-Berganza and L. Leuzzi, Phys. Rev. B 88, 144104 (2013).
* Cescatti _et al._ [2019] F. Cescatti, M. Ibáñez-Berganza, A. Vezzani, and R. Burioni, Phys. Rev. B 100, 054203 (2019), arXiv:1905.06688 [cond-mat.stat-mech] .
* Bermudez _et al._ [2017] A. Bermudez, L. Tagliacozzo, G. Sierra, and P. Richerme, Phys. Rev. B 95, 024431 (2017).
* Botzung _et al._ [2021b] T. Botzung, D. Hagenmüller, G. Masella, J. Dubail, N. Defenu, A. Trombettoni, and G. Pupillo, Phys. Rev. B 103, 155139 (2021b).
|
††thanks: Present Address: Centre for Quantum Dynamics, Griffith University,
Brisbane Queensland 4111, Australia
# Heisenberg-Langevin approach to driven superradiance
Ori Somech Department of Chemical & Biological Physics, Weizmann Institute of
Science, Rehovot 7610001, Israel Yoav Shimshi Department of Chemical &
Biological Physics, Weizmann Institute of Science, Rehovot 7610001, Israel
Ephraim Shahmoon Department of Chemical & Biological Physics, Weizmann
Institute of Science, Rehovot 7610001, Israel
###### Abstract
We present an analytical approach for the study of driven Dicke superradiance
based on a Heisenberg-Langevin formulation. We calculate the steady-state
fluctuations of both the atomic-spin and the light-field operators. While the
atoms become entangled below a critical drive, exhibiting spin squeezing, we
show that the radiated light is in a classical-like coherent state whose
amplitude and spectrum are identical to those of the incident driving field.
Therefore, the nonlinear atomic system scatters light as a linear classical
scatterer. Our results are consistent with the recent theory of coherently
radiating spin states. The presented Heisenberg-Langevin approach should be
simple to generalize for treating superradiance beyond the permutation-
symmetric Dicke model.
## I Introduction
Superradiance describes the cooperative radiation of an ensemble of quantum
emitters into common photonic modes. A conceptually simple case that captures
the essence of cooperative radiation is that of Dicke superradiance, where all
the constituents of an ensemble of two-level atoms are coupled to the common
photonic modes in an identical manner, thus forming an effective “collective
spin” dipole Dicke ; mandel_wolf_1995 ; GH . Superradiance was observed both
in atoms HAR ; TOMs1 ; TOMs2 ; FLD ; BRWsr and artificial emitters MAJ and
plays a role in various quantum phenomena and technologies, ranging from phase
transitions KES ; EMAN to narrowband superradiant lasers HAK ; HOL ; TOM1 ;
MOL1 .
The situation wherein the atoms are additionally driven by a resonant laser
can be studied by a driven-dissipative master equation of the Dicke model.
Mean-field theory yields a second order phase transition of the steady-state
atomic population, or “magnetization”, as a function of the drive
DRUMMOND1978160 ; DRUMMOND1980 ; CAR ; LAW ; LAR ; BAR . Spin squeezing was
recently found in steady state by a numerical solution of the master equation
Alejandro ; yelin ; BAR ; REYt with a supporting analytical result obtained
in yelin . For the radiated light, intensity correlations $g^{(2)}$ were
calculated and found to exhibit bunching correlations above the phase-
transition point, but no correlations below it CAR . More recently, it was
found that the appearance of so-called coherently radiating spin states (CRSS)
as the steady-state of driven Dicke superradiance underlies these results CRSS
.
Here we present a simple analytical approach for driven superradiance based on
Heisenberg-Langevin (HL) equations. While this HL approach is in principle
equivalent to the master equation used previously, the HL equations are
natural for the direct analytical treatment of both spin and field
fluctuations via their operator-form solution. In particular, we account for
spin and field fluctuations around the mean field using the Holstein-Primakoff
approximation. For the spin fluctuations the operator-valued solutions are in
a Bogoliubov transformation form implying quantum correlations, as verified by
the subsequent analytical calculation of spin squeezing. For the field
operator, we find that the fluctuations are proportional to the vacuum field,
thus proving that the radiated field below the phase-transition point is in a
coherent state. We also calculate the two-time correlation of the field,
finding that the spectrum is delta-peaked at the incident-drive frequency.
Surprisingly, the light is thus scattered from the many-atom system as if the
latter is a linear system, although the atomic system is highly nonlinear, as
evident by its phase transition. We discuss the consistency and relation of
these results with the predictions of CRSS theory CRSS .
The paper is organized as follows. In Sec. II we derive the HL equations of
the driven Dicke model, focusing on a relevant cavity-scheme realization.
After recalling the mean-field solution in Sec. III, we treat spin
fluctuations and squeezing in Sec. IV. Sec. V is devoted to the analysis of
the radiated light. Finally, our conclusions are presented in Sec. VI.
## II Model
We begin with the derivation of the HL equations of motion that describe the
driven Dicke model, considering a system of atoms in a damped cavity as
realized in typical experiments HAR ; TOMs1 ; TOMs2 ; MAJ . Realizations of
Dicke physics exist also in other systems wherein many atoms are coupled to a
common photon bath, e.g. in waveguide QED Alejandro or even in an elongated
atomic ensemble in free space GH ; BRWsr ; however, the cavity case considered
here is conceptually the most straightforward one as it directly emphasizes a
single common photonic mode.
Figure 1: Cavity realization of driven superradiance. An atomic ensemble is
trapped inside a cavity, wherein all atoms (green dots) are identically
coupled to a cavity mode (lowering operator $\hat{c}$) and hence described by
a collective-spin dipole (lowering operator $\hat{J}_{-}$). The cavity field
is damped through its mirrors at rate $\kappa$ to the outside propagating
modes $\hat{b}_{k}$, which form the radiated field $\hat{E}$, and is driven by
a laser with Rabi-field amplitude $\Omega_{L}$. Here a single-sided cavity
scheme is presented, with one out-coupling mirror (right-hand side).
### II.1 System and Hamiltonian
We consider the system displayed in Fig. 1: $N$ two-level atoms are trapped
inside an optical cavity driven by external laser light thorough the cavity
mirrors. The atomic positions are such that all atoms are identically coupled
to the cavity mode (i.e. well within the cavity mode waist and at longitudinal
positions that are multiples of cavity wavelength apart). The Hamiltonian of
the atoms and the cavity is given by
$\hat{H}_{S}=\hbar\omega_{a}\hat{J}_{z}+\hbar\omega_{c}\hat{c}^{\dagger}\hat{c}+\hbar\left[\hat{c}^{\dagger}\left(g^{\ast}\hat{J}_{-}+\Omega_{L}e^{-i\omega_{L}t}\right)+\text{h.c.}\right].$
(1)
Here $\hat{c}$ is the boson lowering operator of the cavity mode of frequency
$\omega_{c}$, whereas
$\hat{J}_{\alpha}=(1/2)\sum_{n=1}^{N}\hat{\sigma}_{n}^{\alpha}$
($\alpha\in\\{x,y,z\\}$) are the collective-spin operators of the atomic
ensemble with $\hat{\sigma}_{n}^{\alpha}$ being the Pauli operator of a two-
level atom $n\in\\{1,...,N\\}$ with resonant frequency $\omega_{a}$. The
cavity is driven via its mirrors by an external laser of frequency
$\omega_{L}$ and amplitude $\Omega_{L}$, and is coupled to the atoms via the
dipole coupling $g$ identical to all atoms, where
$\hat{J}_{-}=\hat{J}_{x}-i\hat{J}_{y}=\sum_{n=1}^{N}\hat{\sigma}_{n}^{-}=\hat{J}_{+}^{{\dagger}}$
is the collective-spin lowering operator of the atoms and
$\hat{\sigma}_{n}^{-}=(\hat{\sigma}_{n}^{+})^{{\dagger}}$ the Pauli lowering
operator of atom $n$.
In addition, the cavity mode is coupled through its mirrors to a 1D continuum
of propagating photon modes characterized by the wavenumber $k$ and
corresponding boson modes $\hat{b}_{k}$ and frequencies $vk$ ($v$ being the
speed of light). The Hamiltonians describing this 1D photon reservoir and its
coupling to the system are given by, respectively (here, for one-sided cavity,
Fig. 1),
$\displaystyle\hat{H}_{R}$ $\displaystyle=$ $\displaystyle\sum_{k>0}\hbar
vk\hat{b}_{k}^{\dagger}\hat{b}_{k},$ $\displaystyle\hat{H}_{SR}$
$\displaystyle=$
$\displaystyle\hbar\sum_{k>0}\left(\eta\hat{b}_{k}^{\dagger}\hat{c}+\text{h.c.}\right),\quad\eta\equiv\sqrt{\frac{v}{L}\kappa},$
(2)
where the coupling constant $\eta$ is taken $k$-independent (consistent with
the Markov approximation) and $L$ is the quantization length of the 1D
continuum. The total Hamiltonian is given by
$\hat{H}=\hat{H}_{S}+\hat{H}_{R}+\hat{H}_{SR}$. We note that we neglect here
the direct spontaneous emission from atoms to photon modes in transverse
directions outside the cavity. For a dilute ensemble this is an individual-
atom process that is typically much slower than the relevant Dicke dynamics
discussed here.
### II.2 Heisenberg-Langevin equations
We begin with eliminating the reservoir modes $\hat{b}_{k}$ by inserting the
solution of their Heisenberg equations into the equation for $\hat{c}$,
obtaining within the usual Markov approximation SCU
$\dot{\tilde{c}}=\left(i\delta_{c}-\frac{\kappa}{2}\right)\tilde{c}-ig^{\ast}\tilde{J}_{-}-i\Omega_{L}+\hat{E}_{0}(t),\quad\delta_{c}=\omega_{L}-\omega_{c}.$
(3)
Here the system operators are already written in a rotated frame,
$\tilde{c}=\hat{c}e^{i\omega_{L}t}$ and
$\tilde{J}_{-}=\hat{J}_{-}e^{i\omega_{L}t}$, whereas the Langevin, vacuum
noise of the reservoir is given by
$\hat{E}_{0}(t)=-i\sum_{k}\eta^{\ast}e^{-i(vk-\omega_{L})t}\hat{b}_{k}(0)$,
satisfying (assuming an initial vacuum state)
$\langle\hat{E}_{0}(t)\hat{E}_{0}^{{\dagger}}(t^{\prime})\rangle=\kappa\delta(t-t^{\prime}).$
(4)
Next, we eliminate the cavity mode by assuming that its damping rate $\kappa$
to the 1D continuum is much faster than the typical time scale of variations
in $\tilde{J}_{-}$, i.e. $\kappa\gg|\dot{\tilde{J}}_{-}/\tilde{J}_{-}|$.
Within this coarse-grained dynamical picture and for times $t$ much longer
than $1/\kappa$, the elimination of $\tilde{c}$ is equivalent to setting
$\dot{\tilde{c}}=0$ in Eq. (3) and inserting the solution for $\tilde{c}$ into
the Heisenberg equations for atomic variables such as $\tilde{J}_{-}$ and
$\hat{J}_{z}$. Finally, we obtain (denoting
$\tilde{J}_{\mp}\rightarrow\hat{J}_{\mp}$ for simplicity)
$\displaystyle\dot{\hat{J}}_{-}$ $\displaystyle=$ $\displaystyle
i\delta\hat{J}_{-}+\left(\gamma-i2\Delta\right)\hat{J}_{z}\hat{J}_{-}-i2\hat{J}_{z}\left[\Omega+\hat{f}(t)\right],$
$\displaystyle\dot{\hat{J}}_{z}$ $\displaystyle=$
$\displaystyle-\gamma\hat{J}_{+}\hat{J}_{-}+i\hat{J}_{+}\left[\Omega+\hat{f}(t)\right]-i\left[\Omega^{*}+\hat{f}^{{\dagger}}(t)\right]\hat{J}_{-},$
with the laser-atom detuning $\delta=\omega_{L}-\omega_{a}$, the coefficients
$\gamma=\frac{|g|^{2}\kappa}{\delta_{c}^{2}+(\kappa/2)^{2}},\quad\Delta=\frac{-|g|^{2}\delta_{c}}{\delta_{c}^{2}+(\kappa/2)^{2}},\quad\Omega=\frac{-2g\Omega_{L}}{2\delta_{2}+i\kappa},$
(6)
and the effective Langevin, input-vacuum noise (filtered by the cavity),
$\hat{f}(t)\approx[2g/(\kappa-i2\delta_{c})]\hat{E}_{0}(t)$, satisfying
$\langle\hat{f}(t)\hat{f}^{{\dagger}}(t^{\prime})\rangle=\gamma\delta(t-t^{\prime}).$
(7)
Equations (LABEL:HL) form the HL equations of the driven Dicke model, with an
effective emission rate $\gamma$ of an atom to the outside modes via the
cavity, and an effective laser drive with Rabi frequency $\Omega$. The
collective shift $\Delta$ describes the resonant dipole-dipole interactions
between pairs of atoms LEH , corresponding to an effective Hamiltonian
$\hat{H}_{\text{dd}}=-\hbar\sum_{n}\sum_{m}\Delta_{nm}\hat{\sigma}_{n}^{+}\hat{\sigma}_{m}^{-}$.
Here the dipole-dipole kernel $\Delta_{nm}=\Delta$ is uniform for all atom
pairs $n$ and $m$ since all atoms are coupled identically to the mediating
cavity photon mode. In treatments of superradiance in free space, such
coherent dipole-dipole effects are often ignored in free-space GH whereas
they vanish in a waveguide QED superradiance scheme Alejandro . In the cavity
setting, they exist however if one allows for laser-cavity detuning
$\delta_{c}$ as seen in Eq. (6) for $\Delta$ and noted in Refs. BAR ; REYt .
We note that while this specific derivation was performed starting from the
damped-cavity model, equivalent HL equations (LABEL:HL) can be derived by
considering other models of photon continua to which all atoms are identically
coupled. Here the cavity mode effectively becomes a continuum due to its fast
damping rate $\kappa$.
### II.3 Equivalent master equation
The HL equations (LABEL:HL) are equivalent to the following master equation
for the density matrix of the atoms,
$\displaystyle\frac{d\hat{\rho}}{dt}$ $\displaystyle=$
$\displaystyle-\frac{i}{\hbar}\left[\hat{H}_{\text{eff
}},\hat{\rho}\right]+\gamma\left[\hat{J}_{-}\hat{\rho}\hat{J}_{+}-\frac{1}{2}\left(\hat{J}_{+}\hat{J}_{-}\hat{\rho}+\hat{\rho}\hat{J}_{+}\hat{J}_{-}\right)\right],$
$\displaystyle\hat{H}_{\text{eff}}$ $\displaystyle=$
$\displaystyle-\hbar\Delta\hat{J}_{+}\hat{J}_{-}-\hbar\left(\Omega\hat{J}_{+}+\Omega^{\ast}\hat{J}_{-}\right).$
(8)
Here we have already assumed that the laser drive is resonant with the atoms,
$\delta=\omega_{L}-\omega_{a}=0$. This master equation with $\Delta=0$ is a
typical starting point for the analysis of driven Dicke superradiance
presented in previous works DRUMMOND1978160 ; DRUMMOND1980 ; CAR ; LAW ; LAR ;
Alejandro ; yelin ; RABk , whereas the additional dipole-dipole term $\Delta$
is considered in Refs. BAR ; REYt . Here instead we will use the HL
formulation of Eq. (LABEL:HL), in order to derive analytical results for
fluctuations and correlations of atomic and photonic degrees of freedom. We
will use the master equation as a numerical verification of the one-time
correlation functions of the atoms. Since the total spin
$\hat{J}_{x}^{2}+\hat{J}_{y}^{2}+\hat{J}_{z}^{2}=j(j+1)$ is conserved under
the dynamics of Eqs.(LABEL:HL) and (8), the initial state sets the SU(2) spin
representation $j$. Assuming an initial ground state for the $N$ atoms, we
have $j=N/2$ and the Hilbert space that spans Eqs. (8) is of size $2j+1=N+1$
and can be easily solved numerically for reasonable $N$.
## III Mean-field solution
We begin with the mean-field solution of the model in steady state. To obtain
the mean-field equations, we take the average over the HL equations
(LABEL:HL), such that the Langevin vacuum-noise terms vanish, and perform the
factorization of operator products
$\langle\hat{J}_{\alpha}\hat{J}_{\beta}\rangle\approx\langle\hat{J}_{\alpha}\rangle\langle\hat{J}_{\beta}\rangle$
(with $\alpha,\beta\in\\{x,y,z\\}$). This factorization is justified for
$N\rightarrow\infty$ under the mean-field assumption that fluctuations of
observables are much smaller than their mean. It is important to note that
such a factorization does not mean that there are no correlations between the
atoms that comprise the collective spin $\hat{J}_{\alpha}$ DRUMMOND1978160 :
in fact, we see below that the atoms are entangled Alejandro ; yelin ; BAR ;
REYt . Considering the conservation of the total spin
$\hat{J}_{x}^{2}+\hat{J}_{y}^{2}+\hat{J}_{z}^{2}=j(j+1)$ with $j=N/2\gg 1$ and
taking a resonant drive $\delta=0$, the solution to the mean-field equations
becomes (see also BAR ),
$\displaystyle\langle\hat{J}_{z}\rangle=-\frac{N}{2}\sqrt{1-\frac{|\Omega|^{2}}{\Omega_{c}^{2}}},\quad\langle\hat{J}_{-}\rangle=-\frac{\Omega}{\Delta+i\gamma/2},$
(9)
with the critical driving field defined by
$\displaystyle\Omega_{c}=\Omega_{c}(\Delta)=\frac{N}{4}\sqrt{\gamma^{2}+4\Delta^{2}}.$
(10)
The steady-state population inversion (or “magnetization”)
$\langle\hat{J}_{z}\rangle$ thus exhibits a second order phase transition as a
function of the drive $\Omega$, where it vanishes at the critical value
$\Omega_{c}$. The latter increases with the strength of the dipole-dipole
shift $\Delta$ as seen in Eq. (10). For $|\Omega|>\Omega_{c}$ there exist
oscillatory solutions of the mean-field equations DRUMMOND1978160 which
nevertheless appear to decay to zero at long time scales, upon the
consideration of the full quantum problem BAR . Figure 2 displays
$\langle\hat{J}_{z}\rangle$ obtained by the exact numerical solutions of the
master equation (8) for $N=50$ and different values of $\Delta$. Very good
agreement with the mean-field expression (9) is exhibited when $|\Omega|$ is
not too close to the critical point $\Omega_{c}$. In particular, calculations
with different values of $\Delta$ all collapse to the same curve when $\Omega$
is scaled to the corresponding $\Omega_{c}(\Delta)$ from Eq. (10).
Disagreement between mean-field and numerical results is observed around
$\Omega_{c}$ due to the fact that the mean value of
$\langle\hat{J}_{z}\rangle$ near $\Omega_{c}$ becomes increasingly small while
fluctuations grow, in contradiction to the mean-field assumption. The second
order transition predicted by the mean-field solution in the thermodynamic
limit $N\rightarrow\infty$ then becomes smoother at finite $N$.
Figure 2: Population inversion $\langle\hat{J}_{z}\rangle$ of the collective
atomic system as function of the driving field $\Omega$. Results obtained by
the numerical solution of Eq. (8) with $N=50$ atoms and for different values
of the dipole-dipole shift $2\Delta/\gamma=0,1,2$ all collapse to the same
curve when $\Omega$ is scaled the critical field $\Omega_{c}(\Delta)$ from Eq.
(10). The red line represent the analytical mean-field solution from Eq. (9),
exhibiting a second order phase transition. The exact numerical solutions
agree with the mean-field result until they diverge away a bit before the
transition point due to the finite value of $N$.
The mean-field solution (9) can be also written as a mean of the spin vector
$\hat{\mathbf{J}}=(\hat{J}_{x},\hat{J}_{y},\hat{J}_{z})$ in a Bloch sphere,
$\displaystyle\langle\hat{\mathbf{J}}\rangle=\left(\begin{array}[]{c}\langle\hat{J}_{x}\rangle\\\
\langle\hat{J}_{y}\rangle\\\ \langle\hat{J}_{z}\rangle\\\
\end{array}\right)=-\frac{N}{2}\left(\begin{array}[]{c}\sin\theta\cos\phi\\\
\sin\theta\sin\phi\\\ \cos\theta\\\ \end{array}\right),$ (17)
with the angles in spherical coordinates given by
$\displaystyle\sin\theta=\frac{|\Omega|}{\Omega_{c}},\quad\phi=\mathrm{arg}(\Delta+i\gamma/2)-\mathrm{arg}(\Omega).$
(18)
For later purposes, it is instructive to introduce a rotated coordinate system
at which the mean spin vector is directed to the south pole of the Bloch
sphere and hence appears as a ground state in this rotated system. Spin
operators in the rotated system, described by the vector
$\hat{\mathbf{J}}^{\prime}=(\hat{J}^{\prime}_{x},\hat{J}^{\prime}_{y},\hat{J}^{\prime}_{z})$
are related to the original spin operators
$\hat{\mathbf{J}}=(\hat{J}_{x},\hat{J}_{y},\hat{J}_{z})$ via the rotation
matrix $\mathcal{R}$ as
$\displaystyle\hat{\mathbf{J}}^{\prime}=\mathcal{R}^{-1}\hat{\mathbf{J}},\quad\langle\hat{\mathbf{J}}^{\prime}\rangle=-\frac{N}{2}\left(\begin{array}[]{c}0\\\
0\\\ 1\\\ \end{array}\right),$ (22)
$\displaystyle\mathcal{R}=\left(\begin{array}[]{ccc}\cos\theta\cos\phi&-\sin\phi&\sin\theta\cos\phi\\\
\cos\theta\sin\phi&\cos\phi&\sin\theta\sin\phi\\\ -\sin\theta&0&\cos\theta\\\
\end{array}\right).$ (26)
As required, in the rotated system the mean spin vector
$\langle\hat{\mathbf{J}}^{\prime}\rangle$ points to the south pole, defining
the $-z^{\prime}$ axis as the mean spin direction.
## IV Spin fluctuations and squeezing
We now turn to the analysis of small fluctuations of spin variables around the
mean-field solution (Sec. IV A). This will allow us to estimate atomic
correlations such as spin squeezing (Sec. IV B), and later on also the
fluctuations in the scattered field (Sec. V).
### IV.1 Collective spin fluctuations in the Holstein-Primakoff approximation
We recall that within its representation in the rotated system (26), the mean
spin vector
$\hat{\mathbf{J}}^{\prime}=(\hat{J}^{\prime}_{x},\hat{J}^{\prime}_{y},\hat{J}^{\prime}_{z})$
is directed towards the axis $z^{\prime}$ and vanishes along the
$x^{\prime},y^{\prime}$ axes. In order to analyze fluctuations around this
mean, we first define the spin lowering operator in the rotated basis,
$\hat{J^{\prime}}_{-}=\hat{J}^{\prime}_{x}-i\hat{J}^{\prime}_{y}$, and re-
write the HL equations (LABEL:HL) in terms of the rotated-spin operators
$\hat{J}^{\prime}_{-},\hat{J}^{\prime}_{z}$ using the transformation
$\mathcal{R}$ from (26). As in the original basis, the HL in the rotated basis
are also nonlinear in their relevant variables,
$\hat{J}^{\prime}_{-},\hat{J}^{\prime}_{z}$; however, the linearization of the
equations for small fluctuations around the mean field is simpler in this
rotated basis. To this end, we use the Holstein-Primakoff transformation,
which is an exact representation of SU(2) spin operators (here of spin
$j=N/2$) in terms of a bosonic operator $\hat{a}$ (satisfying
$[\hat{a},\hat{a}^{{\dagger}}]=1$) ASA ,
$\hat{J}^{\prime}_{-}=\sqrt{N-\hat{a}^{\dagger}\hat{a}}\
\hat{a},\quad\hat{J}^{\prime}_{z}=\hat{a}^{\dagger}\hat{a}-\frac{N}{2}.$ (27)
We see that the limit $\hat{a}\rightarrow 0$ is that of the mean-field
solution (17), so that the vacuum of $\hat{a}$ is the mean field and $\hat{a}$
describes fluctuations on top of it. In line with the mean-field assumption,
we consider small fluctuations, $|\hat{a}|\sim O(1)\ll\sqrt{N}$, and expand
the nonlinear HL equation for $\hat{J}^{\prime}_{-}$ to leading orders in the
small parameter $1/\sqrt{N}$. This is achieved by the approximation
$\hat{J}^{\prime}_{-}\approx\sqrt{N}\
\hat{a},\quad\hat{J}^{\prime}_{z}\approx-\frac{N}{2},$ (28)
and the subsequent linearization of the HL equation to first orders of
$\hat{a}$ and the noise $\hat{f}$. Finally, we obtain the HL equation for the
spin fluctuations $\hat{a}$,
$\displaystyle\dot{\hat{a}}$ $\displaystyle=$
$\displaystyle-\left(N\frac{\gamma}{2}\cos\theta-
iN\Delta\frac{1+\cos^{2}\theta}{2}\right)\hat{a}-iN\Delta\frac{\sin^{2}{\theta}}{2}\hat{a}^{{\dagger}}$
(29) $\displaystyle+$ $\displaystyle
i\sqrt{N}\left[\frac{1+\cos\theta}{2}e^{i\phi}\hat{f}(t)-\frac{1-\cos\theta}{2}e^{-i\phi}\hat{f}^{{\dagger}}(t)\right].$
This yields coupled linear equations for $\hat{a}$ and $\hat{a}^{{\dagger}}$
whose solution in the steady state for times $t\gg(N\gamma\cos\theta/2)^{-1}$
is
$\displaystyle\hat{a}(t)$ $\displaystyle=$
$\displaystyle\sqrt{N}\left[\frac{1+\cos\theta}{2}e^{i\phi}\hat{B}(t)+\frac{1-\cos\theta}{2}e^{-i\phi}\hat{B}^{{\dagger}}(t)\right],$
$\displaystyle\hat{B}(t)$ $\displaystyle=$ $\displaystyle
i\int_{0}^{t}dt^{\prime}e^{-N\cos\theta\left[\frac{\gamma}{2}-i\Delta\right](t-t^{\prime})}\hat{f}(t^{\prime}).$
(30)
This operator-form solution, along with the correlation function (7) of the
Langevin vacuum-noise $\hat{f}(t)$, now allows to evaluate correlations of the
collective spin. In fact, even without performing specific calculations, the
operator solution itself is already quite insightful. We see that the lowering
operator of the spin fluctuation $\hat{a}$ exhibits a Bogoliubov
transformation form: it is a linear combination of the integrated vacuum noise
lowering operator $\hat{B}$ and its conjugate $\hat{B}^{{\dagger}}$, with
corresponding Bogoliubov coefficients proportional to $1+\cos\theta$ and
$1-\cos\theta$, respectively. Non-trivial, correlated fluctuations occur
whenever $\hat{a}$ contains the conjugate component $\hat{B}^{{\dagger}}$ (and
not only $\hat{B}$), requiring a non-vanishing coefficient $1-\cos\theta$.
Therefore, quantum correlations are expected to grow with the driving field
$|\Omega|/\Omega_{c}=\sin\theta>0$ , as seen explicitly below.
### IV.2 Spin squeezing
A particulary relevant characterization of collective-spin fluctuations is
provided by the spin squeezing parameter spinsqueezingreview ; KitaUeda . Spin
squeezing quantifies fluctuations of the spin vector perpendicular to its mean
direction, and is linked to the sensitivity of quantum-enhanced metrology with
collections of spins spinsqueezingparameter1 ; spinsqueezingparameter2 ; QSr
and their underlying pairwise entanglement LEW ; SOR . Within the rotated spin
representation from (26), where the mean is directed to $-z^{\prime}$, the
spin squeezing parameter is given by spinsqueezingparameter1 ;
spinsqueezingparameter2 ; spinsqueezingreview
$\xi^{2}=\mathrm{min}_{\varphi}\frac{\mathrm{Var}[\hat{J}^{\prime}_{\varphi}]N}{|\langle\hat{J}^{\prime}_{z}\rangle|^{2}},\quad\hat{J}^{\prime}_{\varphi}=\cos\varphi\hat{J}^{\prime}_{x}+\sin\varphi\hat{J}^{\prime}_{y},$
(31)
i.e. it is proportional to the minimal variance of the fluctuations along the
$x^{\prime}y^{\prime}$ plane. Spin squeezing exists for $\xi^{2}<1$, implying
that the collective-spin has improved phase sensitivity to rotations compared
to the standard quantum limit $\xi^{2}=1$ of an uncorrelated coherent spin
state.
Within our mean-field and small-fluctuations assumption, we use
$|\langle\hat{J}^{\prime}_{z}\rangle|\approx N/2$ and the bosonic
approximation (28) for $\hat{J}^{\prime}_{\mp}=\hat{J}^{\prime}_{x}\mp
i\hat{J}^{\prime}_{y}$, to obtain the spin squeezing parameter in terms of the
bosonic operators $\hat{a}$,
$\xi^{2}=1+2\langle\hat{a}^{\dagger}\hat{a}\rangle-2|\langle\hat{a}^{2}\rangle|.$
(32)
Using the solution for $\hat{a}$, Eq. (30), and the Langevin, vacuum-noise
correlation function (7), we then find
$\left|\langle
a^{2}\rangle\right|=\frac{1-\cos^{2}\theta}{4\cos\theta},\quad\langle
a^{\dagger}a\rangle=\frac{(1-\cos\theta)^{2}}{4\cos\theta},$ (33)
so that the spin squeezing parameter, Eq. (32), becomes
$\xi^{2}=\cos\theta=\sqrt{1-\frac{|\Omega|^{2}}{\Omega_{c}^{2}(\Delta)}}.$
(34)
We observe that the spin squeezing is determined by the ratio between the
driving field and the critical field, $|\Omega|/\Omega_{c}$. It depends on the
dipole-dipole interaction $\Delta$ through the critical field
$\Omega_{c}(\Delta)$ from Eq. (10). This generalizes the analytical result of
Ref. yelin , obtained for the case $\Delta=0$ using a master-equation
approach. When the drive is weak $|\Omega|/\Omega_{c}\rightarrow 0$, no spin
squeezing exists, $\xi^{2}=1$, since the system is in a coherent spin state
wherein all atoms are in the ground state. As the drive increases, population
in the atoms is created, such that collective emission is possible, building
entanglement and spin-squeezing correlations between the atoms, $\xi^{2}<1$.
At the critical point $|\Omega|/\Omega_{c}=1$ the spin-squeezing parameter
vanishes: this result is valid only at the limit $N\rightarrow\infty$ where it
does not contradict the Heisenberg limit $\xi^{2}\geqslant 1/N$
spinsqueezingreview . For finite $N$, our mean-field assumption of small
fluctuations breaks down as we approach the critical point, where fluctuations
become increasingly large (e.g.
$\mathrm{max}_{\varphi}\mathrm{Var}[\hat{J}^{\prime}_{\varphi}]\propto
1/\cos\theta$ diverges near the critical point).
It is instructive to compare the analytical result (34) to that obtained by an
exact numerical solution of the master equation for a finite $N$, as explained
above. In Fig. 3 we observe excellent agreement between the analytical and
numerical solutions up to a driving field somewhat below the critical point
$|\Omega|<\Omega_{c}(\Delta)$, above which the two solutions diverge away. As
in Fig. 2, the dependence on $\Delta$ is captured by plotting the numerical
solutions for different values of $\Delta$, which all collapse to the curve as
a function of the driving field $\Omega$ (e.g. taken real) scaled to the
corresponding critical field $\Omega_{c}(\Delta)$, as anticipated analytically
in Eq. (34). We observe that the exact solution obtains its optimal (minimal)
value for the squeezing $\xi^{2}$ close to the point where it begins to
diverge away from the analytical result. Therefore, this optimal value for
$\xi^{2}$ should improve (become smaller) with increasing $N$ yelin . The
scaling of the optimal $\xi^{2}$, being a finite-size effect, cannot be
accounted for by the above mean-field based results (valid for
$N\rightarrow\infty$). This scaling can be obtained analytically using CRSS
theory, yielding $\xi^{2}\sim N^{-1/3}$ CRSS .
Figure 3: Spin squeezing $\xi^{2}$ as function of the driving field $\Omega$.
Results obtained by the numerical solution of Eq. (8) with $N=50$ atoms and
for different values of the dipole-dipole shift $2\Delta/\gamma=0,1,2$ all
collapse to the same curve by when $\Omega$ is scaled the critical field
$\Omega_{c}(\Delta)$ from Eq. (10). The red line represent the analytical
solution from Eq. (34). The exact numerical solutions agree with the
analytical result until they diverge away close to the transition point, where
$\xi^{2}$ begins degrading (growing) with $\Omega$, see main text.
## V Radiated Light
So far we have treated the field degrees of freedom as a reservoir that
generates driven-dissipative dynamics of the atoms. However, superradiance is
essentially a scattering problem of an input coherent-state field off a
collective dipole $\hat{J}_{-}$ formed by the atoms. As such, the total field
exhibits the general form,
$\hat{E}(t)=\hat{E}_{\mathrm{free}}(t)+G\hat{J}_{-}.$ (35)
The first term is the freely propagating coherent-state field in the absence
of atoms, consisting of an average field and vacuum fluctuations. It may
include the influence of linear optical elements such as the cavity mirrors in
the cavity realization of Fig. 1. The second term is the field component
scattered by the atomic dipole $\hat{J}_{-}$, with a coupling coefficient $G$
(describing field propagation from the atoms to the detector). While the first
term exhibits non-correlated coherent-state statistics of the input field, the
second term may exhibit correlations generated by the nonlinearity of the
atoms QNLOr . In superradiance, the considered atomic system is clearly
nonlinear, as we have already seen that the population inversion
$\langle\hat{J}_{z}\rangle$ is a nonlinear function of the driving field
$\Omega$, see Eq. (9). Nevertheless, we show in the following that,
surprisingly, the scattered component of the field is also a coherent-state
field, linear in the input field. This holds for any driving field $\Omega$
smaller than the critical field $\Omega_{c}$.
Although this result is valid for any realization of superradiance, we focus
for concreteness on the cavity realization considered above. We define the
total observable field as the field propagating out of the cavity (in the
rotated frame $\omega_{L}$)
$\hat{E}(t)=-i\sum_{k>0}\eta^{\ast}\hat{b}_{k}(t)e^{i\omega_{L}t}-i\Omega_{L},$
(36)
where $-i\Omega_{L}$ is the average component of the input coherent field.
Using the same HL approach from Sec. II B, we solve for $\hat{E}(t)$ within
the coarse-grained dynamics at $t\gg 1/\kappa$, obtaining Eq. (35) with (see
Appendix),
$\displaystyle\hat{E}_{\mathrm{free}}(t)$ $\displaystyle=$
$\displaystyle\left(1+\chi\right)\left[\hat{E}_{0}(t)-i\Omega_{L}\right],\quad\chi=\frac{\kappa}{i\delta_{c}-\kappa/2},$
$\displaystyle G$ $\displaystyle=$ $\displaystyle-ig^{\ast}\chi.$ (37)
Here $\chi$ describes the linear response of the cavity to the input field
$\hat{E}_{0}(t)-i\Omega_{L}$ (vacuum + coherent drive), which interferes with
the input, yielding the factor $1+\chi$. Therefore, the atom-free field indeed
has the form of a coherent-state field composed of vacuum + average
components. In the following we will show that this turns out to be the case
also for the total field.
### V.1 Average field
Taking the average of Eq. (35), the vacuum term $\hat{E}_{0}$ does not
contribute so that the free-field component from Eq. (37) gives
$-i\Omega_{L}(1+\chi)$. For the scattered part we use $G=-ig^{\ast}\chi$ from
Eq. (37) and $\langle\hat{J}_{-}\rangle$ from Eq. (9) obtaining
$G\langle\hat{J}_{-}\rangle=i\chi\Omega_{L}$. The total average field then
becomes,
$\langle\hat{E}\rangle=-i\Omega_{L},$ (38)
equal to the incident average field. So, the average radiated field in
superradiance is linear in the incident-field amplitude even though the atomic
system is nonlinear, as discussed above.
### V.2 Field fluctuations
The HL approach allows us to gain direct access to field operators which
entail information on the quantum statistics of the field. We will use it here
to show that the fluctuating part of the radiated field is proportional to
vacuum fluctuations, thus proving that the radiated field is in a coherent
state. We first do this by solving for the operators directly, without the
need to infer the statistics from the calculation of correlations.
To this end, we focus on the scattered component of the field, $G\hat{J}_{-}$.
Using the transformation (26), we write $\hat{J}_{-}$ in terms of the rotated-
system spin operators as,
$\displaystyle\hat{J}_{-}$ $\displaystyle=$ $\displaystyle
e^{-i\phi}\left(\frac{\cos\theta+1}{2}\hat{J}^{\prime}_{-}+\frac{\cos\theta-1}{2}\hat{J}^{\prime}_{+}+\sin\theta\hat{J}^{\prime}_{z}\right)$
$\displaystyle\approx$ $\displaystyle
e^{-i\phi}\left(\frac{\cos\theta+1}{2}\sqrt{N}\hat{a}+\frac{\cos\theta-1}{2}\sqrt{N}\hat{a}^{{\dagger}}-\sin\theta\frac{N}{2}\right).$
In the second line we have used the Holstein-Primakoff linearization, Eq.
(28). Plugging in the solution for $\hat{a}$ from Eq. (30), we then obtain for
the fluctuating part of the field $\hat{E}$ from (35)
$\displaystyle\hat{\mathcal{E}}(t)\equiv\hat{E}-\langle\hat{E}\rangle=\left(1+\chi\right)\hat{E}_{0}(t)+GN\cos\theta\hat{B}(t).$
(40)
The first term describes the vacuum fluctuations $\propto\hat{E}_{0}$ of the
coherent free-field component from Eq. (37). The second term originates from
the fluctuating part of the scattered field $G\hat{J}_{-}$ from Eq.
(LABEL:Jm1) and is also essentially proportional to integrated vacuum
fluctuations $\hat{E}_{0}$ [noting that $\hat{B}$ in Eq. (30) is an integral
of $\hat{f}\propto\hat{E}_{0}$]. This proves that the total radiated field is
in a coherent state, comprised of vacuum fluctuations on top of a mean
coherent amplitude.
### V.3 Light squeezing vs. spin squeezing
Since the radiated field is a classical-like coherent state, it does not
exhibit any quantum correlations. We now show this explicitly for the case of
quantum squeezing correlations. Defining the quadrature operator of the
radiated field,
$\hat{X}_{\varphi}=e^{-i\varphi}\hat{E}+e^{i\varphi}\hat{E}^{{\dagger}}$, the
bosonic squeezing parameter of the field is given by
$\displaystyle\xi^{2}_{E}=\mathrm{min}_{\varphi}\frac{\mathrm{Var}[\hat{X}_{\varphi}]}{V_{0}}=1+\frac{2}{V_{0}}\left(\langle\hat{\mathcal{E}}^{{\dagger}}\hat{\mathcal{E}}\rangle-|\langle\hat{\mathcal{E}}^{2}\rangle|\right),$
(41)
with
$V_{0}\equiv[\hat{E},\hat{E}^{{\dagger}}]=[\hat{E}_{0},\hat{E}^{{\dagger}}_{0}]=\kappa\delta(t=0)$
being the vacuum-noise level. Squeezed quantum noise and correlations exist if
the quadrature noise can become lower than that of the vacuum, i.e. for
$\xi^{2}_{E}<1$. It is seen that this requires the existence of the phase-
dependent correlator $\langle\hat{\mathcal{E}}^{2}\rangle$. Similarly, spin
squeezing in Eq. (32) requires the existence of the phase-dependent correlator
of spin fluctuations $\langle\hat{a}^{2}\rangle$. For either of these
correlators to exist, the corresponding lowering operators $\hat{\mathcal{E}}$
and $\hat{a}$ then must contain a raising field-operator $\hat{B}^{{\dagger}}$
(equivalently, $\hat{E}_{0}^{{\dagger}}$) in addition to $\hat{B}$, since the
average is performed over the initial vacuum state. While the Bogoliubov
coefficient $1-\cos\theta$ in Eq. (30) indeed guarantees that $\hat{a}$
contains $\hat{B}^{{\dagger}}$ for any finite drive
$|\Omega|/\Omega_{c}=\sin\theta<1$, this is not the case for the field
fluctuations $\hat{\mathcal{E}}$: The transformation coefficients in Eq.
(LABEL:Jm1) from $\hat{a},\hat{a}^{{\dagger}}$ to
$\hat{J}_{-}\sim\hat{\mathcal{E}}$, which also depend on $\cos\theta$, lead to
an exact cancellation of the coefficient for $\hat{B}^{{\dagger}}$ in
$\hat{\mathcal{E}}$, as seen in Eq. (40). Therefore, for any drive strength
$|\Omega|/\Omega_{c}=\sin\theta<1$, spin squeezing exists while light
squeezing exactly cancels. This result is equivalent to the geometrical
interpretation given by the so-called dipole-projected squeezing CRSS .
### V.4 Spectrum
Having access to the field operator $\hat{E}(t)$, the HL approach also allows
to directly calculate two-time correlations and spectra. The spectrum of the
radiated field in a steady state time $t$ is given as usual by the Fourier
transform on the time-difference $\tau$ of the two-time correlation,
$\langle\hat{E}^{{\dagger}}(t)\hat{E}(t+\tau)\rangle$. Since this is a normal-
ordered correlator, the fluctuating part of $\hat{E}$ in Eq. (40) drops, as it
is proportional to the lowering operator $\hat{E}_{0}$. This trivially yields
$\langle\hat{E}^{{\dagger}}(t)\hat{E}(t+\tau)\rangle=|\langle\hat{E}\rangle|^{2}=|\Omega_{L}|^{2}$.
The spectrum of the radiated field is then a single delta peak at the incident
frequency $\omega_{L}$ (recalling we work in the laser-rotated frame). This
again shows that the collective atomic dipole scatters light as a linear
optical element even though the atomic population exhibits a strongly
nonlinear dependence on the drive $\Omega_{L}$.
## VI Conclusions
In this work we have presented a HL approach to driven Dicke superradiance in
steady state. The analytical results for steady-state spin squeezing agree and
generalize those obtained in Refs. Alejandro ; yelin ; BAR ; REYt .
Furthermore, our finding that the radiated field is in a coherent state
underlies previous results on uncorrelated photon statistics below the
transition CAR . These HL-based results are consistent with the formation of a
CRSS as described in CRSS . The HL approach is thus complementary to the CRSS
description of superradiance. On the one hand, it is based on the approximate
analysis of small fluctuations around the mean field for $N\rightarrow\infty$,
and did not yield the finite-size scalings with $N$ or the full atomic state
as in CRSS. But on the other hand, it is simpler to generalize for treating
superradiance beyond the permutation-symmetric Dicke case, e.g. by performing
the Holstein-Primakoff approximation for each individual atom separately,
while allowing for direct estimation of atom and field correlations.
*
## Appendix A Output field
Here we elaborate on the derivation of the general expression for the output
field, Eqs. (35) and (37), in the one-sided cavity scheme of Fig. 1. We begin
by defining the outside propagating field
$\displaystyle\hat{E}(x,t)=-i\sum_{k>0}\eta^{\ast}\hat{b}_{k}(t)e^{ikx}e^{i\omega_{L}t}-i\Omega_{L}e^{i\frac{\omega_{L}}{v}x}.$
(42)
Here $x$ is the propagation axis: in the one-sided scheme, $x=0$ denotes the
position of the outcoupling mirror (right-hand side mirror in Fig. 1), so that
$x<0$ denotes incoming left-propagating fields whereas $x>0$ denotes outgoing
right-propagating fields. The radiated field from Eq. (36) is then defined by
taking $x=0^{+}>0$. As in the derivation of the HL equations in Sec. II B, we
first formally solve the Heisenberg equations for $\hat{b}_{k}(t)$, obtaining
in the Markov approximation
$\displaystyle\hat{E}(x,t)$ $\displaystyle=$
$\displaystyle\hat{E}_{0}(x,t)-i\Omega_{L}e^{i\frac{\omega_{L}}{v}x}$ (43)
$\displaystyle-$ $\displaystyle
e^{i\frac{\omega_{L}}{v}x}\kappa\int_{0}^{t}dt^{\prime}\tilde{c}(t^{\prime})\delta(t-x/c-t^{\prime}).$
Here $\hat{E}_{0}(x,t)$ is the vacuum field from Eq. (4) with the exponentials
$e^{ikx}$ in the mode expansion $\hat{b}_{k}(0)$. For $x=0^{-}<0$, the Dirac
delta function does not contribute and we indeed obtain the input field
$\hat{E}_{0}(t)-i\Omega_{L}$. For $x=0^{+}>0$ we obtain
$\displaystyle\hat{E}(t)\equiv\hat{E}_{0}(0^{+},t)=\hat{E}_{0}(t)-i\Omega_{L}-\kappa\tilde{c}(t).$
(44)
Finally, inserting the coarse-grained solution for $\tilde{c}(t)$ [obtained
for simplicity by setting $\dot{\tilde{c}}=0$ in Eq. (3)], we arrive at Eqs.
(35) and (37).
###### Acknowledgements.
We acknowledge financial support from the Israel Science Foundation (ISF)
grant No. 2258/20, the ISF and the Directorate for Defense Research and
Development (DDR&D) grant No. 3491/21, the Center for New Scientists at the
Weizmann Institute of Science, the Council for Higher Education (Israel), and
QUANTERA (PACE-IN). This research is made possible in part by the historic
generosity of the Harold Perlman Family.
## References
* (1) R. H. Dicke, “Coherence in spontaneous radiation processes,” Phys. Rev., vol. 93, pp. 99–110, Jan 1954.
* (2) L. Mandel and E. Wolf, Optical Coherence and Quantum Optics. Cambridge University Press, 1995.
* (3) M. Gross and S. Haroche, “Superradiance: An essay on the theory of collective spontaneous emission,” Physics Reports, vol. 93, no. 5, pp. 301–396, 1982\.
* (4) Y. Kaluzny, P. Goy, M. Gross, J. Raimond, and S. Haroche, “Observation of self-induced rabi oscillations in two-level atoms excited inside a resonant cavity: The ringing regime of superradiance,” Physical review letters, vol. 51, no. 13, p. 1175, 1983.
* (5) M. A. Norcia, M. N. Winchester, J. R. Cline, and J. K. Thompson, “Superradiance on the millihertz linewidth strontium clock transition,” Science advances, vol. 2, no. 10, p. e1601231, 2016.
* (6) M. A. Norcia, J. R. Cline, J. A. Muniz, J. M. Robinson, R. B. Hutson, A. Goban, G. E. Marti, J. Ye, and J. K. Thompson, “Frequency measurements of superradiance from the strontium clock transition,” Physical Review X, vol. 8, no. 2, p. 021036, 2018.
* (7) D. D. Grimes, S. L. Coy, T. J. Barnum, Y. Zhou, S. F. Yelin, and R. W. Field, “Direct single-shot observation of millimeter-wave superradiance in rydberg-rydberg transitions,” Physical Review A, vol. 95, no. 4, p. 043818, 2017.
* (8) G. Ferioli, A. Glicenstein, I. Ferrier-Barbut, and A. Browaeys, “Observation of a non-equilibrium superradiant phase transition in free space,” arXiv preprint arXiv:2207.10361, 2022.
* (9) A. Angerer, K. Streltsov, T. Astner, S. Putz, H. Sumiya, S. Onoda, J. Isoya, W. J. Munro, K. Nemoto, J. Schmiedmayer, et al., “Superradiant emission from colour centres in diamond,” Nature Physics, vol. 14, no. 12, pp. 1168–1172, 2018.
* (10) E. M. Kessler, G. Giedke, A. Imamoglu, S. F. Yelin, M. D. Lukin, and J. I. Cirac, “Dissipative phase transition in a central spin system,” Physical Review A, vol. 86, no. 1, p. 012116, 2012.
* (11) P. Kirton, M. M. Roses, J. Keeling, and E. G. Dalla Torre, “Introduction to the dicke model: From equilibrium to nonequilibrium, and vice versa,” Advanced Quantum Technologies, vol. 2, no. 1-2, p. 1800043, 2019.
* (12) F. Haake, M. I. Kolobov, C. Seeger, C. Fabre, E. Giacobino, and S. Reynaud, “Quantum noise reduction in stationary superradiance,” Physical Review A, vol. 54, no. 2, p. 1625, 1996.
* (13) D. Meiser, J. Ye, D. Carlson, and M. Holland, “Prospects for a millihertz-linewidth laser,” Physical review letters, vol. 102, no. 16, p. 163601, 2009.
* (14) J. G. Bohnet, Z. Chen, J. M. Weiner, D. Meiser, M. J. Holland, and J. K. Thompson, “A steady-state superradiant laser with less than one intracavity photon,” Nature, vol. 484, no. 7392, pp. 78–81, 2012.
* (15) K. Debnath, Y. Zhang, and K. Mølmer, “Lasing in the superradiant crossover regime,” Physical Review A, vol. 98, no. 6, p. 063837, 2018.
* (16) P. Drummond and H. Carmichael, “Volterra cycles and the cooperative fluorescence critical point,” Optics Communications, vol. 27, no. 1, pp. 160–164, 1978.
* (17) P. D. Drummond, “Observables and moments of cooperative resonance fluorescence,” Phys. Rev. A, vol. 22, pp. 1179–1184, Sep 1980.
* (18) H. Carmichael, “Analytical and numerical results for the steady state in cooperative resonance fluorescence,” Journal of Physics B: Atomic and Molecular Physics (1968-1987), vol. 13, no. 18, p. 3551, 1980.
* (19) R. Puri and S. Lawande, “Exact steady-state density operator for a collective atomic system in an external field,” Physics Letters A, vol. 72, no. 3, pp. 200–202, 1979.
* (20) J. Hannukainen and J. Larson, “Dissipation-driven quantum phase transitions and symmetry breaking,” Physical Review A, vol. 98, no. 4, p. 042113, 2018\.
* (21) D. Barberena, R. J. Lewis-Swan, J. K. Thompson, and A. M. Rey, “Driven-dissipative quantum dynamics in ultra-long-lived dipoles in an optical cavity,” Physical Review A, vol. 99, no. 5, p. 053411, 2019.
* (22) A. González-Tudela and D. Porras, “Mesoscopic entanglement induced by spontaneous emission in solid-state quantum optics,” Phys. Rev. Lett., vol. 110, p. 080502, Feb 2013.
* (23) T. E. Lee, C.-K. Chan, and S. F. Yelin, “Dissipative phase transitions: Independent versus collective decay and spin squeezing,” Phys. Rev. A, vol. 90, p. 052109, Nov 2014.
* (24) K. Tucker, D. Barberena, R. J. Lewis-Swan, J. K. Thompson, J. G. Restrepo, and A. M. Rey, “Facilitating spin squeezing generated by collective dynamics with single-particle decoherence,” Phys. Rev. A, vol. 102, p. 051701, Nov 2020.
* (25) O. Somech and E. Shahmoon, “Quantum entangled states of a classically radiating macroscopic spin,” arXiv preprint arXiv:2204.05455, 2022.
* (26) M. O. Scully and M. S. Zubairy, “Quantum optics,” 1999.
* (27) R. Lehmberg, “Radiation from an n-atom system. i. general formalism,” Physical Review A, vol. 2, no. 3, p. 883, 1970.
* (28) J. Huber, P. Kirton, and P. Rabl, “Phase-space methods for simulating the dissipative many-body dynamics of collective spin systems,” SciPost Physics, vol. 10, no. 2, p. 045, 2021.
* (29) A. Auerbach, Interacting electrons and quantum magnetism. Springer Science & Business Media, 2012.
* (30) J. Ma, X. Wang, C. Sun, and F. Nori, “Quantum spin squeezing,” Physics Reports, vol. 509, no. 2, pp. 89–165, 2011.
* (31) M. Kitagawa and M. Ueda, “Squeezed spin states,” Physical Review A, vol. 47, no. 6, p. 5138, 1993.
* (32) D. J. Wineland, J. J. Bollinger, W. M. Itano, F. L. Moore, and D. J. Heinzen, “Spin squeezing and reduced quantum noise in spectroscopy,” Phys. Rev. A, vol. 46, pp. R6797–R6800, Dec 1992.
* (33) D. J. Wineland, J. J. Bollinger, W. M. Itano, and D. J. Heinzen, “Squeezed atomic states and projection noise in spectroscopy,” Phys. Rev. A, vol. 50, pp. 67–88, Jul 1994.
* (34) C. L. Degen, F. Reinhard, and P. Cappellaro, “Quantum sensing,” Reviews of modern physics, vol. 89, no. 3, p. 035002, 2017.
* (35) J. Korbicz, J. I. Cirac, and M. Lewenstein, “Spin squeezing inequalities and entanglement of n qubit states,” Physical review letters, vol. 95, no. 12, p. 120502, 2005.
* (36) A. Sørensen, L.-M. Duan, J. I. Cirac, and P. Zoller, “Many-particle entanglement with bose–einstein condensates,” Nature, vol. 409, no. 6816, pp. 63–66, 2001.
* (37) D. E. Chang, V. Vuletić, and M. D. Lukin, “Quantum nonlinear optics—photon by photon,” Nature Photonics, vol. 8, no. 9, pp. 685–694, 2014.
|
# Qualitative quantum simulation of resonant tunneling and localization with
the shallow quantum circuits
P. Wang<EMAIL_ADDRESS>Department of Mathematics and Physics, North China
Electric Power University, 102206 Beijing, China
###### Abstract
In a circuit-based quantum computer, the computing is performed via the
discrete-time evolution driven by quantum gates. Accurate simulation of
continuous-time evolution requires a large number of quantum gates and
therefore suffers from more noise. In this paper, we find that shallow quantum
circuits are sufficient to qualitatively observe some typical quantum
phenomena in the continuous-time evolution limit, such as resonant tunneling
and localization phenomena. We study the propagation of a spin excitation in
Trotter circuits with a large step size. The circuits are formed of two types
of two-qubit gates, i.e. XY gates and controlled-$R_{x}$ gates, and single-
qubit $R_{z}$ gates. The configuration of the $R_{z}$ gates determines the
distribution of the spin excitation at the end of evolution. We demonstrate
the resonant tunneling with up to four steps and the localization phenomenon
with dozens of steps in Trotter circuits. Our results show that the circuit
depth required for qualitative observation of some significant quantum
phenomena is much smaller than that required for quantitative computation,
suggesting that it is feasible to apply qualitative observations to near-term
quantum computers. We also provide a way to use the physics laws to understand
the error propagation in quantum circuits.
###### pacs:
11.30.Er, 03.65.Nk, 03.65.-w, 42.82.Et
Keywords: Qualitative quantum simulation, shallow quantum circuits, resonant
tunneling, localization, error propagation
## 1 Introduction
Quantum computing can be used to investigate quantum systems as a universal
simulator [1, 2, 3]. In quantum mechanics, the time evolution of quantum
states is driven by a Hamiltonian and described by a unitary operator. In a
digital quantum computer, the computing is carried out by using a set of basic
quantum gates, and usually each gate is a single-qubit or two-qubit unitary
operator. The combination of these basic gates allows us to implement the
evolution operator of a multi-qubit system. A specific approach is Trotter-
Suzuki decomposition [4, 5, 6, 7], in which we approximate the continuous-time
evolution with a discrete-time evolution. For a general local-interaction
Hamiltonian, we can explicitly construct the evolution operator for a short
time, i.e. one time step, from quantum gates. By repetitive gates of one time
step for $N_{T}$ times, we realize the target time evolution. With a smaller
step size, the discrete-time evolution is closer to continuous-time evolution,
however, this requires a larger $N_{T}$, i.e. more quantum gates. Considering
a practical device [8], quantum computing is inaccurate due to decoherence and
imperfect control, and usually the error increases with the gate number [9,
10, 11]. Fault-tolerant quantum computing using quantum error correction is
able to remove the error but impractical using today’s technologies, because
of the large qubit overhand for encoding [12, 13, 14, 15, 16, 17, 18, 19, 20,
21, 22, 23, 24, 25]. A family of practical methods have been developed to
mitigate errors, however the gate number is usually limited due to the finite
error rate on the physical level [26, 27, 35, 29, 30, 31, 32, 33, 34, 35, 36].
Therefore we can only realize the discrete-time Trotter evolution with a small
number of Trotter steps. This motivates researches on the effect of large step
size, i.e. few Trotter steps. Trotter errors induced by large step sizes in
digital quantum simulation have received extensive attention[37, 38, 39, 40,
41]. Some studies show that the Trotter step sizes can separate quantum
chaotic phase from localized phase and comparatively large Trotter steps can
retain controlled errors for local observables[42, 43].
In this paper, we are interested in, when the step size is large (or
equivalently the steps are few), whether some typical physical effects in the
limit of continuous-time can still be observed. The physical effects that we
focus on are the resonant tunneling phenomenon and the localization in
disordered systems [44, 45]. We find that, in a large-step-size Trotter
circuit, the resonant tunneling with $n$ resonant peaks can be observed in
circuits with $n+1$ Trotter steps. Experiments on an IBM quantum computer are
implemented to demonstrate the resonant tunneling with up to three peaks. We
also study the spin transport with the disordered configurations of the
$R_{z}$ gates (we will specify these gates later) using the large step size.
The numerical simulation of circuits with $15$ qubits and tens of Trotter
steps exhibits the localization in the disordered configuration. The results
indicate that shallow quantum circuits on near-term quantum computers are
sufficient to qualitatively simulate some significant physical phenomena. The
localization phenomenon of the spin excitation distribution implies that the
bit-flip error does not affect the measurement on distant qubits if the
configuration of the $R_{z}$ gates is disordered. These conclusions can be
generalized if we replace XY gates with controlled-$R_{x}$ gates, which can
transform one spin excitation into multiple spin excitations.
This paper is organized as follows. In Sec. 2, we discuss the quantum
transverse-field XY model and corresponding quantum circuits, the map between
the Hamiltonian of model and corresponding circuit is established in the limit
of small Trotter step size. In Sec. 3, we discuss the propagation of the spin
excitation, and compare resonant tunneling effects in circuits in the small-
step-size limit and the large-step-size limit. In Sec. 4, we investigate the
transport of the spin excitation in the ordered and disordered configurations
of single-qubit $R_{z}$ gates. Conclusion is given at the end of the paper.
Figure 1: (Color online) (a) The schematic diagram of quantum circuit, which
includes the initialization, $N_{T}$ Trotter steps, and measurements. The
preparation for the initial state is in the red dashed box including a NOT
gate (i.e. the X gate). The blue rectangle represents a layer of two-qubit
gates. The orange rectangles represent single-qubit $R_{z}$ gates. (b) The
schematic diagram of a layer of two-qubit gates. (c) The matrix
representations for two types of 2-qubit gates: XY gates and
controlled-$R_{x}$ gates.
## 2 Model
In this paper, we study the particle transport in the discrete-time evolution
in the quantum transverse-field XY model. The purpose of this study is to
investigate whether some typical quantum phenomena occurring in the
continuous-time limit can be qualitatively observed in discrete-time evolution
when the step size is large. The typical quantum phenomena we concerned here
including resonant tunneling and localization effect, which caused by
interference during the particle transport. Generally, the quantum transverse-
field XY model can be used to describe the propagation of spin excitations, or
particle transport when spin excitations can be treated as particles [46].
Additionally, in the quantum transverse-field XY model, the corresponding
discrete-time evolution operator can be mapped into a quantum circuit
according to the Trotter-Suzuki decomposition. Therefore, we investigate the
particle transport with the quantum transverse-field XY model
$H=\sum_{j}^{N-1}J_{j}H_{j,j+1}^{XY}+\sum_{j}^{N}V_{j}H_{j}^{Z},$ (1)
where $j$ is the lattice site, $N$ is the system size, $J_{j}$ is the
interaction strength and $V_{j}$ is the transverse-field strength.
$H_{j,j+1}^{XY}=(\sigma_{j}^{x}\sigma_{j+1}^{x}+\sigma_{j}^{y}\sigma_{j+1}^{y})/4$
and $H_{j}^{Z}=\sigma_{j}^{z}/2$, where $\sigma_{j}^{i}$ $(i=x,y,z)$
represents the Pauli matrix at the $j$th site. In the case of $J_{j}<0$ and
$V_{j}<0$, the ground state of the transverse-field XY chain is
$\left|00...0\right\rangle$. In this work, we consider the time evolution of
the initial state $\left|\psi(0)\right\rangle=\left|10...0\right\rangle$ which
represents a spin excitation on the first site. The time evolution is in the
subspace of single spin excitation.
The time evolution operator $U(t)$ of quantum transverse-field XY model can be
approximated with a quantum circuit. According to Trotter-Suzuki
decomposition, $U(t)$ can be expanded approximately
$U(t)=e^{-iHt}\approx\left(\prod_{j=1}^{N-1}U_{j,j+1}^{XY}(J_{j}\tau)\prod_{j=1}^{N}U_{j}^{Z}(V_{j}\tau)\right)^{N_{T}},$
(2)
where $N_{T}$ is the number of Trotter steps, $\tau=t/N_{T}$ is the size of
each Trotter step. $U_{j,j+1}^{XY}(\theta_{j})=e^{-iH_{j,j+1}^{XY}\theta_{j}}$
and $U_{j}^{Z}\left(\phi_{j}\right)=e^{-iH_{j}^{Z}\phi_{j}}$, where
$\theta_{j}=J_{j}\tau$ and $\phi_{j}=V_{j}\tau$.
Figure 2: (Color online) (a1)-(d1) The circuit systems. The parameters of
$R_{z}$ gates are marked on the orange squares. For convenience, in (c1) and
(d1) we use single green rectangle to denote a layer of two-qubit gates.
(a2)-(d2) The quantum wells. The energy levels of the wells are denoted by the
horizontal lines in the wells. (a3)-(d3) The tight-binding chain. (a4)-(d4)
Numerical result of the discrete-time evolution. The solid curves represent
the probability of observing the spin excitation on the last qubit after the
time evolution. The curves with circles are the results obtaining on a IBM
quantum computer. The curves with light colors are the numerical result of the
continuous-time evolution.
The corresponding discrete-time evolution is realized with the quantum circuit
as shown in Fig. 1(a). The time evolution of each term, i.e. $U^{XY}$ and
$U^{Z}$, are two-qubit XY gate and single-qubit $R_{z}$ gate, respectively.
The circuit has $N$ qubits $\\{q_{1},q_{2},...,q_{N}\\}$ and $N_{T}$ Trotter
steps, and every Trotter step contains one layer of two-qubit XY gates and one
layer of single-qubit $R_{z}$ gates (We neglect $R_{z}$ gates in the last
Trotter step, because these gates does not have any effect on the distribution
of the spin excitation). In a quantum circuit, a NOT gate (i.e. the Pauli
$\sigma_{x}$ matrix) can flip the qubit $\left|0\right\rangle$ to
$\left|1\right\rangle$, therefore we prepare the initial state
$\left|\psi(0)\right\rangle$ by applying a NOT gate on the first qubit, i.e.
$\left|\psi(0)\right\rangle=\sigma_{1}^{x}\left|0\right\rangle^{N}=\left|10...0\right\rangle$
[see the dashed rectangle in Fig. 1(a)]. The matrix representation of XY gate
is shown in Fig. 1(c).
The behavior of a spin excitation under discrete-time evolution depends on
step size. When the Trotter step size is sufficiently small, the propagation
of a spin excitation in quantum circuit is equivalent to the particle
transport in continuous-time evolution. In this case, the propagation of spin
excitation can exhibit some typical physical phenomena in particle transport.
A natural question to ask is, in shallow quantum circuits with a large Trotter
step size, whether some physical phenomena during continuous-time evolution
nevertheless remains, so that we can observe these phenomena using fewer
quantum gates. In the following text, we show that we observe the resonant
tunneling and localization effect in shallow circuits even if the Trotter step
size is large.
## 3 Resonant tunneling for large Trotter step size
In this section, we study the resonance phenomenon related to the transport of
the spin excitation during the discrete-time evolution with a large step size.
We study the situation where the size of the circuit system is $N=2,3,4,5$
respectively, where the spin excitation is created by a NOT gate and
transmitted through the XY gate. In each one case, we qualitatively find the
resonance tunneling in the limit of continuous-time evolution and give the
corresponding minimum number of Trotter steps.
We also study the transport of a spin excitation through the
controlled-$R_{x}$ gate. In quantum computing, computational errors occur due
to the imperfect control and decoherence always error. The typical one is bit-
flip error corresponding to an unwanted NOT gate. So studying the transmission
of the spin excitation can help us understand the propagating of bit-flip
errors. In actual quantum circuits, errors are not only transmitted, but also
replicated. For example, a single-qubit error will become a multi-qubit error
after passing through the controlled-$R_{x}$ gate. This effect potentially has
a greater impact in quantum computing, because multi-qubit errors can lead to
the failure of quantum error correction. So in this section, we also study the
behavior of a spin excitation propagating through the controlled-$R_{x}$ gate
and find the resonance phenomenon similar to that propagating through the XY
gate.
### 3.1 Resonant tunneling in two-qubit circuit
We first discuss the case of $N=2$. The Hamiltonian of the transverse-field XY
model with 2 sites is
$H=J(\sigma_{1}^{x}\sigma_{2}^{x}+\sigma_{1}^{y}\sigma_{2}^{y})/4+V_{1}\sigma_{1}^{z}/2+V_{2}\sigma_{2}^{z}/2.$
(3)
According to the Trotter-Suzuki decomposition, the time evolution operator can
be approximated using a sequence of quantum gates
$U(t)=e^{-iHt}\approx[U_{1,2}^{XY}(J\tau)U_{1}^{Z}(V_{1}\tau)U_{2}^{Z}(V_{2}\tau)]^{N_{T}},$
(4)
where
$U_{j,j+1}^{XY}(J_{j}\tau)=e^{-i(\sigma_{j}^{x}\sigma_{j+1}^{x}+\sigma_{j}^{y}\sigma_{j+1}^{y})J_{j}\tau}$,
$U_{j}^{Z}\left(V_{j}\tau\right)=e^{-i\sigma_{1}^{z}V_{j}\tau/2}$, and
$\tau=t/N_{T}$. The right side of ”$\approx$” in Eq. (4) represents the
discrete-time evolution. The accuracy of the approximation increases with
$N_{T}$. In Fig. 2(a1), we show the schematic diagram of the quantum circuit
corresponding to discrete-time evolution when $N_{T}=2$ (the last two $R_{z}$
gates have been ignored, and so do the other systems). As we can see, the
yellow squares represent that the qubits are initialized to
$\left|00\right\rangle,$ which is the ground state of $H$ in the case of
$V_{1},V_{2}<0$ and $J\ll V_{1}+V_{2}$. Following the initialization, a NOT
gate on the first qubit flips $\left|00\right\rangle$ into
$\left|10\right\rangle$, which represents that there is a spin excitation on
the first qubit. $\left|10\right\rangle$ is the initial state that we want to
prepare. Due to the symmetry of Hamiltonian, the spin excitation lies in the
subspace spanned by {$\left|10\right\rangle$, $\left|01\right\rangle$}. In
this single-spin-excitation subspace, we can regard a spin excitation as a
particle moving in a $2$-site tight-binding chain [see Fig. 2(a3)], and the
corresponding Hamiltonian is
$H=J\left(\left|10\right\rangle\left\langle
01\right|+\left|01\right\rangle\left\langle
10\right|\right)+V_{1}\left|10\right\rangle\left\langle
10\right|+V_{2}\left|01\right\rangle\left\langle 01\right|,$ (5)
where $\left|10\right\rangle$ or $\left|01\right\rangle$ represents a particle
in the first or second site respectively, $J$ is the tunneling strength,
$V_{1}$ and $V_{2}$ are the on-site potentials. In this chain system with
fixed parameters $J$, $V_{2}$ and variable $V_{1}$, resonance phenomenon can
be observed [47, 48, 49, 50, 51]: Assuming that a particle is on the first
site at $t=0$, the probability, which is denoted by $P_{2}(V_{1},t)$, of
observing the particle on the second site at any time $t$ reaches maximum when
$V_{1}=V_{2}$. We numerically simulate this phenomenon in a discrete-time
evolution with a large $N_{T}$ in Fig. 2(a4). We exhibit $P_{2}$ at $t=15$
(units of $1/J$, $J=0.1$) with the transparent lines for $V_{2}=0,-\pi/2$. As
expected, $P_{2}$ has one resonance peak at $V_{1}=V_{2}$. We can understand
this phenomenon more visually with the help of double-well system as shown in
Fig. 2(a2). Supposing that a particle is bounded in the left well at the
initial time, and then it will tunnel to the right well with a certain
probability. When the potential energies on both sides are equivalent (i.e.
$V_{1}=V_{2}$), the tunneling probability reaches maximization.
We wonder, when $N_{T}$ is small, whether we can qualitatively observe a
similar resonant effect as the large $N_{T}$ limit. Motivated by this, we
discuss the case of $N_{T}=2$. The parameters are redefined as
$\theta=J\tau,\phi=V_{1}\tau,$ and $\alpha=V_{2}\tau$ for convenience. Our
concern is the probability of finding spin excitation on the $2$nd qubit.
Figure 2(a1) shows two propagation paths of the spin excitation from the $1$st
to the $2$nd qubit. The blue path contributes $-i\sin\theta\cos\theta
e^{i(\alpha-\phi)/2}$ to the amplitude, the purple path contributes
$-i\sin\theta\cos\theta e^{-i(\alpha-\phi)/2}$ to the amplitude, so the final
state of the quantum circuit reads
$U_{1,2}^{XY}\left(\theta\right)U_{2}^{Z}\left(\alpha\right)U_{1}^{Z}\left(\phi\right)U_{1,2}^{XY}(\theta)\left|10\right\rangle=A_{10}\left|10\right\rangle+A_{01}\left|01\right\rangle,$
where
$A_{01}=-i\sin\theta\cos\theta(e^{-i(\alpha-\phi)/2}+e^{i(\alpha-\phi)/2})$.
The probability of spin excitation measured on the second qubit is
$P_{01}(\theta,\phi)=2\sin^{2}\theta\cos^{2}\theta\left(1+\cos(\alpha-\phi)\right).$
(6)
In Fig. 2(a4), we plot $P_{01}$ as function of $\phi$ with $\theta=\pi/2$ and
$\alpha=0,-\pi/2$. The accurate results (solid lines) computed using QuESTlink
coincide with the experimental outcomes (solid lines with point symbols)
computed using the IBM quantum device ”ibmq_rome”. The resonance peak is seen
near $\phi=\alpha$, which is related to the interference term
$2\sin^{2}\theta\cos^{2}\theta\cos(\alpha-\phi)$ in Eq. (6). $P_{2}$ coincides
well with $P_{01}$, both of them have only one peak and the position of the
peak is $\alpha=\phi$ (i.e. $V_{1}=V_{2}$). The above analysis indicates that
only two Trotter steps are required for the circuit to exhibit resonant
tunneling similar to the continuous-time evolution (i.e. $N_{T}$ is enough
large).
### 3.2 Multi-qubit system
In this section we discuss the multi-qubit quantum circuits with the system
sizes $N=3,4,5$ respectively.
We first discuss three-qubit system. The Hamiltonian of the transverse-field
XY chain with $3$ sites reads
$\displaystyle H$ $\displaystyle=$ $\displaystyle
J\sum_{j}^{2}(\sigma_{j}^{x}\sigma_{j+1}^{x}+\sigma_{j}^{y}\sigma_{j+1}^{y})/4$
(7)
$\displaystyle+V_{1}\sigma_{1}^{z}/2+V_{2}\sigma_{2}^{z}/2+V_{1}\sigma_{3}^{z}/2.$
The corresponding time evolution can be approximated,
$\displaystyle U(t)$ $\displaystyle=$ $\displaystyle
e^{-iHt}\approx[U_{1,2}^{XY}(J\tau)U_{2,3}^{XY}(J\tau)U_{1}^{Z}(V_{1}\tau)\times$
(8) $\displaystyle U_{2}^{Z}(V_{2}\tau)U_{3}^{Z}(V_{1}\tau)]^{N_{T}},$
where
$U_{j,j+1}^{XY}(J_{j}\tau)=e^{-i(\sigma_{j}^{x}\sigma_{j+1}^{x}+\sigma_{j}^{y}\sigma_{j+1}^{y})J_{j}\tau}$,
$U_{j}^{Z}\left(V_{j}\tau\right)=e^{-i\sigma_{1}^{z}V_{j}\tau/2}$,
$\tau=t/N_{T}$. The right side of above equation represents the discrete-time
evolution. The quantum circuit implementing the discrete-time evolution is
shown in Fig. 2(b1). We concern the time evolution of single spin excitation.
When the single spin excitation can be treated as a particle, the transverse-
field XY chain is equal to a $3$-site tight-binding chain [see Fig. 2(b3)],
$\displaystyle H$ $\displaystyle=$ $\displaystyle
J\left(\left|100\right\rangle\left\langle
010\right|+\left|010\right\rangle\left\langle
100\right|+\mathrm{h.c.}\right)+V_{1}\left|100\right\rangle\left\langle
100\right|$ (9) $\displaystyle+V_{2}\left|010\right\rangle\left\langle
010\right|+V_{1}\left|001\right\rangle\left\langle 001\right|.$
where $J$ is the coupling strength, the potentials on three sites are
$V_{1},V_{2}$ and $V_{1}$ respectively. In the case that $V_{1}$ is the only
variable parameter, the resonance phenomenon means that the probability,
$P_{3}(t,V_{1})$, of finding particle on the $3$th site reaches maximum at
$V_{1}=V_{2}$. We numerically simulate the discrete-time evolution with a
large $N_{T}$ and show the resonance phenomenon. The initial state is
$\left|100\right\rangle$. In Fig. 2(b4), we plot $P_{3}$ as the function of
$V_{1}$ at $t=22$ (units of $1/J$, $J=0.1$) with the transparent lines.
$P_{3}$ has one resonance peak at $V_{1}=V_{2}$. Similarly, we can consider
the resonance phenomenon with a triple-well system [see Fig. 2(b2)], whose
Hamiltonian can be written as Eq. (9). In the triple-well system, the
probability of the particle tunneling from the left well to the right well
reaches maximization at $V_{1}=V_{2},$ when the resonance occurs.
As for the small $N_{T}$, we find that only two Trotter steps are required for
the $3$-qubit circuit to exhibit resonant tunneling. The parameters of the
$3$-qubit circuit are redefined as
$J\tau=\theta,V_{1}\tau=\phi,V_{2}\tau=\alpha$. As shown the blue, red and
purple dashed lines in Fig. 2(b1), the spin excitation goes through three
paths. The blue and red paths contribute
$\cos\theta(-i\sin\theta)(-i\sin\theta)e^{-i\alpha/2}$ to the amplitude, and
the purple path contributes $-i\sin\theta\cos\theta
e^{i\left(\alpha/2-\phi\right)}\cos\theta(-i\sin\theta)$ to the amplitude. The
amplitude of the final state on the third qubit is
$A_{001}=-\sin^{2}\theta\cos\theta\left(2e^{-i\alpha/2}+\cos\theta
e^{i\left(\alpha/2-\phi\right)}\right).$ (10)
Accordingly, the probability of finding the spin excitation on the $3$rd qubit
is
$P_{001}(\theta,\phi)=\sin^{4}\theta\cos^{2}\theta(4+\cos^{2}\theta+4\cos\theta\cos\left(\alpha-\phi\right)).$
(11)
The interference term
$4\sin^{4}\theta\cos^{3}\theta\cos\left(\alpha-\phi\right)$ dominates the
resonant tunneling effect. In Fig. 2(b4), we plot $P_{001}$ as function of
$\phi$ with $\theta=\pi/2$ and $\alpha=0,-\pi/2$. The accurate results (solid
lines) computed using QuESTlink coincide with the experimental outcomes (solid
lines with point symbols) computed using the IBM quantum device ”ibmq_rome”.
As we can see, $P_{3}$ coincides well with $P_{001}$, both of them have only
one peak and the position of the peak is $\alpha=\phi$ (i.e. $V_{1}=V_{2}$).
The above analysis indicates that only two Trotter steps are required for the
$3$-qubit circuit to exhibit similar resonant tunneling in the continuous-time
limit.
We also study the four-qubit system. The Hamiltonian of the $4$-site
transverse-field XY chain we studied is
$\displaystyle H$ $\displaystyle=$ $\displaystyle
J_{1}\sum_{j=1,3}(\sigma_{j}^{x}\sigma_{j+1}^{x}+\sigma_{j}^{y}\sigma_{j+1}^{y})/4+J_{2}(\sigma_{2}^{x}\sigma_{3}^{x}+\sigma_{2}^{y}\sigma_{3}^{y})/4$
(12)
$\displaystyle+V_{1}\sigma_{1}^{z}/2+V_{2}\sigma_{2}^{z}/2-V_{2}\sigma_{3}^{z}/2+V_{1}\sigma_{4}^{z}/2.$
The corresponding discrete-time evolution is in the form
$\displaystyle U(t)$ $\displaystyle=$ $\displaystyle
e^{-iHt}\approx[U_{1,2}^{XY}(J_{1}\tau)U_{2,3}^{XY}(J_{2}\tau)U_{3,4}^{XY}(J_{1}\tau)\times$
(13) $\displaystyle
U_{1}^{Z}(V_{1}\tau)U_{2}^{Z}(V_{2}\tau)U_{3}^{Z}(-V_{2}\tau)U_{3}^{Z}(V_{1}\tau)]^{N_{T}},$
where
$U_{j,j+1}^{XY}(J_{j}\tau)=e^{-i(\sigma_{j}^{x}\sigma_{j+1}^{x}+\sigma_{j}^{y}\sigma_{j+1}^{y})J_{j}\tau}$,
$U_{j}^{Z}\left(V_{j}\tau\right)=e^{-i\sigma_{1}^{z}V_{j}\tau/2}$,
$\tau=t/N_{T}$. We plot the quantum circuit implementing the discrete-time
evolution in Fig. 2(c1). In the single-particle subspace, the equivalent
$4$-site tight-binding chain is
$\displaystyle H$ $\displaystyle=$ $\displaystyle
J_{1}\left(\left|1000\right\rangle\left\langle
0100\right|+\left|0010\right\rangle\left\langle
0001\right|\right)+J_{2}(\left|0100\right\rangle\left\langle 0010\right|$ (14)
$\displaystyle+\left|0010\right\rangle\left\langle
0100\right|)+V_{1}\left|1000\right\rangle\left\langle
1000\right|+V_{2}\left|0100\right\rangle\left\langle 0100\right|$
$\displaystyle-V_{2}\left|0010\right\rangle\left\langle
0010\right|+V_{1}\left|0001\right\rangle\left\langle 0001\right|.$
As we marked in Fig. 2(c4), the coupling strengths between neighboring sites
are $J_{1},J_{2}$, and $J_{1}$ respectively, and the on-site potentials on the
four sites are $V_{1},V_{2},-V_{2},V_{1}$ respectively. In the condition of
$J_{1}\ll J_{2},V_{2}$, we numerically simulate the discrete-time evolution of
one particle with a large $N_{T}$. In Fig. 2(c4), $P_{4}(V_{1},t)$, which is
the probability of finding the particle on the $4$th site, is plotted as
function of $V_{1}$ with the transparent lines. The cases of $V_{2}=10$ and
$V_{2}=20$ are studied when $J_{1}=1$, $t=3$ (units of $1/J_{1}$), $J_{2}=20$.
As we can see, the resonant peaks can be observed near
$\sqrt{J_{2}^{2}+V_{2}^{2}}$ and $-\sqrt{J_{2}^{2}+V_{2}^{2}}$, and the
distance between the resonance peaks varies when $V_{2},-V_{2}$ change. One
can observe the same resonance phenomenon in a triple-well system [see Fig.
2(c2)]. The energy levels of the left and right wells are $V_{1}$, the middle
well has two energy levels: $\sqrt{J_{2}^{2}+V_{2}^{2}}$ and
$-\sqrt{J_{2}^{2}+V_{2}^{2}}$. If there is a particle in the left well at the
initial moment, then we can detect this particle in the right well with a
certain probability. When $V_{1}$ is close to $\sqrt{J_{2}^{2}+V_{2}^{2}}$ or
$-\sqrt{J_{2}^{2}+V_{2}^{2}}$, the probability of finding particle in right
well reaches the maximum. The distance of the two peaks varies with $V_{2}$.
When $N_{T}$ is small, we find that only three Trotter steps are required for
the $4$-qubit circuit to qualitatively exhibit resonant tunneling. We
investigate the propagation of the spin excitation in the circuit. $P_{0001}$,
which is the probability of finding spin excitation on the $4$th qubit, is
shown (the solid lines) in Fig. 2(c1). The parameters are redefined as
$J_{1}\tau=\theta_{1},J_{2}\tau=\theta_{2},V_{1}\tau=\phi,V_{2}\tau=\alpha$.
With $\theta_{1}=\theta_{2}=\pi/1.5$, we plot two cases of $\alpha=\pi/4$ and
$\alpha=-\pi/1.5$. Compare Fig. 2(c4) with (a4) or (b4), we find that the
peaks are not at $\alpha$ and $-\alpha$. However, the distance between every
two peaks is changed when $\alpha$ is adjusted, which is the major
characteristic of the resonant tunneling effect.
Finally, we discuss the five-qubit system. The Hamiltonian of the $5$-site
transverse field XY chain we studied is
$\displaystyle H$ $\displaystyle=$ $\displaystyle
J_{1}\sum_{j=1,4}(\sigma_{j}^{x}\sigma_{j+1}^{x}+\sigma_{j}^{y}\sigma_{j+1}^{y})/4$
(15)
$\displaystyle+J_{2}\sum_{j=2,3}(\sigma_{j}^{x}\sigma_{j+1}^{x}+\sigma_{j}^{y}\sigma_{j+1}^{y})/4$
$\displaystyle+V_{1}\sigma_{1}^{z}/2+V_{2}\sigma_{2}^{z}/2-V_{2}\sigma_{4}^{z}/2+V_{1}\sigma_{5}^{z}/2.$
The quantum circuit implementing the discrete-time evolution is shown in Fig.
2(d1). In the single-particle subspace, the equivalent $5$-site tight-binding
chain is
$\displaystyle H$ $\displaystyle=$
$\displaystyle[J_{1}\left(\left|10000\right\rangle\left\langle
01000\right|+\left|00010\right\rangle\left\langle 00001\right|\right)$
$\displaystyle+J_{2}(\left|01000\right\rangle\left\langle
00100\right|+\left|00100\right\rangle\left\langle
00010\right|)+\mathrm{h.c.}]$
$\displaystyle+V_{1}\left|10000\right\rangle\left\langle
10000\right|+V_{2}\left|01000\right\rangle\left\langle 01000\right|$
$\displaystyle-V_{2}\left|00010\right\rangle\left\langle
00010\right|+V_{1}\left|00001\right\rangle\left\langle 0001\right|.$
The schematic diagram of Hamiltonian Eq. (3.2) is shown in Fig. 2(c4). We
numerically simulate the continuous-time evolution (implemented with the large
$N_{T}$). The system parameters are $J_{1}=0.1$, $t=40$ (units of $1/J_{1}$),
$J_{2}=20$, $V_{2}=10,20$. We focus on the probability $P_{5}(V_{1},t)$ (see
the transparent lines in Fig. 2(d4)) of finding particle on the $5$th site. As
we can see, $P_{5}(V_{1},t)$ will reach the maximum when $V_{1}\approx
0,\pm\sqrt{2J_{2}^{2}+V_{2}^{2}}$, and the distance of the peaks is affected
by $V_{2}$. The same resonance phenomenon can be observed in a triple-well
system (see Fig. 2(d2)). The energy levels on the left and right wells are
$V_{1}$, the middle well has three energy levels:
$\sqrt{2J_{2}^{2}+V_{2}^{2}}$, $0$, and $-\sqrt{2J_{2}^{2}+V_{2}^{2}}$. When
$V_{1}=\sqrt{J_{2}^{2}+V_{2}^{2}}$, $0$, or $-\sqrt{J_{2}^{2}+V_{2}^{2}}$, the
probability of finding particles in right well reaches the maximum. The
distances between three peaks vary with $V_{2}$.
When $N_{T}$ is small, we find that only four Trotter steps are required for
the circuit to qualitatively exhibit resonant tunneling. We investigate the
propagation of the spin excitation in the circuit shown in Fig. 2(d1). We
redefine parameters as
$J_{1}\tau=\theta_{1},J_{2}\tau=\theta_{2},V_{1}\tau=\phi,V_{2}\tau=\alpha,$
and denote the probability of finding spin excitation on the $5$th qubit as
$P_{00001}$. Figure 2(d4) exhibits $P_{00001}$ (the solid lines) as function
of $\phi$ by fixing $\alpha,\theta_{1}=\pi/3$ and $\theta_{2}=\pi/1.2$. We
compare two cases that $\alpha$ is $\pi/1.5,\pi/5$ respectively. The distance
between the peaks is changed when $\alpha$ is adjusted, which is the major
characteristic of the resonant tunneling effect.
### 3.3 CR model
Figure 3: (Color online) (a) 4-qubit circuit. The two-qubit XY gates are
replaced by controlled-$R_{x}$ gates. (b) The probability of finding the spin
excitation on the last qubit as function of $\phi$. (c) The probability of
observing the spin excitation on the last third of the qubits varies with
Trotter step $\eta$. (d) The probability distribution of spin excitation at
$\eta=10$. (c) and (d) use the same legend and drawing parameter
$\theta=\phi=\pi/2$. Figure 4: (Color online) (a) The IPRη varies with Trotter
number $\eta$, the drawing parameters are $N=15,N_{T}=80,\theta=\phi=\pi/2$.
The red and blue lines represent the ordered and disordered case respectively.
The horizontal lines are the average value of IPRη, i.e. IPRave. (b) IPRave as
function of the degree of randomness. For a fixed $R$, we have $20$ data. The
blue line is the average value of the $20$ data, and error bar is the
variance. (c) The physical quantity $P_{t}$ is plotted in disordered
configuration. (d) The probability distribution of spin excitation when
$\eta=10$. (c) and (d) share the same legend.
In this section, the propagation of the spin excitation through
controlled-$R_{x}$ gates is investigated. In the previous section, the spin
excitation is created by a NOT gate and propagated through the XY gate. Here,
we study the behavior of the spin excitation propagating through the
controlled-$R_{x}$ gates. The controlled-$R_{x}$ gate is expressed as
$U^{CR_{x}}=\left|0\right\rangle\left\langle 0\right|\otimes
I+\left|1\right\rangle\left\langle 1\right|\otimes e^{-i\theta\sigma_{x}/2},$
(17)
the matrix representation of $U^{CR_{x}}$ is shown in Fig. 1(c). Consider the
situation that a spin excitation on the first qubit passes through a
controlled-$R_{x}$ gate, we get the following equation
$U^{CR_{x}}\left|10\right\rangle=\cos\frac{\theta}{2}\left|10\right\rangle-i\sin\frac{\theta}{2}\left|11\right\rangle.$
(18)
The above equation indicates that after passing through a controlled-$R_{x}$
gate, the spin excitation becomes a two-qubit entangled state. From the view
point of propagation of the bit-flip error, this indicates that the
controlled-$R_{x}$ gate can transform a single-qubit error to a multi-qubit
error. We take the four-qubit circuit in Fig. 3(a) as a example to study the
behavior of spin excitation propagating through controlled-$R_{x}$ gates under
the discrete-time evolution with a large step size. We observed the
probability of finding the spin excitation on the $4$th qubit and denote the
probability as $P_{0001}$. In Fig. 3(b), $P_{0001}$ shows two resonant peaks,
the distance between the two resonant peaks is changed as the $\phi$ varies.
This indicates that even if the spin excitations are propagated by the
controlled-$R_{x}$ gate, when the step size is large we can qualitatively
observe the resonance phenomenon that occurs in the continuous-time limit.
## 4 Localization for large Trotter step size
In this section, we study whether the localization can be observed in the
discrete-time evolution when the Trotter step size is large. We still study
the transverse field XY model, but the parameters of a layer of $U^{Z}$ (i.e.
a layer of single-qubit $R_{z}$ gates) are random. We compare the probability
distribution of the spin excitation in different configurations of the
parameters of a layer of $R_{z}$ gates. The stronger the randomness of the
parameters, the higher the localization of the distribution of the spin
excitation, which means higher the probability of observing the spin
excitation near a specific qubit. In this study, the propagation of the bit-
flip error (i.e. an unwanted NOT gate) is similar to the transport of the spin
excitation. Therefore, the localization indicates that the bit-flip error may
be localized near a specific qubit and may not affect the measurement on
distant qubits.
The localization phenomenon is studied[52, 53, 54, 55, 56] during the
discrete-time evolution with a large step size. We begin with the Hamiltonian
in Eq. (1) with $J_{j}=J$. For convenience the parameters are defined as
$J\tau=\theta,V_{j}\tau=\phi_{j}.$ For the purpose of investigating
localization, one layer of parameters for $U^{Z}$ in Eq. (2) is
{$\phi_{j}$}$\equiv${$\phi_{1},-\phi_{2},\phi_{3},-\phi_{4},\cdots$}, where
$\phi_{j}=\phi+r_{j}$, $r_{j}\in[-R$, $R]$ is a random number. {$\phi_{j}$} is
the same for each layer of $U^{Z}$. {$\phi_{j}$} is ordered (periodic)
configuration when $R=0$ and disordered configuration when $R>0$. The inverse
participation ratio [57] (IPR) is a measure of localization and defined as
$IPR_{\eta}=\sum_{i=1}^{N}|P_{i}(\eta)|^{4},$ (19)
where
$\displaystyle P_{i}(\eta)$ $\displaystyle=$ $\displaystyle\left\langle
i\right.\left|\psi(\eta)\right\rangle,$
$\displaystyle\left|\psi(\eta)\right\rangle$ $\displaystyle=$
$\displaystyle(\prod_{j=1}^{N-1}U_{j,j+1}^{XY}(\theta_{j})\prod_{j=1}^{N}U_{j}^{Z}(\phi_{j}))^{\eta}\left|\psi(0)\right\rangle.$
(20)
$\left|\psi(\eta)\right\rangle$ represents the quantum state at the $\eta$th
Trotter step, $P_{i}(\eta)$ represents the corresponding amplitude at the
$i$th qubit. In general, IPRη varies from $1/N$ (system size) to $1$ and a
large value of IPRη means a stronger localization effect. The localization of
$\left|\psi(\eta)\right\rangle$ changes with $\eta$, thus the average IPR is
introduced to character the average level of localization during the whole
discrete-time evolution[42],
$IPR_{ave}=\frac{1}{N_{T}}\sum_{\eta=1}^{N_{T}}IPR_{\eta}.$ (21)
In Fig. 4(a), we plot $IPR_{\eta}$ (the solid lines) as function of $\eta$ for
the ordered ($R=0$) and disordered ($R=\pi/2$) configuration respectively. The
drawing parameters are $N=N_{T}=80,\theta=\phi=\pi/2$. $IPR_{\eta}$ for both
the ordered (the red lines) and disordered (the red lines) configuration show
a periodic-like behavior and is larger than $1/N$, which means
$\left|\psi(\eta)\right\rangle$ exhibits localization effect in both cases.
However, the average value (the horizontal line), i.e. $IPR_{ave}$, of the
blue line is larger than the red line, and the peaks of the blue line are
closer to $1$. This indicates stronger localization in the disordered case.
Furthermore, in Fig. 4(b), we show the $IPR_{ave}$ varying with the degree $R$
of the randomness. With the increase of disorder, the localization becomes
stronger.
In this study, the propagation of the bit-flip error (i.e. an unwanted NOT
gate) is similar to the transport of the spin excitation, thus the
localization implies that single bit-flip error propagated by disordered
{$U_{j}^{Z}(\phi_{j})$} does not affect the measurement on a distant qubit. To
illustrate this point, we propose a physical quantity
$P_{\mathrm{t}}\equiv\sum_{q_{i}=2N/3}^{N}p_{q_{i}}$ which is the average
probability of finding the spin excitation on the last third of the qubits,
where $p_{q_{i}}$ denotes the probability of finding the error on the $i$th
qubit. The smaller the $P_{\mathrm{t}}$, the shorter the distance the spin
excitation travels. As shown in Fig. 4(c), $P_{\mathrm{t}}$ is lower when
$r_{i}\neq 0$, which demonstrates that only a little probability is propagated
to the last several qubits. $P_{\mathrm{t}}$ is almost vanishing as the degree
of randomness keeps increasing. In Fig 4(d), we plot $p_{q_{i}}$ at $\eta=10$.
As we can see, more probabilities are propagated to the last few qubits for
$r_{i}=0$ and are localized at the first few qubits for $r_{i}\neq 0$. Figure
4(d) also show that the greater the degree of randomness, the stronger the
localization phenomenon. Above results indicates that the measurement is
almost unaffected by the bit-flip error on the first qubit for $r_{i}\neq 0$.
The above conclusion still holds if XY gates in the circuit are replaced by
controlled-$R_{x}$. As shown in Fig. 3(c), $P_{t}$ grows with the increasing
Trotter steps. However, $P_{\mathrm{t}}$ becomes lower when the random
perturbation is applied to a layer of $R_{z}$ gates, which means less
probabilities are propagated to the end of the circuit. With the increasing
degree of disorder, the inhibiting effect is more significant. In Fig. 3(d),
we plot the probability distribution of spin excitation when
$\eta=10,\theta=\phi=\pi/2$. These results show that, with a strong randomness
of {$\phi_{j}$}, the spin excitation will not be propagated to the farther
qubits. At the same time, this also shows that the randomness of the
parameters will inhibit the propagation of the bit-flip error and protect the
measurement on the distant qubits is not affected.
## 5 Conclusions
We study the transport of the spin excitation in the discrete-time evolution
using Trotter circuits with a large step size, and qualitatively observe the
quantum phenomena in the continuous-time limit, i.e. the resonant tunneling
and localization. We observe the resonance phenomenon during the transport of
the spin excitation in systems of sizes $N=2,3,4,5$ respectively. The
probability distribution of spin excitations propagating through several
Trotter steps agree qualitatively with that in the continuous-time limit. The
corresponding minimum number of Trotter steps is given for each system size.
In a Trotter circuit with random parameters of $R_{z}$ gates, we can observe
the localization phenomenon of the spin excitation distribution even with a
large step size. We study the spin excitation propagating through the XY gates
and also through the controlled-$R_{x}$ gates. Our research indicates that a
discrete-time quantum simulator with a large step size can qualitatively
demonstrate some physical phenomena in the continuous-time limit. Qualitative
observations require fewer quantum gates than quantitative calculations and
therefore are a promising application on near-term quantum computers. In
quantum computing, some errors, such as bit-flip errors, behave like spin
excitations, thus, our finding can be used to understand the propagating of
these errors in the quantum circuits.
I am grateful to Ying Li for the discussions and help in preparing the
manuscript. This work is supported by the National Natural Science Foundation
of China (Grants No. 11875050 and No. 12088101), NSAF (Grant No. U1930403),
and the Special Fund for Theoretical Physics of the National Natural Science
Foundation of China (Grant No. 12047547).
## References
## References
* [1] R. P. Feynman, Simulating physics with computers, Int J Theor Phys 21, 467–488 (1982)
* [2] M. A. Nielsen and I. L. Chuang, Quantum Computation and Quantum Information, Cambridge University Press, Cambridge, (2010)
* [3] F. Arute et al., Quantum supremacy using a programmable superconducting processor, Nature 574, 505 (2019)
* [4] S. Lloyd, Universal quantum simulators, Science 273, 1073 (1996)
* [5] A. M. Childs, D. Maslov, Y. Nam, N. J. Ross, and Y. Su, Toward the first quantum simulation with quantum speedup, Proceedings of the National Academy of Sciences 115, 9456 (2018)
* [6] A. M. Childs, Y. Su, M. C. Tran, N. Wiebe, and S. Zhu, Theory of trotter error with commutator scaling, Phys. Rev. X 11, 011020 (2021)
* [7] C. Yi, Robustness of discretization in digital adiabatic simulation, arXiv:2107.06404 (2021)
* [8] J. Preskill, Quantum computing in the NISQ era and beyond, Quantum 2, 79 (2018)
* [9] R. Landauer, Is quantum mechanics useful? Phil. Tran. R. Soc. Lond. 353, 367 (1995); R. Landauer, The physical nature of information, Phys. Lett. A 217, 188 (1996)
* [10] W. G. Unruh, Maintaining coherence in quantum computers, Phys. Rev. A 51, 992 (1995)
* [11] S. Haroche and J. M. Raimond, Quantum computing: dream or nightmare? Phys. Today 49, 51 (1996)
* [12] P. W. Shor, Scheme for reducing decoherence in quantum computer memory. Phys. Rev. A 52, R2493 (1995)
* [13] A. M. Steane, Error Correcting Codes in Quantum Theory. Phys. Rev. Lett. 77, 793 (1996)
* [14] A. R. Calderbank and P. W. Shor, Good Quantum Error Correcting codes exist, Phys. Rev. A 54, 1098 (1996)
* [15] R. Laamme, C. Miquel, J. P. Paz, and W. H. Zurek, Perfect Quantum Error Correcting Code. Phys. Rev. Lett. 77, 198 (1996)
* [16] C. H. Bennett, D. P. DiVincenzo J. A. Smolin, and W. K. Wooters, Mixed State Entanglement and Quantum Error Correction, Phys. Rev. A 54, 3824, (1996)
* [17] A. M. Steane, Multiple particle interference and quantum error correction, Proc. Royal Society of London A 452, 2551 (1996)
* [18] A. M. Steane, Simple quantum error correcting codes, Phys. Rev. A 54, 4741 (1996)
* [19] E. Knill and R. Laamme, A Theory of Quantum Error Correcting Codes. Phys. Rev. Lett. 84, 2525 (2000)
* [20] P. W. Shor, Fault-Tolerant quantum computation, Proc. 37th IEEE Symp. on Foundations of Computer Science, 56-65 (1996)
* [21] D. Gottesman, Fault-Tolerant Quantum Computation with Local Gates, J. Mod. Opt 47, 333 (2000)
* [22] E. Knill, R. Laamme, and W. H. Zurek, Accuracy threshold for Quantum Computation (1996)
* [23] A. Holmes, M. R. Jokar, G. Pasandi, Y. Ding, M. Pedram, and F. T. Chong, Nisq+: Boosting quantum computing power by approximating quantum error correction, arXiv:2004.04794 (2020)
* [24] Y. Ueno, M. Kondo, M. Tanaka, Y. Suzuki, and Y. Tabuchi, Qecool: On-line quantum error correction with a superconducting decoder for surface code, arXiv:2103.07526 (2021)
* [25] P. Das, A. Locharla, and C. Jones, Lilliput: A lightweight low-latency lookup-table based decoder for near-term quantum error correction, arXiv:2108.06569 (2021)
* [26] Y. Li and S. Benjamin, Practical Quantum Error Mitigation for Near-Future Applications, Phys. Rev. X 8, 031027 (2018)
* [27] S. Endo, Z. Cai, S. C. Benjamin, and X. Yuan, Hybrid quantum-classical algorithms and quantum error mitigation, J. Phys. Soc. Jpn 90. 032001 (2021)
* [28] A. Kandala, K. Temme, A. D. Córcoles, A. Mezzacapo, J. M. Chow, and J. M. Gambetta, Error mitigation extends the computational reach of a noisy quantum processor, Nature 567, 209–212 (2019)
* [29] J. J. Wallman and J. Emerson, Randomized compiling with twirling gates, Phys. Rev. A 94, 052325 (2016)
* [30] A. Strikis, D. Y. Qin, Y. Z. Chen, S. C. Benjamin, and Y. Li, Learning-based quantum error mitigation, PRX Quantum 2, 040330 (2021)
* [31] Z. Wang , Y. Z. Chen , Z. X. Song, D. Qin, H. K. Li, Q. J. Guo, H. Wang, C. Song, and Ying Li, Scalable evaluation of quantum-circuit error loss using Clifford sampling, Phys. Rev. Lett. 126, 080501 (2021)
* [32] V. N. Premakumar and R. Joynt, Error mitigation in quantum computers subject to spatially correlated noise, arXiv:1812.07076 (2018)
* [33] K. Temme, S. Bravyi, and J. M. Gambetta, Error mitigation for short-depth quantum circuits, Phys. Rev. Lett. 119, 180509 (2017)
* [34] C. Song, J. Cui, H. Wang, J. Hao, H. Feng, and Ying Li, Quantum computation with universal error mitigation on a superconducting quantum processor, Sci. Adv 5, eaaw5686 (2019)
* [35] A. Kandala, K. Temme, A. D. C orcoles, A. Mezzacapo, J. M. Chow, and J. M. Gambetta, Error mitigation extends the computational reach of a noisy quantum processor, Nature 567, 491-495 (2019)
* [36] M. Otten and S. K. Gray, Recovering noise-free quantum observables, Phys. Rev. A 99, 012338 (2019)
* [37] A. Tranter, P. J. Love, F. Mintert, N. Wiebe, and P. V. Coveney, Ordering of Trotterization: Impact on Errors in Quantum Simulation of Electronic Structure, Entropy 21, 1218 (2019)
* [38] C. Yi, Success of digital adiabatic simulation with large Trotter step, Phys. Rev. A 104, 052603 (2021)
* [39] P. J. J. O’Malley et al., Scalable Quantum Simulation of Molecular Energies, Phys. Rev. X 6, 031007 (2016)
* [40] A. M. Childs, Y. Su, M. C. Tran, N. Wiebe, and S. Zhu, Theory of Trotter Error with Commutator Scaling, Phys. Rev. X 11, 011020 (2021)
* [41] D. Poulin, M. B. Hastings, D. Wecker, N. Wiebe, A. C. Doherty, and M. Troyer, The trotter step size required for accurate quantum simulation of quantum Chemistry, Quantum Information & Computation 15, 361 (2015)
* [42] M. Heyl, P. Hauke, P. Zoller, Quantum localization bounds Trotter errors in digital quantum simulation, Sci Adv 5, eaau8342 (2019)
* [43] L. M. Sieberer, T. Olsacher, A. Elben, M. Hey, P. Hauke, F. Haake and P. Zoller, Digital quantum simulation, Trotter errors, and quantum chaos of the kicked top, NPJ quantum iniformation 5, 1-11 (2019)
* [44] J. T. Edwards and D. J. Thouless, Numerical studies of localization in disordered systems, J. Phys. C: Solid State Phys. 5 807 (1972)
* [45] M. Segev, Y. Silberberg, and D. N. Christodoulides, Anderson localization of light, Nature Photon 7, 197–204 (2013)
* [46] P. Jordan and E. Wigner, Über das Paulische Äquivalenzverbot, Z. Phys. 47, 631 (1928)
* [47] J. G. Wang, S. J. Yang, Ultracold bosons in a one-dimensional optical lattice chain: Newton’s cradle and Bose enhancement effect, Phys. Lett. A 381, 1665 (2017)
* [48] A. S. Buyskikh , L. Tagliacozzo, D. Schuricht, C. A. Hooley, D. Pekker, and A. J. Daley, Resonant two-site tunneling dynamics of bosons in a tilted optical superlattice, Phys. Rev. A 100, 023627 (2019)
* [49] S. Scherg, T. Kohlert, P. Sala, F. Pollmann, B. H. Madhusudhana, I. Bloch, and M. Aidelsburger, Observing non-ergodicity due to kinetic constraints in tilted Fermi-Hubbard chains, Nat. Commun 12, 4490 (2021)
* [50] R. Tsu and L. Esaki, Tunneling in a finite superlattice, Appl. Phys. Lett. 22, 562 (1973)
* [51] T. C. L. G. Sollner, W. D. Goodhue, P. E. Tannenwald, C. D. Parker, and D. D. Peck, Resonant tunneling through quantum wells at frequencies up to 2.5 THz, Appl. Phys. Lett. 43, 588 (1983)
* [52] P. W. Anderson, Absence of Diffusion in Certain Random Lattices, Phys. Rev. 109, 1492 (1957)
* [53] D. S. Wiersma, P. Bartolini, A. Lagendijk, and R. Righini, Localization of light in a disordered medium, Nature 390, 671–673 (1997)
* [54] B. L. Altshuler, D. Khmel’nitzkii, A. I. Larkin, and P. A. Lee, Magnetoresistance and Hall effect in a disordered two-dimensional electron gas, Phys. Rev. B 22, 5142 (1980)
* [55] D. Vollhardt, P. Wölfle, Scaling Equations from a Self-Consistent Theory of Anderson Localization, Phys. Rev. Lett. 48, 699 (1982)
* [56] A. N. Poddubny, M. V. Rybin, M. F. Limonov, and Y. S. Kivshar, Fano interference governs wave transport in disordered systems, Nat. Commun 3, 914 (2012)
* [57] F. Haake, Quantum Signatures of Chaos (Springer, 2010)
|
# T-Zamfirescu and T-weak contraction mappings on cone metric spaces
José R. Morales and Edixon Rojas Department of Mathematics, Faculty of
Science, University of Los Andes, Mérida-5101, Venezuela<EMAIL_ADDRESS><EMAIL_ADDRESS>
###### Abstract.
The purpose of this paper is to obtain sufficient conditions for the existence
of a unique fixed point of T-Zamfirescu and T-weak contraction mappings in the
framework of complete cone metric spaces.
###### Key words and phrases:
Fixed point, complete cone metric space, $T-$zamficescu mapping, $T-$weak
contraction, subsequentially convergent.
###### 1991 Mathematics Subject Classification:
47H10, 46J10.
## 1\. Introduction
In 2007, Guang and Xiang [11] generalized the concept of metric space,
replacing the set of real numbers by an ordered Banach space and defined a
cone metric space. The authors there described the convergence of sequences in
cone metric spaces and introduced the completeness. Also, they proved some
fixed point theorems of contractive mappings on complete cone metric spaces.
Since then, fixed point theorems for different (classic) classes of mappings
on these spaces have been appeared, see for instance [1], [7], [8], [10],
[15], [16] and [17].
On the other hand, recently A. Beiranvand S. Moradi, M. Omid and H. Pazandeh
[5] introduced the $T-$contraction and $T-$contractive mappings and then they
extended the Banach contraction principle and the Edelstein’s fixed point
Theorem. S. Moradi [12] introduced the $T-$Kannan contractive mappings,
extending in this way the Kannan’s fixed point theorem [9]. The corresponding
version of $T$-contractive, $T$-Kannan mappings and $T-$Chalterjea
contractions on cone metric spaces was studied in [13] and [14] respectively.
In view of these facts, thereby the purpose of this paper is to study the
existence of fixed points of $T-$Zamficescu and $T-$weak contraction mappings
defined on a complete cone metric space $(M,d)$, generalizing consequently the
results given in [11] and [18].
## 2\. General framework
In this section we recall the definition of cone metric space and some of
their properties (see, [11]). The following notions will be useful for us in
order to prove the main results.
###### Definition 2.1.
Let $E$ be a real Banach space. A subset $P$ of $E$ is called a cone if and
only if:
(P1):
$P$ is closed, nonempty and $P\neq\\{0\\}$;
(P2):
$a,b\in\mathbb{R},\,\,a,b\geq 0,\,\,x,y\in P$ imply $ax+by\in P$;
(P3):
$x\in P$ and $-x\in P\Rightarrow x=0$. I.e., $P\cap(-P)=\\{0\\}$.
Given a cone $P\subset E,$ we define a partial ordering $\leq$ with respect to
$P$ by $x\leq y$ if and only if $y-x\in P$. We write $x<y$ to indicate that
$x\leq y$ but $x\neq y$, while $x\ll y$ will stand for
$y-x\in\operatorname{Int}P$. (interior of $P$.)
###### Definition 2.2.
Let $E$ be a Banach space and $P\subset E$ a cone. The cone $P$ is called
normal if there is a number $K>0$ such that for all $x,y\in E,\,\,0\leq x\leq
y$ implies $\|x\|\leq K\|y\|.$ The least positive number satisfying the above
is called the normal constant of $P.$
In the following, we always suppose that $E$ is a Banach space, $P$ is a cone
in $E$ with $\operatorname{Int}P\neq\emptyset$ and $\leq$ is partial ordering
with respect to $P$.
###### Definition 2.3 ([11]).
Let $M$ be a nonempty set. Suppose that the mapping $d:M\times
M\longrightarrow E$ satisfies:
(d1):
$0<d(x,y)$ for all $x,y\in M$, and $d(x,y)=0$ if and only if $x=y$;
(d2):
$d(x,y)=d(y,x)$ for all $x,y\in M$;
(d3):
$d(x,y)\leq d(x,z)+d(z,y)$ for all $x,y,z\in M$.
Then, $d$ is called a cone metric on $M$ and $(M,d)$ is called a cone metric
space.
Note that the notion of cone metric space is more general that the concept of
metric space.
###### Definition 2.4.
Let $(M,d)$ be a cone metric space. Let $(x_{n})$ be a sequence in $M$ and
$x\in M$.
* (i)
$(x_{n})$ converges to $x$ if for every $c\in E$ with $0\ll c$ there is an
$n_{0}$ such that for all $n>n_{0},\,\,d(x_{n},x)\ll c.$ We denote this by
$\displaystyle\lim_{n\rightarrow\infty}x_{n}=x$ or $x_{n}\rightarrow
x,\,\,(n\rightarrow\infty)$.
* (ii)
If for any $c\in E$ with $0\ll c$ there is an $n_{0}$ such that for all
$n,m\geq n_{0}$, $\;d(x_{n},x_{m})\ll c$, then $(x_{n})$ is called a Cauchy
sequence in $M$.
Let $(M,d)$ be a cone metric space. If every Cauchy sequence is convergent in
$M,$ then $M$ is called a complete cone metric space.
###### Lemma 2.1 ([11]).
Let $(M,d)$ be a cone metric space, $P\subset E$ a normal cone with normal
constant $K.$ Let $(x_{n}),\,\,(y_{n})$ be sequences in $M$ and $x,y\in M$.
* (i)
$(x_{n})$ converges to $x$ if and only if
$\displaystyle\lim_{n\rightarrow\infty}d(x_{n},x)=0$.
* (ii)
If $(x_{n})$ converges to $x$ and $(x_{n})$ converges to $y$, then $x=y$.
* (iii)
If $(x_{n})$ converges to $x$, then $(x_{n})$ is a Cauchy sequence.
* (iv)
$(x_{n})$ is a Cauchy sequence if and only if
$\displaystyle\lim_{n,m\rightarrow\infty}d(x_{n},x_{m})=0$.
* (v)
If $x_{n}\longrightarrow x$ and $y_{n}\longrightarrow
y,\,\,(n\rightarrow\infty)$, then $d(x_{n},y_{n})\longrightarrow d(x,y)$.
###### Definition 2.5.
Let $(M,d)$ be a cone metric space, $P$ a normal cone with normal constant $K$
and $T:M\longrightarrow M$. Then
* (i)
$T$ is said to be continuous, if
$\displaystyle\lim_{n\rightarrow\infty}x_{n}=x$ implies that
$\displaystyle\lim_{n\rightarrow\infty}T(x_{n})=T(x)$ for all $(x_{n})$ and
$x$ in $M$.
* (ii)
$T$ is said to be subsequentially convergent if we have, for every sequence
$(y_{n}),$ if $T(y_{n})$ is convergent, then $(y_{n})$ has a convergent
subsequence.
* (iii)
$T$ is said to be sequentially convergent if we have, for every sequence
$(y_{n}),$ if $T(y_{n})$ is convergent then $(y_{n})$ also is convergent.
Examples of cone metric spaces can be found for instance in [11], [17] and
references therein.
## 3\. Main Results
This section is devoted to give fixed point results for $T$-Zamfirescu and
$T$-weak contraction mappings on complete (normal) cone metric spaces, as well
as, their asymptotic behavior. First, we recall the following classes of
contraction type mappings:
###### Definition 3.1.
Let $(M,d)$ be a cone metric space and $T,S:M\longrightarrow M$ two mappings
* (i)
The mapping $S$ is called a $T-$Banach contraction, (TB - Contraction) if
there is $a\in[0,1)$ such that
$d(TSx,TSy)\leq ad(Tx,Ty)$
for all $x,y\in M$.
* (ii)
The mapping $S$ is called a $T-$Kannan contraction, (TK - Contraction) if
there is $b\in[0,1/2)$ such that
$d(TSx,TSy)\leq b[d(Tx,TSx)+d(y,TSy)]$
for all $x,y\in M$.
* (iii)
A mapping $S$ is said to be a Chatterjea contraction, (TC - Contraction) if
there is $c\in[0,1/2)$ such that
$d(TSx,TSy)\leq c[d(Tx,TSy)+d(Ty,TSx)]$
for all $x,y\in M.$
It is clear that if we take $T=I_{d}$ (the identity map) in the Definition 3.1
we obtain the definitions of Banach contraction, Kannan mapping ([9]) and
Chatterjea mapping ([6]).
Now, following the ideas of T. Zamfirescu [18] we introduce the notion of
$T-$Zamfirescu mappings.
###### Definition 3.2.
Let $(M,d)$ be a cone metric space and $T,S:M\longrightarrow M$ two mappings.
$S$ is called a $T-$Zamfirescu mapping, (TZ -mapping), if and only if, there
are real numbers, $0\leq a<1,\,\,0\leq b,c<1/2$ such that for all $x,y\in M,$
at least one of the next conditions are true:
($TZ_{1}$):
$d(TSx,TSy)\leq ad(Tx,Ty)$.
($TZ_{2}$):
$d(TSx,TSy)\leq b[d(Tx,TSx)+d(Ty,TSy)]$.
($TZ_{3}$):
$d(TSx,TSy)\leq c[d(Tx,TSy)+d(Ty,TSx)]$.
If in Definition 3.2 we take $T=I_{d}$ and $E=\mathbb{R}_{+}$ we obtain the
definition of T. Zamfirescu [18].
###### Lemma 3.1.
Let $(M,d)$ be a cone metric space and $T,S:M\longrightarrow M$ two mappings.
If $S$ is a $TZ-$mapping, then there is $0\leq\delta<1$ such that
(3.1) $d(TSx,TSy)\leq\delta d(Tx,Ty)+2\delta d(Tx,TSx)$
for all $x,y\in M$.
###### Proof.
If $S$ is a $TZ-$mapping, then at least one of $(TZ_{1})$, $(TZ_{2})$ o
$(TZ_{3})$ condition is true.
If $(TZ_{2})$ holds, then:
$\begin{array}[]{ccl}d(TSx,TSy)&\leq&b[d(Tx,TSx)+d(Ty,TSy)]\\\ \\\
&\leq&b[d(Tx,TSx)+d(Ty,Tx)+d(Tx,TSx)+d(TSx,TSy)]\end{array}$
thus,
$(1-b)d(TSx,TSy)\leq bd(Tx,Ty)+2bd(Tx,TSx).$
From the fact that $0\leq b<1/2$ we get:
$d(TSx,TSy)\leq\displaystyle\frac{b}{1-b}d(Tx,Ty)+\displaystyle\frac{2b}{1-b}d(Tx,TSx).$
with $\frac{b}{1-b}<1$. If $(TZ_{3})$ holds, then similarly we get
$d(TSx,TSy)\leq\displaystyle\frac{c}{1-c}d(Tx,Ty)+\displaystyle\frac{2c}{1-c}d(Tx,TSx).$
Therefore, denoting by
$\delta:=\max\left\\{a,\,\displaystyle\frac{b}{1-b},\,\displaystyle\frac{c}{1-c}\right\\}$
we have that $0\leq\delta<1$. Hence, for all $x,y\in M,$ the following
inequality holds:
$d(TSx,TSy)\leq\delta d(Tx,Ty)+2\delta d(Tx,TSx).$
∎
###### Remark 1.
Notice that inequality (3.1) in Lemma 3.1 can be replace by
$d(TSx,TSy)\leq\delta d(Tx,Ty)+2\delta d(Tx,TSy)$
for all $x,y\in M$.
###### Theorem 3.2.
Let $(M,d)$ be a complete cone metric space, $P$ be a normal cone with normal
constant $K$. Moreover, let $T:M\longrightarrow M$ be a continuous and one to
one mapping and $S:M\longrightarrow M$ a $T-$Zamfirescu continuous mapping.
Then
* (i)
For every $x_{0}\in M$,
$\displaystyle\lim_{n\rightarrow\infty}d(TS^{n}x_{0},TS^{n+1}x_{0})=0.$
* (ii)
There is $y_{0}\in M$ such that
$\displaystyle\lim_{n\rightarrow\infty}TS^{n}x_{0}=y_{0}.$
* (iii)
If $T$ is subsequentially convergent, then $(S^{n}x_{0})$ has a convergent
subsequence.
* (iv)
There is a unique $z_{0}\in M$ such that $Sz_{0}=z_{0}$.
* (v)
If $T$ is sequentially convergent, then for each $x_{0}\in M$ the iterate
sequence $(S^{n}x_{0})$ converges to $z_{0}$.
###### Proof.
* (i)
Since $S$ is a $T-$Zamfirescu mapping, then by Lemma 3.1, there exists
$0<\delta<1$ such that
$d(TSx,TSy)\leq\delta d(Tx,Ty)+2\delta d(Tx,TSx)$
for all $x,y\in M$.
Suppose $x_{0}\in M$ is an arbitrary point and the Picard iteration associated
to $S,$ $\;(x_{n})$ is defined by
$x_{n+1}=Sx_{n}=S^{n}x_{0},\qquad n=0,1,2,\ldots.$
Thus,
$d(TS^{n+1}x_{0},TS^{n}x_{0})\leq hd(TS^{n}x_{0},TS^{n-1}x_{0})$
where $h=\displaystyle\frac{\delta}{1-2\delta}<1$. Therefore, for all $n$ we
have
$d(TS^{n+1}x_{0},TS^{n}x_{0})\leq h^{n}d(TSx_{0},Tx_{0}).$
From the above, and the fact the cone $P$ is a normal cone we obtain that
$\|d(TS^{n+1}x_{0},TS^{n}x_{0})\|\leq Kh^{n}\|d(TSx_{0},Tx_{0})\|,$
taking limit $n\longrightarrow\infty$ in the above inequality we can conclude
that
$\displaystyle\lim_{n\rightarrow\infty}d(TS^{n+1}x_{0},TS^{n}x_{0})=0.$
* (ii)
Now, for $m,n\in\mathbb{N}$ with $m>n$ we get
$\begin{array}[]{ccl}d(TS^{m}x_{0},TS^{n}x_{0})&\leq&(h^{n}+\ldots+h^{m-1})d(TSx_{0},Tx_{0})\\\
\\\ &\leq&\displaystyle\frac{h^{n}}{1-h}d(TSx_{0},Tx_{0}).\end{array}$
Again; as above, since $P$ is a normal cone we obtain
$\displaystyle\lim_{n,m\rightarrow\infty}d(TS^{m}x_{0},TS^{n}x_{0})=0.$
Hence, the fact that $(M,d)$ is a complete cone metric space, imply that
$(TS^{n}x_{0})$ is a Cauchy sequence in $M$, therefore there is $y_{0}\in M$
such that
$\displaystyle\lim_{n\rightarrow\infty}TS^{n}x_{0}=y_{0}.$
* (iii)
If $T$ is subsequentially convergent, $(S^{n}x_{0})$ has a convergent
subsequence, so there is $z_{0}\in M$ and $(n_{k})_{k=1}^{\infty}$ such that
$\displaystyle\lim_{k\rightarrow\infty}S^{n_{k}}x_{0}=z_{0}.$
* (iv)
Since $T$ and $S$ are continuous mappings we obtain:
$\displaystyle\lim_{k\rightarrow\infty}TS^{n_{k}}x_{0}=Tz_{0},\qquad\displaystyle\lim_{k\rightarrow\infty}TS^{n_{k}+1}x_{0}=TSz_{0}$
therefore, $Tz_{0}=y_{0}=TSz_{0},$ and since $T$ is one to one, then
$Sz_{0}=z_{0}.$ So $S$ has a fixed point.
Now, suppose that $Sz_{0}=z_{0}$ and $Sz_{1}=z_{1}$.
$\begin{array}[]{ccl}d(TSz_{0},TSz_{1})&\leq&\delta d(Tz_{0},Tz_{1})+2\delta
d(Tz_{0},TSz_{0})\\\ \\\ d(Tz_{0},Tz_{1})&\leq&\delta
d(Tz_{0},Tz_{1})\end{array}$
from the fact that $0\leq\delta<1$ and that $T$ is one to one we obtain that
$z_{0}=z_{1}$.
* (v)
It is clear that if $T$ is sequentially convergent, then for each $x_{0}\in
M$, the iterate sequence $(S^{n}x_{0})$ converges to $z_{0}$.
∎
In 2003, V. Berinde (see, [2], [3]) introduced a new class of contraction
mappings on metric spaces, which are called weak contractions. We will extend
these kind of mappings by introducing a new function $T$ and we define it in
the framework of cone metric spaces.
###### Definition 3.3.
Let $(M,d)$ be a cone metric space and $T,S:M\longrightarrow M$ two mappings.
$S$ is called a $T-$weak contraction, (TW- Contraction,
$T_{(S,L)}-$Contraction), if there exist a constant $\delta\in(0,1)$ and some
$L\geq 0$ such that
$d(TSx,TSy)\leq\delta d(Tx,Ty)+Ld(Ty,TSx)$
for all $x,y\in M$.
It is clear that if we take $T=I_{d}$ and $E=\mathbb{R}_{+}$ then we obtain
the notion of Berinde [2].
Due to the symmetry of the metric, the $T-$weak contractive condition
implicitly include the following dual one:
$d(TSx,TSy)\leq\delta d(Tx,Ty)+Ld(Tx,TSy)$
for all $x,y\in M$.
The next proposition gives examples of $T-$weak contraction and it proof is
similar to the proof of Lemma 3.1.
###### Proposition 3.3.
Let $(M,d)$ be a cone metric space and $T,S:M\longrightarrow M$ two mappings.
* (i)
If $S$ is a TB - contraction, then $S$ is a $T-$weak contraction.
* (ii)
If $S$ is a TK - contraction, then $S$ is a $T-$weak contraction.
* (iii)
If $S$ is a TC - contraction, then $S$ is a $T-$weak contraction.
* (iv)
If $S$ is TZ - mapping, then $S$ is a $T-$weak contraction.
Now we have the following result:
###### Theorem 3.4.
Let $(M,d)$ be a complete cone metric space, $P$ a normal cone with normal
constant $K$. Let furthermore $T:M\longrightarrow M$ a continuous and one to
one mapping and $S:M\longrightarrow M$ a continuous $T-$weak contraction. Then
* (i)
For every $x_{0}\in M$,
$\displaystyle\lim_{n\rightarrow\infty}d(TS^{n}x_{0},TS^{n+1}x_{0})=0.$
* (ii)
There is $y_{0}\in M$ such that
$\displaystyle\lim_{n\rightarrow\infty}TS^{n}x_{0}=y_{0}.$
* (iii)
If $T$ is subsequentially convergent, then $(S^{n}x_{0})$ has a convergent
subsequence.
* (iv)
There is $z_{0}\in M$ such that
$Sz_{0}=z_{0}.$
* (v)
If $T$ is sequentially convergent, then for each $x_{0}\in M$ the iterate
sequence $(S^{n}x_{0})$ converges to $z_{0}$.
###### Proof.
Similar to the proof of Theorem 3.2. ∎
As we see in Theorem 3.2, a $T-$Zamfirescu mapping has a unique fixed point.
The next example shows that a $T-$weak contraction may has infinitely fixed
points.
###### Example 1 ([4]).
Let $M=[0,1]$ be the unit interval with the usual metric and
$T,S:M\longrightarrow M$ the identity maps, that is, $Tx=Sx=x$ for all $x\in
M$. Then, taking $0\leq a<1$ and $L\geq 1-a$ we obtain
$\begin{array}[]{ccl}d(TSx,TSy)&=&|TSx-TSy|\\\ \\\
|x-y|&\leq&a|x-y|+L|y-x|\end{array}$
which is valid for all $x,y\in[0,1]$. Thus the set of the fixed points $F_{S}$
of the map $S$ is the interval $[0,1]$. I.e.,
$F_{S}=\\{x\in[0,1]\,/\,Sx=x\\}=[0,1].$
It is possible to force the uniqueness of the fixed point of a $T-$weak
contraction by imposing an additional contractive condition, as is shown in
the next theorem.
###### Theorem 3.5.
Let $(M,d)$ be a complete cone metric space, $P$ be a normal cone with normal
constant $K$. Let furthermore $T:M\longrightarrow M$ a continuous and one to
one mapping and $S:M\longrightarrow M$ a $T-$weak contraction for which there
is $\theta\in(0,1)$ and some $L_{1}\geq 0$ such that
$d(TSx,TSy)\leq\theta d(Tx,Ty)+L_{1}d(Tx,TSx)$
for all $x,y\in M$. Then:
* (i)
For every $x_{0}\in M$
$\displaystyle\lim_{n\rightarrow\infty}d(TS^{n}x_{0},TS^{n+1}x_{0})=0.$
* (ii)
There is $y_{0}\in M$ such that
$\displaystyle\lim_{n\rightarrow\infty}TS^{n}x_{0}=y_{0}.$
* (iii)
It $T$ is subsequentially convergent, then $(S^{n}x_{0})$ has a convergent
subsequence.
* (iv)
There is a unique $z_{0}\in M$ such that
$Sz_{0}=z_{0}.$
* (v)
If $T$ is sequentially convergent, then for each $x_{0}\in M$ the iterate
sequence $(S^{n}x_{0})$ converges to $z_{0}.$
###### Proof.
Assume $S$ has two distinct fixed points $x^{*},y^{*}\in M.$ Then
$d(Tx^{*},Ty^{*})=d(TSx^{*},TSy^{*})\leq\theta
d(Tx^{*},Ty^{*})+L_{1}d(Tx^{*},TSx^{*})$
thus, we get
$d(Tx^{*},Ty^{*})\leq\theta
d(Tx^{*},Ty^{*})\Leftrightarrow(1-\theta)d(Tx^{*},Ty^{*})\leq 0.$
Therefore, $d(Tx^{*},Ty^{*})=0$. Since $T$ is one to one, then $x^{*}=y^{*}$.
The rest of the proof follows as the the proof of Theorem 3.2. ∎
## References
* [1] M. Abbas and B.E. Rhoades, Fixed and periodic results in cone metric space, Appl. Math. Lett., 22, (4), (2009), 511–515.
* [2] V. Berinde, Iterate Approximation of fixed points, lect. Notes Math., Vol 1912, (2nd ed.), Springer Verlag, Berlin, 2007.
* [3] V. Berinde, On the approximation of fixed points of weak contractive mappings, Carpathian J. Math, 19, (1), (2003), 7–22.
* [4] V. Berinde, On the convergence of the Ishikawa Iteration in the class of quasi contractive operators, Acta Math. Univ. Comenianae, 73, (1), (2004), 119–126.
* [5] A. Beiranvand, S. Moradi, M. Omid and H. Pazandeh, Two fixed point theorem for special mapping, arXiv:0903.1504v1 [math.FA].
* [6] S. K. Chatterjea, Fixed point theorems, C. R. Acad. Bulgare Sci., 25, (1972), 727–730.
* [7] D. Ilić and V. Rakočević, Quasi-contraction on a cone metric space, Appl. Math. Lett., 22, (5), (2009), 728–731.
* [8] Z. Kadelburg, S. Radenović and V. Rakočević, Remarks on “Quasi-contraction on a cone metric space”, Appl. Math. Lett., (2009), doi:10.1016/j.aml.2009.06.003.
* [9] R. Kannan, Some results on fixed points, Bull. Calcutta Math. Soc., 60, (1968), 71–76.
* [10] M.S. Khan and M. Samanipour, Fixed point theorems for some discontinuous operators in cone metric space, Mathematica Moravica, Vol 12-2, (2008), 29–34.
* [11] Huan Long - Guang and Zhan Xian, Cone metric spaces and fixed point theorems of contractive mappings, J. Math. Anal. Appl., 332, (2007), 1468–1476.
* [12] S. Moradi, Kannan fixed point theorem on complete metric spaces and on generalized metric spaces depended on another function, arXiv:0903.1577v1 [math.FA].
* [13] J. Morales and E. Rojas, Cone metric spaces and fixed point theorems of $T-$contractive mappings, preprint, 2009.
* [14] J. Morales and E. Rojas, Cone metric spaces and fixed point theorems of $T-$Kannan contractive mappings, arXiv:0907.3949v1 [math.FA].
* [15] H.K. Pathak and N. Shahzad, Fixed points results for generalized quasicontraction mappings in abstract metric spaces, Nonlinear Analysis, (2009), doi:10.1016/j.na.2009.05.052.
* [16] V. Raja and S.M. Vaezpour, Some extension of Banach’s contraction principle in complete cone metric spaces, Fixed Point Theory and Applications, (2008), 11 p.
* [17] Sh. Rezapour and R. Hamlbarani, Some notes on the paper “Cone metric spaces and fixed point theorems of contractive mappings”, J. Math. Anal. Appl., 345, (2), (2008), 719–724.
* [18] T. Zamfirescu, Fixed points theorems in metric spaces, Arch. Math., 23, (1972), 292–298.
|
# Artificial Concepts of Artificial Intelligence: Institutional Compliance and
Resistance in AI Startups
Amy A. Winecoff<EMAIL_ADDRESS>Princeton UniversitySherrerd
HallPrincetonNew JerseyUSA08544 and Elizabeth Anne Watkins
<EMAIL_ADDRESS>Princeton UniversitySherrerd HallPrincetonNew
JerseyUSA08544
(2022)
###### Abstract.
Scholars and industry practitioners have debated how to best develop
interventions for ethical artificial intelligence (AI). Such interventions
recommend that companies building and using AI tools change their technical
practices, but fail to wrangle with critical questions about the
organizational and institutional context in which AI is developed. In this
paper, we contribute descriptive research around the life of ”AI” as a
discursive concept and organizational practice in an understudied
sphere–emerging AI startups–and with a focus on extra-organizational pressures
faced by entrepreneurs. Leveraging a theoretical lens for how organizations
change, we conducted semi-structured interviews with 23 entrepreneurs working
at early-stage AI startups. We find that actors within startups both conform
to and resist institutional pressures. Our analysis identifies a central
tension for AI entrepreneurs: they often valued scientific integrity and
methodological rigor; however, influential external stakeholders either lacked
the technical knowledge to appreciate entrepreneurs’ emphasis on rigor or were
more focused on business priorities. As a result, entrepreneurs adopted hyped
marketing messages about AI that diverged from their scientific values, but
attempted to preserve their legitimacy internally. Institutional pressures and
organizational constraints also influenced entrepreneurs’ modeling practices
and their response to actual or impending regulation. We conclude with a
discussion for how such pressures could be used as leverage for effective
interventions towards building ethical AI.
organizational theory, artificial intelligence, industry practice, qualitative
methods, ethical systems
††journalyear: 2022††copyright: rightsretained††conference: Proceedings of the
2022 AAAI/ACM Conference on AI, Ethics, and Society; August 1–3, 2022; Oxford,
United Kingdom††booktitle: Proceedings of the 2022 AAAI/ACM Conference on AI,
Ethics, and Society (AIES’22), August 1–3, 2022, Oxford, United Kingdom††doi:
10.1145/3514094.3534138††isbn: 978-1-4503-9247-1/22/08††ccs: Social and
professional topics Socio-technical systems††ccs: Applied computing Sociology
## 1\. Introduction
Academic researchers, advocacy groups, and technology companies have created
guidelines and tools for developing ethical artificial intelligence (AI)
(Morley et al., 2020; Fish and Stark, 2021). This research is intended to
ameliorate the considerable negative social impacts produced in AI systems,
such as how AI models encode racial and gender biases (Caliskan et al., 2017;
Noble, 2018; Sweeney, 2013; Rekabsaz and Schedl, 2020; Bolukbasi et al.,
2016), worsen disordered eating and body dysmorphia (Karizat et al., 2021),
and magnify inequality (Obermeyer et al., 2019; Dastin, 2018; Angwin et al.,
2016). However, the real world-utility of available interventions for ethical
AI remains unclear.
As with any research intended for real-world applications, robust
consideration of the context of implementation is critical. An emerging body
of research has begun to recognize that effective change demands non-technical
strategies to contend with organizational context (Rakova et al., 2021), such
as the conditions inside technology firms which might influence, or even
prevent, the effectiveness of interventions for ethical AI. Tight development
timelines, lack of formal organizational processes, and challenging internal
stakeholder dynamics shape how real-world companies can move in the direction
of more ethical AI development (Holstein et al., 2019; Madaio et al., 2020,
2021; Rakova et al., 2021; Hopkins and Booth, 2021). Studies of AI ethics in
organizational contexts often focus on interventions such as model fairness
(Holstein et al., 2019; Madaio et al., 2020, 2021) and model interpretability
techniques (Bhatt et al., 2020; Hong et al., 2020; Kaur et al., 2020).
However, such studies have largely been constrained to organizations that are
mature enough to consider specific AI ethical interventions to begin with.
AI startups constitute a growing portion of the technology sector (Tricot,
2021). As a result, these companies and the ethical practices they embrace are
likely to play a significant role in the impact of future technology on
society. Only a handful of studies have characterized ethical AI development
at smaller firms (Vakkuri et al., 2020; Hopkins and Booth, 2021). This
emerging area of research has begun to illuminate the unique challenges
nascent firms must address when attempting to adopt responsible, transparent,
and accountable AI practices. For example, as with more mature companies,
small firms must navigate complex dynamics amongst stakeholders like clients,
investors, and regulators but unlike more established organizations, they must
do so under significant resource constraints that threaten their very
existence (Hopkins and Booth, 2021). Therefore, the ethical AI practices they
are able to adopt are necessarily limited.
While existing research has illustrated the organizational constraints to
ethical AI, especially intra-organizational dynamics, less is known about how
the inter-organizational or field-level dynamics shape firms’ capacity to
develop ethical approaches. The field-level, i.e., ”institutional” dynamics
and market-based pressures that impact an organization’s chances of survival
inevitably alter the structures and practices firms adopt (Meyer and Rowan,
1977; Salancik and Pfeffer, 1978; DiMaggio and Powell, 1983; Oliver, 1991). By
behaving in ways that conform with institutional expectations, emerging
organizations can improve their social and cultural fitness; however,
institutional expectations sometimes conflict with each other and also with
economic pressures. As a result, emerging organizations such as AI startups
must skillfully navigate a complex gauntlet of social, cultural, and economic
challenges. How these field-level dynamics factor into the ethical choices of
startups, such as their decisions around the use of AI, is an under-explored
area of research.
Here, we contend that before effective ethical AI practices for startups can
be developed, an understanding of the inter-organizational and institutional
dynamics these firms face must be developed. Building on recent scholarship
that takes a contextual and organizational approach to ethical AI, we engage
in descriptive research around the life of ”AI” as a discursive concept and
organizational practice that is situated within an institutional context.
Instead of focusing on ethical practices directly, we take a step back to ask
fundamental questions about the forces that shape the very nature of how
entrepreneurs define, build, and talk about AI itself.
To that end, we ask two research questions:
RQ1: What institutional pressures influence how startup entrepreneurs define,
discuss, and build AI?
RQ2: When do entrepreneurs comply with, avoid, or resist these pressures?
To address these questions, we conducted semi-structured interviews with 23
individuals working at early-stage startups across a range of industry
domains. In our interviews, we focused on the financial, regulatory, and
normative pressures AI startups encounter. Using abductive analysis, we
illustrate how AI entrepreneurs both comply with and resist institutional
pressures through the technological and business practices they employ. We
find that AI entrepreneurs’ face a tension between the expectations of
technology entrepreneurship, which rewards rapid development and optimistic
promises about technology’s potential, and entrepreneurs’ own values of
scientific integrity, which prioritize meticulous practices and encourages
skepticism. This tension was further heightened by external stakeholders’
unrealistic expectations about the potential of AI, particularly when such
stakeholders had limited technical knowledge. We also find that whereas AI
entrepreneurs saw privacy regulation as beneficial and aligned with their own
values of autonomy, they held less uniformly positive views of other AI
regulatory processes such as those employed by the Food and Drug
Administration (FDA) in approving AI medical devices. Drawing from our
theoretical motivations, we conclude with a discussion of how our results
point to both constraints and opportunities for future research on ethical
interventions for AI startups.
## 2\. Related Works
Organizational dynamics are a significant source of influence on the
effectiveness of interventions for ethical and responsible AI. Practitioners
in well-resourced organizations have expressed aspirations for ethics-
supportive structures as such cross-team integration, risk-anticipation
frameworks, and firm-level mission and values (Rakova et al., 2021). Our own
research builds on these recent findings by addressing inter-organizational
and institutional conditions (i.e., external pressures) that are likely to be
sources of change for organizations, especially emerging organizations such as
AI startups, which typically lack formal mechanisms for addressing ethical
concerns (Vakkuri et al., 2020; Hopkins and Booth, 2021). In this section we
provide an overview of the relevant frameworks we draw from in organizational
theory to explain how organizations adopt procedures and adapt over time to
institutional pressures.
### 2.1. Resource Dependency and Institutionalism: How Organizations Change
Organizational theory offers different frameworks to understand how
organizations change. In the early days of the discipline, the dominant
paradigm was of rationality: theorists described organizations as rational
systems, machines for achieving a goal in the market, and that all
organizational decisions were imbued with this same mechanical, systematic
precision. Within this school were early 1900s thinkers like the German
sociologist Max Weber, with his focus on bureaucracies as structural
realizations of rational authority (Weber, 1978), and American mechanical
engineer Frederick Winslow Taylor, with his focus on bringing ”scientific”
methods to management to wring ever-greater ”efficiency” out of a labor force
(Taylor, 1919).
Starting in the 1970s, however, the field took a relational turn, recognizing
that organizations do not operate in a vacuum of rationality but instead
within complex ecologies of other actors. Two schools of thought–resource
dependency theory and institutionalism–both address how firms seek to mitigate
external pressures and uncertainty within their organizational ecosystems.
Resource dependency theorists focus on organizations as their unit of
analysis, i.e. the ”meso” or ”middle” level of institutional change (bigger
than ”micro”, or individual people, but not as large as ”macro”, or field-
level norms or systems). They examine the interactions between these units,
using this perspective to analyze how organizations strategically seek to
manage resources and mitigate dependencies on their exchange partners
(Salancik and Pfeffer, 1978). In doing so, organizations improve their fitness
within the market. Resource dependency theory has been recently used to
analyze the precarity of firms operating within complex supply chains, as they
”require networks to accommodate the interdependencies in product and service
flows, resource flows, and information flows” (Olan et al., 2022).
The theory of institutionalism looks at the ”macro” level, foregrounding
patterns taking place at the level of entire organizational fields or social
orders. Institutionalism focuses on unconscious social and cultural
expectations, contending that these influences lead to widespread changes in
multiple organizations, shaping fields of industry (DiMaggio and Powell, 1983;
Meyer and Rowan, 1977; Zucker, 1977; Van Wijk et al., 2019). A critical
component of institutional theory examines how new organizations establish
legitimacy, where their actions are perceived as ”desirable, proper, or
appropriate within some socially constructed system of norms, values, beliefs,
and definitions” (Suchman, 1995). Amidst technological and market uncertainty,
new firms improve their odds of survival by accruing legitimacy from an
audience of stakeholders in the field including funding entities, regulatory
bodies, and competitor companies. In their pursuit of legitimacy,
organizations change over time, increasingly reflecting the established norms
and practices of the field (DiMaggio and Powell, 1983). Recent scholarship has
used this lens to examine organizational changes around the implementation of
novel technologies and practices such as how news publishers choose to
implement novel cybersecurity tools (Watkins and Anderson, 2019) and how AI
governance and accountability (i.e. algorithmic impact assessment) may become
more widespread (Selbst, 2021).
Both frameworks have limitations in their explanatory power; resource
dependency theory under-emphasizes sociocultural forces and institutionalism
under-emphasizes organizations’ instrumental actions (Scott, 2013). Oliver
(Oliver, 1991) argues that neither institutionalism nor resource dependency
can adequately capture the complexity of organizational action and evolution
within an institutional context. She theorizes that in addition to exercising
agency in response to market demands, organizational actors can also respond
strategically to institutional pressures through a variety of compliance and
resistance tactics. Oliver details five core strategies organizations adopt in
response to institutional pressures varying from passive compliance to active
resistance. On the compliance end of the spectrum, she places ”acquiescence”
in which organizations conform to institutional expectations. She describes
”avoidance” and ”compromise” as partial compliance strategies in which
organizations, for example, attempt to disguise their non-conformity with
institutional norms or attempt to balance the sometimes conflicting pressures
of institutional stakeholders. On the resistance end of the spectrum, she
places ”defiance,” in which organizations actively and openly reject
institutional norms, and ”manipulation,” in which organizations attempt to co-
opt, supplant, or control institutional pressures.
Oliver also theorizes organizations’ likelihood of engaging in compliance or
resistance strategies will depend on a variety of institutional conditions.
The greater the perceived benefit of institutional conformity, the likelier
organizations are to comply with such pressures. On the other hand, competing
stakeholder expectations, multiple conflicting institutional norms, or
conflicting organizational goals and institutional pressures are likelier to
engender resistance. Moreover, when organizations are coerced into conformity
via legal means, they are likelier to resist as compared to when institutional
conformity is effected through diffuse institutional norms that are adopted
voluntarily by organizations. Through the exertion of skillful agency,
organizations can both gain legitimacy through selective compliance with
institutional norms, while also maintaining the practices they adopt in the
service of market fitness (Oliver, 1991; Fligstein, 1997; Suddaby and
Greenwood, 2005).
In our current work, we draw on Oliver’s framework (Oliver, 1991) to
conceptualize how AI startups attempt to manage institutional and market
pressures through a variety of compliance and resistance strategies. We
consider the conditions that contribute to startups’ responses to
institutional and inter-organizational pressures and discuss how such
strategies might interact with or constrain the ethicality of applied AI.
## 3\. Methods
In this section, we describe our methods for gathering data to answer our
research questions. We elected to conduct qualitative interviews, as
interviews are an ideal method for better understanding actors’ cognitive
interpretations of their social reality, and for accessing their own
explanations of their behavioral practices within that social reality.
### 3.1. Participant Recruitment & Sampling
We recruited participants, which we refer to here as ”entrepreneurs,” from US-
based, early-stage startups that involve a significant AI, machine learning,
or predictive analytics component. Our focus on early-stage startups was
driven by theory (Palinkas et al., 2015); because emerging organizations face
many threats to their survival, they are heavily dependent on other
organizations. As a result, inter-organizational and institutional dynamics
are likely to factor significantly into their behavior. We define “early-
stage” as companies that had raised less than $50M in funding from any source
or who were at or before the Series B stage. As our interviews progressed and
patterns related to both regulatory pressures, especially privacy and the FDA,
and related to funding, especially VC and crowdfunding, we began targeting our
recruitment efforts towards additional startups that would further illuminate
these trends. We recruited participants through a variety of methods
including: 1) posts to AI and technology related listservs, message boards,
social media, and Slack groups; 2) messages to general company contact email
addresses or to specific individuals through email or LinkedIn; and 3) through
our own direct or indirect professional contacts. In total, we interviewed 23
entrepreneurs from 20 different companies. Our entrepreneurs’ companies came
from a variety of industry domains including healthcare (n=7), business
analytics (n=6), fitness and wellness (n=5), design and engineering services
(n=2), aviation (n=1), social planning (n=1), and agriculture (n=1). A
breakdown of the self-described demographic, educational, and professional
characteristics of our sample are available in the Supplementary Materials
111https://arxiv.org/abs/2203.01157. Participants were sent a $25 gift
certificate in exchange for their participation.
### 3.2. Interview Protocol
At the beginning of the interview session, we described to our entrepreneur
participants our practices for protecting their privacy and confidentiality,
then read them a verbal consent script, and gave them an opportunity to ask
any questions. After providing consent, the first author asked questions based
on the interview instrument (provided in the Supplementary Materials). In
general the interview instrument was designed to surface data relevant to
several core areas related to our research questions, including questions
about the overarching aims of the company and the entrepreneur’s role within
it, followed by questions tailored to entrepreneur’s area of expertise. For
example, in cases where entrepreneurs were involved in the company’s AI
development, we asked additional technical questions about data collection,
choice of models, evaluation criteria, and infrastructure. In most cases, we
asked entrepreneurs about their companies’ existing sources of financing and
their plans for fundraising. We also asked entrepreneurs for their personal
definition of AI. Lastly, we asked entrepreneurs about the social or ethical
implications of their company. Audio recordings of the interviews were sent to
a third party service for transcription, which were then verified by the first
author. Our study design and protocol were approved by the Princeton
University Institutional Review Board (IRB).
### 3.3. Data Analysis
We adopted an abductive approach to our analysis (Timmermans and Tavory, 2012;
Tavory and Timmermans, 2014), which allowed us to iterate between deductive
analysis guided by relevant theory and inductive analysis guided by emergent
patterns in our data. To facilitate this process, the first author initially
selected 11 transcripts that contained discussions of theoretically-meaningful
themes or that illustrated common patterns across participants. In a
preliminary analysis phase, both authors read each of these transcripts and
applied descriptive codes (Saldana, 2021). After discussing these codes in
detail, transcripts were re-coded using line-by-line in-vivo codes (i.e.,
using participants’ own words) in an effort to better preserve entrepreneurs’
perspectives (Saldana, 2021). For example, one participant discussed the
drawbacks of venture capital (VC), relating VC financing to rocket fuel:
”There are actually very few businesses where rocket fuel is the right thing
[P13].” This excerpt was tagged with the in-vivo code ”rocket fuel.” In-vivo
codes were then aggregated into groupings of similar topics. For example,
codes related to VC funding were grouped with the ”rocket fuel” quote. The
authors subsequently discussed the in-vivo codes as well as relevant theory
and chose to focus the next analysis phase on five core themes: 1) the ”AI
hype cycle” or how ”buzz” surrounding AI drives external stakeholders’
interest in AI companies; 2) practices surrounding the scientific legitimacy
of entrepreneurs’ AI approaches; 3) pressures to raise funds or secure
clients; 4) the impact of regulations; or 5) entrepreneurs’ own personal
beliefs and ethical values that relate to their companies. These themes all
appeared in multiple interviews with participants and had direct relevance to
the study’s theoretical focus on institutional and organizational theory. All
23 transcripts were then coded according to the five themes. The authors
frequently discussed transcripts and code applications to achieve consensus.
We did not measure inter-rater reliability. Inter-rater reliability is a
methodologically unhelpful tool for interpretive research, when codes comes
out of the collaborative process between researchers and consultation with
literature, and not emergent ground-up from data (McDonald et al., 2019).
## 4\. Results
### 4.1. Organizational Responses to Financial Pressures
The need to signal legitimacy to sources of financial support (e.g., investors
and clients) constituted a significant vector of influence on how
entrepreneurs defined, spoke about, and developed practices for AI. A tension
that arose repeatedly in our interviews derived from a conflict between
institutional values, specifically the values of science as a practice, and
the values of technology entrepreneurship. Whereas scientific practice values
systematic, methodological approaches paired with conservative interpretations
of findings, technology entrepreneurship values rapid innovation and
aspirational visions that extend beyond current technological reality, i.e.,
the ”fake it ’til you make it” Silicon Valley culture. In a variety of ways,
AI entrepreneurs attempted to mitigate the conflicts between the values of
entrepreneurship and science by decoupling their external rhetoric from their
day-to-day practices.
One way this decoupling manifested was through the use of the idea of ”AI”
itself. Entrepreneurs leveraged the concept of ”AI” as a symbol of their
technological proficiency even though they personally harbored disdain for the
technical ambiguity of the concept. According to our entrepreneurs, ”AI” had
no precise technical meaning and was instead employed as an operational tool
to signal legitimacy to resource-rich external stakeholders [P1, P2, P5, P6,
P7, P9, P10, P11, P13, P14, P15, P18, P21] rather than as an accurate
descriptor of what their companies actually do. In other words, ”It’s just a
buzz word [P6]” for primarily marketing benefit.
Entrepreneurs described a widespread belief that companies benefited from
marketing themselves as ”AI” companies regardless of the nature of their
underlying technology. They expressed frustration with their peers who ”got to
use the hype term [P18]” without employing any technical practices that
entrepreneurs judged as legitimate. Entrepreneurs described feeling annoyance
with these AI imposters, but nevertheless admitted to employing the same
marketing tactics themselves. Faced with a competitive landscape in which
startups’ technical and business value cannot be objectively verified,
entrepreneurs leverage the institutional expectations around the legitimacy of
AI because it ”gives credibility that we’re on the cutting edge of stuff
[P2].”
Despite the prevalence of this narrative in our interviews, only rarely were
entrepreneurs able to provide specific explanations or concrete examples of
how the abstract idea of AI yielded a tangible benefit. One entrepreneur,
however, pinpointed investors’ fear of missing out on deals as a key driver of
the ”AI hype cycle [P6]”:
> I think [the AI space] feels very confusing to [investors], but they also
> feel like there’s every signal that it’s super lucrative. […] The key thing
> that keeps all of the subordinates [up at night], the ones whose job it is
> to go find those deals and make sure their bosses don’t miss any great deals
> […] is a version of the world, where you passed on Lyft. And then your boss
> comes back to you five years later and is like, ”I would’ve made a billion
> dollars off of Lyft. […] What’s wrong with you? [P10]”
Entrepreneurs emphasized that although some investors and clients have AI
expertise, most do not have the technical background required to adequately
evaluate an AI solution or were simply ”totally disinterested in the technical
details [P6].” As a result, entrepreneurs face institutional pressures to
describe their technology using homogenized, hyped language about AI even
though their underlying algorithmic approaches were heterogeneous and often
carefully devised. Through their instrumental use of hyped AI messaging, AI
startups engage in what Oliver (Oliver, 1991) refers to as ”concealment.”
Externally startups affect the appearance of compliance with institutional
expectations surrounding technology entrepreneurship even as their internal
practices diverged, often significantly, from this affectation; ”AI” became a
discursive tool for avoiding institutional pressures via a process of ”window
dressing” (Oliver, 1991).
In contrast to their external messaging, internally, entrepreneurs sought to
achieve high standards of scientific rigor and validity. For example,
entrepreneurs emphasized their rigorous data selection and curation processes
[P4, P5, P18], described checks on the validity of their systems [P4],
designed algorithm evaluations to appropriately assess the performance of
their systems [P2, P4, P5, P9], and even employed independent validations by
academic collaborators to ensure that their models had good generalizability
[P5]. Entrepreneurs also described translating methods and findings from the
academic literature in their products [P7, P17, P15] and employing scientific
subject matter experts either directly on their development teams or
indirectly through advisors or boards of directors [P7, P17, P18].
Yet these scientific priorities could engender serious conflicts with the
priorities of users, clients, and investors. For example, when entrepreneurs
attempted to use scientific legitimacy as a differentiator in pitches, this
attempt was sometimes regarded by investors or potential clients as confusing
or unconvincing. Two entrepreneurs described remarkably similar experiences,
in which presentations about the scientific merits of their technologies were
dismissed by investors as being merely ”science projects [P5]” or ”high school
projects [P3],” having little business relevance, which from one
entrepreneur’s perspective, felt ”like an anti-science trivialization of what
scientists do [P5].”
External stakeholders’ beliefs about the potential of AI also created a
barrier for entrepreneurs to be readily transparent in their external
messaging about their models’ methodological strengths and weaknesses.
Instead, dovetailing a finding briefly touched upon by (Hopkins and Booth,
2021), we observed institutional pressure around how ”quality” in AI ought to
be reported [P4, P5, P14, P21], specifically in external stakeholders’
arbitrary notions of what constitutes ”good” model accuracy. Though an
algorithm’s accuracy may seem objective, in practice, accuracy metrics involve
many subjective choices. For example, the practical applicability of a measure
of an algorithm’s accuracy is contingent upon its mathematical formulation
(e.g., area under the curve (AUC), F1, sensitivity, etc.) as well as
contextual relevance (e.g., the severity of a false positive versus a false
negative for a medical test) and which data are selected. In our interviews,
entrepreneurs felt that in order to obtain the resources necessary for their
companies to survive, they needed to engage in rhetorical messaging that
complied with stakeholders’ expectations about model performance, even if
these metrics were not the best reflection of the task at hand, nor a valid
reflection of their algorithms’ capabilities.
> How it’s measured is we have to make sure it’s 90% or above […]. So if we
> need to switch from top 3 accuracy to top 5, just people seeing a 9, they
> don’t even think about what it’s measuring … People just have artificial
> concepts of what’s good and what’s bad [P4].
Pressure to present model metrics that have the right ”psychological effect
[P14]” on outside stakeholders was in conflict with entrepreneurs’ desire to
adopt methodologically rigorous AI approaches internally. Strikingly, one
explained that his attempts to include diverse training data in the service of
higher out-of-sample generalization damaged his company’s credibility when his
models’ performance was compared to competitors who use less realistic data.
> Ultimately, our results aren’t going to be as stellar as a lot of others
> because now we have to account for […] all the variability within the data
> set whereas, if we’re just focused on one homogeneous data set, our accuracy
> stats will be higher. So, that has been one sticky, difficult point in terms
> of head-to-head comparisons [P5]
In the service of survival, entrepreneurs sometimes conformed at least
superficially to the pressures of stakeholder expectations, adapting their
external messaging over time to provide a level of scientific detail that was
persuasive to the target audience. In this way, they again engage in what
Oliver refers to as concealment tactics. However, these pressures did not
entirely undermine their desire to externally project the methodological rigor
they prioritized internally. Instead, entrepreneurs chose to target their
products to specific stakeholders who would be more receptive to messages
about the product’s scientific credibility or its technical utility within a
domain. For example, one entrepreneur highlighted that having extensive
scientific references available on their product’s website attracted desirable
early users:
> ”[…] most marketers are like, ”I don’t think that sells the product,” but we
> disagree. […] I’m not sure it makes it so it’s a blockbuster of a product,
> but it brings in the right type of people for your product […]. It brings in
> good early adopters, anyways [P7]”
In this public commitment to scientific integrity, entrepreneurs engage in a
form of ”defiance” (Oliver, 1991). Oliver hypothesizes that defiance is more
likely when the perceived cost of resistance to institutional norms is low and
”when they can demonstrate the rationality or righteousness of their own
alternative convictions or conduct.” AI entrepreneurs who externally project
their strongly-held personal values of the scientific process may do so
because they can promote science as a virtue while still attracting science-
inclined external stakeholders.
Entrepreneurs recognized that in order for their companies to grow, their
products must eventually translate into market success and that their
companies’ investors were ultimately motivated by whether or not the company
would ”provide liquidity [P17]” on their investment. Nevertheless,
entrepreneurs demonstrated a wide range of compliance and resistance tactics
when it came to their decisions regarding financial backing. They described
pursuing a variety of strategies to fund their businesses including revenue
[P8, P13, P21], friends and family raises [P1, P10], grants from government or
private entities [P2, P12, P17], angel investors [P4, P5, P6, P7, P13, P14,
P15, P17, P21], debt financing [P12], VC financing [P4, P10, P14, P16, P17,
P18, P20], and crowdfunding [P7, P9, P13, P21, P23]. Even still, several
entrepreneurs pointed out that VC is regarded as a default financing path:
> We looked at the VC funding route in the beginning because that’s what
> you’re told to do, right? That’s how you get funded. This is the path. You
> go pre-seed, it’s angel investors, after that it’s VCs, and then you go
> through the Series process [P12].
For some, capital from VC firms formed a cornerstone of their strategy for
building their business [P4, P10, P16, P17,P18, P20]. These entrepreneurs saw
investors not only as a source of financial capital, but also of valuable
industry domain and business expertise as well as a mechanism for accessing
important professional networks. Entrepreneurs viewed the fit between the
needs of the company and the expertise of investors as a critical component of
establishing a productive relationship.
> So, we have a number of investors, and the asks really change based on
> business needs. So it’s really, what do we really need today, this week,
> this month, that can help us take the business to the next level, and who do
> we have as investors that we can ask for help in those areas? [P17].
However, entrepreneurs did not blindly acquiesce to the demands of investors.
Where their own goals conflicted with the goals of investors, VC or otherwise,
they would sometimes decline further involvement. For example, one
entrepreneur described ending early conversations with an investor because
their desired exit strategy was not consistent with her own goal to eventually
take the company public [P1]. In another case, an entrepreneur described
evaluating potential investors based on their alignment with the company’s
ethical values:
> We are trying to raise capital from investors that have the same kind of
> values and mindset with us and people who are not afraid to lose certain
> revenue or sales just to follow the same values. We had clients asking us to
> do things that we said, you know what? No. No, this is not something we feel
> comfortable with doing [P23].
One entrepreneur noted that because of intense investor interest in the field
of AI, instead of entrepreneurs doggedly pursuing financiers, ”they find you
[P10].” In a resource-rich environment, entrepreneurs have more latitude to
resist institutional pressures arising from dependencies on investors. To some
extent, we saw this resistance in how entrepreneurs described choosing
specific investors; however, we saw even greater resistance amongst
entrepreneurs who expressed hesitancy about pursuing VC financing [P7, P9,
P12, P13, P16, P21]. In these cases, VCs were seen as having financial goals
that conflicted with entrepreneurs’ long-term business or product objectives.
One entrepreneur noted that unlike traditional software, AI software
development typically requires specialized algorithm expertise,
infrastructure, expensive data labeling, and continuous model performance
monitoring; however, if VC firms’ valuation of AI companies is based on their
knowledge of traditional software startups, they may impose timelines or key
performance indicators that undermine what entrepreneurs believe to be
methodologically-sound practices in AI development. Recognizing that VCs’
profit motives and responsibility to their limited partners (LPs) could
conflict with their own goals, entrepreneurs often described actively avoiding
VC financing.
> That was my whole experience [at a previous startup]. The VCs wanted to hype
> things up, get a lot of press, make a splash, so they could raise the next
> round at a higher valuation and look good to their LPs, which was actually
> contrary to what we needed to do for the slow growth to build the business
> [P13]
Instead of VC funding, several entrepreneurs described using crowdfunding to
finance their businesses [P7, P9, P13, P21, P23]. Crowdfunding and other
financing vehicles that provide investors with little direct control over the
companies’ behavior were viewed by entrepreneurs as a way to maintain autonomy
over their businesses, control progress towards the product vision, and to
maintain the equity of current employees. Moreover, although prior literature
has indicated that the disclosure requirements of crowdfunding platforms can
create a risk for companies’ subsequent financing prospects (Blaseg et al.,
2021), entrepreneurs valued the transparency associated with public
disclosures on crowdfunding platforms and the opportunity to directly engage
with potential crowd investors. The choice of crowdfunding over the ”default”
VC financing path constitutes another form of organizational defiance; these
entrepreneurs challenge the culturally dominant mode of startup funding by
choosing financing paths that they felt would better serve their long-term
objectives. As predicted by Oliver (Oliver, 1991), startups are able to engage
in resistance because doing so does not compromise their chances of survival
since they can rely on alternatives to VC to fund their businesses.
### 4.2. Organizational Responses to Regulatory Pressures
In addition to financial pressures, entrepreneurs also described their
compliance with and resistance to pressures in the form of regulation,
especially regulations surrounding privacy and the Food and Drug
Administration (FDA) approval processes for applications of AI in medicine. On
the whole, entrepreneurs viewed privacy protections as normatively good, a
competitive necessity, or even a competitive advantage compared to industry
peers who are slower to adopt privacy-protecting practices. One entrepreneur
described how his company had chosen to temporarily avoid adapting to the
”landscape of privacy and privacy laws [that are getting] a lot more strict
[P19]” by operating selectively in markets that were subject to less stringent
privacy regulation (e.g., in the US versus European Union where GDPR is
applicable), but even this entrepreneur noted that circumvention was only
tenable in the short term. Yet in contrast to (Hopkins and Booth, 2021), most
entrepreneurs in our study did not express resistance to or subversion of
privacy regulations but openly endorsed them as well as the ethical values
underlying them, such as personal autonomy. In their discussions about
privacy, entrepreneurs sometimes contrasted their own beliefs and policies
with those of large technologies such as Facebook, whose privacy-related
behaviors they generally regarded as reprehensible [P9, P6, P23].
Unlike privacy regulation, the FDA approval processes for AI in healthcare was
viewed less favorably. Entrepreneurs who discussed the FDA viewed these
regulatory requirements as unnecessarily onerous and in some cases,
unscientific [P2, P4, P5, P7, P16]. A theme that arose in our interviews with
multiple entrepreneurs was that the lack of standardization in the FDA
approval process for AI-enabled healthcare products created a serious burden
for startups with an unclear upside. One entrepreneur from a more mature
startup described how challenges posed by the FDA approval process contributed
to her company’s decision to eventually pivot away from the AI products she
felt had life-saving potential to a core product and business model that she
thought was more profitable but ultimately less useful to society. She
highlighted the opportunity and financial costs of pursuing FDA approval as
well as the difficulty of providing evidence to meet FDA performance
standards:
> Whenever you’re thinking about a rule-out device–in our case, [the finding
> that a medical test is normal]–that means you rule out every possible thing,
> it’s statistically insanely hard to do. And in order to get approved, you
> would literally have to do better than humanly possible [P16].
All but one entrepreneur [P18] who discussed relevant FDA regulation found the
ad-hoc approval process to be legally or financially arduous; however, there
was less consensus on whether the FDA’s model performance standards were
unreasonable. One entrepreneur noted:
> The data that needs to be provided in order to get clearance in our opinion
> is relatively low, but it does take a lot of money and other things to get
> [P5]
Similar to this perspective, another noted that if the financial demands for
securing FDA approval were lower, they would begin reallocating their research
and development efforts to undergo FDA approval as soon as possible since
their model performance was already strong.
Entrepreneurs either implied or stated explicitly that the onerous FDA
approval process stifled innovation, but they also noted that such regulations
were necessary to protect users from harmful products. Another entrepreneur
noted that regulatory approval could even be beneficial to his businesses
because ”customers are much more receptive to the FDA stamp than they are to
stats [P5]”. Still, the perception that the FDA approval process was ”a whole
monster [P4]” that was often not worth pursuing motivated AI entrepreneurs to
attempt to operate in regulatory gray areas or exploit loopholes so that they
could continue to pursue technological or product objectives. Such
entrepreneurs described using other, non-regulatory avenues to demonstrate
their legitimacy such as publishing their model details and performance in
academic journals or technical whitepapers. It is important to note that
entrepreneurs’ opposition to the idiosyncratic FDA approval process was not
merely a matter of logistical difficulty; they also viewed the discretionary
nature of the FDA review process as an opportunity for established companies
to be unfairly advantaged:
> It’s also, I think, unfair how the FDA, […] they have existing relationships
> with Pfizer and Johnson & Johnson. I get it. But they’re obnoxiously hard on
> startups because they’re not known [P16]
Thus, entrepreneurs’ personal values only partially aligned with the
institutional pressures from the FDA–they value the consumer protection
intent, but decry its consequences for innovation and believe it to be at odds
with a fair, competitive marketplace. In contrast, entrepreneurs were more
likely to adopt privacy preserving practices, which were consistent with their
own normative beliefs, even in cases when companies were not yet subject to
strict privacy regulation. In other words, consistent with Oliver’s (Oliver,
1991) prediction about legal coercion engendering less compliance than
institutionally diffuse norms, there appears to be an association between
internal adoption of institutional rules derived from regulation when such
rules aligned with field-level values.
### 4.3. Organizational Responses to Technological Pressures
Institutional theorists posit that emerging organizations can bolster their
own legitimacy by adopting the values, structures, and practices of
established organizations (Meyer and Rowan, 1977; DiMaggio and Powell, 1983).
That is, organizations can improve their chances of survival by mimicking what
incumbent organizations already do. Yet this idea is in tension with the
Silicon Valley notion that the most legitimate innovations are those that
”disrupt” existing ways of operating (Hogarth, 2017; Geiger, 2020). So-called
radical innovations are those that break from or are discontinuous with prior
scientific and engineering practices whereas incremental innovations are those
that build upon and extend the existing technological paradigm (Colombo et
al., 2015). In our interviews with entrepreneurs, their rhetoric typically
suggested they viewed deep learning models as constitutive of radical
innovation in that they distinguished deep learning models from other machine
learning techniques; however, their implementations of deep learning
models–typically via transfer learning–were, in contrast, fundamentally
incremental.
Despite the murkiness around definitions of AI more broadly, entrepreneurs
often held deep learning out as distinct, which sometimes manifested in how
entrepreneurs defined AI. Many entrepreneurs provided definitions that
contrasted algorithms’ capabilities with human capabilities [P6, P7, P14,
P16], that differentiated between general and narrow intelligence [P5, P6, P7,
P15, P23], or that described high-level processes that are applicable to any
AI model [P1, P2, P4,P7, P8, P15, P18, P19, P20, P21, P22, P17]. However,
several entrepreneurs implicated deep learning specifically in their
definitions [P2, P9, P13, P11], using deep learning as a threshold of ”real”
AI:
> The most concise answer I can give you is just deep learning. That is almost
> the new cutoff for AI, in my mind at least [P9].
Even for those that did not equate deep learning with AI, the ways
entrepreneurs discussed deep learning relative to other AI approaches suggest
that they consider such techniques separate or superior. For example, some
described using ”machine learning and deep learning [P14]” or explained that
their companies constrained their models to ”classifiers and regression [P3]”
instead of ”doing any deep learning or anything nutty like that [P11]” as
though deep learning models were not a subset of machine learning techniques.
In one case, we interviewed the chief technology officer of a company that
intended to develop an AI-enabled solution but that had not yet begun data
collection. Even in the absence of any empirical evidence to support his
conclusion or expertise in deep learning, he had preemptively concluded that
simple, linear techniques would be insufficient to achieve the high accuracy
he hoped to obtain with deep learning methods [P14]. Thus, entrepreneurs
distinguish the ”magic [P13]” of deep learning from other ”rudimentary data
science [P11]” techniques.
Yet in conflict with widespread framing of deep learning as ”magic” and
perhaps radical innovation, most entrepreneurs’ implementations of deep
learning constituted a more incremental form of development that draws on the
scientific products of researchers in the AI community. Many entrepreneurs
described relying heavily on transfer learning for their deep learning
applications. Transfer learning is technique where models that are initially
trained with massive datasets for one task can be adapted for related tasks
with much lower data requirements. The use of transfer learning can reduce the
computational and data costs associated with training models from scratch
while still affording entrepreneurs ”the accuracy that we feel we need from
that model [P9]”. Through the use of pretrained models initially developed by
AI researchers in academia or at large technology companies, AI startups
”build upon the state of the art, all the advancements that are being driven
by the Googles of the world [P13]”. In a paradoxical way, entrepreneurs’
discursive distinction of deep learning techniques complies with the
institutional pressures to seek rapid and disruptive innovation, even though
the pretrained models used for transfer learning coordinate practices of AI
entrepreneurs, possibly down to the specific pretrained models they employ.
Entrepreneurs’ rhetoric surrounding the distinction of deep learning, which
conforms with institutional expectations about the utility of disruptive
technology, acts to conceal their use of publicly available, incremental
technologies that while not fundamentally disruptive in a scientific or
technical sense, are nevertheless sufficient to meet AI startups’ needs.
Interestingly, the disconnect between entrepreneurs’ rhetoric and practices in
deep learning did not appear to be driven by attempts to appeal to external
stakeholders who would be unlikely to appreciate the difference between deep
learning and other machine learning techniques, but potentially to signal
status to other AI startups or industry peers or to bolster their sense of the
company’s legitimacy internally.
### 4.4. Organizational Responses to Normative Pressures
As an organizational field becomes more institutionalized, professionalization
is enacted through education, membership in professional bodies, and other
aspects of professional culture. These professional mechanisms can drive
organizations within that field to adopt similar norms and values, which
become embedded in their organizational practices (DiMaggio and Powell, 1983).
Recent scholarship on professional norms within the AI research community has
found that pervasive professional norms include efficiency, universality, and
impartiality (Scheuerman et al., 2021). In our interviews, we observed
instances where entrepreneurs articulated personal ethical values that were
either distinct from, or resistant to, professional norms.
The demands of operating within the fast moving technology industry constrain
the extent to which industry practitioners can fully realize ethical values
into substantive practices (Vakkuri et al., 2020; Metcalf and Moss, 2019).
Consistent with these findings, some entrepreneurs hoped to incorporate their
personal ethical values into their product or business model in the future,
but had yet to make much tangible progress towards those ideals [P1, P2, P10,
P13]. Yet, in other cases, entrepreneurs took a strong stance on ethical
issues and described how they built these values into their technology and
organizational cultures. For example, several entrepreneurs described how
their algorithms [P21, P23, P19, P6] or data practices arose from normative
beliefs about the ethics of privacy protection [P2, P9, P11, P14].
Racial bias also came up repeatedly in our interviews, but did not always
inform product or business decisions. In a handful of cases, startups’ AI
approaches had been explicitly designed to ensure that their algorithms would
perform equally well across demographic groups [P4, P21, P23]. Similarly, some
entrepreneurs had designed non-algorithmic elements of their products to
prevent racial bias [P17, P12]. In several cases, entrepreneurs’ motivations
for developing algorithms that perform well across racial groups were not only
based on personal value systems, but also based on the belief that fair
algorithms realized market value. For example, several entrepreneurs pointed
out that in order to serve international clients and diverse users, it was
important for AI-enabled products to be equitable. On the other hand, some
entrepreneurs were aware of the types of racial biases that can be reproduced
by AI algorithms [P6, P7, P9, P11, P22], but either thought that race was
irrelevant to their models [P6, P11] or that racial biases were only a
priority in high-stakes contexts such as healthcare and finance [P9, P11].
Although less common than algorithmic strategies, several entrepreneurs
touched on how they promoted racial and gender equity within their companies
[P12, P22]. Drawing from his own experience of racial marginalization, one
entrepreneur noted:
> A fundamental shift in power from straight white men to the rest of the
> world is really something that needs to happen. […] I want to be able to
> show people that look like me, that they can also use things, and they can
> also build something that’s great and can also help build those communities
> [P22].
Even outside of explicit interview questions about ethics or social impact,
entrepreneurs often espoused values related to democratization and expanded
access to technology [P1, P8, P17, P16, P19, P4, P2] (e.g., ”democratizing
access to data [P17]”). Entrepreneurs described wanting to provide financially
valuable expertise or insights to other businesses, especially small
businesses and startups [P1, P19], to provide needed services to emerging
economies [P16], or to empower users to take on tasks that are more typically
performed by specialized professionals [P4, P8, P16]. In line with these
values, entrepreneurs were critical of insider cultures, implicating ”old
boys’” networks [P6, P7, P22] or ”traditional male VC [P12]” in gatekeeping
behaviors related to client acquisition, external financing, or in other ways
that affected their businesses’ success. Entrepreneurs’ skepticism of
centralization and insider culture was also manifested through their choices
about funding. Whereas VC served as a stand-in for centralized power,
crowdfunding was viewed as consistent with the ideal of democratization since
crowd investors do not need to meet the same financial accreditation standards
required to invest in a VC fund.
Entrepreneurs’ normative values reflect a mix of the techno-libertarian
leanings of Silicon Valley that have been documented elsewhere (Metcalf and
Moss, 2019; Lenhard, 2021; Dahlberg, 2010; Hütten, 2019) as well as beliefs in
social equity and fairness. Sometimes these values conflicted, as is
highlighted by the tension entrepreneurs expressed around racial bias in AI;
they believe that all users should be treated equally, but under the same
resource-constrained system that encourages developers to ”move fast and break
things” (Vardi, 2018), they do not always prioritize development around that
belief.
## 5\. Discussion
Our current study adds to the growing literature on organizational challenges
to ethical AI by describing how broader inter-organizational and institutional
forces shape the practices of AI startups. In this section, we discuss both
the theory-based as well as pragmatic contributions of our research. This
discussion is structured along the same categorical lines of our findings,
discussing in turn financial, regulatory, technological, and normative
pressures.
### 5.1. Financial Pressures
A central tension recurred between entrepreneurs’ desire to preserve the
scientific integrity of their AI approaches and the demands of technology
entrepreneurship that often ran counter to this desire. As one entrepreneur
noted, ”the value in the technology that you use doesn’t necessarily even have
to come from the technology [P9]”. In contrast to purely scientific
enterprises, the import and meaning of novel technologies is not entirely
determined by scientific inventiveness or rigor, but is also constructed
within an economic, social, and cultural context.
The demands of external stakeholders with power to affect the financial
outcomes of AI startups exerted influence over the narratives entrepreneurs
constructed about the benefits of their technology. In response to
stakeholders’ expectations of “silver bullet [P21]” AI solutions,
entrepreneurs tended to adapt their external messaging accordingly, but they
did not necessarily alter their internal practices. In this way, entrepreneurs
engaged in a resistance strategy of concealment (Oliver, 1991), decoupling the
symbolic and homogeneous marketing tactics they adopted to accrue legitimacy
from business partners, from the substantive and often heterogeneous
approaches they employed internally.
That entrepreneurs placed a strong value on scientific integrity points
towards an ethical opportunity within the startup ecosystem. As several
entrepreneurs themselves pointed out, models with inequitable outcomes are
necessarily less valid since they do not generalize well. Moreover, they are
less able to realize business value since they cannot meet the expectations of
diverse clientele. Thus, entrepreneurs’ values of scientific legitimacy might
act as a ”value lever” (Shilton, 2013) through which principles of AI ethics
can be imported into AI startups. On the other hand, external stakeholders’
tendency to treat decontextualized accuracy metrics as a superficial indicator
for AI quality is suggestive of a risk for institutionalization of AI ethical
ideals. For example, the “80% rule” for establishing disparate impact, which
has often been imported into AI fairness research without regard for its
original legal nuance, may have already created an artificial standard within
the research community (Watkins et al., 2022). This metric as a target could
create further ethical risk if stakeholders in the AI startup ecosystem also
adopt it without considering its relevance and caveats within context. Our
observations around the use of ”AI” as a marketing ”buzz word,” reflect recent
concerns around ”AI as snake oil” (Kaltheuner, 2021), and the exploitation of
AI’s vague definition as a loose umbrella term. Thus, strategies that ensure
that AI ethics constitute more than an ethical Potemkin facade are needed
(Cihon et al., 2021).
Consistent with prior literature (Willoughby, 2008), entrepreneurs also
demonstrated more heterogeneity in financing strategies than the culturally
dominant VC-startup narrative would suggest. While some entrepreneurs
conformed with institutional pressures to pursue VC funding and found benefits
beyond financial capital in their partnerships with VCs, others actively
avoided VC funding. This opposition was sometimes based on philosophical
opposition to VC as antithetical to democratic ideals and other times informed
by entrepreneurs’ personal experiences of VCs driving startups away from sound
technological and business practices. Some evidence suggests that
entrepreneurs’ avoidance of VCs could harm their companies’ growth potential
(Baum and Silverman, 2004; Bertoni et al., 2011), but other evidence shows
that the benefits of VC do not always extend to profitability (Rosenbusch et
al., 2013). Moreover, even if VC does improve the financial outcomes of
companies on the whole, this financial benefit does not necessarily redound to
founders themselves since their stake in the company is significantly diluted
by VC investment (Florin, 2005). Thus, resistance to the institutional norm of
VC financing could be conceptualized as economically rational as well as in
line with entrepreneurs’ desire to retain control over their businesses since
VCs sometimes use their power to replace the founding team with professional
executives (Hellmann and Puri, 2002).
Entrepreneurs’ desire to match their financing strategies with their business
goals and normative values presents an opportunity for ethical practices. Even
if AI ethics interventions are seen as antithetical to profit goals as
demonstrated in (von Zahn et al., 2021), entrepreneurs may be able to preserve
AI ethical ideas by matching with investors who share these priorities,
especially if public scrutiny around the ethical implications of investor
strategies increases (International, 2021). It is important to note, however,
that the entrepreneurs who are able to exercise more discretion in terms of
when they seek funding and from whom they seek it are likely already
advantaged in the entrepreneurial ecosystem, as one of our entrepreneurs
himself noted: ”I know we have the luxury that we could decline money. I know
that that is a luxury [P6]”. Black and Latinx founders (Crunchbase, 2020) and
female founders (Teare, 2021) secure less financing than other founders, and
as a result are likely to have fewer options when attempting to find financing
partners who prioritize ethical objectives. Thus, selective matching between
entrepreneurs and investors could also further magnify inequality.
### 5.2. Regulatory Pressures
With respect to regulatory pressures, entrepreneurs typically endorsed privacy
regulations but expressed more frustration with FDA regulations. While privacy
regulations were perceived as aligning with the values of personal freedom and
autonomy, which have been documented in other research on technology sector
actors (Metcalf and Moss, 2019; Lenhard, 2021), FDA regulations were seen as a
barrier to innovation and entrepreneurial autonomy and a mechanism through
which industry insiders receive favor from other institutional actors. These
contrasting results support both theory and evidence that a mismatch between
an organizational field’s normative values and coercive regulatory pressures
will result in less meaningful internalization of policies (Oliver, 1991;
Scott, 2013). However, an alternative resource-based explanation is also
possible. Privacy regulations are likely to apply uniformly, but FDA
regulations are idiosyncratic, and therefore require more expertise and
financial resources to navigate. Regardless of the cause, as legislators
debate proposals to further regulate AI, they should take care to consider
what negative, second-order effects regulations might have on AI startups.
Greater engagement with AI entrepreneurs could improve both policy and its
adoption within startups since active participation from business owners has
been shown to increase regulatory compliance (Malesky and Taussig, 2017).
### 5.3. Technical Pressures
The resource constraints of startups also fed into our findings regarding the
use of deep learning amongst AI startups. Both entrepreneurs who did use deep
learning and those who did not tended to discuss the use of deep learning
techniques with a reverence not afforded to other algorithmic approaches. Yet
deep learning startups most often developed their technology on top of
preexisting pretrained models, especially those developed to perform natural
language processing and computer vision tasks. As with most scientific
advancements, applications developed through transfer learning are incremental
innovations, inextricably tied to established approaches developed by a
broader community of researchers and practitioners. That is, the use of deep
learning in most AI startups is not a radical departure from the dominant
machine learning paradigm, but an endorsement of it. This is not to say that
deep learning applications developed through transfer learning are not
valuable, creative, or innovative. Experts have implicated transfer learning
specifically in the acceleration of AI discovery (Ng, 2016).
The widespread use of pretrained models does raise ethical questions. Word and
image embeddings derived from models trained on human data often encode human-
like biases such as gender, racial, and other harmful stereotypes (Bolukbasi
et al., 2016; Steed and Caliskan, 2021). How entrepreneurs adapt pretrained
models to their applications may obviate transmission of harmful biases from
pretrained models to industry applications; however, some researchers have
suggested that the biases of pretrained models, if not mitigated for their
contextual application, could further propagate harms (Steed and Caliskan,
2021). Thus, our finding supports the call for research to better understand
not only the negative social impacts of models developed in academic contexts,
but also how these impacts are attenuated or magnified by their applications
in industry through transfer learning (Narayanan, 2021).
### 5.4. Normative Pressures
As we have already discussed, entrepreneurs’ beliefs played a significant role
in how they developed their technologies and their business practices. In some
cases, AI entrepreneurs espoused libertarian ideals, such as individual
autonomy and personal responsibility. Much of AI ethics research focuses on
establishing the fairness of model outputs or mitigating unfairness in model
predictions. In this way, AI ethics interventions often center on equity in
outcomes. In contrast, entrepreneurs expressed valuing democratization, which
emphasizes the importance of equality of access, rather than equity in
outcomes. This distinction points to further risks for translating the
technology and ideas developed in AI ethics research contexts into AI startups
or the technology industry on the whole. Institutional pressures deriving from
the technology industry will likely interact or conflict with the values
embedded in ethical AI interventions. Designers of ethical interventions
should consider the normative context in which they are intended to apply.
Otherwise, they could be dismissed as irrelevant by practitioners or be
employed in ways other than how they were designed, which itself constitutes
an ethical risk.
## 6\. Future Research
Our findings on the significant influence exerted by institutional pressures
on AI startups, and the variance in entrepreneurs’ decision-making around
compliance, avoidance, and resistance, open a number of potential research
pathways. First, as mentioned above, more research is needed to better
understand how the social impacts of AI models may be exacerbated through
transfer learning in industry settings. Second, identifying and interviewing
other stakeholders in this sector would allow us to analyze the interactive
dimension of these field-level dynamics, yielding data about how investors,
regulators, competitors, and customers participate in and contribute to
complex system dynamics of institutionalism in AI. Findings around the
alignment between regulatory pressure and normative pressure, further, suggest
that such alignments lead to better take-up within organizations, and so
collaboration with policy researchers could lead to the design of AI policy
better positioned to act as an effective guardrail against the harms of such
systems.
## 7\. Limitations
Our study design presents several limitations which may influence our
findings. First, a limitation of the interview instrument was its exploratory
nature. Due to the broad scope of our research questions, themes could not be
identified prior to the study, but rather were identified in our data as a set
of findings. As a result, we were unable to reach depth within particular
themes, nor did we find that we reached theoretical saturation for any
thematic category. Instead, the exploratory nature of this study identifies
pathways for future research opportunities. Second, our recruitment and
sampling strategies also present limitations. Our sample size was relatively
small, and cannot – and is not intended to be – generalizable to the larger
population of AI startup entrepreneurs. Within qualitative research, sample
size requirements are a subject of debate, and are always a reconciliation
between research interests and goals, access to participants, and maintaining
rigor. We used theoretical sampling, which is intended to ensure that there
are enough participants to surface ”a range of concepts and characteristics
that are deemed critical for emergent findings,” (Dworkin, 2012; Glaser and
Strauss, 1967), which we determined was achieved with our sample.
## 8\. Conclusion
On the whole, our research shows that although institutional forces do shape
AI startups’ beliefs and practices surrounding AI, they do not dictate them
entirely. As a result, while future interventions to support ethical AI should
be mindful of the organizational contexts for which they are intended, they
also should not assume that startup practitioners have no agency to act in the
service of ethical values. Even if the ethical practices adopted by startups
at their outset evolve over time in response to shifting market demands,
founders typically have a lasting influence on startups’ trajectories, even
after they leave the company (Sahaym et al., 2016). As a result, though
startups face more resource constraints than the more mature companies that
have been the focus on most applied AI ethics research, they also may be an
ideal stage for ethical interventions.
## 9\. Acknowledgments
We thank Ranjit Singh, Arvind Narayanan, and Alex Hanna for helpful feedback
on our manuscript and Pedro Gomes for guidance on our research questions. We
gratefully acknowledge financial support from the Schmidt DataX Fund at
Princeton University made possible through a major gift from the Schmidt
Futures Foundation.
## 10\. Bibliography
## References
* (1)
* Angwin et al. (2016) Julia Angwin, Jeff Larson, Surya Mattu, and Lauren Kirchner. 2016. Machine Bias: There’s software used across the country to predict future criminals. And it’s biased against blacks. _ProPublica_ (2016). https://www.propublica.org/article/machine-bias-risk-assessments-in-criminal-sentencing
* Baum and Silverman (2004) Joel AC Baum and Brian S Silverman. 2004. Picking winners or building them? Alliance, intellectual, and human capital as selection criteria in venture financing and performance of biotechnology startups. _Journal of Business Venturing_ 19, 3 (2004), 411–436.
* Bertoni et al. (2011) Fabio Bertoni, Massimo G Colombo, and Luca Grilli. 2011\. Venture capital financing and the growth of high-tech start-ups: Disentangling treatment from selection effects. _Research Policy_ 40, 7 (2011), 1028–1043.
* Bhatt et al. (2020) Umang Bhatt, Alice Xiang, Shubham Sharma, Adrian Weller, Ankur Taly, Yunhan Jia, Joydeep Ghosh, Ruchir Puri, José MF Moura, and Peter Eckersley. 2020. Explainable machine learning in deployment. In _Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency_. 648–657.
* Blaseg et al. (2021) Daniel Blaseg, Douglas Cumming, and Michael Koetter. 2021\. Equity crowdfunding: High-quality or low-quality entrepreneurs? _Entrepreneurship Theory and Practice_ 45, 3 (2021), 505–530.
* Bolukbasi et al. (2016) Tolga Bolukbasi, Kai-Wei Chang, James Y Zou, Venkatesh Saligrama, and Adam T Kalai. 2016. Man is to computer programmer as woman is to homemaker? Debiasing word embeddings. _Conference on Neural Information Processing Systems_ 29 (2016).
* Caliskan et al. (2017) Aylin Caliskan, Joanna J Bryson, and Arvind Narayanan. 2017\. Semantics derived automatically from language corpora contain human-like biases. _Science_ 356, 6334 (2017), 183–186.
* Cihon et al. (2021) Peter Cihon, Moritz J Kleinaltenkamp, Jonas Schuett, and Seth D Baum. 2021. AI certification: Advancing ethical practice by reducing information asymmetries. _IEEE Transactions on Technology and Society_ 2, 4 (2021), 200–209.
* Colombo et al. (2015) Massimo G Colombo, Chiara Franzoni, and Reinhilde Veugelers. 2015\. Going radical: producing and transferring disruptive innovation. _The Journal of Technology Transfer_ 40, 4 (2015), 663–669.
* Crunchbase (2020) Crunchbase. 2020\. Crunchbase diversity spotlight 2020: Funding to Black and Latinx founders. http://about.crunchbase.com/wp-content/uploads/2020/10/crunchbase_diversity_report_2020.pdf
* Dahlberg (2010) Lincoln Dahlberg. 2010\. Cyber-libertarianism 2.0: A discourse theory/critical political economy examination. _Cultural Politics_ 6, 3 (2010), 331–356.
* Dastin (2018) Jeffrey Dastin. 2018\. Amazon scraps secret AI recruiting tool that showed bias against women. _Reuters_ (2018). https://www.reuters.com/article/us-amazon-com-jobs-automation-insight/amazon-scraps-secret-ai-recruiting-tool-that-showed-bias-against-women-idUSKCN1MK08G
* DiMaggio and Powell (1983) Paul J DiMaggio and Walter W Powell. 1983. The iron cage revisited: Institutional isomorphism and collective rationality in organizational fields. _American Sociological Review_ (1983), 147–160.
* Dworkin (2012) Shari L Dworkin. 2012\. Sample size policy for qualitative studies using in-depth interviews. _Archives of Sexual Behavior_ 41, 6 (2012), 1319–1320.
* Fish and Stark (2021) Benjamin Fish and Luke Stark. 2021. Reflexive design for fairness and other human values in formal models. In _Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society_. 89–99.
* Fligstein (1997) Neil Fligstein. 1997\. Social skill and institutional theory. _American Behavioral Scientist_ 40, 4 (1997), 397–405.
* Florin (2005) Juan Florin. 2005\. Is venture capital worth it? Effects on firm performance and founder returns. _Journal of Business Venturing_ 20, 1 (2005), 113–135.
* Geiger (2020) Susi Geiger. 2020\. Silicon Valley, disruption, and the end of uncertainty. _Journal of Cultural Economy_ 13, 2 (2020), 169–184.
* Glaser and Strauss (1967) B Glaser and A Strauss. 1967. The discovery of grounded theory Aldine Publishing Company. _Hawthorne, New York_ (1967).
* Hellmann and Puri (2002) Thomas Hellmann and Manju Puri. 2002. Venture capital and the professionalization of start-up firms: Empirical evidence. _The Journal of Finance_ 57, 1 (2002), 169–197.
* Hogarth (2017) Stuart Hogarth. 2017\. Valley of the unicorns: Consumer genomics, venture capital and digital disruption. _New Genetics and Society_ 36, 3 (2017), 250–272.
* Holstein et al. (2019) Kenneth Holstein, Jennifer Wortman Vaughan, Hal Daumé III, Miro Dudik, and Hanna Wallach. 2019\. Improving fairness in machine learning systems: What do industry practitioners need?. In _Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems_. 1–16.
* Hong et al. (2020) Sungsoo Ray Hong, Jessica Hullman, and Enrico Bertini. 2020\. Human factors in model interpretability: Industry practices, challenges, and needs. _Proceedings of the ACM on Human-Computer Interaction_ 4, CSCW1 (2020), 1–26.
* Hopkins and Booth (2021) Aspen Hopkins and Serena Booth. 2021. Machine learning practices outside Big Tech: How resource constraints challenge responsible development. In _Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society_. 134–145.
* Hütten (2019) Moritz Hütten. 2019\. The soft spot of hard code: Blockchain technology, network governance and pitfalls of technological utopianism. _Global Networks_ 19, 3 (2019), 329–348.
* International (2021) Amnesty International. 2021\. Risky business: How leading venture capital firms ignore human rights when investing in technology. Amnesty International Ltd. https://www.amnestyusa.org/wp-content/uploads/2021/07/Risky-Business-How-Leading-VC-Firms-Ignore-Human-Rights-When-investing-in-Tech-Final-1.pdf
* Kaltheuner (2021) Frederike Kaltheuner. 2021\. AI snake oil, pseudoscience, and hype: An interview with Arvind Narayanan. In _Fake AI_. Meatspace Press, 18–39.
* Kapiszewski and Karcher (2021) Diana Kapiszewski and Sebastian Karcher. 2021. Transparency in practice in qualitative research. _Political Science & Politics_ 54, 2 (2021), 285–291.
* Karizat et al. (2021) Nadia Karizat, Dan Delmonaco, Motahhare Eslami, and Nazanin Andalibi. 2021. Algorithmic folk theories and identity: How TikTok users co-produce Knowledge of identity and engage in algorithmic resistance. _Proceedings of the ACM on Human-Computer Interaction_ 5, CSCW2 (2021), 1–44.
* Kaur et al. (2020) Harmanpreet Kaur, Harsha Nori, Samuel Jenkins, Rich Caruana, Hanna Wallach, and Jennifer Wortman Vaughan. 2020. Interpreting interpretability: Understanding data scientists’ use of interpretability tools for machine learning. In _Proceedings of the 2020 CHI conference on Human Factors in Computing Systems_. 1–14.
* Lenhard (2021) Johannes Lenhard. 2021\. _Work, Society, and the Ethical Self: Chimeras of Freedom in the Neoliberal Era_. Berghahn Books, Chapter Unicorn-makers working for freedom (and monopolies), 221.
* Madaio et al. (2021) Michael Madaio, Lisa Egede, Hariharan Subramonyam, Jennifer Wortman Vaughan, and Hanna Wallach. 2021. Assessing the Fairness of AI Systems: AI Practitioners’ Processes, Challenges, and Needs for Support. _arXiv preprint arXiv:2112.05675_ (2021).
* Madaio et al. (2020) Michael A Madaio, Luke Stark, Jennifer Wortman Vaughan, and Hanna Wallach. 2020. Co-designing checklists to understand organizational challenges and opportunities around fairness in AI. In _Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems_. 1–14.
* Malesky and Taussig (2017) Edmund Malesky and Markus Taussig. 2017. The danger of not listening to firms: Government responsiveness and the goal of regulatory compliance. _Academy of Management Journal_ 60, 5 (2017), 1741–1770.
* McDonald et al. (2019) Nora McDonald, Sarita Schoenebeck, and Andrea Forte. 2019\. Reliability and inter-rater reliability in qualitative research: Norms and guidelines for CSCW and HCI practice. _Proceedings of the ACM on Human-Computer Interaction_ 3, CSCW (2019), 1–23.
* Metcalf and Moss (2019) Jacob Metcalf and Emanuel Moss. 2019. Owning ethics: Corporate logics, Silicon Valley, and the institutionalization of ethics. _Social Research: An International Quarterly_ 86, 2 (2019), 449–476.
* Meyer and Rowan (1977) John W Meyer and Brian Rowan. 1977. Institutionalized organizations: Formal structure as myth and ceremony. _Amer. J. Sociology_ 83, 2 (1977), 340–363.
* Morley et al. (2020) Jessica Morley, Luciano Floridi, Libby Kinsey, and Anat Elhalal. 2020. From what to how: An initial review of publicly available AI ethics tools, methods and research to translate principles into practices. _Science and Engineering Ethics_ 26, 4 (2020), 2141–2168.
* Murphy et al. (2021) Alexandra K Murphy, Colin Jerolmack, and DeAnna Smith. 2021\. Ethnography, data transparency, and the information age. _Annual Review of Sociology_ 47 (2021), 41–61.
* Narayanan (2021) Arvind Narayanan. 2021\. The ethics of datasets: Moving forward requires stepping back. Keynote talk at the AAAI/ACM Conference on AI, Ethics, and Society.
* Ng (2016) Andrew Ng. 2016\. Nuts and bolts of building AI applications using Deep Learning. Keynote Talk at the Conference on Neural Information Processing Systems.
* Noble (2018) Safiya Umoja Noble. 2018\. _Algorithms of oppression_. New York University Press.
* Obermeyer et al. (2019) Ziad Obermeyer, Brian Powers, Christine Vogeli, and Sendhil Mullainathan. 2019. Dissecting racial bias in an algorithm used to manage the health of populations. _Science_ 366, 6464 (2019), 447–453.
* Olan et al. (2022) Femi Olan, Emmanuel Ogiemwonyi Arakpogun, Uchitha Jayawickrama, Jana Suklan, and Shaofeng Liu. 2022. Sustainable supply chain finance and supply networks: The role of artificial intelligence. _IEEE Transactions on Engineering Management_ (2022).
* Oliver (1991) Christine Oliver. 1991\. Strategic responses to institutional processes. _Academy of Management Review_ 16, 1 (1991), 145–179.
* Palinkas et al. (2015) Lawrence A Palinkas, Sarah M Horwitz, Carla A Green, Jennifer P Wisdom, Naihua Duan, and Kimberly Hoagwood. 2015. Purposeful sampling for qualitative data collection and analysis in mixed method implementation research. _Administration and Policy in Mental Health and Mental Health Services rRsearch_ 42, 5 (2015), 533–544.
* Pratt et al. (2020) Michael G Pratt, Sarah Kaplan, and Richard Whittington. 2020\. The tumult over transparency: Decoupling transparency from replication in establishing trustworthy qualitative research. _Administrative Science Quarterly_ 65, 1 (2020), 1–19.
* Rakova et al. (2021) Bogdana Rakova, Jingying Yang, Henriette Cramer, and Rumman Chowdhury. 2021. Where responsible AI meets reality: Practitioner perspectives on enablers for shifting organizational practices. _Proceedings of the ACM on Human-Computer Interaction_ 5, CSCW1 (2021), 1–23.
* Reinhart (2016) Martin Reinhart. 2016\. Reproducibility in the Social Sciences. _Reproducibility: Principles, Problems, Practices, and Prospects_ 407, 423 (2016), 419.
* Rekabsaz and Schedl (2020) Navid Rekabsaz and Markus Schedl. 2020. Do neural ranking models intensify gender bias?. In _Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval_. 2065–2068.
* Rhoads (2020) Robert A Rhoads. 2020\. “Whales tales” on the run: Anonymizing ethnographic data in an age of openness. _Cultural Studies and Critical Methodologies_ 20, 5 (2020), 402–413.
* Rosenbusch et al. (2013) Nina Rosenbusch, Jan Brinckmann, and Verena Müller. 2013\. Does acquiring venture capital pay off for the funded firms? A meta-analysis on the relationship between venture capital investment and funded firm financial performance. _Journal of Business Venturing_ 28, 3 (2013), 335–353.
* Sahaym et al. (2016) Arvin Sahaym, Michael D Howard, Sandip Basu, and Warren Boeker. 2016. The parent’s legacy: Firm founders and technological choice. _Journal of Business Research_ 69, 8 (2016), 2624–2633.
* Salancik and Pfeffer (1978) Gerald R Salancik and Jeffrey Pfeffer. 1978. A social information processing approach to job attitudes and task design. _Administrative Science Quarterly_ (1978), 224–253.
* Saldana (2021) J Saldana. 2021\. _The Coding Manual for Qualitative Research, 4th Edition_. SAGE Publications Ltd.
* Scheuerman et al. (2021) Morgan Klaus Scheuerman, Alex Hanna, and Emily Denton. 2021\. Do datasets have politics? Disciplinary values in computer vision dataset development. _Proceedings of the ACM on Human-Computer Interaction_ 5, CSCW2 (2021), 1–37.
* Scott (2013) W Richard Scott. 2013\. _Institutions and organizations: Ideas, interests, and identities_. Sage publications.
* Seale (1999) Clive Seale. 1999\. Quality in qualitative research. _Qualitative Inquiry_ 5, 4 (1999), 465–478.
* Selbst (2021) Andrew D Selbst. 2021\. An institutional view Of algorithmic impact assessments. (2021).
* Shilton (2013) Katie Shilton. 2013\. Values levers: Building ethics into design. _Science, Technology, & Human Values_ 38, 3 (2013), 374–397.
* Steed and Caliskan (2021) Ryan Steed and Aylin Caliskan. 2021. Image representations learned with unsupervised pre-training contain human-like biases. In _Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency_. 701–713.
* Suchman (1995) Mark C Suchman. 1995\. Managing legitimacy: Strategic and institutional approaches. _Academy of Management Review_ 20, 3 (1995), 571–610.
* Suddaby and Greenwood (2005) Roy Suddaby and Royston Greenwood. 2005. Rhetorical strategies of legitimacy. _Administrative Science Quarterly_ 50, 1 (2005), 35–67.
* Sweeney (2013) Latanya Sweeney. 2013\. Discrimination in Online Ad Delivery: Google Ads, Black Names and White Names, Racial Discrimination, and Click Advertising. _Queue_ 11, 3 (mar 2013), 10–29. https://doi.org/10.1145/2460276.2460278
* Tavory and Timmermans (2014) Iddo Tavory and Stefan Timmermans. 2014. _Abductive analysis: Theorizing qualitative research_. University of Chicago Press.
* Taylor (1919) Frederick Winslow Taylor. 1919\. _The principles of scientific management_. Harper & brothers.
* Teare (2021) Gené Teare. 2021\. Sole female founders raised $1B less in 2020 Despite record venture funding Surge in the US. _Crunchbase_ (2021). https://news.crunchbase.com/news/sole-female-founders-raised-1b-less-in-2020-despite-record-venture-funding-surge-in-the-us/
* Timmermans and Tavory (2012) Stefan Timmermans and Iddo Tavory. 2012. Theory construction in qualitative research: From grounded theory to abductive analysis. _Sociological Theory_ 30, 3 (2012), 167–186.
* Tricot (2021) Roland Tricot. 2021\. Venture capital investments in artificial intelligence. 319 (2021). https://doi.org/https://doi.org/10.1787/f97beae7-en
* Vakkuri et al. (2020) Ville Vakkuri, Kai-Kristian Kemell, Marianna Jantunen, and Pekka Abrahamsson. 2020. “This is Just a prototype”: How ethics Are ignored in software startup-like environments. In _International Conference on Agile Software Development_. Springer, Cham, 195–210.
* Van Wijk et al. (2019) Jakomijn Van Wijk, Charlene Zietsma, Silvia Dorado, Frank GA De Bakker, and Ignasi Marti. 2019\. Social innovation: Integrating micro, meso, and macro level insights from institutional theory. _Business & Society_ 58, 5 (2019), 887–918.
* Vardi (2018) Moshe Y Vardi. 2018\. Move fast and break things. , 7–7 pages.
* von Zahn et al. (2021) Moritz von Zahn, Stefan Feuerriegel, and Niklas Kuehl. 2021\. The cost of fairness in AI: Evidence from e-commerce. _Business & Information Systems Engineering_ (2021), 1–14.
* Watkins and Anderson (2019) Elizabeth Anne Watkins and CW Anderson. 2019. Managing journalistic innovation and source security in the age of the weaponized Internet. In _Media Management and Digital Transformation_. Routledge, 119–131.
* Watkins et al. (2022) Elizabeth Anne Watkins, Michael McKenna, and Jiahao Chen. 2022\. The four-fifths rule is not disparate impact: A woeful tale of epistemic trespassing in algorithmic fairness. _arXiv preprint arXiv:2202.09519_ (2022).
* Weber (1978) Max Weber. 1978\. _Economy and society: An outline of interpretive sociology_. Vol. 2. University of California press.
* Willoughby (2008) Kelvin W Willoughby. 2008\. How do entrepreneurial technology firms really get financed, and what difference does it make? _International Journal of Innovation and Technology Management_ 5, 01 (2008), 1–28.
* Zucker (1977) Lynne G Zucker. 1977\. The role of institutionalization in cultural persistence. _American Sociological Review_ (1977), 726–743.
## Appendix A Interview Instrument
The interview instrument we used to loosely structure our interview is below.
We note that the primary questions we asked participants from the ideals and
values section of our instrument were about social and ethical implications.
Typically, we asked follow up questions based on participants’ responses to
this main question rather than other questions in this instrument.
Background
Can you tell me a little bit about your company?
[if not already answered] What is the problem your company is trying to solve?
What is your role in the company?
Can you walk me through what you did at work on a specific day recently?
Product & AI
How do you define AI?
How does your company use/want to use machine learning, artificial
intelligence, or predictive analytics?
Why did you/your team decide ML/AI was the right approach?
What was your/your team’s experience in AI/ML before starting this company?
Funding
How is your business funded?
Can you describe your experience trying to secure funding?
[If funded]
How do you typically interact with your funders? What happens in these
interactions?
Have you discussed how your company uses AI with your funders?
What do you think your funders think about AI?
[If not funded]
Once you do secure funding, how do you anticipate you will interact with your
funders?
Have you discussed how your company uses AI with potential funders?
What do you think potential funders think about AI?
What is your company’s exit strategy?
Ideals & Values
What would you say are the core values of your company?
What do you think differentiates a successful startup from an unsuccessful
one?
Are there other companies that you think are good examples for your own
company to follow?
What role does AI/machine learning play in the technology industry as a whole?
Has your team ever discussed the ethical or social implications of the AI you
use in your product?
Demographics & Background
What is your title at your company?
What is your gender?
What is your race?
What is your age?
What is your educational background?
## Appendix B Participant Characteristics
Participants’ self-described demographic characteristics and company roles are
listed in Table 1. Note that some participants listed more than one race, and
many listed more than one role. The methods through which participants were
recruited are listed in 2. The breakdown of participants by industry are
available in Table 3
| n
---|---
Race |
White/Caucasian | 14
Asian | 2
South Asian/Indian | 5
Black/African American | 2
Middle Eastern | 1
|
Gender |
Male | 17
Female | 6
|
Education (Highest) |
High School | 1
Bachelor’s | 7
Master’s | 9
PhD | 4
MD (or equivalent) | 3
|
Role |
C-Suite (e.g., CEO, CTO) | 11
Founder/Co-Founder | 15
Division Head | 3
Other | 4
Table 1. Participant Demographics | n
---|---
1st degree contact of authors | 1
2nd degree contact of authors | 5
Slack groups | 4
In-person networking event | 4
University alumni message board | 2
Cold contact | 7
Table 2. Recruitment Methods | n
---|---
Healthcare | 7
Aviation | 1
Fitness & Wellness | 5
Business Intelligence & Analytics | 6
Social Planning | 1
Design & Engineering Services | 2
Agriculture | 1
Table 3. Participant Industry
## Appendix C Supporting Quotes
Scholars have hotly debated whether the goals of open science that have been
gaining traction in the quantitative sciences are also relevant for
qualitative research (Pratt et al., 2020; Reinhart, 2016; Kapiszewski and
Karcher, 2021; Seale, 1999; Rhoads, 2020; Murphy et al., 2021). Unlike
quantitative research, qualitative human subjects often participate in
research on the condition of anonymity, which precludes complete transparency.
Moreover, qualitative work is often premised on the idea that the
interpretation lens the researcher brings to research is itself a valuable
component of any qualitative scientific pursuit. Here, we attempt to create a
balance between these values by offering our own interpretation of our
findings in the body of the article and offering as much transparency as
possible without compromising our participants’ anonymity through supporting
quotes. We omit quotes from the supplement that that directly or indirectly
could identify participants. To reduce the possibility that participants could
be identified by patterns across their quotes, we do not provide a participant
identifier for each quote, but we include the participants whose quotes are
listed within each section. To align with the organization of the main
article, quotes are organized according to financial pressures, regulatory
pressures, and normative pressures and are in no particular order within
sections.
#### C.0.1. Financial Pressures
Quotes are derived from P1, P2, P5, P6, P7, P8, P9, P10, P11, P12, P13, P14,
P15, P17, P18, P20, P21.
* •
I think it’s because there’s a little bit of a sense of AI being a like
magical silver bullet type solution. AI is just like this loosely defined
thing that if you give to somebody, it could potentially make them more money
or give you better insights or something like that, that from a more public
perspective, as far as a company saying that they are an AI based company
providing a service that may not use AI at all as better investment and also
people, it shows or it signals that the solution could be more scalable than
it is in its current fashion.”
* •
We’re going to talk about our machine learning algorithms, because from a
marketing standpoint, it connotes this next generation high tech, God, it has
to be good.
* •
And I think part of the biggest issue, and this may not be unspoken, is that
everyone and their brother wants to have AI in their product, especially in
healthcare right now, because it’s the buzzword du jour. So basically, if
you’re in the physical space, you want two things: you want to say your
product has AI and you want to say your product has a robot.
* •
Because [AI is] the sexy thing for investors. And quite frankly, coming from
an engineering perspective of what I know of AI and what I know of machine
learning, I actually think a lot of it is overblown and a lot of things that
are called AI is not actually AI. It’s actually machine learning or a learning
algorithm that is kind of tweaked and people are bringing up AI just to say
they have it.
* •
Just including a small amount of AI gives you a marketing edge.
* •
Plus it’s a feedback loop, I guess. That’s what you see in industry, everyone
doing this. And getting good results by saying that, ”Look, we use AI.” It
almost seems like, ”Why wouldn’t you?” I guess, whether or not you have it.
While I did mention it as a pet peeve of mine it’s also, I guess,
understandable. Especially for people in startups, is a very competitive
space. So you’re trying to get every edge you can.
* •
It is similar thing as saying ”blockchain.” You’re well aware, but it’s a
super common thing with startups, that startups are doing. Just trying to
catch people’s attention with the hottest new tech.
* •
When I was doing research, it was just super focused on like, ”Here’s this new
thing that I did that’s novel. And it’s state of the art and it gets 0.001%
better accuracy than this other guy’s thing. It’s not reproducible, but it’s
AI, and it’s really cool.” So I don’t know. The field is super legitimate. How
do I say this? You end up getting a lot of people who are just trying to ride
that wave of legitimacy and not contribute anything substantial.
* •
I think the basics of it is – a large dataset, train a machine learning model,
you can predict many things. I think that’s permeated a lot of just public
understanding, scientific popular science. I think that basic equation
resonates with people, even if they don’t really care to understand okay
exactly how does a neural network work or what is an LSTM or something like
that. They might not care about those specifics but they probably see the
results in their daily lives. I think there’s been a lot of remarkable changes
in the last 5 or 10 years with real products that people are using AI,
improving them. And so, I think they can appreciate its ability, while not
necessarily caring to delve too deep on the technical side.
* •
But AI, machine learning, they’re buzzwords, they make us sound smart, like
we’ve got access to something because to them, they don’t know any of this
stuff […] INTERVIEWER: Who is that a buzzword for? Who’s impressed by the word
AI or machine learning? PARTICIPANT: It just puts us in a box of, and I’m out
here in Silicon Valley – so it’s the nerds, the data people, they’re in touch
with tech, they associate us with the tech industry
* •
When it comes to funders, I think AI is a buzzword that everybody likes
hearing on the VC side, big data, machine learning, those types of things. But
I think from a marketing perspective, we’re marketing to work with
collaborators and hospitals or potential customers, it’s really focusing on
the clinical impact, that’s the most important.
* •
At least, to me it’s pretty obvious that machine learning has transformed a
lot of different tech sectors. The larger scale ones certainly, from Google to
Facebook. But down to even more specialized sectors. So I think there’s this…
I think the feeling is that AI has the capability of transforming or paradigm
shifting different sectors, and if you get in on it early you can be part of
that wave. And I think it’s certainly been a very successful methodology in
many fields across many domains. So it has a lot of demonstrated success.
* •
[…] We have these hype cycles for AI throughout history of at least the last
80 years, 70, 80 years. And if you look back into the history of what we now
understand as artificial intelligence, we have these just incredible claims
what will be possible tomorrow or at least next year. And then we had these AI
winters and all of that again. And I think sometimes in the last years, the
last 10, 15 years, people started to recognize, ”Hey, it starts working.” And
then again, all these claims we came up with, we will have robot butlers at
home and automatically driving cars and all that. And people really jumped on
it. But I think it was the first time that some of the promises get fulfilled.
To what degree, is another topic, but that you could, as a non-technical
academic person involved, see that is something happening. So you have Siri on
your phone, or if you have this incredible Google voice assistant that is
completely AI driven. So when the people first recognized it and then said,
”Hey, if I can use something that’s smart for my business case, I will make a
gazillion dollars.” And everyone jumped on this early AI thing. And now
companies think, ”Hey, I must do something with AI.”
* •
Five years ago, if you had an AI startup, you’d just get stupid money without
any proof.
* •
Because hype is nothing logical. So you can see it if you really look at
startup financing, lots of the larger VCs moved away from funding AI
companies, or solidly AI companies. Now the hype is biotech, of course,
everyone wants to be the next Pfizer. And I think that’s how humans work.
* •
I guess the one dirty little truth is that I care a little about the fact that
it’s AI per se, but it gets a lot of resonance and interest when I use the
word AI as opposed to machine learning, as opposed to algorithm.
* •
Well, at a certain point, it certainly resonated with investors. Maybe we’re
at a certain point in society now where it’s almost overused, and so there’s a
certain backlash against just the general use of the word AI, but certainly
for a period there, no matter what business you’re in, in society, you had to
drop the word AI in order to seem relevant and to seem like you’re doing
something important. Ultimately, again, from a fundraising perspective, from a
customer perspective, so there’s different stages of adoption of different
technologies that generalize across a lot of different sectors of the economy.
AI or whatever it is, is relatively early in [my company’s domain], and so
now, we’re at kind of like what’s called like the early adopters’ phase, but I
think there’s an even more extreme form of that which is some [businesses that
could be clients] want to adopt. It seems like they have an incentive to adopt
AI for the sake of the fact that it’s AI, and that they can then now say that
they’re using AI
* •
It’s a check box. INTERVIEWER: What does that mean that it’s a checkbox?
PARTICIPANT: You have it or you don’t. It’s binary. They really don’t give a
rip of how good it is. They just want to say that you have it because it’s a
marketing buzz.
* •
I mean, [investors believe you have AI] because you say it and you can talk
about the algorithm. I think the vast majority don’t dig in because I think
the vast majority truly don’t understand it. I mean, there are some investors
that… if you want to maybe get to talk to them, there are some that focus just
on AI, and they have experts who know it inside out because they came from
that space. And they’re going to be savvy enough to know the difference. There
are a lot of investors who don’t. And so they’re trying to catch up with the
next big thing in tech and they’re just following whatever the buzz is.
* •
I want to say I have machine learning and AI because it makes me sound like
I’m on the cutting edge.
* •
I don’t think it’s a specific message. […] I mean, listen to CNBC, look at the
investors in startups. And I think if you scan the vast majority of startups,
anything related with tech is going to talk about their AI engine or their
machine learning engine. […] But the fact is at the end of the day, regardless
of what you’re doing, the end goal is that you’re meeting an unmet need, and
AI and machine learning is just a way that you’re getting there. So saying
that you’re using it in your product is the way that you solve the problem.
And right now, because it’s the buzz du jour and everyone wants to do it to
say, ”Yeah, we’re on the cutting edge.” So that’s why you’d see more of it. To
be fair, there are some investors that see it as, okay, this is cutting edge.
And with multiple evolutions of this, we are going to get to that point where
this starts to overtake humans in terms of their function and intelligence.
And that’ll happen at some point in the future. But I believe that with 95% of
the uses of AI and machine learning, it’s just a way to solve a particular
problem and meet a need. And it’s just a slightly better algorithm.
* •
I have my own personal definition of AI. I’ll tell you. I think it’s just the
marketing term for the ability to do massive amounts of equations in order to
make predictions. I don’t view it as actual artificial intelligence. I view it
as a marketing term in order to … The ability to use very clever algorithms to
do massive amounts of statistical calculations to make better predictions.
That’s what I view as artificial intelligence.
* •
I think when we talk to most of our customers, they’re quite aware of AI and
how this stuff works, because our target is a mostly technical audience. We
end up having this sort of conversation with mostly people from tech companies
who are likely to be our customers, and they’ve a fair understanding of AI and
how it could work. [..] Yeah, we rarely mention, the AI does it. We are AI
powered, that’s sort of understood in most cases.
* •
We speak about AI value but we don’t mention AI so much.
* •
Because most of my experience with that is in the B2B space, where everyone is
somewhat technically inclined, or at least people making the decisions are
familiar with the general typology of what’s out there. And they understand at
least and in broad details, about the benefits that AI can offer a business.
But in terms of users… Just giving an example of, when I tell new people that
I meet what my job is, what I do, they’re a lot of times, ”What’s that? I have
no idea what that is.” It’s definitely not as strong of a marketing tool
directly to users.
* •
So we’re not fundraising right now, still bootstrapping. So I don’t know if it
will help or perhaps even the opposite, be not helpful anymore. When we
fundraise, I think with, to be honest, most customers don’t recognize it.
* •
If you have something that you say is a predictor and your name is something
.ai, you lose a lot of credibility when your predictors are not quality. So it
was just this growing pains. It happened really early on when there’s probably
like a couple hundred users. We adapted to it pretty quickly but it’s
something that the real … When I see it every day in my cloud customers, it’s
a real concern.
* •
I would say from the perspective of founders that would be [a signal] to
investors and clients as well, that’s definitely not at what we do simply
because this is something that the founder, [Founder Name] went to school for
and really has a passion about computer vision. And so he wanted a true
computer vision solution moving from the get-go. But yeah, even without the AI
moniker, [compared to] before we were considered [not an AI] company, before
we [developed our AI features…] there’s more attraction from investors if you
are labeled as a AI company.
* •
…one of the things I’ve learned about what makes this distinct is that when
you’re fundraising, or if you’re even thinking about fundraising for something
even remotely related to AI/ML and the market potential is as big as it is and
you’re at a time right now where the liquidity is really, really high in the
market, they find you. [… ] It’s a very different kind of power dynamic.
* •
there oftentimes questions [from investors…] it’s just like, ”Okay, how does
your AI work?” So then we have to describe the backend processes. We have
questions about where does the data come from? So that’s an often question. I
think those are the two most common questions.
* •
INTERVIEWER: So you said that you hate that AI as in your name. Then why is it
in your name? PARTICIPANT: Pretty much because we want to ride the hype cycle
too. So let’s be honest here, of course. And I don’t see that negative as I
might’ve sounded. It helps, labels help people.
* •
[…] there are lots of already set up AI frameworks that you can just connect
to and apply to your own products and then you have an AI product. Or other
cases that I see personally most often testing different tools as a marketer
is that lots of tools are automating something, and then they say, ”Oh, I have
an AI.”
* •
I actually wouldn’t classify any of it [as AI]. Anything I see today I think
artificial intelligence is not applied because none of the algorithms are
self-aware. So I actually think that the whole… You see this a lot in tech.
I’ll use autonomous driving as an example, right now where it’s like, ”Oh, I
got full autonomous driving.” Bullshit. That’s two decades away. I mean,
anyone who really can look at those algorithms and see what’s happening
without having some sort of sensors in the road, we’re so far away from that
and all these edge cases to get there. I mean, same thing with AI. People hear
of AI and they think of robots or the character in the Marvel movies who’s
actually a fully humanoid, thinking being. No. And I could be wrong on this
because it’s not my area of expertise, but I just feel that there’s so much…
Right now it’s one of those buzzwords that everybody’s jumping on to say they
have it, but very few actually do. In reality, what it is are better
algorithms to figure things out. And to be fair, I mean, there are some things
that are really good, like the voice recognition. And if you think about the
AI like with Siri and Amazon’s Alexa and the amount of language processing
that’s happening to pull things out, I mean, that’s very impressive. But at
the end of the day, they’re not self-aware. Yet, I hope. I mean, I don’t know.
But at the end of the day, it’s just a really good algorithm. And so I think a
lot of this people are glomming on to that futuristic view of it. And it’s
that next big, futuristic thing that they can do. And I think it ranges from,
oh yeah, we’re using machine learning and AI. On the one hand it could be just
it’s an adaptive algorithm for something fairly simple that’s looking at a
relationship with two variables all the way up to now we actually have a whole
platform like Siri and it’s a different thing. And that entire continuum
contains AI. So everyone wants to say, ”Yep, I got AI. I’m just like Apple or
I’m just like Amazon.”
* •
I’m so bad with names, but if you look into the acquisitions of Salesforce
from the last three years, and I think it was 12 companies, 11 claim to do AI.
And Salesforce bought them and with big marketing, ”Hey, we bought another AI
company to do our Einstein platform,” I think it’s called, ”To make it better,
smarter, faster.” And after half a year, if you looked into it, they just
discontinued all the companies because you find a little press release or if
you know someone who works at Salesforce in San Francisco, and once again,
they couldn’t do what they claimed it could do. So I know that it’s very
episodical. INTERVIEWER: Yeah, but it sounds like then from your experience, a
lot of the companies that are claiming to do AI are telling their customers,
they can do AI, ultimately those solutions are failing. Is that accurate?
PARTICIPANT: I don’t know if they are failing, but at least they’re not
succeeding with AI technology I’d say.
* •
[…] Artificial general intelligence. They think of something that actually can
think like a human being can think, but an AI model doesn’t actually think
right? It’s just, it can maybe make you, fool you into thinking one day, but
it’s actually not really intelligence. It could do massive amounts of
calculations and use that to make predictions. It’s not the same thing as
actually, intelligence. You, as a human being, I could suddenly tell you
something completely unrelated and you could apply what you learned to figure
that out. Is there an AI model on the planet that can really do that? No.
Could you walk and run, and suddenly I teach you about, tell you a little bit
about what a movie is, and watch a movie, and you’d make comments on a movie?
That’s real intelligence. A human being has intelligence. What’s amazing about
the human brain isn’t so much that it can do things, one thing at a time. It’s
the fact that it can take things it’s learned in those things and apply them
to completely different situations and come up with its own ideas. It’s not
like an AI suddenly is going to come up with a brand new statistical [domain
of application] model for you. That requires, still, human intelligence.
That’s what I view as true intelligence. That idea that creativity, that an
idea suddenly pops in your head and then you can implement it and come up with
something new that does not exist. […] To me, that’s real intelligence.
* •
I tend never to use the word artificial intelligence. Internally I think we
almost exclusively use the word machine learning. I think the FDA now kind of
labels some of it as AI so I use it a little bit more now publicly but there
was a time where I probably only used the word machine learning because AI
feels so amorphous.
* •
I think absolutely, when anybody hears the term artificial intelligence,
they’re thinking of HAL from 2001. They’re thinking of all the science fiction
novels they’ve ever read, or all the movies they’ve ever watched, or the
Terminator from Terminator 2. I think they’re thinking of something that
mimics human behavior, that has a consciousness. INTERVIEWER: Some people
would think that evoking something that, or using a term that evokes The
Terminator would be a bad thing, would not be a good marketing tactic.
PARTICIPANT: I think that’s what they think. It’s like it can actually replace
human intelligence. But they can’t. I do not see any AI model that’s even
close to human intelligence right now. Even a child’s intelligence. Even a
child’s intelligence, I do not see it.
* •
So we’ve been around [for several years]. I think there was some pitch
competitions in the beginning when we were a small startup and I remember
there was a few other companies that were pitching and a few of them were
using the term AI and I remember just listening and realizing I don’t think
there’s any AI at all happening. I can’t quite remember the application but I
remember being kind of annoyed that they got to use the hype term whereas I
didn’t really think there was any of that happening. I think that’s probably
less so now. I think a lot of companies really are using machine learning more
than they were five years ago because if you have a lot of data, that’s the
right thing to do.
* •
So I definitely think that in the same vein as like crypto and blockchain, as
buzzwords, as companies will spin up the idea of providing a service that can
be automated with artificial intelligence and what they wind up doing is doing
a bunch of manual work to make it seem like they can provide that service. But
once they go to scale, it doesn’t scale very well because you’re still doing a
lot of things manually and not doing a AI data driven approach first, because
from day one, we started with a, our proprietary model that we began training
to make sure that we weren’t like, ”All right, well, submit us [the raw data]
and then we’ll process it and then give it back to you.” That kind of thing,
because it’s just not a scalable solution.
* •
But in terms of why we decided to include a bunch of real AI, deep learning,
all that kind of stuff, it’s the only solution that will enable our grander
vision.
* •
Personally, it’s a pet peeve of mine, that some companies will call their
solution AI. And it’s like, ”Okay. You used a random forest. Congratulations.”
But no. There’s that. You can get value. The value in the technology that you
use doesn’t necessarily even have to come from the technology. Just saying you
use AI. In the same way that if you’re interviewing, saying that there’s like,
”Oh yeah. I know like ML, I know these ML frameworks, whatever.” Even if you
don’t really. It’s a huge bonus. I think the value we deliver is like, ”We do
use AI/ML technologies. Here is exactly where we use them. And then here’s how
it’s making your life easier.” And being able to actually substantiate that
with results from our application, I think it’s more than just a marketing
edge and it substantiates the claim.
* •
On [a recent date], I did a pitch event. It was a virtual pitch event with
about five or six investors and an audience as well. It was a three-minute
pitch. I ran through a slide deck, talked about the core technologies and the
products that we’re building. After that, there’s was a two-minute QA where
the investors asked me different questions. So from that, four out of five of
the investors reached out to me to say, hey, I’m interested in what you’re
doing. So I said, hey, thanks for your interest. I’m not currently raising
funds, but I’m more than happy to keep you informed about the progress we’re
doing. Here’s some materials you can read over. We’ll reach out to you every
month with our progress.
* •
The reason we use the word is just it sounds cool and people like that in the
marketplace. Like, ”Oh, you’re going to do AI on my data.” I’m like, ”Well
we’re going to do machine learning on your data.” I don’t really have the need
to do AI where I would with these kind of things, right? It’s not that
complex. I’m sure someone could. Like if you’re trying to say, ”Hey, I’m
looking at this data and maybe can I impute whether you have an illness or
something?” That probably would take something a little bit more nuanced and
AI-centric. But that’s not what we’re trying to do.
* •
So our overall strategy is to take as little investment as possible. And
there’s two sides to that. One is to keep the burn as low as possible, two is
to get to revenue as quickly as possible. And we began with seed investors. We
actually did [a University angels funding event], that was some of our first
money in. We had great experience with a crowdfunding platform. Easiest
fundraising I’ve ever done in my life. We raised a [an approximate sum] in
$10,000 checks in one day. One day, I couldn’t believe it. And it made me look
at these things differently. It allows you to control your destiny a lot more.
* •
There’s angel investors. There’s VCs. It depends. I can’t pick the kinds of
investors that come to these [pitch] events. I only know the panel maybe after
I sign up, but it’s a good mix. It’s fairly classified angel investors,
accredited angel investors. Only those accredited angel investors can actually
invest in startups. There’s a key distinction. Not everybody can give you
money. I have to say no to a lot of people, actually, because we don’t have
enough of a net worth to do business. But if they’re accredited investors or
venture capitalists, we’ll keep them on our Rolodex.
* •
I always try to speak to the dumbest person in the room. Based on the
questions, I’ll go into the level of detail that’s appropriate. I never like
to start up here because you just lose the audience immediately. The biggest
barrier to data science is being able to tell the story. I think a lot of data
scientists really are not good at that. So if you have really technical
founders, they don’t know how to relate their product to laymen, and so I
always try to be cognizant of the fact that I’m talking to people that are not
industry experts. They’re experts at determining value propositions, and so I
need to be able to accurately say what my value proposition is.
* •
So we’re a venture backed company, so we’ve raised a Series A in the past, and
as a company we have to continue raising. So our funders are typically other
venture capitalists, we talk to some strategics, so other big companies in the
area that are really interested in what we’re doing, but kind of want to play
an observer role. So we talk to them and then we also try to make use of
government funding as well, so there’s some kind of programs available for
non-dilutive funding for grants. So we make use of those three avenues of
funding.
* •
I think we certainly want to make sure any VCs that we work with are well-
aligned with us and can provide added value. We’re not just looking for source
of funding but also some support and oftentimes VCs or funders, equity funders
will take a board seat, and so we want to make sure we work well with them. So
there certainly is a mutual vetting process any time that there’s a potential
relationship in the works. But in general, the VCs that we work with tend to
be in the healthcare space and be from backgrounds that could help us, they
have a lot of contacts or they know how to build these types of devices or how
I think through pricing models, commercial strategy, things like that.
* •
So we’re funded through investors. So [Company Name] has done a seed round,
raised about [an approximate sum of money]. And so we’re very lean. I have
done venture backed businesses. And I actually try and avoid that business
model because I think oftentimes, the VC’s goals are at cross purposes with
building a long term successful business.
* •
And [at a previous startup] the VCs wanted to hype things up, get a lot of
press, make a splash, so they could raise the next round at a higher valuation
and look good to their LPs, which was actually contrary to what we needed to
do for the slow growth to build the business. In [current startup name], our
goal is to get to revenue and cashflow positive as quickly as possible. In
many ways, VCs, they won’t say this directly, but as an entrepreneur, once you
have revenue, it’s problematic from a fundraising standpoint. Because before
you have revenue, you’re all about promise and potential. Once you have
revenue, there are metrics. And the question becomes, why aren’t you growing
faster? And it’s very rare that something comes along with the true hockey
stick growth that VCs are looking for. So that actually puts you in this mode
where you want to go on the hype, as long as you can, put off revenue as far
as you can, which puts the entrepreneur in a defensive position, because the
only option then is to raise VC money for the next round, and of course, the
VCs want you to spend more and more, because that’s what their metrics to
their LPs look like. So I found that for many, many businesses, it’s not the
right model. There certainly are cases where if you talk to Peter Wendell at
Sierra, he’ll say, ”I sell rocket fuel.” If you’re not going to Mars, you
don’t need rocket fuel. There are actually very few businesses where rocket
fuel is the right thing. So I’m mindful of that, and that’s something I’ve
learned. INTERVIEWER: It sounds like that can create a vicious cycle for the
entrepreneur in terms of interrupting good product development. PARTICIPANT:
Totally. And the burn gets higher and higher, your runway gets shorter and
shorter, the expectations diverge from reality faster and faster. Yet the
entrepreneur is in a situation where, I’d feel this very viscerally where I’d
been one thing to my customers, they care about what I’m doing today. And my
investors would care about where’s this going to be in five years? And as that
gap got bigger and bigger, that’s a huge source of stress.
* •
VCs are doing just fine these days. And it’s also like, where people are in
their careers and their experience levels, there will always be a pipeline of
people coming fresh out of college early in their career, where they want that
rocket fuel. And they don’t realize that it’s only going to work out for 5
percent of them. And as you get later in your career, you realize there’s a
lot more options than VCs. There’s other ways to build a business. Whereas
when you start out, that seems like the, maybe it’s the way we’re taught, but
that seems to me like the only option. I didn’t even know there were other
options besides that, to build a business.
* •
I personally do not want to pursue funding just because then it will be like a
real job. Like I’ll be in debt to someone and someone else will influence what
we’re doing. If I didn’t work every day, that might be interesting, but it’s
not interesting right now. It’s like an endless stream of people who want me
to go present and talk to people but it’s like if I do that, what’s the
benefit to me? It could be useful to solve some problems, but I think it adds
a lot of complexity to where it’s like all of a sudden, ”Eh,” like I’m having
to spend several hours a day on this as opposed to if I’m working hard at [my
main job], this stuff goes on pause. If [my main job] stuff is chill, I can
work nights and weekends on this. That’s the other thing is the more
interesting more for me is to have an equity partner who is really good at
user experience and maybe one who is really good at data science as opposed to
someone who just has deep pockets, right? Then it’s all we’re doing this for
our passions and what our shared interests are as opposed to trying to get
rich. Maybe some day it will be something I could sell to someone if it has a
large user base and has a proven track record of profitability, that’s
probably interesting. But until then I don’t need somebody to give me a bunch
of money.
* •
The best investors and the most helpful ones have really been staying up to
date with what we’re doing, and we have a number of asks that we ask of our
investors and we have great investors who will follow through with that. And
it really depends on who they are, so if we’re looking for advisors in a
specific, like [one aspect of the industry domain], we know who to ask. If
we’re looking for advisors in brand building or marketing, there’s people to
ask those questions. Or how do we think about press, something like that. And
so, it’s really folks that are engaged and willing to help in the areas where
they’re able to.
* •
I think there’s not really a formula for it, it’s kind of a feel of would I
want this person on the board and likewise on their end, it’s do I have enough
belief in this company that I think they’re really going to succeed and I want
to work with these founders or work with the management team.
* •
We did it pretty simple, we put up a website, said what we have as a product
and looked who signed up and then went from there. And yeah, we had both
pretty well-paying jobs before, so we didn’t have to charge a lot, we wanted
just to get iteration speed and did that. And then we somehow noticed, ”Hey,
we now have X customers and the company is running itself more or less.” And
yeah, so we stumbled into it.
* •
So, in our case, we raised pre-seed. So not yet to seed. I think that’s still
something we’re working on. So for the pre-seed, we applied for an accelerator
program. We actually applied for a few. And then luckily we had a choice at
the end, so that was great. But yeah, we basically just applied for a few
accelerator programs, because we felt like we wanted to go through a program
to get mentors on top of just the funding for the early stage, and this is
what they did. That’s kind of was our focus.
* •
And because I think Silicon Valley has this idea of, you’ve got to grow like a
rocket ship speed and huge margins.
* •
So we had some great conversations; a few investors who actually were
interested in joining, but I didn’t feel like they were in line with our own
vision, so we didn’t agree to proceed there. INTERVIEWER: What about them made
you feel like they weren’t aligning with your vision? PARTICIPANT: Yeah, so
like one of the very prominent examples is that a VC’s focus specifically to
just do the seed round and then make sure that the startup doesn’t get
additional investment and just exits, so basically sells the company right
after, which is not what we want to do. I want to develop our technology to
have a [the fully developed product] for everyone, and then go for an IPO.
* •
But I also don’t think we would have built ML if we didn’t have VC money. And
I think you’re right. We would have probably gone to more stable, ”Let’s build
a good [core services business] with good software. And once that is on sure
footing, then let’s maybe some special projects.”
* •
This is something we actually see with our own customer base. We recently had
a fundraising round, and instead of going through VCs, we’ve generated … It’s
everybody who’s invested in us uses our product. It’s very simple. You come in
here, you see the recommendations we get. You try them out and they work, and
you’re like, ”Oh, wow. This is legit stuff. This is very, very legit stuff.”
You know?
* •
What maximizes efficiency, and what’s something that people are willing to pay
for? And oftentimes it’s efficiency. I’m so cynical. I don’t think it’s
quality.
* •
So the biggest thing about ML is like, oh, take a step back. Software is
pretty cheap to make and run. Other than your engineers, the most expensive
part. But for machine learning, you have to label a lot of data. You have to
label thousands of images for your training, for your validation. And then you
have to pay [specialists to label your data]. That’s what’s hard. You have to
pay people who are [specialists] otherwise then you get bad ML.
* •
So angels, we consider any individual that’s not part of an institutional
fund. And the strategic part of that, are really people who have subject
matter expertise in something having to do with our business. […]. And so, we
have angels that really have that diversity of expertise in those areas.
* •
INTERVIEWER: What do you think motivates the questions they ask you about your
modeling approach or data approaches? PARTICIPANT: Revenue generation
opportunities. INTERVIEWER: Can you tell me more about that? PARTICIPANT: I
would say first and foremost, investors are investing in a company because
they believe that it could grow and scale, and ultimately exit and provide
liquidity. So, the questions or conversations that we’re having from a
business operations, whatever standpoint, is all thinking about how are we
building the best business that we can, so that we can grow as quickly as we
can to create some sort of liquidity for investors? And so, I would say that’s
the basis of the questions.
* •
I would say both because institutional investors that want to invest in AI
companies that are very mindful of how they’re applying AI and doing the
thinking for the investors essentially being like, look we are aware of all of
the problems at large and upcoming regulation and all of the confusion around
it. And we want to stay ahead of it and educate people about it and do it in
as transparently as possible manner to make sure that people are comfortable
with the solution before it is deployed en mass.
* •
We’re VC backed, which means that we believe that we can IPO at some point. If
the IPO doesn’t happen, then there’s a number of strategic exit opportunities
that would make sense for this kind of company.
* •
We’ve had private investment and then we also have a crowd funding round that
we raised. I think we’re almost at [a sum less than 1M], anyway it’s on [a
publicly available site…]. But that has a lot of our publicly available
information as to share count a number of investors and all that kind of
stuff.
* •
So we went the route of equity crowd funding, so it’s a little bit different
than like Kickstarter or something like that. Yeah. And so one of the main
benefits is we’re like, ”Hey, if you invest in the company, then you own a
share account. And so there’s potential for you to have a return on that
investment.” And that’s the major selling point of doing equity crowd funding
like this. And was it really only possible because of newer federal laws that
allow that kind of funding. The main benefit was just a little bit more
transparency to individual people. And we learned a lot from it. There’s a lot
of people that we talk to that are business sales, engineering, whatever. And
they’re like, ”Okay, we’ll take this to our leadership and talk about doing
business with you, but at a personal level we would like to invest.”
* •
Yeah, I would say the downside of it is just that funding, it trickles in and
it’s less lump sum payments, but it really just depends. The main downside
that we face through this equity crowdfund funding is through the company that
|
# Reduced Dirac Equation And Lamb Shift As An Off-mass-shell
Effect In Quantum Electrodynamics
Ni Guang-jiong a,b<EMAIL_ADDRESS>a Department of Physics, Fudan University,
Shanghai, 200433, China
b Department of Physics, Portland State University, Portland, OR97207, U. S.
A. Xu Jianjun<EMAIL_ADDRESS>Department of Physics, Fudan University,
Shanghai, 200433, China Lou Senyuec,d<EMAIL_ADDRESS>c Department of
Physics, Shanghai Jiao Tong University, Shanghai, 200030, China
d Department of Physics, Ningbo University, Ningbo 315211, China
###### Abstract
Based on the precision experimental data of energy-level differences in
hydrogenlike atoms, especially the $1S-2S$ transition of hydrogen and
deuterium, the necessity of introducing a reduced Dirac equation with reduced
mass as the substitution of original electron mass is stressed. Based on new
cognition about the essence of special relativity, we provide a reasonable
argument for reduced Dirac equation to have two symmetries, the invariance
under the (newly defined) space-time inversion and that under the pure space
inversion, in a noninertial frame. By using reduced Dirac equation and within
the framework of quantum electrodynamics in covariant form, the Lamb shift can
be evaluated (at one-loop level) as the radiative correction on a bound
electron staying in an off-mass-shell state–a new approach eliminating the
infrared divergence. Hence the whole calculation, though with limited
accuracy, is simplified, getting rid of all divergences and free of ambiguity.
Keywords: Reduced Dirac Equation, Lamb shift, off-mass-shell
PACC: 0365, 1110G, 1220D
## I Introduction
As is well known, the Dirac equation for electron in a hydrogenlike atom is
usually treated as a one-body equation with the nucleus being an inert core
having infinite mass and exerting a potential $V(r)=-\frac{Z\alpha}{r}\ \
(\hbar=c=1)$ on the electron. Then the rigorous solution of energy levels
reads[1]:
$\displaystyle E_{nj}$ $\displaystyle=$ $\displaystyle m_{e}f(n,j)$ (1)
$\displaystyle f(n,j)$ $\displaystyle=$
$\displaystyle\left[1+\frac{(Z\alpha)^{2}}{(n-\beta)^{2}}\right]^{-\frac{1}{2}}$
(2) $\displaystyle\beta$ $\displaystyle=$ $\displaystyle
j+\frac{1}{2}-\sqrt{(j+\frac{1}{2})^{2}-(Z\alpha)^{2}}$ (3)
where $j$ is the total angular momentum. The expansion of $f(n,j)$ to the
power of $(Z\alpha)^{6}$ is given as 1
$\begin{array}[]{l}f(n,j)=1-\frac{(Z\alpha)^{2}}{2n^{2}}-\frac{(Z\alpha)^{4}}{2n^{3}}\left(\frac{1}{j+\frac{1}{2}}-\frac{3}{4n}\right)\\\
-\frac{(Z\alpha)^{6}}{8n^{3}}\left[\frac{1}{(j+\frac{1}{2})^{3}}+\frac{3}{n(j+\frac{1}{2})^{2}}+\frac{5}{2n^{3}}-\frac{6}{n^{2}(j+\frac{1}{2})}\right]+\cdots\end{array}$
(4)
Obviously, besides the rest energy of the electron given by the first term,
the second term has exactly the form of Bohr energy level except that the mass
$m_{e}$ must be replaced by the reduced mass
$\mu=\frac{m_{e}m_{N}}{m_{e}+m_{N}}\equiv\frac{m_{e}m_{N}}{M}$ (5)
with $m_{N}$ being the mass of the nucleus and $M=m_{e}+m_{N}$.
However, as discussed in Refs.1 and 2 , the concept of reduced mass in
relativistic quantum mechanics (RQM) is ambiguous to some extent. Beginning
from 1950’s, a number of authors have been devoting a great effort at the
level of two-body RQM and that of quantum electrodynamics (QED) to take
account of the recoil effect 3 ; 4 ; 1 , incorporating their results in a
compact form (to order of $\alpha^{4}$):
$E=M+\mu[f(n,j)-1]-\frac{{\mu}^{2}}{2M}[f(n,j)-1]^{2}+\frac{(Z\alpha)^{4}{\mu}^{3}}{2n^{3}m_{N}^{2}}\left[\frac{1}{j+\frac{1}{2}}-\frac{1}{l+\frac{1}{2}}\right](1-{\delta}_{l0})$
(6)
A comprehensive review on the theory of hydrogenlike atoms can be found in
Ref.27 (27). The aim of this paper is two-fold: First, based on the
experimental data of hydrogen $1S-2S$ transition frequency 5 and its isotope
shift of hydrogen and deuterium 6 , we stress the necessity of the
introduction of reduced mass $\mu$ (section II) before we are able to argue
the reasonableness of introducing a ”reduced Dirac equation” with $\mu$ as the
substitution of $m_{e}$ (section III). Second, based on above conception, we
will present a calculation of Lamb Shift (LS) as an off-mass-shell effect by
performing the evaluation of self-energy diagrams of electron (section IV) and
photon (section V) as well as the vertex function (section VI) at the one-loop
level of QED in covariant form. The new insight of our calculation is focused
on the regularization renormalization method (RRM). As initiated by J-F Yang 7
and elaborated in a series of papers (8a ; 8b ; 9 (9, 24, 25, 26) and
references therein), we can get rid of all ultra violet divergences in the
calculation of quantum field theory (QFT). Furthermore, in this paper, we will
be able to get rid of the annoying infrared divergence in the vertex function
by treating the electron moving off its mass-shell to certain extent which is
fixed through the evaluation of self-energy diagram or by the Virial theorem.
Based on above improvements, the one-loop calculation yields values of LS in a
simple but semi-quantitative way (section VII and VIII). Although the accuracy
is limited at one-loop level, we hope our approach could be served as a new
starting point for calculations at high-loop order to get accurate results at
a comparably low labor cost. The final section IX and Appendix will contain a
summary and discussion.
## II The $1S-2S$ Transition of Atomic Hydrogen and Deuterium
In the last decade, thanks to remarkable advances in high resolution laser
spectroscopy and optical frequency metrology, the $1S-2S$ two-photon
transition in atomic hydrogen $H$ (or deuterium $D$) with its natural
linewidth of only $1.3Hz$ had been measured to a very high precision. In 1997,
Udem et al.determined the $1S-2S$ interval of $H$ being 5
$f^{(H)}(1S-2S)=2466061413187.34(84)\ \ kHz$ (7)
Even four years earlier, Schmidt-Kalar et al. measured the isotope-shift of
the $1S-2S$ transition of $H$ and $D$ to an accuracy of $3.7\times 10^{-8}$6 ,
giving (as quoted in 10 (10)):
$\Delta f\equiv f^{(D)}(2S-1S)-f^{(H)}(2S-1S)=670994337(22)\ \ kHz$ (8)
(In 1998, Huber et al.measured a more accurate data 28 (28):
$670994334.64(15)\ kHz$). which is of the order of $10^{-4}$ in comparison
with Eq.(7). As pointed out in Ref.6 , this $671\ GHz$ isotope-shift can be
ascribed almost entirely to the different masses of proton ($p$) and deuteron
($d$). And the nuclear volume effects become important because the QED effects
cancel considerably in the isotope shift.
Here, we wish to emphasize that in the first approximation, both experimental
data (7) and (8) can be well accounted for by simply resorting to Eq.(1) with
$m_{e}$ replaced by the reduced mass
$\mu_{H}=\frac{m_{e}m_{p}}{m_{e}+m_{p}},\ \ \
\mu_{D}=\frac{m_{e}m_{d}}{m_{e}+m_{d}}$ (9)
for $H$ and $D$ respectively.
Indeed, adopting the following updated values 10 (10, 11, 12, 13)
$\displaystyle\alpha$ $\displaystyle=$ $\displaystyle(137.03599944)^{-1},\ \ \
{\alpha}^{2}=0.532513542\times 10^{-4}$ (10) $\displaystyle{\alpha}^{4}$
$\displaystyle=$ $\displaystyle 0.283570673\times 10^{-8},\ \ \
{\alpha}^{6}=0.151005223\times 10^{-12}$ (11) $\displaystyle m_{e}$
$\displaystyle=$ $\displaystyle 0.51099906\ \ MeV=1.2355897\times 10^{20}\ \
Hz$ (12) $\displaystyle R_{\infty}$ $\displaystyle=$
$\displaystyle\frac{1}{2}{\alpha}^{2}m_{e}=3.28984124\times 10^{15}\ \ Hz$
(13) $\displaystyle\frac{m_{p}}{m_{e}}$ $\displaystyle=$ $\displaystyle
1836.1526665$ (14)
and denoting
$\displaystyle\frac{m_{e}}{m_{p}}$ $\displaystyle=$ $\displaystyle
b_{H}=5.446170255\times 10^{-4},\ \ \ \frac{1}{1+b_{H}}=0.999455679$ (15)
$\displaystyle\frac{m_{e}}{m_{d}}$ $\displaystyle=$ $\displaystyle
b_{D}=2.724436319\times 10^{-4},\ \ \ \frac{1}{1+b_{D}}=0.99972763$ (16)
we can calculate the energy difference of $2S$ and $1S$ of $H$ through Eq.(1)
with $m_{e}$ replaced by $\mu_{H}$ (the superscript RDE refers to the reduced
Dirac equation)
$\displaystyle\Delta E^{RDE}_{H}(2S-1S)$ $\displaystyle=$
$\displaystyle\mu_{H}[f(2,1/2)-f(1,1/2)]$ (17) $\displaystyle=$
$\displaystyle\frac{m_{e}}{1+b_{H}}(1.996950159\times 10^{-5})$
$\displaystyle=$ $\displaystyle 1.2355897\times 10^{20}\times
0.999455679\times 1.996950159\times 10^{-5}$ $\displaystyle=$ $\displaystyle
2.466067984\times 10^{15}\ \ Hz$
which is only a bit larger than the experimental data Eq.(7) with accuracy
$3\times 10^{-6}$. However, a more stringent test of RDE should be the isotope
shift of $H$ and $D$. We have
$\frac{1}{1+b_{D}}-\frac{1}{1+b_{H}}=(b_{H}-b_{D})-(b^{2}_{H}-b^{2}_{D})+(b^{3}_{H}-b^{3}_{D})+\cdots=2.719511528\times
10^{-4}$ (18)
$\Delta E^{RDE}_{D-H}=(\mu_{D}-\mu_{H})[f(2,1/2)-f(1,1/2)]=6.7101527879\times
10^{11}\ \ Hz$ (19)
which has only a discrepancy larger than the experimental data, Eq.(8) by
$20.941\ MHz$ with accuracy $3\times 10^{-5}$. Of course, it is still not
satisfied in an analysis of high precision 6 . Let us resort to the Eq.(6),
where the third term does provide a further modification:
$\displaystyle-$
$\displaystyle\frac{1}{2}m_{e}[\frac{b_{D}}{(1+b_{D})^{3}}-\frac{b_{H}}{(1+b_{H})^{3}}]\\{[f(2,1/2)-1]^{2}-[f(1,1/2)-1]^{2}\\}$
(20) $\displaystyle=$
$\displaystyle\frac{1}{2}m_{e}[(b_{H}-b_{D})-3(b^{2}_{H}-b^{2}_{D})+\cdots](-6.646361554\times
10^{-10})=-11.176\ MHz$
which brings the discrepancy between the theory and experimental down to less
than $10\ MHz$.
Although the detail explanation for this discrepancy remains quite
complicated6 , the above comparison is enough to convince us that the
inevitable appearance of reduced mass in the RDE or Eq.(6) is by no means a
simple fortune. It must have a deep reason from a theoretical point of view.
Notice further that once the conditions $m_{e}\ll m_{p}$ and $m_{e}\ll m_{d}$
hold, the difference of spin between $p$ and $d$ seems not so important. So in
next section, we will strive to justify the reduced Dirac equation on a
reasonable basis. Of course, it is still an approximate one, but seems much
better than the original Dirac equation when dealing with hydrogenlike atoms.
## III Reduced Mass and Reduced Dirac Equation
Consider a system of two particles with rest masses $m_{1}$ and $m_{2}$. Their
coordinates in the center-of-mass (CM) system are ${\mathbf{r}}_{1}$ and
${\mathbf{r}}_{2}$ respectively, as shown in Fig.1. If there is a potential
$V(r)=V(|{\mathbf{r}_{1}-r_{2}}|)$ between them, two equations
$m_{1}\ddot{\mathbf{r}}_{1}=-\nabla_{r}V(r)$ and
$m_{2}\ddot{\mathbf{r}}_{2}=\nabla_{r}V(r)$ will reduce to one:
$\mu\frac{d^{2}{\mathbf{r}}}{dt^{2}}=-\nabla_{r}V(r),\ \ \
(\mu=\frac{m_{1}m_{2}}{m_{1}+m_{2}})$ (21)
At first sight, the definition of center-of -mass (CM) in classical mechanics
$m_{1}r_{1}=m_{2}r_{2}$ becomes doubtful in the theory of special relativity
(SR) because the mass is no longer a constant. But actually, we can still
introduce the coordinate of CM in the laboratory coordinate system (LCS) (with
${\mathbf{r}}^{\prime}_{1}$ and ${\mathbf{r}}^{\prime}_{2}$ being the
coordinates of $m_{1}$ and $m_{2}$):
$R=\frac{1}{M}(m_{1}{\mathbf{r}}^{\prime}_{1}+m_{2}{\mathbf{r}}^{\prime}_{2})=(X,Y,Z),\
\ \ (M=m_{1}+m_{2})$ (22)
and the relative coordinate of $m_{1}$ and $m_{2}$
(${\mathbf{r}}_{i}={\mathbf{r}}^{\prime}_{i}-{\mathbf{R}},\,i=1,2$) :
${\mathbf{r}}={\mathbf{r}}^{\prime}_{1}-{\mathbf{r}}^{\prime}_{2}={\mathbf{r}}_{1}-{\mathbf{r}}_{2}=(x,y,z)$
(23)
Here the motion of CM in the LCS is assumed to be slow and so
$\frac{\partial}{\partial
x^{\prime}_{1}}=\frac{m_{1}}{M}\frac{\partial}{\partial
X}+\frac{\partial}{\partial x},\ \ \ \frac{\partial}{\partial
x^{\prime}_{2}}=\frac{m_{2}}{M}\frac{\partial}{\partial
X}-\frac{\partial}{\partial x}$ (24)
Notice that the momentum $\mathbf{P}$ of CM and the relative momentum
${\mathbf{p}}_{r}$ becomes operator in quantum mechanics (QM) without explicit
dependence on mass:
${\mathbf{P}}=-i\hbar\nabla_{\mathbf{R}},\ \ \
{\mathbf{p}}_{r}=-i\hbar\nabla_{\mathbf{r}}$ (25)
Thus the momenta of $m_{1}$ and $m_{2}$ in laboratory coordinate system (LCS)
read:
${\mathbf{p}}^{\prime}_{1}=-i\hbar\nabla_{{\mathbf{r}}^{\prime}_{1}}=\frac{m_{1}}{M}{\mathbf{P}}+{\mathbf{p}}_{r},\
\ \
{\mathbf{p}}^{\prime}_{2}=-i\hbar\nabla_{{\mathbf{r}}^{\prime}_{2}}=\frac{m_{2}}{M}{\mathbf{P}}-{\mathbf{p}}_{r}$
(26)
Since the center-of-mass coordinate system (CMCS) is also an inertial frame
which can be transformed from the LCS via a linear Lorentz transformation, it
is defined by the condition that $\mathbf{P}=0$ in CMCS. In other words, CMCS
is defined by the condition ${\mathbf{p}}_{1}+{\mathbf{p}}_{2}=0$, or from
Eq.(26):
${\mathbf{p}}^{\prime}_{1}=-i\hbar\nabla_{{\mathbf{r}}^{\prime}_{1}}={\mathbf{p}}_{r},\
\ \
{\mathbf{p}}^{\prime}_{2}=-i\hbar\nabla_{{\mathbf{r}}^{\prime}_{2}}=-{\mathbf{p}}_{r}$
(27)
Evidently, the above definition of CMCS remains valid in the realm of
relativistic QM (RQM) even the exact meaning of CM seems obscure to some
extent due to the conjugation relation of $a$ particle’s position and its
momentum, see Fig.1.
Now, from Eq.(27), it is natural to replace $\mathbf{p}_{1}$ and
$\mathbf{p}_{2}$ by ${\mathbf{p}}_{r}$, reducing the two-particle degrees of
freedom to one. In the meantime, the origin of CMCS is discarded, it is
substituted by the position of $m_{2}$
(${\mathbf{r}}={\mathbf{r}}_{1}-{\mathbf{r}}_{2}$). We will call the system
associated with $\mathbf{r}$ the relative motion coordinate system (RMCS),
which should be viewed as a deformation of CMCS. The transformation from CMCS
to RMCS is by no means a linear one. Rather, the origin of RMCS ($m_{2}$) is
moving non-uniformly in the CMCS. Therefore, while rest masses $m_{1}$ and
$m_{2}$ remain the same in both LCS and CMCS, they reduce to one mass
$\mu=\frac{m_{1}m_{2}}{m_{1}+m_{2}}$ for $m_{1}$ in RMCS (or for $m_{2}$ if
$m_{1}$ is chosen as the origin of RMCS).
Let us express the total energy $E=E_{1}+E_{2}$ in CMCS in terms of $p_{r}$
and reduced mass $\mu$ ($\mu=\frac{m_{1}m_{2}}{M},\ M=m_{1}+m_{2}$), where
$E_{1}=\sqrt{m_{1}^{2}+p_{1}^{2}}=\sqrt{m_{1}^{2}+p_{r}^{2}},\ \ \
E_{2}=\sqrt{m_{2}^{2}+p_{2}^{2}}=\sqrt{m_{2}^{2}+p_{r}^{2}}$ (28)
Treating all $p_{1},p_{2}$ and $p_{r}$ being $c$-numbers, we have
$E^{2}=(E_{1}+E_{2})^{2}=M^{2}+\frac{M}{\mu}p_{r}^{2}+\frac{1}{4\mu^{2}}p_{r}^{4}(4-\frac{M}{\mu})+\cdots$
(29)
where the expansion in $p_{r}$ is kept to the order of $p_{r}^{4}$. Two
extreme cases will be considered separately:
A. $m_{2}\gg m_{1},\ \ \mu\lesssim m_{1},\ \ M\gg\mu$:
$\displaystyle E^{2}$ $\displaystyle=$ $\displaystyle
M^{2}\left[1+\frac{1}{\mu
M}p_{r}^{2}-\frac{1}{4M\mu^{3}}p_{r}^{4}(1-\frac{4\mu}{M})+\cdots\right]$
$\displaystyle E$ $\displaystyle=$ $\displaystyle M\left[1+\frac{1}{2\mu
M}p_{r}^{2}-\frac{1}{8M\mu^{3}}p_{r}^{4}(1-\frac{3\mu}{M})+\cdots\right]$ (30)
$\displaystyle=$ $\displaystyle
M+\frac{1}{2\mu}p_{r}^{2}-\frac{1}{8\mu^{3}}p_{r}^{4}+\cdots$
$\displaystyle\simeq$ $\displaystyle M-\mu+\sqrt{\mu^{2}+p_{r}^{2}}\simeq
m_{2}+(m_{1}-\mu)+\sqrt{\mu^{2}+p_{r}^{2}}$
$E^{\prime}\equiv E-m_{2}=(m_{1}-\mu)+\sqrt{\mu^{2}+p_{r}^{2}}$ (31)
B. $m_{1}=m_{2}=m,\ \ \mu=\frac{m}{2},\ \ M=2m=4\mu$ Then to the accuracy of
$p_{r}^{4}$, we have :
$\displaystyle E^{2}$ $\displaystyle=$ $\displaystyle
M^{2}+\frac{M}{\mu}p_{r}^{2}=4m^{2}+4p_{r}^{2}$ $\displaystyle E$
$\displaystyle=$ $\displaystyle
2m+\frac{1}{2\mu}p_{r}^{2}-\frac{1}{32\mu^{3}}p_{r}^{4}+\cdots$ $\displaystyle
E^{\prime}$ $\displaystyle\equiv$ $\displaystyle
E-M=\frac{1}{2\mu}p_{r}^{2}-\frac{1}{32\mu^{3}}p_{r}^{4}\simeq\frac{1}{2\mu}p_{r}^{2},\
\ \ (\mbox{if}\ \ p_{r}^{2}\ll\mu^{2})$ (32)
It is interesting to see that after introducing $\mu$ and $p_{r}$, the energy
$E^{\prime}$ in RMCS looks quite ”relativistic” in the case A whereas it looks
rather ”non-relativistic” in the case B even both of them are derived from the
relativistic expressions, Eq.(28), approximately.
Since the RMCS is not an inertial system, the original mass of $m_{1}$ in CM
changes abruptly to $\mu$ as shown in Eq.(31). How can we derive the reduced
Dirac equation (RDE) in RMCS? Fortunately, we already found a basic symmetry,
the space-time inversion symmetry, which not only serves as the essence of
special relativity (SR), but also goes beyond it to derive the original Dirac
equation and the tachyon theory for neutrinos 14 (14, 15, 16, 17). Based on
this symmetry, we are going to derive the equation in RQM for case either A or
B respectively.
Let us consider case B ($m_{1}\simeq m_{2}$) first. The motivation is stemming
from the success of using the Schrödinger equation to heavy-quarkoniums like
$c\bar{c}$ and $b\bar{b}$ in particle physics (18 (18), see also 15 (15) §9.5
D). Ignoring the spin of both $m_{1}$ and $m_{2}$, we assume the coupling
equations in laboratory system for the two-particle system as:
$\left\\{\begin{array}[]{ll}i\hbar\frac{\partial\varphi}{\partial
t}&=(m_{1}+m_{2})c^{2}\varphi+V(|{\mathbf{r}^{\prime}_{1}-\mathbf{r}^{\prime}_{2}}|)(\varphi+\chi)-(\frac{\hbar^{2}}{2m_{1}}\nabla^{2}_{\mathbf{r}^{\prime}_{1}}+\frac{\hbar^{2}}{2m_{2}}\nabla^{2}_{\mathbf{r}^{\prime}_{2}})(\varphi+\chi)\\\
i\hbar\frac{\partial\chi}{\partial
t}&=-(m_{1}+m_{2})c^{2}\chi-V(|{\mathbf{r}^{\prime}_{1}-\mathbf{r}^{\prime}_{2}}|)(\varphi+\chi)+(\frac{\hbar^{2}}{2m_{1}}\nabla^{2}_{\mathbf{r}^{\prime}_{1}}+\frac{\hbar^{2}}{2m_{2}}\nabla^{2}_{\mathbf{r}^{\prime}_{2}})(\varphi+\chi)\end{array}\right.$
(33)
where $\varphi=\varphi({\mathbf{r}^{\prime}_{1},\mathbf{r}^{\prime}_{2}},t)$
and $\chi=\chi({\mathbf{r}^{\prime}_{1},\mathbf{r}^{\prime}_{2}},t)$ are
hidden ”particle” and ”antiparticle” fields of the two-particle system (From
now on, the ${\mathbf{r}}^{\prime}_{i}(i=1,2)$ is the flowing coordinate of
”fields” in QM, i.e., that of ”fictitious point particles”. See Fig.1).
Eq.(33) remains invariant under the (newly defined) space-time inversion
(${\mathbf{r}^{\prime}_{1}\rightarrow-\mathbf{r}^{\prime}_{1},\mathbf{r}^{\prime}_{2}\rightarrow-\mathbf{r}^{\prime}_{2}},t\rightarrow-t$):
$\left\\{\begin{array}[]{ll}\varphi({-\mathbf{r}^{\prime}_{1},-\mathbf{r}^{\prime}_{2}},-t)&\longrightarrow\chi({\mathbf{r}^{\prime}_{1},\mathbf{r}^{\prime}_{2}},t)\\\
\chi({-\mathbf{r}^{\prime}_{1},-\mathbf{r}^{\prime}_{2}},-t)&\longrightarrow\varphi({\mathbf{r}^{\prime}_{1},\mathbf{r}^{\prime}_{2}},t)\end{array}\right.$
(34)
$V({-\mathbf{r}^{\prime}_{1},-\mathbf{r}^{\prime}_{2}},-t)\longrightarrow
V({\mathbf{r}^{\prime}_{1},\mathbf{r}^{\prime}_{2}},t)$ (35)
Note that, however, the time $t$ is not contained in $V$ explicitly. Eq.(35)
merely means that both $m_{1}$ and $m_{2}\ (m_{1}\approx m_{2})$ transform
into their antiparticles under the space-time inversion. Actually, the hidden
antiparticle field $\chi$ enhances in nearly equal strength in $m_{1}$ and
$m_{2}$ when fictitious particles’ velocities increase with the enhancement of
attractive potential $V(r)$.
After introducing the CM coordinate
${\mathbf{R}}=\frac{1}{M}(m_{1}{\mathbf{r}^{\prime}}_{1}+m_{2}{\mathbf{r}^{\prime}}_{2})$,
($M=m_{1}+m_{2}$) and relative coordinate
$\mathbf{r}=\mathbf{r}^{\prime}_{1}-\mathbf{r}^{\prime}_{2}$, and setting
$\varphi=\Phi+i\frac{\hbar}{Mc^{2}}\dot{\Phi},\ \ \ \
\chi=\Phi-i\frac{\hbar}{Mc^{2}}\dot{\Phi}$ (36)
we find ($\mu=\frac{m_{1}m_{2}}{M}$)
$\ddot{\Phi}-c^{2}\nabla^{2}_{R}\Phi-c^{2}\frac{M}{\mu}\nabla^{2}_{r}\Phi+\frac{1}{\hbar^{2}}(M^{2}c^{4}+2VMc^{2})\Phi=0$
(37)
Its stationary solution reads
$\Phi({\mathbf{R},\mathbf{r}},t)=\psi(\mathbf{r})\exp\left[\frac{i}{\hbar}({\mathbf{P}\cdot\mathbf{R}}-Et)\right]$
(38)
where $E$ is the total energy of the system while $\mathbf{P}$ the momentum of
CM. The reduced ”one-body” equation for $\psi(\mathbf{r})$ turns out to be:
111With Eq.(35), Eq.(37) is invariant under the space-time inversion (${\bf
r}\to-{\bf r},t\to-t$). Equivalently, under the mass inversion ($m_{1}\to-
m_{1},m_{2}\to-m_{2}$), Eq.(37) and Eq.(39) remain invariant in the sense that
not only $\mu\to-\mu,M\to-M$, but also $V({\bf r})\to-V({\bf
r}),\varepsilon\to-\varepsilon$. Notice that, however, the simultaneous
inversion of $m_{1}$ and $m_{2}$ implies $m_{1}\simeq m_{2}$, so both
particles change under their mutual interaction $V({\bf r})$ simultaneously.
Here $V$, being the ”internal potential energy” of two-body system, was called
as a ”scalar potential”. We see that either the invariance under the space-
time inversion or that under the mass inversion is capable of showing the
particle-antiparticle symmetry (i.e., relativistic nature) of a system
essentially.
$\left\\{\begin{array}[]{ll}&\left[-\dfrac{\hbar^{2}}{2\mu}\nabla^{2}_{\mathbf{r}}+V(\mathbf{r})\right]\psi(\mathbf{r})=\varepsilon\psi(\mathbf{r})\\\\[11.38109pt]
&\varepsilon=\dfrac{1}{2Mc^{2}}(E^{2}-M^{2}c^{4}-{\mathbf{P}}^{2}c^{2})\end{array}\right.$
(39)
We set $\mathbf{P}=0$ (i.e.turn to CMCS) and denote the binding energy
$B=Mc^{2}-E$, yielding:
$B=Mc^{2}\left[1-(1+\frac{2\varepsilon}{Mc^{2}})^{1/2}\right]=-\varepsilon+\frac{1}{2}\frac{\varepsilon^{2}}{Mc^{2}}-\cdots$
(40)
Notice that although Eq.(39) looks like a ”non-relativistic” stationary
Schrödinger equation, it is essentially relativistic. This can be seen from
its remarkable property that the eigenvalue $\varepsilon$ has a lower bound
$-\frac{1}{2}Mc^{2}$, corresponding to $E_{\mbox{min}}=0\ (B_{\mbox{max}}=M)$!
An example is: consider ”positronium” composed of $e^{+}$ and $e^{-}$ with
charge $Ze$ and $-Ze$ respectively. Once when the ”fictitious charge number”
$Z$ increases from $1$ to $Z_{max}=(\frac{4}{\alpha^{2}})^{1/4}=16.555$, the
whole bound system would have lowest ground energy $E_{min}=0$! So Eq.(39) is
really a relativistic QM equation capable of giving a nonperturbative solution
under the strong coupling.
Eq.(39) provides a justification (realization) of conjecture Eq.(32) relevant
to case B ($m_{1}\simeq m_{2}$) where the spin of both particles is merely of
second importance.
Now let us turn to case A where $m_{2}\gg m_{1}$, taking the spin of $m_{1}$
into account but ignoring that of $m_{2}$ as before. Based on the experience
in case B, also because of great difficulty to derive the equation starting
from the laboratory system for this case A, we directly introduce the reduced
Dirac equation (RDE) in the RMCS as a pair of coupled equations of two-
component spinors $\varphi({\mathbf{r}},t)$ and $\chi({\mathbf{r}},t)$,
($c=\hbar=1$)
$\left\\{\begin{array}[]{ll}i\dot{\varphi}&=i{\mathbf{\sigma}}_{1}\cdot\nabla_{\mathbf{r}}\chi+\mu\varphi+V({\mathbf{r}})\varphi\\\
i\dot{\chi}&=i{\mathbf{\sigma}}_{1}\cdot\nabla_{\mathbf{r}}\varphi-\mu\chi+V({\mathbf{r}})\chi\end{array}\right.$
(41)
with $\mu$ replacing $m_{1}$. Here $\mathbf{\sigma}_{1}$ are Pauli matrices
acting on the spin space of particle $m_{1}$. Eq.(41) is invariant under the
space-time inversion $({\mathbf{r}\rightarrow-\mathbf{r}},t\rightarrow-t),\
\varphi(-{\mathbf{r}},-t)\rightarrow\chi({\mathbf{r}},t),\
\chi(-{\mathbf{r}},-t)\rightarrow\varphi({\mathbf{r}},t)$ whereas we assume
$V(-{\mathbf{r}},-t)\longrightarrow-V({\mathbf{r}},t)$ (42)
here in contrast to Eq.(35) for the case B. 222For a hydrogenlike atom,
$V(r)=-\frac{Ze^{2}}{r}$ does not contain time $t$ explicitly. Eq.(42) merely
means that under the space-time inversion, the electron transforms into a
position whereas the nucleus remains unchanged. See point (a) of section IX.
Previously, the $V$ in Eq.(42) was called as a ”vector potential”, meaning the
”potential energy” of the electron in an ”external field” of nucleus. Note
that, formally, Eq.(41) remains invariant under a mass inversion as
$\mu\to-\mu,\phi\to\chi,\chi\to\phi$ ($V({\bf r})$ remains unchanged) in a
noninertial frame $RMCS$. Actually, since $m_{1}=m_{e}\to-m_{e}$, but
$m_{2}=m_{N}\to m_{N},\mu\to-\mu(1+\frac{2m_{e}}{M})$. So Eq.(41) has an
inaccuracy up to $\frac{2m_{e}}{M}$ ($<1.1\times 10^{-3}$ for $H$).
The reasons are as follows: (a) Eq.(41) should degenerate into the original
Dirac equation when $m_{2}\rightarrow\infty$,
$\mu=\frac{m_{1}m_{2}}{m_{1}+m_{2}}\rightarrow m_{1}$. (b) Since now $m_{2}\gg
m_{1}$ (but $m_{2}\neq\infty$), $m_{1}$ is moving much faster than $m_{2}$ in
the CMCS. Hence the antiparticle field $\chi$ enhances much appreciably in
$m_{1}$ than that in $m_{2}$, a situation totally different from that in the
case B where $m_{1}\approx m_{2}$. (c) If instead of Eq.(42), we still assume
$V(-{\mathbf{r}},-t)\longrightarrow V({\mathbf{r}},t)$ like Eq.(35) and change
the sign before $V(\mathbf{r})$ in the second equation of Eq.(41) to keep its
invariance under the space-time inversion, then we would get an equation which
would lead to a reversed fine-structure of atom (e.g., the $P_{1/2}$ state
would lie above the $P_{3/2}$ state), a wrong prediction obviously excluded by
experiments.
However, one kind of invariance is not enough to fix an equation. Indeed, the
beauty of Dirac equation or RDE is hidden in two symmetries: besides the
symmetry of space-time inversion, it has another left-right (parity) symmetry.
To see it, we define
$\xi=\frac{1}{\sqrt{2}}(\varphi+\chi),\ \ \
\eta=\frac{1}{\sqrt{2}}(\varphi-\chi)$ (43)
and recast Eq.(41) into:
$\left\\{\begin{array}[]{ll}i\dot{\xi}&=i{\mathbf{\sigma}}_{1}\cdot\nabla_{\mathbf{r}}\xi+\mu\eta+V(\mathbf{r})\xi\\\
i\dot{\eta}&=-i{\mathbf{\sigma}}_{1}\cdot\nabla_{\mathbf{r}}\eta+\mu\xi+V(\mathbf{r})\eta\end{array}\right.$
(44)
which is invariant under a pure space inversion
(${\mathbf{r}\rightarrow-\mathbf{r}},t\rightarrow t$) if assuming
$\xi({-\mathbf{r}},t)\rightarrow\eta({\mathbf{r}},t),\ \
\eta({-\mathbf{r}},t)\rightarrow\xi({\mathbf{r}},t),\ \
V(-{\mathbf{r}})\rightarrow V({\mathbf{r}})=V(r)$ (45)
The parity invariance of Dirac equation or RDE has a far-reaching consequence
that the Dirac particle is always a subluminal one. By contrast, once the
parity is violated to maximum, a superluminal particle (tachyon) will emerge.
Interestingly enough, any theory capable of treating particle and antiparticle
on an equal footing must respect to the common basic symmetry—the invariance
of space-time inversion. The new insight of this section is this symmetry can
be applied even in a noninertial frame—the RMCS. Of course, the validity of
RDE can only be verified by experiments as discussed in section II, although
it is still an approximate description of nature like any other theory in
physics. For further discussion, see section IX.
## IV Self-Energy Correction of a Bound Electron in Atom
In our understanding, one important reason why the calculations of QED for
electron in a hydrogenlike atom is so complicated lies in the fact that while
calculations are performed in the CMCS, the center of potential (the nucleus
with mass $m_{2}=m_{N}$) undergoes a complex motion. So the recoil effect
interwinds with the high-loop correction of QED, as discussed in many chapters
of the books 1 and 2 . We will try to find an alternative approach by
adopting the RDE and doing calculation in the RMCS. Let us begin with the
Feynman diagram integral (FDI) of electron self-energy at one-loop level,
adopting the Bjorken-Drell metric and rationalized Gaussian units with
electron charge $-e(e>0)$, see Fig.2(a) (8a ).
$-i\Sigma(p)=(ie)^{2}\int\frac{d^{4}k}{(2\pi)^{4}}\frac{g_{\mu\nu}}{ik^{2}}\gamma^{\mu}\frac{i}{\not\\!\\!p-\not\\!\\!k-\mu}\gamma^{\nu}$
(46)
Here a free electron with reduced mass $\mu$ is moving at a four-dimensional
momentum $p$, whose spatial component is just the relative momentum
$\mathbf{p}_{r}$ discussed in the previous section, $k$ is the momentum of
virtual photon. As usual, a Feynman parameter $x$ will bring Eq.(46) into
$-i\Sigma(p)=-e^{2}\int\frac{d^{4}k}{(2\pi)^{4}}\frac{N}{D}$ (47)
$\frac{1}{D}=\int^{1}_{0}\frac{dx}{[k^{2}-2p\cdot kx+(p^{2}-\mu^{2})x]^{2}},\
\ \ N=-2(\not\\!\\!p-\not\\!\\!k)+4\mu$ (48)
($\not\\!\\!p=p^{\mu}\gamma_{\mu},\ p\cdot k=p^{\mu}k_{\mu}$). A shift in
momentum integration:$k\rightarrow K=k-xp$ recast Eq.(47) into
$-i\Sigma(p)=-e^{2}\int^{1}_{0}dx[-2(1-x)\not\\!\\!p+4\mu]I$ (49)
with a logarithmically divergent integral (in Minkowski momentum space):
$I=\int\frac{d^{4}K}{(2\pi)^{4}}\frac{1}{[K^{2}-M^{2}]^{2}},\ \ \
M^{2}=p^{2}x^{2}+(\mu^{2}-p^{2})x$ (50)
Our new regularization-renormalization method (RRM) is based on a cognition
that the virtual process in the self-energy diagram does provide a radiative
correction to the electron mass but only when the electron is off the mass
shell, i.e., $p^{2}\not=\mu^{2}$. When it is on the mass shell,
$p^{2}=\mu^{2}$, the appearance of a divergent integral like $I$ in Eq.(50) is
essentially a warning on the fact that to calculate the mass of electron is
beyond the ability of perturbative QED.
Let us consider the converse: if $\Sigma(p)$ does modify the electron mass
$\mu$ to some extent, it must comes from the divergent integral $I$. However,
the latter is a dimensionless number, we can change the unit of $M$ (and $k$)
at our disposal without any change in the value of $I$. So any real change of
$\mu$ (on the mass shell) is incredible. The deeper reason lies in a
”principle of relativity” in epistemology: everything is moving and becomes
recognizable only in relationship with other things. What we can understand is
either no mass scale or two mass scales, but never one mass scale. For
instance, in the famous Gross-Neveu model 19 (19), a massive fermion is
created only in accompanying with the change (phase transition) of its
environment (vacuum) which provides another mass scale (a standard weight).
Another example is just the change of electron mass from $m_{e}$ to $\mu_{H}$
in a hydrogen atom due to the coexistence of atom nucleus—the proton, this
change is also a nonperturbative effect.
Therefore, we expected too much in the past. There is no way to evaluate
Eq.(50) unambiguously or pick out some finite and fixed modification on the
mass $\mu$. What we can do is to separate the valuable information carried by
Eq.(50) from an arbitrary constant which will be introduced by a simple trick
and then fixed by the experimental data of $\mu$. We will see the information
telling us exactly how the value of $I$ changes when the electron is moving
off the mass shell.
To handle Eq.(50), we perform a differentiation with respect to the mass-
square parameter $M^{2}$, then the integration with respect to $K$ becomes
convergent, yielding:
$\frac{\partial I}{\partial M^{2}}=\frac{-i}{(4\pi)^{2}}\frac{1}{M^{2}}$ (51)
which tells us that while the exact value of $I$ remains obscure, its change
linked with $M^{2}$ has a definite meaning. So we reintegrate Eq.(51) with
respect to $M^{2}$ and arrive at
$I=\frac{-i}{(4\pi)^{2}}(\ln
M^{2}+C_{1})=\frac{-i}{(4\pi)^{2}}\ln\frac{M^{2}}{\mu_{2}^{2}}$ (52)
where an arbitrary constant $C_{1}=-\ln\mu_{2}^{2}$ is introduced ($\mu_{2}$
should not be confused with the reduced mass $\mu$). Further integration with
respect to Feynman parameter $x$ leads to
$\begin{array}[]{ll}\Sigma(p)&=A+B\not\\!\\!p\\\
A&=\frac{\alpha}{\pi}\mu\left[2-2\ln\frac{\mu}{\mu_{2}}+\frac{(\mu^{2}-p^{2})}{p^{2}}\ln\frac{(\mu^{2}-p^{2})}{\mu^{2}}\right]\\\
B&=\frac{\alpha}{4\pi}\left[2\ln\frac{\mu}{\mu_{2}}-3-\frac{(\mu^{2}-p^{2})}{p^{2}}\left[1+\frac{(\mu^{2}+p^{2})}{p^{2}}\ln\frac{(\mu^{2}-p^{2})}{\mu^{2}}\right]\right]\end{array}$
(53)
Using the chain approximation, we can derive the modification of electron
propagator as
$\frac{i}{\not\\!\\!p-\mu}\rightarrow\frac{i}{\not\\!\\!p-\mu}\frac{1}{1-\frac{\Sigma(p)}{\not\\!\\!p-\mu}}=\frac{iZ_{2}}{\not\\!\\!p-\mu_{R}}$
(54)
where
$Z_{2}=\frac{1}{1-B}$ (55)
is the renormalization factor for wave function of electron and
$\mu_{R}=\frac{\mu+A}{1-B}$ (56)
is the renormalized mass of $\mu$. The increment of mass reads
$\delta\mu=\mu_{R}-\mu=\frac{A+\mu B}{1-B}$ (57)
For a free electron (in the atom), the mass-shell condition $p^{2}=\mu^{2}$
should lead to
$\delta\mu|_{p^{2}=\mu^{2}}=\frac{\alpha\mu}{4\pi}(5-6\ln\frac{\mu}{\mu_{2}})=0$
(58)
as discussed above333We will keep the same mass symbol $\mu$ through out high-
loop calculations of QED and reconfirm (renormalize) it at every step by
experiment. Just like one has to reconfirm his plane ticket before his
departure from the airport, he must use the same name through out his entire
journey 8b .. So we must set $\mu_{2}=\mu e^{-5/6}$ which in turn fixes
$Z_{2}|_{p^{2}=\mu^{2}}=\frac{1}{1+\frac{\alpha}{3\pi}}\approx
1-\frac{\alpha}{3\pi}$ (59)
However, the above evaluation further provides us with important knowledge of
$\delta\mu$ when electron is moving off the mass-shell. Consider the similar
diagram in Fig.(2b), we can set on an average meaning that
$p^{2}=\mu^{2}(1-\zeta)$ (60)
with $\zeta>0$, which implies from Eq.(57) with Eq.(53) that 20 (20):
$\delta\mu=\frac{\alpha\mu}{4\pi}\frac{(-\zeta+2\zeta\ln\zeta)}{1+\alpha/3\pi}$
(61)
where some terms of the order of $\zeta^{2}$ or $\zeta^{2}\ln\zeta$ are
neglected since $\zeta\ll 1$. Eq.(61) establishes the correspondence between
the mass modification $\delta\mu$ and the parameter $\zeta$ describing the
off-mass-shell extent of electron in the bound state. For a hydrogenlike atom,
we may ascribe $\delta\mu$ to the (minus) binding energy of electron in the
Bohr theory:
$\delta\mu=\varepsilon_{n}=-\frac{Z^{2}\alpha^{2}}{2n^{2}}\mu$ (62)
Then Eq.(61) gives the value of $\zeta$ for fixed values $Z$ and $n$. We will
see from the vertex function that these values of $\zeta$ are crucial to the
calculation of Lamb shift (sections VII and VIII).
## V Photon Self-energy
As discussed in various text books 21 (21, 22, 23), we encounter the FDI of
vacuum polarization Fig.2(c) as 8a :
$\Pi_{\mu\nu}(q)=-(-ie)^{2}Tr\int\frac{d^{4}\bar{p}}{(2\pi)^{4}}\gamma_{\mu}\frac{i}{\not\\!\\!\bar{p}-m}\gamma_{\nu}\frac{i}{\not\\!\\!\bar{k}+\not\\!\\!q-m}$
(63)
Here $q$ is the momentum transfer along the photon line and $m$ the mass of
electron. Introducing the Feynman parameter $x$ as in previous section and
performing a shift in momentum integration: $\bar{p}\rightarrow K=\bar{p}+xq$,
we get:
$\Pi_{\mu\nu}(q)=-4e^{2}\int_{0}^{1}dx(I_{1}+I_{2})$ (64)
where
$I_{1}=\int\frac{d^{4}K}{(2\pi)^{4}}\frac{2K_{\mu}K_{\nu}-g_{\mu\nu}K^{2}}{(K^{2}-M^{2})^{2}}$
(65)
with
$M^{2}=m^{2}+q^{2}(x^{2}-x)$ (66)
is quadratically divergent while
$I_{2}=\int\frac{d^{4}K}{(2\pi)^{4}}\frac{(x^{2}-x)(2q_{\mu}q_{\nu}-g_{\mu\nu}q^{2})+m^{2}g_{\mu\nu}}{(K^{2}-M^{2})^{2}}$
(67)
is only logarithmically divergent like that in Eq.(50). An elegant way to
handle $I_{1}$, Eq.(65), is modifying $M^{2}$ to
$M^{2}(\sigma)=m^{2}+q^{2}(x^{2}-x)+\sigma$ (68)
and differentiating $I_{1}$ with respect to $\sigma$. After integration with
respect to $K$, we reintegrate it with respect to $\sigma$ twice, arriving at
the limit $\sigma\rightarrow 0$:
$I_{1}=\frac{ig_{\mu\nu}}{(4\pi)^{2}}\left\\{[m^{2}+q^{2}(x^{2}-x)]\ln\frac{m^{2}+q^{2}(x^{2}-x)}{\mu_{3}^{2}}+C_{2}\right\\}$
(69)
with two arbitrary constant: $C_{1}=-\ln\mu_{3}^{2}$ and $C_{2}$. Combining
$I_{1}$ and $I_{2}$ together, we find:
$\Pi_{\mu\nu}(q)=\frac{8ie^{2}}{(4\pi)^{2}}(q_{\mu}q_{\nu}-g_{\mu\nu}q^{2})\int_{0}^{1}dx(x^{2}-x)\ln\frac{m^{2}+q^{2}(x^{2}-x)}{\mu_{3}^{2}}-\frac{4ie^{2}}{(4\pi)^{2}}g_{\mu\nu}C_{2}$
(70)
The continuity equation of current induced in the vacuum polarization 21 (21)
$q^{\mu}\Pi_{\mu\nu}(q)=0$ (71)
is ensured by the factor $(q_{\mu}q_{\nu}-g_{\mu\nu}q^{2})$. So we set
$C_{2}=0$. Consider the scattering between two electrons via the exchange of a
photon with momentum transfer $q\rightarrow 0$. Adding the contribution of
$\Pi_{\mu\nu}(q)$ to the tree diagram amounts to modify the charge square:
$e^{2}\longrightarrow e^{2}_{R}=Z_{3}e^{2},\ \ \
Z_{3}=1+\frac{\alpha}{3\pi}(\ln\frac{m^{2}}{\mu_{3}^{2}}-\frac{q^{2}}{5m^{2}}+\cdots)$
(72)
As in Ref.8b , we will set $\mu_{3}=m$ so that at the Thomson
limit:$\lim_{q\rightarrow 0}e_{R}^{2}=e^{2}$. However, for the purpose of
calculating Lamb shift (LS) below, the second term in the parenthesis of
$Z_{3}$ is important because for a bound state it contributes a term of
effective potential (adding to Coulomb potential), called the Uehling
potential (23 (23),p.253):
$-\frac{4\alpha^{2}}{15m^{2}}\delta(\mathbf{r})$ (73)
## VI The Off-Mass-Shell Vertex Function
Consider an electron (see Fig.2(d)) moving in a hydrogen atom, its momentum
changes from $p$ to $p^{\prime}$ via the scattering by the proton and an
exchange of virtual photon with momentum $k$. The FDI at one-loop level reads
$\Lambda_{\mu}(p^{\prime},p)=(-ie)^{2}\int\frac{d^{4}k}{(2\pi)^{4}}\frac{-i}{k^{2}}\gamma_{\nu}\frac{i}{\not\\!\\!p^{\prime}-\not\\!\\!k^{\prime}-\mu}\gamma_{\mu}\frac{i}{\not\\!\\!p-\not\\!\\!k-\mu}\gamma^{\nu}$
(74)
However, different from 8b and many other literatures, not only the reduced
mass $\mu$ (instead of $m$) of electron is used, but also a new approach will
be adopted. We assume that the electron is moving off-mass-shell in the sense
of (as in section IV):
$p^{2}=p^{\prime 2}=\mu^{2}(1-\zeta)$ (75)
We still have
$p^{\prime}-p=q,\ \ \ p\cdot q=-\frac{1}{2}q^{2}$ (76)
Introducing Feynman parameters $u=x+y$ and $v=x-y$, we perform a shift in the
momentum integration:$k\rightarrow K=k-(p+q/2)u-(q/2)v$, thus
$\Lambda_{\mu}=-ie^{2}[I_{3}\gamma_{\mu}+I_{4}]$ (77)
$\displaystyle I_{3}$ $\displaystyle=$
$\displaystyle\int_{0}^{1}du\int_{-u}^{u}dv\int\frac{d^{4}K}{(2\pi)^{4}}\frac{K^{2}}{(K^{2}-M^{2})^{3}}$
(78) $\displaystyle M^{2}$ $\displaystyle=$
$\displaystyle[\mu^{2}(1-\zeta)-\frac{q^{2}}{4}]u^{2}+\frac{q^{2}}{4}v^{2}+\zeta\mu^{2}u$
(79) $\displaystyle I_{4}$ $\displaystyle=$
$\displaystyle\int_{0}^{1}du\int_{-u}^{u}dv\int\frac{d^{4}K}{(2\pi)^{4}}\frac{A_{\mu}}{(K^{2}-M^{2})^{3}}$
(80) $\displaystyle A_{\mu}\\!\\!\\!$ $\displaystyle=$
$\displaystyle\\!\\!\\!(4\\!\\!-\\!\\!4u\\!\\!-\\!\\!2u^{2})\mu^{2}(1\\!\\!-\\!\\!\zeta)\gamma_{\mu}\\!+\\!2i(u^{2}\\!\\!-\\!\\!u)\mu
q^{\nu}\\!\sigma_{\mu\nu}\\!-\\!(2\\!-\\!2u\\!+\\!\frac{u^{2}}{2}\\!-\\!\frac{v^{2}}{2})q^{2}\gamma_{\mu}\\!-\\!(2\\!+\\!2u)v\mu
q_{\mu}$ (81)
Set $K^{2}=K^{2}-M^{2}+M^{2}$, then $I_{3}=I^{\prime}_{3}-\frac{i}{32\pi^{2}}$
and $I^{\prime}_{3}$ is only logarithmically divergent and so can be treated
as in previous sections, yielding:
$I^{\prime}_{3}=\frac{-i}{(4\pi)^{2}}\int_{0}^{1}du\int_{-u}^{u}dv\ln\frac{M^{2}}{\mu_{1}^{2}}$
(82)
with $\mu_{1}$ an arbitrary constant.
However, unlike Ref.8b where the calculation was conducted on the mass-shell,
now the off-mass-shell integration in Eq.(82) can be performed in the
approximation that $\frac{Q^{2}}{4\mu^{2}}\ll 1$ and $\zeta\ll 1$
($Q^{2}=-q^{2}$, $Q$ is the three-dimensional momentum transfer) which will be
enough to calculate the Lamb shift (LS). Denoting
$a=[\mu^{2}(1-\zeta)+\frac{Q^{2}}{4}]u^{2}+\zeta\mu^{2}u,\ \ \
b=\frac{Q^{2}}{4}$ (83)
we will perform the integration with respect to $v$ and $u$ rigorously:
$\int_{-u}^{u}dv\ln(a-bv^{2})=2u[\ln\mu^{2}+\ln
u+\ln[(1-\zeta)u+\zeta]-4u+2\sqrt{\frac{4a}{Q^{2}}}\ln\frac{1+\sqrt{Q^{2}/4a}u}{1-\sqrt{Q^{2}/4a}u}$
(84)
Expanding the last term and keeping only up to the order of $\zeta$ and
$Q^{2}/4\mu^{2}$, we obtain
$\int_{0}^{1}du\int_{-u}^{u}dv\ln(a-bv^{2})\simeq\ln\mu^{2}-1+\zeta+\frac{Q^{2}}{6\mu^{2}}(1-\zeta)$
(85)
To our great pleasure, throughout the evaluation of $I_{4}$, there is no any
infrared divergence which would appear in previous literatures when
integrating with respect to $u$ with lower limit zero. To avoid the infrared
divergence, e.g., in 8b , a cutoff was introduced at the lower limit. Now the
infrared divergence disappears due to the existence of off-mass-shell
parameter $\zeta$. For example, we encounter the following integral, in which
no cutoff is needed ($\lambda=(1-\zeta)+Q^{2}/4\mu^{2}\sim 1$):
$\int_{0}^{1}\frac{du}{u+\zeta/\lambda}=\frac{\zeta}{\lambda}-\ln\frac{\zeta}{\lambda}$
(86)
Hence, after elementary but tedious calculation, we find:
$\begin{array}[]{l}\Lambda_{\mu}(p^{\prime},p)=\frac{\alpha}{4\pi}[\frac{11}{2}-\ln\frac{\mu^{2}}{\mu_{1}^{2}}-3\zeta+4(1+\zeta)\ln\zeta]\gamma_{\mu}+\frac{\alpha}{4\pi}\frac{Q^{2}}{\mu^{2}}\gamma_{\mu}(\frac{1}{6}+\frac{1}{2}\zeta+\frac{4}{3}\ln\zeta+2\zeta\ln\zeta)\\\
+i\frac{\alpha}{4\pi}\frac{q^{\nu}}{\mu}\sigma_{\mu\nu}(1+3\zeta+2\zeta\ln\zeta)\end{array}$
(87)
## VII Calculation of Lamb Shift as an Off-Mass-Shell Effect at One-Loop
Level
There are three parts in Eq.(87). The first part in combination with the
vertex $\gamma_{\mu}$ at tree level provides a renormalization factor as
$Z_{1}^{-1}=1+\frac{\alpha}{4\pi}[\frac{11}{2}-\ln\frac{\mu^{2}}{\mu_{1}^{2}}-3\zeta+4(1+\zeta)\ln\zeta]$
(88)
Further combination with $Z_{2}$ in Eq.(55) and $Z_{3}$ in Eq.(72) leads to a
renormalized charge (at one-loop level, see Fig.2):
$e_{R}=\frac{Z_{2}}{Z_{1}}Z_{3}^{1/2}e$ (89)
However the Ward identity implies that 21 (21, 22, 23)
$Z_{1}=Z_{2}$ (90)
Therefore
$\alpha_{R}=\frac{e_{R}^{2}}{4\pi}=Z_{3}\alpha$ (91)
Note that Ward identity holds not only for an electron on the mass-shell, but
also for off-mass-shell case. Hence for every bound state in hydrogenlike atom
with a definite value of $\zeta$ ($Z_{1}$ and $Z_{2}$ are functions of
$\zeta$), the arbitrary constant $\mu_{1}$ in Eq.(88) plays a flexible role to
guarantee the validity of Eq.(90) (other two constants $\mu_{2}$ and $\mu_{3}$
had been fixed in Eq.(58) and (72) respectively). For further discussion, see
section IX.
The second part of Eq.(87) contains $Q^{2}\gamma_{\mu}$. Just like the Uehling
potential in Eq.(72) (with $q^{2}=-Q^{2}$), it contributes an effective
potential of $\delta$ function type as
$\frac{\alpha^{2}}{\mu^{2}}[-\frac{1}{6}-\frac{1}{2}\zeta-\frac{4}{3}\ln\zeta-2\zeta\ln\zeta]\delta(\mathbf{r})$
(92)
Finally, the third part of Eq.(87) amounts to a modification of electron
magnetic moment in the atom, the gyromagnetic ratio of electron reads:
$g=2[1+\frac{\alpha}{2\pi}(1+3\zeta+2\zeta\ln\zeta)]$ (93)
We will call the anomalous part of magnetic moment
$a=\frac{\tilde{\alpha}}{2\pi}$,
$\tilde{\alpha}=\alpha(1+3\zeta+2\zeta\ln\zeta)$. The radiative correction on
the magnetic moment of an electron has two consequences. One is a modification
to the L-S coupling in a hydrogenlike atom (with charge number $Z$) 21 (21,
22):
$H_{LS}^{rad}=2(\frac{\tilde{\alpha}}{2\pi})\frac{\alpha
Z}{4\mu^{2}r^{3}}{\mathbf{\sigma}\cdot L}$ (94)
Here the electron mass has been modified from $m$ (see, e.g., 15 (15)) to
$\mu$ which can be derived from the reduced Dirac equation.
Another consequence of anomalous magnetic moment of electron exhibits itself
as an additional potential of $\delta$ function type like Eq.(73)21 (21, 22)
$\frac{Z\alpha\tilde{\alpha}}{2\mu^{2}}\delta(\mathbf{r})$ (95)
Note that Eqs.(94) and (95) are only effective to states with $L\not=0$ and
$S$ state with $L=0$ respectively.
Adding the results of Eqs.(94), (95) and the sum of Eqs.(73) and (92)
multiplied by $Z$ together to get all radiative corrections (at one-loop
level) on electron in the hydrogenlike atom, then we get the effective
potential as
$\begin{array}[]{ll}V_{eff}^{rad}&=\frac{Z\alpha^{2}}{\mu^{2}}[-\frac{4}{3}\ln\zeta-\frac{1}{2}\zeta-2\zeta\ln\zeta-\frac{1}{6}-\frac{4}{15}\frac{\mu^{2}}{m^{2}}+\frac{1}{2}(1+3\zeta+2\zeta\ln\zeta)]\delta(\mathbf{r})\\\
&+\frac{Z\alpha^{2}}{4\pi\mu^{2}r^{3}}(1+3\zeta+2\zeta\ln\zeta){\mathbf{\sigma}\cdot
L}\\\
&\simeq\frac{Z\alpha^{2}}{\mu^{2}}[-\frac{4}{3}\ln\zeta+\frac{1}{15}+\zeta-\zeta\ln\zeta]\delta(\mathbf{r})+\frac{Z\alpha^{2}}{4\pi\mu^{2}r^{3}}(1+3\zeta+2\zeta\ln\zeta){\mathbf{\sigma}\cdot
L}\end{array}$ (96)
where we take $\mu^{2}/m^{2}\approx 1$ in the Uehling potential to make the
formula simpler for a semi-quantitative calculation. Eq.(96) leads to the
energy modification of a bound state (with quantum numbers $n,l,j$) in a
hydrogenlike atom:
$\delta({\mathbf{r}})\longrightarrow|\psi_{ns}(0)|^{2}=\frac{Z^{3}\alpha^{3}}{\pi
n^{3}}\mu^{3},\ \ \ (l=0)$ (97)
$\Delta E^{rad}=\Delta E^{rad}(ns)+\Delta E^{rad}_{LS}$ (98)
$\Delta E^{rad}(ns)=\frac{Z^{4}\alpha^{3}}{\pi
n^{3}}R_{y}[\frac{8}{3}\ln\frac{1}{\zeta}+\frac{2}{15}+2\zeta(1-\ln\zeta)]\delta_{l0}$
(99)
$\Delta E^{rad}_{LS}=\frac{Z^{4}\alpha^{3}}{\pi
n^{3}}R_{y}\frac{1+\zeta(3+2\ln\zeta)}{l(2l+1)(l+1)}\left\\{\begin{array}[]{ll}&l,\
\ \ \ \ \ (j=l+1/2)\\\ -&(l+1),\ \ \ (j=l-1/2)\end{array}\right.$ (100)
where
$R_{y}=\frac{1}{2}\alpha^{2}\mu=\frac{\mu}{m}R_{\infty}$ (101)
## VIII Energy-Level Difference in Hydrogenlike Atom: Theory vs. Experiment
We will study some energy-level differences near the ground state of
hydrogenlike atoms, where precise experimental data are available.
Theoretically, the energy level is fixed primarily by the formula derived from
the reduced Dirac equation (RDE), i.e., Eq.(1) with $m_{e}$ substituted by
$\mu_{A}$ where the subscript $A$ refers to atom $H$, $D$ or $He^{+}$, et
al..:
$\displaystyle E_{A}^{RDE}$ $\displaystyle=$
$\displaystyle\mu_{A}[f(n,j)-1]=\frac{m_{e}}{1+b_{A}}[f(n,j)-1]$ (102)
$\displaystyle=$ $\displaystyle\frac{1}{1+b_{A}}(1.2355897\times
10^{20})[-\frac{(Z\alpha)^{2}}{2n^{2}}-\frac{(Z\alpha)^{4}}{3n^{3}}(\frac{1}{j+1/2}-\frac{3}{4n})-\cdots]\
\ Hz$
Further recoil corrections Eq.(6) derived by previous authors will be divided
into two terms:
$\displaystyle\Delta E_{A}^{recoil-1}(n,j)$ $\displaystyle=$
$\displaystyle-\frac{\mu^{2}_{A}}{2M_{A}}[f(n,j)-1]^{2}=-\frac{m_{e}b_{A}}{2(1+b_{A})^{3}}[f(n,j)-1]^{2}$
(103) $\displaystyle\Delta E_{A}^{recoil-2}(n,j,l)$ $\displaystyle=$
$\displaystyle\frac{(Z\alpha)^{4}\mu^{3}_{A}}{2n^{3}m_{N}^{(A)^{2}}}(\frac{1}{j+\frac{1}{2}}-\frac{1}{l+\frac{1}{2}})(1-\delta_{l0})$
(104)
Next comes the radiative correction calculated by QED at one-loop level,
Eq.(98):
$\Delta
E_{A}^{rad}(n,j,l)=\frac{1}{1+b_{A}}\frac{Z^{4}}{n^{3}}(\frac{\alpha^{3}}{\pi}R_{\infty})[(-\frac{8}{3}\ln\zeta+\frac{2}{15}+2\zeta(1-\ln\zeta))\delta_{l0}+\frac{1+\zeta(3+2\ln\zeta)}{2l+1}C_{jl}(1-\delta_{l0})]$
(105)
where
$C_{jl}=\left\\{\begin{array}[]{ll}&\frac{1}{l+1},\ \ \ j=l+\frac{1}{2}\\\
&-\frac{1}{l},\ \ \ j=l-\frac{1}{2}\end{array}\right.$ (106)
Finally, the finite nucleus size (NS) with radius $r_{N}^{(A)}$ brings a
correction 10 (10):
$\begin{array}[]{ll}\Delta
E_{A}^{NS}(n,j)&=\frac{4}{3}(\frac{\mu_{A}}{m_{e}})^{3}\frac{Z^{4}}{n^{3}}(\frac{r_{N}^{(A)}}{a_{\infty}})^{2}R_{\infty}\delta_{l0}\\\
&=(\frac{1}{1+b_{A}})^{3}\frac{Z^{4}}{n^{3}}(4.386454987\times
10^{7})[\frac{r_{N}^{(A)}(fm)}{5.2917725}]^{2}\delta_{l0}\ \ Hz\end{array}$
(107)
As explained in Eq.(61) with Eq.(62), the value of off-mass-shell parameter
$\zeta$ in Eq.(105) can be calculated from the electron self-energy at one-
loop level:
$\frac{Z^{2}\alpha}{n^{2}}=\frac{1}{2\pi}\frac{(\zeta^{<S>}-2\zeta^{<S>}\ln\zeta^{<S>})}{1+\alpha/3\pi}$
(108)
where the superscript $<S>$ refers to ”self-energy”. However, we may derive
the value of $\zeta$ in an alternative way. Divide the square average of four-
dimensional momentum $p$ into two parts:
$<p^{2}>=<E^{2}>-<{\mathbf{p}}^{2}>$ (109)
where
$<E^{2}>=E^{2}=(\mu-B)^{2}\simeq\mu^{2}-2\mu B,$ (110)
since the binding energy
$B=\frac{Z^{2}\alpha^{2}}{2n^{2}}\mu\ll\mu$ (111)
The square average of three-dimensional momentum $\mathbf{p}$,
$<{\mathbf{p}}^{2}>$, can be evaluated by the Virial theorem (e.g., 15 (15)).
In a Coulomb field, an electron has potential energy $V=-\frac{Ze^{2}}{4\pi
r}$ and kinetic energy $T=\frac{1}{2\mu}{\mathbf{p}}^{2}$. Then
$\displaystyle<{\mathbf{p}}^{2}>$ $\displaystyle=$ $\displaystyle
2\mu<T>=2\mu[-B-<V>]=2\mu B$ $\displaystyle<p^{2}>$ $\displaystyle=$
$\displaystyle\mu^{2}-4\mu B=\mu^{2}(1-\frac{4B}{\mu})$ (112)
Comparing Eq.(112) with $<p^{2}>=\mu^{2}(1-\zeta^{<V>})$, we find
$\zeta^{<V>}=\frac{4B}{\mu}=\frac{2Z^{2}\alpha^{2}}{n^{2}}$ (113)
where the superscript $<V>$ refers to ”Virial theorem”. Table 1 gives the
values of $\zeta^{<S>}$ and $\zeta^{<V>}$ with their logarithm values as well
as two kinds of ”average”,
$\zeta^{<S+V>}=\frac{1}{2}(\zeta^{<S>}+\zeta^{<V>})$ and
$\zeta^{<SV>}=\sqrt{\zeta^{<S>}\zeta^{<V>}}$, to be used in Eq.(105).
Table 1. Off-mass-shell parameter $\zeta$ and $\ln\zeta$ |
---|---
$\frac{Z^{2}}{n^{2}}$ | $\zeta^{<S>}\times 10^{4}$ | -$\ln\zeta^{<S>}$ | $\zeta^{<V>}\times 10^{6}$ | -$\ln\zeta^{<V>}$ | $\zeta^{<S+V>}\times 10^{5}$ | -$\ln\zeta^{<S+V>}$ | $\zeta^{<SV>}\times 10^{5}$ | $-\ln\zeta^{<SV>}$
$\frac{1}{16}$ | $1.546093458$ | $8.77461$ | $\frac{\alpha^{2}}{8}=6.6564192$ | 11.91992886 | $8.0632$ | 9.425609 | 3.2080284 | 10.34727
$\frac{1}{4}$ | $7.446539697$ | 7.20259 | $\frac{\alpha^{2}}{2}=26.6256771$ | 10.5336345 | $38.5639$ | 7.860609 | 14.0808 | 8.86816225
1 | $37.73719345$ | 5.57969 | $2\alpha^{2}=106.502$ | 9.147340142 | $194.011$ | 6.2450103 | 63.39626 | 7.36351521
Now we are in a position to discuss a number of cases:
(a) The so-called classic Lamb shift of hydrogen atom was measured
experimentally as 10 (10):
$L_{H}^{exp}(2S-2P)\equiv E_{H}(2S_{1/2})-E_{H}(2P_{1/2})=1057.845\ \ MHz$
(114)
Theoretically, in this case ($b_{H}=5.446170255\times 10^{-4},\
r_{N}^{H}=r_{p}=0.862fm$), Eqs.(102) and (103) make no contributions while
Eqs.(104) and (107) only contribute
$\Delta
E_{H}^{recoil-2}(2S_{1/2}-2P_{1/2})=-E_{H}^{recoil-2}(2,1/2,1)=-2.16156\ \
kHz$ (115)
and
$\Delta E_{H}^{NS}(2S-2P)=0.14525347\ \ MHz$ (116)
respectively. The dominant contribution comes from Eq.(105). If using
$\zeta^{<S>}$, we obtain
$\begin{array}[]{l}\Delta
E_{H}^{Rad<S>}(2S-2P)=\frac{1}{1+b_{H}}\frac{1}{8}(4.06931316\times
10^{8})[-\frac{8}{3}\ln\zeta^{<S>}+\frac{7}{15}+3\zeta^{<S>}-\frac{4}{3}\zeta^{<S>}\ln\zeta^{<S>}]\\\
=1000.6567\ MHz\end{array}$ (117)
If we use another three values of $\ln\zeta$ in Table 1, we get
$\displaystyle\Delta E_{H}^{Rad<V>}(2S-2P)$ $\displaystyle=$ $\displaystyle
1451.7912\ \ MHz$ (118) $\displaystyle\Delta E_{H}^{Rad<S+V>}(2S-2P)$
$\displaystyle=$ $\displaystyle 1089.6513\ \ MHz$ (119) $\displaystyle\Delta
E_{H}^{Rad<SV>}(2S-2P)$ $\displaystyle=$ $\displaystyle 1226.0871\ \ MHz$
(120)
It seems that Eq.(117) is smaller whereas Eq.(118) too large. So as an
empirical rule in our semiquantitative calculation, we may use Eq.(119) to get
$L_{H}^{theor.}(2S-2P)=1089.651+0.145-0.002=1089.794\ \ MHz$ (121)
which is larger than Eq.(114) by $3\%$.
(b) The Lamb shift of $He^{+}$ atom has been measured as (quoted from 27
(27)):
$L_{He^{+}}^{exp}(2S-2P)=14041.13(17)\ \ MHz$ (122)
Similar to the case of hydrogen atom but with $Z=2$ and
$b_{He^{+}}=\frac{m_{e}}{m_{\alpha}}=0.0001371$, we find
$\displaystyle\Delta E_{He^{+}}^{Rad<S>}(2S-2P)$ $\displaystyle=$
$\displaystyle 1.252680693\times 10^{10}\ \ Hz$ $\displaystyle\Delta
E_{He^{+}}^{Rad<V>}(2S-2P)$ $\displaystyle=$ $\displaystyle 2.023083608\times
10^{10}\ \ Hz$ $\displaystyle\Delta E_{He^{+}}^{Rad<S+V>}(2S-2P)$
$\displaystyle=$ $\displaystyle 1.369980830\times 10^{10}\ \ Hz$
$\displaystyle\Delta E_{He^{+}}^{Rad<SV>}(2S-2P)$ $\displaystyle=$
$\displaystyle 1.636521214\times 10^{10}\ \ Hz$ (123)
As in the case of $H$ atom, we take the $<S+V>$ scheme and add
$\displaystyle\Delta E_{He^{+}}^{recoil-2}(2S-2P)$ $\displaystyle=$
$\displaystyle-2.165\ \ kHz$ (124) $\displaystyle\Delta
E_{He^{+}}^{NS}(2S-2P)$ $\displaystyle=$ $\displaystyle 4.514\ \ MHz$ (125)
($r_{\alpha}\simeq 1.2fm$), to find the theoretical value:
$L_{He^{+}}^{theor.}(2S-2P)=13704.220\ \ MHz$ (126)
which is smaller than Eq.(122) by $2.41\%$.
(c) The following energy-level difference is related to the ”hyper Lamb shift
(HLS)” 10 (10):
$\Delta_{H}^{exp}\equiv
E_{H}(4S)-E_{H}(2S)-\frac{1}{4}[E_{H}(2S)-E_{H}(1S)]=4797.338(10)\ \ MHz$
(127)
Theoretically, now Eq.(102) makes the main contribution:
$\Delta E_{H}^{RDE}[(4S)-\frac{5}{4}(2S)+\frac{1}{4}(1S)]=3923.95\ \ MHz$
(128)
(The notation in parenthesis is self-evident). Eq.(103) and Eq.(105)
contribute
$\Delta E_{H}^{recoil-1}[(4S)-\frac{5}{4}(2S)+\frac{1}{4}(1S)]=-4.186\ \ MHz$
(129)
and
$\displaystyle\Delta E_{H}^{Rad<S>}$ $\displaystyle=$ $\displaystyle
451.229097\ \ MHz$ $\displaystyle\Delta E_{H}^{Rad<S+V>}$ $\displaystyle=$
$\displaystyle 529.288296\ \ MHz$ $\displaystyle\Delta E_{H}^{Rad<SV>}$
$\displaystyle=$ $\displaystyle 675.907131\ \ MHz$ $\displaystyle\Delta
E_{H}^{Rad<V>}$ $\displaystyle=$ $\displaystyle 903.266275\ \ MHz$ (130)
respectively. Adding a small contribution from Eq.(107)
$\Delta E_{H}^{NS}[(4S)-\frac{5}{4}(2S)+\frac{1}{4}(1S)]=0.1270967854\ \ MHz$
(131)
we get
$\displaystyle\Delta_{H}^{Theor.<S>}$ $\displaystyle=$ $\displaystyle
4371.120197\ \ MHz$ $\displaystyle\Delta_{H}^{Theor.<S+V>}$ $\displaystyle=$
$\displaystyle 4449.179396\ \ MHz$ $\displaystyle\Delta_{H}^{Theor.<SV>}$
$\displaystyle=$ $\displaystyle 4595.798231\ \ MHz$
$\displaystyle\Delta_{H}^{Theore.<V>}$ $\displaystyle=$ $\displaystyle
3923.95-4.186+903.266275+0.1271=4823.1574\ \ MHz$ (132)
The $<V>$ scheme is only larger than Eq.(127) by $0.54\%$. All other schemes
would be too small. So we guess that for $S$ states $<V>$ scheme is better
than $<S>$ scheme.
(d) The following energy-level difference was also measured as 10 (10):
${\Delta^{\prime}}_{H}^{exp}\equiv
E_{H}(4D_{5/2})-E_{H}(2S)-\frac{1}{4}[E_{H}(2S)-E_{H}(1S)]=6490.144(24)\ \
MHz$ (133)
Theoretically, Eq.(102) also makes the main contribution:
$\Delta E_{H}^{RDE}[(4D_{5/2})-\frac{5}{4}(2S)+\frac{1}{4}(1S)]=5747.92\ \
MHz$ (134)
while
$\Delta E_{H}^{recoil-1}[(4D_{5/2})-\frac{5}{4}(2S)+\frac{1}{4}(1S)]=-4.18611\
\ MHz$ (135)
$\Delta E_{H}^{recoil-2}[(4D_{5/2})]=\alpha^{4}m_{e}(5.446170255\times
10^{-4})^{2}(\frac{1}{3}-\frac{2}{5})=-6.9283\ \ kHz$ (136)
are all small, we will have
$\displaystyle{\Delta^{\prime}}E_{H}^{rad<S>}[(4D_{5/2})-\frac{5}{4}(2S)+\frac{1}{4}(1S)]$
$\displaystyle=$ $\displaystyle 302.088631\ \ MHz$
$\displaystyle{\Delta^{\prime}}E_{H}^{rad<V>}[(4D_{5/2})-\frac{5}{4}(2S)+\frac{1}{4}(1S)]$
$\displaystyle=$ $\displaystyle 700.843464\ \ MHz$
$\displaystyle{\Delta^{\prime}}E_{H}^{rad<S+V>}[(4D_{5/2})-\frac{5}{4}(2S)+\frac{1}{4}(1S)]$
$\displaystyle=$ $\displaystyle 369.124660\ \ MHz$
$\displaystyle{\Delta^{\prime}}E_{H}^{rad<SV>}[(4D_{5/2})-\frac{5}{4}(2S)+\frac{1}{4}(1S)]$
$\displaystyle=$ $\displaystyle 500.131264\ \ MHz$ (137)
Finally, the nucleus size effect gives
${\Delta^{\prime}}E_{H}^{NS}[(4D_{5/2})-\frac{5}{4}(2S)+\frac{1}{4}(1S)]=11.62027752\times
10^{5}(\frac{1}{4}-\frac{5}{4}\times\frac{1}{8})=0.10894\ \ MHz$ (138)
In sum, we have
$\displaystyle{\Delta^{\prime}}_{H}^{<S>}$ $\displaystyle=$
$\displaystyle{\Delta^{\prime}}E_{H}^{RDE}+{\Delta^{\prime}}E_{H}^{recoil-1}+{\Delta^{\prime}}E_{H}^{recoil-2}+{\Delta^{\prime}}E_{H}^{rad<S>}+{\Delta^{\prime}}E_{H}^{rad<NS>}=6045.925\
\ MHz$ $\displaystyle{\Delta^{\prime}}_{H}^{<V>}$ $\displaystyle=$
$\displaystyle 6444.679\ \ MHz$ $\displaystyle{\Delta^{\prime}}_{H}^{<S+V>}$
$\displaystyle=$ $\displaystyle 6112.961\ \ MHz$
$\displaystyle{\Delta^{\prime}}_{H}^{<SV>}$ $\displaystyle=$ $\displaystyle
6243.967\ \ MHz$ (139)
which are smaller than the experimental value (133) by $6.8\%,\ 0.7\%,\ 5.8\%$
and $3.8\%$ respectively. .
(e) Experimentally, the combination of Eq.(127) with Eq.(133) yields:
${\Delta^{\prime\prime}}_{H}^{exp}\equiv E(4D_{5/2})-E(4S_{1/2})=1692.806\ \
MHz$ (140)
Then, theoretically, we have
$\displaystyle{\Delta^{\prime\prime}}_{H}^{RDE}(4D_{5/2}-4S)$ $\displaystyle=$
$\displaystyle 1.823886903\times 10^{9}\ \ Hz$ (141)
$\displaystyle{\Delta^{\prime\prime}}_{H}^{recoil-1}(4D_{5/2}-4S)$
$\displaystyle=$ $\displaystyle 1.1008\ \ Hz$ (142)
$\displaystyle{\Delta^{\prime\prime}}_{H}^{recoil-2}(4D_{5/2}-4S)$
$\displaystyle=$ $\displaystyle-6.9283\ \ kHz$ (143)
$\displaystyle{\Delta^{\prime\prime}}_{H}^{NS}(4D_{5/2}-4S)$ $\displaystyle=$
$\displaystyle-0.0181605862\ \ MHz$ (144)
and
$\displaystyle{\Delta^{\prime\prime}}_{H}^{rad<S>}(4D_{5/2}-4S)$
$\displaystyle=$ $\displaystyle-149.1404661\ \ MHz$ (145)
$\displaystyle{\Delta^{\prime\prime}}_{H}^{rad<V>}(4D_{5/2}-4S)$
$\displaystyle=$ $\displaystyle-202.4228107\ \ MHz$ (146)
$\displaystyle{\Delta^{\prime\prime}}_{H}^{rad<S+V>}(4D_{5/2}-4S)$
$\displaystyle=$ $\displaystyle-160.1636366\ \ MHz$ (147)
$\displaystyle{\Delta^{\prime\prime}}_{H}^{rad<SV>}(4D_{5/2}-4S)$
$\displaystyle=$ $\displaystyle-175.7758676\ \ MHz$ (148)
Altogether, we have
$\displaystyle{\Delta^{\prime\prime}}_{H}^{theore.<S>}(4D_{5/2}-4S)$
$\displaystyle=$ $\displaystyle 1674.721349\ \ MHz$
$\displaystyle{\Delta^{\prime\prime}}_{H}^{theore.<S+V>}(4D_{5/2}-4S)$
$\displaystyle=$ $\displaystyle 1663.716339\ \ MHz$
$\displaystyle{\Delta^{\prime\prime}}_{H}^{theore.<SV>}(4D_{5/2}-4S)$
$\displaystyle=$ $\displaystyle 1648.104108\ \ MHz$
$\displaystyle{\Delta^{\prime\prime}}_{H}^{theore.<V>}(4D_{5/2}-4S)$
$\displaystyle=$ $\displaystyle 1621.439105\ \ MHz$ (149)
which are smaller than Eq.(140) by $1.1\%,\ 1.7\%,\ 2.6\%$ and $4.2\%$
respectively.
(f) It’s time to go back to the precision data of $2S-1S$ transition in
hydrogen atom as discussed in section II. Rewrite Eq.(7) as (see also 43
(43)):
$\Delta E_{H}^{exp}(2S-1S)=2.46606141318734\times 10^{15}\ \ Hz$ (150)
Theoretically, we have [see Eq.(17)]:
$\Delta E_{H}^{RDE}(2S-1S)=2.466067984\times 10^{15}\ \ Hz$ (151)
$\Delta E_{H}^{recoil-1}(2S-1S)=22.32598676\ \ MHz$ (152)
$\displaystyle\Delta E_{H}^{rad<S>}(2S-1S)$ $\displaystyle=$
$\displaystyle-5142.081146\ \ MHz$ $\displaystyle\Delta
E_{H}^{rad<S+V>}(2S-1S)$ $\displaystyle=$ $\displaystyle-5765.958928\ \ MHz$
$\displaystyle\Delta E_{H}^{rad<SV>}(2S-1S)$ $\displaystyle=$
$\displaystyle-6835.535314\ \ MHz$ $\displaystyle\Delta E_{H}^{rad<V>}(2S-1S)$
$\displaystyle=$ $\displaystyle-8541.095068\ \ MHz$ (153)
$\Delta E_{H}^{NS}(2S-1S)=11.62027752\times
10^{5}(\frac{1}{8}-1)=-1.016774283\ \ MHz$ (154)
If taking the value of $\Delta E_{H}^{rad}(2S-1S)$, we get
$\displaystyle\Delta E_{H}^{theore.<S>}(2S-1S)$ $\displaystyle=$
$\displaystyle 2.466062836\times 10^{15}\ \ Hz$ $\displaystyle\Delta
E_{H}^{theore.<S+V>}(2S-1S)$ $\displaystyle=$ $\displaystyle 2.466062239\times
10^{15}\ \ Hz$ $\displaystyle\Delta E_{H}^{theore.<SV>}(2S-1S)$
$\displaystyle=$ $\displaystyle 2.466061169\times 10^{15}\ \ Hz$
$\displaystyle\Delta E_{H}^{theore.<V>}(2S-1S)$ $\displaystyle=$
$\displaystyle 2.466059464\times 10^{15}\ \ Hz$ (155)
They are larger than Eq.(150) by $1450\ MHz,\ 826\ MHz$ and smaller than
Eq.(150) by $244\ MHz,\ 1949\ MHz$ respectively. Or, their discrepancies are
$+5.9\times 10^{-7},\ +3.3\times 10^{-7},\ -1.0\times 10^{-7},\ -7.9\times
10^{-7}$, respectively. This discrepancy is basically stemming from the
uncertainty in the calculation of $\Delta E_{H}^{rad}(2S-1S)$.
(g) Let us turn to the isotope-shift of $2S-1S$ transition. Rewrite Eq.(8) as
$\Delta E_{D-H}^{exp}(2S-1S)=6.70994337\times 10^{11}\ \ Hz$ (156)
Theoretically, rewrite Eqs.(19) and (20) as
$\Delta E_{D-H}^{RDE}(2S-1S)=6.7101527879\times 10^{11}\ \ Hz$ (157)
and
$\Delta E_{D-H}^{recoil-1}(2S-1S)=-11.176\ \ MHz$ (158)
$\displaystyle\Delta E_{D-H}^{rad<S>}(2S-1S)$ $\displaystyle=$
$\displaystyle-1.399158\ \ MHz$ $\displaystyle\Delta E_{D-H}^{rad<V>}(2S-1S)$
$\displaystyle=$ $\displaystyle-2.324028\ \ MHz$ $\displaystyle\Delta
E_{D-H}^{rad<S+V>}(2S-1S)$ $\displaystyle=$ $\displaystyle-1.568915\ \ MHz$
$\displaystyle\Delta E_{D-H}^{rad<SV>}(2S-1S)$ $\displaystyle=$
$\displaystyle-1.859945\ \ MHz$ (159)
$\Delta E_{D-H}^{NS}(2S-1S)=-5.11384949\ \ MHz$ (160)
Altogether, we find [using $<V>$ scheme in Eq.(159)]:
$\Delta E_{D-H}^{theore.<V>}(2S-1S)=6.709966701\times 10^{11}\ \ Hz$ (161)
which is larger than Eq.(156) by $2.333\ MHz$ or only $3.5\times 10^{-6}$.
Evidently, even Eq.(157) solely deviates from Eq.(156) by $3\times 10^{-5}$
only. And as expected, the different schemes for $\Delta E_{D-H}^{rad}(2S-1S)$
have little influence on the theoretical value, because any one of Eq.(159) is
much smaller than the nucleus size effect Eq.(160)
($r_{N}^{D}=r_{d}=2.115fm$).
(h) Finally, the so-called absolute Lamb-shift of $1S$ state in hydrogen atom
was determined by Weitz et al.10 (10) from the measured value Eq.(127) or
(133). In our notation, using Eq.(133), we will write it as follows:
$\begin{array}[]{ll}L_{H}(1S)=&4\\{{\Delta^{\prime}}_{H}^{exp}-\Delta
E_{H}^{RDE}[(4D_{5/2})-\frac{5}{4}(2S)+\frac{1}{4}(1S)]-\Delta
E_{H}^{recoil-1}[(4D_{5/2})-\frac{5}{4}(2S)+\frac{1}{4}(1S)]\\\ &-\Delta
E_{H}^{recoil-2}(4D_{5/2})+\frac{5}{4}L_{H}(2S)-L_{H}(4D_{5/2})\\}\end{array}$
(162)
Here the Lamb shift of $2S$ state $L_{H}(2S)$ can be determined from the
experimental value of Eq.(114) with $L_{H}(2P_{1/2})$ being calculated from
Eq.(105):
$L_{H}(2S)=L_{H}^{exp}(2S-2P)-\Delta
E_{H}^{recoil-2}(2S-2P_{1/2})+L_{H}(2P_{1/2})=1040.901\ \ MHz$ (163)
And $L_{H}(4D_{5/2})$ can also be calculated from Eq.(105), so
$L_{H}(1S)=8188.478\ \ MHz$ (164)
which is in agreement with $8172.874(60)\ MHz$ given by 10 (10) within an
accuracy $\lesssim 0.2\%$. If we use Eq.(127) to derive $L_{H}(1S)$, we would
have to calculate $L_{H}(4S)$ which is much larger than $L_{H}(4D_{5/2})$ and
its derivation from Eq.(105) seems not reliable. Similarly, the theoretical
value of $L_{H}(1S)$ turns out to be
$L_{H}^{theore.}(1S)=\Delta E_{H}^{rad}(1S)+\Delta E_{H}^{NS}(1S)$ (165)
with $\Delta E_{H}^{NS}(1S)=0.14525347\ MHz$. However, the value of $\Delta
E_{H}^{rad}(1S)$ strongly depends on the scheme we used in Eq.(105), which
must be narrowed in a high-loop calculation. The theoretical prediction was
given in 27 (27) as:
$L_{H}^{theor.}(1S)=8172754(14)(32)\ kHz$ (166)
Further discussions can be found in Refs. 5 ; 44 (44, 45).
## IX Summary and Discussion
The remarkable progress of the experimental research on energy-level
differences in hydrogenlike atoms has been making this field an ideal
theoretical laboratory for physics:
(a) The inevitable and successful use of reduced Dirac equation (RDE) to
hydrogenlike atoms, especially to the isotope-shift of $2S-1S$ transition as
reflected by Eqs.(156) through (161), is by no means an accidental fortune. It
implies that the argument in section III for introducing RDE, Eq.(41), is
correct to a high accuracy. In particular, the basic principle of invariance
under space-time inversion Eq.(42) (with original mass $m$) could remain valid
even for a noninertial frame. This implication has a far-reaching consequence
that a generalization at the above symmetry to a localized curved space-time
may be served as a possible road to quantize the general theory of relativity
16 (16).
However, there are two realizations of potential $V$ under the space-time
inversion, Eq.(35)(”scalar” type) and Eq.(42)(”vector” type). While Eq.(42)
does dominant in an atom like $H$ with $m_{p}\gg m_{e}$, the remaining
discrepancy of $2.333\ MHz$ between theory and experiment [Eq.(161) versus
Eq.(156)] strongly hints that an important and subtle effect had been ignored.
(To consider the contribution of the deuteron polarizability merely accounts
for about $20\ kHz$ 6 ). We think what neglected must be a tiny excitation of
antiparticle field in the nucleus due to its interaction with electron in the
CMCS. So when we reduce the degrees of freedom of two-body system from two to
one, the RDE should be modified to take account of the tiny mixture of
”scalar” potentials (see the page note after Eq.(42)). We don’t know how to
improve $RDE$ yet. However, an experimental evidence for the above conjecture
could be the following prediction: The discrepancy between present theory
(with RDE) and experiment must be smaller for the isotope shift in $2S-1S$
transition of atoms ${}^{4}He$ and ${}^{3}He$ than that of atoms $H$ and $D$.
Recently, by using Dirac’s method, Marsch rigorously solved the hydrogen atom
as a two-Dirac particle system bound by Coulomb force 34 (34). His solutions
are composed of positive and negative pairs, corresponding respectively to
hydrogen and anti-hydrogen as expected. However, surprisingly, in the hydrogen
spectrum, besides the normal type-1 solution with reduced mass $\mu$, there is
another anomalous type$-2$ solution with energy levels:
${E^{\prime}}_{n}=Mc^{2}-2\mu c^{2}+\frac{1}{2}\mu
c^{2}(\frac{\alpha}{n})^{2}+\cdots\ (n=1,2,\ldots)$ and ”strange enough, the
type$-2$ ground state $(n=1)$ does not have lowest energy but the continuum
$(n=\infty)$”. In our opinion, based on what we learnt from the Dirac equation
and RDE, these anomalous solutions imply a positron moving in the field of
proton. So all discrete states with energy ${E^{\prime}}_{n}$ are actually
unbound, they should be and can be ruled out in physics either by the ”square
integrable condition” or the ”orthogonality criterion” acting on their
rigorous wave functions (for one-body Dirac equation, see 35 (35), also
p.$28-31$, $50$ of 36 (36)). On the other hand, all continuum states
($n=\infty$) with energies lower than $Mc^{2}-2\mu c^{2}$ correspond to
scattering wave functions with negative phase shifts , showing the repulsive
force between positron and proton. (see 37 (37), section 1.5 in 36 (36) or
section 9.5 of 15 (15)). Marsch’s discovery precisely reflects two things: (a)
the negative energy state of a particle just describes its antiparticle state.
(b) The Coulomb potential allows a complete set of solutions comprising of two
symmetric sectors,hydrogen and antihydrogen.In the hydrogen sector, the proton
remains unchanged regardless of the changing process of electron into positron
under the Coulomb interaction.
The above particle-antiparticle symmetry (including Eq.(42) showing the
unequal treatment between electron and nucleus), together with the parity
symmetry, is hidden in the Dirac’s four-component theory in covariant form so
they were overlooked to some extent in the past. The advantage or flexibility
of two-component noncovariant form of Dirac equation or RDE (as discussed in
this paper) lies in the fact that the above two symmetries become accurate and
so easily to be extended (or violated) in an explicit manner. For
completeness, let us stress again that for antiparticle, one should use the
momentum and energy operators being ${\mathbf{p}}_{c}=i\nabla$ and
$E_{c}=-i\frac{\partial}{\partial t}$ versus ${\mathbf{p}}=-i\nabla$ and
$E=i\frac{\partial}{\partial t}$ for particle as required by the space-time
inversion symmetry. The historical mission of the conception to imagine the
positron as a ”hole” in the sea of negative energy electrons is already over.
Since the CPT invariance had been further verified 39 (39), the relation
between a particle $|a\rangle$ and its antiparticle $|\bar{a}\rangle$ is well-
established as: 444To our knowledge, the correct definition, Eq.(167), was
first given by T. D. Lee and C. S. Wu at Ann. Rev. Nucl. Sci. 15, 381(1965).
See also G. J. Ni at J. Fudan Univ. (Natural Science) No.3-4, 125(1974).
$|\bar{a}\rangle=CPT|a\rangle$ (167)
with their wave-functions (in free motion) being respectively:
$\langle{\bf x},t|a\rangle\sim\exp[\frac{i}{\hbar}({\bf p}\cdot{\bf x}-Et)]$
(168) $\langle{\bf x},t|\bar{a}\rangle\sim\exp[-\frac{i}{\hbar}({\bf
p}\cdot{\bf x}-Et)]$ (169)
Note that in Eqs.(168) and (169), they have the same momentum $\bf p$ and
positive energy $E$. Either a newly defined space-time inversion (${\bf
x}\to-{\bf x},\,t\to-t$) or a simple change of $i\to-i$ will transform
Eq.(168) into Eq.(169) (or vice versa).
(b) Throughout this paper, the electron bound in an atom is just treated like
a stationary ”ball” with nucleus at its center and having a (Bohr) radius
($\sim 1/\alpha m_{e}$). However, it is in an off-mass-shell state (In some
sense, our atom model is just the opposite to J. J. Thomson’s atom model 100
years ago). In fact, the electron’s mass is reduced suddenly from $m_{e}$ to
$\mu$ in the RMCS when it is captured by a nucleus at the far remote orbit
with quantum number $n\longrightarrow\infty$ and further reduced to
$\mu+\delta\mu\simeq\mu-\frac{Z^{2}\alpha^{2}}{2n^{2}}\mu$ until $n$
decreasing to the lowest limit $n=1$. The Lamb shift should be viewed as a
further modification on the mass of an off-mass-shell electron due to
radiative correction.
Notice that the parameter $Q^{2}$ in the vertex function, Eq.(87), means the
square of (three-dimensional) momentum transfer when a free electron is on its
mass-shell and collides with some other particle as discussed in Ref.8b . By
contrast, now $Q^{2}$ exhibits itself as an effective potential of
$\delta$-function type exerted by the nucleus to the bound (and so off-mass-
shell) electron as shown by Eq.(92). To bind an electron to a nucleus is a
nonperturbative effect. Hence we can understand why the discrepancy between
$\zeta^{<S>}$ (calculated by perturbative QED at one-loop order) and
$\zeta^{<V>}$ (evaluated via nonperturbative Virial theorem) is so large.
Fortunately, they lead to discrepancies in the calculated values of Lamb shift
being not so large as shown in Section VIII. When $\zeta^{<V>}$ or
$\zeta^{<S+V>}$ (or $\zeta^{<SV>}$) is substituted into the Eq.(105) which is
derived from perturbative ($L=1$) theory, we should always be aware of some
theoretical inconsistency in such a semi-empirical treatment. But as a whole,
we believe that the concept of Lamb shift as an off-mass-shell effect in
covariant QED is basically correct.
(c) For a free on-mass-shell electron, its charge square $e_{R}^{2}$ will
increase with the increase of $Q^{2}$ as shown by Eq.(72) (with
$\mu_{3}=m_{e},\ q^{2}=-Q^{2}$) and was calculated in detail in 8b ,
coinciding with the experimental data. Note that, however, the Ward identity
$Z_{1}=Z_{2}$ had been used. An interesting question arises for a bound
electron: as its $e_{R}^{2}$ is not a function of $Q^{2}$, will $e_{R}^{2}$
change with the variation of the quantum number $n$? To answer this question,
let us put Ward identity aside for a while and write down the renormalized
$\alpha_{R}=\frac{e_{R}^{2}}{4\pi}$ as
$\alpha_{R}=\frac{Z_{2}^{2}}{Z_{1}^{2}}Z_{3}\alpha\longrightarrow\frac{Z_{2}^{2}}{Z_{1}^{2}}\alpha$
(170)
Let us work in the CMCS, so $Z_{2}=\frac{1}{1-B}$ and $B$ is shown in Eq.(53)
but with $\mu$ replaced by $m_{e}=m$. Similarly, $Z_{1}$ is given by the first
part of Eq.(87) with $\mu\longrightarrow m$:
$Z_{1}\simeq
1+\frac{\alpha}{4\pi}[\frac{11}{2}-\ln\frac{m^{2}}{\mu_{1}^{2}}-3\xi+4(1+\xi)\ln\xi]$
(171)
where the off-mass-shell parameter $\xi$ in CMCS is defined by
$p^{2}=m^{2}(1-\xi)=m^{2}(1-\eta-\zeta^{\prime})=m^{2}(1-\eta)-m^{2}\zeta^{\prime}=\mu^{2}-m^{2}\zeta^{\prime}=\mu^{2}(1-\zeta)$
(172)
with
$m^{2}(1-\eta)=\mu^{2},\ \ \eta=1-\frac{\mu^{2}}{m^{2}},\ \
\zeta^{\prime}=\frac{\mu^{2}}{m^{2}}\zeta$ (173)
and $\zeta$ is exactly that in Eq.(88). If we ignore the dependence of $(1-B)$
on $\zeta$, Eq.(170) would give ($\zeta\ll 1$):
$\alpha_{R}=\alpha[1+\frac{2\alpha}{\pi}\ln(1+\frac{\zeta}{\eta})]$ (174)
after renormalization by adjusting the arbitrary constant $\mu_{1}$ so that
$\alpha_{R}|_{\zeta\rightarrow 0}=\alpha$ (175)
which connects to the Thomson limit $\alpha_{R}|_{Q\rightarrow 0}=\alpha$ for
a free electron continuously but not smoothly. Then for two lowest bound
states with $n=1$ and $n=2$, we would have (in $<V>$ scheme):
$\alpha_{R}|_{n=1}\simeq\alpha(1.000433832),\ \
\alpha_{R}|_{n=2}\simeq\alpha(1.0001123)$ (176)
This would modify the Bohr energy level in hydrogenlike atom A to
$\tilde{E}_{A}^{Bohr}(n)=-\frac{Z^{2}\alpha_{n}^{2}}{2n^{2}}\mu_{A}$ (177)
and make an extra contribution to the isotope-shift as
$\Delta\tilde{E}_{D-H}^{Bohr}(2S-1S)\simeq 726\ \ MHz$ (178)
which is definitely excluded by the experiment. Hence the above consideration
from Eq.(170) till Eq.(178) is wrong. We learn concretely once again that the
Ward identity $Z_{1}=Z_{2}$ is valid not only for an electron on its mass-
shell, but also for off-mass-shell case. Thus we use the same value of
$\alpha$ throughout the whole calculation.
(d) In Ref.8b , using our RRM and new renormalization group equation (RGE) for
QCD derived from it and keeping all masses of 6 quarks ($m_{c}=1.031\;GeV$,
$m_{b}=4.326\;GeV$, $m_{t}=175\;GeV$, $m_{s}=200\;MeV$, $m_{u}=8\;MeV$,
$m_{d}=10\;MeV$), we calculated the strong coupling constant
$\alpha_{s_{i}}(Q)$ for $i=u,d,s,c,b$ respectively. Their running curves
(starting from the common renormalization point $\alpha_{s}(M_{Z})=0.118$)
follow the trend of experimental data (as shown on p.158 of 39 (39)) quite
well but separate at the low $Q$ region. Interesting enough, each of them
rises to a maximum $\alpha_{s_{i}}^{max}$ and then suddenly drops to zero at
$Q=0$ corresponding to a threshold energy scale $E_{i}^{th}$ which could be
explained as the excitation energy scale for breaking the quark pair. For
example, we find $E_{b}^{th}=1.13\;GeV$ which is just the hadronization energy
scale of Upsilon $\Upsilon(b\bar{b})$ against its dissociation into two
bosons. Experimentally, $M(\Upsilon(4s))-M(\Upsilon)=1.12\;GeV$ and
$\Upsilon(4s)\to B^{+}B^{-}$ or $B^{0}\bar{B}^{0}$. similarly,
$E_{c}^{th}=0.398\;GeV$ while $M(\psi(3770))-M(\psi(3097))=673\;MeV$ and
$\psi(3770)\to D^{+}D^{-}$ or $D^{0}\bar{D}^{0}$. it seems that
$E_{s}^{th}\sim 90\,MeV$ and $E_{u,d}^{th}\sim 0.4\,MeV$ are not so reliable
but still reasonable.
Actually, our calculation on QCD is backed by that on QED. In [8]b, using our
RRM and improved RGE, keeping all contributions from 9 charged leptons and
quarks we were able to calculate the running fine-structure constant
$\alpha_{R}(Q)$ from the renormalization point
$\alpha_{R}(Q)|_{Q=0}=\alpha=(137.036)^{-1}$ until it coincides with the
experimental value of $\alpha_{exp}(M_{Z})=(128.89)^{-1}$. We fitted quark’s
masses as mentioned above and found no further room left for extra charged
elementary particles (say, of 4th generation).
(e) In 1989, we had estimated the upper and lower bounds on Higgs mass $M_{H}$
by using a nonperturbative approach in QFT — the Gaussian effective potential
(GEP) method, yielding 40 (40):
$76\;GeV<M_{H}<170\;GeV$ (179)
Like many other authors, we were bothered a lot by divergences. After a deeper
study on the $\lambda\phi^{4}$ model by using our new RRM 8a , we restudied
this problem by combination GEP with RRM, yielding24 (24):
$M_{H}=138\;GeV$ (180)
This is not a upper or lower bound but a prediction based on the input of
experimental data:
$\begin{array}[]{l}M_{W}=80.359\;GeV,\;M_{Z}=91.1884\;GeV,\alpha^{-1}=\dfrac{4\pi}{g^{2}\sin^{2}\theta_{W}}=128.89,\\\
\sin^{2}\theta_{W}=\dfrac{{g^{\prime}}^{2}}{g^{2}+{g^{\prime}}^{2}}=0.2317\end{array}$
(181)
where $\theta_{W}$ is the weak mixing (Weinberg) angle. Because of getting rid
of all divergences, our calculation is well under control at every step. As
now the search for Higgs particle becomes so urgent experimentally but the
theoretical estimation about its mass still remains uncertain39 (39), we think
our approach with the prediction (180) deserves to be reconsidered.
(f) Moreover, our RRM can be used in $D+1$ space-time without limitation on
the space dimension $D$. A detailed analysis of sinh(sine)-Gordon models with
$D=1,2$ and $3$ (also using GEPM) is given by Ref.25 (25). Another example is
again the Lamb shift but calculated by QED in noncovariant form and by using
RRM similar to that in this paper, see Appendix (26 (26), see also the
Appendix 9A in 15 (15)). The theoretical value (A.20) seems better than (121),
showing that for dealing with the Lamb shift, the noncovariant theory may be
more suitable than the covariant one at least in the lowest order.
(g) Previously, the theories for Lamb shift or generally for calculating
energy levels in hydrogenlike atoms are rather complicated as reviewed in refs
1 ; 2 and 27 (27), some of them have been discussed in the Appendix of this
paper. For further clarity, let us try to summarize the main obstacles, or
challenges in four points:
(1) Different masses of nuclei must be taken into account;
(2) Relativistic effects of the electron (not nucleus) are important;
(3) In calculating radiative corrections, the divergence becomes severer and
severer with the increase of loop number;
(4) Since nuclei’s properties are different from one atom to another, to treat
each atom as a two-body system individually would be a daunting task, it
couldn’t be rigorous eventually too. This can be clearly seen from the recent
work by Marsch 34 (34).
Facing these challenges and learning from lessons and experiences of previous
authors, we see that the clue point is to replace the electron mass $m$ by
reduced mass $\mu$ and work in the noninertial frame (RMCS) throughout the
entire calculation. As is well known, this can be handled in nonrelativistic
QM by a mathematical trick but is impossible in relativistic case. So what we
need is a new understanding on the essence of special relativity — the
invariance (of theory) under the (newly defined) space-time inversion in one
inertial frame. Then we are able to claim the same invariance in the RMCS with
$\mu$ replacing $m$ for establishing the RDE, ignoring a small centripetal
acceleration of the nucleus in $CMCS$ (see page note after Eq.(42)). The
approximation in $RDE$ is some price paid for the much bigger gain—improving
the original Dirac equation (unable to treat different nuclei) and avoiding
the confusion in QED calculation because of the entanglement of two frames:
CMCS with RMCS ( i.e., the radiative corrections are entangled with the recoil
effect as we can see from previous literatures). In some senses, we jump over
obstacles (1) ,(2) and (4)at the least labor cost (by constructing RDE). In
the meantime, we hope RDE would help to ease the difficulty in point (3). And
it’s a great pleasure to see that the essential correctness of our
understanding has been validated by Marsch’s work as well as puzzles raised in
his paper 34 (34). Please see also Ref 41 (41).
As to challenge (3), only after we puzzled over the ”divergence” for decades,
could we suddenly realize that we misread its implication as a ”large number”.
Rather, it means the ”uncertainty”. Let us look at the calculation in section
IV again. Previously, many authors treated the divergent integral I in Eq.(50)
by different tricks of regularization , arriving at Eq.(56). Because both A
and B are divergent, it was thought that the original mass ($\mu$ here) does
receive some radiative corrections (via the self-energy diagrams in Fig 2(a)
and (b)) and becomes a ”renormalized” mass ($\mu_{R}$ here). The latter should
be the observed mass in experiment or physical mass (of electron). So the
original mass was called as the ”bare mass”, which was written into the
Lagrangian density as an input parameter of QFT. Then in constructing Feynman
diagrams of certain perturbative calculation, one needs to further introduce
some (divergent) ”counter terms” for cancelling the divergence stemming from
the bare mass. Based on that understanding, the renormalization factor for
wavefunction, $Z_{2}$ in Eq.(55), would be a divergent quantity too (in sharp
contrast to here Eq.(59) being a fixed and finite number). Previously, In
Eq.(72), while the $e_{R}$ on the left handed side is the observed charge
which should be finite, the $e$ on the right handed side was regarded as a
”bare charge” which, together with the $Z_{3}$, was a divergent quantity. (see
Fig. 7.8 in 23 (23). By contrast, here both $Z_{3}$ and $e$ are finite.
Actually, here $e$ is defined as the physical charge observed at the Thomson
limit in experiment).
In our opinion, the reason why we encountered so many superfluous troubles in
the past is because we overlooked what Bethe said in 194729 (29). Please read
his words quoted after Eq. (A.2) in the Appendix. Let us explain via our
Eq.(46). The (reduced) mass ($\mu$) already contains some contributions from
self-energy diagrams like Fig. 2(a) and (b). When we evaluate the (divergent)
integral, Eq.(50), trying to find the radiative corrections on the electron,
the latter is bound to confuse with that already contained in the mass. In
other words, the dividing line between them is blurred inevitably. The
emergence of explicit divergence is essentially a warning that the effect you
want to evaluate has been entangled with the mass itself, rendering both of
them uncertain. Hence the aim of so-called renormalization is nothing but to
redraw the dividing line between them such that the values of mass
(reconfirmed by the experiment) and the new effect (e.g., the mass increment
when the electron is moving off-mass-shell, Eq.(61)) can be clearly separated.
In short, what we have been learning in the past decade is: At the level of
QM, in the Hamiltonian like Eq. (A.1), the parameters $m$ and $e$ can be
regarded as well-defined. But they are not so at the level of QFT. Once the
calculation is made beyond the tree level, i.e., with loop number $L\geq 1$,
the divergence occurs and the meaning of parameters becomes obscure
immediately.
We need to reconfirm all parameters contained in the Lagrangian density before
they can be linked with experiments. In this sense, a model of QFT is at most
an ”effective field theory”. According to the above point of view, we believe
that our RRM just provides a natural way to carry out these processes of
reconfirmation 8a , getting rid of divergences and ambiguities. Please see
also Ref 42 (42).
(h) Last, but not least, during the learning and teaching of graduate courses
on QFT for decades, we have been sharing the joy and puzzle with our students
all the time. We hope that the presentation of this paper could be useful as a
teaching reference to render the QFT course more understandable, interesting
and attractive.
## Acknowledgements
We thank S.Q. Chen, S.S. Feng, R.T. Fu, P.T. Leung, W.F. Lu, X.T. Song, F.
Wang, H.B. Wang, J. Yan, G.H. Yang and J.F. Yang for close collaboration and
helpful discussions. We are also indebted to referees whose comments provided
us opportunities to improve our manuscript.
## Appendix: Comparison Between Noncovariant and Covariant Theories for Lamb
Shift
1\. To our knowledge, the precision theory for Lamb shift was based on a
combination of noncovariant (nonrelativistic or old-fashioned) QED with
covariant (or relativistic) QED as discussed in Ref.27 (27). As explained
clearly by Sakurai in Ref.21 (21), in perturbative QFT of noncovariant form,
all virtual particles are ”on-mass-shell”. Here we wish to emphasize that a
rigorous reconfirmation procedure of mass parameter was often overlooked in
previous literatures. The theory for hydrogenlike atom begins with a
Hamiltonian:
$H_{0}=\frac{1}{2m}{\mathbf{p}}^{2}+\frac{1}{2m_{N}}{\mathbf{p}}^{2}-\frac{Z\alpha}{r}$
$None$
(${\mathbf{p}}=-i\nabla$, see Eq.(34) in 27 (27)). As Bethe 29 (29) first
pointed out that the effect of electron’s interaction with the vector
potential $\mathbf{A}$ of radiation field (see 21 (21),15 (15))
$H_{int}^{(1)}=\frac{e}{mc}{\mathbf{A}}\cdot{\mathbf{p}}$ $None$
should properly be regarded as already included in the observed mass $m_{obs}$
of the electron, which is denoted by $m$ in (A1). However, once a concrete
calculation is made with (A2) being taken into account, the divergence emerges
immediately. What does it mean? Mathematicians teach us that there are three
implications for a divergence:
(a)It is a dimensionless number; (b)It is a large number; (c)It is uncertain.
While we physicists often emphasized the point (b), we didn’t pay enough
attention to the points (a) and (c). We often talked about a quadratically (or
linearly) divergent integral without noticing that it has a dimension (say,
mass dimension) and thus meaningless in mathematics unless a mass parameter
(say, $m$) in the integral is already fixed as a mass ”unit” so that the
integral can be divided by $m^{2}$ (or $m$) to become dimensionless.
Alternatively, a logarithmically divergent integral is dimensionless and thus
unaffected by the choice of unit [like Eq.(50), see also Eq.(A6) below], it
just implies an uncertainty waiting to be fixed. The implication of
uncertainty of a divergence will never vanish even after we introduced a
cutoff by hand to curb it. For example, in a pioneering paper to explain the
Lamb shift, Welton (30 (30), see section 9.6B in Ref.15 (15)) encountered an
integral $I=\int_{\omega_{min}}^{\omega_{max}}\frac{d\omega}{\omega}$ with
$\omega$ being the (angular) frequency of virtual photon (vacuum fluctuation).
He simply set $\omega_{min}\sim mZ\alpha=Z/a$, ($a$ is Bohr radius) and
$\omega_{max}\sim m$ so that $I\simeq\ln(1/Z\alpha)=4.92$ (for $Z=1$) which
leads to an estimation of Lamb shift $L_{H}^{theor.}(2S_{1/2}-2P_{1/2})\simeq
668\ MHz$. If instead of Bohr radius, the lower cutoff is provided by the
electron binding energy, one should get $I\simeq\ln(Z\alpha)^{-2}$ and
$L_{H}^{theor.}\simeq 1336\ MHz$. (see Eq.(30) in 27 (27)). The above
arbitrariness just reflects what essential in a divergent integral is not its
large magnitude ($\ln(Z\alpha)^{-1}$ is merely of the order of 10) but its
uncertainty. So what important in handling the integral is not to curb (or to
hide) its divergence but let the divergence exhibits itself as some arbitrary
constants explicitly (as shown in section IV-VI). We will show later how to do
this way for noncovariant QED.
2\. While Eqs.(A1) and (A2) only describe a spinless particle, the electron
has spin which endows it with the relativistic nature as shown by
Eqs.(41)-(45). For two-particle system, based on Bethe-Salpeter equation, an
effective Dirac equation (EDE) was derived as shown by Eq.(23) in 27 (27).
When the electromagnetic field interaction is taken into account, the Breit
potential $V_{Br}$ was derived as shown by Eq.(35) in 27 (27). Then the total
Breit Hamiltonian reads (Eq.(36) in 27 (27)):
$H_{Br}=H_{0}+V_{Br}$ $None$
However, the electron mass $m^{\prime}$ (in our notation here) appeared in EDE
or $V_{Br}$ should be that in the Dirac equation, also that in the definition
of reduced mass $\mu=\frac{m^{\prime}m_{N}}{m^{\prime}+m_{N}}$, eventually
$m^{\prime}$ could be identified with the observed mass $m_{obs}$, which is
not equal to the $m$ in Eq.(A1). This is because besides (A2) there is an
extra interaction due to electron spin with the radiation field:
$H_{int}^{(2)}=\frac{ge\hbar}{4\mu
c}{\mathbf{\sigma}}\cdot\nabla\times{\mathbf{A}}$ $None$
($g=2\times 1.0011596522$ is the gyromagnetic ratio of electron, see
Eq.(9A.15) in 15 (15)). The difference between $m$ and $m^{\prime}$ will be
calculated in (A16) below. It turns out to be of the order of $\alpha m$ and
cannot be ignored at the level of QED, especially for the explanation of Lamb
shift. We guess this must be one of the reasons why all calculations based on
Eq.(A3) became so complicated.
3\. In noncovariant theory, the leading contribution to the Lamb shift comes
from the one-photon electron self-energy. The nomenclature here is different
from that in the covariant theory. Roughly speaking, so-called electron self-
energy often corresponds to the vertex function in covariant theory (Fig.2(d)
in this paper) or to Figs.8 and 11 in Ref.27 (27) and its evaluations have
extended over 50 years 31 (31). More precisely, it is identified with the
radiative insertions in the electron line and the Dirac form factor
contribution. Further contributions from the Pauli form factor and the vacuum
polarization 27 (27) will add to a theoretical value of classic Lamb shift
being $1050.559\ MHz$. If taking more high-order corrections into account, the
theoretical value coincides with the experimental value $1057.845\ MHz$ rather
accurately (see Table 20 in 27 (27)). However, the above calculation looks
quite complicated due to two reasons: (a) The difficulty of dealing with two
masses in two coordinate systems, the electron mass $m$ and the reduced mass
$\mu$; (b) The introduction of an auxiliary parameter $\sigma\
[m(Z\alpha)\gg\sigma\gg m(Z\alpha)^{2}]$ to separate the radiative photon
integration region into two parts. In the low momentum region, the Bethe
Logarithm 32 (32) in noncovariant form makes the main contribution. In the
high momentum region, the evaluation is resorting to some relativistic
covariant form 22 (22). Then two expressions are matched together to get the
correct result. It seems to us that the matching trick used is doubtful
because both ultraviolet and infrared divergences were ambiguously handled by
some cutoff which missed the main point of renormalization—to reconfirm the
mass parameter in the presence of radiative corrections as shown in section IV
(covariant form) or below.
4\. A simple calculation for Lamb shift in noncovariant form was proposed in
Ref.26 (26) (see also Appendix 9A of Ref.15 (15)). Consider the self-energy
diagram of an electron with reduced mass $\mu$ and (three-dimensional)
momentum $\mathbf{p}$ in the RMCS of a hydrogenlike atom. Similar to Fig.2(a),
but also different in the virtual state, now a photon has energy
$\omega_{k}=k=|{\mathbf{k}}|$ while the electron has momentum
${\mathbf{q}}={\mathbf{p}}-{\mathbf{k}}$ and energy
$\varepsilon_{q}=\frac{1}{2\mu}q^{2}$. The electron in plane-wave state
$|{\mathbf{p}}>$ has two interactions with the radiative field at each vertex
as shown by (A2) ($m\rightarrow\mu$) and (A4), acquiring an increase in energy
respectively (see FIG. 3):
$\Delta
E_{p}^{(j)}=\sum_{i}\frac{|<i|H_{int}^{(j)}|{\mathbf{p}}>|^{2}}{\varepsilon_{p}-\varepsilon_{i}},\quad(j=1,2)$
$None$
Here $\varepsilon_{i}=\varepsilon_{q}+\omega_{k}$ is the energy of the
intermediate virtual state $|i>$. Simple evaluation leads to
$\Delta E_{p}^{(1)}=-\frac{\alpha
p}{\pi\mu}\int_{-1}^{1}d\eta(1-\eta^{2})I,\quad
I=\int_{0}^{\infty}\frac{dk}{k+\xi}$ $None$
where $\eta=\cos\theta$ with $\theta$ being the angle between $\mathbf{k}$ and
$\mathbf{p}$, $\xi=2(\mu-p\eta)$. Like Eq.(50), we take partial derivative of
the divergent integral $I$ with respect to $\xi$ (then the integration of $k$)
and integrate back to $I$ again, yielding:
$\Delta E_{p}^{(1)}=b_{1}^{(1)}p^{2}+b_{2}^{(1)}p^{4}+\cdots$ $None$
$b_{1}^{(1)}=\frac{\alpha}{\pi\mu}(\frac{4}{3}\ln
2+\frac{4}{3}\ln\mu-\frac{4}{3}C_{1})$ $None$
$b_{2}^{(1)}=\frac{\alpha}{\pi\mu^{3}}(-\frac{2}{15})$ $None$
Note that the term $b_{1}^{(1)}p^{2}$ will combine with the kinetic energy
$\frac{1}{2\mu}p^{2}$ of a (”spinless”) electron, they are indistinguishable.
The appearance of an arbitrary constant $C_{1}$ precisely reflects the fact
that we cannot find the reduced mass via the valuation of $\Delta E_{p}^{(1)}$
in perturbation theory. So we must choose $b_{1}^{(1)}=0$ to reconfirm the
value of $\mu$ (which is still not the final observed mass, see below).
Similar evaluation on $H_{int}^{(2)}$ (of the real electron with spin $1/2$)
which would induce the spin flip process between states
$|{\mathbf{p}},\pm\frac{1}{2}>$ and $|{\mathbf{q}},\pm\frac{1}{2}>$, leads to
$\Delta E_{p}^{(2)}=\frac{1}{2}\sum_{i,s_{z}=\pm
1/2}\frac{|<i|H_{int}^{(2)}|{\mathbf{p}},s_{z}>|^{2}}{\varepsilon_{p}-\varepsilon_{i}}=-\frac{\alpha
g^{2}}{8\pi\mu}\int_{-1}^{1}d\eta J$
$J=\int_{0}^{\infty}\frac{k^{2}dk}{k+\xi}$ $None$
Being a quadratically divergent integral, $J$ needs partial derivative of
third order with respect to $\xi$, yielding:
$\Delta E_{p}^{(2)}=b_{0}^{(2)}+b_{1}^{(2)}p^{2}+b_{2}^{(2)}p^{4}+\cdots$
$None$
$b_{0}^{(2)}=\frac{g^{2}}{4}\frac{\alpha\mu}{\pi}[4(\ln
2+\ln\mu)-4C_{2}-\frac{2C_{3}}{\mu}-\frac{C_{4}}{\mu^{2}}]$ $None$
$b_{1}^{(2)}=\frac{g^{2}}{4}\frac{\alpha}{\pi\mu}(\frac{4}{3}\ln
2+2+\frac{4}{3}\ln\mu-\frac{4}{3}C_{2})$ $None$
$b_{2}^{(2)}=\frac{g^{2}}{4}\frac{\alpha}{\pi\mu^{3}}(-\frac{1}{15})$ $None$
Let’s manage to fix three arbitrary constants $C_{2},C_{3}$ and $C_{4}$.
First, the term $b_{1}^{(2)}p^{2}$ should be combined with
$\frac{1}{2\mu}p^{2}$ term. Since $\mu$ is already fixed, further modification
on $\mu$ due to electron spin should be finite and fixed. So the only possible
choice of $C_{2}$ is to cancel $\ln\mu$ which is ambiguous in dimension:
$C_{2}=\ln\mu$, yielding
$b_{1}^{(2)}=\frac{\beta}{2\mu},\quad\beta=\frac{g^{2}\alpha}{2\pi}(\frac{4}{3}\ln
2+2)$ $None$
Then the dimensional constants $C_{3}$ and $C_{4}$ must be chosen such that
$b_{0}^{(2)}=0$, implying that the starting point of this theory is the
nonrelativistic Hamiltonian $H_{0}$ in Eq.(A1) without rest energy term while
both masses of the nucleus and the electron (with spin) are fixed by
experiments. Hence now $\mu$ acquires a modification via $b_{1}^{(2)}p^{2}$
term and becomes an observable one:
$\mu\longrightarrow\mu_{obs}=\frac{\mu}{1+\beta}$ $None$
However, we have to consider the relativistic energy of electron shown in
Eq.(30), where the term $(-\frac{1}{8\mu^{3}}p^{4})$ goes beyond Eq.(A1). Yet
the modification of $\mu$ shown as (A16) does induce a corresponding change
$-\frac{1}{8}(\frac{1}{\mu_{obs}^{3}}-\frac{1}{\mu^{3}})p^{4}$, which should
be regarded as an invisible ”background” and subtracted from the $p^{4}$ term
induced by radiative corrections. (The relativistic correction is brought in
via the RDE as discussed in section VIII). As a whole, the combination of
contributions from $H_{int}^{(1)}$ and $H_{int}^{(2)}$ leads to
$b_{1}=b_{1}^{(1)}+b_{1}^{(2)}=b_{1}^{(2)}$ $None$
and a ”renormalized” $b_{2}$:
$b_{2}^{R}=b_{2}^{(1)}+b_{2}^{(2)}+\frac{1}{8\mu^{3}}(3\beta+3\beta^{2}+\beta^{3})\simeq\frac{\alpha}{\pi\mu_{obs}^{3}}(1.99808)$
$None$
Here we only keep the lowest approximation at the last step. Hence the
electron self-energy-diagram contributes a radiative correction to the energy
level of the stationary state $|Z,n,l>$ in a hydrogenlike atom:
$\Delta E^{rad}(Z,n,l)=\langle
Z,n,l|b_{2}^{R}p^{4}|Z,n,l\rangle=[\frac{8n}{2l+1}-3]\frac{b_{2}^{R}Z^{4}\alpha^{4}}{n^{4}}\mu^{4}_{obs}$
$None$
This form, together with contributions from the vacuum polarization and
nuclear size effect, gives a theoretical value for classic Lamb shift:
$L_{H}^{theor.}(2S_{1/2}-2P_{1/2})\approx 1056.52\ MHz$ $None$
which is smaller than the experimental value by $0.13\%$. Despite its
approximation involved, the above method clearly shows that so-called
renormalization is nothing but a reconfirmation process of mass. We must
reconfirm the mass before it could be modified via radiative corrections.
Either ”skipping over the first step” or ”combining two steps into one” is not
allowed.
5\. In noncovariant theory, the (three-dimensional) momentum $\mathbf{p}$ is
combined with the reduced mass $\mu$ to form a kinetic energy term
$\frac{1}{2\mu}{\mathbf{p}}^{2}$ on the mass shell. Once the energy is
modified whereas $\mathbf{p}$ is conserved at the vertex, $\mu$ is bound to be
modified. On the other hand, in covariant theory, the electron energy turns to
a component of four-dimensional momentum $p$ and the latter is conserved at
the vertex. So the (reduced) mass $\mu$ cannot be modified on the mass shell
$(p^{2}=\mu^{2})$. Therefore, the renormalization as some reconfirmation has
different meaning in covariant theory versus that in noncovariant theory. We
guess this is why the matching procedure of these two formalisms into one
theory for Lamb shift proves so difficult.
6\. Every theory in physics is not only a discovery of natural law, but also
an invention of human being 33 (33). Hence the comparison among various
theories, in many cases, is not about a problem of being right or wrong.
Rather, it’s about a choice of simplicity, harmony (self-consistency) and
beauty. Only time can tell.
## References
* (1) Sapirstein J R and Yennie D R 1990 in Quantum Electrodynamics (Editor: Kinoshita T, World Scientific ) p.560.
* (2) Pachucki K et al. 1996 J. Phys. B: Atom Mol. Opt. Phys. 29 177 .
* (3) Barker W A and Chraplyvy Z V 1953 Phys. Rev. 89 446 ;
Chraplyvy Z V 1953 Phys. Rev. 91 388 ; 1953 Phys. Rev. 92 1310 .
* (4) Barker W A and Glover F N 1955 Phys. Rev. 99 317 .
* (5) Udem Th et al. 1997 Phys. Rev. Lett. 79 2646 .
* (6) Schmidt-Kaler F et al. 1993 Phys. Rev. Lett. 70 2261 .
* (7) Yang J F 1994 Thesis for PhD (Fudan University); hep-th/9708104;
Yang J F and Ni G J 1995 Acta Physica Sinica (Overseas Edition) 4 88\.
* (8) Ni G J and Chen S Q 1998 Acta Physica Sinica (Overseas Edition) 7 401\.
* (9) Ni G J, Yang G H, Fu R T and Wang H B 2001 Int. J. Mod. Phys. A 16 2873; hep-th/9906254.
* (10) Yang J F 1998 Phys. Rev. D 57 4686 .
* (11) Weitz M et al. 1995 Phys. Rev. A 52 2664 .
* (12) Kinoshita T 1995 Phys. Rev. Lett. 75 4728 .
* (13) Farnham D L et al. 1995 Phys. Rev. Lett. 75 3598 .
* (14) Review of Particle Physics 2000 Euro. Phys. J. C 15 350\.
* (15) Ni G J 2003 Progress in Physics (Nanjing, China) 23 484 ; hep-th/0206250.
* (16) Ni G J and Chen S Q 2003 Advanced Quantum Mechanics 2nd. (Fudan University Press); English Edition was published by the Rinton Press, 2002.
* (17) Ni G J 2003 in Relativity, Gravitation, Cosmology, Edit by Dvoeglazov V V and Espinoza Garrido A A, Chapter 10 (NOVA Science Publisher); physics/0302038.
* (18) Ni G J 2003 in Relativity, Gravitation, Cosmology, Edit by Dvoeglazov V V and Espinoza Garrido A A, Chapter 11 (NOVA Science Publisher); hep-ph/0306028; preprint, hep-ph/0404030, to be published in Relativity, Gravitation, Cosmology.
* (19) Ni G J and Chen S Q 1997 J. Fudan University (Natural Science) 36 247; hep-th/9708156;
in Proceedings of International Workshop, Lorentz Group, CPT and Neutrinos,
Zacatocas, Maxico, June 23-26, 1999 (World Scientific, 2000), p.450.
* (20) Gross D J and Neveu A 1974 Phys. Rev. D 10 3235 .
* (21) Ni G J and Wang H B 1998 Physics Since Parity Breaking, edited by Wang F (World Scientific, Singapore) p.436; hep-ph/9708457.
* (22) Sakurai J J 1967 Advanced Quantum Mechanics (Addison-Wesley Publishing Company).
* (23) Itzykson C and Zuber J-B 1980 Quantum Field Theory (McGraw-Hill Book Company).
* (24) Peskin M E and Schroeder D V 1995 An Introduction to Quantum Field Theory (Addison-Wesley Publishing Company).
* (25) Ni G J, Lou S Y, Lu W F and Yang J F 1998 Science in China (Series A), 41 1206; hep-ph/9801264.
* (26) Feng S S and Ni G J 1999 Int. J. Mod. Phys. A 14 4259\.
* (27) Ni G J, Wang H B, Yan J and Li H L 2000 High Energy Physics and Nuclear Physics24 400\.
* (28) Eides M I, Grotch H and Shelyuto V A 2001 Phys. Rep. 342 63-261 .
* (29) Huber A , Udem Th, Gross B et al. 1998 Phys. Rev. Lett. 80 468 .
* (30) Bethe H A 1947 Phys. Rev. 72 339 .
* (31) T. A. Welton 1948 Phys. Rev. 74 1157 .
* (32) Jentschura U D , Mohr P J and Soff G 1999 Phys. Rev. Lett. 82 53 .
* (33) Drake G W F and Swainson R A 1990 Phys. Rev. A 41 1243 .
* (34) Ni G J 2004 physics/0407092 accepted by ”Relativity, Gravitation, Cosmology”.
* (35) Marsch E 2005 Ann. Phys. (Leipzig) 14 No.5 324.
* (36) Dai Xianxi 1977 Journal of Fudan University No.1 100.
* (37) Ni G J and Chen S Q 1995 Levinson Theorem, Anomaly and the Phase Transition of Vacuum (Shanghai Scientific & Technical Publishers).
* (38) Ma Z Q and Ni G J 1985 Phys. Rev. D 31 1482 .
* (39) Weinberg S 1995 The Quantum Theory of Fields Vol.I-II (Cambridge University Press, Cambridge)
* (40) Amsler C et al. Particle Physics Booklet 2008 Phys. Lett. B 667 1
* (41) Lou S Y and Ni G J 1989 Phys. Rev. D 40 3040
* (42) Ni G J, Chen S Q, Lou S Y and Xu J J Essence of Special Relativity, Reduced Dirac Equation and Antigravity, Preprint.
* (43) Ni G J, Xu J J and Lou S Y Puzzles of Divergence and Renormalization in Quantum Field Theory, Preprint.
* (44) Niering M et al., 2000 Phys. Rev. Lett. 84 5496
* (45) Schwob C et al., 1999 Phys. Rev. Lett. 82 4960
* (46) de Beauvoir B et al., 2000 Eur. Phys. J D 12 61
Figure 1: A hydrogenlike atom in quantum mechanical description. The nucleus
with mass $m_{2}$ occupies a small sphere with radius $r_{N}$ (greatly
exaggerated in the diagram) while the electron with mass $m_{1}$ spreads over
a larger sphere with radius $R_{e}$ (i.e.atomic radius). Their common center
is the atom’s center of mass (CM). The wavefunction $\psi({\bf r})e^{-iEt}$
with ${\bf r}={\bf r}_{1}-{\bf r}_{2}$ shows the electron’s amplitude under a
”fictitious measurement” 15 (15), during which the electron and nucleus shrink
into two ”fictitious point particles ” located at ${\bf r}_{1}$ and ${\bf
r}_{2}$ simultaneously. The Coulomb potential $V(r)=-\frac{Ze^{2}}{r}$ between
them is a static one. The probability to find the electron at ${\bf r}$ is
$|\psi({\bf r})|^{2}$ while that to find its momentum being ${\bf p}$ is
$|\phi({\bf p})|^{2}$ with $\phi({\bf p})$ being the Fourier transform of
$\psi({\bf r})$.
Figure 2: Four Feynman diagrams at one-loop level (in covariant form). (a) and
(b) are self-energy diagrams of the electron. (c) is vacuum polarization. (d)
is vertex function. Solid lines and wavy lines refer to electron and photon
respectively, while X denotes the nucleus. Here $p,q$ and $k$ are four-
dimensional momenta.
Figure 3: The electron self-energy (radiative correction) diagram at one-loop
level of perturbative QCD in noncovariant form. $H_{int}^{(1)}$ (A.2) or
$H_{int}^{(2)}$ (A.4) is inserted into two vertices. Here $\bf p,q$ and $\bf
k$ are three dimensional momenta.
|
# Large deviation principle for geometric and topological functionals and
associated point processes
Christian<EMAIL_ADDRESS>[ Takashi
<EMAIL_ADDRESS>[ Department of Mathematics
Aarhus University
Ny Munkegade, 118, 8000, Aarhus C, Denmark, Department of Statistics
Purdue University
West Lafayette, 47907, USA,
###### Abstract
We prove a large deviation principle for the point process associated to
$k$-element connected components in ${\mathbb{R}}^{d}$ with respect to the
connectivity radii $r_{n}\to\infty$. The random points are generated from a
homogeneous Poisson point process or the corresponding binomial point process,
so that $(r_{n})_{n\geq 1}$ satisfies $n^{k}r_{n}^{d(k-1)}\to\infty$ and
$nr_{n}^{d}\to 0$ as $n\to\infty$ (i.e., sparse regime). The rate function for
the obtained large deviation principle can be represented as relative entropy.
As an application, we deduce large deviation principles for various
functionals and point processes appearing in stochastic geometry and topology.
As concrete examples of topological invariants, we consider persistent Betti
numbers of geometric complexes and the number of Morse critical points of the
min-type distance function.
60F10,
60D05, 60G55, 55U10,
Large deviation principle,
point process,
stochastic geometry,
stochastic topology,
persistent Betti number,
Morse critical point,
###### keywords:
[class=MSC]
###### keywords:
, and
## 1 Introduction
The objective of this paper is to examine large deviation behaviors of a point
process associated to a configuration of random points generated by a
homogeneous Poisson point process or the corresponding binomial point process.
The _Donsker–Varadhan large deviation principle (LDP)_ (see [14])
characterizes the limiting behavior of a family of probability measures
$(\mu_{n})_{n\geq 1}$ on a measurable space $(\mathcal{X},\mathcal{B})$. We
say that $(\mu_{n})_{n\geq 1}$ satisfies an LDP with rate $a_{n}$ and _rate
function_ $I(x)$, if for any $A\in\mathcal{B}$,
${-\inf_{x\in\text{int}(A)}I(x)}\leq\liminf_{n\to\infty}\frac{1}{a_{n}}\log\mu_{n}(A)\leq\limsup_{n\to\infty}\frac{1}{a_{n}}\log\mu_{n}(A)\leq-\inf_{x\in\bar{A}}I(x),$
(1.1)
where $\text{int}(A)$ denotes the interior of $A$ and $\bar{A}$ the closure of
$A$.
Let $\mathcal{P}_{n}$ be a homogeneous Poisson point process with intensity
$n$ on the unit cube $[0,1]^{d}$, $d\geq 2$, and $(r_{n})_{n\geq 1}$ be a
sequence of positive numbers decreasing to $0$ as $n\to\infty$. We put our
focus on the spatial distribution of $k$-tuples
${\mathcal{Y}}\subset\mathcal{P}_{n}$, which consist of components with
respect to $r_{n}$. More specifically, given a parameter $t\in(0,\infty)$, we
consider the following point process in $({\mathbb{R}}^{d})^{k}$:
$\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n},\,|{\mathcal{Y}}|=k}s_{n}({\mathcal{Y}},\mathcal{P}_{n};t)\,\delta_{r_{n}^{-1}({\mathcal{Y}}-\ell({\mathcal{Y}}))},$
(1.2)
where $\delta_{\bf z}$ denotes the Dirac measure at ${\bf
z}=(z_{1},\dots,z_{k})\in({\mathbb{R}}^{d})^{k}$ and $\ell({\mathcal{Y}})$ is
a “center” point of ${\mathcal{Y}}$ such as the left most point (in the
lexicographic ordering). Moreover, $s_{n}$ denotes an indicator function,
requiring that the diameter of $k$-tuples
${\mathcal{Y}}\subset\mathcal{P}_{n}$ be at most a constant multiple of
$r_{n}$, and further, such $k$-tuples must be distant at least $r_{n}t$ from
all the remaining points in $\mathcal{P}_{n}$. In this setup, our main theorem
(i.e., Theorem 2.1) aims to describe the LDP for the process (1.2) in the so-
called _sparse regime_ : $nr_{n}^{d}\to 0$ as $n\to\infty$.
As a primary application, we also provide the LDP for geometric and
topological statistics possessing the structure as component counts. Given
$t\in(0,\infty)$ and measurable functions $H\mathrel{\mathop{\mathchar
58\relax}}({\mathbb{R}}^{d})^{k}\to[0,\infty)$ that are symmetric in the $k$
arguments of ${\mathbb{R}}^{d}$, we define
$G_{n}({\mathcal{Y}},\mathcal{P}_{n};t)\mathrel{\mathop{\mathchar
58\relax}}=H(r_{n}^{-1}{\mathcal{Y}})\,{\mathbbm{1}}\big{\\{}\|y-z\|\geq
r_{n}t,\ \text{for all }y\in{\mathcal{Y}}\text{ and
}z\in\mathcal{P}_{n}\setminus{\mathcal{Y}}\big{\\}},$ (1.3)
where ${\mathbbm{1}}\\{\cdot\\}$ denotes an indicator function and $\|\cdot\|$
is the Euclidean norm. We then consider the point process
$\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n},\,|{\mathcal{Y}}|=k}\delta_{G_{n}({\mathcal{Y}},\mathcal{P}_{n};t)}.$
(1.4)
The process (1.4) is associated to scaled $k$-tuples
$r_{n}^{-1}{\mathcal{Y}}$, satisfying the geometric condition implicit in $H$,
with the additional restriction that geometric objects generated by
$r_{n}^{-1}{\mathcal{Y}}$ be distant at least $t$ from all the remaining
points of $r_{n}^{-1}\mathcal{P}_{n}$. A more precise and general setup for
the process (1.4) is given in Section 3.
With an appropriate choice of scaling regimes, the large deviations of point
processes as in (1.4) have been studied by Sanov’s theorem and its variant
[16, 37, 17]. In recent times, the process (1.4) also found applications in
stochastic geometry [13, 26, 9]. In particular, the authors of [9] explored
the Poisson process approximation of the point process for general stabilizing
functionals, including (1.4) as its special case, by deriving the rate of
convergence in terms of the Kantorovich-Rubinstein distance. Furthermore, [27]
discussed large deviations of the probability distribution of (1.4) under the
$M_{0}$-topology, with the assumption that the connectivity radii are even
smaller than those considered in [13, 9].
As another main application, we also examine the LDP for statistics of the
form
$\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n},\,|{\mathcal{Y}}|=k}G_{n}({\mathcal{Y}},\mathcal{P}_{n};t).$
(1.5)
A variety of functionals in stochastic geometry can be considered as such
statistics; see [5, 9, 26, 22]. Moreover, with an appropriate choice of $H$,
the statistics (1.5) can be used to investigate the behavior of topological
invariants of a geometric complex [6, 8, 20, 30, 38, 28]. Along this line of
research, Section 4 below discusses the LDP for persistent Betti numbers and
the number of Morse critical points of the min-type distance function. Loosely
speaking, the persistent Betti number is a quantifier for topological
complexity, capturing the creation / destruction of topological cycles. The
Morse critical points under our consideration provide a good approximation of
homological changes in the geometric complex.
The large deviation behavior of the processes (1.2), (1.4), and (1.5) heavily
depends on how rapidly the sequence $(r_{n})_{n\geq 1}$ decays to $0$ as
$n\to\infty$. In the existing literature (not necessarily relating to large
deviations), the configuration of geometric objects (determined by $H$ in
(1.3)) splits into multiple regimes [31, 20, 30]. In the context of random
geometric complexes (resp. random geometric graphs), if $nr_{n}^{d}\to 0$,
called the _sparse regime_ , the spatial distribution of complexes (resp.
graphs) is sparse, so that they are mostly observed as isolated components. In
the critical phase $nr_{n}^{d}\to c\in(0,\infty)$, called the _critical
regime_ for which $r_{n}$ decreases to $0$ at a slower rate than the sparse
regime, the complex (resp. graph) begins to coalesce, forming much larger
components. Finally, the case when $nr_{n}^{d}\to\infty$ is the _dense regime_
, for which the complex (resp. graph) is even more connected and may even
consist of a single giant component.
Among the regimes described above, the present paper focuses on the sparse
regime. In this case, there have been a number of studies on the “average”
behavior and the likely deviations from the “average” behavior of geometric
functionals and topological invariants, such as subgraph/component counts and
Betti numbers. More specifically, a variety of strong laws of large numbers
and central limit theorems for these quantities have already been established.
The readers may refer to the monograph [31] for the limit theorems for
subgraph/component counts, while the works in [20, 30, 38] provide the limit
theorems for topological invariants.
In contrast, however, there have been very few attempts made at examining the
large deviation behaviors of these quantities, especially from the viewpoints
of an LDP. In fact, even for a simple edge count in a random geometric graph,
determining the rate $a_{n}$ and the rate function $I(x)$ in (1.1) is a highly
non-trivial problem ([11]). Although there are several works (e.g., [3, 33,
39]) that deduced concentration inequalities for subgraph counts, the length
power functionals, and Betti numbers, these papers were not aimed to derive
the LDP. As a consequence, the obtained upper bounds in these studies do not
seem to be tight. Furthermore, [35] studied the LDP for the functional of
spatial point processes satisfying a weak dependence condition characterized
by a radius of stabilization. One of the major assumptions in their study is
that the contribution of any particular vertex must be uniformly bounded (see
condition (L1) therein). However, our functionals and point processes may not
fulfill such conditions. More importantly, the study in [35] treated only the
critical regime, whereas the main focus of our study is the sparse regime. One
of the benefits of studying the sparse regime is that geometric and
topological objects have a relatively simpler structure in the limit, which
will allow us to explicitly identify the structure of a rate function. In some
cases, we may even solve the variational problem for the rate function; see
Remark 3.6. We also note that [36] provides a framework to establish LDPs when
the uniform boundedness assumption in [35] is replaced by the finiteness of
suitably scaled logarithmic moment generating functions (see Equ. (2.5)
therein). However, the LDPs in [36] are designed only for real-valued random
variables, whereas our main results include LDPs for random measures, such as
Theorems 2.1 and 3.1 below. As a final remark, we note that the work in [19],
which is still under preparation, seems to utilize the ideas of [36] to derive
the LDP for Betti numbers and persistence diagrams of cubical complexes.
The remainder of this paper is structured as follows. Section 2 gives a
rigorous description of the LDP for the point process in (1.2). In Section 3,
we shall deduce the LDP for the processes in (1.4) and (1.5). Additional
examples on persistent Betti numbers and the number of Morse critical points
will be offered in Section 4. All the proofs are deferred to Section 5. For
the proof of our main theorems, we partition the unit cube $[0,1]^{d}$ into
multiple smaller sub-cubes and consider a family of point processes restricted
to each of these small sub-cubes. Subsequently, with the aid of the main
theorem in [13], we establish weak convergence of such point processes in
terms of the total variation distance. We then exploit Cramér’s theorem in
Polish spaces [14, Theorem 6.1.3] to identify the structure of a rate
function. After that, Proposition 2.2 ensures that the rate function can be
represented in terms of relative entropy. The required approximation argument
relies on the standard technique on the maximal coupling ([21, Lemma 4.32]).
For the proof of the theorems in Section 3, we shall utilize an extension of
the contraction principle, provided in [14, Theorem 4.2.23].
Before concluding the Introduction, we comment on our setup and assumptions.
First, we assume that the Poisson point process $\mathcal{P}_{n}$ is
homogeneous. It is clear, however, that in many applications (e.g., [1, 29]),
it is important to understand geometric and topological effects of lack of
homogeneity. A possible starting point for introducing inhomogeneity is to
look at point clouds arising from “inhomogeneous” Poisson point processes. In
this case, there is no doubt that a new and more involved machinery must be
developed to analyze such data; a detailed discussion will be postponed to a
future publication. Another possible extension seems to investigate the
regimes other than the sparse case. At least for the critical case (i.e.,
$nr_{n}^{d}\to c\in(0,\infty)$) however, it is impossible to directly
translate our proof techniques. In particular, Lemma 5.3 does not hold
anymore; see Remark 5.4 for more details on this point. Finally, the LDPs in
Section 4 are proven only when the dimensions of the Euclidean space, as well
as those of topological invariants, are small enough. This is due to the fact
that our proof techniques can apply only to a lower-dimensional case. It is
still unclear whether the LDP holds in higher-dimensions; this is also left as
a future topic of our research.
## 2 Model and main results
Let $\mathcal{P}_{n}$ be a homogeneous Poisson point process on $[0,1]^{d}$,
$d\geq 2$, with intensity $n$. Choose an integer $k\geq 2$, which will remain
fixed for the remainder of this section, and let
$\mathsf{diam}(x_{1},\dots,x_{k})\mathrel{\mathop{\mathchar
58\relax}}=\max_{1\leq i<j\leq k}\|x_{i}-x_{j}\|,\ \ \
x_{i}\in{\mathbb{R}}^{d}.$
Taking a sequence $(r_{n})_{n\geq 1}$ decaying to $0$ as $n\to\infty$, we
focus on the _sparse regime_ :
$\rho_{k,n}\mathrel{\mathop{\mathchar
58\relax}}=n^{k}r_{n}^{d(k-1)}\to\infty,\ \ \ nr_{n}^{d}\to 0,\ \ \text{ as
}n\to\infty.$ (2.1)
For a subset ${\mathcal{Y}}$ of $k$ points in ${\mathbb{R}}^{d}$, a finite set
$\mathcal{Z}\supset{\mathcal{Y}}$ in ${\mathbb{R}}^{d}$ and $t\in(0,\infty)$,
we define
$c({\mathcal{Y}},\mathcal{Z};t)\mathrel{\mathop{\mathchar
58\relax}}={\mathbbm{1}}\big{\\{}\|y-z\|\geq t,\ \text{for all
}y\in{\mathcal{Y}}\text{ and }z\in\mathcal{Z}\setminus{\mathcal{Y}}\big{\\}}.$
We then define a scaled version of $c$ by
$c_{n}({\mathcal{Y}},\mathcal{Z};t)\mathrel{\mathop{\mathchar
58\relax}}=c(r_{n}^{-1}{\mathcal{Y}},r_{n}^{-1}\mathcal{Z};t)={\mathbbm{1}}\big{\\{}\|y-z\|\geq
r_{n}t,\ \text{for all }y\in{\mathcal{Y}}\text{ and
}z\in\mathcal{Z}\setminus{\mathcal{Y}}\big{\\}}.$ (2.2)
and also, for a fixed $L\in(t,\infty)$,
$s_{n}({\mathcal{Y}},\mathcal{Z};t)\mathrel{\mathop{\mathchar
58\relax}}=c_{n}({\mathcal{Y}},\mathcal{Z};t)\,{\mathbbm{1}}\big{\\{}\mathsf{diam}({\mathcal{Y}})\leq
r_{n}L\big{\\}}.$ (2.3)
Let $M_{+}\big{(}({\mathbb{R}}^{d})^{k}\big{)}$ be the space of Radon measures
on $({\mathbb{R}}^{d})^{k}$. For a finite set
${\mathcal{Y}}\subset{\mathbb{R}}^{d}$ of $k$ points in general position,
denote by $\ell({\mathcal{Y}})$ a _center point_ of ${\mathcal{Y}}$. For
example, it may represent the left most point of ${\mathcal{Y}}$ in the
lexicographic ordering. Another way of defining it is that one may set
$\ell({\mathcal{Y}})$ to be a center of the unique $(k-2)$-dimensional sphere
containing ${\mathcal{Y}}$. In either case, we write
$\overline{\mathcal{Y}}\mathrel{\mathop{\mathchar
58\relax}}={\mathcal{Y}}-\ell({\mathcal{Y}})$.
The primary objective of this section is to describe the LDP for the point
process
$\xi_{k,n}\mathrel{\mathop{\mathchar
58\relax}}=\frac{1}{\rho_{k,n}}\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n},\,|{\mathcal{Y}}|=k}s_{n}({\mathcal{Y}},\mathcal{P}_{n};t)\,\delta_{r_{n}^{-1}\overline{\mathcal{Y}}},\
\ n\geq 1.$ (2.4)
The process $\xi_{k,n}$ counts scaled $k$-tuples
$r_{n}^{-1}\overline{\mathcal{Y}}$, that are locally concentrated in the sense
of $\mathsf{diam}({\mathcal{Y}})\leq r_{n}L$, under an additional restriction
that $r_{n}^{-1}{\mathcal{Y}}$ be distant at least $t$ from all the remaining
points of $r_{n}^{-1}\mathcal{P}_{n}$. Notice that
$\mathsf{diam}({\mathcal{Y}})\leq r_{n}L$ indicates
$\|r_{n}^{-1}\overline{\mathcal{Y}}\|\leq C$ for some constant
$C\in(0,\infty)$. Thus, one can fix a compact subset $E$ of
$({\mathbb{R}}^{d})^{k}$ such that the process (2.4) can be viewed as an
element of $M_{+}(E)$. Because of this restriction, $M_{+}(E)$ is now
equivalent to the space of _finite_ measures on $E$. Write
$\mathcal{M}_{+}(E)$ for the Borel $\sigma$-field generated by weak topology
on $M_{+}(E)$.
The proofs of the main results below are deferred to Section 5.1.
###### Theorem 2.1.
The sequence $(\xi_{k,n})_{n\geq 1}$ satisfies an LDP in the weak topology
with rate $\rho_{k,n}$ and rate function
$\begin{split}\Lambda_{k}^{*}(\rho)&\mathrel{\mathop{\mathchar
58\relax}}=\sup_{f\in C_{b}(E)}\Big{\\{}\int_{E}f({\bf
x})\rho(\operatorname{d\\!}{\bf x})\\\
&\qquad\qquad-\frac{1}{k!}\int_{({\mathbb{R}}^{d})^{k-1}}\big{(}e^{f(\overline{({\bf
0}_{d},{\bf y})})}-1\big{)}\,{\mathbbm{1}}\big{\\{}\mathsf{diam}({\bf
0}_{d},{\bf y})\leq L\big{\\}}\operatorname{d\\!}{\bf y}\Big{\\}},\ \ \rho\in
M_{+}(E),\end{split}$ (2.5)
where ${\bf 0}_{d}=(0,\dots,0)\in{\mathbb{R}}^{d}$, ${\bf
y}\in({\mathbb{R}}^{d})^{k-1}$ and $C_{b}(E)$ is the collection of continuous
and bounded functions on $E$.
The rate function (2.5) can be associated to the notion of _relative entropy_.
Writing $\lambda_{m}$ for Lebesgue measure on $({\mathbb{R}}^{d})^{m}$, we
define
$\tau_{k}(A)=\frac{1}{k!}\,\lambda_{k-1}\big{\\{}{\bf
y}\in({\mathbb{R}}^{d})^{k-1}\mathrel{\mathop{\mathchar
58\relax}}\mathsf{diam}({\bf 0}_{d},{\bf y})\leq L,\,\overline{({\bf
0}_{d},{\bf y})}\in A\big{\\}},\ \ A\subset E.$ (2.6)
For a measure $\rho\in M_{+}(E)$,
$H_{k}(\rho|\tau_{k})\mathrel{\mathop{\mathchar
58\relax}}=\begin{cases}\int_{E}\log\frac{\operatorname{d\\!}\rho}{\operatorname{d\\!}\tau_{k}}({\bf
x})\rho(\operatorname{d\\!}{\bf x})-\rho(E)+\tau_{k}(E)&\text{ if
}\rho\ll\tau_{k},\\\ \infty&\text{ otherwise, }\end{cases}$ (2.7)
denotes the relative entropy of $\rho$ with respect to $\tau_{k}$. In the
special case when $\rho$ and $\tau_{k}$ are probability measures, (2.7)
reduces to the relative entropy defined for the space of probability measures;
see, for example, [14, Equ. (6.2.8)]. In our setup, however, $\rho$ and
$\tau_{k}$ are not necessarily probability measures, so we need to extend the
definition of relative entropy as in (2.7).
###### Proposition 2.2.
Under the setup of Theorem 2.1, we have that
$\Lambda_{k}^{*}(\rho)=H_{k}(\rho|\tau_{k}),\ \ \ \rho\in M_{+}(E).$
Finally, we also prove the analog of Theorem 2.1 when the Poisson point
process is replaced by a binomial point process. To make this precise, we put
$\xi_{k,n}^{\mathsf{B}}\mathrel{\mathop{\mathchar
58\relax}}=\frac{1}{\rho_{k,n}}\sum_{{\mathcal{Y}}\subset\mathcal{B}_{n},\,|{\mathcal{Y}}|=k}s_{n}({\mathcal{Y}},\mathcal{B}_{n};t)\,\delta_{r_{n}^{-1}\overline{\mathcal{Y}}},\
\ n\geq 1,$
where $\mathcal{B}_{n}=\\{X_{1},\dots,X_{n}\\}$ is a binomial point process
consisting of $n$ i.i.d. uniform random vectors in $[0,1]^{d}$.
###### Corollary 2.3.
The sequence $(\xi_{k,n}^{\mathsf{B}})_{n\geq 1}$ satisfies an LDP in the weak
topology with rate $\rho_{k,n}$ and rate function
$\Lambda_{k}^{*}=H_{k}(\cdot|\tau_{k})$.
## 3 Large deviation principles for geometric and topological functionals
In this section, we provide the LDP for point processes relating more directly
to geometric and topological functionals. Choose integers $k\geq 2$ and $m\geq
1$, which remain fixed throughout the section. Define
$H\mathrel{\mathop{\mathchar
58\relax}}=(h^{(1)},\dots,h^{(m)})\mathrel{\mathop{\mathchar
58\relax}}({\mathbb{R}}^{d})^{k}\to[0,\infty)^{m}$ to be a non-negative
measurable function satisfying the following conditions:
(H1) $H$ is symmetric with respect to permutations of variables in
${\mathbb{R}}^{d}$.
(H2) $H$ is translation invariant:
$H(x_{1},\dots,x_{k})=H(x_{1}+y,\dots,x_{k}+y),\ \ \
x_{i},y\in{\mathbb{R}}^{d}.$
(H3) $H$ is locally determined:
$H(x_{1},\dots,x_{k})={\bf 0}_{m}\ \ \text{whenever
}\mathsf{diam}(x_{1},\dots,x_{k})>L,$
where $L>0$ is a constant and ${\bf 0}_{m}=(0,\dots,0)\in{\mathbb{R}}^{m}$.
(H4) For every ${\bf a}=(a_{1},\dots,a_{m})\in{\mathbb{R}}^{m}$,
$\int_{({\mathbb{R}}^{d})^{k-1}}e^{\langle{\bf a},H({\bf 0}_{d},{\bf
y})\rangle}\Big{(}{\mathbbm{1}}\big{\\{}H({\bf 0}_{d},{\bf y})\neq{\bf
0}_{m}\big{\\}}+\sum_{1\leq i\leq j\leq m}h^{(i)}({\bf 0}_{d},{\bf
y})h^{(j)}({\bf 0}_{d},{\bf y})\Big{)}\operatorname{d\\!}{\bf y}<\infty,$
where ${\bf y}\in({\mathbb{R}}^{d})^{k-1}$ and $\langle\cdot,\cdot\rangle$
denotes the Euclidean inner product.
(H5) For every ${\bf a}=(a_{1},\dots,a_{m})\in{\mathbb{R}}^{m}\setminus\\{{\bf
0}_{m}\\}$, it holds that
$\int_{({\mathbb{R}}^{d})^{k-1}}\big{|}\sum_{i=1}^{m}a_{i}h^{(i)}({\bf
0}_{d},{\bf y})\big{|}\operatorname{d\\!}{\bf y}>0$.
If $h^{(i)}$ are all bounded functions, we can drop condition (H4) because it
can be implied by (H3). Moreover, note that (H3) remains true even when $L$ is
increased. Hence, we may assume that $L$ is larger than a fixed value $t>0$.
Taking up a sequence $(r_{n})_{n\geq 1}$ in (2.1), we define a scaled version
of $H$ by
$H_{n}(x_{1},\dots,x_{k})\mathrel{\mathop{\mathchar
58\relax}}=H(r_{n}^{-1}x_{1},\dots,r_{n}^{-1}x_{k}),\ \ \ n=1,2,\dots.$ (3.1)
For a $k$-point subset ${\mathcal{Y}}\subset{\mathbb{R}}^{d}$, a finite subset
$\mathcal{Z}\supset{\mathcal{Y}}$ in ${\mathbb{R}}^{d}$ and ${\bf
t}=(t_{1},\dots,t_{m})\in[0,\infty)^{m}$, define
$c({\mathcal{Y}},\mathcal{Z};{\bf t})\mathrel{\mathop{\mathchar
58\relax}}=\big{(}c({\mathcal{Y}},\mathcal{Z};t_{i})\big{)}_{i=1}^{m},$ (3.2)
and also,
$G({\mathcal{Y}},\mathcal{Z};{\bf t})\mathrel{\mathop{\mathchar
58\relax}}=H({\mathcal{Y}})\odot c({\mathcal{Y}},\mathcal{Z};{\bf t}),$
where $\odot$ means the Hadamard product; that is, for $A=(a_{ij})$ and
$B=(b_{ij})$, we have $A\odot B=(a_{ij}b_{ij})$. We can define the scaled
version of these functions by
$c_{n}({\mathcal{Y}},\mathcal{Z};{\bf t})\mathrel{\mathop{\mathchar
58\relax}}=c(r_{n}^{-1}{\mathcal{Y}},r_{n}^{-1}\mathcal{Z};{\bf t}),\ \ \text{
and }\ \ G_{n}({\mathcal{Y}},\mathcal{Z};{\bf t})\mathrel{\mathop{\mathchar
58\relax}}=H_{n}({\mathcal{Y}})\odot c_{n}({\mathcal{Y}},\mathcal{Z};{\bf
t}).$ (3.3)
For notational convenience, the $i$th entries of $H_{n}({\mathcal{Y}})$ and
$G_{n}({\mathcal{Y}},\mathcal{Z};{\bf t})$ are denoted respectively as
$h_{n}^{(i)}({\mathcal{Y}})$ and
$g_{n}^{(i)}({\mathcal{Y}},\mathcal{Z};t_{i})$.
Henceforth, our aim is to explore the large deviations of the process
$U_{k,n}\mathrel{\mathop{\mathchar
58\relax}}=\frac{1}{\rho_{k,n}}\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n},\,|{\mathcal{Y}}|=k}\delta_{G_{n}({\mathcal{Y}},\mathcal{P}_{n};{\bf
t})}\in M_{+}(E^{\prime}),$ (3.4)
where $E^{\prime}=[0,\infty)^{m}\setminus\\{{\bf 0}_{m}\\}$. Although the
process (3.4) aggregates the contributions of a certain score function over
all $k$-tuples, it is not a pure $U$-statistic since this score function
depends also on the remaining points of $\mathcal{P}_{n}$. Finally, as an
analog of (2.6), we define
$\tau_{k}^{\prime}(A)\mathrel{\mathop{\mathchar
58\relax}}=\frac{1}{k!}\,\lambda_{k-1}\big{\\{}{\bf
y}\in({\mathbb{R}}^{d})^{k-1}\mathrel{\mathop{\mathchar 58\relax}}H({\bf
0}_{d},{\bf y})\in A\big{\\}},\ \ \ A\subset E^{\prime}.$
The proofs of all the theorems and propositions below are deferred to Sections
5.2 and 5.3.
###### Theorem 3.1.
The sequence $(U_{k,n})_{n\geq 1}$ satisfies an LDP in the weak topology with
rate $\rho_{k,n}$ and rate function
$\begin{split}\bar{\Lambda}_{k}^{*}(\rho)&\mathrel{\mathop{\mathchar
58\relax}}=\sup_{f\in C_{b}(E^{\prime})}\Big{\\{}\int_{E^{\prime}}f({\bf
x})\rho(\operatorname{d\\!}{\bf x})\\\
&\qquad\qquad-\frac{1}{k!}\int_{({\mathbb{R}}^{d})^{k-1}}\big{(}e^{f(H({\bf
0}_{d},{\bf y}))}-1\big{)}\,{\mathbbm{1}}\big{\\{}H({\bf 0}_{d},{\bf
y})\neq{\bf 0}_{m}\big{\\}}\operatorname{d\\!}{\bf y}\Big{\\}},\ \ \rho\in
M_{+}(E^{\prime}),\end{split}$ (3.5)
where ${\bf y}\in({\mathbb{R}}^{d})^{k-1}$. Furthermore,
$\bar{\Lambda}_{k}^{*}(\rho)=H_{k}^{\prime}(\rho|\tau_{k}^{\prime}),\ \
\rho\in M_{+}(E^{\prime}),$ (3.6)
where $H_{k}^{\prime}$ is the relative entropy in $M_{+}(E^{\prime})$ with
respect to $\tau_{k}^{\prime}$; that is, if $\rho\ll\tau_{k}^{\prime}$,
$H_{k}^{\prime}(\rho|\tau_{k}^{\prime})\mathrel{\mathop{\mathchar
58\relax}}=\int_{E^{\prime}}\log\frac{\operatorname{d\\!}\rho}{\operatorname{d\\!}\tau_{k}^{\prime}}({\bf
x})\rho(\operatorname{d\\!}{\bf
x})-\rho(E^{\prime})+\tau_{k}^{\prime}(E^{\prime})$ (3.7)
and $H_{k}^{\prime}(\rho|\tau_{k}^{\prime})=\infty$ otherwise.
As in Corollary 2.3, we can extend Theorem 3.1 to the case of a binomial
input. We define
$U_{k,n}^{\mathsf{B}}\mathrel{\mathop{\mathchar
58\relax}}=\frac{1}{\rho_{k,n}}\sum_{{\mathcal{Y}}\subset\mathcal{B}_{n},\,|{\mathcal{Y}}|=k}\delta_{G_{n}({\mathcal{Y}},\mathcal{B}_{n};{\bf
t})}.$
###### Corollary 3.2.
The sequence $(U_{k,n}^{\mathsf{B}})_{n\geq 1}$ satisfies an LDP in the weak
topology with rate $\rho_{k,n}$ and rate function
$\bar{\Lambda}_{k}^{*}=H_{k}^{\prime}(\cdot|\tau_{k}^{\prime})$.
Subsequently, we also consider statistics of the form
$T_{k,n}\mathrel{\mathop{\mathchar
58\relax}}=\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n},\,|{\mathcal{Y}}|=k}G_{n}({\mathcal{Y}},\mathcal{P}_{n};{\bf
t}).$ (3.8)
One can describe the corresponding LDP as follows. We define
$\mathcal{D}=\\{A\subset E^{\prime}\mathrel{\mathop{\mathchar
58\relax}}\bar{A}=\overline{\text{int}(A)}\\}$.
###### Theorem 3.3.
For every measurable $A\in\mathcal{D}$, we have as $n\to\infty$,
$\frac{1}{\rho_{k,n}}\log\mathbb{P}\big{(}\rho_{k,n}^{-1}T_{k,n}\in
A\big{)}\to-\inf_{{\bf x}\in A}I_{k}({\bf x}),$ (3.9)
where $I_{k}$ is a rate function given as
$I_{k}({\bf x})=\sup_{{\bf a}\in{\mathbb{R}}^{m}}\Big{\\{}\langle{\bf a},{\bf
x}\rangle-\frac{1}{k!}\,\int_{({\mathbb{R}}^{d})^{k-1}}\big{(}e^{\langle{\bf
a},H({\bf 0}_{d},{\bf y})\rangle}-1\big{)}\operatorname{d\\!}{\bf
y}\Big{\\}},\ \ {{\bf x}=(x_{1},\dots,x_{m})\in{\mathbb{R}}^{m}},$ (3.10)
with ${\bf y}\in({\mathbb{R}}^{d})^{k-1}$.
Again, we can extend the above result to the case of binomial point processes.
We define
$T_{k,n}^{\mathsf{B}}\mathrel{\mathop{\mathchar
58\relax}}=\sum_{{\mathcal{Y}}\subset\mathcal{B}_{n},\,|{\mathcal{Y}}|=k}G_{n}({\mathcal{Y}},\mathcal{B}_{n};{\bf
t}).$
###### Corollary 3.4.
For every measurable $A\in\mathcal{D}$, we have as $n\to\infty$,
$\displaystyle\frac{1}{\rho_{k,n}}\log\mathbb{P}\big{(}\rho_{k,n}^{-1}T_{k,n}^{\mathsf{B}}\in
A\big{)}\to-\inf_{{\bf x}\in A}I_{k}({\bf x}).$
The rate function $I_{k}$ satisfies the following properties.
###### Proposition 3.5.
$(i)$ $I_{k}$ is continuously differentiable on $E^{\prime}$.
$(ii)$ $I_{k}$ is strictly convex on $E^{\prime}$.
$(iii)$ $I_{k}({\bf x})=0$ if and only if ${\bf x}=(\mu_{1},\dots,\mu_{m})$,
where
$\mu_{i}\mathrel{\mathop{\mathchar
58\relax}}=\frac{1}{k!}\,\int_{({\mathbb{R}}^{d})^{k-1}}h^{(i)}({\bf
0}_{d},{\bf y})\operatorname{d\\!}{\bf y},\ \ \ i=1,\dots,m,$
with ${\bf y}\in({\mathbb{R}}^{d})^{k-1}$.
###### Remark 3.6.
If $m=1$ and $h^{(1)}$ is an indicator function, one can explicitly solve the
variational problem in (3.10), to get that
$I_{k}(x)=x\log(x/\mu_{1})-x+\mu_{1}$. In this case, $I_{k}$ coincides with a
rate function in the LDP for $\big{(}n^{-1}\sum_{i=1}^{n}X_{i}\big{)}_{n\geq
1}$, where $(X_{i})$ are i.i.d. Poisson random variables with mean $\mu_{1}$.
###### Example 3.7 (Čech complex component counts).
We consider an application to the Čech complex component counts. Let
$\check{C}({\mathcal{X}},r)$ be the Čech complex on a point set
${\mathcal{X}}=\\{x_{1},\dots,x_{m}\\}\subset{\mathbb{R}}^{d}$ with
connectivity radius $r>0$. Namely,
* •
The $0$-simplices of $\check{C}({\mathcal{X}},r)$ are the points in
${\mathcal{X}}$.
* •
The $p$-simplex $\\{x_{i_{0}},\dots,x_{i_{p}}\\}\subset{\mathcal{X}}$ with
$1\leq i_{0}<\dots<i_{p}\leq m$, belongs to $\check{C}({\mathcal{X}},r)$ if
$\bigcap_{\ell=0}^{p}B(x_{i_{\ell}},r/2)\neq\emptyset$, where $B(x,r)$ is the
$d$-dimensional closed ball of radius $r$ centered at $x\in{\mathbb{R}}^{d}$.
We then explore the LDP for
$S_{k,n}\mathrel{\mathop{\mathchar
58\relax}}=\Big{(}\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n},\,|{\mathcal{Y}}|=k+1}{\mathbbm{1}}\big{\\{}\check{C}({\mathcal{Y}},r_{n}t_{i})\cong\Gamma_{i}\big{\\}}\,c_{n}({\mathcal{Y}},\mathcal{P}_{n};t_{i})\Big{)}_{i=1}^{m},$
(3.11)
where $\Gamma_{i}$ is a _connected_ Čech complex on $k+1$ vertices, and
$\cong$ means isomorphism between simplicial complexes, and $c_{n}$ is defined
in (2.2). Assume that $\Gamma_{i}\not\cong\Gamma_{j}$ for $i\neq j$. Then, the
$i$th entry of $S_{k,n}$ represents the number of components isomorphic to
$\Gamma_{i}$ in the Čech complex $\check{C}(\mathcal{P}_{n},r_{n}t_{i})$. In
this setting, it is easy to check that the function
$H=(h^{(1)},\dots,h^{(m)})$ with
$h^{(i)}(x_{1},\dots,x_{k+1})={\mathbbm{1}}\Big{\\{}\check{C}\big{(}\\{x_{1},\dots,x_{k+1}\\},t_{i}\big{)}\cong\Gamma_{i}\Big{\\}},\
\ \ (x_{1},\dots,x_{k+1})\in({\mathbb{R}}^{d})^{k+1},$
satisfies conditions (H1)–(H5).
Finally, we also define $S_{k,n}^{\mathsf{B}}$ to be the statistics analogous
to (3.11), that are generated by the binomial point process $\mathcal{B}_{n}$.
The following corollary can be obtained as a direct application of the above
results.
###### Corollary 3.8.
Assume that $\rho_{k+1,n}\to\infty$ and $nr_{n}^{d}\to 0$ as $n\to\infty$.
Then, for any measurable $A\in\mathcal{D}$, as $n\to\infty$,
$\displaystyle\frac{1}{\rho_{k+1,n}}\log\mathbb{P}\big{(}\rho_{k+1,n}^{-1}S_{k,n}\in
A\big{)}\to-\inf_{{\bf x}\in A}I_{k+1}({\bf x}),$
$\displaystyle{\frac{1}{\rho_{k+1,n}}\log\mathbb{P}\big{(}\rho_{k+1,n}^{-1}S_{k,n}^{\mathsf{B}}\in
A\big{)}\to-\inf_{{\bf x}\in A}I_{k+1}({\bf x}).}$
Here, $I_{k+1}$ is the rate function from (3.10). The unique minimizer of
$I_{k+1}$ equals $(\mu_{1},\dots,\mu_{m})$, where
$\mu_{i}=\frac{1}{(k+1)!}\,\int_{({\mathbb{R}}^{d})^{k}}{\mathbbm{1}}\big{\\{}\check{C}\big{(}\\{{\bf
0}_{d},{\bf y}\\},t_{i}\big{)}\cong\Gamma_{i}\big{\\}}\operatorname{d\\!}{\bf
y},\ \ i=1,\dots,m,$
with ${\bf y}\in({\mathbb{R}}^{d})^{k}$.
###### Example 3.9 (Component counts in a random geometric graph).
We next deal with the component counts in a random geometric graph
$G(\mathcal{P}_{n},r_{n}t_{i})$ with radius $r_{n}t_{i}$, $i=1,\dots,m$.
Define the function $H=(h^{(1)},\dots,h^{(m)})$ by
$h^{(i)}(x_{1},\dots,x_{k})={\mathbbm{1}}\Big{\\{}G\big{(}\\{x_{1},\dots,x_{k}\\},t_{i}\big{)}\cong\Gamma_{i}\Big{\\}},\
\ \ (x_{1},\dots,x_{k})\in({\mathbb{R}}^{d})^{k},$
where $\Gamma_{i}$ is a _connected_ graph on $k$ vertices and $\cong$ denotes
a graph isomorphism. Assume that $\Gamma_{i}\not\cong\Gamma_{j}$ for $i\neq
j$. Then, the collection of component counts defined by
$S_{k,n}\mathrel{\mathop{\mathchar
58\relax}}=\Big{(}\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n},\,|{\mathcal{Y}}|=k}{\mathbbm{1}}\big{\\{}G({\mathcal{Y}},r_{n}t_{i})\cong\Gamma_{i}\big{\\}}\,c_{n}({\mathcal{Y}},\mathcal{P}_{n};t_{i})\Big{)}_{i=1}^{m},$
satisfies the following LDP.
###### Corollary 3.10.
Assume that $\rho_{k,n}\to\infty$ and $nr_{n}^{d}\to 0$ as $n\to\infty$. Then,
for any measurable $A\in\mathcal{D}$, as $n\to\infty$,
$\displaystyle\frac{1}{\rho_{k,n}}\log\mathbb{P}\big{(}\rho_{k,n}^{-1}S_{k,n}\in
A\big{)}\to-\inf_{{\bf x}\in A}I_{k}({\bf x}),$
$\displaystyle{\frac{1}{\rho_{k,n}}\log\mathbb{P}\big{(}\rho_{k,n}^{-1}S_{k,n}^{\mathsf{B}}\in
A\big{)}\to-\inf_{{\bf x}\in A}I_{k}({\bf x}).}$
Once again, $I_{k}$ is the rate function with its unique minimizer
$(\mu_{1},\dots,\mu_{m})$ given by
$\mu_{i}=\frac{1}{k!}\,\int_{({\mathbb{R}}^{d})^{k-1}}{\mathbbm{1}}\big{\\{}G\big{(}\\{{\bf
0}_{d},{\bf y}\\},t_{i}\big{)}\cong\Gamma_{i}\big{\\}}\operatorname{d\\!}{\bf
y},\ \ i=1,\dots,m,$
where ${\bf y}\in({\mathbb{R}}^{d})^{k-1}$.
## 4 Applications in stochastic topology
In this section, we elucidate how to apply our main results to derive LDPs for
two key quantities in stochastic topology, namely _persistent Betti numbers_
(Section 4.1) and _Morse critical points_ (Section 4.2). In both cases, we
assume $d=2$ and $k=3$ in the notation of Section 2. In other words, we work
with 1-dimensional topological features generated by $3$ random points in the
plane. Although the proposed methods can also be applied to $0$-dimensional
features in the plane, we focus only on $1$-dimensional quantities, because
they are much more non-trivial than $0$-dimensional ones. We conjecture that
our findings can be generalized, at least partially, to higher-dimensional
cases. The remark after Theorem 4.1 below explains why such extensions are
challenging.
### 4.1 Persistent Betti number of alpha complex
First, we deal with the persistent Betti numbers in the alpha complex. For
readers not familiar with algebraic topology, we briefly discuss conceptual
ideas behind persistent Betti numbers. We suggest [10, 18] as a good
introductory reading, while a more rigorous coverage of algebraic topology is
in [25].
The Betti numbers $(\beta_{k})_{k\geq 0}$ are fundamental invariants of
topological spaces counting the number of $k$-dimensional cycles (henceforth
we call it $k$-cycle) as the boundary of a $(k+1)$-dimensional body. In the
$3$-dimensional space, $\beta_{1}$ and $\beta_{2}$ can be viewed as the number
of loops and cavities, respectively. Figure 1 illustrates a sphere in
${\mathbb{R}}^{3}$, which encompasses one central cavity; therefore,
$\beta_{2}=1$. The Betti number $\beta_{1}$ of this sphere is zero; even if we
wind a closed loop around the sphere, the loop ultimately vanishes as it moves
upward (or downward) along the sphere until the pole. Figure 1 also
illustrates a torus in ${\mathbb{R}}^{3}$, for which there are two distinct
non-contractible loops; therefore, $\beta_{1}=2$. Moreover, the torus has a
cavity, meaning that $\beta_{2}=1$.
Figure 1: Illustration of a sphere (left) and torus (right).
We consider the case when the (persistent) Betti numbers are built on the
alpha complex. The alpha complex is similar, in nature, to the Čech complex
(see Example 3.7), but it has a more natural geometric realization. Given a
finite set ${\mathcal{X}}$ of points in ${\mathbb{R}}^{d}$ and
$r\in[0,\infty]$, the alpha complex $\alpha({\mathcal{X}},r)$ is defined as a
collection of subsets $\sigma\subset{\mathcal{X}}$ such that
$\bigcap_{x\in\sigma}\big{(}B(x,r/2)\cap V_{x}\big{)}\neq\emptyset$, where
$V_{x}$ is a Voronoi cell of $x$; that is,
$V_{x}=\big{\\{}y\in{\mathbb{R}}^{d}\mathrel{\mathop{\mathchar
58\relax}}\|y-x\|\leq\inf_{z\in{\mathcal{X}}}\|y-z\|\\}$. Clearly, it holds
that $\alpha({\mathcal{X}},r)\subset\check{C}({\mathcal{X}},r)$. Moreover, we
have inclusions
$\alpha({\mathcal{X}},r)\subset\alpha({\mathcal{X}},r^{\prime})$ for all
$r\leq r^{\prime}$, which indicates that $\alpha({\mathcal{X}},r)$ is a
subcomplex of the Delaunay complex
$\text{Del}({\mathcal{X}})\mathrel{\mathop{\mathchar
58\relax}}=\alpha({\mathcal{X}},\infty)$. This property plays a crucial role
in our analysis; see Remark 4.2.
Returning to the setup in Section 2, we consider the filtration induced by a
collection of alpha complexes over a scaled Poisson point process
$r_{n}^{-1}\mathcal{P}_{n}$,
$\big{(}\alpha(r_{n}^{-1}\mathcal{P}_{n},t),\,t\geq
0\big{)}=\big{(}\alpha(\mathcal{P}_{n},r_{n}t),\,t\geq 0\big{)}.$ (4.1)
Topological changes in the filtration (4.1) can be captured by the $k$th
persistent Betti number $\beta_{k,n}(s,t)$ for $0\leq s\leq t\leq\infty$.
Loosely speaking, $\beta_{k,n}(s,t)$ represents the number of $k$-cycles, that
appear in (4.1) before time $r_{n}s$ and remain alive at time $r_{n}t$. More
formally, $\beta_{k,n}(s,t)$ is defined as
$\beta_{k,n}(s,t)=\text{dim}\frac{Z_{k}\big{(}\alpha(\mathcal{P}_{n},r_{n}s)\big{)}}{Z_{k}\big{(}\alpha(\mathcal{P}_{n},r_{n}s)\big{)}\cap
B_{k}\big{(}\alpha(\mathcal{P}_{n},r_{n}t)\big{)}},$
where $Z_{k}(\cdot)$ is the $k$th cycle group of an alpha complex and
$B_{k}(\cdot)$ denotes its $k$th boundary group. Here, the homology
coefficients are taken from an arbitrary field. In the special case $s=t$,
$\beta_{k,n}(s,t)$ reduces to the ordinary $k$th Betti number.
As mentioned before, in the following we restrict ourselves to a lower-
dimensional case $d=2$ and $k=3$ (in the notation of Section 2). For
$(x_{1},x_{2},x_{3})\in({\mathbb{R}}^{2})^{3}$ and $r>0$, define
$\displaystyle h_{r}(x_{1},x_{2},$ $\displaystyle
x_{3})\mathrel{\mathop{\mathchar
58\relax}}={\mathbbm{1}}\Big{\\{}\beta_{1}\big{(}\alpha(\\{x_{1},x_{2},x_{3}\\},r)\big{)}=1\Big{\\}}$
(4.2) $\displaystyle={\mathbbm{1}}\bigg{\\{}\Big{\\{}\bigcap_{j=1,\,j\neq
j_{0}}^{3}B(x_{j},r/2)\neq\emptyset\ \text{for all
}j_{0}\in\\{1,2,3\\}\Big{\\}}\cap\Big{\\{}\bigcap_{j=1}^{3}B(x_{j},r/2)=\emptyset\Big{\\}}\bigg{\\}}.$
For a subset ${\mathcal{Y}}$ of $3$ points in ${\mathbb{R}}^{2}$, a finite set
$\mathcal{Z}\supset{\mathcal{Y}}$ in ${\mathbb{R}}^{2}$, and $0\leq s\leq
t<\infty$,
$\displaystyle g_{r_{n}s,r_{n}t}({\mathcal{Y}},\mathcal{Z})$
$\displaystyle\mathrel{\mathop{\mathchar
58\relax}}=h_{r_{n}s}({\mathcal{Y}})\,h_{r_{n}t}({\mathcal{Y}})\,{\mathbbm{1}}\big{\\{}\|y-z\|\geq
r_{n}t,\text{ for all }y\in{\mathcal{Y}}\text{ and
}z\in\mathcal{Z}\setminus{\mathcal{Y}}\big{\\}}$
$\displaystyle=h_{r_{n}s}({\mathcal{Y}})\,h_{r_{n}t}({\mathcal{Y}})\,{\mathbbm{1}}\big{\\{}\alpha({\mathcal{Y}},r_{n}t)\text{
is a connected component of }\alpha(\mathcal{Z},r_{n}t)\big{\\}}.$
Note that $g_{r_{n}s,r_{n}t}({\mathcal{Y}},\mathcal{Z})=1$ if and only if a
set ${\mathcal{Y}}$ in ${\mathbb{R}}^{2}$ with $|{\mathcal{Y}}|=3$ forms a
single $1$-cycle before time $r_{n}s$, such that this $1$-cycle remains alive
at time $r_{n}t$ and isolated from all the remaining points in $\mathcal{Z}$
at that time. The proof of the next theorem is presented in Section 5.4.
###### Theorem 4.1.
Assume that $\rho_{3,n}\to\infty$ and $nr_{n}^{2}\to 0$ as $n\to\infty$. Then,
for every $0\leq s_{i}\leq t_{i}<\infty$, $i=1,\dots,m$, with
$(s_{i},t_{i})\neq(s_{j},t_{j})$ for $i\neq j$, and $A\in\mathcal{D}$,
$\frac{1}{\rho_{3,n}}\log\mathbb{P}\Big{(}\big{(}\rho_{3,n}^{-1}\beta_{1,n}(s_{i},t_{i}),\,i=1,\dots,m\big{)}\in
A\Big{)}\to-\inf_{{\bf x}\in A}I_{3}({\bf x}),\ \ \text{as }n\to\infty,$ (4.3)
where $I_{3}$ is the rate function from (3.10). The unique minimizer of
$I_{3}$ equals $(\mu_{1},\dots,\mu_{m})$, where
$\mu_{i}=\frac{1}{6}\,\int_{({\mathbb{R}}^{2})^{2}}h_{s_{i}}({\bf 0}_{2},{\bf
y})\,h_{t_{i}}({\bf 0}_{2},{\bf y})\operatorname{d\\!}{\bf y},\ \
i=1,\dots,m,$
with ${\bf 0}_{2}=(0,0)\in{\mathbb{R}}^{2}$ and ${\bf
y}\in({\mathbb{R}}^{2})^{2}$.
Moreover, define $\beta_{1,n}^{\mathsf{B}}(s,t)$ to be the first-order
persistent Betti number generated by the binomial point process. Then,
$(\beta_{1,n}^{\mathsf{B}})_{n\geq 1}$ satisfies the same LDP as (4.3).
###### Remark 4.2 (Extensions to higher-dimensions).
In the proof of Theorem 4.1, we shall exploit the Morse inequality, which
enables to bound the first-order persistent Betti number by the number of
$1$-simplex counts (i.e., edge counts). A key property for dealing with the
corresponding exponential moments is that the $1$-simplex count in the planar
Delaunay triangulation grows at most linearly in the number of vertices (see
Equ. (5.47) for details). Unfortunately, this property breaks down in higher-
dimensions ([2]); this is the reason why our discussion must be restricted to
a special case $d=2$, $k=3$. Notice, however, that one needs this assumption
only for showing (5.47); the rest of our analyses holds true for general $d$
and $k$. For the LDP of higher-order persistent Betti numbers in a higher-
dimensional space, we need to develop a new machinery that does not rely on
the Morse inequality. More concretely, we need to detect the parameters $d$
and $k$, such that the (persistent) Betti number grows at most linearly in the
number of vertices.
### 4.2 Morse critical points of min-type distance functions
The objective of this example is to deduce the LDP for the number of Morse
critical points of a certain min-type distance function. The behavior of Morse
critical points of such distance functions have been intensively investigated
in the context of central limit theorems and the Poisson process approximation
[6, 8, 9, 38]. In addition to its intrinsic interests, this concept has served
as a practical quantifier of homological changes in random geometric
complexes, especially in the field of Topological Data Analysis ([6, 7]).
Once again, we treat only the special case $d=2$ and $k=3$ (in the notation of
Section 2). Given a homogeneous Poisson point process $\mathcal{P}_{n}$ on
$[0,1]^{2}$ with intensity $n$, we define min-type distance functions by
$d_{\mathcal{P}_{n}}(x)\mathrel{\mathop{\mathchar
58\relax}}=\min_{y\in\mathcal{P}_{n}}\|x-y\|,\ \ \ x\in{\mathbb{R}}^{2}.$
Though $d_{\mathcal{P}_{n}}$ is not differentiable, one can still define the
notion of Morse critical points in the following sense. A point
$c\in{\mathbb{R}}^{2}$ is said to be a Morse critical point of
$d_{\mathcal{P}_{n}}$ with index $2$ if there exists a subset
${\mathcal{Y}}\subset\mathcal{P}_{n}$ of three points such that
$(i)$ The points in ${\mathcal{Y}}$ are in general position.
$(ii)$ $d_{\mathcal{P}_{n}}(c)=\|c-y\|$ for all $y\in{\mathcal{Y}}$ and
$d_{\mathcal{P}_{n}}(c)<\min_{z\in\mathcal{P}_{n}\setminus{\mathcal{Y}}}\|c-z\|$.
$(iii)$ The interior of the convex hull spanned by the points in
${\mathcal{Y}}$, denoted $\text{conv}^{\circ}({\mathcal{Y}})$, contains $c$.
By the Nerve theorem (see, e.g., Theorem 10.7 in [4]), for each $r>0$ the
sublevel set $d_{\mathcal{P}_{n}}(-\infty,r]$ is homotopy equivalent to a Čech
complex $\check{C}(\mathcal{P}_{n},2r)$. For this reason, the number of
critical points of $d_{\mathcal{P}_{n}}$ with index $2$ whose critical values
are less than $r_{n}$, behaves very similarly to the first-order Betti number
of $\check{C}(\mathcal{P}_{n},2r_{n})$ (see [6]). A similar analysis was
conducted for the case that distributions are supported on a closed manifold
embedded in the ambient Euclidean space [8]. Moreover, [38] studied the case
when random points are sampled from a stationary point process.
Given a point set ${\mathcal{Y}}\subset{\mathbb{R}}^{2}$ with
$|{\mathcal{Y}}|=3$ in general position, let $\gamma({\mathcal{Y}})$ be the
center of a unique $1$-dimensional sphere containing ${\mathcal{Y}}$, and
define
$\mathcal{R}({\mathcal{Y}})=d_{\mathcal{P}_{n}}\big{(}\gamma({\mathcal{Y}})\big{)}$.
If $\gamma({\mathcal{Y}})$ gives a critical point of $d_{\mathcal{P}_{n}}$
(with index $2$), then $\mathcal{R}({\mathcal{Y}})$ represents its critical
value. Additionally, $\mathcal{U}({\mathcal{Y}})$ denotes an open ball in
${\mathbb{R}}^{2}$ with radius $\mathcal{R}({\mathcal{Y}})$ centered at
$\gamma({\mathcal{Y}})$. Given $t_{i}\in[0,\infty)$, $i=1,\dots,m$, with
$t_{i}\neq t_{j}$ for $i\neq j$, we define
$N_{n}\mathrel{\mathop{\mathchar
58\relax}}=\Big{(}\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n},\,|{\mathcal{Y}}|=3}{\mathbbm{1}}\big{\\{}\gamma({\mathcal{Y}})\in\text{conv}^{\circ}({\mathcal{Y}}),\
\mathcal{R}({\mathcal{Y}})\leq r_{n}t_{i},\
\mathcal{U}({\mathcal{Y}})\cap\mathcal{P}_{n}=\emptyset\big{\\}}\Big{)}_{i=1}^{m}.$
(4.4)
In particular, the $i$th entry of $N_{n}$ represents the number of Morse
critical points of index $2$ with critical values less than or equal to
$r_{n}t_{i}$.
The result below is essentially a consequence of Theorem 3.3 and Proposition
3.5. Unlike Theorem 4.1, we do not provide the result for the version of
binomial point processes; the required extension actually involves more
complicated machinery. A formal proof is given in Section 5.4.
###### Theorem 4.3.
Assume that $\rho_{3,n}\to\infty$ and $nr_{n}^{2}\to 0$ as $n\to\infty$. Then,
for every measurable $A\in\mathcal{D}$,
$\frac{1}{\rho_{3,n}}\,\log\mathbb{P}\big{(}\rho_{3,n}^{-1}N_{n}\in
A\big{)}\to-\inf_{{\bf x}\in A}I_{3}({\bf x}),\ \ \text{as }n\to\infty,$
where $I_{3}$ is the rate function from (3.10). The unique minimizer of
$I_{3}$ equals $(\mu_{1},\dots,\mu_{m})$, where
$\mu_{i}=\frac{1}{6}\,\int_{({\mathbb{R}}^{2})^{2}}{\mathbbm{1}}\big{\\{}\gamma({\bf
0}_{2},{\bf y})\in\text{conv}^{\circ}({\bf 0}_{2},{\bf y}),\,\mathcal{R}({\bf
0}_{2},{\bf y})\leq t_{i}\big{\\}}\operatorname{d\\!}{\bf y},\ \ i=1,\dots,m,$
with ${\bf y}\in({\mathbb{R}}^{2})^{2}$.
## 5 Proofs
As preparation, we partition the unit cube $[0,1]^{d}$ into $\rho_{k,n}$ sub-
cubes $Q_{1},\dots,Q_{\rho_{k,n}}$ of volume $\rho_{k,n}^{-1}$. To avoid
notational complication we assume that $\rho_{k,n}$ takes only positive
integers for all $n\in{\mathbb{N}}$. This assumption applies to many of the
sequences and functions throughout this section. In particular, we set
$Q_{1}=[0,\rho_{k,n}^{-1/d}]^{d}$. Given a finite set ${\mathcal{Y}}$ of $k$
points in ${\mathbb{R}}^{d}$, we take $\ell({\mathcal{Y}})$ to be the left
most point of ${\mathcal{Y}}$ in the lexicographic ordering. As discussed in
Section 2, one may set $\ell({\mathcal{Y}})$ to be a center of the unique
$(k-2)$-dimensional sphere containing ${\mathcal{Y}}$. In this case, however,
the description will be slightly more involved. Hence, for ease of
description, we prefer to define $\ell({\mathcal{Y}})$ as the left most point.
In the below, $C^{*}$ denotes a generic constant, which is independent of $n$
and may vary between and within the lines.
### 5.1 Proofs of Theorem 2.1, Proposition 2.2, and Corollary 2.3
The proof of Theorem 2.1 can be completed via Propositions 5.1 and 5.6 below.
Proposition 5.1 is aimed to give the LDP for
$\eta_{k,n}=\frac{1}{\rho_{k,n}}\sum_{\ell=1}^{\rho_{k,n}}\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}|_{Q_{\ell}},\,|{\mathcal{Y}}|=k}s_{n}({\mathcal{Y}},\mathcal{P}_{n}|_{Q_{\ell}};t)\,\delta_{r_{n}^{-1}\overline{\mathcal{Y}}}=\mathrel{\mathop{\mathchar
58\relax}}\frac{1}{\rho_{k,n}}\sum_{\ell=1}^{\rho_{k,n}}\eta_{k,n}^{(\ell)}\in
M_{+}(E),$ (5.1)
where $\mathcal{P}_{n}|_{Q_{\ell}}$ denotes the restriction of
$\mathcal{P}_{n}$ to the cube $Q_{\ell}$. Setting up a “blocked” point process
as in (5.1) is a standard approach in the literature ([36, 35]). The process
(5.1) is of course different from the original process $(\xi_{k,n})_{n\geq
1}$. In fact, if geometric/topological objects exist spreading over multiple
cubes in $(Q_{\ell})_{\ell\geq 1}$, then these objects are counted possibly by
$(\xi_{k,n})_{n\geq 1}$, whereas they are never counted by
$(\eta_{k,n})_{n\geq 1}$. Despite such a difference, one may justify in
Proposition 5.6 that the difference between $(\xi_{k,n})_{n\geq 1}$ and
$(\eta_{k,n})_{n\geq 1}$ is exponentially negligible in terms of the total
variation distance. By virtue of this proposition, as well as Theorem 4.2.13
in [14], our task can be reduced to prove the LDP for $(\eta_{k,n})_{n\geq
1}$.
The main benefit of working with $(\eta_{k,n})_{n\geq 1}$ is that it breaks
down into a collection of i.i.d. point processes
$(\eta_{k,n}^{(\ell)})_{\ell\geq 1}$. Note that
$(\eta_{k,n}^{(\ell)})_{\ell\geq 1}$ are defined in the space $M_{p}(E)$ of
_finite_ point measures on $E$. We here equip $M_{p}(E)$ with a Borel
$\sigma$-field $\mathcal{M}_{p}(E)\mathrel{\mathop{\mathchar
58\relax}}=\mathcal{M}_{+}(E)\cap M_{p}(E)$. Because of this decomposition,
one can clarify the correspondence between (5.1) and the process
$\zeta_{k,n}\mathrel{\mathop{\mathchar
58\relax}}=\frac{1}{\rho_{k,n}}\sum_{\ell=1}^{\rho_{k,n}}\zeta_{k}^{(\ell)}\in
M_{+}(E),\ \ \ n=1,2,\dots,$ (5.2)
where $(\zeta_{k}^{(\ell)})_{\ell\geq 1}$ are i.i.d. Poisson random measures
on $E$ with intensity measure $\tau_{k}$. Write
$(\Omega^{\prime},\mathcal{F}^{\prime},\mathbb{P}^{\prime})$ for the
probability space on which (5.2) is defined, and $\mathbb{E}^{\prime}$ denotes
the corresponding expectation.
In Lemma 5.2 below, we first prove that (5.2) satisfies the required LDP in
Theorem 2.1. After that, Lemmas 5.3 and 5.5 are aimed to establish exponential
equivalence between (5.1) and (5.2). Now, we formally state one of the main
results of this section.
###### Proposition 5.1.
The sequence $(\eta_{k,n})_{n\geq 1}$ fulfills an LDP in the weak topology
with rate $\rho_{k,n}$ and the rate function $\Lambda_{k}^{*}$ in (2.5).
Since the proof of Proposition 5.1 is rather long, we divide it into several
lemmas. Combining these lemmas can conclude Proposition 5.1.
###### Lemma 5.2.
The sequence $(\zeta_{k,n})_{n\geq 1}$ satisfies an LDP in the weak topology
with rate $\rho_{k,n}$ and rate function $\Lambda_{k}^{*}$.
###### Proof.
It follows from Theorem 5.1 in [34] that for every $f\in C_{b}(E)$,
$\displaystyle\Lambda_{k}(f)$ $\displaystyle\mathrel{\mathop{\mathchar
58\relax}}=\log\mathbb{E}^{\prime}\Big{[}e^{\zeta_{k}^{(1)}(f)}\Big{]}=\frac{1}{k!}\int_{({\mathbb{R}}^{d})^{k-1}}\big{(}e^{f(\overline{({\bf
0}_{d},{\bf y})})}-1\big{)}\,{\mathbbm{1}}\big{\\{}\mathsf{diam}({\bf
0}_{d},{\bf y})\leq L\big{\\}}\operatorname{d\\!}{\bf y}.$
According to Cramér’s theorem in Polish spaces in [14, Theorem 6.1.3] (see
also [14, Corollary 6.2.3]), it turns out that $(\zeta_{k,n})_{n\geq 1}$
satisfies a _weak_ LDP in $M_{+}(E)$ with rate function $\Lambda_{k}^{*}$.
In order to extend this to a _full_ LDP, we need to demonstrate that
$(\zeta_{k,n})_{n\geq 1}$ is exponentially tight in the space $M_{+}(E)$. The
proof is analogous to that in Lemma 6.2.6 of [14]. Since $\tau_{k}$ is tight
in the space $E$, for every $\ell\geq 1$ there exists a compact subset
$\Gamma_{\ell}\subset E$, such that
$\tau_{k}(E\setminus\Gamma_{\ell})\leq\ell/(e^{2\ell^{2}}-1)$. Define
$K_{\ell}\mathrel{\mathop{\mathchar 58\relax}}=\big{\\{}\nu\in
M_{+}(E)\mathrel{\mathop{\mathchar
58\relax}}\nu(E\setminus\Gamma_{\ell})\leq\ell^{-1},\,\nu(E)\leq\ell\big{\\}}.$
By Portmanteau’s theorem for weak convergence, one can deduce that $K_{\ell}$
is weakly closed in $M_{+}(E)$. For $m\geq 1$, let
$L_{m}\mathrel{\mathop{\mathchar
58\relax}}=\bigcap_{\ell=m}^{\infty}K_{\ell}$. Obviously, $L_{m}$ is weakly
closed in $M_{+}(E)$. Furthermore, Prohorov’s theorem (see, e.g., Theorem
A2.4.I in [12]) ensures that $L_{m}$ is relatively compact. Since $L_{m}$ is
now compact in $M_{+}(E)$, the exponential tightness of $(\zeta_{k,n})_{n\geq
1}$ follows from
$\limsup_{n\to\infty}\frac{1}{\rho_{k,n}}\log\mathbb{P}^{\prime}\big{(}\zeta_{k,n}\in
M_{+}(E)\setminus L_{m}\big{)}\leq-\big{(}m-\tau_{k}(E)(e-1)\big{)},$ (5.3)
for every $m\geq 1$. By the union bound,
$\mathbb{P}^{\prime}\big{(}\zeta_{k,n}\in M_{+}(E)\setminus
L_{m}\big{)}\leq\sum_{\ell=m}^{\infty}\Big{\\{}\mathbb{P}^{\prime}\big{(}\zeta_{k,n}(E\setminus\Gamma_{\ell})>\ell^{-1}\big{)}+\mathbb{P}^{\prime}\big{(}\zeta_{k,n}(E)>\ell\big{)}\Big{\\}}.$
By Markov’s inequality and the fact that
$\zeta_{k}^{(1)}(E\setminus\Gamma_{\ell})$ is Poisson distributed with mean
$\tau_{k}(E\setminus\Gamma_{\ell})$,
$\displaystyle\sum_{\ell=m}^{\infty}\mathbb{P}^{\prime}\big{(}\zeta_{k,n}(E\setminus\Gamma_{\ell})>\ell^{-1}\big{)}\leq\sum_{\ell=m}^{\infty}e^{-2\rho_{k,n}\ell}\Big{(}\mathbb{E}^{\prime}\big{[}e^{2\ell^{2}\zeta_{k}^{(1)}(E\setminus\Gamma_{\ell})}\big{]}\Big{)}^{\rho_{k,n}}$
(5.4)
$\displaystyle=\sum_{\ell=m}^{\infty}e^{-2\rho_{k,n}\ell+\rho_{k,n}\tau_{k}(E\setminus\Gamma_{\ell})(e^{2\ell^{2}}-1)}\leq\sum_{\ell=m}^{\infty}e^{-\rho_{k,n}\ell}\leq
2e^{-\rho_{k,n}m}.$
By the similar calculation,
$\sum_{\ell=m}^{\infty}\mathbb{P}^{\prime}\big{(}\zeta_{k,n}(E)>\ell\big{)}\leq
2e^{-(m-\tau_{k}(E)(e-1))\rho_{k,n}}.$ (5.5)
Putting (5.4) and (5.5) together gives (5.3), as desired. ∎
By virtue of Lemma 5.2, our goal is now to show that $(\eta_{k,n})_{n\geq 1}$
in (5.1) satisfies the same LDP as $(\zeta_{k,n})_{n\geq 1}$. As a first step,
we claim that for every $\ell\geq 1$, the total variation distance between the
laws of $\eta_{k,n}^{(\ell)}$ and $\zeta_{k}^{(\ell)}$ tends to $0$ as
$n\to\infty$. Write $\mathcal{L}(\xi)$ for the probability law of a random
element $\xi$.
###### Lemma 5.3.
For every $\ell\geq 1$,
$d_{\mathsf{TV}}\big{(}\mathcal{L}(\eta_{k,n}^{(\ell)}),\mathcal{L}(\zeta_{k}^{(\ell)})\big{)}\mathrel{\mathop{\mathchar
58\relax}}=\sup_{A\in\mathcal{M}_{p}(E)}\big{|}\mathbb{P}(\eta_{k,n}^{(\ell)}\in
A)-\mathbb{P}^{\prime}(\zeta_{k}^{(\ell)}\in A)\big{|}\to 0,\ \ \text{as
}n\to\infty.$ (5.6)
###### Remark 5.4.
One of the possible extensions of Lemma 5.3 seems to consider the case, for
which $(r_{n})_{n\geq 1}$ belongs to the critical regime, i.e., $nr_{n}^{d}\to
c\in(0,\infty)$ as $n\to\infty$. In this case, however, the asymptotic
behaviors of (5.1) and (5.7) below are essentially different. More concretely,
in the critical regime (5.10) no longer holds, since the last term in (5.11)
does not vanish as $n\to\infty$.
###### Proof.
We begin with defining a sequence of random measures,
$\widetilde{\eta}_{k,n}=\frac{1}{\rho_{k,n}}\sum_{\ell=1}^{\rho_{k,n}}\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}|_{Q_{\ell}},\,|{\mathcal{Y}}|=k}{\mathbbm{1}}\big{\\{}\mathsf{diam}({\mathcal{Y}})\leq
r_{n}L\big{\\}}\,\delta_{r_{n}^{-1}\overline{\mathcal{Y}}}=\mathrel{\mathop{\mathchar
58\relax}}\frac{1}{\rho_{k,n}}\sum_{\ell=1}^{\rho_{k,n}}\widetilde{\eta}_{k,n}^{(\ell)},\
\ \ n\geq 1.$ (5.7)
Note that $(\widetilde{\eta}_{k,n}^{(\ell)})_{\ell\geq 1}$ are i.i.d. point
processes in the space $M_{p}(E)$. We first show that
$d_{\mathsf{TV}}\big{(}\mathcal{L}(\widetilde{\eta}_{k,n}^{(\ell)}),\mathcal{L}(\zeta_{k}^{(\ell)})\big{)}\to
0,\ \ \text{as }n\to\infty.$ (5.8)
According to Theorem 3.1 in [13], (5.8) follows if one can verify that
$\sup_{A\mathrel{\mathop{\mathchar 58\relax}}\text{Borel in
}E}\Big{|}\mathbb{E}\big{[}\widetilde{\eta}_{k,n}^{(\ell)}(A)\big{]}-\mathbb{E}^{\prime}\big{[}\zeta^{(\ell)}_{k}(A)\big{]}\Big{|}\to
0,\ \ \ n\to\infty,$
and
$\displaystyle\gamma_{n}$ $\displaystyle\mathrel{\mathop{\mathchar
58\relax}}=\max_{1\leq p\leq
k-1}n^{2k-p}\int_{(Q_{\ell})^{2k-p}}{\mathbbm{1}}\big{\\{}\mathsf{diam}(x_{1},\dots,x_{k})\leq
r_{n}L\big{\\}}$ (5.9)
$\displaystyle\qquad\qquad\qquad\qquad\times{\mathbbm{1}}\big{\\{}\mathsf{diam}(x_{1},\dots,x_{p},x_{k+1},\dots,x_{2k-p})\leq
r_{n}L\big{\\}}\operatorname{d\\!}{\bf x}\to 0,\ \ n\to\infty.$
By the multivariate Mecke formula for Poisson point processes (see, e.g.,
Chapter 4 in [23]),
$\mathbb{E}\big{[}\widetilde{\eta}_{k,n}^{(\ell)}(A)\big{]}=\frac{n^{k}}{k!}\,\int_{(Q_{\ell})^{k}}{\mathbbm{1}}\big{\\{}\mathsf{diam}(x_{1},\dots,x_{k})\leq
r_{n}L,\ r_{n}^{-1}\overline{(x_{1},\dots,x_{k})}\in
A\big{\\}}\operatorname{d\\!}{\bf x}.$
Performing the change of variables $x_{i}=x+r_{n}y_{i-1}$, $i=1,\dots,k$ (with
$y_{0}\equiv 0$),
$\displaystyle\mathbb{E}\big{[}\widetilde{\eta}_{k,n}^{(\ell)}(A)\big{]}$
$\displaystyle=\frac{\rho_{k,n}}{k!}\,\int_{Q_{\ell}}\int_{({\mathbb{R}}^{d})^{k-1}}{\mathbbm{1}}\big{\\{}\mathsf{diam}({\bf
0}_{d},{\bf y})\leq L,\,\overline{({\bf 0}_{d},{\bf y})}\in A\big{\\}}$
$\displaystyle\qquad\qquad\qquad\qquad\times\prod_{i=1}^{k-1}{\mathbbm{1}}\\{x+r_{n}y_{i}\in
Q_{\ell}\\}\operatorname{d\\!}{\bf y}\operatorname{d\\!}x.$
On the other hand,
$\mathbb{E}^{\prime}\big{[}\zeta^{(\ell)}_{k}(A)\big{]}=\tau_{k}(A)=\frac{\rho_{k,n}}{k!}\,\int_{Q_{\ell}}\int_{({\mathbb{R}}^{d})^{k-1}}{\mathbbm{1}}\big{\\{}\mathsf{diam}({\bf
0}_{d},{\bf y})\leq L,\ \overline{({\bf 0}_{d},{\bf y})}\in
A\big{\\}}\operatorname{d\\!}{\bf y}\operatorname{d\\!}x.$
Thus, as $n\to\infty$,
$\displaystyle\sup_{A}\Big{|}\mathbb{E}\big{[}\widetilde{\eta}_{k,n}^{(\ell)}(A)\big{]}-\mathbb{E}^{\prime}\big{[}\zeta^{(\ell)}_{k}(A)\big{]}\Big{|}$
$\displaystyle\leq\frac{\rho_{k,n}}{k!}\,\int_{Q_{\ell}}\int_{({\mathbb{R}}^{d})^{k-1}}{\mathbbm{1}}\big{\\{}\mathsf{diam}({\bf
0}_{d},{\bf y})\leq L\big{\\}}$
$\displaystyle\qquad\qquad\times\Big{(}1-\prod_{i=1}^{k-1}{\mathbbm{1}}\\{x+r_{n}y_{i}\in
Q_{\ell}\\}\Big{)}\operatorname{d\\!}{\bf y}\operatorname{d\\!}x\to 0.$
As for (5.9), the change of variables $x_{i}=x+r_{n}y_{i-1}$, $i=1,\dots,2k-p$
(with $y_{0}\equiv 0$), yields that
$\displaystyle\gamma_{n}$ $\displaystyle\mathrel{\mathop{\mathchar
58\relax}}=\max_{1\leq p\leq
k-1}n^{2k-p}r_{n}^{d(2k-p-1)}\int_{Q_{\ell}}\int_{({\mathbb{R}}^{d})^{2k-p-1}}{\mathbbm{1}}\big{\\{}\mathsf{diam}({\bf
0}_{d},y_{1},\dots,y_{k-1})\leq L\big{\\}}$
$\displaystyle\qquad\times{\mathbbm{1}}\big{\\{}\mathsf{diam}({\bf
0}_{d},y_{1},\dots,y_{p-1},y_{k},\dots,y_{2k-p-1})\leq
L\big{\\}}\prod_{i=1}^{2k-p-1}{\mathbbm{1}}\\{x+r_{n}y_{i}\in
Q_{\ell}\\}\operatorname{d\\!}{\bf y}\operatorname{d\\!}x$
$\displaystyle\leq\max_{1\leq p\leq
k-1}(nr_{n}^{d})^{k-p}\int_{({\mathbb{R}}^{d})^{2k-p-1}}\prod_{i=1}^{2k-p-1}{\mathbbm{1}}\big{\\{}\|y_{i}\|\leq
L\big{\\}}\operatorname{d\\!}{\bf y}\to 0,\ \ \ n\to\infty.$
Thus, (5.8) has been established, and it remains to verify that
$d_{\mathsf{TV}}\big{(}\mathcal{L}(\widetilde{\eta}_{k,n}^{(\ell)}),\mathcal{L}(\eta_{k,n}^{(\ell)})\big{)}\to
0,\ \ \ n\to\infty.$ (5.10)
Noting that $\widetilde{\eta}_{k,n}^{(\ell)}$ and $\eta_{k,n}^{(\ell)}$ are
both defined in the same probability space, we have
$d_{\mathsf{TV}}\big{(}\mathcal{L}(\widetilde{\eta}_{k,n}^{(\ell)}),\mathcal{L}(\eta_{k,n}^{(\ell)})\big{)}=\frac{1}{2}\sup_{\|f\|_{\infty}\leq
1}\Big{|}\mathbb{E}\big{[}f(\widetilde{\eta}_{k,n}^{(\ell)})\big{]}-\mathbb{E}\big{[}f(\eta_{k,n}^{(\ell)})\big{]}\Big{|},$
where the supremum is taken over all $f\mathrel{\mathop{\mathchar
58\relax}}M_{p}(E)\to{\mathbb{R}}$ with
$\|f\|_{\infty}\mathrel{\mathop{\mathchar 58\relax}}=\text{esssup}_{x\in
M_{p}(E)}|f(x)|\leq 1$. Whenever
$f(\widetilde{\eta}_{k,n}^{(\ell)})-f(\eta_{k,n}^{(\ell)})$ is non-zero, there
exists a subset ${\mathcal{Y}}$ of $k$ points in
$\mathcal{P}_{n}|_{Q_{\ell}}$, such that $\mathsf{diam}({\mathcal{Y}})\leq
r_{n}L$ and $c_{n}({\mathcal{Y}},\mathcal{P}_{n}|_{Q_{\ell}};t)=0$. This
implies that
$\displaystyle
d_{\mathsf{TV}}\big{(}\mathcal{L}(\widetilde{\eta}_{k,n}^{(\ell)}),\mathcal{L}(\eta_{k,n}^{(\ell)})\big{)}$
$\displaystyle\leq\frac{1}{2}\sup_{\|f\|_{\infty}\leq
1}\mathbb{E}\bigg{[}\,\big{|}\,f(\widetilde{\eta}_{k,n}^{(\ell)})-f(\eta_{k,n}^{(\ell)})\,\big{|}$
$\displaystyle\qquad\times{\mathbbm{1}}\bigg{\\{}\bigcup_{{\mathcal{Y}}\subset\mathcal{P}_{n}|_{Q_{\ell}}}\Big{\\{}\mathsf{diam}({\mathcal{Y}})\leq
r_{n}L,\,c_{n}({\mathcal{Y}},\mathcal{P}_{n}|_{Q_{\ell}};t)=0\Big{\\}}\bigg{\\}}\bigg{]}$
$\displaystyle\leq\mathbb{E}\bigg{[}\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}|_{Q_{\ell}}}{\mathbbm{1}}\big{\\{}\mathsf{diam}({\mathcal{Y}})\leq
r_{n}L\big{\\}}\,\big{(}1-c_{n}({\mathcal{Y}},\mathcal{P}_{n}|_{Q_{\ell}};t)\big{)}\bigg{]}.$
Writing
$\mathcal{B}\big{(}\\{x_{1},\dots,x_{k}\\};r\big{)}\mathrel{\mathop{\mathchar
58\relax}}=\bigcup_{i=1}^{k}B(x_{i},r)$ and appealing to the Mecke formula for
Poisson point processes, the rightmost term above is equal to
$\begin{split}&\frac{n^{k}}{k!}\int_{(Q_{\ell})^{k}}{\mathbbm{1}}\big{\\{}\mathsf{diam}(x_{1},\dots,x_{k})\leq
r_{n}L\big{\\}}\Big{(}1-e^{-n\text{vol}\big{(}\mathcal{B}(\\{x_{1},\dots,x_{k}\\};r_{n}t)\cap
Q_{\ell}\big{)}}\Big{)}\operatorname{d\\!}{\bf x}\\\
&=\frac{\rho_{k,n}}{k!}\int_{Q_{\ell}}\int_{({\mathbb{R}}^{d})^{k-1}}{\mathbbm{1}}\big{\\{}\mathsf{diam}({\bf
0}_{d},{\bf y})\leq L\big{\\}}\prod_{i=1}^{k-1}{\mathbbm{1}}\\{x+r_{n}y_{i}\in
Q_{\ell}\\}\\\
&\qquad\qquad\qquad\times\Big{(}1-e^{-n\text{vol}\big{(}\mathcal{B}(\\{x,x+r_{n}y_{1},\dots,x+r_{n}y_{k-1}\\};r_{n}t)\cap
Q_{\ell}\big{)}}\Big{)}\operatorname{d\\!}{\bf y}\operatorname{d\\!}x\\\
&\leq\frac{1}{k!}\int_{({\mathbb{R}}^{d})^{k-1}}{\mathbbm{1}}\big{\\{}\mathsf{diam}({\bf
0}_{d},{\bf y})\leq
L\big{\\}}\Big{(}1-e^{-nr_{n}^{d}\text{vol}\big{(}\mathcal{B}(\\{{\bf
0}_{d},{\bf y}\\};t)\big{)}}\Big{)}\operatorname{d\\!}{\bf y}.\end{split}$
(5.11)
Since $nr_{n}^{d}\to 0$ as $n\to\infty$, the last integral tends to $0$ as
$n\to\infty$. Now, (5.10) is obtained and the proof of (5.6) is completed. ∎
By the standard argument on the maximal coupling (see [21, Lemma 4.32]), for
every $\ell\geq 1$, there exists a coupling
$(\hat{\eta}_{k,n}^{(\ell)},\hat{\zeta}^{(\ell)}_{k})$ on some probability
space
$(\hat{\Omega}_{\ell},\hat{\mathcal{F}}_{\ell},\hat{\mathbb{P}}_{\ell})$, such
that $\hat{\eta}_{k,n}^{(\ell)}\stackrel{{\scriptstyle
d}}{{=}}\eta_{k,n}^{(\ell)}$, $\hat{\zeta}^{(\ell)}_{k}\stackrel{{\scriptstyle
d}}{{=}}\zeta^{(\ell)}_{k}$, and
$\hat{\mathbb{P}}_{\ell}\big{(}\hat{\eta}_{k,n}^{(\ell)}\neq\hat{\zeta}^{(\ell)}_{k}\big{)}=d_{\mathsf{TV}}\big{(}\mathcal{L}(\eta_{k,n}^{(\ell)}),\mathcal{L}(\zeta_{k}^{(\ell)})\big{)}\to
0,\ \ \ n\to\infty,$ (5.12)
where the last convergence follows from Lemma 5.3. Define
$\hat{\Omega}=\prod_{\ell=1}^{\infty}\hat{\Omega}_{\ell}$,
$\hat{\mathcal{F}}=\bigotimes_{\ell=1}^{\infty}\hat{\mathcal{F}}_{\ell}$,
$\hat{\mathbb{P}}=\bigotimes_{\ell=1}^{\infty}\hat{\mathbb{P}}_{\ell}$, and
$\hat{\mathbb{E}}$ is the corresponding expectation; then
$(\hat{\eta}_{k,n}^{(\ell)},\hat{\zeta}^{(\ell)}_{k})_{\ell\geq 1}$ becomes a
sequence of i.i.d. random vectors under $\hat{\mathbb{P}}$. Define
$\hat{\eta}_{k,n}$ and $\hat{\zeta}_{k,n}$ analogously to (5.1) and (5.2).
According to [14, Theorem 4.2.13], if one can show that
$(\hat{\eta}_{k,n})_{n\geq 1}$ and $(\hat{\zeta}_{k,n})_{n\geq 1}$ are
exponentially equivalent under a coupled probability $\hat{\mathbb{P}}$ (in
terms of the total variation distance), then it will be concluded that
$(\eta_{k,n})_{n\geq 1}$ and $(\zeta_{k,n})_{n\geq 1}$ exhibit the same LDP. A
more precise statement is given as follows.
###### Lemma 5.5.
For every $\delta>0$,
$\lim_{n\to\infty}\frac{1}{\rho_{k,n}}\log\hat{\mathbb{P}}\Big{(}d_{\mathsf{TV}}\big{(}\hat{\eta}_{k,n},\hat{\zeta}_{k,n}\big{)}\geq\delta\Big{)}=-\infty.$
(5.13)
###### Proof.
By Markov’s inequality, we have, for any $a>0$,
$\displaystyle\hat{\mathbb{P}}\Big{(}d_{\mathsf{TV}}\big{(}\hat{\eta}_{k,n},\hat{\zeta}_{k,n}\big{)}\geq\delta\Big{)}$
$\displaystyle\leq\hat{\mathbb{P}}\Big{(}\sum_{\ell=1}^{\rho_{k,n}}d_{\mathsf{TV}}\big{(}\hat{\eta}_{k,n}^{(\ell)},\hat{\zeta}^{(\ell)}_{k}\big{)}\geq\delta\rho_{k,n}\Big{)}$
$\displaystyle\leq
e^{-a\delta\rho_{k,n}}\Big{(}\hat{\mathbb{E}}\Big{[}e^{ad_{\mathsf{TV}}\big{(}\hat{\eta}_{k,n}^{(1)},\,\hat{\zeta}_{k}^{(1)}\big{)}}\Big{]}\Big{)}^{\rho_{k,n}}.$
Hence, (5.13) follows if we can show that, for every $a>0$,
$\lim_{n\to\infty}\hat{\mathbb{E}}\Big{[}e^{ad_{\mathsf{TV}}\big{(}\hat{\eta}_{k,n}^{(1)},\,\hat{\zeta}_{k}^{(1)}\big{)}}\Big{]}=1.$
(5.14)
Because of (5.12), we get that
$d_{\mathsf{TV}}\big{(}\hat{\eta}_{k,n}^{(1)},\,\hat{\zeta}_{k}^{(1)}\big{)}\stackrel{{\scriptstyle\hat{\mathbb{P}}}}{{\to}}0$
as $n\to\infty$. It thus remains to demonstrate that
$\limsup_{n\to\infty}\hat{\mathbb{E}}\Big{[}e^{ad_{\mathsf{TV}}\big{(}\hat{\eta}_{k,n}^{(1)},\,\hat{\zeta}_{k}^{(1)}\big{)}}\Big{]}<\infty,\
\ \text{ for all }a>0.$ (5.15)
In fact, (5.15) ensures the uniform integrability for the convergence (5.14).
By the Cauchy–Schwarz inequality,
$\displaystyle\hat{\mathbb{E}}\Big{[}e^{ad_{\mathsf{TV}}\big{(}\hat{\eta}_{k,n}^{(1)},\,\hat{\zeta}_{k}^{(1)}\big{)}}\Big{]}$
$\displaystyle\leq\Big{\\{}\mathbb{E}\Big{[}e^{2a\eta_{k,n}^{(1)}(E)}\Big{]}\Big{\\}}^{1/2}\Big{\\{}\mathbb{E}^{\prime}\Big{[}e^{2a\zeta_{k}^{(1)}(E)}\Big{]}\Big{\\}}^{1/2}$
$\displaystyle=\Big{\\{}\mathbb{E}\big{[}e^{2a\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}|_{Q_{1}}}s_{n}({\mathcal{Y}},\mathcal{P}_{n}|_{Q_{1}};t)}\big{]}\Big{\\}}^{1/2}\Big{\\{}\exp\big{\\{}\tau_{k}(E)(e^{2a}-1)\big{\\}}\Big{\\}}^{1/2}.$
Now we need to verify that for all $a>0$,
$\limsup_{n\to\infty}\mathbb{E}\Big{[}e^{a\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}|_{Q_{1}}}s_{n}({\mathcal{Y}},\mathcal{P}_{n}|_{Q_{1}};t)}\Big{]}<\infty.$
(5.16)
For the proof of (5.16), we consider the diluted family of cubes
$G\mathrel{\mathop{\mathchar
58\relax}}=\big{\\{}4Lr_{n}z+[0,tr_{n}/\sqrt{d}]^{d}\subset
Q_{1}\mathrel{\mathop{\mathchar 58\relax}}z\in{\mathbb{Z}}^{d}\big{\\}}$
(5.17)
(recall that we have taken $L>t$). Then, $Q_{1}$ can be covered by at most
$(4L\sqrt{d}/t)^{d}$ many translates of $G$. Denote these translates as
$G_{1},\dots,G_{(4L\sqrt{d}/t)^{d}}$ (with $G_{1}\equiv G$). Let
$b_{n}\mathrel{\mathop{\mathchar 58\relax}}=\rho_{k,n}^{-1}/(4Lr_{n})^{d}$
denote the number of cubes (of side length $tr_{n}/\sqrt{d}$) that are
contained in $G$. As mentioned at the beginning of Section 5, we assume,
without loss of generality, that $(4L\sqrt{d}/t)^{d}$ and $b_{n}$ take only
positive integers. Suppose ${\mathcal{Y}}$ is a set of $k$ points with
$\mathsf{diam}({\mathcal{Y}})\leq r_{n}L$, then there is a unique
$j\in\\{1,\dots,(4L\sqrt{d}/t)^{d}\\}$, so that the left most point
$\ell({\mathcal{Y}})$ belongs to one of the cubes in $G_{j}$. Hence,
$\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}|_{Q_{1}}}s_{n}\big{(}{\mathcal{Y}},\mathcal{P}_{n}|_{Q_{1}};t\big{)}=\sum_{j=1}^{(4L\sqrt{d}/t)^{d}}\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}|_{Q_{1}}}s_{n}\big{(}{\mathcal{Y}},\mathcal{P}_{n}|_{Q_{1}};t\big{)}{\mathbbm{1}}\Big{\\{}\ell({\mathcal{Y}})\in\bigcup_{J\in
G_{j}}J\Big{\\}}.$ (5.18)
Write $G_{1}=\\{J_{1},\dots,J_{b_{n}}\\}$ with
$J_{1}=\big{[}0,tr_{n}/\sqrt{d}\big{]}^{d}$. By (5.18), Hölder’s inequality,
and the spatial independence and homogeneity of $\mathcal{P}_{n}$, we need to
demonstrate that
$\limsup_{n\to\infty}\bigg{(}\mathbb{E}\Big{[}e^{a\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}|_{Q_{1}}}s_{n}({\mathcal{Y}},\mathcal{P}_{n}|_{Q_{1}};t){\mathbbm{1}}\\{\ell({\mathcal{Y}})\in
J_{1}\\}}\Big{]}\bigg{)}^{b_{n}}<\infty.$ (5.19)
A key observation for the proof of (5.19) is that there is at most a single
$k$-point subset ${\mathcal{Y}}\subset\mathcal{P}_{n}|_{Q_{1}}$ such that
$c_{n}({\mathcal{Y}},\mathcal{P}_{n}|_{Q_{1}};t)=1$ and
$\ell({\mathcal{Y}})\in J_{1}$. In fact, if there are two distinct point sets
${\mathcal{Y}},{\mathcal{Y}}^{\prime}\subset\mathcal{P}_{n}|_{Q_{1}}$ with
$\ell({\mathcal{Y}})\in J_{1}$ and $\ell({\mathcal{Y}}^{\prime})\in J_{1}$,
then it must be that
$c_{n}({\mathcal{Y}},\mathcal{P}_{n}|_{Q_{1}};t)=c_{n}({\mathcal{Y}}^{\prime},\mathcal{P}_{n}|_{Q_{1}};t)=0$.
It turns out from this observation that
$\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}|_{Q_{1}}}s_{n}\big{(}{\mathcal{Y}},\mathcal{P}_{n}|_{Q_{1}};t\big{)}{\mathbbm{1}}\big{\\{}\ell({\mathcal{Y}})\in
J_{1}\big{\\}}$
is a $\\{0,1\\}$-valued random variable; hence,
$\displaystyle\bigg{(}\mathbb{E}\Big{[}e^{a\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}|_{Q_{1}}}s_{n}({\mathcal{Y}},\mathcal{P}_{n}|_{Q_{1}};t){\mathbbm{1}}\\{\ell({\mathcal{Y}})\in
J_{1}\\}}\Big{]}\bigg{)}^{b_{n}}$
$\displaystyle=\bigg{(}1+(e^{a}-1)\mathbb{P}\Big{(}\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}|_{Q_{1}}}s_{n}\big{(}{\mathcal{Y}},\mathcal{P}_{n}|_{Q_{1}};t\big{)}{\mathbbm{1}}\big{\\{}\ell({\mathcal{Y}})\in
J_{1}\big{\\}}=1\Big{)}\bigg{)}^{b_{n}}$
$\displaystyle\leq\bigg{(}1+(e^{a}-1)\mathbb{E}\Big{[}\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}}{\mathbbm{1}}\big{\\{}\mathsf{diam}({\mathcal{Y}})\leq
r_{n}L,\,\ell({\mathcal{Y}})\in J_{1}\big{\\}}\Big{]}\bigg{)}^{b_{n}}.$
Repeating the same calculations as before, it is not hard to see that
$\mathbb{E}\Big{[}\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}}{\mathbbm{1}}\big{\\{}\mathsf{diam}({\mathcal{Y}})\leq
r_{n}L,\,\ell({\mathcal{Y}})\in
J_{1}\big{\\}}\Big{]}=C^{*}\big{(}\rho_{k,n}r_{n}^{d}+o(1)\big{)}.$
Now, we obtain
$\bigg{(}1+(e^{a}-1)\mathbb{E}\Big{[}\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}}{\mathbbm{1}}\big{\\{}\mathsf{diam}({\mathcal{Y}})\leq
r_{n}L,\,\ell({\mathcal{Y}})\in J_{1}\big{\\}}\Big{]}\bigg{)}^{b_{n}}\to
e^{C^{*}(e^{a}-1)/(4L)^{d}}<\infty,$
as desired. ∎
Now that the proof of Proposition 5.1 has been completed, we next verify that
the difference between $(\xi_{k,n})_{n\geq 1}$ and $(\eta_{k,n})_{n\geq 1}$ in
terms of the total variation distance, is exponentially negligible.
###### Proposition 5.6.
For every $\delta>0$,
$\lim_{n\to\infty}\frac{1}{\rho_{k,n}}\log\mathbb{P}\big{(}d_{\mathsf{TV}}(\xi_{k,n},\eta_{k,n})\geq\delta\big{)}=-\infty.$
(5.20)
###### Proof.
For $\ell=1,\dots,\rho_{k,n}$, define a collection of points in $Q_{\ell}$
that are distance at most $r$ from the boundary of $Q_{\ell}$:
$Q^{\partial}_{\ell}(r)\mathrel{\mathop{\mathchar 58\relax}}=\big{\\{}x\in
Q_{\ell}\mathrel{\mathop{\mathchar 58\relax}}\inf_{y\in\partial
Q_{\ell}}\|x-y\|\leq r\big{\\}},\ \ \ r>0.$
For a subset $A\subset E$, we discuss two distinct cases for which a $k$-point
set ${\mathcal{Y}}\subset[0,1]^{d}$ with $\mathsf{diam}({\mathcal{Y}})\leq
r_{n}L$, makes different contributions to $\xi_{k,n}(A)$ and $\eta_{k,n}(A)$.
The first case is that the point set ${\mathcal{Y}}$ “crosses a boundary”
between two neighboring sub-cubes, i.e., there exist distinct $\ell_{1}$ and
$\ell_{2}$ such that
${\mathcal{Y}}\cap Q_{\ell_{1}}\neq\emptyset,\ \ \ {\mathcal{Y}}\cap
Q_{\ell_{2}}\neq\emptyset.$
Then, ${\mathcal{Y}}$ must be contained in
$\bigcup_{\ell=1}^{\rho_{k,n}}Q^{\partial}_{\ell}\big{(}(L+t)r_{n}\big{)}$
such that
${\mathcal{Y}}\cap\bigcup_{\ell=1}^{\rho_{k,n}}Q_{\ell}^{\partial}(tr_{n})\neq\emptyset$.
Furthermore, ${\mathcal{Y}}$ may increase the value of $\xi_{k,n}(A)$, while
the value of $\eta_{k,n}(A)$ is unchanged. The second case is that there are
two neighboring sub-cubes $Q_{\ell_{1}}$ and $Q_{\ell_{2}}$, together with two
point sets ${\mathcal{Y}}_{i}\subset Q_{\ell_{i}}$, $i=1,2$, of cardinality
$k$ with $\mathsf{diam}({\mathcal{Y}}_{i})\leq r_{n}L$, such that
$\inf_{y_{i}\in{\mathcal{Y}}_{i},\,i=1,2}\|y_{1}-y_{2}\|\leq r_{n}t$. It then
holds that ${\mathcal{Y}}_{i}\subset
Q^{\partial}_{\ell_{i}}\big{(}(L+t)r_{n}\big{)}$ with ${\mathcal{Y}}_{i}\cap
Q_{\ell}^{\partial}(tr_{n})\neq\emptyset$ for $i=1,2$. Moreover,
${\mathcal{Y}}_{i}$ may increase the value of $\eta_{k,n}(A)$, but the value
of $\xi_{k,n}(A)$ is unchanged. Putting these observations together, we
conclude that
$\displaystyle d_{\mathsf{TV}}(\xi_{k,n},\eta_{k,n})$
$\displaystyle\leq\frac{1}{\rho_{k,n}}\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}}s_{n}\big{(}{\mathcal{Y}},\mathcal{P}_{n}|_{\bigcup_{\ell=1}^{\rho_{k,n}}Q_{\ell}^{\partial}((L+t)r_{n})};t\big{)}{\mathbbm{1}}\Big{\\{}{\mathcal{Y}}\cap\bigcup_{\ell=1}^{\rho_{k,n}}Q_{\ell}^{\partial}(tr_{n})\neq\emptyset\Big{\\}}$
(5.21)
$\displaystyle\qquad\qquad+\frac{1}{\rho_{k,n}}\sum_{\ell=1}^{\rho_{k,n}}\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}}s_{n}\big{(}{\mathcal{Y}},\mathcal{P}_{n}|_{Q^{\partial}_{\ell}((L+t)r_{n})};t\big{)}{\mathbbm{1}}\big{\\{}{\mathcal{Y}}\cap
Q_{\ell}^{\partial}(tr_{n})\neq\emptyset\big{\\}}.$
Now, (5.20) will follow if one can show that for every $\delta>0$,
$\displaystyle\lim_{n\to\infty}\frac{1}{\rho_{k,n}}\,\log\mathbb{P}\Big{(}\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}}s_{n}\big{(}{\mathcal{Y}},\mathcal{P}_{n}|_{\bigcup_{\ell=1}^{\rho_{k,n}}Q_{\ell}^{\partial}((L+t)r_{n})};t\big{)}$
(5.22)
$\displaystyle\qquad\qquad\qquad\qquad\qquad\times{\mathbbm{1}}\Big{\\{}{\mathcal{Y}}\cap\bigcup_{\ell=1}^{\rho_{k,n}}Q_{\ell}^{\partial}(tr_{n})\neq\emptyset\Big{\\}}\geq\delta\rho_{k,n}\Big{)}=-\infty,$
and
$\lim_{n\to\infty}\frac{1}{\rho_{k,n}}\,\log\mathbb{P}\Big{(}\sum_{\ell=1}^{\rho_{k,n}}\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}}s_{n}\big{(}{\mathcal{Y}},\mathcal{P}_{n}|_{Q^{\partial}_{\ell}((L+t)r_{n})};t\big{)}{\mathbbm{1}}\big{\\{}{\mathcal{Y}}\cap
Q_{\ell}^{\partial}(tr_{n})\neq\emptyset\big{\\}}\geq\delta\rho_{k,n}\Big{)}=-\infty.$
(5.23)
The proof techniques for (5.22) and (5.23) are similar, so we show only
(5.22). To begin, for each $1\leq j\leq d$, denote a collection of ordered
$j$-tuples by
$\mathcal{I}_{j}=\big{\\{}{\bm{\ell}}=(\ell_{1},\dots,\ell_{j})\mathrel{\mathop{\mathchar
58\relax}}1\leq\ell_{1}<\dots<\ell_{j}\leq d\big{\\}}.$
For ${\bm{\ell}}=(\ell_{1},\dots,\ell_{j})\in\mathcal{I}_{j}$, we define a
collection of disjoint hyper-rectangles by
$\displaystyle J$ $\displaystyle\mathrel{\mathop{\mathchar
58\relax}}=\bigg{\\{}\Big{(}\rho_{k,n}^{-1/d}z+\big{[}0,\rho_{k,n}^{-1/d}\big{]}^{\ell_{1}-1}\times\big{[}-(L+t)r_{n},(L+t)r_{n}\big{]}\times\big{[}0,\rho_{k,n}^{-1/d}\big{]}^{\ell_{2}-\ell_{1}-1}$
(5.24)
$\displaystyle\qquad\times\big{[}-(L+t)r_{n},(L+t)r_{n}\big{]}\times\big{[}0,\rho_{k,n}^{-1/d}\big{]}^{\ell_{3}-\ell_{2}-1}\times\dots\times\big{[}-(L+t)r_{n},(L+t)r_{n}\big{]}$
$\displaystyle\qquad\times\big{[}0,\rho_{k,n}^{-1/d}\big{]}^{\ell_{j}-\ell_{j-1}-1}\times\big{[}-(L+t)r_{n},(L+t)r_{n}\big{]}\times\big{[}0,\rho_{k,n}^{-1/d}\big{]}^{d-\ell_{j}}\Big{)}\cap[0,1]^{d}\mathrel{\mathop{\mathchar
58\relax}}z\in{\mathbb{Z}}_{+}^{d}\bigg{\\}}.$
By construction, all the rectangles in $J$ are contained in
$\bigcup_{\ell=1}^{\rho_{k,n}}Q_{\ell}^{\partial}((L+t)r_{n})$. Moreover,
since the number of rectangles in $J$ is $\rho_{k,n}$, one can enumerate $J$
in a way that $J=(I_{p,n}^{\bm{\ell}},\,p=1,\dots,\rho_{k,n})$. Then, one can
offer the following bound:
$\displaystyle\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}}s_{n}\big{(}{\mathcal{Y}},\mathcal{P}_{n}|_{\bigcup_{\ell=1}^{\rho_{k,n}}Q_{\ell}^{\partial}((L+t)r_{n})};t\big{)}{\mathbbm{1}}\Big{\\{}{\mathcal{Y}}\cap\bigcup_{\ell=1}^{\rho_{k,n}}Q_{\ell}^{\partial}(tr_{n})\neq\emptyset\Big{\\}}$
(5.25)
$\displaystyle\leq\sum_{j=1}^{d}\sum_{{\bm{\ell}}\in\mathcal{I}_{j}}\sum_{p=1}^{\rho_{k,n}}\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}|_{I_{p,n}^{\bm{\ell}}}}s_{n}\big{(}{\mathcal{Y}},\mathcal{P}_{n}|_{I_{p,n}^{\bm{\ell}}};t\big{)}.$
Owing to this bound, it remains to show that for every $j\in\\{1,\dots,d\\}$,
${\bm{\ell}}\in\mathcal{I}_{j}$, and $\delta>0$,
$\lim_{n\to\infty}\frac{1}{\rho_{k,n}}\,\log\mathbb{P}\Big{(}\sum_{p=1}^{\rho_{k,n}}\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}|_{I_{p,n}^{\bm{\ell}}}}s_{n}\big{(}{\mathcal{Y}},\mathcal{P}_{n}|_{I_{p,n}^{\bm{\ell}}};t\big{)}\geq\delta\rho_{k,n}\Big{)}=-\infty.$
(5.26)
Since $(I_{p,n}^{\bm{\ell}},\,p=1,\dots,\rho_{k,n})$ are disjoint, the spatial
independence of $\mathcal{P}_{n}$ ensures that
$\bigg{(}\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}|_{I_{p,n}^{\bm{\ell}}}}s_{n}\big{(}{\mathcal{Y}},\mathcal{P}_{n}|_{I_{p,n}^{\bm{\ell}}};t\big{)},\,p=1,\dots,\rho_{k,n}\bigg{)}$
are i.i.d. random variables. Hence, we have for every $a>0$,
$\displaystyle\mathbb{P}\Big{(}\sum_{p=1}^{\rho_{k,n}}\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}|_{I_{p,n}^{\bm{\ell}}}}s_{n}\big{(}{\mathcal{Y}},\mathcal{P}_{n}|_{I_{p,n}^{\bm{\ell}}};t\big{)}\geq\delta\rho_{k,n}\Big{)}$
$\displaystyle\quad\leq
e^{-a\delta\rho_{k,n}}\bigg{(}\mathbb{E}\Big{[}\exp\Big{\\{}a\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}|_{I_{1,n}^{\bm{\ell}}}}s_{n}\big{(}{\mathcal{Y}},\mathcal{P}_{n}|_{I_{1,n}^{\bm{\ell}}};t\big{)}\Big{\\}}\Big{]}\bigg{)}^{\rho_{k,n}}.$
For the proof of (5.26), it is now sufficient to show that for every $a>0$,
$\mathbb{E}\bigg{[}\exp\Big{\\{}a\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}|_{I_{1,n}^{\bm{\ell}}}}s_{n}\big{(}{\mathcal{Y}},\mathcal{P}_{n}|_{I_{1,n}^{\bm{\ell}}};t\big{)}\Big{\\}}\bigg{]}\to
1,\ \ \text{as }n\to\infty.$ (5.27)
By the same argument as that for (5.16), one can see that for every $a>0$,
$\limsup_{n\to\infty}\mathbb{E}\bigg{[}\exp\Big{\\{}a\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}|_{I_{1,n}^{\bm{\ell}}}}s_{n}\big{(}{\mathcal{Y}},\mathcal{P}_{n}|_{I_{1,n}^{\bm{\ell}}};t\big{)}\Big{\\}}\bigg{]}<\infty,$
which implies the required uniform integrability. Now, (5.27) will follow if
we can verify that
$\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}|_{I_{1,n}^{\bm{\ell}}}}s_{n}\big{(}{\mathcal{Y}},\mathcal{P}_{n}|_{I_{1,n}^{\bm{\ell}}};t\big{)}\stackrel{{\scriptstyle\mathbb{P}}}{{\to}}0,\
\ \text{as }n\to\infty.$ (5.28)
To show this, we have as $n\to\infty$,
$\displaystyle\mathbb{E}\Big{[}\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}|_{I_{1,n}^{\bm{\ell}}}}s_{n}\big{(}{\mathcal{Y}},\mathcal{P}_{n}|_{I_{1,n}^{\bm{\ell}}};t\big{)}\Big{]}\leq\mathbb{E}\Big{[}\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}|_{I_{1,n}^{\bm{\ell}}}}{\mathbbm{1}}\big{\\{}\mathsf{diam}({\mathcal{Y}})\leq
r_{n}L\big{\\}}\Big{]}$
$\displaystyle\qquad\qquad=\frac{n^{k}}{k!}\,\int_{(I_{1,n}^{\bm{\ell}})^{k}}{\mathbbm{1}}\big{\\{}\mathsf{diam}(x_{1},\dots,x_{k})\leq
r_{n}L\big{\\}}\operatorname{d\\!}{\bf x}$
$\displaystyle\qquad\qquad=\frac{\rho_{k,n}}{k!}\,\int_{I_{1,n}^{\bm{\ell}}}\int_{({\mathbb{R}}^{d})^{k-1}}{\mathbbm{1}}\big{\\{}\mathsf{diam}({\bf
0}_{d},{\bf y})\leq
L\big{\\}}\prod_{i=1}^{k-1}{\mathbbm{1}}\big{\\{}x+r_{n}y_{i}\in
I_{1,n}^{\bm{\ell}}\big{\\}}\operatorname{d\\!}{\bf y}\operatorname{d\\!}x$
$\displaystyle\qquad\qquad\leq\frac{\rho_{k,n}}{k!}\,\text{vol}(I_{1,n}^{\bm{\ell}})\int_{({\mathbb{R}}^{d})^{k-1}}{\mathbbm{1}}\big{\\{}\mathsf{diam}({\bf
0}_{d},{\bf y})\leq L\big{\\}}\operatorname{d\\!}{\bf
y}=\mathcal{O}\big{(}(nr_{n}^{d})^{k/d}\big{)}\to 0.$
Hence (5.28) has been established, as desired. ∎
Before concluding this subsection, we present the proof of Proposition 2.2.
###### Proof of Proposition 2.2.
Our proof is closely related to Theorem 5.4 in [32]. Our goal here is to show
that, for every $f\in C_{b}(E)$,
$\int_{E}\big{(}e^{f({\bf x})}-1\big{)}\tau_{k}(\operatorname{d\\!}{\bf
x})=\sup_{\rho\in M_{+}(E)}\big{\\{}\langle\rho,f\rangle-
H_{k}(\rho|\tau_{k})\big{\\}},$ (5.29)
where $\langle\rho,f\rangle\mathrel{\mathop{\mathchar
58\relax}}=\int_{E}f({\bf x})\rho(\operatorname{d\\!}{\bf x})$. Given $f\in
C_{b}(E)$, we define
$\operatorname{d\\!}\rho=e^{f}\operatorname{d\\!}\tau_{k}$. Then, $\rho\in
M_{+}(E)$ with $\rho\ll\tau_{k}$. By (2.7), it is elementary to calculate that
$\langle\rho,f\rangle-H_{k}(\rho|\tau_{k})=\int_{E}\big{(}e^{f({\bf
x})}-1\big{)}\tau_{k}(\operatorname{d\\!}{\bf x})$, which has shown that the
left hand side in (5.29) is bounded by the right hand side.
Next, let us take $\nu\in M_{+}(E)$ with $\nu\ll\tau_{k}$, say with density
$\varphi$. Then, by Jensen’s inequality,
$\int_{E}\log\bigg{(}\frac{e^{f({\bf x})}}{\varphi({\bf
x})}\bigg{)}\,\frac{\nu(d{\bf
x})}{\nu(E)}\leq\log\bigg{(}\int_{E}\frac{e^{f({\bf x})}}{\varphi({\bf
x})}\frac{\nu(d{\bf x})}{\nu(E)}\bigg{)}=\log\bigg{(}\int_{E}e^{f({\bf
x})}\frac{\tau_{k}(d{\bf x})}{\nu(E)}\bigg{)}.$
Hence, by (2.7) and the elementary inequality: $1+\log x\leq x$ for $x>0$,
$\langle\nu,f\rangle-
H_{k}(\nu|\tau_{k}){\leq\nu(E)\bigg{\\{}1+\log\bigg{(}\int_{E}e^{f({\bf
x})}\frac{\tau_{k}(dx)}{\nu(E)}\bigg{)}\bigg{\\}}-\tau_{k}(E)}\leq\int_{E}\big{(}e^{f({\bf
x})}-1\big{)}\tau_{k}(\operatorname{d\\!}{\bf x}).$
Now (5.29) is obtained, and the rest of the argument after (5.29) is
essentially the same as Theorem 5.4 in [32], so we will omit it. ∎
Finally, we prove Corollary 2.3. We can deduce its assertion from Theorem 2.1
by showing that for every $\varepsilon_{0}>0$,
$\lim_{n\to\infty}\rho_{k,n}^{-1}\log\mathbb{P}\big{(}d_{\mathsf{TV}}(\xi_{k,n},\xi_{k,n}^{\mathsf{B}})\geq\varepsilon_{0}\big{)}=-\infty.$
###### Proof of Corollary 2.3.
Similarly as in the proof of Lemma 5.5, we consider a family of diluted cubes.
In the current setting, we take
$G\mathrel{\mathop{\mathchar
58\relax}}=\big{\\{}8Lr_{n}z+[0,4Lr_{n}]^{d}\subset[0,1]^{d}\mathrel{\mathop{\mathchar
58\relax}}z\in{\mathbb{Z}}^{d}\big{\\}}.$ (5.30)
Then, $[0,1]^{d}$ can be covered by $2^{d}$ translates of $G$. As before, we
write $G=\\{J_{1},\dots,J_{b_{n}^{\prime}}\\}$ with
$J_{1}=\big{[}0,4Lr_{n}\big{]}^{d}$, where
$b_{n}^{\prime}\mathrel{\mathop{\mathchar 58\relax}}=(8Lr_{n})^{-d}$ denotes
the number of cubes that are contained in $G$. Since there are at most
finitely many translates of $G$, it suffices to prove that as $n\to\infty$,
$\displaystyle\rho_{k,n}^{-1}\log\mathbb{P}\bigg{(}\sup_{A\subset
E}\Big{|}\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n},\,|{\mathcal{Y}}|=k}s_{n}({\mathcal{Y}},\mathcal{P}_{n};t){\mathbbm{1}}\Big{\\{}\ell({\mathcal{Y}})\in\bigcup_{i=1}^{b_{n}^{\prime}}J_{i}\Big{\\}}\,\delta_{r_{n}^{-1}\overline{\mathcal{Y}}}(A)$
$\displaystyle\qquad\qquad\qquad-\sum_{{\mathcal{Y}}\subset\mathcal{B}_{n},\,|{\mathcal{Y}}|=k}s_{n}({\mathcal{Y}},\mathcal{B}_{n};t){\mathbbm{1}}\Big{\\{}\ell({\mathcal{Y}})\in\bigcup_{i=1}^{b_{n}^{\prime}}J_{i}\Big{\\}}\,\delta_{r_{n}^{-1}\overline{\mathcal{Y}}}(A)\Big{|}\geq\varepsilon_{0}\rho_{k,n}\bigg{)}\to-\infty$
(recall that $\ell({\mathcal{Y}})$ is the left most point of ${\mathcal{Y}}$
in the lexicographic ordering).
We say that $J_{i}$ is an _$n$ -bad cube_ if one of the following events
happens.
* •
There exists a $k$-element subset ${\mathcal{Y}}\subset\mathcal{P}_{n}$ with
$\ell({\mathcal{Y}})\in J_{i}$, $s_{n}({\mathcal{Y}},\mathcal{P}_{n};t)=1$,
such that ${\mathcal{Y}}\not\subset\mathcal{B}_{n}$ or
$s_{n}\big{(}{\mathcal{Y}},\mathcal{B}_{n};t\big{)}=0$ holds.
* •
There exists a $k$-element subset ${\mathcal{Y}}\subset\mathcal{B}_{n}$ with
$\ell({\mathcal{Y}})\in J_{i}$, $s_{n}({\mathcal{Y}},\mathcal{B}_{n};t)=1$,
such that ${\mathcal{Y}}\not\subset\mathcal{P}_{n}$ or
$s_{n}\big{(}{\mathcal{Y}},\mathcal{P}_{n};t\big{)}=0$ holds.
In this setting, the key observation is that there exists a constant $M>0$,
such that
$\displaystyle\sup_{A\subset
E}\Big{|}\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n},\,|{\mathcal{Y}}|=k}s_{n}({\mathcal{Y}},\mathcal{P}_{n};t){\mathbbm{1}}\Big{\\{}\ell({\mathcal{Y}})\in\bigcup_{i=1}^{b_{n}^{\prime}}J_{i}\Big{\\}}\,\delta_{r_{n}^{-1}\overline{\mathcal{Y}}}(A)$
$\displaystyle\qquad\qquad\qquad-\sum_{{\mathcal{Y}}\subset\mathcal{B}_{n},\,|{\mathcal{Y}}|=k}s_{n}({\mathcal{Y}},\mathcal{B}_{n};t){\mathbbm{1}}\Big{\\{}\ell({\mathcal{Y}})\in\bigcup_{i=1}^{b_{n}^{\prime}}J_{i}\Big{\\}}\,\delta_{r_{n}^{-1}\overline{\mathcal{Y}}}(A)\Big{|}$
$\displaystyle\leq M\sum_{i=1}^{b_{n}^{\prime}}{\mathbbm{1}}\\{J_{i}\text{ is
$n$-bad}\\}.$
Thus, it suffices to show that for every $\varepsilon_{0}>0$,
$\lim_{n\to\infty}\frac{1}{\rho_{k,n}}\log\mathbb{P}\Big{(}\sum_{i=1}^{b_{n}^{\prime}}{\mathbbm{1}}\\{J_{i}\text{
is $n$-bad}\\}\geq\varepsilon_{0}\rho_{k,n}\Big{)}=-\infty.$
For $0<\varepsilon\leq 1$, let $\mathcal{P}_{n}^{(\varepsilon)}$ be a
homogeneous Poisson point process on $[0,1]^{d}$ with intensity
$n\varepsilon$, independent of $\mathcal{P}_{n}$. Then,
$\mathcal{P}_{n}^{(\varepsilon,\mathsf{a})}\mathrel{\mathop{\mathchar
58\relax}}=\mathcal{P}_{n}\cup\mathcal{P}_{n}^{(\varepsilon)}$ represents an
_augmented_ version of $\mathcal{P}_{n}$ with intensity $n(1+\varepsilon)$.
Moreover, $\mathcal{P}_{n}^{(\varepsilon,\mathsf{t})}$ denotes a _thinned_
version of $\mathcal{P}_{n}$, obtained by removing points with probability
$\varepsilon$. Notice that
$\mathcal{P}_{n}^{(\varepsilon,\mathsf{a})}\stackrel{{\scriptstyle
d}}{{=}}\mathcal{P}_{n(1+\varepsilon)}$ and
$\mathcal{P}_{n}^{(\varepsilon,\mathsf{t})}\stackrel{{\scriptstyle
d}}{{=}}\mathcal{P}_{n(1-\varepsilon)}$. In this setting, we introduce the
event
$F_{n,\varepsilon}\mathrel{\mathop{\mathchar
58\relax}}=\big{\\{}\mathcal{P}_{n}^{(\varepsilon,\mathsf{t})}\subset\mathcal{B}_{n}\subset\mathcal{P}_{n}^{(\varepsilon,\mathsf{a})}\big{\\}},$
and note by the Poisson concentration bound from [31, Lemma 1.2],
$\displaystyle\limsup_{n\to\infty}\rho_{k,n}^{-1}\log\mathbb{P}\big{(}F_{n,\varepsilon}^{c}\big{)}\leq-C^{*}\lim_{n\to\infty}\rho_{k,n}^{-1}n=-\infty.$
(5.31)
The key advantage of the event $F_{n,\varepsilon}$ is that it allows to
simplify the property of being $n$-bad. Indeed, if $J_{i}$ is $n$-bad and
$F_{n,\varepsilon}$ holds, then $J_{i}$ becomes _$(n,\varepsilon)$ -special_,
in the sense that
$T\cap(\mathcal{P}_{n}^{(\varepsilon,\mathsf{a})}\setminus\mathcal{P}_{n}^{(\varepsilon,\mathsf{t})})\neq\emptyset,\
\ \text{ and }\ \ \mathcal{P}_{n}^{(\varepsilon,\mathsf{a})}(T)\geq k,$
where
$T\mathrel{\mathop{\mathchar
58\relax}}=\big{\\{}x\in{\mathbb{R}}^{d}\mathrel{\mathop{\mathchar
58\relax}}\inf_{y\in J_{i}}\|x-y\|\leq 2r_{n}L\big{\\}},\ r>0.$
Since we work with diluted cubes in (5.30), one can see that
$\big{(}{\mathbbm{1}}\\{J_{i}\text{ is
}(n,\varepsilon)\text{-special}\\}\big{)}_{i=1}^{b_{n}^{\prime}}$ are i.i.d.
Bernoulli random variables. Thus, the number of $(n,\varepsilon)$-special
cubes is a binomial random variable with $b_{n}^{\prime}$ trials and success
probability $p_{n,\varepsilon}\mathrel{\mathop{\mathchar
58\relax}}=\mathbb{P}\big{(}J_{1}\text{ is $(n,\varepsilon)$-special}\big{)}.$
Then, one can bound $p_{n,\varepsilon}$ as follows:
$\displaystyle p_{n,\varepsilon}$
$\displaystyle\leq\mathbb{P}\big{(}T\cap(\mathcal{P}_{n}^{(\varepsilon,\mathsf{a})}\setminus\mathcal{P}_{n}^{(\varepsilon,\mathsf{t})})\neq\emptyset\,\big{|}\,\mathcal{P}_{n}^{(\varepsilon,\mathsf{a})}(T)=k\big{)}\mathbb{P}\big{(}\mathcal{P}_{n}^{(\varepsilon,\mathsf{a})}(T)=k\big{)}$
$\displaystyle\qquad\qquad\qquad\qquad\qquad\qquad\qquad\qquad+\mathbb{P}\big{(}\mathcal{P}_{n}^{(\varepsilon,\mathsf{a})}(T)\geq
k+1\big{)}$ $\displaystyle\leq
2k\varepsilon\mathbb{P}\big{(}\mathcal{P}_{n(1+\varepsilon)}(T)=k\big{)}+\mathbb{P}\big{(}\mathcal{P}_{n(1+\varepsilon)}(T)\geq
k+1\big{)}$ $\displaystyle\leq
C^{*}\big{(}\varepsilon(nr_{n}^{d})^{k}+(nr_{n}^{d})^{k+1}\big{)}.$
In conclusion, if one takes sufficiently small $\varepsilon>0$, then
$b_{n}^{\prime}p_{n,\varepsilon}\leq
C^{*}b_{n}^{\prime}\big{(}\varepsilon(nr_{n}^{d})^{k}+(nr_{n}^{d})^{k+1}\big{)}\leq\varepsilon_{0}\rho_{k,n}$
for large $n$ enough. Therefore, the binomial concentration inequality [31,
Lemma 1.1] gives that
$\displaystyle\limsup_{n\to\infty}\frac{1}{\rho_{k,n}}\log\mathbb{P}\big{(}\text{Bin}(b_{n}^{\prime},p_{n,\varepsilon})\geq\varepsilon_{0}\rho_{k,n}\big{)}$
$\displaystyle\leq-\lim_{n\to\infty}\frac{\varepsilon_{0}}{2}\,\log\bigg{\\{}\frac{\varepsilon_{0}\rho_{k,n}}{C^{*}b_{n}^{\prime}(\varepsilon(nr_{n}^{d})^{k}+(nr_{n}^{d})^{k+1})}\bigg{\\}}$
$\displaystyle=-\frac{\varepsilon_{0}}{2}\log\Big{\\{}\frac{(8L)^{d}\varepsilon_{0}}{C^{*}\varepsilon}\Big{\\}}.$
The last term tends to $-\infty$ as $\varepsilon\to 0$. Hence, combining this
result with (5.31) concludes the proof of Corollary 2.3. ∎
### 5.2 Proofs of Theorem 3.1 and Corollary 3.2
First we point out that the proof of Corollary 3.2 is almost identical to that
of Corollary 2.3, so we skip it here. For the proof of Theorem 3.1, we define
$V_{k,n}\mathrel{\mathop{\mathchar
58\relax}}=\frac{1}{\rho_{k,n}}\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n},\,|{\mathcal{Y}}|=k}s_{n}({\mathcal{Y}},\mathcal{P}_{n};t)\,\delta_{H_{n}({\mathcal{Y}})}\in
M_{+}(E^{\prime}).$
The key step of our proof is to show that the assertions of Theorem 3.1 still
hold even when $(U_{k,n})_{n\geq 1}$ is replaced by $(V_{k,n})_{n\geq 1}$. To
clarify our presentation, we state this step as a separate proposition.
###### Proposition 5.7.
The sequence $(V_{k,n})_{n\geq 1}$ satisfies an LDP in the weak topology with
rate $\rho_{k,n}$ and the rate function $\bar{\Lambda}_{k}^{*}$ defined in
(3.5).
###### Proof.
The process $(V_{k,n})_{n\geq 1}$ has structure very similar to that of
$(\xi_{k,n})_{n\geq 1}$ in (2.4); thus, the proof techniques for Proposition
5.7 are parallel to those for Theorem 2.1. More precisely, as an analog of
(5.1), we define
$W_{k,n}=\frac{1}{\rho_{k,n}}\sum_{\ell=1}^{\rho_{k,n}}\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}|_{Q_{\ell}},\,|{\mathcal{Y}}|=k}s_{n}({\mathcal{Y}},\mathcal{P}_{n}|_{Q_{\ell}};t)\,\delta_{H_{n}({\mathcal{Y}})}\in
M_{+}(E^{\prime}).$
It then follows from the same argument as Proposition 5.1 that
$(W_{k,n})_{n\geq 1}$ satisfies an LDP in the weak topology with rate
$\rho_{k,n}$ and the rate function $\bar{\Lambda}_{k}^{*}$. Subsequently, by
repeating the proof of Proposition 5.6, one can also establish that for every
$\delta>0$,
$\lim_{n\to\infty}\frac{1}{\rho_{k,n}}\log\mathbb{P}\big{(}d_{\mathsf{TV}}(V_{k,n},W_{k,n})\geq\delta\big{)}=-\infty.$
This concludes the proof of Proposition 5.7. ∎
###### Proof of Theorem 3.1.
We now have to show that $(U_{k,n})_{n\geq 1}$ exhibits the same LDP as
$(V_{k,n})_{n\geq 1}$ above. We take, without loss of generality, $0<t_{1}\leq
t_{2}\leq\dots\leq t_{m}<\infty$ for time parameters of $(U_{k,n})_{n\geq 1}$,
whereas we fix the parameter of $(V_{k,n})_{n\geq 1}$ at $t=t_{1}$. In this
setup, we need to verity that for every $\delta>0$,
$\lim_{n\to\infty}\frac{1}{\rho_{k,n}}\log\mathbb{P}\big{(}{d_{\mathsf{TV}}(U_{k,n},V_{k,n})}\geq\delta\big{)}=-\infty.$
It is straightforward to see that
${d_{\mathsf{TV}}(U_{k,n},V_{k,n})}\leq\frac{1}{\rho_{k,n}}\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n},\,|{\mathcal{Y}}|=k}s_{n}({\mathcal{Y}},\mathcal{P}_{n};t_{1})\big{(}1-c_{n}({\mathcal{Y}},\mathcal{P}_{n};t_{m})\big{)}.$
By virtue of the partition of $[0,1]^{d}$ into multiple sub-cubes
$Q_{1},\dots,Q_{\rho_{k,n}}$ as in the proof of Proposition 5.1, what need to
be shown are
$\lim_{n\to\infty}\frac{1}{\rho_{k,n}}\log\mathbb{P}\Big{(}\sum_{\ell=1}^{\rho_{k,n}}\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}|_{Q_{\ell}}}s_{n}({\mathcal{Y}},\mathcal{P}_{n}|_{Q_{\ell}};t_{1})\big{(}1-c_{n}({\mathcal{Y}},\mathcal{P}_{n}|_{Q_{\ell}};t_{m})\big{)}\geq\delta\rho_{k,n}\Big{)}=-\infty,$
(5.32)
and
$\displaystyle\lim_{n\to\infty}\frac{1}{\rho_{k,n}}\log\,$
$\displaystyle\mathbb{P}\Big{(}\,\Big{|}\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}}s_{n}({\mathcal{Y}},\mathcal{P}_{n};t_{1})\big{(}1-c_{n}({\mathcal{Y}},\mathcal{P}_{n};t_{m})\big{)}$
$\displaystyle\qquad-\sum_{\ell=1}^{\rho_{k,n}}\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}|_{Q_{\ell}}}s_{n}({\mathcal{Y}},\mathcal{P}_{n}|_{Q_{\ell}};t_{1})\big{(}1-c_{n}({\mathcal{Y}},\mathcal{P}_{n}|_{Q_{\ell}};t_{m})\big{)}\Big{|}\geq\delta\rho_{k,n}\Big{)}=-\infty,$
for every $\delta>0$. Of the last two conditions, the latter can be
established by the same argument as Proposition 5.6. By Markov’s inequality,
(5.32) will follow if we can show that for every $a>0$,
$\mathbb{E}\Big{[}e^{a\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}|_{Q_{1}}}s_{n}({\mathcal{Y}},\mathcal{P}_{n}|_{Q_{1}};t_{1})(1-c_{n}({\mathcal{Y}},\mathcal{P}_{n}|_{Q_{1}};t_{m}))}\Big{]}\to
1,\ \ n\to\infty.$ (5.33)
Repeating the same calculations as in (5.11), while using an obvious bound
$s_{n}({\mathcal{Y}},\mathcal{P}_{n}|_{Q_{1}};t_{1})\leq{\mathbbm{1}}\big{\\{}\mathsf{diam}({\mathcal{Y}})\leq
r_{n}L\big{\\}}$, as well as the Mecke formula for Poisson point processes,
$\displaystyle\mathbb{E}\Big{[}\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}|_{Q_{1}}}s_{n}({\mathcal{Y}},\mathcal{P}_{n}|_{Q_{1}};t_{1})(1-c_{n}({\mathcal{Y}},\mathcal{P}_{n}|_{Q_{1}};t_{m}))\Big{]}$
$\displaystyle\leq\frac{1}{k!}\,\int_{({\mathbb{R}}^{d})^{k-1}}{\mathbbm{1}}\big{\\{}\mathsf{diam}({\bf
0}_{d},{\bf y})\leq
L\big{\\}}\Big{(}1-e^{-nr_{n}^{d}\text{vol}(\mathcal{B}(\\{{\bf 0}_{d},{\bf
y}\\};t_{m}))}\Big{)}\operatorname{d\\!}{\bf y}\to 0,\ \ \text{as
}n\to\infty.$
Since we have already shown the uniform integrability by (5.16), we now obtain
(5.33), as desired.
Finally, by the same proof as Proposition 2.2, we can obtain (3.6). ∎
### 5.3 Proofs of Theorem 3.3, Corollary 3.4, and Proposition 3.5
In this section, we first prove Proposition 3.5, because the proof of Theorem
3.3 makes use of Proposition 3.5.
###### Proof of Proposition 3.5.
For every ${\bf x}=(x_{1},\dots,x_{m})\in E^{\prime}$, one can choose ${\bf
a}=(a_{1},\dots,a_{m})\in{\mathbb{R}}^{m}$ so that
$u_{i}({\bf x},{\bf a})\mathrel{\mathop{\mathchar
58\relax}}=\frac{1}{k!}\int_{({\mathbb{R}}^{d})^{k-1}}h^{(i)}({\bf 0}_{d},{\bf
y})e^{\sum_{r=1}^{m}a_{r}h^{(r)}({\bf 0}_{d},{\bf y})}\operatorname{d\\!}{\bf
y}-x_{i}=0,\ \ i=1,\dots,m,$
with ${\bf y}\in({\mathbb{R}}^{d})^{k-1}$. By condition (H4), $u_{i}$ is
continuously differentiable with respect to $({\bf x},{\bf a})$. Furthermore,
(H5) ensures that the matrix $\big{(}\partial u_{i}/\partial
a_{j}\big{)}_{i,j=1}^{m}$ is positive definite. Thus, by the implicit function
theorem, there exist continuously differentiable functions
$\alpha_{r}\mathrel{\mathop{\mathchar 58\relax}}E^{\prime}\to{\mathbb{R}}$,
$r=1,\dots,m$, such that
$I_{k}({\bf x})=\sum_{i=1}^{m}\alpha_{i}({\bf
x})x_{i}-\frac{1}{k!}\,\int_{({\mathbb{R}}^{d})^{k-1}}\Big{(}e^{\sum_{i=1}^{m}\alpha_{i}({\bf
x})h^{(i)}({\bf 0}_{d},{\bf y})}-1\Big{)}\operatorname{d\\!}{\bf y},$ (5.34)
and
$x_{i}=\frac{1}{k!}\int_{({\mathbb{R}}^{d})^{k-1}}h^{(i)}({\bf 0}_{d},{\bf
y})e^{\sum_{r=1}^{m}\alpha_{r}({\bf x})h^{(r)}({\bf 0}_{d},{\bf
y})}\operatorname{d\\!}{\bf y},\ \ i=1,\dots,m.$
In particular, (5.34) implies that $I_{k}$ is continuously differentiable on
$E^{\prime}$.
For Part $(ii)$, we show that the Hessian of $I_{k}$, denoted
$\mathcal{H}(I_{k})$, is positive definite. First it is elementary to check
that $\partial I_{k}({\bf x})/\partial x_{j}=\alpha_{j}({\bf x})$ for
$j=1,\dots,m$. Using this, it is not hard to calculate that
$\mathcal{H}(I_{k})=A^{-1}$ where $A=(A_{ij})$ is given by
$A_{ij}=\frac{1}{k!}\,\int_{({\mathbb{R}}^{d})^{k-1}}h^{(i)}({\bf 0}_{d},{\bf
y})h^{(j)}({\bf 0}_{d},{\bf y})e^{{\sum_{r=1}^{m}\alpha_{r}({\bf
x})h^{(r)}({\bf 0}_{d},{\bf y})}}\operatorname{d\\!}{\bf y}.$
We conclude from condition (H5) that $A$ is positive definite; thus, so is
$A^{-1}$ as required. Finally, if we set $0=\partial I_{k}({\bf x})/\partial
x_{j}$ for every $j=1,\dots,m$, then $\alpha_{j}({\bf x})=0$, and so,
$x_{j}=\mu_{j}$ and $I_{k}({\bf x})=0$. As $I_{k}$ is strictly convex,
$(\mu_{1},\dots,\mu_{m})$ is a unique minimizer of $I_{k}$. ∎
The proof of Theorem 3.3 is based on an extension of the contraction
principle, provided in [14, Theorem 4.2.23]. We begin with defining a map
$F\mathrel{\mathop{\mathchar 58\relax}}M_{+}(E^{\prime})\to[0,\infty)^{m}$ by
$F(\rho)=\big{(}\int_{E^{\prime}}x_{i}\rho(\operatorname{d\\!}{\bf
x})\big{)}_{i=1}^{m}$. Then, $F(U_{k,n})=T_{k,n}/\rho_{k,n}$, where $U_{k,n}$
and $T_{k,n}$ are defined in (3.4) and (3.8), respectively. Since $F$ is not
continuous in the weak topology, we need to introduce a family of continuous
maps: $F_{K}(\rho)\mathrel{\mathop{\mathchar
58\relax}}=\big{(}\int_{E^{\prime}}s_{i}^{(K)}({\bf
x})\rho(\operatorname{d\\!}{\bf x})\big{)}_{i=1}^{m}$, where
$s_{i}^{(K)}(x_{1},\dots,x_{m})=\begin{cases}x_{i}&\text{ if }0\leq x_{i}\leq
K,\\\ -K^{2}(x_{i}-K)+K&\text{ if }K\leq x_{i}\leq K+K^{-1},\\\ 0&\text{ if
}x_{i}\geq K+K^{-1}.\end{cases}$
Clearly, $s_{i}^{(K)}$ is continuous and bounded on $E^{\prime}$, and
consequently, $F_{K}$ becomes continuous in the weak topology.
In order to apply [14, Theorem 4.2.23], we need to demonstrate the following
auxiliary results.
###### Lemma 5.8.
$(i)$ Let $\delta>0$. Then,
$\limsup_{K\to\infty}\limsup_{n\to\infty}\frac{1}{\rho_{k,n}}\log\mathbb{P}\Big{(}\big{\|}F(U_{k,n})-F_{K}(U_{k,n})\big{\|}>\delta\Big{)}=-\infty,$
$(ii)$ Let $a>0$. Then,
$\limsup_{K\to\infty}\sup_{\rho\in
M_{+}(E^{\prime}),\,\bar{\Lambda}_{k}^{*}(\rho)\leq
a}\big{\|}F(\rho)-F_{K}(\rho)\big{\|}=0.$
Assuming the assertions of Lemma 5.8 temporarily, we first explain how to
conclude the proof of Theorem 3.3.
###### Proof of Theorem 3.3.
Note that one can conclude from Theorem 3.1, [14, Theorem 4.2.23], and Lemma
5.8 that $(T_{k,n}/\rho_{k,n})_{n\geq 1}$ satisfies an LDP with rate
$\rho_{k,n}$ and the rate function
$\inf_{\nu\in M_{+}(E^{\prime}),\,F(\nu)={\bf
x}}H_{k}^{\prime}(\nu|\tau_{k}^{\prime}),\ \ \ {\bf x}\in{\mathbb{R}}^{m},$
where $H_{k}^{\prime}$ is a relative entropy given at (3.7). It thus remains
to show that
$\inf_{\nu\in M_{+}(E^{\prime}),\,F(\nu)={\bf
x}}H_{k}^{\prime}(\nu|\tau_{k}^{\prime})=I_{k}({\bf x}),\ \ \ {\bf
x}\in{\mathbb{R}}^{m}.$ (5.35)
If $x_{i}<0$ for some $i\in\\{1,\dots,m\\}$, both sides above are equal to
infinity. So from onward, we consider ${\bf x}\in E^{\prime}$ and prove first
that $\inf_{\nu\in M_{+}(E^{\prime}),\,F(\nu)={\bf
x}}H_{k}^{\prime}(\nu|\tau_{k}^{\prime})\geq I_{k}({\bf x})$. Fix $\nu\in
M_{+}(E^{\prime})$ such that $F(\nu)={\bf x}$. For every ${\bf
a}\in{\mathbb{R}}^{m}$, we set $f({\bf z})=\langle{\bf a},{\bf z}\rangle$ for
${\bf z}\in E^{\prime}$. Approximating $f$ via a sequence of continuous and
bounded functions, we have
$\displaystyle H_{k}^{\prime}(\nu|\tau_{k}^{\prime})$
$\displaystyle=\bar{\Lambda}_{k}^{*}(\nu)\geq\langle\nu,f\rangle-\frac{1}{k!}\int_{({\mathbb{R}}^{d})^{k-1}}\big{(}e^{f(H({\bf
0}_{d},{\bf y}))}-1\big{)}\,{\mathbbm{1}}\big{\\{}H({\bf 0}_{d},{\bf
y})\neq{\bf 0}_{m}\big{\\}}\operatorname{d\\!}{\bf y}$
$\displaystyle=\langle{\bf a},{\bf
x}\rangle-\frac{1}{k!}\int_{({\mathbb{R}}^{d})^{k-1}}\big{(}e^{\langle{\bf
a},H({\bf 0}_{d},{\bf y})\rangle}-1\big{)}\operatorname{d\\!}{\bf y};$
thus, $H_{k}^{\prime}(\nu|\tau_{k}^{\prime})\geq I_{k}({\bf x})$ holds. Next,
for every ${\bf x}\in E^{\prime}$, there exists ${\bf
a}_{0}\in{\mathbb{R}}^{m}$ such that
$I_{k}({\bf x}){=\langle{\bf a}_{0},{\bf
x}\rangle-\frac{1}{k!}\int_{({\mathbb{R}}^{d})^{k-1}}\big{(}e^{\langle{\bf
a}_{0},H({\bf 0}_{d},{\bf y})\rangle}-1\big{)}\operatorname{d\\!}{\bf
y}}=\langle{\bf a}_{0},{\bf x}\rangle-\int_{E^{\prime}}\big{(}e^{\langle{\bf
a}_{0},{\bf z}\rangle}-1\big{)}\tau_{k}^{\prime}(\operatorname{d\\!}{\bf z}),$
where $x_{i}=\int_{E^{\prime}}z_{i}e^{\langle{\bf a}_{0},{\bf
z}\rangle}\tau_{k}^{\prime}(\operatorname{d\\!}{\bf z})$ for $i=1,\dots,m$.
Define $\nu(A)\mathrel{\mathop{\mathchar 58\relax}}=\int_{A}e^{\langle{\bf
a}_{0},{\bf z}\rangle}\tau_{k}^{\prime}(\operatorname{d\\!}{\bf z})$,
$A\subset E^{\prime}$. It then follows that
$\int_{E^{\prime}}z_{i}\nu(\operatorname{d\\!}{\bf z})=x_{i}$ for
$i=1,\dots,m$; equivalently, $F(\nu)={\bf x}$. Moreover,
$H_{k}^{\prime}(\nu|\tau_{k}^{\prime})=\int_{E^{\prime}}\log\frac{\operatorname{d\\!}\nu}{\operatorname{d\\!}\tau_{k}^{\prime}}({\bf
z})\nu(\operatorname{d\\!}{\bf
z})-\nu(E^{\prime})+\tau_{k}^{\prime}(E^{\prime})=\langle{\bf a}_{0},{\bf
x}\rangle-\int_{E^{\prime}}\big{(}e^{\langle{\bf a}_{0},{\bf
z}\rangle}-1\big{)}\tau_{k}^{\prime}(\operatorname{d\\!}{\bf z})=I_{k}({\bf
x}).$
This implies $\inf_{\nu\in M_{+}(E^{\prime}),\,F(\nu)={\bf
x}}H_{k}^{\prime}(\nu|\tau_{k}^{\prime})\leq I_{k}({\bf x})$, and thus, the
proof of (5.35) has been completed.
Finally, Proposition 3.5 $(i)$ ensures that $I_{k}$ is continuous; therefore,
the LDP for $(T_{k,n}/\rho_{k,n})$ implies the convergence in (3.9).
∎
Now, we present the proof of Lemma 5.8.
###### Proof of Lemma 5.8 $(i)$.
Note first that the result follows from
$\limsup_{K\to\infty}\limsup_{n\to\infty}\frac{1}{\rho_{k,n}}\log\mathbb{P}\Big{(}\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}}g_{n}^{(i)}({\mathcal{Y}},\mathcal{P}_{n};t_{i})\,{\mathbbm{1}}\big{\\{}h_{n}^{(i)}({\mathcal{Y}})>K\big{\\}}\geq\delta\rho_{k,n}\Big{)}=-\infty,$
(5.36)
for each $i\in\\{1,\dots,m\\}$. In the sequel, we prefer to use the notation
$G_{n}$ and $H_{n}$, instead of $g_{n}^{(i)}$ and $h_{n}^{(i)}$, with the
assumption $m=1$, so that the range of $G_{n}$ and $H_{n}$ is $[0,\infty)$.
Additionally, we drop the subscript $i$ from $t_{i}$. Then, (5.36) can be
rephrased as
$\limsup_{K\to\infty}\limsup_{n\to\infty}\frac{1}{\rho_{k,n}}\log\mathbb{P}\Big{(}\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}}G_{n}({\mathcal{Y}},\mathcal{P}_{n};t)\,{\mathbbm{1}}\big{\\{}H_{n}({\mathcal{Y}})>K\big{\\}}\geq\delta\rho_{k,n}\Big{)}=-\infty.$
(5.37)
Clearly, (5.37) is implied by
$\limsup_{K\to\infty}\limsup_{n\to\infty}\frac{1}{\rho_{k,n}}\log\mathbb{P}\Big{(}\sum_{\ell=1}^{\rho_{k,n}}\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}|_{Q_{\ell}}}G_{n}({\mathcal{Y}},\mathcal{P}_{n}|_{Q_{\ell}};t)\,{\mathbbm{1}}\big{\\{}H_{n}({\mathcal{Y}})>K\big{\\}}\geq\delta\rho_{k,n}\Big{)}=-\infty,$
(5.38)
and
$\displaystyle\limsup_{K\to\infty}\limsup_{n\to\infty}\frac{1}{\rho_{k,n}}\log\mathbb{P}\bigg{(}\,\Big{|}\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}}G_{n}({\mathcal{Y}},\mathcal{P}_{n};t)\,{\mathbbm{1}}\big{\\{}H_{n}({\mathcal{Y}})>K\big{\\}}$
(5.39)
$\displaystyle\qquad\qquad\qquad\qquad\qquad-\sum_{\ell=1}^{\rho_{k,n}}\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}|_{Q_{\ell}}}G_{n}({\mathcal{Y}},\mathcal{P}_{n}|_{Q_{\ell}};t)\,{\mathbbm{1}}\big{\\{}H_{n}({\mathcal{Y}})>K\big{\\}}\Big{|}\geq\delta\rho_{k,n}\bigg{)}=-\infty,$
for every $\delta>0$. As for (5.38), Markov’s inequality yields that
$\displaystyle\frac{1}{\rho_{k,n}}\log\mathbb{P}\Big{(}\sum_{\ell=1}^{\rho_{k,n}}\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}|_{Q_{\ell}}}G_{n}({\mathcal{Y}},\mathcal{P}_{n}|_{Q_{\ell}};t)\,{\mathbbm{1}}\big{\\{}H_{n}({\mathcal{Y}})>K\big{\\}}\geq\delta\rho_{k,n}\Big{)}$
$\displaystyle\leq-a\delta+\log\mathbb{E}\Big{[}e^{a\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}|_{Q_{1}}}G_{n}({\mathcal{Y}},\mathcal{P}_{n}|_{Q_{1}};t)\,{\mathbbm{1}}\\{H_{n}({\mathcal{Y}})>K\\}}\Big{]}.$
It is thus sufficient to demonstrate that for every $a>0$,
$\limsup_{K\to\infty}\limsup_{n\to\infty}\mathbb{E}\Big{[}e^{a\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}|_{Q_{1}}}G_{n}({\mathcal{Y}},\mathcal{P}_{n}|_{Q_{1}};t){\mathbbm{1}}\\{H_{n}({\mathcal{Y}})>K\\}}\Big{]}=1.$
(5.40)
Although the proof techniques for (5.40) are mostly the same as those for
(5.16), we still need to update it because $G_{n}$ in (5.40) is not
necessarily an indicator function. Using the same logic as in the proof of
(5.16), our goal is to prove that for every $a>0$,
$\limsup_{K\to\infty}\limsup_{n\to\infty}\bigg{(}\mathbb{E}\Big{[}e^{a\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}|_{Q_{1}}}G_{n}({\mathcal{Y}},\mathcal{P}_{n}|_{Q_{1}};t){\mathbbm{1}}\\{H_{n}({\mathcal{Y}})>K,\,\ell({\mathcal{Y}})\in
J_{1}\\}}\Big{]}\bigg{)}^{b_{n}}\leq 1,$
where $J_{1}=\big{[}0,tr_{n}/\sqrt{d}\big{]}^{d}$ and
$b_{n}=\rho_{k,n}^{-1}/(4Lr_{n})^{d}$. Observe, once again, that there exists
at most a single $k$-point set ${\mathcal{Y}}\subset\mathcal{P}_{n}|_{Q_{1}}$
satisfying $c_{n}({\mathcal{Y}},\mathcal{P}_{n}|_{Q_{1}};t)=1$ and
$\ell({\mathcal{Y}})\in J_{1}$. Therefore, as $n\to\infty$,
$\displaystyle\mathbb{E}\Big{[}e^{a\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}|_{Q_{1}}}G_{n}({\mathcal{Y}},\mathcal{P}_{n}|_{Q_{1}};t){\mathbbm{1}}\\{H_{n}({\mathcal{Y}})>K,\,\ell({\mathcal{Y}})\in
J_{1}\\}}\Big{]}$ (5.41) $\displaystyle\leq
1+\mathbb{E}\Big{[}\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}}e^{aH_{n}({\mathcal{Y}})}{\mathbbm{1}}\big{\\{}H_{n}({\mathcal{Y}})>K,\,\ell({\mathcal{Y}})\in
J_{1}\big{\\}}\Big{]}$
$\displaystyle=1+\frac{n^{k}}{k!}\int_{([0,1]^{d})^{k}}e^{aH_{n}(x_{1},\dots,x_{k})}{\mathbbm{1}}\big{\\{}H_{n}(x_{1},\dots,x_{k})>K,\,\ell(x_{1},\dots,x_{k})\in
J_{1}\big{\\}}\operatorname{d\\!}{\bf x}$
$\displaystyle=1+\frac{\rho_{k,n}}{k!}\int_{[0,1]^{d}}\int_{({\mathbb{R}}^{d})^{k-1}}e^{aH({\bf
0}_{d},{\bf y})}{\mathbbm{1}}\big{\\{}H({\bf 0}_{d},{\bf
y})>K,\,\ell(x,x+r_{n}y_{1},\dots,x+r_{n}y_{k-1})\in J_{1}\big{\\}}$
$\displaystyle\qquad\qquad\qquad\qquad\qquad\qquad\qquad\qquad\qquad\qquad\times\prod_{i=1}^{k-1}{\mathbbm{1}}\big{\\{}x+r_{n}y_{i}\in[0,1]^{d}\big{\\}}\operatorname{d\\!}{\bf
y}\operatorname{d\\!}x$ $\displaystyle\leq
1+C^{*}\rho_{k,n}r_{n}^{d}\int_{({\mathbb{R}}^{d})^{k-1}}e^{aH({\bf
0}_{d},{\bf y})}{\mathbbm{1}}\big{\\{}H({\bf 0}_{d},{\bf
y})>K\big{\\}}\operatorname{d\\!}{\bf y}.$
Here, the last integral is finite due to property (H4). Now, it follows from
(5.41) that
$\displaystyle\limsup_{K\to\infty}\limsup_{n\to\infty}\bigg{(}\mathbb{E}\Big{[}e^{a\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}|_{Q_{1}}}G_{n}({\mathcal{Y}},\mathcal{P}_{n}|_{Q_{1}};t){\mathbbm{1}}\\{H_{n}({\mathcal{Y}})>K,\,\ell({\mathcal{Y}})\in
J_{1}\\}}\Big{]}\bigg{)}^{b_{n}}$
$\displaystyle\leq\limsup_{K\to\infty}\exp\Big{\\{}\frac{C^{*}}{(4L)^{d}}\int_{({\mathbb{R}}^{d})^{k-1}}e^{aH({\bf
0}_{d},{\bf
y})}{\mathbbm{1}}\big{\\{}H_{n}({\mathcal{Y}})>K\big{\\}}\operatorname{d\\!}{\bf
y}\Big{\\}}=1$
Next, turning to condition (5.39), we deduce an inequality analogous to
(5.21):
$\displaystyle\Big{|}\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}}G_{n}({\mathcal{Y}},\mathcal{P}_{n};t)\,{\mathbbm{1}}\big{\\{}H_{n}({\mathcal{Y}})>K\big{\\}}-\sum_{\ell=1}^{\rho_{k,n}}\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}|_{Q_{\ell}}}G_{n}({\mathcal{Y}},\mathcal{P}_{n}|_{Q_{\ell}};t)\,{\mathbbm{1}}\big{\\{}H_{n}({\mathcal{Y}})>K\big{\\}}\Big{|}$
$\displaystyle\leq\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}}G_{n}\big{(}{\mathcal{Y}},\mathcal{P}_{n}|_{\bigcup_{\ell=1}^{\rho_{k,n}}Q_{\ell}^{\partial}((L+t)r_{n})};t\big{)}{\mathbbm{1}}\Big{\\{}{\mathcal{Y}}\subset\bigcup_{\ell=1}^{\rho_{k,n}}Q_{\ell}^{\partial}\big{(}(L+t)r_{n}\big{)}\Big{\\}}$
$\displaystyle\qquad\qquad\qquad\qquad\qquad\qquad+\sum_{\ell=1}^{\rho_{k,n}}\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}|_{Q^{\partial}_{\ell}((L+t)r_{n})}}G_{n}\big{(}{\mathcal{Y}},\mathcal{P}_{n}|_{Q^{\partial}_{\ell}((L+t)r_{n})};t\big{)}.$
Proceeding the same argument as those after (5.21), while using property (H4),
we can get (5.39). Now, the proof of Lemma 5.8 $(i)$ is completed. ∎
###### Proof of Lemma 5.8 $(ii)$.
Suppose, for contradiction, that we can choose $a>0$ and $\delta>0$ such that
$\limsup_{K\to\infty}\sup_{\rho\in
M_{+}(E^{\prime}),\,\bar{\Lambda}_{k}^{*}(\rho)\leq
a}\big{\|}F(\rho)-F_{K}(\rho)\big{\|}>\delta.$ (5.42)
For $K^{\prime}\geq K+K^{-1}$, define
$F_{K,K^{\prime}}(\rho)\mathrel{\mathop{\mathchar
58\relax}}=\Big{(}\int_{E^{\prime}}u_{i}^{(K,K^{\prime})}({\bf
x})\rho(\operatorname{d\\!}{\bf x})\Big{)}_{i=1}^{m},\ \ \ \rho\in
M_{+}(E^{\prime}),$
where
$u_{i}^{(K,K^{\prime})}(x_{1},\dots,x_{m})=\begin{cases}0&\text{ if
}x_{i}\in[0,K]\cup[K^{\prime}+(K^{\prime})^{-1},\infty),\\\
(K^{2}+1)(x_{i}-K)&\text{ if }K\leq x_{i}\leq K+K^{-1},\\\ x_{i}&\text{ if
}K+K^{-1}\leq x_{i}\leq K^{\prime},\\\
-(K^{\prime})^{2}(x_{i}-K^{\prime})+K^{\prime}&\text{ if }K^{\prime}\leq
x_{i}\leq K^{\prime}+(K^{\prime})^{-1}.\end{cases}$
Then, $F_{K,K^{\prime}}$ is continuous in the weak topology, and further,
$\big{\|}F_{K,K^{\prime}}(\rho)\big{\|}\nearrow\big{\|}F(\rho)-F_{K}(\rho)\big{\|},\
\ \text{as }K^{\prime}\uparrow\infty.$ (5.43)
By Lemma 5.8 $(i)$ and (5.43), we can choose $K_{1}>0$ so that
$\limsup_{n\to\infty}\frac{1}{\rho_{k,n}}\log\mathbb{P}\Big{(}\big{\|}F_{K,K^{\prime}}(U_{k,n})\big{\|}>\delta\Big{)}<-a$
(5.44)
for any $K^{\prime}$ and $K$, with $K^{\prime}\geq K+K^{-1}>K\geq K_{1}$. By
virtue of the continuity of $F_{K,K^{\prime}}$ together with (5.44), one can
apply the contraction principle to the LDP for $(U_{k,n})_{n\geq 1}$ (see
Theorem 3.1), to obtain that
$\inf_{\rho\in
M_{+}(E^{\prime}),\,\|F_{K,K^{\prime}}(\rho)\|>\delta}\bar{\Lambda}_{k}^{*}(\rho)>a,$
for all $K^{\prime}\geq K+K^{-1}>K\geq K_{1}$.
Next, let us turn to (5.42) and fix $K\geq K_{1}$, so that $\sup_{\rho\in
M_{+}(E^{\prime}),\,\bar{\Lambda}_{k}^{*}(\rho)\leq
a}\big{\|}F(\rho)-F_{K}(\rho)\big{\|}>\delta$. Then, we can choose $\nu\in
M_{+}(E^{\prime})$ with $\bar{\Lambda}_{k}^{*}(\nu)\leq a$ and
$\big{\|}F(\nu)-F_{K}(\nu)\big{\|}>\delta$. Furthermore, (5.43) implies that
there exists $K^{\prime}\geq K+K^{-1}$ with
$\big{\|}F_{K,K^{\prime}}(\nu)\big{\|}>\delta$. Since this is contradiction,
we have established Lemma 5.8 $(ii)$, as desired. ∎
Finally, we prove Corollary 3.4. The proof is very similar to that of
Corollary 2.3. Hence, we focus only on some of its key differences. The goal
is to show that for every $\varepsilon_{0}>0$,
$\lim_{n\to\infty}\rho_{k,n}^{-1}\log\mathbb{P}\big{(}\|T_{k,n}-T_{k,n}^{\mathsf{B}}\|\geq\varepsilon_{0}\rho_{k,n}\big{)}=-\infty.$
###### Proof of Corollary 3.4.
As in the proof of Lemma 5.8 $(i)$, we may put the assumption $m=1$ and use
the notation $G_{n}$ and $H_{n}$ (instead of $g_{n}^{(1)}$ and $h_{n}^{(1)}$).
For $K>0$, we decompose $T_{k,n}^{\mathsf{B}}=T_{\leq
K,k,n}^{\mathsf{B}}+T_{>K,k,n}^{\mathsf{B}}$, where
$T_{\leq K,k,n}^{\mathsf{B}}\mathrel{\mathop{\mathchar
58\relax}}=\sum_{{\mathcal{Y}}\subset\mathcal{B}_{n},\,|{\mathcal{Y}}|=k}G_{n}({\mathcal{Y}},\mathcal{B}_{n};t){\mathbbm{1}}\\{H_{n}({\mathcal{Y}})\leq
K\\},$
and
$T_{>K,k,n}^{\mathsf{B}}\mathrel{\mathop{\mathchar
58\relax}}=\sum_{{\mathcal{Y}}\subset\mathcal{B}_{n},\,|{\mathcal{Y}}|=k}G_{n}({\mathcal{Y}},\mathcal{B}_{n};t){\mathbbm{1}}\\{H_{n}({\mathcal{Y}})>K\\}.$
Similarly, we write $T_{k,n}=T_{\leq K,k,n}+T_{>K,k,n}$. Since all the
summands in $T_{\leq K,k,n}$ and $T_{\leq K,k,n}^{\mathsf{B}}$ are bounded by
$K$, a simple repetition of the proof of Corollary 2.3 can yield that
$\frac{1}{\rho_{k,n}}\log\mathbb{P}\big{(}\,|\,T_{\leq K,k,n}-T_{\leq
K,k,n}^{\mathsf{B}}\,|\geq\varepsilon_{0}\rho_{k,n}\big{)}\to-\infty,\ \ \
n\to\infty,$
for every $K>0$. Additionally, the proof of Lemma 5.8 $(i)$ has already shown
that
$\limsup_{K\to\infty}\limsup_{n\to\infty}\frac{1}{\rho_{k,n}}\log\mathbb{P}(T_{>K,k,n}\geq\varepsilon_{0}\rho_{k,n})=-\infty;$
so it remains to verify that
$\limsup_{K\to\infty}\limsup_{n\to\infty}\frac{1}{\rho_{k,n}}\log\mathbb{P}\big{(}T_{>K,k,n}^{\mathsf{B}}\geq\varepsilon_{0}\rho_{k,n}\big{)}=-\infty.$
(5.45)
Using the same diluted cubes from the proof of Corollary 2.3, (5.45) will
follow, provided that
$\displaystyle\limsup_{K\to\infty}\limsup_{n\to\infty}\frac{1}{\rho_{k,n}}\log\mathbb{P}\Big{(}\sum_{i=1}^{b_{n}^{\prime}}\sum_{{\mathcal{Y}}\subset\mathcal{B}_{n},\,|{\mathcal{Y}}|=k}G_{n}({\mathcal{Y}},\mathcal{B}_{n};t)$
$\displaystyle\qquad\qquad\qquad\qquad\qquad\qquad\qquad\times{\mathbbm{1}}\big{\\{}H_{n}({\mathcal{Y}})>K,\,\ell({\mathcal{Y}})\in
J_{i}\big{\\}}\geq\varepsilon_{0}\rho_{k,n}\Big{)}=-\infty.$
Observe again that there exists a constant $M>0$ such that
$\displaystyle\begin{split}&\sum_{{\mathcal{Y}}\subset\mathcal{B}_{n},\,|{\mathcal{Y}}|=k}G_{n}({\mathcal{Y}},\mathcal{B}_{n};t){\mathbbm{1}}\\{H_{n}({\mathcal{Y}})>K,\,\ell({\mathcal{Y}})\in
J_{i}\\}\\\ &\quad\leq
M\max_{{\mathcal{Y}}\subset\mathcal{B}_{n},\,|{\mathcal{Y}}|=k}\Big{(}H_{n}({\mathcal{Y}}){\mathbbm{1}}\\{H_{n}({\mathcal{Y}})>K,\,\ell({\mathcal{Y}})\in
J_{i}\\}\Big{)}.\end{split}$ (5.46)
Moreover, as in the proof of Corollary 2.3, we may work under the event
$\big{\\{}\mathcal{B}_{n}\subset\mathcal{P}_{n}^{(1,\mathsf{a})}\big{\\}}$,
where
$\mathcal{P}_{n}^{(1,\mathsf{a})}=\mathcal{P}_{n}\cup\mathcal{P}_{n}^{(1)}\stackrel{{\scriptstyle
d}}{{=}}\mathcal{P}_{2n}$. By Markov’s inequality and (5.46), we have, for
every $a>0$,
$\displaystyle\mathbb{P}\Big{(}\Big{\\{}\sum_{i=1}^{b_{n}^{\prime}}\sum_{{\mathcal{Y}}\subset\mathcal{B}_{n},\,|{\mathcal{Y}}|=k}G_{n}({\mathcal{Y}},\mathcal{B}_{n};t){\mathbbm{1}}\\{H_{n}({\mathcal{Y}})>K,\,\ell({\mathcal{Y}})\in
J_{i}\\}\geq\varepsilon_{0}\rho_{k,n}\Big{\\}}\cap\big{\\{}\mathcal{B}_{n}\subset\mathcal{P}_{n}^{(1,\mathsf{a})}\big{\\}}\Big{)}$
$\displaystyle\quad\leq\mathbb{P}\Big{(}\sum_{i=1}^{b_{n}^{\prime}}M\max_{{\mathcal{Y}}\subset\mathcal{P}_{n}^{(1,\mathsf{a})},\,|{\mathcal{Y}}|=k}\Big{(}H_{n}({\mathcal{Y}}){\mathbbm{1}}\\{H_{n}({\mathcal{Y}})>K,\,\ell({\mathcal{Y}})\in
J_{i}\\}\Big{)}\geq\varepsilon_{0}\rho_{k,n}\Big{)}$ $\displaystyle\quad\leq
e^{-a\varepsilon_{0}\rho_{k,n}}\big{(}\mathbb{E}\big{[}e^{aZ_{n}}\big{]}\big{)}^{b_{n}^{\prime}},$
where
$Z_{n}\mathrel{\mathop{\mathchar
58\relax}}=M\max_{{\mathcal{Y}}\subset\mathcal{P}_{2n},\,|{\mathcal{Y}}|=k}\Big{(}H_{n}({\mathcal{Y}}){\mathbbm{1}}\\{H_{n}({\mathcal{Y}})>K,\,\ell({\mathcal{Y}})\in
J_{1}\\}\Big{)}.$
Now, proceeding as in (5.41),
$\displaystyle\mathbb{E}\big{[}e^{aZ_{n}}\big{]}$ $\displaystyle\leq
1+\mathbb{E}\Big{[}\sum_{{\mathcal{Y}}\subset\mathcal{P}_{2n},\,|{\mathcal{Y}}|=k}e^{aMH_{n}({\mathcal{Y}})}{\mathbbm{1}}\big{\\{}H_{n}({\mathcal{Y}})>K,\,\ell({\mathcal{Y}})\in
J_{1}\big{\\}}\Big{]}$ $\displaystyle\leq
1+C^{*}\rho_{k,n}r_{n}^{d}\int_{({\mathbb{R}}^{d})^{k-1}}e^{aMH({\bf
0}_{d},{\bf y})}{\mathbbm{1}}\big{\\{}H({\bf 0}_{d},{\bf
y})>K\big{\\}}\operatorname{d\\!}{\bf y}$
$\displaystyle\leq\exp\Big{\\{}C^{*}\rho_{k,n}r_{n}^{d}\int_{({\mathbb{R}}^{d})^{k-1}}e^{aMH({\bf
0}_{d},{\bf y})}{\mathbbm{1}}\big{\\{}H({\bf 0}_{d},{\bf
y})>K\big{\\}}\operatorname{d\\!}{\bf y}\Big{\\}}.$
This, together with property (H4), implies that
$\displaystyle\limsup_{K\to\infty}\limsup_{n\to\infty}\frac{1}{\rho_{k,n}}\log\Big{\\{}e^{-a\varepsilon_{0}\rho_{k,n}}\Big{(}\mathbb{E}\big{[}e^{aZ_{n}}\big{]}\Big{)}^{b_{n}^{\prime}}\Big{\\}}$
$\displaystyle\quad\leq-a\varepsilon_{0}+C^{*}\limsup_{K\to\infty}\int_{({\mathbb{R}}^{d})^{k-1}}e^{aMH({\bf
0}_{d},{\bf y})}{\mathbbm{1}}\big{\\{}H({\bf 0}_{d},{\bf
y})>K\big{\\}}\operatorname{d\\!}{\bf y}=-a\varepsilon_{0}.$
Since $a$ is arbitrary, letting $a\to\infty$ concludes the entire proof. ∎
### 5.4 Proofs of Theorem 4.1 and Theorem 4.3
###### Proof of Theorem 4.1 (Poisson input).
We first deal with the case of a Poisson input. Since the function
$\big{(}h_{s_{i}}(x_{1},x_{2},x_{3})h_{t_{i}}(x_{1},x_{2},x_{3})\big{)}_{i=1}^{m}$
defined at (4.2) satisfies conditions (H1)–(H5), a direct application of
Theorem 3.3 yields that as $n\to\infty$,
$\frac{1}{\rho_{3,n}}\,\log\mathbb{P}\bigg{(}\Big{(}\rho_{3,n}^{-1}\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n},\,|{\mathcal{Y}}|=3}g_{r_{n}s_{i},r_{n}t_{i}}({\mathcal{Y}},\mathcal{P}_{n}),\,i=1,\dots,m\Big{)}\in
A\bigg{)}\to-\inf_{{\bf x}\in A}I_{3}({\bf x}).$
To complete the proof we show that for every $0\leq s\leq t<\infty$ and
$\delta>0$,
$\lim_{n\to\infty}\frac{1}{\rho_{3,n}}\,\log\mathbb{P}\bigg{(}\beta_{1,n}(s,t)-\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n},\,|{\mathcal{Y}}|=3}g_{r_{n}s,r_{n}t}({\mathcal{Y}},\mathcal{P}_{n})\geq\delta\rho_{3,n}\bigg{)}=-\infty.$
By definition, $\beta_{1,n}(s,t)$ represents the number of persistent
$1$-cycles in the region $[0,s]\times[t,\infty]$ of the first-order
persistence diagram. In particular, $\beta_{1,n}(s,t)$ accounts for subsets of
$3$ points in ${\mathbb{R}}^{2}$ that form a single $1$-cycle before time
$r_{n}s$, such that this 1-cycle remains alive at time $r_{n}t$ and isolated
from all the remaining points in $\mathcal{P}_{n}$ at that time. Note that
these subsets of $3$ points in ${\mathbb{R}}^{2}$ are counted by
$\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n},\,|{\mathcal{Y}}|=3}g_{r_{n}s,r_{n}t}({\mathcal{Y}},\mathcal{P}_{n})$
as well. Thus, all the remaining points in $[0,s]\times[t,\infty]$ of the
first-order persistence diagram are associated to the 1-cycles on connected
components of size greater than or equal to $4$ at time $r_{n}t$. From this
point of view,
$\beta_{1,n}(s,t)-\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n},\,|{\mathcal{Y}}|=3}g_{r_{n}s,r_{n}t}({\mathcal{Y}},\mathcal{P}_{n})$
can be bounded by the first-order Betti number at time $r_{n}t$, associated
only to connected components of size greater than or equal to $4$. Moreover,
this Betti number is further bounded by the corresponding $1$-simplex counts
(i.e., edge counts). By Lemma 2.1 in [15], one can bound such $1$-simplex
counts by three times the number of vertices that are contained in connected
components of size greater than or equal to $4$. In conclusion, we have that
$\beta_{1,n}(s,t)-\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n},\,|{\mathcal{Y}}|=3}g_{r_{n}s,r_{n}t}({\mathcal{Y}},\mathcal{P}_{n})\leq
3\sum_{i=4}^{\infty}i\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n},\,|{\mathcal{Y}}|=i}v_{n}^{(i)}({\mathcal{Y}},\mathcal{P}_{n}),$
(5.47)
where
$v_{n}^{(i)}({\mathcal{Y}},\mathcal{P}_{n})={\mathbbm{1}}\big{\\{}\alpha({\mathcal{Y}},r_{n}t)\text{
is a connected component of }\alpha(\mathcal{P}_{n},r_{n}t)\big{\\}}.$
As in the proof of Proposition 5.1, we partition $[0,1]^{2}$ into sub-cubes
$Q_{1},\dots,Q_{\rho_{3,n}}$ of volume $\rho_{3,n}^{-1}$ so that
$Q_{1}=[0,\rho_{3,n}^{-1/2}]^{2}$. We claim that
$\sum_{i=4}^{\infty}i\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n},\,|{\mathcal{Y}}|=i}v_{n}^{(i)}({\mathcal{Y}},\mathcal{P}_{n})\leq\sum_{{\bf
z}\in\\{0,\pm
1\\}^{2}}\sum_{\ell=1}^{\rho_{3,n}}\sum_{i=4}^{\infty}i\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}|_{(Q_{\ell}+3r_{n}t{\bf
z})\cap[0,1]^{2}}}v_{n}^{(i)}\big{(}{\mathcal{Y}},\mathcal{P}_{n}|_{(Q_{\ell}+3r_{n}t{\bf
z})\cap[0,1]^{2}}\big{)}.$ (5.48)
To show (5.48), suppose there exists a connected component ${\mathcal{Y}}$ of
cardinality at least $4$, whose points are counted by the left hand side of
(5.48). If ${\mathcal{Y}}$ is contained in one of the cubes in
$(Q_{\ell})_{\ell=1}^{\rho_{3,n}}$, all the points in ${\mathcal{Y}}$ can also
be counted by the statistics on the right hand side of (5.48) with ${\bf
z}=(0,0)$. If we observe a subset $\mathcal{Z}\subset{\mathcal{Y}}\cap
Q_{\ell}$ for some $\ell$ with $|\mathcal{Z}|\leq 3$, such that $\mathcal{Z}$
itself forms a connected component within $Q_{\ell}$, with respect to the
process $\mathcal{P}_{n}|_{Q_{\ell}}$, then the points in $\mathcal{Z}$ will
be missed from the above statistics with ${\bf z}=(0,0)$. Even in that case,
however, all the points in $\mathcal{Z}$ can eventually be counted by the
statistics in (5.48) with other choice of ${\bf z}\in\\{0,\pm
1\\}^{2}\setminus(0,0)$. Since there are at most finitely many choices of
${\bf z}$, the entire proof will be complete, provided that for any
$\delta>0$,
$\lim_{n\to\infty}\frac{1}{\rho_{3,n}}\,\log\mathbb{P}\bigg{(}\sum_{\ell=1}^{\rho_{3,n}}\sum_{i=4}^{\infty}i\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}|_{Q_{\ell}},\,|{\mathcal{Y}}|=i}v_{n}^{(i)}\big{(}{\mathcal{Y}},\mathcal{P}_{n}|_{Q_{\ell}}\big{)}\geq\delta\rho_{3,n}\bigg{)}=-\infty.$
(5.49)
For the proof of (5.49) we apply Markov’s inequality to obtain that
$\displaystyle\frac{1}{\rho_{3,n}}\,\log\mathbb{P}\bigg{(}\sum_{\ell=1}^{\rho_{3,n}}\sum_{i=4}^{\infty}i\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}|_{Q_{\ell}},\,|{\mathcal{Y}}|=i}v_{n}^{(i)}\big{(}{\mathcal{Y}},\mathcal{P}_{n}|_{Q_{\ell}}\big{)}\geq\delta\rho_{3,n}\bigg{)}$
$\displaystyle\leq-a\delta+\log\mathbb{E}\bigg{[}\exp\Big{\\{}a\sum_{i=4}^{\infty}i\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}|_{Q_{1}},\,|{\mathcal{Y}}|=i}v_{n}^{(i)}\big{(}{\mathcal{Y}},\mathcal{P}_{n}|_{Q_{1}}\big{)}\Big{\\}}\bigg{]}.$
From this, it suffices to show that for every $a>0$,
$\mathbb{E}\bigg{[}\exp\Big{\\{}a\sum_{i=4}^{\infty}i\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}|_{Q_{1}},\,|{\mathcal{Y}}|=i}v_{n}^{(i)}\big{(}{\mathcal{Y}},\mathcal{P}_{n}|_{Q_{1}}\big{)}\Big{\\}}\bigg{]}\to
1,\ \ \text{as }n\to\infty.$ (5.50)
To show this, we first claim that
$\sum_{i=4}^{\infty}i\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}|_{Q_{1}},\,|{\mathcal{Y}}|=i}v_{n}^{(i)}\big{(}{\mathcal{Y}},\mathcal{P}_{n}|_{Q_{1}}\big{)}\stackrel{{\scriptstyle\mathbb{P}}}{{\to}}0,\
\ \ n\to\infty.$ (5.51)
Taking an expectation of (5.51),
$\displaystyle\sum_{i=4}^{\infty}i\,\mathbb{E}\Big{[}\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}|_{Q_{1}},\,|{\mathcal{Y}}|=i}v_{n}^{(i)}\big{(}{\mathcal{Y}},\mathcal{P}_{n}|_{Q_{1}}\big{)}\Big{]}\leq\sum_{i=4}^{\infty}i\,\mathbb{E}\Big{[}\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}|_{Q_{1}},\,|{\mathcal{Y}}|=i}u_{n}^{(i)}({\mathcal{Y}})\Big{]},$
where $u_{n}^{(i)}({\mathcal{Y}})\mathrel{\mathop{\mathchar
58\relax}}={\mathbbm{1}}\big{\\{}\alpha({\mathcal{Y}},r_{n}t)\text{ is
connected}\big{\\}}.$ By the Mecke formula for Poisson point processes, along
with the customary change of variables, the right hand side of the above is
equal to
$\displaystyle\sum_{i=4}^{\infty}i\,\frac{n^{i}}{i!}\,\int_{(Q_{1})^{i}}u_{n}^{(i)}(x_{1},\dots,x_{i})\operatorname{d\\!}{\bf
x}$
$\displaystyle=\sum_{i=4}^{\infty}\frac{n^{i}r_{n}^{2(i-1)}}{(i-1)!}\int_{Q_{1}}\int_{({\mathbb{R}}^{2})^{i-1}}{\mathbbm{1}}\big{\\{}\alpha\big{(}\\{{\bf
0}_{2},y_{1},\dots,y_{i-1}\\},t\big{)}\text{ is
connected}\big{\\}}\prod_{\ell=1}^{i-1}{\mathbbm{1}}\\{x+r_{n}y_{\ell}\in
Q_{1}\\}\operatorname{d\\!}{\bf y}\operatorname{d\\!}x$
$\displaystyle\leq\sum_{i=4}^{\infty}\frac{(nr_{n}^{2})^{i-3}}{(i-1)!}\,\int_{({\mathbb{R}}^{2})^{i-1}}{\mathbbm{1}}\Big{\\{}\alpha\big{(}\\{{\bf
0}_{2},y_{1},\dots,y_{i-1}\\},t\big{)}\text{ is
connected}\Big{\\}}\operatorname{d\\!}{\bf y}.$
Because of an elementary fact that there exist at most $i^{i-2}$ spanning
trees on $i$ vertices,
$\int_{({\mathbb{R}}^{2})^{i-1}}{\mathbbm{1}}\Big{\\{}\alpha\big{(}\\{{\bf
0}_{2},y_{1},\dots,y_{i-1}\\},t\big{)}\text{ is
connected}\Big{\\}}\operatorname{d\\!}{\bf y}\leq i^{i-2}(t^{2}\pi)^{i-1}.$
Referring this bound back into the above and appealing to Stirling’s formula,
i.e., $(i-1)!\geq\big{(}(i-1)/e\big{)}^{i-1}$ for large $i$, we conclude that
as $n\to\infty$,
$\displaystyle\sum_{i=4}^{\infty}i\,\mathbb{E}\Big{[}\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}|_{Q_{1}},\,|{\mathcal{Y}}|=i}u_{n}^{(i)}({\mathcal{Y}})\Big{]}$
$\displaystyle\leq\sum_{i=4}^{\infty}\frac{(nr_{n}^{2})^{i-3}}{(i-1)!}\,i^{i-2}(t^{2}\pi)^{i-1}\leq
C^{*}\sum_{i=4}^{\infty}\big{(}et^{2}\pi nr_{n}^{2}\big{)}^{i-3}\to 0.$
To conclude (5.50) from (5.51), one needs to show the uniform integrability:
for every $a>0$,
$\limsup_{n\to\infty}\mathbb{E}\bigg{[}\exp\Big{\\{}a\sum_{i=4}^{\infty}i\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}|_{Q_{1}},\,|{\mathcal{Y}}|=i}v_{n}^{(i)}\big{(}{\mathcal{Y}},\mathcal{P}_{n}|_{Q_{1}}\big{)}\Big{\\}}\bigg{]}<\infty.$
(5.52)
For the proof, we consider the family $G$ of diluted cubes of side length
$8r_{n}$ that are contained in $Q_{1}=\big{[}0,\rho_{3,n}^{-1/2}\big{]}^{2}$.
Then, the total number of such cubes in $G$ is
$c_{n}\mathrel{\mathop{\mathchar 58\relax}}=\rho_{3,n}^{-1}/(8r_{n})^{2}$,
which is assumed without loss of generality to be integer-valued for every
$n$. Observe now that for any connected component of size greater than or
equal to $4$, there exist ${\bf z}=(z_{1},z_{2})\in\big{\\{}0,\pm 1,\dots,\pm
4\big{\\}}^{2}$ and $J\in G$ such that $J+r_{n}{\bf z}\subset Q_{1}$ and
$\big{|}\mathcal{P}_{n}\cap(J+r_{n}{\bf z})\big{|}\geq 4$. In conclusion, we
have
$\displaystyle\sum_{i=4}^{\infty}i\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}|_{Q_{1}},\,|{\mathcal{Y}}|=i}v_{n}^{(i)}\big{(}{\mathcal{Y}},\mathcal{P}_{n}|_{Q_{1}}\big{)}$
$\displaystyle\leq\sum_{{\bf z}\in\\{0,\pm 1,\dots,\pm 4\\}^{2}}\sum_{J\in
G,\,J+r_{n}{\bf z}\subset Q_{1}}\big{|}\mathcal{P}_{n}\cap(J+r_{n}{\bf
z})\big{|}\,{\mathbbm{1}}\Big{\\{}\big{|}\mathcal{P}_{n}\cap(J+r_{n}{\bf
z})\big{|}\geq 4\Big{\\}}.$
Now, according to Hölder’s inequality as well as the homogeneity of
$\mathcal{P}_{n}$, (5.52) follows from
$\limsup_{n\to\infty}\mathbb{E}\Big{[}e^{a\sum_{J\in G}|\mathcal{P}_{n}\cap
J|\,{\mathbbm{1}}\\{|\mathcal{P}_{n}\cap J|\geq 4\\}}\Big{]}<\infty,$
for every $a>0$. Writing $G=\\{J_{1},\dots,J_{c_{n}}\\}$ with
$J_{1}=[0,8r_{n}]^{2}$, we have that
$\mathbb{E}\Big{[}e^{a\sum_{J\in G}|\mathcal{P}_{n}\cap
J|\,{\mathbbm{1}}\\{|\mathcal{P}_{n}\cap J|\geq
4\\}}\Big{]}=\bigg{\\{}\Big{(}\mathbb{E}\Big{[}e^{a\,|\mathcal{P}_{n}\cap
J_{1}|\,{\mathbbm{1}}\\{|\mathcal{P}_{n}\cap J_{1}|\geq
4\\}}\Big{]}\Big{)}^{1/(\rho_{3,n}r_{n}^{2})}\bigg{\\}}^{1/64}$
It is elementary to calculate that
$\mathbb{E}\Big{[}e^{a\,|\mathcal{P}_{n}\cap
J_{1}|\,{\mathbbm{1}}\\{|\mathcal{P}_{n}\cap J_{1}|\geq 4\\}}\Big{]}\leq
1+\sum_{\ell=4}^{\infty}e^{a\ell}\mathbb{P}\big{(}|\mathcal{P}_{n}\cap
J_{1}|=\ell\big{)}\leq 1+C^{*}(nr_{n}^{2})^{4}.$
Since $\rho_{3,n}r_{n}^{2}=(nr_{n}^{2})^{3}$, one can obtain that
$\displaystyle\limsup_{n\to\infty}\Big{(}\mathbb{E}\Big{[}e^{t\,|\mathcal{P}_{n}\cap
J_{1}|\,{\mathbbm{1}}\\{|\mathcal{P}_{n}\cap J_{1}|\geq
4\\}}\Big{]}\Big{)}^{1/(\rho_{3,n}r_{n}^{2})}\leq\limsup_{n\to\infty}\big{(}1+C^{*}(nr_{n}^{2})^{4}\big{)}^{1/(nr_{n}^{2})^{3}}=1.$
∎
###### Proof of Theorem 4.1 (binomial input).
In the case of a binomial input, instead of (5.47), we deduce that
$\beta_{1,n}^{\mathsf{B}}(s,t)-\sum_{{\mathcal{Y}}\subset\mathcal{B}_{n},\,|{\mathcal{Y}}|=3}g_{r_{n}s,r_{n}t}({\mathcal{Y}},\mathcal{B}_{n})\leq
3\sum_{i=4}^{n}i\sum_{{\mathcal{Y}}\subset\mathcal{B}_{n},\,|{\mathcal{Y}}|=i}v_{n}^{(i)}({\mathcal{Y}},\mathcal{B}_{n}).$
As in the proofs of Corollaries 2.3 and 3.4, we may work under the event
$\\{\mathcal{B}_{n}\subset\mathcal{P}_{n}^{(1,\mathsf{a})}\\}$, where
$\mathcal{P}_{n}^{(1,\mathsf{a})}=\mathcal{P}_{n}\cup\mathcal{P}_{n}^{(1)}\stackrel{{\scriptstyle
d}}{{=}}\mathcal{P}_{2n}$. Then, for every $\delta>0$,
$\displaystyle\begin{split}&\frac{1}{\rho_{3,n}}\log\mathbb{P}\bigg{(}\Big{\\{}\sum_{i=4}^{n}i\sum_{{\mathcal{Y}}\subset\mathcal{B}_{n},\,|{\mathcal{Y}}|=i}v_{n}^{(i)}({\mathcal{Y}},\mathcal{B}_{n})\geq\delta\rho_{3,n}\Big{\\}}\cap\\{\mathcal{B}_{n}\subset\mathcal{P}_{n}^{(1,\mathsf{a})}\\}\bigg{)}\\\
&\leq\frac{1}{\rho_{3,n}}\log\mathbb{P}\Big{(}\sum_{i=4}^{\infty}i\sum_{{\mathcal{Y}}\subset\mathcal{P}_{2n},\,|{\mathcal{Y}}|=i}v_{n}^{(i)}({\mathcal{Y}},\mathcal{P}_{2n})\geq\delta\rho_{3,n}\Big{)}.\end{split}$
(5.53)
Here, we used that the double sum counts the number of vertices in connected
components of size at least 4. Moreover, adding further points in
$\mathcal{P}_{n}^{(1,\mathsf{a})}\setminus\mathcal{B}_{n}$ increases the
number of points in the components associated to $\mathcal{B}_{n}$. This is
clear for the Čech complex and follows by the nerve lemma for the alpha
complex. Repeating the same argument as in the Poisson input case, one can
show that the right-hand side in (LABEL:e:binomial.betti.to.Poisson) goes to
$-\infty$ as $n\to\infty$. ∎
###### Proof of Theorem 4.3.
We start by formulating the function $H\mathrel{\mathop{\mathchar
58\relax}}=(h^{(1)},\dots,h^{(m)})\mathrel{\mathop{\mathchar
58\relax}}({\mathbb{R}}^{2})^{3}\to[0,\infty)^{m}$ in the current setup:
$h^{(i)}(x_{1},x_{2},x_{3})\mathrel{\mathop{\mathchar
58\relax}}={\mathbbm{1}}\big{\\{}\gamma(x_{1},x_{2},x_{3})\in\text{conv}^{\circ}(x_{1},x_{2},x_{3}),\,\mathcal{R}(x_{1},x_{2},x_{3})\leq
t_{i}\big{\\}}.$
Clearly, $H$ satisfies conditions (H1)–(H5). In particular, fix a constant $L$
determined by property (H3). Define the scaled version $H_{n}$ of $H$ as in
(3.1). For a $3$-point subset ${\mathcal{Y}}\subset{\mathbb{R}}^{2}$ and a
finite subset $\mathcal{Z}\supset{\mathcal{Y}}$ in ${\mathbb{R}}^{2}$, define
$c({\mathcal{Y}},\mathcal{Z})\mathrel{\mathop{\mathchar
58\relax}}=\big{(}{\mathbbm{1}}\big{\\{}\mathcal{U}({\mathcal{Y}})\cap\mathcal{Z}=\emptyset\big{\\}}\big{)}_{i=1}^{m}.$
(5.54)
In contrast to (3.2), the function (5.54) does not involve a time parameter
${\bf t}=(t_{1},\dots,t_{m})$. Furthermore, unlike (3.3),
$c_{n}({\mathcal{Y}},\mathcal{Z})\mathrel{\mathop{\mathchar
58\relax}}=c(r_{n}^{-1}{\mathcal{Y}};r_{n}^{-1}\mathcal{Z})=c({\mathcal{Y}},\mathcal{Z})$
does not depend on $n\geq 1$. Finally we define
$s_{n}({\mathcal{Y}},\mathcal{Z})\mathrel{\mathop{\mathchar
58\relax}}=c_{n}({\mathcal{Y}},\mathcal{Z})\,{\mathbbm{1}}\big{\\{}\mathsf{diam}({\mathcal{Y}})\leq
r_{n}L\big{\\}}$ as in (2.3).
The required large deviations in Theorem 4.3 can be deduced by an application
of Theorem 3.3 to the number of Morse critical points in (4.4). Before doing
so, however, one must properly modify the proof of Theorem 2.1 under the setup
of Theorem 4.3. After that, one must also modify the proofs of Theorems 3.1
and 3.3; however, the required modification for these two theorems seems to be
sufficiently simple, so we focus our attention only to Theorem 2.1. In the
below, we discuss two specific points. First, one has to modify the
calculation in (5.11) as follows: as $n\to\infty$,
$\displaystyle\frac{n^{3}}{6}\int_{(Q_{\ell})^{3}}{\mathbbm{1}}\big{\\{}\mathsf{diam}(x_{1},x_{2},x_{3})\leq
r_{n}L\big{\\}}\Big{(}1-e^{-n\text{vol}\big{(}\mathcal{U}(x_{1},x_{2},x_{3})\cap
Q_{\ell}\big{)}}\Big{)}\operatorname{d\\!}{\bf x}$
$\displaystyle=\frac{\rho_{3,n}}{6}\int_{Q_{\ell}}\int_{({\mathbb{R}}^{2})^{2}}{\mathbbm{1}}\big{\\{}\mathsf{diam}({\bf
0}_{2},{\bf y})\leq L\big{\\}}\prod_{i=1}^{2}{\mathbbm{1}}\\{x+r_{n}y_{i}\in
Q_{\ell}\\}$
$\displaystyle\qquad\qquad\qquad\qquad\times\Big{(}1-e^{-n\text{vol}\big{(}\mathcal{U}(x,x+r_{n}y_{1},x+r_{n}y_{2})\cap
Q_{\ell}\big{)}}\Big{)}\operatorname{d\\!}{\bf y}\operatorname{d\\!}x$
$\displaystyle\leq\frac{1}{6}\int_{({\mathbb{R}}^{2})^{2}}{\mathbbm{1}}\big{\\{}\mathsf{diam}({\bf
0}_{2},{\bf y})\leq L\big{\\}}\Big{(}1-e^{-nr_{n}^{2}\mathcal{R}({\bf
0}_{2},{\bf y})^{2}\pi}\Big{)}\operatorname{d\\!}{\bf y}\to 0.$
Second, we also need to modify the proof of (5.16). Specifically, we need to
show that
$\limsup_{n\to\infty}\mathbb{E}\Big{[}e^{a\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}|_{Q_{1}}}s_{n}({\mathcal{Y}},\mathcal{P}_{n}|_{Q_{1}})}\Big{]}<\infty.$
For this purpose, we again consider the diluted families of cubes
$G_{1},G_{2},\dots$ as in (5.17) (with $t\equiv 1$). Then, as an analog of
(5.19), our task is reduced to showing that for every $a>0$,
$\limsup_{n\to\infty}\bigg{(}\mathbb{E}\Big{[}e^{a\sum_{{\mathcal{Y}}\subset\mathcal{P}_{n}|_{Q_{1}}}s_{n}({\mathcal{Y}},\mathcal{P}_{n}|_{Q_{1}}){\mathbbm{1}}\\{\ell({\mathcal{Y}})\in
J_{1}\\}}\Big{]}\bigg{)}^{b_{n}}<\infty,$ (5.55)
where $J_{1}=[0,r_{n}/\sqrt{2}]^{2}$ and
|
Tractable and Intractable Entailment Problems in Separation Logic
# Tractable and Intractable Entailment Problems in Separation Logic with
Inductively Defined Predicates
Mnacho Echenim This work has been partially funded by the the French National
Research Agency (ANR-21-CE48-0011). Nicolas Peltier1
Univ. Grenoble Alpes CNRS Grenoble INP LIG 38000 Grenoble France
###### Abstract
We establish various complexity results for the entailment problem between
formulas in Separation Logic (SL) with user-defined predicates denoting
recursive data structures. The considered fragments are characterized by
syntactic conditions on the inductive rules that define the semantics of the
predicates. We focus on so-called $\mathtt{P}$-rules, which are similar to
(but simpler than) the so-called bounded treewidth fragment of SL studied by
Iosif et al. [14]. In particular, for a specific fragment where predicates are
defined by so-called _$\mathtt{loc}$ -deterministic_ inductive rules, we
devise a sound and complete cyclic proof procedure running in polynomial time.
Several complexity lower bounds are provided, showing that any relaxing of the
provided conditions makes the problem intractable.
###### keywords:
Separation logic, Inductive reasoning, Decision procedures, Cyclic proofs
††terms: def=††terms: _
#### ACM Computing Classification System:
$\bullet$ Theory of computation, Logic, Automated reasoning $\bullet$ Theory
of computation, Logic, Separation logic
## 1 Introduction
Separation Logic [16, 22] (SL) is widely used in verification to reason on
programs manipulating pointer-based data structures. It forms the basis of
several automated static program analyzers such as Smallfoot [1], Infer [3]
(Facebook) or SLAyer [2] (Microsoft Research) and several correctness proofs
were carried out by embedding SL in interactive theorem provers such as Coq
[23], see for instance [17]. SL uses a special connective $*$, called
separating conjunction, modeling heap compositions and allowing for concise
and natural specifications. More precisely, atoms in SL are expressions of the
forms $x\mapsto(y_{1},\dots,y_{n})$, where $x,y_{1},\dots,y_{n}$ are variables
denoting locations (i.e., memory addresses), asserting that location $x$ is
allocated and refers to the tuple (record) $(y_{1},\dots,y_{n})$. The special
connective $\phi*\psi$ asserts that formulas $\phi$ and $\psi$ hold on
disjoint parts of the memory. Recursive data structures may then be described
by considering predicates associated with inductive rules, such as:
$\begin{array}[]{llllll}{\mathtt{ls}}(x,y)&\Leftarrow&x\mapsto
y&{\mathtt{tree}}(x)&\Leftarrow&x\mapsto()\\\
{\mathtt{ls}}(x,y)&\Leftarrow&x\mapsto
z*{\mathtt{ls}}(z,y)&{\mathtt{tree}}(x)&\Leftarrow&x\mapsto(y,z)*{\mathtt{tree}}(y)*{\mathtt{tree}}(z)\end{array}$
where ${\mathtt{ls}}(x,y)$ denotes a nonempty list segment and
${\mathtt{tree}}(x)$ denotes a tree. For the sake of genericity, such rules
are not built-in but may be provided by the user. Due to the expressive power
of such inductive definitions, the input language is usually restricted in
this context to so-called symbolic heaps, i.e., existentially quantified
conjunctions and separating conjunctions of atoms (dismissing for instance
universal quantifications, negations and separating implications). Many
problems in verification require to solve entailment problems between such SL
formulas, for instance when these formulas denote pre- or post-conditions of
programs. Unfortunately, the entailment problem between symbolic heaps is
undecidable in general [20], but it is decidable if the considered inductive
rules satisfy the so-called PCE conditions (standing for Progress,
Connectivity and Establishment) [14]. However even for the PCE fragment the
complexity of the entailment problem is still very high; more precisely, this
problem is $2$-ExpTime-complete [9, 10, 18]. Less expressive fragments have
thus been considered, for which more efficient algorithms were developed. In
[15] a strict subclass of PCE entailments is identified with an ExpTime
complexity based on a reduction to the language inclusion problem for tree
automata [7]. In [12], an algorithm is developed to handle various kinds of
(possibly nested) singly linked lists based on a reduction to the membership
problem for tree automata. The complexity of the procedure is dominated by the
boolean satisfiability and unsatisfiability tests, that are NP and co-NP
complete, respectively. A polynomial proof procedure has been devised for the
specific case of singly linked lists [8]. In [5], the tractability result is
extended to more expressive fragments, with formulas defined on some unique
nonlinear compositional inductive predicate with distinguished source,
destination, and static parameters. The compositional properties satisfied by
the considered predicate (as originally introduced in [13]) ensure that the
entailment problem can be solved efficiently. Recently [19] introduced a
polynomial-time cyclic proof to solve entailment problem efficiently, under
some condition on the inductive rules.
In the present paper, we study the complexity of the entailment problem for a
specific fragment that is similar to the PCE fragment, but simpler. The
fragment inherits most of the conditions given in [14] and admits an
additional restriction that is meant to ensure that entailment problems can be
solved in a more efficient way111At the cost, of course, of a loss of
expressivity.: every predicate is bound to allocate exactly one of its
parameters (forbidding for instance predicates denoting doubly linked list
segments from $x$ to $y$, as both $x$ and $y$ would be allocated). This means
that the rules do not allow for multiple pointers into a data structure
(whereas multiple pointers out of the structure are allowed). We first show
that this additional restriction is actually not sufficient to ensure
tractability. More precisely, we establish several lower-bound complexity
results for the entailment problem under various additional hypotheses.
Second, we define a new class of inductive definitions for which the
entailment problem can be solved in polynomial time, based mainly on the two
following additional restrictions: (i) the arity of the predicates is bounded;
and (ii) the rules defining the same predicate do not overlap, in a sense that
will be formally defined below. Both conditions are rather natural
restrictions in the context of programming. Indeed, the number of parameters
is usually small in this context. Also, data structures are typically defined
using a finite set of free constructors, which yields inductive definitions
that are trivially non-overlapping.
If Condition $(i)$ is not satisfied, then the complexity is simply
exponential. In contrast with other polynomial-time algorithms, the formulas
we consider may contain several inductive predicates, and these predicates are
possibly non-compositional (in the sense of [12]). The algorithm for testing
entailment is defined as a sequent-like cyclic proof procedure, with standard
unfolding and decomposition rules, together with a specific strategy ensuring
efficiency and additional syntactic criteria to detect and dismiss non-
provable sequents. Our approach is close to that of [19], in the sense that
the two procedures use cyclic proof procedures with non-disjunctive
consequents, however the conditions on the rules are completely different: our
definition allows for multiple inductive rules with mutually recursive
definitions, yielding richer recursive data structures. On the other hand, the
SHLIDe rules in [19] support ordering and equality relations on non-
addressable values, whereas the predicate we consider are purely spatial.
Moreover, the base cases of the rules in [19] correspond to empty heaps, which
are forbidden in our approach.
To provide some intuition on what can and cannot be expressed in the fragment
we consider, we provide some examples (formal definitions will be given
later); consider the predicate $\mathtt{P}$ defined by the following rules,
which encode a combination of lists and trees, possibly looping on an initial
element $y$, and ending with an empty tuple:
$\mathtt{P}(x,y)$ | $\Leftarrow$ | $x\mapsto(\mathtt{list},u)*\mathtt{P}(u,y)$
---|---|---
$\mathtt{P}(x,y)$ | $\Leftarrow$ | $x\mapsto(\mathtt{tree},u_{1},u_{2})*\mathtt{P}(u_{1},y)*\mathtt{P}(u_{2},y)$
$\mathtt{P}(x,y)$ | $\Leftarrow$ | $x\mapsto(\mathtt{loop},y)$
$\mathtt{P}(x,y)$ | $\Leftarrow$ | $x\mapsto()$
The constants $\mathtt{list},\mathtt{tree}$ and $\mathtt{loop}$ may be viewed
as constructors for the data structure. This predicate does not fall in the
scope of the fragment considered in [19] since it involves a definition with
several inductive rules, but it falls in the scope of the fragment considered
in the present paper. Our restrictions require that the definition must be
deterministic, in the sense that there can be no overlap between distinct
rules. This is the case here, as the tuples $(\mathtt{list},u)$,
$(\mathtt{tree},u_{1},u_{2})$ and $(\mathtt{loop},y)$ are pairwise distinct
(not unifiable), but replacing for instance the constant $\mathtt{loop}$ by
$\mathtt{list}$ in the third rule would not be possible, as the resulting rule
would overlap with the first one (both rules could allocate the same heap
cell). As explained above, a key limitation of our fragment (compared to that
of [14]) is that it does not allow predicates allocating several parameters,
such as the following predicate ${\mathtt{dllseg}}(x,y,z,u)$ defining a doubly
linked list segment from $x$ to $z$ (each cell points to a pair containing the
previous and next element and $y$ and $u$ denote the previous and next element
in the list, respectively):
${\mathtt{dllseg}}(x,y,z,u)$ | $\Leftarrow$ | $(x\mapsto(y,x^{\prime})*{\mathtt{dllseg}}(x^{\prime},x,z,u))\wedge x\not\approx z$
---|---|---
${\mathtt{dllseg}}(x,y,z,u)$ | $\Leftarrow$ | $x\mapsto(y,u)\wedge x\approx z$
Other definitions of ${\mathtt{dllseg}}$ are possible, but none would fit in
with our restrictions: in every case, both $x$ (the beginning of the list) and
$z$ (its end) must be eventually allocated, which is not permitted in the
fragment we consider. On the other hand, the following predicate, defining a
doubly linked list, ending with $()$, can be defined ($y$ denotes the previous
element in the list):
${\mathtt{dll}}(x,y)$ | $\Leftarrow$ | $x\mapsto(y,z)*{\mathtt{dll}}(z,x)$
---|---|---
${\mathtt{dll}}(x,y)$ | $\Leftarrow$ | $x\mapsto()$
The rest of the paper is organised as follows. In Section 2, the syntax and
semantics of the logic are defined. The definitions are mostly standard,
although we consider a multisorted framework, with a special sort
$\mathtt{loc}$ denoting memory locations and additional sorts for data or
constructors. We then introduce a class of inductive definitions called
$\mathtt{P}$-rules. In Section 3, various lower bounds on the complexity of
the entailment problem for SL formulas with $\mathtt{P}$-rules are established
which allow one to motivate additional restrictions on the inductive rules.
These lower bounds show that all the restrictions are necessary to ensure that
the entailment problem is tractable. This leads to the definition of the
notion of a $\mathtt{loc}$-deterministic set of rules, that is a subset of
$\mathtt{P}$-rules for which entailment can be decided in polynomial time. The
proof procedure is defined in Section 4. For the sake of readability and
generality we first define generic inference rules and establish their
correctness, before introducing a specific strategy to further restrict the
application of the rules that is both complete and efficient. Section 5
contains all soundness, completeness and complexity results and Section 6
concludes the paper.
## 2 Definitions
### 2.1 Syntax
We use a multisorted framework, which is essentially useful to distinguish
locations from data. Let $\mathfrak{S}$ be a set of sorts, containing a
special sort $\mathtt{loc}$, denoting memory locations. Let
$\mathcal{V}_{{\mathtt{s}}\in\mathfrak{S}}$ be a family of countably infinite
disjoint sets of variables of sort ${\mathtt{s}}$, with
$\mathcal{V}\overset{\text{\tiny\it
def}}{=}\bigcup_{{\mathtt{s}}\in\mathfrak{S}}\mathcal{V}_{\mathtt{s}}$. Let
$\mathcal{C}_{{\mathtt{s}}\in\mathfrak{S}}$ be a family of disjoint sets of
constant symbols of sort ${\mathtt{s}}$, also disjoint from $\mathcal{V}$,
with $\mathcal{C}\overset{\text{\tiny\it
def}}{=}\bigcup_{{\mathtt{s}}\in\mathfrak{S}}\mathcal{C}_{\mathtt{s}}$. The
set of terms of sort ${\mathtt{s}}$ is ${\mathtt{s}}\overset{\text{\tiny\it
def}}{=}\mathcal{V}_{\mathtt{s}}\cup\mathcal{C}_{\mathtt{s}}$, and we let
$\bigcup_{{\mathtt{s}}\in\mathfrak{S}}{\mathtt{s}}$. Constants are especially
useful in our framework to denote constructors in data structures. To simplify
technicalities, we assume that there is no constant of sort $\mathtt{loc}$,
i.e., $\mathcal{C}_{\mathtt{loc}}=\emptyset$.
An equation (resp. a disequation) is an expression of the form $t\approx s$
(resp. $t\not\approx s$) where $t,s\in{\mathtt{s}}$ for some
${\mathtt{s}}\in\mathfrak{S}$. The set of pure formulas ${\cal F}_{P}$ is the
set of formulas of the form $e_{1}\wedge\dots\wedge e_{n}$, where every
expression $e_{i}$ is either an equation or a disequation. Such formulas are
considered modulo contraction, e.g., a pure formula $\xi\wedge\xi$ is
considered identical to $\xi$, and also modulo associativity and commutativity
of conjunction. We denote by $\bot$ (false) any formula of the form
$t\not\approx t$. If $n=0$, then $\bigwedge_{i=1}^{n}e_{i}$ may be denoted by
$\top$ (true). If $(t_{1},\dots,t_{n})$ and $(s_{1},\dots,s_{m})$ are vectors
of terms, then $(t_{1},\dots,t_{n})\approx(s_{1},\dots,s_{m})$ denotes the
formula $\bot$ if either $n\not=m$ or $n=m$ and there exists
$i\in\\{1,\dots,n\\}$ such that $s_{i}$ and $t_{i}$ are of different sorts;
and denotes $\bigwedge_{i=1}^{n}t_{i}\approx s_{i}$ otherwise.
Let $\mathcal{P}$ be a set of predicate symbols. Each symbol in $\mathcal{P}$
is associated with a unique profile of the form
$({\mathtt{s}}_{1},\dots,{\mathtt{s}}_{n})$ with $n\geq 1$,
${\mathtt{s}}_{1}=\mathtt{loc}$ and ${\mathtt{s}}_{i}\in\mathfrak{S}$, for all
$i\in\\{2,\dots,n\\}$. We write $p:{\mathtt{s}}_{1},\dots,{\mathtt{s}}_{n}$ to
denote a symbol with profile $({\mathtt{s}}_{1},\dots,{\mathtt{s}}_{n})$ and
we write $p:{\mathtt{s}}_{1},\dots,{\mathtt{s}}_{n}\in\mathcal{P}$ to state
that $p$ is a predicate symbol of profile
${\mathtt{s}}_{1},\dots,{\mathtt{s}}_{n}$ in $\mathcal{P}$. A spatial atom
$\alpha$ is either a points-to atom $x\mapsto(t_{1},\dots,t_{n})$ with
$x\in\mathcal{V}_{\mathtt{loc}}$ and $t_{1},\dots,t_{n}\in,ora{\em
predicateatom}oftheform$p(x,t_1,…,t_n)$,where$p$isapredicateofprofile$loc,s_1,…,s_n$in$P$,theterm$x$isavariablein$V_loc$and$t_i
∈s_i$forall$i ∈{ 1,…, n}$.Inbothcases,thevariable$x$iscalledthe{\em
root}of$α$andisdenotedby$root(α)$.\par Thesetof{\em
spatialformulas}$F_S$isthesetofformulasoftheform$β_1 * …*
β_n$,whereeveryexpression$β_i$isaspatialatom.If$n = 0$then$β_1 * …*
β_n$isdenotedby$emp$.Thenumber$n$ofoccurrencesofspatialatomsinaspatialformula$ϕ=
β_1 * …* β_n$isdenotedby$len(ϕ)$.Wewrite$ϕ⊑ψ$if$ψ$isoftheform$ϕ*
ϕ’$,moduloassociativityandcommutativityof$*$.Thesetof(nonquantified){\em
symbolicheaps}$F_H$isthesetofexpressionsoftheform$ϕ⋏ξ$,where$ϕ∈F_S$and$ξ∈F_P$.Notethatforclarityweuse$⋏$todenoteconjunctionsbetweenspatialandpureformulasand$∧$todenoteconjunctionsoccurringwithinpureformulas.If$ξ=
⊤$,then$ϕ⋏ξ$maybewritten$ϕ$(i.e.,anyspatialformulamaybeviewedasasymbolicheap).Foranyformula$λ,—λ—$denotesthesizeof$λ$(whichisdefinedinductivelyasusual).Notethat$⊤$isnotasymbolicheap(but$emp⋏⊤$isasymbolicheap).\par
Wedenoteby$V(β)$(resp.\ $V_s(β)$)thesetofvariables(resp.\
ofvariablesofsort$s$)occurringinavariableorformula$β$.A{\em
substitution}isasort-preservingtotalmappingfrom$V$to$. We denote by
$\mathit{dom}(\sigma)$ the set of variables such that $\sigma(x)\not=x$, and
by $\mathit{codom}(x)$ the set $\\{\sigma(x)\mid x\in\mathit{dom}(\sigma)\\}$.
The substitution $\sigma$ such that $\sigma(x_{i})=y_{i}$ for all
$i=1,\dots,n$ and $\mathit{dom}(\sigma)\subseteq\\{x_{1},\dots,x_{n}\\}$ is
denoted by $\replall{}{x_{i}}{y_{i}}{i=1,\dots,n}$. For any expression $\beta$
and substitution $\sigma$, we denote by $\beta\sigma$ the expression obtained
from $\beta$ by replacing every variable $x$ by $\sigma(x)$. A unifier of two
expressions or tuples of expressions $\beta$ and $\beta^{\prime}$ is a
substitution $\sigma$ such that $\beta\sigma=\beta^{\prime}\sigma$.
An inductive rule is an expression of the form
$p(x_{1},\dots,x_{n})\Leftarrow\lambda$, where
$p:{\mathtt{s}}_{1},\dots,{\mathtt{s}}_{n}\in\mathcal{P}$, $x_{1},\dots,x_{n}$
are pairwise distinct variables of sorts
${\mathtt{s}}_{1},\dots,{\mathtt{s}}_{n}$ respectively and $\lambda\in{\cal
F}_{H}$. The set of variables in
$\mathcal{V}(\lambda)\setminus\\{x_{1},\dots,x_{n}\\}$ are the existential
variables of the rule. Let $\mathfrak{R}$ be a set of inductive rules. We
write $p(t_{1},\dots,t_{n})\Leftarrow_{\mathfrak{R}}\lambda$ iff
$\mathfrak{R}$ contains (up to a renaming, and modulo AC) a rule of the form
$p(y_{1},\dots,y_{n})\Leftarrow\gamma$ and
$\lambda=\replall{\gamma}{y_{i}}{t_{i}}{i=1,\dots,n}$. We assume by renaming
that $\gamma$ contains no variable in $\left\\{t_{1},\dots,t_{n}\right\\}$. We
write $p(t_{1},\dots,t_{n})\rightsquigarrow_{\mathfrak{R}}E$ if $E$ is the set
of symbolic heaps $\lambda$ such that
$p(t_{1},\dots,t_{n})\Leftarrow_{\mathfrak{R}}\lambda$. Note that if
$\mathfrak{R}$ is finite then $E$ is finite up to a renaming of variables not
occurring in $\\{t_{1},\dots,t_{n}\\}$. Note also that the considered logic
does not allow for negations (hence entailment is not reducible to
satisfiability) or separating implications, as this would make satisfiability
undecidable (see for instance [21]).
The symbol $\subseteq_{m}$ denotes the inclusion relation between multisets.
With a slight abuse of notations, we will sometimes identify sequences with
sets when the order and number of repetitions is not important, for instance
we may write $\bm{x}\subseteq\bm{y}$ to state that every element of $\bm{x}$
occurs in $\bm{y}$.
In the present paper, we shall consider entailment problems between symbolic
heaps.
### 2.2 Semantics
We assume for technical convenience that formulas are interpreted over a fixed
universe and that constants are interpreted as pairwise distinct elements. Let
$\mathfrak{U}_{{\mathtt{s}}\in\mathfrak{S}}$ be pairwise disjoint countably
infinite sets and let $\mathfrak{U}\overset{\text{\tiny\it
def}}{=}\bigcup_{{\mathtt{s}}\in\mathfrak{S}}\mathfrak{U}_{\mathtt{s}}$. We
assume that an injective function is given, mapping every constant
$c\in\mathcal{C}_{\mathtt{s}}$ to an element of $\mathfrak{U}_{{\mathtt{s}}}$,
denoted by $\dot{c}$.
A heap is a partial finite function from $\mathfrak{U}_{\mathtt{loc}}$ to
$\mathfrak{U}^{*}$, where $\mathfrak{U}^{*}$ denotes as usual the set of
finite sequences of elements of $\mathfrak{U}$. An element
${\ell}\in\mathfrak{U}_{\mathtt{loc}}$ is allocated in a heap $\mathfrak{h}$
if ${\ell}\in\mathit{dom}(\mathfrak{h})$. Two heaps $\mathfrak{h}$ and
$\mathfrak{h}^{\prime}$ are disjoint if
$\mathit{dom}(\mathfrak{h})\cap\mathit{dom}(\mathfrak{h}^{\prime})=\emptyset$,
in which case $\mathfrak{h}\uplus\mathfrak{h}^{\prime}$ denotes their disjoint
union. We write $\mathfrak{h}\subseteq\mathfrak{h}^{\prime}$ if there is a
heap $\mathfrak{h}^{\prime\prime}$ such that
$\mathfrak{h}^{\prime}=\mathfrak{h}\uplus\mathfrak{h}^{\prime\prime}$. For
every heap $\mathfrak{h}$, we denote by $\mathit{ref}(\mathfrak{h})$ the set
of elements $\ell\in\mathfrak{U}_{\mathtt{loc}}$ such that there exists
$\ell_{0},\dots,\ell_{n}$ with $\ell_{0}\in\mathit{dom}(\mathfrak{h})$,
$\mathfrak{h}(\ell_{0})=(\ell_{1},\dots,\ell_{n})$ and $\ell=\ell_{i}$ for
some $i=0,\dots,n$. We write $\ell\rightarrow_{\mathfrak{h}}\ell^{\prime}$ if
$(\ell,\ell^{\prime})\in\mathfrak{U}_{\mathtt{loc}}^{2}$,
$\ell\in\mathit{dom}(\mathfrak{h})$,
$\mathfrak{h}(\ell)=(\ell_{1},\dots,\ell_{n})$ and $\ell^{\prime}=\ell_{i}$,
for some $i=1,\dots,n$.
###### Proposition 2.1
Let $\mathfrak{h},\mathfrak{h}^{\prime}$ be two heaps such that
$\mathfrak{h}\subseteq\mathfrak{h}^{\prime}$. For all
$\ell,\ell^{\prime}\in\mathfrak{U}_{\mathtt{loc}}$, if
$\ell\rightarrow_{\mathfrak{h}}^{*}\ell^{\prime}$ then
$\ell\rightarrow_{\mathfrak{h}^{\prime}}^{*}\ell^{\prime}$.
###### Proof 2.2
By definition of the relation $\rightarrow_{\mathfrak{h}}$ we have
$\rightarrow_{\mathfrak{h}}\subseteq\rightarrow_{\mathfrak{h}^{\prime}}$, thus
$\rightarrow_{\mathfrak{h}}^{*}\subseteq\rightarrow_{\mathfrak{h}^{\prime}}^{*}$
A store $\mathfrak{s}$ is a total function mapping every term in
${\mathtt{s}}$ to an element in $\mathfrak{U}_{\mathtt{s}}$ such that
$\mathfrak{s}(c)=\dot{c}$, for all $c\in\mathcal{C}$ (note that this entails
that $\mathfrak{s}$ is injective on $\mathcal{C}$). A store $\mathfrak{s}$ is
injective on a multiset of variables $V$ if
$\\{x,y\\}\subseteq_{m}V\implies\mathfrak{s}(x)\not=\mathfrak{s}(y)$. When a
store is injective on a multiset of variables, this entails that the latter is
a set, i.e., that contains at most one occurrence of each variable. For any
$V\subseteq\mathcal{V}$, and for any store $\mathfrak{s}$, a store
$\mathfrak{s}^{\prime}$ is an associate of $\mathfrak{s}$ w.r.t. $V$ if
$\mathfrak{s}(x)=\mathfrak{s}^{\prime}(x)$ holds for all $x\not\in V$.
###### Definition 2.3
An SL-structure is a pair $(\mathfrak{s},\mathfrak{h})$, where $\mathfrak{s}$
is a store and $\mathfrak{h}$ is a heap.
The satisfiability relation on SL-formulas is defined inductively as follows:
###### Definition 2.4
An SL-structure $(\mathfrak{s},\mathfrak{h})$ validates a formula (pure
formula, spatial formula, or symbolic heap) $\lambda$ modulo a set of
inductive rules $\mathfrak{R}$, written
$(\mathfrak{s},\mathfrak{h})\models_{\mathfrak{R}}\lambda$, if one of the
following conditions holds:
* •
$\lambda=\mathit{emp}$ and $\mathfrak{h}=\emptyset$;
* •
$\lambda=(t\approx s)$ and $\mathfrak{s}(t)=\mathfrak{s}(s)$;
* •
$\lambda=(t\not\approx s)$ and $\mathfrak{s}(t)\not=\mathfrak{s}(s)$;
* •
$\lambda=x\mapsto(t_{1},\dots,t_{k})$ and
$\mathfrak{h}=\\{(\mathfrak{s}(x),\mathfrak{s}(t_{1}),\dots,\mathfrak{s}(t_{k}))\\}$;
* •
either $\lambda=\lambda_{1}\wedge\lambda_{2}$ or
$\lambda=\lambda_{1}\curlywedge\lambda_{2}$, and
$(\mathfrak{s},\mathfrak{h})\models_{\mathfrak{R}}\lambda_{i}$ for all
$i=1,2$;
* •
$\lambda=\lambda_{1}*\lambda_{2}$ and there exist disjoint heaps
$\mathfrak{h}_{1}$ and $\mathfrak{h}_{2}$ such that
$\mathfrak{h}=\mathfrak{h}_{1}\uplus\mathfrak{h}_{2}$ and
$(\mathfrak{s},\mathfrak{h}_{i})\models_{\mathfrak{R}}\lambda_{i}$ for all
$i=1,2$;
* •
$\lambda=p(t_{1},\dots,t_{n})$ and
$p(t_{1},\dots,t_{n})\Leftarrow_{\mathfrak{R}}\gamma$, where there exists an
associate $\mathfrak{s}^{\prime}$ of $\mathfrak{s}$ w.r.t. the set
$\mathcal{V}(\gamma)\setminus\mathcal{V}(\lambda)$ such that
$(\mathfrak{s}^{\prime},\mathfrak{h})\models_{\mathfrak{R}}\gamma$.
An $\mathfrak{R}$-model of a formula $\lambda$ is a structure
$(\mathfrak{s},\mathfrak{h})$ such that
$(\mathfrak{s},\mathfrak{h})\models_{\mathfrak{R}}\lambda$. A formula is
satisfiable (w.r.t. $\mathfrak{R}$) if it admits at least one
$\mathfrak{R}$-model.
###### Remark 2.5
Note that a formula $x\mapsto(t_{1},\dots,t_{k})$ asserts not only that $x$
refers to $(t_{1},\dots,t_{k})$ but also that $x$ is the only allocated
location. This fits with usual definitions (see, e.g., [16]). The assertions
are meant to describe elementary heaps, which can be combined afterwards using
the connective $*$. Simply asserting that $x$ refers to $(t_{1},\dots,t_{k})$
could be done in full SL using the following formula:
$x\mapsto(t_{1},\dots,t_{k})*\top$, but such a formula is not a symbolic heap
and is thus outside of the fragment we consider in the present paper.
We write $(\mathfrak{s},\mathfrak{h})\models\lambda$ instead of
$(\mathfrak{s},\mathfrak{h})\models_{\mathfrak{R}}\lambda$ if $\lambda$
contains no predicate symbol, since the relation is independent of
$\mathfrak{R}$ in this case. Similarly, if $\lambda$ is a pure formula then
the relation does not depend on the heap, thus we may simply write
$\mathfrak{s}\models\lambda$.
###### Remark 2.6
Note that there is no symbolic heap that is true in every structure. For
instance $\mathit{emp}\curlywedge\top$ is true only in structures with an
empty heap.
###### Proposition 2.7
Let $(\mathfrak{s},\mathfrak{h})$ be a structure and let $\sigma$ be a
substitution. If
$(\mathfrak{s},\mathfrak{h})\models_{\mathfrak{R}}\lambda\sigma$ then we have
${(\mathfrak{s}\circ\sigma,\mathfrak{h})\models_{\mathfrak{R}}\lambda}$.
###### Proof 2.8
By an immediate induction on the satisfiability relation.
In the present paper, we shall consider inductive rules of a particular form,
defined below.
###### Definition 2.9
An inductive rule is a $\mathtt{P}$-rule if it is of the form
$p(x_{1},\dots,x_{n})\Leftarrow
x_{1}\mapsto(y_{1},\dots,y_{k})*q_{1}(z_{1},\bm{u}_{1})*\dots
q_{m}(z_{m},\bm{u}_{m})\curlywedge\xi$
possibly with $m=0$, where:
1. 1.
$\xi$ is a conjunction of disequations of the form $u\not\approx v$, where
$u\in\\{x_{1},\dots,x_{n},y_{1},\dots,y_{k}\\}$ and
$v\in\\{y_{1},\dots,y_{k}\\}\setminus\\{x_{1},\dots,x_{n}\\}$;
2. 2.
$\\{z_{1},\dots,z_{m}\\}=(\\{y_{1},\dots,y_{k}\\}\setminus\\{x_{1},\dots,x_{n}\\})\cap\mathcal{V}_{\mathtt{loc}}$,
and $z_{1},\dots,z_{m}$ are pairwise distinct;
3. 3.
All the elements of $\bm{u}_{i}$ occurs in
$\\{x_{1},\dots,x_{n}\\}\cup\\{y_{1},\dots,y_{k}\\}\cup\mathcal{C}$.
The predicate symbol $p$ is called the head of the rule.
###### Example 2.10
The rules associated with ${\mathtt{ls}}$ and ${\mathtt{tree}}$ in the
introduction are $\mathtt{P}$-rules, as well as the following rules.
Intuitively ${\mathtt{als}}(x,y)\curlywedge x\not\approx y$ denotes an acyclic
list (${\mathtt{als}}(x,y)$ is thus “quasi-acyclic”, in the sense that it may
loop only on the first element). Note that the constraint $x\not\approx y$
cannot be added to the right-hand side of the rules because the obtained rule
would not be a $\mathtt{P}$-rule, hence it must be added in the formula. The
atom ${\mathtt{dll}}(x,y)$ denotes a doubly linked list starting at $x$, with
the convention that each element of the list points to a pair containing the
previous and next elements. The parameter $y$ denotes the element before $x$
and the last element points to the empty tuple $()$. The atom
${\mathtt{tptr}}(x,y,z)$ denotes a binary tree in which every node refers to
its two successors and to its parent and brother nodes (the parameter $y$ and
$z$ denote the brother and parent nodes, respectively). Leaves point to $()$.
${\small\begin{array}[]{llll}{\mathtt{als}}(x,y)&\Leftarrow&(x\mapsto(z)*{\mathtt{als}}(z,y))\curlywedge
y\not\approx z&\text{\tt\% (quasi-)acyclic list}\\\
{\mathtt{als}}(x,y)&\Leftarrow&x\mapsto(y)\\\
{\mathtt{tll}}(x,y)&\Leftarrow&x\mapsto(y,z)*{\mathtt{tree}}(z)&\text{\tt\%
binary trees with}\\\
{\mathtt{tll}}(x,y)&\Leftarrow&(x\mapsto(z,u)*{\mathtt{tll}}(z,y)*{\mathtt{tree}}(u))&\text{\tt\%
leftmost leaf $y$}\\\ &&\qquad\curlywedge(y\not\approx z)&\\\
{\mathtt{dll}}(x,y)&\Leftarrow&x\mapsto(y,z)*{\mathtt{dll}}(z,x)&\text{\tt\%
doubly linked lists}\\\ {\mathtt{dll}}(x,y)&\Leftarrow&x\mapsto()\\\
{\mathtt{tptr}}(x,y,z)&\Leftarrow&x\mapsto(u,v,y,z)*{\mathtt{tptr}}(u,v,x)&\text{\tt\%
binary trees with}\\\ &&\qquad*{\mathtt{tptr}}(v,u,x)&\text{\tt\% pointers to
brother}\\\ {\mathtt{tptr}}(x,y,z)&\Leftarrow&x\mapsto()&\text{\tt\% and
parent nodes}\\\ \end{array}}$
The following rules are not $\mathtt{P}$-rules (if all variables are of sort
$\mathtt{loc}$):
$\begin{array}[]{llll}p(x)&\Leftarrow&x\mapsto(z)&\text{Condition 2
violated}\\\ p(x)&\Leftarrow&{\mathtt{ls}}(x,z)*p(z)&\text{No points-to
atom}\\\ q(x,y)&\Leftarrow&x\mapsto(z)\curlywedge y\approx
z&\text{\text{Condition 2 violated}}\\\
q(x,y)&\Leftarrow&{\mathtt{ls}}(x,y)&\text{No points-to atom}\\\
{\mathtt{als}}(x,y)&\Leftarrow&(x\mapsto(z)*{\mathtt{als}}(z,y))\curlywedge
x\not\approx y&\text{Condition $1$ violated}\\\
{\mathtt{als}}(x,y)&\Leftarrow&x\mapsto(y)\curlywedge x\not\approx
y&\text{Condition $1$ violated}\\\ \end{array}$
###### Remark 2.11
As evidenced by the rules in Example 2.10, the tuple $()$ is frequently used
as a base case, to end a data structure. This departs from standard
conventions in SL in which a non-allocated constant $\mathtt{nil}$ is
frequently used instead. We avoid considering constants of sort $\mathtt{loc}$
in our framework because this would complicate definitions: one would have to
keep track of allocated and non allocated constants and/or to add syntactic
conditions on the formulas and rules to ensure that such constants are never
allocated.
Note that $\mathtt{P}$-rules are progressing and connected (in the sense of
[14]): every rule allocates exactly one location –the first parameter of the
predicate– and the first parameter of every predicate in the body of the rule
occurs in the right-hand side of the (necessarily unique) points-to atom of
the rule. They are not necessarily established (again in the sense of[14]) as
non-allocated existential variables are allowed provided they are not of sort
$\mathtt{loc}$.
###### Example 2.12
The following (non established, in the sense of [14]) rules, denoting list
segments with unallocated elements are $\mathtt{P}$-rules iff
$u\not\in\mathcal{V}_{\mathtt{loc}}$:
$\begin{array}[]{lllll}{\mathtt{ls}}(x,y)&\Leftarrow
x\mapsto(u,y)&\quad{\mathtt{ls}}(x,y)&\Leftarrow
x\mapsto(u,z)*{\mathtt{ls}}(z,y)\end{array}$
The heap of any model of ${\mathtt{ls}}(x,y)$ is of the form
$\\{\ell_{i}\mapsto(u_{i},\ell_{i+1})\mid i\in\\{1,\dots,n\\}\\}$, where
$u_{1},\dots,u_{n}$ denote arbitrary elements (of a sort distinct from
$\mathtt{loc}$).
$\mathtt{P}$-Rules containing no variable of a sort distinct from
$\mathtt{loc}$ are established. $\mathtt{P}$-Rules also differ from PCE rules
in that every predicate allocates exactly one of its parameters, namely the
first one (the other allocated locations are associated with existential
variables). In other words, there may be only one “entry point” to the
structure allocated by a predicate, namely its root. For instance the rule
$p(x,y)\Leftarrow x\mapsto(y)*q(y)$ (along with another rule for symbol $q$,
e.g., $q(y)\Leftarrow y\mapsto()$) is PCE but it is not a $\mathtt{P}$-rule,
whereas $p(x)\Leftarrow x\mapsto(y)*q(y)$ is a $\mathtt{P}$-rule. Such a
restriction makes the entailment problem easier to solve because it rules out
data structures that can be constructed in different orders (for instance
doubly linked lists with a reference to the end of the list).
We introduce some useful notations and measures on sets of $\mathtt{P}$-rules.
For every set of $\mathtt{P}$-rules $\mathfrak{R}$, we denote by
$\mathcal{P}(\mathfrak{R})$ the set of predicate symbols occurring in a rule
in $\mathfrak{R}$. We define:
$\mathit{ar}_{\mathit{max}}(\mathfrak{R})=\max\\{n\mid
p:{\mathtt{s}}_{1},\dots,{\mathtt{s}}_{n}\in\mathcal{P}(\mathfrak{R})\\}$
and
$\mathit{record}_{\mathit{max}}(\mathfrak{R})=\max\\{k\mid
p(x_{1},\dots,x_{n})\Leftarrow(x\mapsto(t_{1},\dots,t_{k})*\phi)\curlywedge\xi\in\mathfrak{R}\\}$
The numbers $\mathit{ar}_{\mathit{max}}(\mathfrak{R})$ and
$\mathit{record}_{\mathit{max}}(\mathfrak{R})$ respectively denote the maximum
arity of the predicate symbols in $\mathfrak{R}$ and the maximum number of
record fields in a points-to atom occurring in $\mathfrak{R}$. The width of
$\mathfrak{R}$ is defined as follows:
$\mathit{width}(\mathfrak{R})\overset{\text{\tiny\it
def}}{=}\max(\mathit{ar}_{\mathit{max}}(\mathfrak{R}),\mathit{record}_{\mathit{max}}(\mathfrak{R}))$.
We make two additional assumptions about the considered set of rules: we
assume that every predicate is productive (Assumption 2.2) and that no
parameter is useless (Assumption 2.16). More precisely, the set of productive
predicate symbols is inductively defined as follows: $p\in\mathcal{P}$ is
productive w.r.t. a set of inductive rules $\mathfrak{R}$ if $\mathfrak{R}$
contains a rule $p(\bm{x})\Leftarrow\lambda$ such that all the predicate
symbols occurring in $\lambda$ are productive. In particular, a rule with no
predicate in its right-hand side is always productive (base case). Productive
rules can easily be computed using a straightforward least fixpoint algorithm.
###### Example 2.13
Let $\mathfrak{R}=\\{p(x)\Leftarrow q(x),q(x)\Leftarrow p(x),r(x)\Leftarrow
x\mapsto(y)*p(y)\\}$. The predicates $p,q,r$ are not productive.
It is easy to check that every formula containing at least one non-productive
predicate symbol is unsatisfiable. Indeed, if a predicate symbol $p$ is non-
productive, then an atom $p(\bm{x})$ cannot be unfolded into a formula
containing no predicate symbol. This justifies the following:
For all sets of $\mathtt{P}$-rules $\mathfrak{R}$, we assume that all the
predicate symbols are productive w.r.t. $\mathfrak{R}$.
For all predicates $p:{\mathtt{s}}_{1},\dots,{\mathtt{s}}_{n}$, the set
$\mathit{out}_{\mathfrak{R}}(p)$ denotes the least set of indices $i$ in
$\\{1,\dots,n\\}$ such that ${\mathtt{s}}_{i}=\mathtt{loc}$ and there exists a
rule $p(x_{1},\dots,x_{n})\Leftarrow\lambda$ in $\mathfrak{R}$ such that
$\lambda$ contains either a points-to atom $x_{1}\mapsto(t_{1},\dots,t_{k})$
where $x_{i}\in\\{t_{1},\dots,t_{k}\\}$ or a predicate atom
$q(t_{1},\dots,t_{m})$ with $t_{j}=x_{i}$, for some
$j\in\mathit{out}_{\mathfrak{R}}(q)$. Intuitively,
$\mathit{out}_{\mathfrak{R}}(p)$ denote the set of “out-going” nodes of the
structures corresponding to $p$, i.e., the set of parameters corresponding to
locations that can be referred to but not necessarily allocated.
###### Example 2.14
Consider the following rules:
$p(x,y,z)\Leftarrow x\mapsto(x,y)$ | $p(x,y,z)\Leftarrow x\mapsto(x,u)*q(u,z,z)$ | $q(x,y,z)\Leftarrow x\mapsto(y)$
---|---|---
Then $\mathit{out}_{\mathfrak{R}}(p)=\\{1,2\\}$ and
$\mathit{out}_{\mathfrak{R}}(q)=\\{2\\}$.
###### Proposition 2.15
Let $\mathfrak{R}$ be a set of $\mathtt{P}$-rules. If
$(\mathfrak{s},\mathfrak{h})\models_{\mathfrak{R}}p(t_{1},\dots,t_{n})$ and
the index $i\neq 1$ is such that $i\not\in\mathit{out}_{\mathfrak{R}}(p)$
(i.e., $i$ is not an outgoing parameter of $p$) and
${\mathtt{s}}_{i}=\mathtt{loc}$, then the entailment
$(\mathfrak{s},\mathfrak{h})\models_{\mathfrak{R}}p(t_{1},\dots,t_{i-1},s,t_{i+1},t_{n})$
holds for all terms $s$.
###### Proof 2.16
By an induction on the satisfiability relation. By hypothesis
$p(t_{1},\dots,t_{n})\Leftarrow_{\mathfrak{R}}\gamma$, where $\gamma$ is of
the form $t_{1}\mapsto(t_{1}^{\prime},\ldots,t_{k}^{\prime})*\gamma^{\prime}$,
and there exists an associate $\mathfrak{s}^{\prime}$ of $\mathfrak{s}$ w.r.t.
$\mathcal{V}(\gamma)\setminus\mathcal{V}(\lambda)$ such that
$(\mathfrak{s}^{\prime},\mathfrak{h})\models_{\mathfrak{R}}\gamma$. By
hypothesis $i\neq 1$, and since $i\not\in\mathit{out}_{\mathfrak{R}}(p)$,
$t_{i}$ cannot occur in
$\left\\{t_{1}^{\prime},\ldots,t_{k}^{\prime}\right\\}$. This entails that
$p(t_{1},\dots,t_{i-1},s,t_{i+1},t_{n})\Leftarrow_{\mathfrak{R}}t_{1}\mapsto(t_{1}^{\prime},\ldots,t_{k}^{\prime})*\gamma^{\prime\prime}$.
If $\gamma^{\prime}$ contains a predicate $q(s_{1},\ldots,s_{m})$ and there
exists an index $j$ such that $s_{j}=t_{i}$, then we cannot have $j=1$ because
$t_{i}\notin\left\\{t_{1}^{\prime},\ldots,t_{k}^{\prime}\right\\}$ and the
rule under consideration is a $\mathtt{P}$-rule. Since
$i\not\in\mathit{out}_{\mathfrak{R}}(p)$ by hypothesis, $j$ cannot belong to
$\mathit{out}_{\mathfrak{R}}(q)$ and by induction, we deduce that
$(\mathfrak{s}^{\prime},\mathfrak{h})\models_{\mathfrak{R}}t_{1}\mapsto(t_{1}^{\prime},\ldots,t_{k}^{\prime})*\gamma^{\prime\prime}$,
hence the result.
Proposition 2.15 states that if
$i\not\in\mathit{out}_{\mathfrak{R}}(p)\cup\\{1\\}$ and
${\mathtt{s}}_{i}=\mathtt{loc}$, then the semantics of $p(t_{1},\dots,t_{n})$
does not depend on $t_{i}$, thus the $i$-th argument of $p$ is redundant and
can be removed. This justifies the following: For all sets of
$\mathtt{P}$-rules $\mathfrak{R}$ and for all predicate symbols
$p:\mathtt{loc},{\mathtt{s}}_{1},\dots,{\mathtt{s}}_{n}\in\mathcal{P}$, we
assume that $\mathit{out}_{\mathfrak{R}}(p)\supseteq\\{2\leq i\leq
n\mid{\mathtt{s}}_{i}=\mathtt{loc}\\}$.
###### Definition 2.17
For any formula $\lambda$, we write $x\rightarrow_{\lambda}y$ if
$x,y\in\mathcal{V}_{\mathtt{loc}}$ and $\lambda$ contains an atom
$p(t_{1},\dots,t_{n})$ (resp. $t_{1}\mapsto(t_{2},\dots,t_{n})$) such that
$t_{1}=x$ and $t_{i}=y$, for some $i\in\mathit{out}_{\mathfrak{R}}(p)$ (resp.
for some $i\in\\{2,\dots,n\\}$).
A structure $(\mathfrak{s},\mathfrak{h})$ is called a $\rightarrow$-compatible
model of a formula $\lambda$ iff
$(\mathfrak{s},\mathfrak{h})\models_{\mathfrak{R}}\lambda$ and for every
$x,y\in\mathcal{V}_{\mathtt{loc}}$,
$\mathfrak{s}(x)\rightarrow_{\mathfrak{h}}^{*}\mathfrak{s}(y)\implies
x\rightarrow_{\lambda}^{*}y$.
Intuively, $x\rightarrow_{\lambda}y$ states that the formula $\lambda$
allocates an edge from $x$ to $y$.
###### Definition 2.18
A sequent is an expression of the form
$\lambda\vdash_{\mathfrak{R}}^{V}\gamma$, where $\lambda,\gamma$ are symbolic
heaps, $V$ is a multiset of variables of sort $\mathtt{loc}$ and
$\mathfrak{R}$ is a finite set of inductive rules. If $V=\emptyset$ then the
sequent is written $\lambda\vdash_{\mathfrak{R}}\gamma$. A sequent is
equality-free if $\lambda$ and $\gamma$ contain no atoms of the form $u\approx
v$. A counter-model of a sequent $\lambda\vdash_{\mathfrak{R}}^{V}\gamma$ is a
structure $(\mathfrak{s},\mathfrak{h})$ such that:
* •
$(\mathfrak{s},\mathfrak{h})\models_{\mathfrak{R}}\lambda$ and
$(\mathfrak{s},\mathfrak{h})\not\models_{\mathfrak{R}}\gamma$,
* •
$\forall x\in V,\,\mathfrak{s}(x)\not\in\mathit{dom}(\mathfrak{h})$,
* •
$\mathfrak{s}$ is injective on the multiset $V$.
A sequent is valid iff it has no counter-model.
## 3 Lower Bounds
We establish various lower bounds for the validity problems for sequents
$\lambda\vdash_{\mathfrak{R}}\gamma$, where $\mathfrak{R}$ satisfies some
additional conditions. These lower bounds will motivate the additional
restrictions that are imposed to devise a polynomial-time proof procedure.
Checking the validity of sequents $\lambda\vdash_{\mathfrak{R}}\gamma$ where
$\mathfrak{R}$ is a set of $\mathtt{P}$-rules is actually undecidable in
general. This result can established using an argument similar to the one used
in [11] to prove the undecidability of PCE entailments modulo theories; it is
not given here for the sake of conciseness, and because the goal of this paper
is to investigate tractable cases. The undecidability proof relies on the
existence of variables of a sort distinct from $\mathtt{loc}$. If such
variables are forbidden, then the rules are PCE hence entailment is decidable
[14], but we still get an ExpTime lower bound:
###### Proposition 3.1
Checking the validity of sequents $\lambda\vdash_{\mathfrak{R}}\gamma$ is
ExpTime-hard, even if $\mathfrak{R}$ is a set of $\mathtt{P}$-rules and all
the variables occurring in $\lambda\vdash_{\mathfrak{R}}\gamma$ are of sort
$\mathtt{loc}$.
###### Proof 3.2
The proof is by a straightforward reduction from the inclusion problem for
languages accepted by tree automata (see [6]). Indeed, a tree automaton
$(Q,V,\\{q_{0}\\},R)$ can be straightforwardly encoded as a set of
$\mathtt{P}$-rules, where each rule $q\rightarrow f(q_{1},\dots,q_{n})$ in $R$
is encoded by an inductive rule of the form ${q(x)\Leftarrow
x\mapsto(f,x_{1},\dots,x_{n})}*\scaleobj{1.5}{*}_{i=1}^{n}q_{i}(x_{i})$. Each
function symbol $f$ is considered as a constant of a sort
${\mathtt{s}}\not=\mathtt{loc}$, and a term $f(t_{1},\dots,t_{n})$ is
represented as a heap
$\mathfrak{h}_{1}\uplus\dots\uplus\mathfrak{h}_{n}\uplus\\{(\ell_{0},\ell_{1},\dots,\ell_{n})\\}$,
where $\ell_{0},\dots,\ell_{n}$ are pairwise distinct locations and
$\mathfrak{h}_{1},\dots,\mathfrak{h}_{n}$ are disjoint representations of
$t_{i}$ with ${\ell_{0}\not\in\mathit{dom}(\mathfrak{h}_{i})}$, for
$i=1,\dots,n$. It is straightforward to verify that the language accepted by
$(Q,V,\\{q_{0}\\},R)$ is included in that of
$(Q^{\prime},V,\\{q_{0}^{\prime}\\},R^{\prime})$ iff the sequent
$q_{0}(x)\vdash_{\mathfrak{R}}q_{0}^{\prime}(x)$ is valid.
Since the inclusion problem is polynomial for top-down deterministic tree
automata [6], it is natural to further restrict the considered rules to make
them deterministic, in the following sense:
###### Definition 3.3
A set of $\mathtt{P}$-rules $\mathfrak{R}$ is deterministic if for all pairs
of distinct rules of the form
$p(\bm{x}_{i})\Leftarrow(y_{i}\mapsto\bm{t}_{i}*\phi_{i})\curlywedge\xi_{i}$
(where $i=1,2$) occurring in $\mathfrak{R}$, the formula
$\bm{x}_{1}\approx\bm{x}_{2}\wedge\bm{t}_{1}\approx\bm{t}_{2}\wedge\xi_{1}\wedge\xi_{2}$
is unsatisfiable (we assume by renaming that the rules share no variable).
For instance the rules associated with the predicate ${\mathtt{ls}}$ in the
introduction are not deterministic, whereas the rules associated with
${\mathtt{tree}}$ are deterministic, as well as all those in Example 2.10. For
the predicate ${\mathtt{ls}}$, the formula $x\approx x^{\prime}\wedge y\approx
z$ is satisfiable, whereas for the predicate ${\mathtt{tree}}$, the formula
$x\approx x^{\prime}\wedge()\approx(y,z)$ is unsatisfiable (in both cases the
variable $x$ is renamed by $x^{\prime}$ in the second rule).
The following proposition shows that the restriction to deterministic sets of
$\mathtt{P}$-rules is still not sufficient to obtain a tractable validity
problem:
###### Proposition 3.4
Checking the validity of sequents $\lambda\vdash_{\mathfrak{R}}\gamma$ is
PSpace-hard, even if $\mathfrak{R}$ is a deterministic set of
$\mathtt{P}$-rules and all variables in $\lambda\vdash_{\mathfrak{R}}\gamma$
are of sort $\mathtt{loc}$.
###### Proof 3.5
Let “$w\in E$” be any problem in PSpace. By definition, there exists a Turing
machine $M=(Q,\Sigma,B,\Gamma,\delta,q_{0},F)$ accepting exactly the words in
$E$ and a polynomial $R$ such that $M$ runs in space $R(n)$ on all words
$w\in\Sigma^{n}$. The set $Q$ denotes the set of states of $M$, $\Sigma$ is
the input alphabet, $B$ is the blank sumbol, $\Gamma$ is the tape alphabet,
$\delta$ is the transition function, $q_{0}$ is the initial state and $F$ is
the set of final states. We shall reduce the problem “$w\in E$” to the
entailment problem, for a sequent fulfilling the conditions above. Consider a
word $w$ of length $n$, and let $N=R(n)$. Assume that $\mathcal{C}$ contains
all the elements in $\Gamma$. We consider $\mathit{card}(Q)\cdot N$ predicates
$q^{i}$ of arity $N+3$, for all $q\in Q$ and $i\in\\{1,\dots,N\\}$, associated
with the following rules:
$\displaystyle q^{i}(x,y_{1},\dots,y_{N},u,v)$
$\displaystyle\Leftarrow\begin{multlined}x\mapsto(x^{\prime},u,v,a)*p^{i+\mu}(x^{\prime},y_{1},\dots,y_{i-1},b,y_{i+1},\dots,y_{N},y_{i},a)\\\
{\text{if $q\not\in F$ and $\delta$ contains a rule
$(q,a)\rightarrow(p,b,\mu)$}}\\\ {\text{ with $i+\mu\in\\{1,\dots,N\\}$}}\\\
\end{multlined}x\mapsto(x^{\prime},u,v,a)*p^{i+\mu}(x^{\prime},y_{1},\dots,y_{i-1},b,y_{i+1},\dots,y_{N},y_{i},a)\\\
{\text{if $q\not\in F$ and $\delta$ contains a rule
$(q,a)\rightarrow(p,b,\mu)$}}\\\ {\text{ with $i+\mu\in\\{1,\dots,N\\}$}}\\\ $
$\displaystyle q^{i}(x,y_{1},\dots,y_{N},u,v)$ $\displaystyle\Leftarrow
x\mapsto(x,u,v,B)\text{,\ if $q\in F$.}$
Intuitively, $q$ is the state of the machine, the arguments
$y_{1},\dots,y_{N}$ denote the tape (that is of length $N$ by hypothesis) and
$i$ denotes the position of the head on the tape. The constants $a,b$ denote
the symbols read and written on the tape, respectively, $p$ is the final state
of the transition rule and the integer $\mu$ denotes the move, i.e., an
element of $\\{-1,0,+1\\}$, so that $i+\mu$ is the final position of the head
on the tape. Note that at this point the inductive rule does not test whether
the symbol $a$ is indeed identical to the symbol at position $i$, namely
$y_{i}$. Instead, it merely stores both $y_{i}$ and $a$ within the next tuple
of the heap, by passing them as parameters to $p^{i+\mu}$. The arguments $u$
and $v$ are used to encode respectively the symbol read on the tape at the
previous state and the symbol that was expected. By definition of the above
rules, it is clear that $q^{i}(x,y_{1},\dots,y_{N},B,B)$ holds if the heap is
a list of tuples $(x_{j},u_{j},v_{j},a_{j})$, for $j=1,\ldots,k$, linked on
the first argument (the last element loops on itself). The heap encodes a
“candidate run” of length $k$ of $M$, i.e., a run for which one does not
check, when applying a transition $(p,a)\rightarrow(q,b,\mu)$, that the symbol
read on the tape is identical to the expected symbol $a$. The symbols
$u_{i},v_{i},a_{i}$ stored at each node are precisely the symbols that are
read ($u_{i}$) and expected ($v_{i}$) at the previous step, respectively,
along with the symbol $(a_{i}$) that is expected at the current step (this
last symbol is added to ensure that the rules are deterministic). Note that
for $i=1$ there is no previous step and for $i=k$ no symbol is read since the
state is final; thus by convention, $a_{k}$ is set to $B$ (see the last rule
of $q^{i}$). Furthermore, $u_{1},v_{1}$ will also be set to $B$ by invoking
the initial state predicate with $B$ as last and before last argument (see the
definition of the sequent below). To check that the list corresponds to an
actual run of $M$, it thus suffices to check that $u_{i}=v_{i}$ holds for all
$i=1,\dots,k$.
The right-hand side of the sequent will allocate all structures not satisfying
this condition. To this purpose, we associate with each state $r\in Q$ two
predicate symbols $r_{0}$ and $r_{1}$ defined by the following rules (where
$i,j\in\\{0,1\\}$):
$\displaystyle r_{i}(x)$
$\displaystyle\Leftarrow\begin{multlined}x\mapsto(x^{\prime},a,b,c)*r_{j}(x^{\prime})\\\
{\text{for all $r\not\in F$, $a,b,c\in\Gamma$, where $j=1$ iff either $i=1$ or
$a\not=b$;}}\end{multlined}x\mapsto(x^{\prime},a,b,c)*r_{j}(x^{\prime})\\\
{\text{for all $r\not\in F$, $a,b,c\in\Gamma$, where $j=1$ iff either $i=1$ or
$a\not=b$;}}$ $\displaystyle r_{i}(x)$ $\displaystyle\Leftarrow
x\mapsto(x,a,b,B)\text{\quad if $a,b\in\Gamma$, $r\in F$ and either $i=1$ or
$a\not=b$.}$
Intuitively, the index $i$ in predicate $r_{i}$ is equal to $1$ iff a faulty
location has been encountered (i.e., a tuple $(x^{\prime},a,b)$ with
$a\not=b$).
Note that the number of rules is polynomial w.r.t. $N$, since the machine $M$
is fixed. Also, the obtained set of rules is deterministic, because $M$ is
deterministic and the expected symbol is referred to by the location allocated
by each predicate symbol $q^{i}$, thus the tuples $(x^{\prime},u,v,a)$
corresponding to distinct rules associated with the same symbol $q^{i}$ cannot
be unifiable.
Let $w_{1}.\dots,w_{N}=w.B^{N-n}$ be the initial tape (where $w$ is completed
by blank symbols $B$ to obtain a word of length $N$). It is clear that the
sequent $q_{0}^{1}(x,w_{1},\dots,w_{N},B,B)\vdash_{\mathfrak{R}}r_{0}(x)$ is
valid iff all the “candidate runs” of $M$ fulfill the conditions of the right-
hand side, i.e., falsify at least one equality between read and expected
symbols. Thus
$q_{0}^{1}(x,w_{1},\dots,w_{N},B,B)\vdash_{\mathfrak{R}}r_{0}(x)$ is valid iff
$M$ does not accept $w$.
In view of this result, it is natural to investigate the complexity of the
entailment problem when the maximal arity of the predicates is bounded.
However, this is still insufficient to get a tractable problem, as the
following lemma shows.
###### Lemma 3.6
Checking the validity of sequents $\lambda\vdash_{\mathfrak{R}}\gamma$ is co-
Np-hard, even if $\mathit{width}(\mathfrak{R})\leq 4$ (i.e., if the symbols
and tuples are of arity at most $4$) and $\mathfrak{R}$ is a deterministic set
of $\mathtt{P}$-rules.
###### Proof 3.7
The proof is by a reduction from the complement of the $3$-coloring problem,
that is well-known to be NP-complete. Let $G=(V,E)$ be a graph, where
$V=\\{v_{1},\dots,v_{n}\\}$ is a finite set of vertices and $E$ is a set of
undirected edges, i.e., a set of unordered pairs of vertices. Let
$\mathtt{Colors}=\\{a,b,c\\}$ be a set of colors, with
$\mathit{card}(\mathtt{Colors})=3$. We recall that a solution of the
$3$-coloring problem is a function $f:V\rightarrow\mathtt{Colors}$ such that
$(x,y)\in E\implies f(x)\not=f(y)$. We assume, w.l.o.g., that all vertices
occur in at least one edge. We consider two distinct sorts $\mathtt{loc}$ and
${\mathtt{s}}$. We assume, w.l.o.g., that
$V\cup\mathtt{Colors}\subseteq\mathcal{V}_{{\mathtt{s}}}$ (i.e., $a,b,c$, as
well as the set of vertices in $G$, are variables) and
$V\cap\mathtt{Colors}=\emptyset$.
Let $E=\\{(x_{i},y_{i})\mid i=1,\dots,m\\}$ (where the edges are ordered
arbitrarily) and let $u_{1},\dots,u_{m+1}$ be pairwise distinct variables of
sort $\mathtt{loc}$. Let $\phi$ be the formula: $\scaleobj{1.5}{*}_{i=1}^{m}\
u_{i}\mapsto(x_{i},y_{i},u_{i+1})*u_{m+1}\mapsto()$. Let $p$ and $q$ be
predicate symbols associated with the following rules:
$\begin{array}[]{lll}p(u,a,b,c)&\Leftarrow&u\mapsto(v,v,u^{\prime})*q(u^{\prime})\\\
p(u,a,b,c)&\Leftarrow&u\mapsto(v_{1},v_{2},u^{\prime})*v_{1}\not\approx
v_{2}*v_{1}\not\approx a*v_{1}\not\approx b*v_{1}\not\approx
c*q(u^{\prime})\\\
p(u,a,b,c)&\Leftarrow&u\mapsto(d,v_{2},u^{\prime})*v_{2}\not\approx
a*v_{2}\not\approx b*v_{2}\not\approx c*q(u^{\prime})\\\ &&\text{\quad for all
$d\in\\{a,b,c\\}$}\\\
p(u,a,b,c)&\Leftarrow&u\mapsto(d_{1},d_{2},u^{\prime})*p(u^{\prime},a,b,c)\\\
&&\text{\quad for all $d_{1},d_{2}\in\\{a,b,c\\}$ where $d_{1}\not=d_{2}$}\\\
q(u)&\Leftarrow&u\mapsto(v_{1},v_{2},u^{\prime})*q(u^{\prime})\\\
q(u)&\Leftarrow&u\mapsto()\end{array}$
Intuitively, any model $(\mathfrak{s},\mathfrak{h})$ of $\phi$ encodes a
candidate solution of the $3$-coloring problem, where each variable
$z\in\\{x_{i}\mid i=1,\dots,m\\}\cup\\{y_{i}\mid i=1,\dots,m\\}$ is mapped to
an element $\mathfrak{s}(z)$ in $\mathfrak{U}_{{\mathtt{s}}}$. The heap
$\mathfrak{h}$ is a list of tuples linked on the last element and containing a
tuple
$(\mathfrak{s}(u_{i}),\mathfrak{s}(x_{i}),\mathfrak{s}(y_{i}),\mathfrak{s}(u_{i+1}))$
for all $(x_{i},y_{i})\in E$. To check that this candidate solution indeed
fulfills the required properties, one has to verify that all the pairs
$(\mathfrak{s}(x_{i}),\mathfrak{s}(y_{i}))$ are composed of distinct elements
in $\\{a,b,c\\}$.
By definition of the rules for predicate $p$, $(\mathfrak{s},\mathfrak{h})$ is
a model of $p(u_{1},a,b,c)$ iff the list contains a pair
$(\mathfrak{s}(x),\mathfrak{s}(y))$ such that one of the following holds:
* •
$\mathfrak{s}(x)=\mathfrak{s}(y)$ (first rule of $p$),
* •
$\mathfrak{s}(x)\not=\mathfrak{s}(y)$ and
$\mathfrak{s}(x)\not\in\\{\mathfrak{s}(a),\mathfrak{s}(b),\mathfrak{s}(c)\\}$
(second rule of $p$),
* •
$\mathfrak{s}(x)\in\\{\mathfrak{s}(a),\mathfrak{s}(b),\mathfrak{s}(c)\\}$ and
$\mathfrak{s}(y)\not\in\\{\mathfrak{s}(a),\mathfrak{s}(b),\mathfrak{s}(c)\\}$
(third rule of $p$).
After the cell corresponding to this faulty pair is allocated, $q$ is invoked
to allocate the remaining part of the list. Thus $p(u_{1},a,b,c)$ holds iff
the model does not encode a solution of the $3$-coloring problem, either
because $\mathfrak{s}(x_{i})=\mathfrak{s}(y_{i})$ for some $(x_{i},y_{i})\in
E$ or because one of the variables is mapped to an element distinct from
$a,b,c$ – note that by the above assumption, each of these variables occurs in
the list. Consequently $\phi\vdash_{\mathfrak{R}}p(u_{1},a,b,c)$ admits a
counter-model if there exists a model of $\phi$ that does not satisfy
$p(u_{1},a,b,c)$, i.e., iff the $3$-coloring problem admits a solution.
The results above motivate the following definition, that strengthens the
notion of a deterministic set of rules.
###### Definition 3.8
A set of $\mathtt{P}$-rules $\mathfrak{R}$ is $\mathtt{loc}$-deterministic if
it is deterministic and all the disequations occurring in the rules in
$\mathfrak{R}$ are of the form $x\not\approx y$ with
$x,y\in\mathcal{V}_{\mathtt{loc}}$.
The intuition behind $\mathtt{loc}$-deterministic rules is that, to get an
efficient proof procedure, we have to restrict the amount of equational
reasoning needed to establish the validity of the sequents. Disequations
between locations are relatively easy to handle because (by definition of
$\mathtt{P}$-rules) all existential variables of sort $\mathtt{loc}$ must be
pairwise distinct (as they are allocated in distinct atoms). However, dealing
with disequations between data is much more difficult, as evidenced by the
proof of Lemma 3.6. Thus we restrict such disequations to those occurring in
the initial sequent.
The rules associated with ${\mathtt{als}}$, ${\mathtt{tree}}$,
${\mathtt{tll}}$, ${\mathtt{tptr}}$ or ${\mathtt{dll}}$ in the Introduction
and in Example 2.10 are $\mathtt{loc}$-deterministic. In contrast, the
following rules are deterministic, but not $\mathtt{loc}$-deterministic (where
$u,v$ denote variables of some sort distinct from $\mathtt{loc}$):
$\begin{array}[]{llllll}p(x,u)&\Leftarrow&x\mapsto(v)\curlywedge v\not\approx
u&\quad p(x,u)&\Leftarrow&x\mapsto(u)\end{array}$
Rules that are $\mathtt{loc}$-deterministic are well-suited to model
constructor-based data structures used in standard programming languages; for
instance, lists could be represented as follows (where $\mathtt{cons}$ is a
constant symbol denoting a constructor and $y$ is a variable of some sort
distinct from $\mathtt{loc}$, denoting data stored in the list):
$\begin{array}[]{llllll}{\mathtt{ls}}(x)\Leftarrow&x\mapsto(\mathtt{cons},y,z)*{\mathtt{ls}}(z)&\quad{\mathtt{ls}}(x)\Leftarrow&x\mapsto()\end{array}$
We end this section by establishing a key property of deterministic set of
rules, namely the fact that every spatial formula $\phi$ is precise, in the
sense of [4]: it is fulfilled on at most one subheap within a given structure.
###### Lemma 3.9
Let $\mathfrak{R}$ be a deterministic set of rules. For every spatial formula
$\phi$, for every store $\mathfrak{s}$ and for every heap $\mathfrak{h}$ there
exists at most one heap $\mathfrak{h}^{\prime}$ such that
$\mathfrak{h}^{\prime}\subseteq\mathfrak{h}$ and
$(\mathfrak{s},\mathfrak{h}^{\prime})\models_{\mathfrak{R}}\phi$.
###### Proof 3.10
The proof is by induction on the satisfiability relation
$\models_{\mathfrak{R}}$. Note that by hypothesis $\phi$ is a spatial formula,
hence contains no occurrences of $\approx$, $\not\approx$, $\curlywedge$ or
$\wedge$. Assume that there exist two heaps
$\mathfrak{h}_{1}^{\prime},\mathfrak{h}_{2}^{\prime}$ such that
$\mathfrak{h}_{i}^{\prime}\subseteq\mathfrak{h}$ and
$(\mathfrak{s},\mathfrak{h}_{i}^{\prime})\models_{\mathfrak{R}}\phi$ (for
$i=1,2$). We show that $\mathfrak{h}_{1}^{\prime}=\mathfrak{h}_{2}^{\prime}$.
* •
If $\phi=\mathit{emp}$ then necessarily $\mathfrak{h}_{i}=\emptyset$ for
$i=1,2$ thus $\mathfrak{h}_{1}^{\prime}=\mathfrak{h}_{2}^{\prime}$.
* •
If $\phi=y_{0}\mapsto(y_{1},\dots,y_{n})$ then by Definition 2.4 we have
$\mathfrak{h}_{i}^{\prime}=\\{(\mathfrak{s}(y_{0}),\dots,\mathfrak{s}(y_{n}))\\}$
for $i=1,2$ thus $\mathfrak{h}_{1}^{\prime}=\mathfrak{h}_{2}^{\prime}$.
* •
If $\phi=\phi_{1}*\phi_{2}$ then for all $i=1,2$ there exist two disjoint
heaps $\mathfrak{h}_{i}^{j}$ (for $j=1,2$) such that
$\mathfrak{h}_{i}^{\prime}=\mathfrak{h}_{i}^{1}\uplus\mathfrak{h}_{i}^{2}$ for
$i=1,2$ and
$(\mathfrak{s},\mathfrak{h}_{i}^{j})\models_{\mathfrak{R}}\phi_{j}$, for
$i,j\in\\{1,2\\}$. Since
$\mathfrak{h}_{i}^{j}\subseteq\mathfrak{h}_{i}^{\prime}\subseteq\mathfrak{h}$
we get by the induction hypothesis $\mathfrak{h}_{1}^{j}=\mathfrak{h}_{2}^{j}$
for $j=1,2$. Therefore
$\mathfrak{h}_{1}^{1}\uplus\mathfrak{h}_{1}^{2}=\mathfrak{h}_{2}^{1}\uplus\mathfrak{h}_{2}^{2}$,
i.e., $\mathfrak{h}_{1}^{\prime}=\mathfrak{h}_{2}^{\prime}$.
* •
If $\phi$ is a predicate atom of root $x$, then for $i=1,2$ we have
$\phi\Leftarrow_{\mathfrak{R}}\lambda_{i}$, and there exists an associate
$\mathfrak{s}_{i}$ of $\mathfrak{s}$ w.r.t.
$\mathcal{V}(\psi_{i})\setminus\mathcal{V}(\phi)$ such that
$(\mathfrak{s}_{i},\mathfrak{h}_{i}^{\prime})\models_{\mathfrak{R}}\lambda_{i}$.
Since $\mathfrak{R}$ is a set of $\mathtt{P}$-rules, $\lambda_{i}$ is of the
form $(x_{i}\mapsto\bm{y}_{i}*\phi_{i})\curlywedge\xi_{i}$ and there exist
disjoint heaps $\mathfrak{h}_{i}^{j}$ (for $j=1,2$) such that the following
conditions are satisfied: (i)
$\mathfrak{h}_{i}^{\prime}=\mathfrak{h}_{i}^{1}\uplus\mathfrak{h}_{i}^{2}$;
(ii)
$(\mathfrak{s}_{i},\mathfrak{h}_{i}^{1})\models_{\mathfrak{R}}x_{i}\mapsto(\bm{y}_{i})$;
(iii) $(\mathfrak{s}_{i},\mathfrak{h}_{i}^{2})\models_{\mathfrak{R}}\phi_{i}$;
(iv) and $\mathfrak{s}_{i}\models\xi_{i}$. Furthermore, $x_{i}$ must be the
root of $\phi$, thus $x_{1}=x_{2}=x$. For $i=1,2$ we have
$\mathfrak{h}_{i}^{1}=\\{(\mathfrak{s}(x_{i}),\mathfrak{s}_{i}(\bm{y}_{i}))\\}$,
and since $\mathfrak{h}_{i}^{1}\subseteq\mathfrak{h}$, necessarily
$\mathfrak{s}_{1}(\bm{y}_{1})=\mathfrak{s}_{2}(\bm{y}_{2})$ and
$\mathfrak{h}_{1}^{1}=\mathfrak{h}_{2}^{1}$. The heap $\mathfrak{h}_{i}^{2}$
is the restriction of $\mathfrak{h}_{i}^{\prime}$ to the locations distinct
from $\mathfrak{s}(x)$. We distinguish two cases.
* –
Assume that the inductive rules applied on $\phi$ to respectively derive
$\lambda_{1}$ and $\lambda_{2}$ are different. We may assume by
$\alpha$-renaming that
$(\mathcal{V}(\lambda_{1})\setminus\mathcal{V}(\phi))\cap(\mathcal{V}(\lambda_{2})\setminus\mathcal{V}(\phi))=\emptyset$,
which entails that there exists a store $\mathfrak{s}^{\prime}$ that coincides
with $\mathfrak{s}_{i}$ on $\mathcal{V}(\lambda_{i})$ (since
$\mathfrak{s}_{1}$ and $\mathfrak{s}_{2}$ coincide on the variables in
$\mathcal{V}(\phi)$). Then we have
$\mathfrak{s}^{\prime}\models\bm{y}_{1}\approx\bm{y}_{2}$ (since
$\mathfrak{s}_{1}(\bm{y}_{1})=\mathfrak{s}_{2}(\bm{y}_{2})$) and
$\mathfrak{s}^{\prime}\models\xi_{i}$ (since
$\mathfrak{s}_{i}\models\xi_{i}$), which entails that the formula
$\bm{y}_{1}\approx\bm{y}_{2}\wedge\xi_{1}\wedge\xi_{2}$ is satisfiable,
contradicting the fact that $\mathfrak{R}$ is deterministic.
* –
Assume that the same rule is used to derive both $\lambda_{1}$ and
$\lambda_{2}$. We may assume in this case (again by $\alpha$-renaming) that
the vector of variables occurring in $\lambda_{1}$ and $\lambda_{2}$ are the
same, so that $\bm{y}_{1}=\bm{y}_{2}$ and $\phi_{1}=\phi_{2}$. Since
$\mathfrak{R}$ is a set of $\mathtt{P}$-rules, all variables $z$ in
$\mathcal{V}(\phi_{i})\setminus\mathcal{V}(\phi)$ occur in $\bm{y}_{i}$. As
$\mathfrak{s}_{1}(\bm{y}_{1})=\mathfrak{s}_{2}(\bm{y}_{2})$, this entails that
$\mathfrak{s}_{1}(z)=\mathfrak{s}_{2}(z)$ holds for all such variables, thus
$\mathfrak{s}_{1}=\mathfrak{s}_{2}$. Consequently,
$(\mathfrak{s}_{1},\mathfrak{h}_{i}^{2})\models_{\mathfrak{R}}\phi_{1}$, for
all $i=1,2$ with
$\mathfrak{h}_{i}^{2}\subseteq\mathfrak{h}_{i}^{\prime}\subseteq\mathfrak{h}$.
By the induction hypothesis this entails that
$\mathfrak{h}_{1}^{2}=\mathfrak{h}_{2}^{2}$, thus
$\mathfrak{h}_{1}^{\prime}=\mathfrak{h}_{2}^{\prime}$.
###### Example 3.11
Lemma 3.9 does not hold if the rules are not deterministic. For instance, the
formula ${\mathtt{ls}}(x,y)$ (with the rules given in the introduction) has
two models $(\mathfrak{s},\mathfrak{h})$ and
$(\mathfrak{s},\mathfrak{h}^{\prime})$ where $\mathfrak{h}^{\prime}$ is a
strict subheap of $\mathfrak{h}$: $\mathfrak{s}(x)=\ell_{1}$,
$\mathfrak{s}(y)=\ell_{2}$,
$\mathfrak{h}=\\{\ell_{1}\mapsto(\ell_{2}),\ell_{2}\mapsto(\ell_{2})\\}$ and
$\mathfrak{h}^{\prime}=\\{\ell_{1}\mapsto(\ell_{2})\\}$. Intuitively, the
formula ${\mathtt{ls}}(y,y)$ (which is useful to derive ${\mathtt{ls}}(x,y)$)
can be derived by any of the two rules of ${\mathtt{ls}}$, yielding two
different models. In contrast ${\mathtt{als}}(x,y)$ (with the rules of Example
2.10) has only one model with the store $\mathfrak{s}$ and a heap included in
$\mathfrak{h}$, namely $(\mathfrak{s},\mathfrak{h}^{\prime})$.
## 4 Proof Procedure
From now on, we consider a fixed $\mathtt{loc}$-deterministic set of
$\mathtt{P}$-rules $\mathfrak{R}$, satisfying Assumptions 2.2 and 2.16. For
technical convenience, we also assume that $\mathfrak{R}$ is nonempty and that
every constant in $\mathcal{C}$ occurs in a rule in $\mathfrak{R}$.
### 4.1 Some Basic Properties of $\mathtt{P}$-Rules
We begin by introducing some definitions and deriving straightforward
consequences of the definition of $\mathtt{P}$-rules. We shall denote by
$\mathit{alloc}(\lambda)$ the multiset of variables allocated by a formula
$\lambda$:
###### Definition 4.1
For every formula $\lambda$, we denote by $\mathit{alloc}(\lambda)$ the
multiset of variables $x$ such that $\lambda$ contains a spatial atom with
root $x$.
Lemma 4.2 states that the variables in $\mathit{alloc}(\lambda)$ are
necessarily allocated in every model of $\lambda$, which entails (Corollary
4.4) that they must be associated with pairwise distinct locations. Moreover,
a formula distinct from $\mathit{emp}$ has at least one root, hence allocates
at least one variable (Corollary 4.6).
###### Lemma 4.2
Let $\lambda$ be a formula. If
$(\mathfrak{s},\mathfrak{h})\models_{\mathfrak{R}}\lambda$ and
$x\in\mathit{alloc}(\lambda)$ then
$\mathfrak{s}(x)\in\mathit{dom}(\mathfrak{h})$.
###### Proof 4.3
By hypothesis, $\lambda$ is of the form $(\alpha*\phi)\curlywedge\xi$ where
$\alpha$ is a spatial atom with root $x$. Thus
$(\mathfrak{s},\mathfrak{h})\models\alpha*\phi$ and there exist disjoint heaps
$\mathfrak{h}_{1},\mathfrak{h}_{2}$ such that
$(\mathfrak{s},\mathfrak{h}_{1})\models_{\mathfrak{R}}\alpha$,
$(\mathfrak{s},\mathfrak{h}_{2})\models_{\mathfrak{R}}\phi$, and
$\mathfrak{h}=\mathfrak{h}_{1}\uplus\mathfrak{h}_{2}$. Since the root of
$\alpha$ is $x$, $\alpha$ is either of the form $x\mapsto\bm{y}$ or of the
form $p(x,\bm{y})$ where $p\in\mathcal{P}$. In the former case, it is clear
that
$\mathfrak{s}(x)\in\mathit{dom}(\mathfrak{h}_{1})\subseteq\mathit{dom}(\mathfrak{h})$
since $(\mathfrak{s},\mathfrak{h}_{1})\models_{\mathfrak{R}}\alpha$. In the
latter case, we have $\alpha\Leftarrow_{\mathfrak{R}}\lambda^{\prime}$ and
$(\mathfrak{s}^{\prime},\mathfrak{h})\models_{\mathfrak{R}}\lambda^{\prime}$,
where $\mathfrak{s}^{\prime}$ is an associate of $\mathfrak{s}$ w.r.t.
$\mathcal{V}(\lambda^{\prime})\setminus\mathcal{V}(\alpha)$. Since
$\mathfrak{R}$ is a set of $\mathtt{P}$-rules, necessarily $\lambda^{\prime}$
contains a points-to atom of the form $x\mapsto\bm{z}$, which entails that
$\mathfrak{s}^{\prime}(x)\in\mathit{dom}(\mathfrak{h}_{1})$, hence
$\mathfrak{s}(x)\in\mathit{dom}(\mathfrak{h})$.
###### Corollary 4.4
Let $\lambda$ be a formula and let $(\mathfrak{s},\mathfrak{h})$ be an
$\mathfrak{R}$-model of $\lambda$. If
$\\{x,y\\}\subseteq_{m}\mathit{alloc}(\lambda)$ then
$\mathfrak{s}(x)\not=\mathfrak{s}(y)$. In particular, if
$\mathit{alloc}(\lambda)$ contains two occurrences of the same variable $x$
then $\lambda$ is unsatisfiable.
###### Proof 4.5
By definition, $\lambda$ is of the form
$(\alpha_{1}*\alpha_{2}*\phi)\curlywedge\xi$, where $\alpha_{1}$ and
$\alpha_{2}$ are spatial atoms of roots $x$ and $y$, respectively, with
$\mathit{alloc}(\alpha_{1})=\\{x\\}$ and $\mathit{alloc}(\alpha_{2})=\\{y\\}$.
If $\lambda$ admits a model $(\mathfrak{s},\mathfrak{h})$, then there exists
disjoint heaps $\mathfrak{h}_{1},\mathfrak{h}_{2},\mathfrak{h}^{\prime}$ such
that
$\mathfrak{h}=\mathfrak{h}_{1}\uplus\mathfrak{h}_{2}\uplus\mathfrak{h}^{\prime}$,
$(\mathfrak{s},\mathfrak{h}_{i})\models_{\mathfrak{R}}\alpha_{i}$ (for
$i=1,2$) and $(\mathfrak{s},\mathfrak{h}^{\prime})\models_{\mathfrak{R}}\phi$.
By Lemma 4.2 we have $\mathfrak{s}(x)\in\mathit{dom}(\mathfrak{h}_{1})$ and
$\mathfrak{s}(y)\in\mathit{dom}(\mathfrak{h}_{2})$, thus
$\mathfrak{s}(x)\not=\mathfrak{s}(y)$ since $\mathfrak{h}_{1}$ and
$\mathfrak{h}_{2}$ are disjoint.
###### Corollary 4.6
Let $\phi$ be a spatial formula and let $(\mathfrak{s},\mathfrak{h})$ be an
$\mathfrak{R}$-model of $\phi$. If $\phi\not=\mathit{emp}$ then
$\mathfrak{h}\not=\emptyset$.
###### Proof 4.7
Since $\phi\not=\mathit{emp}$, necessarily $\phi$ contains at least one atom
$\alpha$, thus $\mathit{root}(\alpha)\in\mathit{alloc}(\phi)$. Then the result
follows immediately from Lemma 4.2.
Corollary 4.4 motivates the following definition, which provides a simple
syntactic criterion to identify some formulas that cannot be satisfiable, due
to the fact that the same variable is allocated twice.
###### Definition 4.8
A formula $\lambda$ is heap-unsatisfiable if $\mathit{alloc}(\lambda)$
contains two occurrences of the same variable. Otherwise, it is heap-
satisfiable.
The next proposition states that every location that is referred to in the
heap of some model of $\lambda$ must be reachable from one of the roots of
$\lambda$. This follows from the fact that, by definition of
$\mathtt{P}$-rules, the set of allocated locations has a tree-shaped
structure: the root of each atom invoked in an inductive rule must be
connected to the location allocated by the rule (see Condition 2 in Definition
2.9).
###### Proposition 4.9
Let $\lambda$ be a symbolic heap and let $(\mathfrak{s},\mathfrak{h})$ be an
$\mathfrak{R}$-model of $\lambda$. For every
$\ell\in\mathit{ref}(\mathfrak{h})$, there exists
$x\in\mathit{alloc}(\lambda)$ such that
$\mathfrak{s}(x)\rightarrow_{\mathfrak{h}}^{*}\ell$.
###### Proof 4.10
The proof is by induction on the satisfiability relation. We establish the
result also for spatial formulas and pure formulas.
* •
If $\lambda=\mathit{emp}$ or is $\lambda$ is a pure formula then
$\mathit{ref}(\lambda)=\emptyset$ hence the proof is immediate.
* •
If $\lambda=y_{0}\mapsto(y_{1},\dots,y_{n})$, then
$\mathfrak{h}=\\{(\mathfrak{s}(y_{0}),\dots,\mathfrak{s}(y_{n}))\\}$, by
Definition 2.4, thus $\mathit{ref}(\mathfrak{h})=\\{\mathfrak{s}(y_{i})\mid
i=0,\dots,n,\text{and $y_{i}$ is of sort $\mathtt{loc}$}\\}$ and
$\mathfrak{s}(y_{0})\rightarrow_{\mathfrak{h}}^{*}\mathfrak{s}(y_{i})$, for
all $i=1,\dots,n$ such that $y_{i}$ is of sort $\mathtt{loc}$. By Definition
4.1 $\mathit{alloc}(\lambda)=\\{y_{0}\\}$, thus the proof is completed.
* •
If $\lambda=\phi\curlywedge\xi$, where $\phi$ is a spatial formula and $\xi$
is a pure formula distinct from $\top$, then we have
$(\mathfrak{s},\mathfrak{h})\models_{\mathfrak{R}}\phi$ and
$\mathit{alloc}(\lambda)=\mathit{alloc}(\phi)$, hence the result follows
immediately from the induction hypothesis.
* •
If $\lambda=\phi_{1}*\phi_{2}$, then there exist two disjoint heaps
$\mathfrak{h}_{1},\mathfrak{h}_{2}$ such that
$\mathfrak{h}=\mathfrak{h}_{1}\uplus\mathfrak{h}_{2}$ and
$(\mathfrak{s},\mathfrak{h}_{i})\models_{\mathfrak{R}}\phi_{i}$, for all
$i=1,2$. If $\ell\in\mathit{ref}(\mathfrak{h})$ then necessarily
$\ell\in\mathit{ref}(\mathfrak{h}_{i})$ for some $i=1,2$. By the induction
hypothesis, we deduce that there exists $x\in\mathit{alloc}(\phi_{i})$ such
that $\mathfrak{s}(x)\rightarrow_{\mathfrak{h}_{i}}^{*}\ell$. Since
$\mathit{alloc}(\phi_{i})\subseteq_{m}\mathit{alloc}(\phi)$ and
$\rightarrow_{\mathfrak{h}_{i}}^{*}\subseteq\rightarrow_{\mathfrak{h}}^{*}$ by
Proposition 2.1, we obtain the result.
* •
If $\lambda$ is a predicate atom, then
$\lambda\Leftarrow_{\mathfrak{R}}\gamma$ and
$(\mathfrak{s}^{\prime},\mathfrak{h})\models_{\mathfrak{R}}\gamma$ for some
formula $\gamma$ and some associate $\mathfrak{s}^{\prime}$ of $\mathfrak{s}$
w.r.t. $\mathcal{V}(\gamma)\setminus\mathcal{V}(\lambda)$. Let
$\ell\in\mathit{ref}(\mathfrak{h})$. By the induction hypothesis, there exists
$x\in\mathit{alloc}(\gamma)$ such that
$\mathfrak{s}(x)\rightarrow_{\mathfrak{h}}^{*}\ell$. By definition, $x$ is the
root of some atom $\alpha$ in $\gamma$. If $\alpha$ is a points-to atom, then
since $\lambda\Leftarrow_{\mathfrak{R}}\gamma$ is an instance of a rule in
$\mathfrak{R}$ and all rules are $\mathtt{P}$-rules, $x$ must be the root of
$\lambda$; in this case $x\in\mathit{alloc}(\lambda)$ and the proof is
completed. Otherwise, $x$ is the root of a spatial atom in $\gamma$, and,
because all rules are $\mathtt{P}$-rules, $\gamma$ must contain an atom of the
form $y_{0}\mapsto(y_{1},\dots,y_{n})$, such that
$y_{0}=\mathit{root}(\lambda)$ and $y_{i}=x$, for some $i=1,\dots,n$. Since
$(\mathfrak{s}^{\prime},\mathfrak{h})\models_{\mathfrak{R}}\gamma$, we have
$(\mathfrak{s}(y_{0}),\dots,\mathfrak{s}(y_{n}))\in\mathfrak{h}$, hence
$\mathfrak{s}(y_{0})\rightarrow_{\mathfrak{h}}\mathfrak{s}(x)$. Using the fact
that $\mathfrak{s}(x)\rightarrow_{\mathfrak{h}}^{*}\ell$, we deduce that
$\mathfrak{s}(y_{0})\rightarrow_{\mathfrak{h}}^{*}\ell$, hence the proof is
completed since $\mathit{alloc}(\lambda)=\\{y_{0}\\}$.
The next lemma asserts a key property of the considered formulas: all the
locations occurring in the heap of a model of some formula $\phi$ are either
allocated or associated with a variable that is free in $\phi$. This follows
from the definition of $\mathtt{P}$-rules: all variables of sort
$\mathtt{loc}$ that are existentially quantified in an inductive rule must be
allocated (at the next recursive call). Recall that spatial formulas contain
no quantifiers.
###### Lemma 4.11
Let $\phi$ be a spatial formula and let $(\mathfrak{s},\mathfrak{h})$ be an
$\mathfrak{R}$-model of $\phi$. Then the following inclusion holds:
${\mathit{ref}(\mathfrak{h})\subseteq\mathit{dom}(\mathfrak{h})\cup\mathfrak{s}(\mathcal{V}(\phi))}$.
###### Proof 4.12
The proof is by induction on the relation $\models_{\mathfrak{R}}$. Note that
as $\phi$ is spatial, it contains no occurrence of $\approx$, $\not\approx$,
$\curlywedge$ and $\wedge$.
* •
If $\phi$ is of the form $y_{0}\mapsto(y_{1},\dots,y_{n})$ then by definition
$\mathfrak{h}=\\{(\mathfrak{s}(y_{0}),\dots,\mathfrak{s}(y_{n}))\\}$ and
$\mathit{ref}(\mathfrak{h})=\\{\mathfrak{s}(y_{0}),\dots,\mathfrak{s}(y_{n})\\}=\mathfrak{s}(\mathcal{V}(\phi))$.
* •
If $\phi=\mathit{emp}$ then $\mathfrak{h}=\emptyset$ thus
$\mathit{ref}(\mathfrak{h})=\emptyset$ and the proof is immediate.
* •
If $\phi$ is a predicate atom then we have
$\phi\Leftarrow_{\mathfrak{R}}\psi\curlywedge\xi$, and
$(\mathfrak{s}^{\prime},\mathfrak{h})\models_{\mathfrak{R}}\psi$, for some
associate $\mathfrak{s}^{\prime}$ of $\mathfrak{s}$ w.r.t.
$\mathcal{V}(\psi\curlywedge\xi)\setminus\mathcal{V}(\phi)$. Let
$\ell\in\mathit{ref}(\mathfrak{h})\setminus\mathit{dom}(\mathfrak{h})$. By the
induction hypothesis, $\ell=\mathfrak{s}^{\prime}(x)$ for some variable
$x\in\mathcal{V}(\psi)$. If $x\in\mathcal{V}(\phi)$ then necessarily
$\mathfrak{s}^{\prime}(x)=\mathfrak{s}(x)$, thus
$\ell\in\mathfrak{s}(\mathcal{V}(\phi))$ and the proof is completed.
Otherwise, by Definition 2.9, since all the rules in $\mathfrak{R}$ are
$\mathtt{P}$-rules; $x$ occurs as the root of some predicate atom in $\psi$,
i.e., $x\in\mathit{alloc}(\psi)$. By Lemma 4.2 we deduce that
$\mathfrak{s}^{\prime}(x)\in\mathit{dom}(\mathfrak{h})$, i.e.,
$\ell\in\mathit{dom}(\mathfrak{h})$ which contradicts our assumption.
* •
If $\phi$ is of the form $\phi_{1}*\phi_{2}$ then there exist disjoint heaps
$\mathfrak{h}_{1},\mathfrak{h}_{2}$ such that
$(\mathfrak{s},\mathfrak{h}_{i})\models_{\mathfrak{R}}\phi_{i}$ (for $i=1,2$)
and $\mathfrak{h}=\mathfrak{h}_{1}\uplus\mathfrak{h}_{2}$. Let
$\ell\in\mathit{ref}(\mathfrak{h})\setminus\mathit{dom}(\mathfrak{h})$.
Necessarily we have $\ell\in\mathit{ref}(\mathfrak{h}_{i})$, for some $i=1,2$
and $\ell\not\in\mathit{dom}(\mathfrak{h}_{i})$, hence by the induction
hypothesis we deduce that $\ell=\mathfrak{s}(x)$, for some
$x\in\mathcal{V}(\phi_{i})$. Since
$\mathcal{V}(\phi_{i})\subseteq\mathcal{V}(\phi)$, the proof is completed.
### 4.2 A Restricted Entailment Relation
We introduce a simple syntactic criterion, used in the inference rules of
Section 4.3, that is sufficient to ensure that a given pure formula $\xi$
holds in every counter-model of a sequent with left-hand side $\lambda$ and
multiset of variables $V$. The idea is to test that $\xi$ either occurs in
$\lambda$, is trivial, or is a disequation entailed by the fact that the
considered store must be injective on $\mathit{alloc}(\lambda)\cup V$ (using
Definition 2.18 and Corollary 4.4). Lemma 4.15 states that the relation
satisfies the expected property.
###### Definition 4.13
Let $\lambda$ be a symbolic heap, $\xi$ be a pure formula and let $V$ be a
multiset of variables. We write $\lambda\triangleright_{V}\xi$ if for every
atom $\zeta$ occurring in $\xi$, one of the following conditions holds:
1. 1.
$\zeta$ occurs in $\lambda$;
2. 2.
either $\zeta=(t\approx t)$ for some variable $t$, or $\zeta=(t_{1}\not\approx
t_{2})$ and $t_{1},t_{2}$ are distinct constants;
3. 3.
$\zeta=(x_{1}\not\approx x_{2})$ modulo commutativity, and one of the
following holds: $\\{x_{1},x_{2}\\}\subseteq_{m}\mathit{alloc}(\lambda)$,
($x_{1}\in\mathit{alloc}(\lambda)$ and $x_{2}\in V$) or
$\\{x_{1},x_{2}\\}\subseteq_{m}V$. This is equivalent to stating that
$\\{x_{1},x_{2}\\}\subseteq_{m}\mathit{alloc}(\lambda)+V$ where
$\mathit{alloc}(\lambda)+V$ denotes as usual the union of the multisets
$\mathit{alloc}(\lambda)$ and $V$.
###### Example 4.14
Consider the symbolic heap $\lambda=(p(x,y)*q(z))\curlywedge x\not\approx y$,
and let $V=\\{u\\}$. We have $\lambda\triangleright_{V}x\not\approx y\wedge
x\not\approx z\wedge x\not\approx u$. Indeed, $x$ and $z$ are necessarily
distinct since they are allocated by distinct atoms $p(x,y)$ and $q(z)$ (as,
by definition of the $\mathtt{P}$-rules, every predicate allocates it first
parameter) $x$ cannot be equal to $u$ as $u\in V$ and $V$ is intended to
denote a set of non-allocated variables (see Definition 2.18) and $x$ is
allocated, and $x$ cannot be equal to $y$ as the disequation $x\not\approx y$
occurs in $\lambda$.
###### Lemma 4.15
Let $\lambda$ be a symbolic heap and let $\xi$ be a pure formula such that
$\lambda\triangleright_{V}\xi$. For every structure
$(\mathfrak{s},\mathfrak{h})$, if
$(\mathfrak{s},\mathfrak{h})\models_{\mathfrak{R}}\lambda$,
$\mathfrak{s}(V)\cap\mathit{dom}(\mathfrak{h})=\emptyset$ and $\mathfrak{s}$
is injective on $V$ then $\mathfrak{s}\models_{\mathfrak{R}}\xi$.
###### Proof 4.16
We show that $(\mathfrak{s},\mathfrak{h})\models_{\mathfrak{R}}\zeta$, for all
atoms $\zeta$ in $\xi$. We consider each case in Definition 4.13 separately:
1. 1.
If $\zeta$ occurs in $\lambda$ then since
$(\mathfrak{s},\mathfrak{h})\models\gamma$, necessarily
$\mathfrak{s}\models_{\mathfrak{R}}\zeta$.
2. 2.
If $\zeta=(t\approx t)$ then it is clear that $\mathfrak{s}\models\zeta$. If
$\zeta=(t_{1}\not\approx t_{2})$ and $t_{1},t_{2}$ are distinct constants then
$\mathfrak{s}(t_{1})\not=\mathfrak{s}(t_{2})$ since all stores are injective
on constants by definition.
3. 3.
If $\zeta=x_{1}\not\approx x_{2}$ and
$\\{x_{1},x_{2}\\}\subseteq_{m}\mathit{alloc}(\lambda)$ then by Corollary 4.4,
we get $\mathfrak{s}(x_{1})\not=\mathfrak{s}(x_{2})$ since
$\mathfrak{s}(x_{1})$ and $\mathfrak{s}(x_{2})$ must be allocated in disjoint
heaps. Thus $\mathfrak{s}\models_{\mathfrak{R}}\zeta$. If
$x_{1}\in\mathit{alloc}(\phi)$ and $x_{2}\in V$ then we have
$\mathfrak{s}(x_{1})\in\mathit{dom}(\mathfrak{h})$ by Lemma 4.2, and since
$\mathfrak{s}(V)\cap\mathit{dom}(\mathfrak{h})=\emptyset$, we deduce that
$\mathfrak{s}\models_{\mathfrak{R}}\zeta$. Finally, if
$\\{x_{1},x_{2}\\}\subseteq_{m}V$ then
$\mathfrak{s}(x_{1})\not=\mathfrak{s}(x_{2})$, because $\mathfrak{s}$ is
injective on $V$ by hypothesis.
### 4.3 Inference Rules
We consider the rules represented in Figure 1. The rules apply modulo AC, they
are intended to be applied bottom-up: a rule is applicable on a sequent
$\lambda\vdash_{\mathfrak{R}}^{V}\gamma$ if there exists an instance of the
rule the conclusion of which is $\lambda\vdash_{\mathfrak{R}}^{V}\gamma$. We
recall that an inference rule is sound if the validity of the premises entails
the validity of the conclusion, and invertible if the converse holds.
###### Remark 4.17
The application conditions given in Figure 1 are the most general ones
ensuring that the rules are sound. Additional application conditions will be
provided afterwards (see Definition 4.43) to obtain a proof procedure that
runs in polynomial time. The latter conditions are rather complex, and in our
opinion introducing them at this point could hinder the understanding of the
rules.
$\repl{\phi}{x}{t}\curlywedge\repl{\xi}{x}{t}\vdash_{\mathfrak{R}}^{\repl{V}{x}{t}}\repl{\gamma}{x}{t}$
R: if $x\in\mathcal{V}$ $\phi\curlywedge(x\approx
t\wedge\xi)\vdash_{\mathfrak{R}}^{V}\gamma$
$\lambda\vdash_{\mathfrak{R}}^{V}\psi\curlywedge\zeta$ E: if
$\lambda\triangleright_{V}\zeta^{\prime}$
$\lambda\vdash_{\mathfrak{R}}^{V}\psi\curlywedge(\zeta\wedge\zeta^{\prime})$
$\phi_{1}\curlywedge\xi_{1}\vdash_{\mathfrak{R}}^{V\cup\mathit{alloc}(\phi_{2})}\psi_{1}\curlywedge\zeta_{1}$
$\phi_{2}\curlywedge\xi_{2}\vdash_{\mathfrak{R}}^{V\cup\mathit{alloc}(\phi_{1})}\psi_{2}\curlywedge\zeta_{2}$
S: if $\phi_{1}\not=\mathit{emp}$ and $\phi_{2}\not=\mathit{emp}$
$(\phi_{1}*\phi_{2})\curlywedge(\xi_{1}\wedge\xi_{2})\vdash_{\mathfrak{R}}^{V}(\psi_{1}*\psi_{2})\curlywedge(\zeta_{1}\wedge\zeta_{2})$
$(\phi^{\prime}_{1}*\phi)\curlywedge(\xi^{\prime}_{1}\wedge\xi)\vdash_{\mathfrak{R}}^{V}\gamma$
$\dots$
$(\phi^{\prime}_{n}*\phi)\curlywedge(\xi^{\prime}_{n}\wedge\xi)\vdash_{\mathfrak{R}}^{V}\gamma$
U: if
$p(\bm{t})\rightsquigarrow_{\mathfrak{R}}\\{\phi^{\prime}_{i}\curlywedge\xi^{\prime}_{i}\mid
i=1,\dots,n\\}$
$(p(\bm{t})*\phi)\curlywedge\xi\vdash_{\mathfrak{R}}^{V}\gamma$
$(x\mapsto(y_{1},\dots,y_{k})*\phi)\curlywedge\xi\vdash_{\mathfrak{R}}^{V}(x\mapsto(y_{1},\dots,y_{k})*\psi^{\prime}\sigma*\psi)\curlywedge\zeta$
I: if all the conditions below hold:
$(x\mapsto(y_{1},\dots,y_{k})*\phi)\curlywedge\xi\vdash_{\mathfrak{R}}^{V}(p(x,\bm{z})*\psi)\curlywedge\zeta$
(i)
$p(x,\bm{z})\Leftarrow_{\mathfrak{R}}(x\mapsto(u_{1},\dots,u_{k})*\psi^{\prime})\curlywedge\zeta^{\prime}$;
(ii) $\sigma$ is a substitution such that
$\mathit{dom}(\sigma)\subseteq\\{u_{1},\dots,u_{k}\\}\setminus(\\{x\\}\cup\bm{z})$;
(iii) $\sigma(u_{i})=y_{i}$, for all $i\in\\{1,\dots,k\\}$; and (iv)
$(x\mapsto(y_{1},\dots,y_{k})*\phi)\curlywedge\xi\triangleright_{V}\zeta^{\prime}\sigma$
$\phi\curlywedge\xi\vdash_{\mathfrak{R}}^{V}\gamma$ W: if one of the
conditions below holds:
$\phi\curlywedge(\xi\wedge\xi^{\prime})\vdash_{\mathfrak{R}}^{V}\gamma$
(i) $\xi^{\prime}=x\not\approx y$ and $\phi\triangleright_{V}x\not\approx y$;
or (ii) $\xi^{\prime}=\bigwedge_{i=1}^{n}x\not\approx y_{i}$ and
$x\not\in\mathcal{V}(\phi)\cup\mathcal{V}(\xi)\cup\mathcal{V}(\gamma)\cup\\{y_{1},\dots,y_{n}\\}$
$\repl{\phi}{x}{y}\curlywedge\repl{\xi}{x}{y}\vdash_{\mathfrak{R}}^{\repl{V}{x}{y}}\repl{\gamma}{x}{y}$
$\phi\curlywedge(\xi\wedge x\not\approx y)\vdash_{\mathfrak{R}}^{V}\gamma$ C:
if $x,y\in\mathcal{V}_{\mathtt{loc}}$
$\phi\curlywedge\xi\vdash_{\mathfrak{R}}^{V}\gamma$
$\lambda\vdash_{\mathfrak{R}}^{V}\gamma$ V: if
$x\not\in\mathcal{V}(\lambda)\cup\mathcal{V}(\gamma)$
$\lambda\vdash_{\mathfrak{R}}^{V\cup\\{x\\}}\gamma$
Figure 1: The Inference rules
We provide some examples and explanations concerning the inference rules.
###### Example 4.18
Rules R (replacement) and E (elimination) handle equational reasoning. For
instance, given the sequent $(p(x,u)*p(y,u))\curlywedge(u\approx
v)\vdash_{\mathfrak{R}}^{\emptyset}(p(x,u)*p(y,v))\curlywedge(x\not\approx
y)$, one may first apply R, yielding:
$p(x,u)*p(y,u)\vdash_{\mathfrak{R}}^{\emptyset}(p(x,u)*p(y,u))\curlywedge(x\not\approx
y)$. As $\\{x,y\\}\subseteq_{m}\mathit{alloc}(p(x,u)*p(y,x))$, E applies,
which yields the trivially valid sequent
$p(x,u)*p(y,u)\vdash_{\mathfrak{R}}^{\emptyset}p(x,u)*p(y,u)$.
###### Example 4.19
Rule S (separation) applies on the sequent
$p(x)*p(y)\vdash_{\mathfrak{R}}^{\emptyset}q(x)*r(y)$, yielding
$p(x)\vdash_{\mathfrak{R}}^{\\{y\\}}q(x)$ and
$p(y)\vdash_{\mathfrak{R}}^{\\{x\\}}r(y)$ (as $\\{x\\}=\mathit{alloc}(p(x))$
and $\\{y\\}=\mathit{alloc}(p(y))$). The addition of $y$ (resp. $x$) to the
variables associated with the sequent allows us to keep track of the fact that
these variables cannot be allocated in $p(x)$ (resp. $p(y)$) as they are
already allocated in the other part of the heap. Note that the rule also
yields $p(x)\vdash_{\mathfrak{R}}^{\\{y\\}}r(y)$ and
$p(y)\vdash_{\mathfrak{R}}^{\\{x\\}}q(x)$, however as we shall see later (see
Definition 4.34) the latter premises cannot be valid and this application can
be dismissed.
###### Example 4.20
Rules U (unfolding) and I (imitation) both unfold inductively defined
predicate symbols. U unfolds predicates occurring on the left-hand side of a
sequent, yielding one premise for each inductive rule. In contrast, I applies
on the right-hand side and selects one rule in a non-deterministic way
(provided it fulfills the rule application condition), yielding exactly one
premise. Let $\mathfrak{R}$ be the following set of rules, where
$\mathtt{a},\mathtt{b}$ are constant symbols and $z,z_{1},z_{2}$ are variables
of the same sort as $\mathtt{a}$ and $\mathtt{b}$:
$p(x)$ | $\Leftarrow$ | $x\mapsto(\mathtt{a},x)$
---|---|---
$p(x)$ | $\Leftarrow$ | $x\mapsto(\mathtt{b},x)$
$q(x)$ | $\Leftarrow$ | $x\mapsto(z,y)$
$q(x)$ | $\Leftarrow$ | $x\mapsto(z_{1},z_{2},x)$
Rule U applies on $p(x,y)\vdash_{\mathfrak{R}}^{\emptyset}q(x)$, yielding
$x\mapsto(\mathtt{a},x)\vdash_{\mathfrak{R}}^{\emptyset}q(x)$ and
$x\mapsto(\mathtt{b},x)\vdash_{\mathfrak{R}}^{\emptyset}q(x)$. Then I applies
on both sequents, with the respective substitutions $\\{y\leftarrow
x,z\leftarrow\mathtt{a}\\}$ and $\\{y\leftarrow x,z\leftarrow\mathtt{b}\\}$,
yielding
$x\mapsto(\mathtt{a},x)\vdash_{\mathfrak{R}}^{\emptyset}x\mapsto(\mathtt{a},x)$
and
$x\mapsto(\mathtt{b},x)\vdash_{\mathfrak{R}}^{\emptyset}x\mapsto(\mathtt{b},x)$.
Note that I cannot be applied with the inductive rule $q(x)\Leftarrow
x\mapsto(z_{1},z_{2},x)$, as $(\mathtt{a},x)$ and $(\mathtt{b},x)$ are not
instances of $(z_{1},z_{2},x)$.
###### Example 4.21
Rules W (weakening) and V (variable weakening) allow to get rid of
irrelevant information, which is essential for termination. For instance one
may deduce $p(x)\vdash_{\mathfrak{R}}^{\emptyset}q(x)$ from
$p(x)\curlywedge(u\not\approx v\wedge u\not\approx
w)\vdash_{\mathfrak{R}}^{\emptyset}q(x)$. Indeed, if the former sequent admits
a counter-model, then one gets a counter-model of the latter one by
associating $u,v,w$ with pairwise distinct elements.
###### Example 4.22
Rule C performs a case analysis on $x\approx y$. It is essential to allow
further applications of Rule I in some cases. Consider the sequent
$u\mapsto(x,x)*p(x)\vdash_{\mathfrak{R}}^{\emptyset}q(u,y)$, with the rules
$q(u,y)\Leftarrow u\mapsto(y,z)*p(z)$, and
$q(u,y)\Leftarrow(u\mapsto(z,z)*p(z))\curlywedge z\not\approx y$. Rule I does
not apply because there is no substitution $\sigma$ with domain $\\{z\\}$ such
that $(x,x)=(y,z)\sigma$, and
${u\mapsto(x,x)*p(x)\not\\!\triangleright_{V}x\not\approx y}$. The rule C
yields $u\mapsto(y,y)*p(y)\vdash_{\mathfrak{R}}^{\emptyset}q(u,y)$ and
$(u\mapsto(x,x)*p(x))\curlywedge x\not\approx
y\vdash_{\mathfrak{R}}^{\emptyset}q(u,y)$. Then the rule I applies on both
sequents, yielding the premisses
${u\mapsto(y,y)*p(y)\vdash_{\mathfrak{R}}^{\emptyset}u\mapsto(y,y)*p(y)}$ and
${(u\mapsto(x,x)*p(x))\curlywedge x\not\approx
y\vdash_{\mathfrak{R}}^{\emptyset}u\mapsto(x,x)*p(x)}$.
We have the following facts, which can be verified by an inspection of the
inference rules:
###### Proposition 4.23
Consider an equality-free sequent $\lambda\vdash_{\mathfrak{R}}^{V}\gamma$
that is the conclusion of an inference rule, of which
$\lambda^{\prime}\vdash_{\mathfrak{R}}^{V^{\prime}}\gamma^{\prime}$ is a
premise.
1. 1.
No rule can introduce any equality to
$\lambda^{\prime}\vdash_{\mathfrak{R}}^{V^{\prime}}\gamma^{\prime}$.
2. 2.
If $x\in V^{\prime}\setminus V$, then the inference rule is either V or C.
3. 3.
If $\mathit{alloc}(\gamma^{\prime})\subsetneq\mathit{alloc}(\gamma)$ then the
inference rule is either S or C.
4. 4.
The only inference rules that can introduce new variables to $\gamma^{\prime}$
are I and C.
5. 5.
No rule introduces a disequation between terms of a sort distinct from
$\mathtt{loc}$.
6. 6.
The only rule that introduces a predicate atom to the right-hand side of a
premise is I.
7. 7.
If $v\in(\mathit{alloc}(\gamma)\cup
V)\setminus(\mathit{alloc}(\gamma^{\prime})\cup V^{\prime})$, then the
inference rule must be C.
###### Proof 4.24
The first six facts are straightforward to verify. Fact 7 holds because Rule
R cannot apply if no equality occurs; if rule S is applied then the variables
in $\mathit{alloc}(\gamma)\setminus\mathit{alloc}(\gamma^{\prime})$ must occur
in $V^{\prime}$; Rule I deletes a predicate atom but introduces a points-to
atom with the same root and rule V cannot be applied on variables occurring
in $\mathcal{V}{\gamma}$.
We establish additional basic properties of the inference rules. All the rules
are sound and invertible, except for rule Sthat is only sound. The results
follow easily from the semantics, except for the invertibility of I, which
crucially depends on the fact that rules are deterministic. Indeed, I unfolds
one atom on the right-hand side by selecting one specific inductive rule. In
our case, at most one rule can be applied, which ensures that equivalence is
preserved. This is a crucial point because otherwise one would need to
consider disjunctions of formulas on the right-hand side of the sequent (one
disjunct for each possible rule), which would make the procedure much less
efficient.
###### Example 4.25
Consider the sequent
$x\mapsto(y)*y\mapsto(z)\vdash_{\mathfrak{R}}^{\emptyset}p(x,z)$, with the
rules
$p(x,z)$ | $\Leftarrow$ | $x\mapsto(y)*q(y,z)$ | $p(x,z)$ | $\Leftarrow$ | $x\mapsto(y)*q^{\prime}(y,y)$
---|---|---|---|---|---
$q(y,z)$ | $\Leftarrow$ | $y\mapsto(z)\curlywedge y\not\approx z$ | $q^{\prime}(y,z)$ | $\Leftarrow$ | $y\mapsto(z)$
It is clear that the rules are not deterministic, as there is an overlap
between the two rules associated with $p$. Applying rule I on the above
sequent yields either
$x\mapsto(y)*y\mapsto(z)\vdash_{\mathfrak{R}}^{\emptyset}x\mapsto(y)*q(y,z)$
or
$x\mapsto(y)*y\mapsto(z)\vdash_{\mathfrak{R}}^{\emptyset}x\mapsto(y)*q^{\prime}(y,y)$.
None of these two possible premises is valid, although the initial sequent is
valid. This shows that I is not invertible (although it is still sound) when
$\mathfrak{R}$ is not deterministic. The intuition is that it is not possible
to decide which rule must be applied on $p$ before deciding whether $z$ is
equal to $y$ or not.
###### Lemma 4.26
The rules R, E, U, W, V and C and I are sound and invertible. More
specifically, for all heaps $\mathfrak{h}$, the conclusion of the rule admits
a counter-model $(\mathfrak{s},\mathfrak{h})$ iff one of the premises admits a
counter-model $(\mathfrak{s}^{\prime},\mathfrak{h})$.
###### Proof 4.27
We consider each rule separately (we refer to the definitions of the rules for
notations) and establish the equivalence of the lemma. We recall (Definition
2.18) that a counter-model of a sequent is a structure
$(\mathfrak{s},\mathfrak{h})$ that validates the antecedent, falsifies the
conclusion, and is such that the store is injective on $V$ and no variable in
$V$ is allocated.
* R
Assume that
$(\mathfrak{s},\mathfrak{h})\models_{\mathfrak{R}}\phi\curlywedge(x\approx
t\wedge\xi)$, that
$(\mathfrak{s},\mathfrak{h})\not\models_{\mathfrak{R}}\gamma$, that
$\mathfrak{s}(V)\cap\mathit{dom}(\mathfrak{h})=\emptyset$ and that
$\mathfrak{s}$ is injective on $V$. Then we have
$\mathfrak{s}(x)=\mathfrak{s}(t)$, thus
$(\mathfrak{s},\mathfrak{h})\models_{\mathfrak{R}}\repl{\phi}{x}{t}\curlywedge\repl{\xi}{x}{t}$
and
$(\mathfrak{s},\mathfrak{h})\not\models_{\mathfrak{R}}\repl{\gamma}{x}{t}$.
For all $y\in\repl{V}{x}{t}$, we have either $y\in V$ and
$\mathfrak{s}(y)\not\in\mathit{dom}(\mathfrak{h})$, or $x\in V$ and $y=t$,
thus
$\mathfrak{s}(y)=\mathfrak{s}(t)=\mathfrak{s}(x)\not\in\mathit{dom}(\mathfrak{h})$.
Finally, assume (for the sake of contradiction) that
$\\{u,v\\}\subseteq\repl{V}{x}{t}$ with $\mathfrak{s}(u)=\mathfrak{s}(v)$.
Then there exist variables $u^{\prime},v^{\prime}$ such that
$\\{u^{\prime},v^{\prime}\\}\subseteq V$, with $\repl{u^{\prime}}{x}{t}=u$ and
$\repl{v^{\prime}}{x}{t}=v$. If $u^{\prime}$ and $v^{\prime}$ are both equal
to $x$ or both distinct from $x$ then we have
$\mathfrak{s}(u^{\prime})=\mathfrak{s}(v^{\prime})$, which contradicts the
fact that $\mathfrak{s}$ is injective on $V$. If $u^{\prime}=x$ and
$v^{\prime}\neq x$, then we have
$\mathfrak{s}(v^{\prime})=\mathfrak{s}(t)=\mathfrak{s}(x)$, again
contradicting the fact that $\mathfrak{s}$ is injective on $V$. The proof is
similar when $u^{\prime}\neq x$ and $v^{\prime}=x$. Consequently,
$(\mathfrak{s},\mathfrak{h})$ is also a counter-model of the premise.
Conversely, assume that
$(\mathfrak{s},\mathfrak{h})\models_{\mathfrak{R}}\repl{\phi}{x}{t}\curlywedge\repl{\xi}{x}{t}$;
$\mathfrak{s}(\repl{V}{x}{t})\cap\mathit{dom}(\mathfrak{h})=\emptyset$; the
store $\mathfrak{s}$ is injective on $\repl{V}{x}{t}$; and
$(\mathfrak{s},\mathfrak{h})\not\models_{\mathfrak{R}}\repl{\gamma}{x}{t}$. We
consider the store $\mathfrak{s}^{\prime}$ such that
$\mathfrak{s}^{\prime}(x)=\mathfrak{s}(t)$ and
$\mathfrak{s}^{\prime}(y)=\mathfrak{s}(y)$ if $y\not=x$. It is clear that
$\mathfrak{s}^{\prime}\models_{\mathfrak{R}}x\approx t$,
$(\mathfrak{s}^{\prime},\mathfrak{h})\models_{\mathfrak{R}}\phi\curlywedge\xi$
and $(\mathfrak{s}^{\prime},\mathfrak{h})\not\models_{\mathfrak{R}}\gamma$.
For all $y\in V$, if $y\not=x$ then $y\in\repl{V}{x}{t}$ and
$\mathfrak{s}^{\prime}(y)=\mathfrak{s}(y)\not\in\mathit{dom}(\mathfrak{h})$;
otherwise $y=x$ and $t\in\repl{V}{x}{t}$, thus
$\mathfrak{s}^{\prime}(y)=\mathfrak{s}^{\prime}(x)=\mathfrak{s}(t)\not\in\mathit{dom}(\mathfrak{h})$.
There only remains to check that $\mathfrak{s}^{\prime}$ is injective on $V$,
which is done by contradiction: if this is not the case then there exists
$\\{u,v\\}\subseteq_{m}V$ such that
$\mathfrak{s}^{\prime}(u)=\mathfrak{s}^{\prime}(v)$. Hence
$\\{\repl{u}{x}{t},\repl{v}{x}{t}\\}\subseteq_{m}\repl{V}{x}{t}$, and we have
$\mathfrak{s}^{\prime}(\repl{u}{x}{t})=\mathfrak{s}(\repl{u}{x}{t})$ and
$\mathfrak{s}^{\prime}(\repl{v}{x}{t})=\mathfrak{s}(\repl{v}{x}{t})$. This
contradicts the fact that $\mathfrak{s}$ is injective on $\repl{V}{x}{t}$.
* E
Let $(\mathfrak{s},\mathfrak{h})$ be a structure such that
$(\mathfrak{s},\mathfrak{h})\models_{\mathfrak{R}}\lambda$;
$\mathfrak{s}(V)\cap\mathit{dom}(\mathfrak{h})=\emptyset$; the store
$\mathfrak{s}$ is injective on $V$; and
$(\mathfrak{s},\mathfrak{h})\not\models_{\mathfrak{R}}\psi\curlywedge(\zeta\wedge\zeta^{\prime})$.
By the application condition of the rule we have
$\lambda\triangleright_{V}\zeta^{\prime}$, thus by Lemma 4.15, we deduce that
$(\mathfrak{s},\mathfrak{h})\models_{\mathfrak{R}}\zeta^{\prime}$. Therefore
$(\mathfrak{s},\mathfrak{h})\not\models_{\mathfrak{R}}\psi\curlywedge\zeta$.
Conversely, it is clear that any counter-model of
$\lambda\vdash_{\mathfrak{R}}^{V}\psi\curlywedge\zeta$ is a counter-model of
$\lambda\vdash_{\mathfrak{R}}^{V}\psi\curlywedge(\zeta\wedge\zeta^{\prime})$.
* U
Assume that
$(\mathfrak{s},\mathfrak{h})\models_{\mathfrak{R}}(p(\bm{t})*\phi)\curlywedge\xi$,
$\mathfrak{s}(V)\cap\mathit{dom}(\mathfrak{h})=\emptyset$, $\mathfrak{s}$ is
injective on $V$ and
$(\mathfrak{s},\mathfrak{h})\not\models_{\mathfrak{R}}\gamma$. By definition,
$\mathfrak{s}\models\xi$, and there exist disjoint heaps
$\mathfrak{h}_{1},\mathfrak{h}_{2}$ such that
${(\mathfrak{s},\mathfrak{h}_{1})\models_{\mathfrak{R}}p(\bm{t})}$,
${(\mathfrak{s},\mathfrak{h}_{2})\models_{\mathfrak{R}}\phi}$ and
$\mathfrak{h}=\mathfrak{h}_{1}\uplus\mathfrak{h}_{2}$. We get that
$p(\bm{t})\Leftarrow_{\mathfrak{R}}\phi^{\prime}\curlywedge\xi^{\prime}$ and
$(\mathfrak{s}^{\prime},\mathfrak{h}_{1})\models_{\mathfrak{R}}\phi^{\prime}\curlywedge\xi^{\prime}$,
for some associate of $\mathfrak{s}$ w.r.t.
$\mathcal{V}(\phi^{\prime}\curlywedge\xi^{\prime})\setminus\mathcal{V}(p(\bm{t}))$.
We assume that
${\mathcal{V}(\phi^{\prime}\curlywedge\xi^{\prime})\cap{\mathcal{V}((p(\bm{t})*\phi)\curlywedge\xi)\subseteq\bm{t}}}$
(by $\alpha$-renaming). We get $\mathfrak{s}^{\prime}\models\xi^{\prime}$ and
$(\mathfrak{s}^{\prime},\mathfrak{h}_{1}\uplus\mathfrak{h}_{2})\models_{\mathfrak{R}}\phi^{\prime}*\phi$,
hence
$(\mathfrak{s}^{\prime},\mathfrak{h})\models_{\mathfrak{R}}(\phi^{\prime}*\phi)\curlywedge(\xi^{\prime}\wedge\xi)$.
Furthermore, the formula $\phi^{\prime}\curlywedge\xi^{\prime}$ occurs in
$\\{\phi_{i}^{\prime}\curlywedge\xi_{i}^{\prime}\mid i=1,\dots,n\\}$, up to a
renaming of the variables not occurring in $\bm{t}$, by definition of
$\rightsquigarrow_{\mathfrak{R}}$. Thus $(\mathfrak{s}^{\prime},\mathfrak{h})$
is a counter-model of a sequent
$(\phi^{\prime}_{i}*\phi)\curlywedge(\xi^{\prime}_{i}\wedge\xi)\vdash_{\mathfrak{R}}^{V}\gamma$,
for some $i=1,\dots,n$.
Conversely, let $(\mathfrak{s},\mathfrak{h})$ be a structure such that
$(\mathfrak{s},\mathfrak{h})\models_{\mathfrak{R}}(\phi^{\prime}_{i}*\phi)\curlywedge(\xi^{\prime}_{i}\wedge\xi)$,
$\mathfrak{s}(V)\cap\mathit{dom}(\mathfrak{h})=\emptyset$, $\mathfrak{s}$ is
injective on $V$ and
$(\mathfrak{s},\mathfrak{h})\not\models_{\mathfrak{R}}\gamma$. We deduce that
$\mathfrak{s}\models\xi^{\prime}_{i}\wedge\xi$ and there exist disjoint heaps
$\mathfrak{h}_{1},\mathfrak{h}_{2}$ such that
$(\mathfrak{s},\mathfrak{h}_{1})\models_{\mathfrak{R}}\phi^{\prime}_{i}$ and
$(\mathfrak{s},\mathfrak{h}_{2})\models_{\mathfrak{R}}\phi$. By the
application condition of the rule we have
$p(\bm{t})\Leftarrow_{\mathfrak{R}}\phi^{\prime}_{i}\curlywedge\xi^{\prime}_{i}$,
thus $(\mathfrak{s},\mathfrak{h}_{1})\models_{\mathfrak{R}}p(\bm{t})$ by
definition of the semantics of predicate atoms (since $\mathfrak{s}$ is an
extension of itself). Consequently,
$(\mathfrak{s},\mathfrak{h}_{1}\uplus\mathfrak{h}_{2})\models_{\mathfrak{R}}p(\bm{t})*\phi$,
hence
$(\mathfrak{s},\mathfrak{h})\models_{\mathfrak{R}}(p(\bm{t})*\phi)\curlywedge\xi$,
and $(\mathfrak{s},\mathfrak{h})$ is a counter-model of
$(p(\bm{t})*\phi)\curlywedge\xi\vdash_{\mathfrak{R}}^{V}\gamma$.
* W
Assume that
$(\mathfrak{s},\mathfrak{h})\models_{\mathfrak{R}}\phi\curlywedge\xi$,
$\mathfrak{s}(V)\cap\mathit{dom}(\mathfrak{h})=\emptyset$, $\mathfrak{s}$ is
injective on $V$ and
$(\mathfrak{s},\mathfrak{h})\not\models_{\mathfrak{R}}\gamma$. If
$\xi^{\prime}$ is $x\not\approx y$, with $\phi\triangleright_{V}x\not\approx
y$, then, by Lemma 4.15, we have $\mathfrak{s}(x)\not=\mathfrak{s}(y)$, thus
$(\mathfrak{s},\mathfrak{h})\models_{\mathfrak{R}}\phi\curlywedge(\xi\wedge\xi^{\prime})$,
hence $(\mathfrak{s},\mathfrak{h})$ is a counter-model of the conclusion of
the rule. Assume that $\xi^{\prime}=\bigwedge_{i=1}^{n}x\not\approx y_{i}$
with
$x\not\in\mathcal{V}(\phi)\cup\mathcal{V}(\xi)\cup\mathcal{V}(\gamma)\cup\\{y_{1},\dots,y_{n}\\}$.
Let $\mathfrak{s}^{\prime}$ be a store such that
$\mathfrak{s}^{\prime}(y)=\mathfrak{s}(y)$ if $y\not=x$ and
$\mathfrak{s}^{\prime}(x)$ is an arbitrarily chosen location not occurring in
$\mathit{dom}(\mathfrak{h})\cup\\{\mathfrak{s}(y_{i})\mid i=1,\dots,n\\}$
(such a location exists since $\mathfrak{h}$ is finite and
$\mathfrak{U}_{\mathtt{loc}}$ is infinite). By definition, we have
$\mathfrak{s}^{\prime}(V)\cap\mathit{dom}(\mathfrak{h})=\emptyset$,
$\mathfrak{s}^{\prime}\models\bigwedge_{i=1}^{n}x\not\approx y_{i}$,
$(\mathfrak{s}^{\prime},\mathfrak{h})\models_{\mathfrak{R}}\phi\curlywedge\xi$
(since $\mathfrak{s}^{\prime}$ and $\mathfrak{s}$ coincide on
$\mathcal{V}(\phi)\cup\mathcal{V}(\xi)$) and
$(\mathfrak{s}^{\prime},\mathfrak{h})\not\models_{\mathfrak{R}}\gamma$ (since
$\mathfrak{s}^{\prime}$ and $\mathfrak{s}$ coincide on $\mathcal{V}(\gamma)$).
Thus $(\mathfrak{s}^{\prime},\mathfrak{h})$ is a counter-model of
$\phi\curlywedge(\xi\wedge\bigwedge_{i=1}^{n}x\not\approx
y_{i})\vdash_{\mathfrak{R}}^{V}\gamma$.
Conversely, it is clear that any counter-model of
$\phi\curlywedge(\xi\wedge\xi^{\prime})\vdash_{\mathfrak{R}}^{V}\gamma$ is a
counter-model of $\phi\curlywedge\xi\vdash_{\mathfrak{R}}^{V}\gamma$.
* C
Let $(\mathfrak{s},\mathfrak{h})$ such that
$(\mathfrak{s},\mathfrak{h})\models_{\mathfrak{R}}\phi\curlywedge\xi$,
$\mathfrak{s}(V)\cap\mathit{dom}(\mathfrak{h})=\emptyset$, $\mathfrak{s}$ is
injective on $V$ and
$(\mathfrak{s},\mathfrak{h})\not\models_{\mathfrak{R}}\gamma$. We distinguish
two cases.
* –
If $\mathfrak{s}(x)=\mathfrak{s}(y)$ then
$(\mathfrak{s},\mathfrak{h})\models_{\mathfrak{R}}\repl{\phi}{x}{y}$,
$(\mathfrak{s},\mathfrak{h})\models_{\mathfrak{R}}\repl{\xi}{x}{y}$, and
$(\mathfrak{s},\mathfrak{h})\not\models_{\mathfrak{R}}\repl{\gamma}{x}{y}$.
Moreover,
$\mathfrak{s}(\repl{V}{x}{y})\cap\mathit{dom}(\mathfrak{h})=\mathfrak{s}(V)\cap\mathit{dom}(\mathfrak{h})=\emptyset$,
and $\mathfrak{s}$ must be injective on $\repl{V}{x}{y}$, hence
$(\mathfrak{s},\mathfrak{h})$ is a counter-model of
$\repl{\phi}{x}{y}\curlywedge\repl{\xi}{x}{y}\vdash_{\mathfrak{R}}^{\repl{V}{x}{y}}\repl{\lambda}{x}{y}$.
* –
Otherwise, we have $\mathfrak{s}\models x\not\approx y$, thus
$(\mathfrak{s},\mathfrak{h})\models_{\mathfrak{R}}\phi\curlywedge(\xi\wedge
x\not\approx y)$, and $(\mathfrak{s},\mathfrak{h})$ is a counter-model of
$\phi\curlywedge(\xi\wedge x\not\approx y)\vdash_{\mathfrak{R}}^{V}\lambda$.
Conversely, if $(\mathfrak{s},\mathfrak{h})$ is a counter-model of
$\phi\curlywedge(\xi\wedge x\not\approx y)\vdash_{\mathfrak{R}}^{V}\gamma$
then it is clear that it is also a counter-model of
$\phi\curlywedge\xi\vdash_{\mathfrak{R}}^{V}\gamma$. If
$(\mathfrak{s},\mathfrak{h})$ is a counter-model of
$\repl{\phi}{x}{y}\curlywedge\repl{\xi}{x}{y}\vdash_{\mathfrak{R}}^{\repl{V}{x}{y}}\repl{\gamma}{x}{y}$,
then consider the store $\mathfrak{s}^{\prime}$ such that
$\mathfrak{s}^{\prime}(x)=\mathfrak{s}(y)$ and
$\mathfrak{s}^{\prime}(z)=\mathfrak{s}(z)$ if $z\not=x$. By definition,
$(\mathfrak{s}^{\prime},\mathfrak{h})\models_{\mathfrak{R}}\phi\curlywedge\xi$
and $(\mathfrak{s}^{\prime},\mathfrak{h})\not\models_{\mathfrak{R}}\gamma$.
The set
${\mathfrak{s}^{\prime}(V)\cap\mathit{dom}(\mathfrak{h})=\mathfrak{s}(\repl{V}{x}{y})\cap\mathit{dom}(\mathfrak{h})}$
is empty, and $\mathfrak{s}^{\prime}$ is injective on $V$, since
$\mathfrak{s}$ is injective on $\repl{V}{x}{y}$ and by definition
$\mathfrak{s}^{\prime}(u)=\mathfrak{s}(\repl{u}{x}{y})$, for all variables
$u$. Therefore $(\mathfrak{s}^{\prime},\mathfrak{h})$ is a counter-model of
$\phi\curlywedge\xi\vdash_{\mathfrak{R}}^{V}\gamma$.
* V
Let $(\mathfrak{s},\mathfrak{h})$ be a counter-model of
$\lambda\vdash_{\mathfrak{R}}^{V}\gamma$. Then
$(\mathfrak{s},\mathfrak{h})\models_{\mathfrak{R}}\lambda$ and
$(\mathfrak{s},\mathfrak{h})\not\models_{\mathfrak{R}}\gamma$. Let
$\mathfrak{s}^{\prime}$ be a store such that
$\mathfrak{s}^{\prime}(x)\not\in\mathit{dom}(\mathfrak{h})\cup\mathfrak{s}(V)$
and $\mathfrak{s}^{\prime}(y)=\mathfrak{s}(y)$, if $y\not=x$. Since
$x\not\in\mathcal{V}(\lambda)\cup\mathcal{V}(\gamma)$ we have
$(\mathfrak{s}^{\prime},\mathfrak{h})\models_{\mathfrak{R}}\lambda$, and
$(\mathfrak{s}^{\prime},\mathfrak{h})\not\models_{\mathfrak{R}}\gamma$.
Moreover $\mathfrak{s}^{\prime}(x)\not\in\mathit{dom}(\mathfrak{h})$ and
$\mathfrak{s}^{\prime}$ is injective on $V\cup\\{x\\}$ (since $\mathfrak{s}$
is injective on $V$ and $\mathfrak{s}^{\prime}(x)\not\in V$), thus
$(\mathfrak{s}^{\prime},\mathfrak{h})$ is a counter-model of
$\lambda\vdash_{\mathfrak{R}}^{V\cup\\{x\\}}\gamma$.
Conversely, it is clear that any counter-model of
$\lambda\vdash_{\mathfrak{R}}^{V\cup\\{x\\}}\gamma$ is also a counter-model of
$\lambda\vdash_{\mathfrak{R}}^{V}\gamma$.
* I
Let $(\mathfrak{s},\mathfrak{h})$ be a counter-model of
$(x\mapsto(y_{1},\dots,y_{k})*\phi)\curlywedge\xi\vdash_{\mathfrak{R}}^{V}(p(x,\bm{z})*\psi)\curlywedge\zeta$.
By definition
$(\mathfrak{s},\mathfrak{h})\models_{\mathfrak{R}}(x\mapsto(y_{1},\dots,y_{k})*\phi)\curlywedge\xi$,
$(\mathfrak{s},\mathfrak{h})\not\models_{\mathfrak{R}}(p(x,\bm{z})*\psi)\curlywedge\zeta$,
the set ${\mathfrak{s}(V)\cap\mathit{dom}(\mathfrak{h})}$ is empty and
$\mathfrak{s}$ is injective on $V$. By the application conditions of the rule
and Lemma 4.15, we have $\mathfrak{s}\models\zeta^{\prime}\sigma$. Assume for
the sake of contradiction that $(\mathfrak{s},\mathfrak{h})$ is not a counter-
model of
$(x\mapsto(y_{1},\dots,y_{k})*\phi)\curlywedge\xi\vdash_{\mathfrak{R}}^{V}(x\mapsto(y_{1},\dots,y_{k})*\psi^{\prime}\sigma*\psi)\curlywedge\zeta$.
This entails that
${(\mathfrak{s},\mathfrak{h})\models_{\mathfrak{R}}(x\mapsto(y_{1},\dots,y_{k})*\psi^{\prime}\sigma*\psi)\curlywedge\zeta}$,
hence $\mathfrak{s}\models\zeta$, and there exist disjoint heaps
$\mathfrak{h}_{1},\mathfrak{h}_{2}$ and $\mathfrak{h}_{3}$ such that
$\mathfrak{h}=\mathfrak{h}_{1}\uplus\mathfrak{h}_{2}\uplus\mathfrak{h}_{3}$,
$(\mathfrak{s},\mathfrak{h}_{1})\models_{\mathfrak{R}}x\mapsto(y_{1},\dots,y_{k})$,
$(\mathfrak{s},\mathfrak{h}_{2})\models_{\mathfrak{R}}\psi^{\prime}\sigma$ and
$(\mathfrak{s},\mathfrak{h}_{3})\models_{\mathfrak{R}}\psi$. Let
$\lambda=x\mapsto(u_{1},\dots,u_{k})*\psi^{\prime})\curlywedge\zeta^{\prime}$,
so that $p(x,\bm{z})\Leftarrow_{\mathfrak{R}}\lambda$. Since all the rules in
$\mathfrak{R}$ are $\mathtt{P}$-rules, necessarily
$\mathcal{V}(\lambda)\setminus\mathcal{V}(p(x,\bm{z}))\subseteq\\{u_{1},\dots,u_{k}\\}$.
Let $\mathfrak{s}^{\prime}$ be the store defined by:
$\mathfrak{s}^{\prime}(y)=\mathfrak{s}(\sigma(y))$, for all $y\in\mathcal{V}$.
By the application condition of the rule, $\mathit{dom}(\sigma)$ is a subset
of $\\{u_{1},\dots,u_{k}\\}\setminus(\left\\{x\right\\}\cup\bm{z})$, hence
$\mathfrak{s}^{\prime}$ coincides with $\mathfrak{s}$ on all variables in
$x,\bm{z}$. We deduce that $\mathfrak{s}^{\prime}$ is an associate of
$\mathfrak{s}$ w.r.t.
$\\{u_{1},\dots,u_{k}\\}\setminus\mathcal{V}(p(x,\bm{z}))$. We have
${(\mathfrak{s},\mathfrak{h}_{1})\models_{\mathfrak{R}}x\mapsto(y_{1},\dots,y_{k})}$,
with $\sigma(x)=x$ and $\sigma(u_{i})=y_{i}$, thus
$(\mathfrak{s},\mathfrak{h}_{1})\models_{\mathfrak{R}}(x\mapsto(u_{1},\dots,u_{k}))\sigma$.
Moreover, we also have
$(\mathfrak{s},\mathfrak{h}_{2})\models_{\mathfrak{R}}\psi^{\prime}\sigma$,
hence
$(\mathfrak{s},\mathfrak{h}_{1}\uplus\mathfrak{h}_{2})\models_{\mathfrak{R}}\lambda\sigma$.
By Proposition 2.7, we get
$(\mathfrak{s}^{\prime},\mathfrak{h}_{1}\uplus\mathfrak{h}_{2})\models_{\mathfrak{R}}\lambda$.
Since $p(x,\bm{z})\Leftarrow_{\mathfrak{R}}\lambda$, we deduce that
$(\mathfrak{s},\mathfrak{h}_{1}\uplus\mathfrak{h}_{2})\models_{\mathfrak{R}}p(x,\bm{z})$,
hence $(\mathfrak{s},\mathfrak{h})\models_{\mathfrak{R}}p(x,\bm{z})*\psi$.
Thus
$(\mathfrak{s},\mathfrak{h})\models_{\mathfrak{R}}(p(x,\bm{z})*\psi)\curlywedge\zeta$,
which contradicts our hypothesis.
Conversely, assume that
$(x\mapsto(y_{1},\dots,y_{k})*\phi)\curlywedge\xi\vdash_{\mathfrak{R}}^{V}(p(x,\bm{z})*\psi)\curlywedge\zeta$
is valid, and let $(\mathfrak{s},\mathfrak{h})$ be an $\mathfrak{R}$-model of
$(x\mapsto(y_{1},\dots,y_{k})*\phi)\curlywedge\xi$ such that
$\mathfrak{s}(V)\cap\mathit{dom}(\mathfrak{h})=\emptyset$ and $\mathfrak{s}$
is injective on $V$. We deduce that
$(\mathfrak{s},\mathfrak{h})\models_{\mathfrak{R}}(p(x,\bm{z})*\psi)\curlywedge\zeta$,
thus $\mathfrak{s}\models\zeta$, and there exist disjoint heaps
$\mathfrak{h}_{1},\mathfrak{h}_{2}$ such that
$(\mathfrak{s},\mathfrak{h}_{1})\models_{\mathfrak{R}}(p(x,\bm{z})$ and
$(\mathfrak{s},\mathfrak{h}_{2})\models_{\mathfrak{R}}\psi$. This entails that
there exists a symbolic heap $\lambda$ and a associate $\mathfrak{s}^{\prime}$
of $\mathfrak{s}$ w.r.t.
$\mathcal{V}(\lambda)\setminus\mathcal{V}(p(x,\bm{z}))$ such that
$p(x,\bm{z})\Leftarrow_{\mathfrak{R}}\lambda$, and
$(\mathfrak{s}^{\prime},\mathfrak{h}_{1})\models_{\mathfrak{R}}\lambda$. Since
the rules in $\mathfrak{R}$ are $\mathtt{P}$-rules, $\lambda$ is of the form
$(x\mapsto(v_{1},\dots,v_{m})*\psi^{\prime\prime})\curlywedge\zeta^{\prime\prime}$.
Moreover, it is clear that
$\mathfrak{h}(\mathfrak{s}(x))=(\mathfrak{s}(y_{1}),\dots,\mathfrak{s}(y_{k}))$,
so that $m=k$ and $\mathfrak{s}^{\prime}(v_{i})=\mathfrak{s}(y_{i})$, for all
$i=1,\dots,k$. Let $\sigma^{\prime}$ be the substitution mapping every
variable in $\\{v_{1},\dots,v_{k}\\}$ not occurring in $x,\bm{z}$ to the first
variable $y_{j}$ such that $\mathfrak{s}^{\prime}(v_{i})=\mathfrak{s}(y_{j})$.
By definition, we have
$\mathfrak{s}^{\prime}=\mathfrak{s}\circ\sigma^{\prime}$, thus by Proposition
2.7, we get
$(\mathfrak{s},\mathfrak{h}_{1})\models_{\mathfrak{R}}\lambda\sigma^{\prime}$.
Assume for the sake of contradiction that the inductive rule used to derive
$\lambda$ is distinct from the one used to derive the formula
$\lambda^{\prime}=(x\mapsto(u_{1},\dots,u_{k})*\psi^{\prime})\curlywedge\zeta^{\prime}$
in the application condition of the rule. We may assume by renaming that
$\mathcal{V}(\lambda^{\prime})\cap\mathcal{V}(\lambda)\subseteq\mathcal{V}(p(x,\bm{z})$.
Let $\mathfrak{s}^{\prime\prime}$ be a store coinciding with
$\mathfrak{s}^{\prime}$ on all constants and on all variables in
$\mathcal{V}(\lambda)$, and such that, for all variables
$y\in\mathcal{V}(\lambda^{\prime})\setminus\mathcal{V}(p(x,\bm{z}))$,
$\mathfrak{s}^{\prime\prime}(y)=\mathfrak{s}(\sigma(y))$. Since
$\mathfrak{s}^{\prime}\models\zeta^{\prime\prime}$ we have
$\mathfrak{s}^{\prime\prime}\models\zeta^{\prime\prime}$. By the application
condition of the rule $\sigma(u_{i})=y_{i}$, thus
$\mathfrak{s}^{\prime\prime}\models u_{i}\approx v_{i}$, for all
$i=1,\dots,k$. Since $\mathfrak{s}\models\xi$ and
$\xi\models\zeta^{\prime}\sigma$ (still by the application condition of the
rule), we get $\mathfrak{s}\models\zeta^{\prime}\sigma$, and by Proposition
2.7 we deduce that $\mathfrak{s}\models\zeta^{\prime}$. Thus
$(u_{1},\dots,u_{k})\approx(v_{1},\dots,v_{k})\wedge\zeta^{\prime}\wedge\zeta^{\prime\prime}$
is satisfiable, which contradicts the fact that $\mathfrak{R}$ is
deterministic.
This entails that the rules applied to derive $\lambda$ and $\lambda^{\prime}$
are the same, and by renaming we may assume in this case that
$(u_{1},\dots,u_{k})=(v_{1},\dots,v_{k})$ (which entails that
${(x\mapsto(v_{1},\dots,v_{k}))\sigma}={x\mapsto(y_{1},\dots,y_{k})}$), with
$\psi^{\prime}=\psi^{\prime\prime}$ and $\zeta^{\prime}=\zeta^{\prime\prime}$.
Using the fact that $\sigma(u_{i})=y_{i}$ and
$\mathfrak{s}^{\prime}(v_{i})=\mathfrak{s}(y_{i})$, for all $i=1,\dots,k$, it
is easy to check that $\mathfrak{s}^{\prime}(y)=\mathfrak{s}(\sigma(y))$, for
all variables $y$. Since
$(\mathfrak{s}^{\prime},\mathfrak{h}_{2})\models_{\mathfrak{R}}\lambda$ we get
by Proposition 2.7,
$(\mathfrak{s},\mathfrak{h}_{2})\models_{\mathfrak{R}}\lambda\sigma$, i.e.,
$(\mathfrak{s},\mathfrak{h}_{2})\models_{\mathfrak{R}}(x\mapsto(y_{1},\dots,y_{k})*\psi^{\prime}\sigma)$.
Therefore
$(\mathfrak{s},\mathfrak{h})\models_{\mathfrak{R}}(x\mapsto(y_{1},\dots,y_{k})*\psi^{\prime}\sigma*\psi)\curlywedge\zeta$.
Rule S is sound but in contrast with the other rules, it is not invertible in
general (intuitively, there is no guarantee that the decomposition of the
left-hand side of the sequent corresponds to that of the right-hand side).
###### Example 4.28
Consider the (valid) sequent
$x\mapsto(y)*y\mapsto(x)\vdash_{\mathfrak{R}}^{\emptyset}p(x,y)*p(y,x)$ with
the rule $p(u,v)\Leftarrow u\mapsto(v)$. Rule S applies, yielding the valid
premises $x\mapsto(y)\vdash_{\mathfrak{R}}^{\emptyset}p(x,y)$ and
$y\mapsto(x)\vdash_{\mathfrak{R}}^{\emptyset}p(y,x)$. However, since the rules
apply modulo commutativity of $*$ we may also get the premises:
$x\mapsto(y)\vdash_{\mathfrak{R}}^{\emptyset}p(y,x)$ and
$y\mapsto(x)\vdash_{\mathfrak{R}}^{\emptyset}p(x,y)$ which are not valid.
###### Lemma 4.29
Rule S is sound. More specifically, if $(\mathfrak{s},\mathfrak{h})$ is a
counter-model of the conclusion, then one of the premises admits a counter-
model $(\mathfrak{s},\mathfrak{h}^{\prime})$, where $\mathfrak{h}^{\prime}$ is
a proper subheap of $\mathfrak{h}$.
###### Proof 4.30
Let $(\mathfrak{s},\mathfrak{h})$ be an $\mathfrak{R}$-model of
$(\phi_{1}*\phi_{2})\curlywedge(\xi_{1}\wedge\xi_{2})$, where
$\mathfrak{s}(V)\cap\mathit{dom}(\mathfrak{h})=\emptyset$ and $\mathfrak{s}$
is injective on $V$. Assume that the premises admit no counter-model of the
form $(\mathfrak{s},\mathfrak{h}^{\prime})$ with
$\mathfrak{h}^{\prime}\subset\mathfrak{h}$. By definition, there exist
disjoint heaps $\mathfrak{h}_{1},\mathfrak{h}_{2}$ such that
$\mathfrak{h}=\mathfrak{h}_{1}\uplus\mathfrak{h}_{2}$, and
$(\mathfrak{s},\mathfrak{h}_{i})\models_{\mathfrak{R}}\phi_{i}$, for $i=1,2$.
Since $\phi_{i}\not=\mathit{emp}$, $\phi_{i}$ contains at least one predicate
atom, with a root $x_{i}$. By Lemma 4.2, necessarily
$\mathfrak{s}(x_{i})\in\mathit{dom}(\mathfrak{h}_{i})$, so $\mathfrak{h}_{i}$
is not empty and $\mathfrak{h}_{i}\subset\mathfrak{h}$ for $i=1,2$. Still by
Lemma 4.2,
$\mathfrak{s}(\mathit{alloc}(\phi_{i}))\subseteq\mathit{dom}(\mathfrak{h}_{i})$,
and since $\mathfrak{h}_{1}$ and $\mathfrak{h}_{2}$ are disjoint, we deduce
that
$\mathfrak{s}(\mathit{alloc}(\phi_{3-i}))\cap\mathit{dom}(\mathfrak{h}_{i})=\emptyset$
for $i=1,2$. Thus
$\mathfrak{s}(V\cup\mathit{alloc}(\phi_{3-i}))\cap\mathit{dom}(\mathfrak{h}_{i})=\emptyset$.
By Corollary 4.4 $\mathfrak{s}$ is injective on
$\mathit{alloc}(\phi_{1}*\phi_{2})$, hence since $\mathfrak{s}$ is injective
on $V$, we deduce that $\mathfrak{s}$ is injective on
$V\cup\mathit{alloc}(\phi_{3-i})$. Since $(\mathfrak{s},\mathfrak{h}_{i})$
cannot be a counter-model of the premises because
$\mathfrak{h}_{i}\subset\mathfrak{h}$, this entails that
$(\mathfrak{s},\mathfrak{h}_{i})\models_{\mathfrak{R}}\psi_{i}\curlywedge\zeta_{i}$
for $i=1,2$, thus
$(\mathfrak{s},\mathfrak{h})\models_{\mathfrak{R}}(\psi_{1}*\psi_{2})\curlywedge(\zeta_{1}\wedge\zeta_{2})$.
### 4.4 Axioms and Anti-Axioms
We define two sets of syntactic criteria on sequents that allow to quickly
conclude that such sequents are respectively valid or non-valid. This will be
useful to block the application of the inference rules in these cases. Axioms
(i.e., necessarily valid sequents) are defined as follows.
###### Definition 4.31
An axiom is a sequent that is of one of the following forms modulo AC:
1. 1.
$\phi\curlywedge(\xi\wedge\xi^{\prime})\vdash_{\mathfrak{R}}^{V}\phi\curlywedge\xi$;
2. 2.
$\phi\curlywedge(\xi\wedge x\not\approx x)\vdash_{\mathfrak{R}}^{V}\gamma$;
3. 3.
$\phi\curlywedge\xi\vdash_{\mathfrak{R}}^{V}\gamma$ where $\phi$ is heap-
unsatisfiable;
4. 4.
$\phi\curlywedge\xi\vdash_{\mathfrak{R}}^{V}\gamma$ where either
$\mathit{alloc}(\phi)\cap V\not=\emptyset$ or $V$ contains two occurrences of
the same variable.
Intuitively, a sequent is valid if the right-hand side is a trivial
consequence of the left-hand side, if the left-hand side is (trivially)
unsatisfiable, or if $V$ contains a variable that is allocated by the left-
hand side or two occurrences of the same variable (since by hypothesis
counter-models must be injective on $V$).
###### Lemma 4.32
All axioms are valid.
###### Proof 4.33
We consider each case separately (using the same notations as in Definition
4.31):
1. 1.
It is clear that every model of $\phi\curlywedge(\xi\wedge\xi^{\prime})$ is a
model of $\phi\curlywedge\xi$.
2. 2.
By definition, $\phi\curlywedge(\xi\wedge x\not\approx x)$ has no model, hence
$\phi\curlywedge(\xi\wedge x\not\approx x)\vdash_{\mathfrak{R}}^{V}\gamma$ has
no counter-model.
3. 3.
If $\phi$ is heap-unsatisfiable then $\mathit{alloc}(\phi)$ contains two
occurrences of the same variable, which by Corollary 4.4, entails that $\phi$
has no model. Thus $\phi\curlywedge\xi\vdash_{\mathfrak{R}}^{V}\gamma$ has no
counter-model.
4. 4.
Let $(\mathfrak{s},\mathfrak{h})$ be a counter-model of
$\phi\curlywedge\xi\vdash_{\mathfrak{R}}^{V}\gamma$. By definition
$\mathfrak{s}$ is injective on $V$ hence we cannot have
$\\{x,x\\}\subseteq_{m}V$. Also, by definition, if $x\in V$ then we cannot
have $\mathfrak{s}(x)\in\mathit{dom}(\mathfrak{h})$, and if
$x\in\mathit{alloc}(\phi)$ then if $x\in\mathit{alloc}(\phi)\cap V$ then
$\mathfrak{s}(x)\in\mathit{dom}(\mathfrak{h})$ by Lemma 4.2. We conclude that
it is impossible to have $x\in\mathit{alloc}(\phi)\cap V$ either.
We also introduce the notion of an anti-axiom, which is a sequent satisfying
some syntactic conditions that prevent it from being valid.
###### Definition 4.34
A sequent $\phi\curlywedge\xi\vdash_{\mathfrak{R}}^{V}\psi\curlywedge\zeta$ is
an anti-axiom if it is not an axiom, $\xi$ contains no equality, $\zeta=\top$
and one of the following conditions holds:
1. 1.
$\mathit{alloc}(\psi)\not\subseteq\mathit{alloc}(\phi)$;
2. 2.
$\psi=\mathit{emp}$ and $\phi\not=\mathit{emp}$.
3. 3.
There exists a variable
$x\in\mathit{alloc}(\phi)\setminus\mathit{alloc}(\psi)$, such that
$y\not\rightarrow_{\phi}^{*}x$ holds, for all $y\in\mathit{alloc}(\psi)$;
4. 4.
$V\cap(\mathcal{V}(\phi)\setminus\mathcal{V}(\psi))$ is not empty.
5. 5.
$\mathcal{V}_{\mathtt{loc}}(\phi)\setminus(\mathcal{V}(\psi)\cup\mathit{alloc}(\phi))$
is not empty.
We provide examples illustrating every case in Definition 4.34:
###### Example 4.35
The following sequents (where $p$ is some arbitrary predicate) are anti-
axioms:
$\begin{array}[]{llllcllll}1:&p(x,y)&\vdash_{\mathfrak{R}}^{\emptyset}&p(y,x)&&2:&p(x,y)&\vdash_{\mathfrak{R}}^{\emptyset}&\mathit{emp}\\\
3:&p(x,y)*p(z,y)&\vdash_{\mathfrak{R}}^{\emptyset}&q(x,y)&&4:&p(x,y)&\vdash_{\mathfrak{R}}^{\\{y\\}}&r(x)\\\
5:&p(x,y)&\vdash_{\mathfrak{R}}^{\emptyset}&r(x)\end{array}$
Intuitively, $1$ cannot be valid because there exist models of $p(x,y)$ in
which $y$ is not allocated whereas $y$ is allocated in all models of $p(y,x)$
Note that by Assumption 2.2, all predicates are productive, hence $p(x,y)$
admits at least one model. Furthermore, a predicate cannot allocate any of its
arguments other than the root, for instance rules of the form
$p(x,y)\Leftarrow x\mapsto(y)*p(y,x)$, indirectly allocating $y$, are not
allowed. $2$ cannot be valid because the models of $p(x,y)$ allocate at least
$x$. For $3$, assuming that all variables are associated with distinct
locations, one can construct a model of $p(x,y)*p(z,y)$ in which there is no
path from $x$ to $z$ and by Proposition 4.9 all locations occurring in the
heap of any model of $q(x,y)$ must be reachable from $x$. For $4$ and $5$, we
can construct a counter-model by considering any structure in which $y$ occurs
in the heap but is not allocated, and by Lemma 4.11, all the locations
occurring in the heap of any model of $r(x)$ must be allocated.
To show that all anti-axioms admit counter-models, we use the following lemma,
which will also play a key rôle in the completeness proof. It states that all
the formulas that are heap-satisfiable admit a model satisfying some
particular properties:
###### Lemma 4.36
Let $\phi$ be a spatial formula, containing a variable $x$ of sort
$\mathtt{loc}$. Let $\mathfrak{s}$ be a store that is injective on
$\mathit{alloc}(\phi)$. Let $U$ be an infinite subset of
$\mathfrak{U}_{\mathtt{loc}}$ such that
${U\cap\mathfrak{s}(\mathcal{V}_{\mathtt{loc}})}=\emptyset$. If $\phi$ is
heap-satisfiable, then it admits an $\mathfrak{R}$-model of the form
$(\mathfrak{s},\mathfrak{h})$, where
$\mathfrak{s}(x)\in\mathit{ref}(\mathfrak{h})$, the set
$\mathit{dom}(\mathfrak{h})$ is a subset of
$U\cup\mathfrak{s}(\mathit{alloc}(\phi))$ and
${\mathit{ref}(\mathfrak{h})\subseteq U\cup\mathfrak{s}(\mathcal{V}(\phi))}$.
Moreover, if $\mathfrak{s}$ is injective then $(\mathfrak{s},\mathfrak{h})$ is
a $\rightarrow$-compatible model of $\phi$.
###### Proof 4.37
The proof is by induction on the formulas. Note that we cannot have
$\phi=\mathit{emp}$ since $x\in\mathcal{V}(\phi)$.
* •
If $\phi=y_{0}\mapsto(y_{1},\dots,y_{n})$ then we set:
$\mathfrak{h}\overset{\text{\tiny\it
def}}{=}\\{(\mathfrak{s}(y_{0}),\dots,\mathfrak{s}(y_{n}))\\}$. By definition,
${(\mathfrak{s},\mathfrak{h})\models_{\mathfrak{R}}\phi}$. Moreover,
$x\in\\{y_{0},\dots,y_{n}\\}$, and
$\mathit{ref}(\mathfrak{h})=\\{\mathfrak{s}(y_{i})\mid i=0,\dots,n\text{\ and\
}y_{i}\in\mathcal{V}_{\mathtt{loc}}\\}$, hence
$\mathfrak{s}(x)\in\mathit{ref}(\mathfrak{h})$ and
$\mathit{ref}(\mathfrak{h})\subseteq\mathfrak{s}(\mathcal{V}(\phi))$.
Furthermore, $\mathit{dom}(\mathfrak{h})=\\{\mathfrak{s}(y_{0})\\}$ and
$\mathit{alloc}(\phi)=\\{y_{0}\\}$ thus
$\mathit{dom}(\mathfrak{h})\subseteq\mathfrak{s}(\mathit{alloc}(\phi))$.
Finally, assume that $\mathfrak{s}$ is injective. We have by definition
$\rightarrow_{\mathfrak{h}}=\\{(\mathfrak{s}(y_{0}),\mathfrak{s}(y_{i}))\mid
i=1,\dots,n\\}$, thus if
$\mathfrak{s}(u)\rightarrow_{\mathfrak{h}}^{*}\mathfrak{s}(v)$ for
$u,v\in\mathcal{V}(\lambda)$ then we must have either
$\mathfrak{s}(u)=\mathfrak{s}(v)$, so that $u=v$ because $\mathfrak{s}$ is
injective, in which case it is clear that $u\rightarrow_{\phi}^{*}v$; or
$\mathfrak{s}(u)=\mathfrak{s}(y_{0})$ and
$\mathfrak{s}(v)=\mathfrak{s}(y_{i})$ for some $i=1,\dots,n$. Since
$\mathfrak{s}$ is injective this entails that $u=y_{0}$, $v=y_{i}$, thus
$u\rightarrow_{\phi}v$ by definition of $\rightarrow_{\phi}$.
* •
Assume that $\phi=p(y_{0},\dots,y_{n})$ is a predicate atom. Then, since by
Assumption 2.2 every predicate symbol is productive, there exists a symbolic
heap $\gamma$ such that $\phi\Leftarrow_{\mathfrak{R}}\gamma$. If $x=y_{0}$
then since the rules in $\mathfrak{R}$ are $\mathtt{P}$-rules, $\gamma$
contains a points-to-atom with root $x$. Otherwise, by Assumption 2.16,
$x=y_{i}$ for some $i\in\mathit{out}_{\mathfrak{R}}(p)$, hence there exists a
rule application $\phi\Leftarrow_{\mathfrak{R}}\gamma$ such that $x$ occurs in
some predicate atom in $\gamma$. Thus in both cases we may assume that $x$
occurs in a spatial atom in $\gamma$. Note that $\gamma$ must be heap-
satisfiable, since all considered rules are $\mathtt{P}$-rules and by
Definition 2.9 the roots of the predicate symbols in $\gamma$ are pairwise
distinct existential variables, thus also distinct from the root $y_{0}$ of
the points-to atom. Furthermore, $\gamma$ is of the form
$\psi\curlywedge\zeta$, where $x\in\mathcal{V}(\psi)$ and $\zeta$ is a
conjunction of disequations $u\not\approx v$, with $u\not=v$.
Let $\mathfrak{s}^{\prime}$ be an associate of $\mathfrak{s}$ w.r.t.
$\mathcal{V}(\gamma)\setminus\mathcal{V}(\phi)$ mapping the variables in
$\mathcal{V}(\gamma)\setminus\mathcal{V}(\phi)$ to pairwise distinct locations
in $U$. Since by hypothesis $U\cap\mathfrak{s}(\mathcal{V})=\emptyset$,
$\mathfrak{s}^{\prime}$ is injective on any set of the form
$E\cup(\mathcal{V}(\gamma)\setminus\mathcal{V}(\phi))$ when $\mathfrak{s}$ is
injective on $E$. Let $U^{\prime}\overset{\text{\tiny\it
def}}{=}U\setminus\mathfrak{s}^{\prime}(\mathcal{V}(\gamma)\setminus\mathcal{V}(\phi))$.
By the induction hypothesis, there exists a heap $\mathfrak{h}$ such that
$(\mathfrak{s}^{\prime},\mathfrak{h})\models_{\mathfrak{R}}\psi$, with
$\mathfrak{s}^{\prime}(x)\in\mathit{ref}(\mathfrak{h})$,
${\mathit{dom}(\mathfrak{h})\subseteq
U^{\prime}\cup\mathfrak{s}^{\prime}(\mathit{alloc}(\psi))}$ and
$\mathit{ref}(\mathfrak{h})\subseteq
U^{\prime}\cup\mathfrak{s}^{\prime}(\mathcal{V}(\psi))$. Now if $\mathfrak{s}$
is injective then (as $\mathfrak{s}^{\prime}$ is also injective in this case),
$(\mathfrak{s}^{\prime},\mathfrak{h})$ is a $\rightarrow$-compatible
$\mathfrak{R}$-model of $\psi$. We show that $(\mathfrak{s},\mathfrak{h})$
fulfills all the properties of the lemma.
* –
Since $\mathfrak{s}^{\prime}$ maps the variables in
$\mathcal{V}(\gamma)\setminus\mathcal{V}(\phi)$ to pairwise distinct locations
in $U$ and ${\mathfrak{s}(\mathcal{V})\cap U}$ is $\emptyset$, necessarily
$\mathfrak{s}^{\prime}\models_{\mathfrak{R}}\zeta$, thus
$(\mathfrak{s}^{\prime},\mathfrak{h})\models_{\mathfrak{R}}\gamma$ which
entails that $(\mathfrak{s},\mathfrak{h})\models_{\mathfrak{R}}\phi$. We also
have $\mathfrak{s}(x)=\mathfrak{s}^{\prime}(x)\in\mathit{ref}(\mathfrak{h})$.
* –
Let $\ell\in\mathit{ref}(\mathfrak{h})$. We show that $\ell\in
U\cup\mathfrak{s}(\mathcal{V}(\phi))$. By the induction hypothesis, we have
$\ell\in U^{\prime}\cup\mathfrak{s}^{\prime}(\mathcal{V}(\psi))$. If $\ell\in
U^{\prime}\subseteq U$ then the proof is completed, otherwise we have
$\ell=\mathfrak{s}^{\prime}(y)$ for some $y\in\mathcal{V}(\psi)$. If
$y\in\mathcal{V}(\phi)$ then $\mathfrak{s}(y)=\mathfrak{s}^{\prime}(y)$ thus
$\ell\in\mathfrak{s}(\mathcal{V}(\phi))$. Otherwise, we must have
$y\in\mathcal{V}(\gamma)\setminus\mathcal{V}(\phi)$, thus
$\mathfrak{s}^{\prime}(y)\in U$ by definition of $\mathfrak{s}^{\prime}$ and
the result holds.
* –
Let $\ell\in\mathit{dom}(\mathfrak{h})$, we show that $\ell\in
U\cup\mathfrak{s}(\mathit{alloc}(\phi))$. By the induction hypothesis we have
$\ell\in U^{\prime}\cup\mathfrak{s}^{\prime}(\mathit{alloc}(\psi))$. If
$\ell\in U^{\prime}\subseteq U$ then the proof is completed. Otherwise,
$\ell=\mathfrak{s}^{\prime}(y)$ with $y\in\mathit{alloc}(\psi)$. Since the
rules in $\mathfrak{R}$ are $\mathtt{P}$-rules, $y$ is either the root $y_{0}$
of $\phi$, in which case we have $y\in\mathit{alloc}(\phi)$ and
$\mathfrak{s}(y)=\mathfrak{s}^{\prime}(y)$, thus the result holds; or $y$
occurs in $\mathcal{V}(\gamma)\setminus\mathcal{V}(\phi)$, in which that we
have $\mathfrak{s}^{\prime}(y)\in U$, by definition of
$\mathfrak{s}^{\prime}$.
* –
There remains to show that $(\mathfrak{s},\mathfrak{h})$ is a
$\rightarrow$-compatible $\mathfrak{R}$-model of $\phi$, in the case where
$\mathfrak{s}$ is injective. Assume that
$\mathfrak{s}(u)\rightarrow_{\mathfrak{h}}^{*}\mathfrak{s}(v)$. If
$\mathfrak{s}(u)=\mathfrak{s}(v)$ then $u=v$ by injectivity of $\mathfrak{s}$,
hence $u\rightarrow_{\mathfrak{h}}^{*}v$. Otherwise, we must have
$\\{\mathfrak{s}(u),\mathfrak{s}(v)\\}\subseteq\mathit{ref}(\mathfrak{h})\subseteq
U^{\prime}\cup\mathfrak{s}^{\prime}(\mathcal{V}(\psi))$, and since
$U^{\prime}\subseteq U$ and $U\cap\mathfrak{s}(\mathcal{V})=\emptyset$, we
have
$\\{\mathfrak{s}(u),\mathfrak{s}(v)\\}\subseteq\mathfrak{s}^{\prime}(\mathcal{V}(\psi))$.
We also have
$\mathfrak{s}^{\prime}(\mathcal{V}(\gamma)\setminus\mathcal{V}(\phi))\subseteq
U$, which entails that
$\\{\mathfrak{s}(u),\mathfrak{s}(v)\\}\subseteq\mathfrak{s}^{\prime}(\mathcal{V}(\phi))$.
Since $\mathfrak{s}^{\prime}$ is an associate of $\mathfrak{s}$ w.r.t.
$\mathcal{V}(\gamma)\setminus\mathcal{V}(\phi)$, $\mathfrak{s}$ and
$\mathfrak{s}^{\prime}$ coincide on all variables in $\mathcal{V}(\phi)$ and
we deduce that
$\\{\mathfrak{s}(u),\mathfrak{s}(v)\\}\subseteq\mathfrak{s}(\mathcal{V}(\phi))$.
Because $\mathfrak{s}$ is injective, this entails that
$u,v\in\mathcal{V}(\phi)$, so that $\mathfrak{s}(u)=\mathfrak{s}^{\prime}(u)$
and $\mathfrak{s}(v)=\mathfrak{s}^{\prime}(v)$. By hypothesis
$(\mathfrak{s}^{\prime},\mathfrak{h})$ is a $\rightarrow$-compatible
$\mathfrak{R}$-model of $\psi$, and we deduce that $u\rightarrow_{\psi}^{*}v$.
Since $u\not=v$, necessarily $u\in\mathit{alloc}(\psi)$ (by definition of
$\rightarrow_{\psi}$), and since the rules in $\mathfrak{R}$ are
$\mathtt{P}$-rules, and $u\in\mathcal{V}(\phi)$, this entails that
$u=\mathit{roots}(\phi)$. Since $v\in\mathcal{V}(\phi)$, by Assumption 2.16 we
have $u\rightarrow_{\phi}^{*}v$.
* •
Assume that $\phi=\phi_{1}*\phi_{2}$, with $\phi_{i}\not=\mathit{emp}$. Let
$U_{1},U_{2}$ be disjoint infinite subsets of $U$. Let $\\{x_{1},x_{2}\\}$ be
some arbitrary chosen variables such that $x_{i}\in\mathcal{V}(\phi_{i})$ and
$x\in\\{x_{1},x_{2}\\}$ (it is easy to check that such a pair of variables
always exists). By the induction hypothesis, there exist heaps
$\mathfrak{h}_{i}$ such that $(\mathfrak{s},\mathfrak{h}_{i})\models\phi_{i}$
where $\mathfrak{s}(x_{i})\in\mathit{ref}(\mathfrak{h}_{i})$,
${\mathit{dom}(\mathfrak{h}_{i})\subseteq
U_{i}\cup\mathfrak{s}(\mathit{alloc}(\phi_{i}))}$ and
$\mathit{ref}(\mathfrak{h}_{i})\subseteq
U_{i}\cup\mathfrak{s}(\mathcal{V}(\phi_{i}))$. Moreover, if $\mathfrak{s}$ is
injective, then $(\mathfrak{s},\mathfrak{h}_{i})$ is an
$\rightarrow$-compatible $\mathfrak{R}$-model of $\phi_{i}$. We first show
that $\mathfrak{h}_{1}$ and $\mathfrak{h}_{2}$ are disjoint. Assume for the
sake of contradiction that
$\ell\in\mathit{dom}(\mathfrak{h}_{1})\cap\mathit{dom}(\mathfrak{h}_{2})$.
Since $U_{1}\cap U_{2}=\emptyset$, necessarily $\ell=\mathfrak{s}(y_{i})$ (for
$i=1,2$), with $y_{i}\in\mathit{alloc}(\phi_{i})$. Since $\mathfrak{s}$ is
injective on $\mathit{alloc}(\phi)$, we deduce that $y_{1}=y_{2}$. We have
$\\{y_{1},y_{2}\\}\subseteq_{m}\mathit{alloc}(\phi)$, hence $\phi$ is heap-
unsatisfiable, which contradicts the hypotheses of the lemma. Thus
$\mathfrak{h}_{1}$ and $\mathfrak{h}_{2}$ are disjoint.
Let $\mathfrak{h}=\mathfrak{h}_{1}\uplus\mathfrak{h}_{2}$. We have
$\mathfrak{s}(x)\in\\{\mathfrak{s}(x_{1}),\mathfrak{s}(x_{2})\\}\subseteq\mathit{ref}(\mathfrak{h}_{1})\cup\mathit{ref}(\mathfrak{h}_{2})=\mathit{ref}(\mathfrak{h})$.
Moreover,
$\mathit{dom}(\mathfrak{h})=\mathit{dom}(\mathfrak{h}_{1})\cup\mathit{dom}(\mathfrak{h}_{2})\subseteq
U_{1}\cup
U_{2}\cup\mathfrak{s}(\mathit{alloc}(\phi_{1}))\cup\mathfrak{s}(\mathit{alloc}(\phi_{2})){\subseteq
U\cup\mathfrak{s}(\mathit{alloc}(\phi))}$, and
$\mathit{ref}(\mathfrak{h})=\mathit{ref}(\mathfrak{h}_{1})\cup\mathit{ref}(\mathfrak{h}_{2})\subseteq
U_{1}\cup
U_{2}\cup\mathfrak{s}(\mathcal{V}(\phi_{i}))\cup\mathfrak{s}(\mathcal{V}(\phi_{2})){\subseteq
U\cup\mathfrak{s}(\mathcal{V}(\phi))}$. Furthermore,
$(\mathfrak{s},\mathfrak{h})\models_{\mathfrak{R}}\phi_{1}*\phi_{2}=\phi$.
There only remains to prove that $(\mathfrak{s},\mathfrak{h})$ is a
$\rightarrow$-compatible $\mathfrak{R}$-model of $\phi$ when $\mathfrak{s}$ is
injective. Assume that this is not the case, and let $u,v$ be variables such
that $\mathfrak{s}(u)\rightarrow_{\mathfrak{h}}^{*}\mathfrak{s}(v)$ and
$u\not\rightarrow_{\phi}^{*}v$. This entails that $u\not=v$. By definition,
there exist $\ell_{0},\dots,\ell_{m}$ such that $\ell_{0}=\mathfrak{s}(u)$,
$\ell_{m}=\mathfrak{s}(v)$, and $\forall
i=1,\dots,m,\,\ell_{i-1}\rightarrow_{\mathfrak{h}}\ell_{i}$. We assume,
w.l.o.g., that $m$ is miminal, i.e., that there is no sequence
$\ell_{0}^{\prime},\dots,\ell_{k}^{\prime}$ and no variables $x_{0},x_{k}$
such that $k<m$, $\ell_{0}^{\prime}=\mathfrak{s}(x_{0})$,
$\ell_{k}^{\prime}=\mathfrak{s}(x_{k})$ and
$x_{0}\not\rightarrow_{\phi}^{*}x_{k}$. We may also assume, by symmetry, that
$\ell_{0}\in\mathit{dom}(\mathfrak{h}_{1})$. If all the locations
$\ell_{1},\dots,\ell_{m-1}$ occur in $\mathit{dom}(\mathfrak{h}_{1})$ then
$\mathfrak{s}(u)\rightarrow_{\mathfrak{h}_{1}}^{*}\mathfrak{s}(v)$, thus
$u\rightarrow_{\phi_{i}}^{*}v$ because $(\mathfrak{s},\mathfrak{h}_{i})$ is an
|
# Exponential asymptotics and the generation of free-surface flows by
submerged point vortices
Josh Shelton<EMAIL_ADDRESS>Philippe H. Trinh<EMAIL_ADDRESS>Department of Mathematical Sciences, University of Bath, Bath BA2 7AY, UK
( [Draft])
###### Abstract
There has been significant recent interest in the study of water waves coupled
with non-zero vorticity. Here, we derive analytical approximations for the
exponentially-small free-surface waves generated by one or several submerged
point vortices when driven at low Froude numbers. The vortices are fixed in
place, and a boundary-integral formulation in the arclength along the surface
allows the study of nonlinear waves and strong point vortices. We demonstrate
that for a single point vortex, techniques in exponential asymptotics
prescribe the formation of waves in connection with the presence of Stokes
lines originating from the vortex. When multiple point vortices are placed
within the fluid, trapped waves may occur, which are confined to lie between
the vortices. We also demonstrate that for the two-vortex problem, the
phenomena of trapped waves occurs for a countably infinite set of values of
the Froude number. This work will form a basis for other asymptotic
investigations of wave-structure interactions where vorticity plays a key role
in the formation of surface waves.
## 1 Introduction
In this paper we study the nonlinear flow of an ideal fluid past a submerged
point vortex. The two-dimensional inviscid and incompressible fluid of
infinite depth is assumed to be irrotational everywhere, with the exception of
the point vortices themselves. For a flow in the complex
$z=x+\mathrm{i}y$-plane, with a vortex at $z=z^{*}$, the complex potential
behaves as
$f=\phi+\mathrm{i}\psi\sim cz-\frac{\mathrm{i}\Gamma}{2\pi}\log{(z-z^{*})},$
(1)
where $\Gamma$ is the circulation of the vortex, and the background flow is of
speed $c$. The non-dimensional system is then characterised by two key
parameters: $\Gamma_{\text{c}}=\Gamma/(cH)$, relating vortex strength(s) to
inertial effects, and the Froude number, $F=c/\sqrt{gH}$, relating inertial
effects to gravitational effects. Here, $H$ is the depth of the point vortex
and $g$ is the constant acceleration due to gravity.
The study of such vortex-driven potential flows is complicated by the
following fact. Typically, the formulation of the free surface system would
involve the inversion of the velocity potential and streamfunction $\phi(x,y)$
and $\psi(x,y)$, which are both a function of the domain $x$ and $y$. In
expressing these instead as $x(\phi,\psi)$ and $y(\phi,\psi)$, we note that
the free surface is a streamline along which $\psi$ takes constant values, and
so $x(\phi)$ and $y(\phi)$ represent the free surface. However, near the point
vortex, the local behaviour (1), can not be inverted analytically to give
$z(f)$. This motivated the work of Forbes (1985), who re-formulated the
boundary-integral formulation in terms of a free-surface arclength, $s$, and a
more complex set of governing equations results.
The imposition of a uniform stream as $x\to-\infty$ results in the generation
of downstream free-surface waves, as shown in figure 1(a). As hinted in the
preliminary numerical investigations of Forbes (1985), the wave amplitude
tends to zero as $F\to 0$. In this work, we confirm this behaviour and
demonstrate, both numerically and analytically, that the amplitude is
exponentially-small in the low-Froude limit. For instance, the amplitude
versus $1/F^{2}$ graph shown in figure 2 demonstrates the fit between our
asymptotic predictions of §3 and numerical results of §4. We note that this
theory is nonlinear in the vortex strength, $\Gamma_{\text{c}}$, and the
assumption of small $\Gamma_{\text{c}}$ need not apply.
The purpose of this paper is to thus characterise the formation of water waves
using the framework of exponential asymptotics. We show that these
exponentially-small waves smoothly switch-on as the fluid passes beyond the
vortex, resulting in oscillations as $x\to\infty$ in the far field. When two
submerged vortices are considered, the waves switched-on due to each of the
vortices may be out of phase with one another and cancel for certain values of
the Froude number. This yields trapped waves between the vortices, and a free
surface whose derivative decays to zero as $x\to\infty$. A trapped wave
solution is depicted in figure 1(b). This phenomenon of trapped waves has
previously been studied for obstructions both within the fluid, and for flows
of finite depth past lower topography. For instance, both Gazdar (1973) and
Vanden-Broeck & Tuck (1985) detected these numerically for flows over a
specified lower topography. More recent works, such as those by Dias & Vanden-
Broeck (2004), Hocking et al. (2013), and Holmes et al. (2013), have focused
on detecting parameter values for which these trapped wave solutions occur in
various formulations.
Figure 1: The two physical regimes of underlying point vortices considered
within this paper are shown. In (a), a single point vortex with circulation
$\Gamma$ is placed within the fluid. In (b), two point vortices, each with
circulation $\Gamma$, are located at the same depth within the fluid. These
solutions have been computed using the numerical scheme detailed in §4. Figure
2: The amplitude, $\bar{y}$, of the free-surface waves is shown for
$\log(\bar{y})$ vs $1/F^{2}$ for the analytical (line) and numerical (dots)
solutions of §3 and §4. These have a fixed value of the nondimensional vortex
strength, $\Gamma_{\text{c}}=0.25$. The graph confirms exponential smallness
of the waves. The solid line has a gradient of $\approx 0.7395$, computed
using the exponential asymptotic theory of §4.
The work in this paper provides a first step towards extending many of the
existing ideas and techniques of exponential asymptotics, previously developed
for purely gravity- or capillary-driven waves (e.g. Chapman & Vanden-Broeck
2002, 2006) to wave phenomena with vortices. As noted above, because the
governing equations require an alternative formulation (originally developed
by Miksis et al. 1981) the asymptotic formulation we present can be extended
to other wave-structure interactions where the more general arc-length
formulation of the water-wave equations is required. In addition, there has
been significant recent interest in the study of water-wave phenomena with
dominant vorticity effects, and we reference the recent extensive survey by
Haziot et al. (2022) and references therein. The exponential-asymptotic
techniques developed in this work can also be extended to situations where
capillary ripples are forced on the surface of steep vortex-driven waves. The
leading order solution in these asymptotic regimes would then be known
analytically from the works of e.g. (Crowdy & Nelson, 2010; Crowdy & Roenby,
2014; Crowdy, 2022). These, and other exciting future directions, we shall
discuss in §6.
## 2 Mathematical formulation and outline
We consider the typical configurations shown in 1. Following Forbes (1985), in
nondimensional form, the system is formulated in terms of the arclength, $s$,
along the free surface, with unknown velocity potential $\phi=\phi(s)$, and
free-surface positions, $(x(s),y(s))$. Then, the governing equations are given
by Bernoulli’s equation, an arclength relation between $x$ and $y$, and a
boundary-integral equation.
For a single submerged point vortex at $(x,y)=(0,-1)$, the three equations are
$\displaystyle\frac{F^{2}}{2}\big{[}\phi^{\prime}(s)\big{]}^{2}+y(s)=\frac{F^{2}}{2},$
(2a)
$\displaystyle\big{[}x^{\prime}(s)\big{]}^{2}+\big{[}y^{\prime}(s)\big{]}^{2}=1,$
(2b)
$\displaystyle\phi^{\prime}(s)x^{\prime}(s)-1=\frac{\Gamma_{\text{c}}}{\pi}\frac{y(s)+1}{[x(s)]^{2}+[y(s)+1]^{2}}+\mathcal{I}[x,y,\phi^{\prime}].$
(2c)
In the above, two nondimensional parameters appear: the Froude number, $F$,
and the vortex strength, $\Gamma_{\text{c}}$, defined by
$F=\frac{c}{\sqrt{gH}}\quad\text{and}\quad\Gamma_{\text{c}}=\frac{\Gamma}{cH}.$
(3)
Here, $c$ is the speed of the fluid, $H$ is the depth of the submerged point
vortex, $g$ is the constant acceleration due to gravity, and $\Gamma$ is the
circulation of the point vortex. Furthermore, we have also introduced
$\mathcal{I}$ as the nonlinear principle-valued integral defined by
$\mathcal{I}[x,y,\phi^{\prime}]=\frac{1}{\pi}\mathchoice{{\vbox{\hbox{$\textstyle-$}}\kern-4.86108pt}}{{\vbox{\hbox{$\scriptstyle-$}}\kern-3.25pt}}{{\vbox{\hbox{$\scriptscriptstyle-$}}\kern-2.29166pt}}{{\vbox{\hbox{$\scriptscriptstyle-$}}\kern-1.875pt}}\\!\int_{-\infty}^{\infty}\frac{[\phi^{\prime}(t)-x^{\prime}(t)][y(t)-y(s)]+y^{\prime}(t)[x(t)-x(s)]}{[x(t)-x(s)]^{2}+[y(t)-y(s)]^{2}}\,\mathrm{d}t.$
(4)
When the configuration with two point vortices is considered in §3.5, the
boundary-integral equation (2c) will need to be modified to (29).
### 2.1 Analytic continuation
In the exponential asymptotic procedure of §3, we study the exponentially
small terms that display the Stokes phenomenon across Stokes lines of the
problem. These Stokes lines originate from singularities of the leading order
asymptotic solution, which are located in the analytic continuation of the
domain, the arclength $s$. The analytic continuation of the governing
equations (2a)-(2c) is studied in this section.
We now analytically continue the domain $s\mapsto\sigma$, where
$\sigma\in\mathbb{C}$. Bernoulli’s equation (2a) and the arclength relation
(2b) may be analytically continued in a straightforward manner, with all
dependence on $s$ replaced by the complex valued variable $\sigma$. The
analytic continuation of the boundary integral equation (2c) is more
complicated, due to the principal value integral $\mathcal{I}$ defined in (4).
The analytic continuation of this integral is given by
$\mathcal{I}[x,y,\phi^{\prime}]=\widehat{\mathcal{I}}[x,y,\phi^{\prime}]-a\mathrm{i}\phi^{\prime}(\sigma)y^{\prime}(\sigma),$
(5)
where $a=\pm 1$ denotes the direction of analytic continuation into
$\text{Im}[\sigma]>0$ or $\text{Im}[\sigma]<0$, respectively, and
$\widehat{\mathcal{I}}$ is the complex-valued integral. Equation (5) may be
verified by taking the limit of either $\text{Im}[\sigma]\to 0^{+}$, or
$\text{Im}[\sigma]\to 0^{-}$, which yields half of a residue contribution
associated with the singular point at $t=s$ of the integrand.
Substitution of (5) into (2c) then yields the analytically continued
equations, given by
$\displaystyle\frac{F^{2}}{2}\big{[}\phi^{\prime}(\sigma)\big{]}^{2}+y(\sigma)=\frac{F^{2}}{2},$
(6a)
$\displaystyle\big{[}x^{\prime}(\sigma)\big{]}^{2}+\big{[}y^{\prime}(\sigma)\big{]}^{2}=1,$
(6b)
$\displaystyle\phi^{\prime}(\sigma)x^{\prime}(\sigma)-1+a\mathrm{i}\phi^{\prime}(\sigma)y^{\prime}(\sigma)=\frac{\Gamma_{\text{c}}}{\pi}\frac{y(\sigma)+1}{[x(\sigma)]^{2}+[y(\sigma)+1]^{2}}+\widehat{\mathcal{I}}[x,y,\phi^{\prime}].$
(6c)
The analytic continuation for situations with multiple point vortices is
similarly done, with the only difference being the inclusion of additional
point vortices in (6c).
### 2.2 Outline of paper
In this work, we will consider the following two regimes depicted in figure 1:
1. (i)
A single submerged point vortex, which is the formulation originally
considered by Forbes (1985). Imposing free stream conditions as $x\to-\infty$
results in surface waves generated by the vortex. Their amplitude is
exponentially-small as $F\to 0$. This is the limit considered by Chapman &
Vanden-Broeck (2006) in the absence of vortical effects.
2. (ii)
Two submerged point vortices of the same circulation. For certain critical
values of the Froude number, $F$, the resultant waves are confined to lie
between the two vortices. The amplitude of these is also exponentially small
as $F\to 0$.
We begin in §3 by determining these exponentially small waves using the
techniques of exponential asymptotics. This relies of the optimal truncation
of an algebraic asymptotic series for small Froude number, $F$, and deriving
the connection of this to the Stokes phenomenon that acts on the exponentially
small waves. The case for two submerged point vortices is then studied in
§3.5, where we derive the critical values of the Froude number for which the
waves are trapped. Numerical solutions are computed in §4, where comparison
occurs with the exponential asymptotic predictions for the single vortex and
double vortex cases.
## 3 Exponential asymptotics
### 3.1 Early orders of the solution
We begin by considering the following asymptotic expansions, in powers of
$F^{2}$, for the solutions, which are given by
$x(\sigma)=\sum_{n=0}^{\infty}F^{2n}x_{n}(\sigma),\quad
y(\sigma)=\sum_{n=0}^{\infty}F^{2n}y_{n}(\sigma),\quad\phi^{\prime}(\sigma)=\sum_{n=0}^{\infty}F^{2n}\phi_{n}^{\prime}(\sigma).$
(7)
Substitution of expansions (7) into equations (6a)-(6c) yields at leading
order three equations for the unknowns $x_{0}$, $y_{0}$, and
$\phi_{0}^{\prime}$. The first of these, Bernoulli’s equation (6a), yields
$y_{0}(\sigma)=0$. This may be substituted into the second equation, (6b), to
find $(x_{0}^{\prime})^{2}=1$, for which we consider $x_{0}^{\prime}=1$
without any loss of generality. This may integrated to find $x_{0}=\sigma$,
where the constant of integration has been chosen to set the origin at
$x_{0}(0)=0$. Next, $\phi_{0}^{\prime}$ is determined from equation (6c).
Since $y_{0}=0$, the integral $\widehat{\mathcal{I}}$ does not enter the
leading order equation. This yields the leading order solutions as
$y_{0}(\sigma)=0,\qquad
x_{0}(\sigma)=\sigma,\qquad\phi^{\prime}_{0}(\sigma)=1+\frac{\Gamma_{\text{c}}}{\pi}\frac{1}{(1+\sigma^{2})}.$
(8)
Note that there is a singularity in $\phi_{0}^{\prime}$ above whenever
$\sigma^{2}=-1$. This corresponds to the point vortex within the fluid at
$\sigma=-\mathrm{i}$, as well as another singularity at $\sigma=\mathrm{i}$,
which will produce a complex-conjugate contribution to the exponentially-small
solution along the free surface.
Next at order $O(F^{2})$, $y_{1}$ is found explicitly from (6a). We then find
the equation $x_{1}^{\prime}=0$ from (6b), and $\phi_{1}$ is determined
explicitly from (6c). This yields
$\left.\quad\begin{aligned}
y_{1}(\sigma)&=\frac{1}{2}\Big{(}1-\big{(}\phi_{0}^{\prime}\big{)}^{2}\Big{)},\qquad
x_{1}(\sigma)=0,\\\
\phi^{\prime}_{1}(\sigma)&=-a\mathrm{i}\phi_{0}^{\prime}y_{1}^{\prime}+\frac{\Gamma_{\text{c}}(\sigma^{2}-1)}{\pi(1+\sigma^{2})^{2}}y_{1}+\widehat{\mathcal{I}}_{1}(\sigma),\end{aligned}\quad\right\\}$
(9)
where $\widehat{\mathcal{I}}_{1}$ is the $O(F^{2})$ component of the complex-
valued integral $\widehat{\mathcal{I}}$, originally defined along the real
axis in equation (4).
### 3.2 Late-term divergence
Our derivation of the exponentially-small terms and associated Stokes
phenomenon of §3.4 requires the knowledge of the late-terms of the solution
expansion (7), $x_{n}$, $y_{n}$, and $\phi_{n}^{\prime}$, as $n\to\infty$. We
begin by determining the $O(F^{2n})$ components of equations (6a)-(6c). The
late-terms of Bernoulli’s equation are given by
$y_{n}+\phi_{0}^{\prime}\phi_{n-1}^{\prime}+\phi_{1}^{\prime}\phi_{n-2}^{\prime}+\cdots=0,$
(10a) for the arclength relation we have
$x_{0}^{\prime}x_{n}^{\prime}+x_{1}^{\prime}x_{n-1}^{\prime}+\cdots+y_{1}^{\prime}y_{n-1}^{\prime}+y_{2}^{\prime}y_{n-2}^{\prime}+\cdots=0,\\\
$ (10b) and finally the boundary integral equation yields
$x_{0}^{\prime}\phi_{n}^{\prime}+x_{1}^{\prime}\phi_{n-1}^{\prime}+\phi_{0}^{\prime}x_{n}^{\prime}+\cdots+a\mathrm{i}\big{[}\phi_{0}^{\prime}y_{n}^{\prime}+\phi_{1}^{\prime}y_{n-1}^{\prime}+y_{1}^{\prime}\phi_{n-1}^{\prime}+\cdots\big{]}\\\
+\frac{\Gamma_{\text{c}}}{\pi}\bigg{[}\frac{y_{n}}{1+x_{0}^{2}}-\frac{2y_{n}}{(1+x_{0}^{2})^{2}}+\cdots\bigg{]}-\widehat{\mathcal{I}}_{n}(\sigma)=0.$
(10c)
In (10a)-(10c) above, only the terms that will appear at the first two orders
of $n$ as $n\to\infty$ have been included.
In (10c), the $O(F^{2n})$ component of the complex-valued integral,
$\widehat{\mathcal{I}}$ has been denoted by $\widehat{\mathcal{I}}_{n}$. The
dominant components of this integral, as $n\to\infty$, require the integration
of late-term asymptotic solutions that are either a function of the real
valued integration domain, such as $y_{n}(t)$, or a function of the complex
domain, such as $y_{n}(\sigma)$. The first of these, $y_{n}(t)$, is integrated
along the real-valued free surface, away from any singular behaviour. It is
thus subdominant to the other terms appearing in equation (10c). This is
analogous to the neglection of the late terms of the complex-valued Hilbert
transform in similar free-surface problems in exponential asymptotics [c.f.
Xie & Tanveer (2002), Chapman & Vanden-Broeck (2002), Chapman & Vanden-Broeck
(2006)]. All that remains is to integrate the components of
$\widehat{\mathcal{I}}_{n}$ that involve late-term solutions evaluated in the
complex-valued domain. Of these, only that involving $y_{n}(\sigma)$ appears
in the two leading orders, as $n\to\infty$, of equation (10c). This component
is given by
$\widehat{\mathcal{I}}_{n}\sim-\frac{y_{n}(\sigma)}{\pi}\int_{-\infty}^{\infty}\frac{\phi_{0}^{\prime}(t)-1}{(t-\sigma)^{2}}\mathrm{d}t=-\frac{\Gamma_{\text{c}}}{\pi}\frac{y_{n}(\sigma)}{(\sigma+a\mathrm{i})^{2}},$
(11)
for which the integral was evaluated by substituting for $\phi_{0}^{\prime}$
from equation (8). Note that integration of $y_{n}(\sigma)$ was not required
due to the lack of any dependence on the domain of integration, $t$.
Recall that the leading order solutions were singular at $\sigma=\pm i$. For
each of the three solution expansions, this singularity first appeared in
$\phi^{\prime}_{0}$, $y_{1}$, and $x_{2}$. Since successive terms in the
asymptotic expansion involve differentiation of previous terms (for instance,
equation (10a) for $y_{n}$ involves $\phi^{\prime}_{n-1}$, whose determination
in equation (10c) requires knowledge of $y_{n-1}^{\prime}$), the strength of
this singularity will grow as we proceed into the asymptotic series.
Furthermore, this growing singular behaviour will also lead to the divergence
of the late-term solutions as $n\to\infty$, which we capture analytically with
the factorial-over-power ansatzes of
$x_{n}\sim
X(\sigma)\frac{\Gamma(n+\alpha-1)}{[\chi(\sigma)]^{n+\alpha-1}},\quad
y_{n}\sim
Y(\sigma)\frac{\Gamma(n+\alpha)}{[\chi(\sigma)]^{n+\alpha}},\quad\phi_{n}\sim\Phi(\sigma)\frac{\Gamma(n+\alpha)}{[\chi(\sigma)]^{n+\alpha}}.$
(12)
Here, $\alpha$ is a constant, $\chi$ is the singulant function that will
capture the singular behaviour of the solution at $\sigma=\pm\mathrm{i}$, and
$X$, $Y$, and $\Phi$ are functional prefactors of the divergent solutions. It
can be seen from the dominant balance as $n\to\infty$ of equations (10a) and
(10b) that $x_{n+1}=O(y_{n})$ and $y_{n}=O(\phi_{n})$, which has motivated our
precise ordering in $n$ in the ansatzes (12).
Substitution of ansatzes (12) into the $O(F^{2n})$ equations (10a)-(10c)
yields at leading order in $n$ the three equations
$Y-\phi_{0}^{\prime}\chi^{\prime}\Phi=0,\qquad\chi^{\prime}\Big{(}X+y_{1}^{\prime}Y\Big{)}=0,\qquad\chi^{\prime}\Big{(}\Phi+a\mathrm{i}\phi_{0}^{\prime}Y\Big{)}=0.$
(13)
While the last two of these equations permit the solution $\chi^{\prime}=0$,
this is unable to satisfy the first equation in (13). The remaining solutions
can be solved to give $\chi^{\prime}=a\mathrm{i}(\phi_{0}^{\prime})^{-2}$,
which we integrate to find
$\chi_{a}(\sigma)=a\mathrm{i}\int_{a\mathrm{i}}^{\sigma}\bigg{[}1+\frac{\Gamma_{\text{c}}}{\pi}\frac{1}{(1+t^{2})}\bigg{]}^{-2}\mathrm{d}t.$
(14)
Here, we have introduced the notation $\chi_{a}=\chi$, where $a=\pm 1$, to
discern between each singulant generated by the two singular points of
$\phi_{0}^{\prime}$, which are given by $\sigma=\mathrm{i}$ and
$\sigma=-\mathrm{i}$. The starting point of integration in (14) is
$\sigma=\pm\mathrm{i}$ to ensure that $\chi_{a}(a\mathrm{i})=0$. This
condition is required in order to match with an inner solution near this
singular point. Integration of (14) yields
$\displaystyle\chi_{a}(\sigma)=$ $\displaystyle
a\mathrm{i}\bigg{[}\sigma+\frac{\Gamma_{\text{c}}^{2}\sigma}{2(\Gamma_{\text{c}}+\pi)(\pi\sigma^{2}+\Gamma_{\text{c}}+\pi)}-\frac{\Gamma_{\text{c}}(3\Gamma_{\text{c}}+4\pi)}{2\sqrt{\pi}(\Gamma_{\text{c}}+\pi)^{3/2}}\tan^{-1}{\bigg{(}\frac{\sqrt{\pi}\sigma}{\sqrt{(\Gamma_{\text{c}}+\pi)}}\bigg{)}}\bigg{]}$
(15)
$\displaystyle+1+\frac{\Gamma_{\text{c}}}{2(\Gamma_{\text{c}}+\pi)}-\frac{\Gamma_{\text{c}}(3\Gamma_{\text{c}}+4\pi)}{2\sqrt{\pi}(\Gamma_{\text{c}}+\pi)^{3/2}}\tanh^{-1}{\bigg{(}\frac{\sqrt{\pi}}{\sqrt{\Gamma_{\text{c}}+\pi}}\bigg{)}}.$
### 3.3 Solution of the late-term amplitude equations
We now determine the amplitude functions, $\Phi$, $X$, and $Y$, of the late
term solutions. Note that if one of these amplitude functions is known, then
the other two may be determined by the last two equations in (13). Thus, only
one equation is required for the amplitude functions, which we find at the
next order of $n$ in the late term equation (10a). This equation is given by
$\phi_{0}^{\prime}\Phi^{\prime}=\phi_{1}^{\prime}\chi^{\prime}\Phi,$ (16)
which may be integrated to find the solution
$\Phi(\sigma)=\Lambda\exp{\bigg{(}a\mathrm{i}\int_{0}^{\sigma}\frac{\phi_{1}^{\prime}(t)}{[\phi_{0}^{\prime}(t)]^{3}}\mathrm{d}t\bigg{)}}.$
(17)
In the above, $\Lambda$ is a constant of integration, which is determined by
matching with an inner solution near the singular points $\sigma=a\mathrm{i}$.
Once $\Phi$ is known, the remaining amplitude functions are determined by the
equations $Y=a\mathrm{i}(\phi_{0}^{\prime})^{-1}\Phi$ and
$X=a\mathrm{i}\phi_{0}^{\prime\prime}\Phi$.
We now calculate the constant, $\alpha$, that appears in the factorial-over-
power ansatzes (12). This is determined by ensuring that the singular
behaviour, as $\sigma\to a\mathrm{i}$, of each ansatz is consistent with the
anticipated singular behaviours of
$x_{n}=O\Big{(}(\sigma-a\mathrm{i})^{1-3n}\Big{)},\qquad
y_{n}=O\Big{(}(\sigma-a\mathrm{i})^{1-3n}\Big{)},\qquad\phi_{n}=O\Big{(}(\sigma-a\mathrm{i})^{-3n}\Big{)}.$
(18)
In taking the inner limit of $\Phi$ from (17), we have
$\Phi=O(\sigma-a\mathrm{i})^{3/2}$. Furthermore since
$\chi=O\big{(}(\sigma-a\mathrm{i})^{3}\big{)}$, derived later in equation
(40), equating the power of the singularities for $\phi_{n}$ between the
ansatz (12) and the anticipated singular behaviour above in (18) yields the
value of $\alpha=1/2$. The constant of integration, $\Lambda$, that appears in
solution (17) for the amplitude function, $\Phi$, is derived in Appendix A by
matching the inner limit of the divergent solution, $\phi_{n}$, with an inner
solution at $\sigma=a\mathrm{i}$. This yields
$\alpha=\frac{1}{2}\qquad\text{and}\qquad\Lambda=-\frac{\Gamma_{\text{c}}(-a\mathrm{i})^{1/2}\mathrm{e}^{-\mathcal{P}(a\mathrm{i})}}{6\pi}\bigg{(}-\frac{4a\mathrm{i}\pi^{2}}{3\Gamma_{\text{c}}^{2}}\bigg{)}^{\alpha}\lim_{n\to\infty}\bigg{(}\frac{\hat{\phi}_{n}}{\Gamma(n+\alpha+1)}\bigg{)},$
(19)
where $\hat{\phi}_{n}$, determined via recurrence relation (45), is a constant
appearing in the series expansion for the outer limit of the inner solution
for $\phi$, and $\mathcal{P}(\sigma)$ is defined in equation (48).
To conclude, the late-term divergence of the asymptotic expansions (7) diverge
in a factorial-over-power manner specified by the ansatzes (12). Evaluation of
this divergence requires the constants $\alpha$ and $\Lambda$ from equation
(19), as well as the singulant function $\chi(\sigma)$ from (15) and amplitude
function $\Phi(\sigma)$ from (17). These will be required in the derivation of
the exponentially-small terms considered in the next section.
### 3.4 Stokes smoothing and Stokes lines
The exponentially-small components of the solutions are now determined. We
truncate the asymptotic expansions (7) at $n=N-1$ and considering a remainder,
yielding
$x=\underbrace{\sum_{n=0}^{N-1}F^{2n}x_{n}}_{x_{r}}+~{}\bar{x},\qquad
y=\underbrace{\sum_{n=0}^{N-1}F^{2n}y_{n}}_{y_{r}}+~{}\bar{y},\qquad\phi^{\prime}=\underbrace{\sum_{n=0}^{N-1}F^{2n}\phi_{n}^{\prime}}_{\phi^{\prime}_{r}}+~{}\bar{\phi},$
(20)
where the truncated asymptotic expansions have been denoted by $x_{r}$,
$y_{r}$, and $\phi_{r}^{\prime}$. When $N$ is chosen optimally at the point at
which the divergence expansions reorder as $n\to\infty$, given by
$N\sim\frac{\lvert\chi\rvert}{F^{2}}+\rho,$ (21)
where $0\leq\rho<1$ to ensure that $N$ is an integer, the remainders to the
asymptotic expansions (20) will be exponentially-small.
Equations for these remainders are found by substituting the truncated
expansions (20) into the analytically continued equations (6a)–(6c). These are
given by
$\displaystyle(F^{2}\phi_{0}^{\prime}+F^{4}\phi_{1}^{\prime})\bar{\phi}^{\prime}+\bar{y}$
$\displaystyle=-\xi_{\text{a}},$ (22a) $\displaystyle
2\bar{x}^{\prime}+2F^{2}y_{1}^{\prime}\bar{y}^{\prime}$
$\displaystyle=-\xi_{\text{b}},$ (22b)
$\displaystyle\bar{\phi}^{\prime}+a\mathrm{i}\phi_{0}^{\prime}\bar{y}^{\prime}$
$\displaystyle=-\xi_{\text{c}}.$ (22c)
In equations (22) above, nonlinear terms such as $\bar{x}^{2}$ were neglected
as they will be exponentially subdominant. In anticipating that
$\bar{x}=O(F^{2}\bar{y})=O(F^{2}\bar{\phi})$, terms of the first two orders of
$F^{2}$ have been retained on the left-hand side of (22a). Motivated by the
late-term analysis, in which equations for the amplitude functions were
obtained at leading order for the last two governing equations, we have only
retained the leading order terms in equations (22b) and (22c). Furthermore,
the forcing terms introduced in equations (22) are defined by
$\left.\quad\begin{aligned}
\xi_{\text{a}}&=\frac{F^{2}}{2}(\phi_{r}^{\prime})^{2}+y_{r}-\frac{F^{2}}{2},\qquad\xi_{\text{b}}=\big{(}x_{r}^{\prime}\big{)}^{2}+\big{(}y_{r}^{\prime}\big{)}^{2}-1,\\\
\xi_{\text{c}}&=\phi_{r}^{\prime}x_{r}^{\prime}-1+a\mathrm{i}\phi_{r}^{\prime}y_{r}^{\prime}-\frac{\Gamma_{\text{c}}}{\pi}\frac{y_{r}+1}{(x_{r})^{2}+(y_{r}+1)^{2}}-\widehat{\mathcal{I}}[x_{r},y_{r},\phi_{r}^{\prime}].\end{aligned}\quad\right\\}$
(23)
Since each order of these forcing terms will be identically zero up to and
including $O(F^{2(N-1)})$, they will be of $O(F^{2N})$. Only knowledge of
$\xi_{\text{a}}$ will be required in the Stokes smoothing procedure of this
section, which is given by
$\xi_{\text{a}}\sim\phi_{0}^{\prime}\phi_{N-1}^{\prime}F^{2N}.$ (24)
Homogeneous solutions to equations (22), for which the forcing terms on the
right-hand sides are omitted, are $\bar{x}\sim
F^{2}X\mathrm{e}^{-\chi/F^{2}}$, $\bar{y}\sim Y\mathrm{e}^{-\chi/F^{2}}$, and
$\bar{\phi}\sim\Phi\mathrm{e}^{-\chi/F^{2}}$, where the singulant $\chi$ and
amplitude functions $X$, $Y$, and $\Phi$ satisfy the same equations as those
found for the late-term solutions in §3.2. Next, we solve for the particular
solutions of equations (22) through variation of parameters by multiplying the
homogeneous solutions by an unknown function, $\mathcal{S}(\sigma)$, giving
$\left.\quad\begin{aligned}
\bar{x}\sim\mathcal{S}(\sigma)X(\sigma)\mathrm{e}^{-\chi(\sigma)/F^{2}},\\\
\bar{y}\sim\mathcal{S}(\sigma)Y(\sigma)\mathrm{e}^{-\chi(\sigma)/F^{2}},\\\
\bar{\phi}\sim\mathcal{S}(\sigma)\Phi(\sigma)\mathrm{e}^{-\chi(\sigma)/F^{2}},\end{aligned}\quad\right\\}$
(25)
where $Y=a\mathrm{i}\Phi/\phi_{0}^{\prime}$ and $X=-y_{1}^{\prime}Y$. The
function $\mathcal{S}$ is called the Stokes multiplier, as it will display the
Stokes phenomenon across Stokes lines of the problem, which is demonstrated
next. An equation for $\mathcal{S}$ is obtained by substituting (25) and
similar expressions for $\bar{x}$ and $\bar{y}$ into equation (22a), yielding
$F^{2}\phi_{0}^{\prime}\Phi\mathrm{e}^{-\chi/F^{2}}\mathcal{S}^{\prime}(\sigma)\sim-\xi_{\text{a}}$.
In substituting for the dominant behaviour of $\xi_{\text{a}}$ from (24) and
the factorial-over-power divergence of $\phi^{\prime}_{N-1}$ from (12), and
changing derivatives of $\mathcal{S}$ from $\sigma$ to $\chi$, we find
$\frac{\mathrm{d}\mathcal{S}}{\mathrm{d}\chi}\sim\frac{\Gamma(N+\alpha)}{\chi^{N+\alpha}}F^{2(N-1)}\mathrm{e}^{\chi/F^{2}}.$
(26)
Figure 3: The Stokes lines (bold) lie along the imaginary axis between the
two singular points of $\sigma=-\mathrm{i}$ and $\sigma=\mathrm{i}$. Branch
cuts are shown with a wavy line.
In expanding as $N\to\infty$, and substituting for
$N\sim\lvert\chi\rvert/F^{2}+\rho$ from equation (21), the right-hand side of
equation (25) is seen to be exponentially-small, except for in a boundary
layer close to contours satisfying
$\text{Im}[\chi]=0\quad\text{and}\quad\text{Re}[\chi]>0.$ (27)
These are the Stokes line conditions originally derived by Dingle (1973).
Across the Stokes lines, the solution for the Stokes multiplier $\mathcal{S}$,
$\mathcal{S}(\sigma)=S_{a}+\frac{\sqrt{2\pi}\mathrm{i}}{F^{2\alpha}}\int_{-\infty}^{\sqrt{\lvert\chi\rvert}\tfrac{\arg{(\chi)}}{F}}\exp{(-t^{2}/2)}\mathrm{d}t,$
(28)
rapidly varies from the constant $S_{a}$ to
$S_{a}+2\pi\mathrm{i}/F^{2\alpha}$. This is the Stokes phenomenon, and the
contours satisfying the Dingle conditions (27) are shown in figure 3 to lie
along the imaginary axis. For the one vortex case studied in this section, the
upstream condition as $\text{Re}[\sigma]\to-\infty$ requires that $S_{1}=0$
and $S_{-1}=-2\pi\mathrm{i}/F^{2\alpha}$.
### 3.5 Trapped waves generated by two submerged vortices
We have so far studied the case of a single submerged point vortex. When
multiple point vortices are placed within the fluid, the only change is to the
boundary integral equation, previously specified in (6c) for a single vortex.
In this section we study the formulation of two submerged point vortices of
the same nondimensional strength, $\Gamma_{\text{c}}$, located at
$z=x+\mathrm{i}y=\pm\lambda-\mathrm{i}$, for which the analytically continued
boundary integral equation is given by
$\displaystyle\phi^{\prime}(\sigma)x^{\prime}(\sigma)-1+a\mathrm{i}\phi^{\prime}(\sigma)y^{\prime}(\sigma)=\frac{\Gamma_{\text{c}}}{\pi}$
$\displaystyle\bigg{[}\frac{y(\sigma)+1}{[x(\sigma)-\lambda]^{2}+[y(\sigma)+1]^{2}}$
(29)
$\displaystyle+\frac{y(\sigma)+1}{[x(\sigma)+\lambda]^{2}+[y(\sigma)+1]^{2}}\bigg{]}+\widehat{\mathcal{I}}[x,y,\phi].$
Unlike the case for a single submerged point vortex that produces waves in the
far field for $x\to\infty$, two identical point vortices can produce solutions
for which the waves are confined to lie between the vortices,
$-\lambda<\text{Re}[\sigma]<\lambda$. This occurs for critical values of the
Froude number, which we now predict using the techniques of exponential
asymptotics developed in the previous sections.
The first two orders of the asymptotic solution for $\phi$ are now given by
$\displaystyle\phi_{0}^{\prime}(\sigma)$
$\displaystyle=1+\frac{\Gamma_{\text{c}}}{\pi}\bigg{[}\frac{1}{1+(\sigma+\lambda)^{2}}+\frac{1}{1+(\sigma-\lambda)^{2}}\bigg{]},$
(30a) $\displaystyle\phi_{1}^{\prime}(\sigma)$
$\displaystyle=-a\mathrm{i}\phi_{0}^{\prime}y_{1}^{\prime}+\frac{\Gamma_{\text{c}}y_{1}}{\pi}\bigg{[}\frac{(\sigma+\lambda)^{2}-1}{[1+(\sigma+\lambda)^{2}]^{2}}+\frac{(\sigma-\lambda)^{2}-1}{[1+(\sigma-\lambda)^{2}]^{2}}\bigg{]}+\widehat{\mathcal{I}}_{n}(\sigma),$
(30b)
which are singular at the four locations $\sigma=-\lambda+a\mathrm{i}$ (from
the vortex at $z=-\lambda-\mathrm{i}$) and $\sigma=\lambda+a\mathrm{i}$ (from
the vortex at $z=\lambda-\mathrm{i}$). Note that we have again defined $a=\pm
1$ to indicate whether $\text{Im}[\sigma]>0$ or $\text{Im}[\sigma]<0$. These
four singular points each have associated Stokes lines, shown in figure 4.
Figure 4: The Stokes lines (bold) generated by the four singular points are
shown.
In general, the waves switched on across the first Stokes lines, emanating
from the points $\sigma=-\lambda+a\mathrm{i}$, will be out of phase with the
waves switched on across the second Stokes lines, from
$\sigma=\lambda+a\mathrm{i}$. However, for certain values of $F$, the wave
switched on across the first Stokes line is then switched off by the second
Stokes line, yielding solutions with no waves for $\text{Re}[\sigma]>\lambda$.
An example of this trapped solution was shown earlier in figure 1(b).
Thus, in using the Stokes switching prediction for $\bar{\phi}$ shown in
figure 4 and writing $\bar{y}=a\mathrm{i}\bar{\phi}/\phi_{0}^{\prime}$, we
require for the two contributions of
$\left.\begin{aligned}
\bar{y}_{1}&\sim-\frac{2\pi}{F^{2\alpha}\phi^{\prime}_{0}}\Phi_{1}(\sigma)\exp{\Big{(}-\frac{\chi_{1}(\sigma)}{F^{2}}\Big{)}}+c.c.,\\\
\bar{y}_{2}&\sim-\frac{2\pi}{F^{2\alpha}\phi^{\prime}_{0}}\Phi_{2}(\sigma)\exp{\Big{(}-\frac{\chi_{2}(\sigma)}{F^{2}}\Big{)}}+c.c.,\end{aligned}\quad\right\\}$
(31)
to cancel with one another for $\text{Re}[\sigma]>\lambda$. Here, we denoted
$\chi_{1}$ and $\Phi_{1}$ as the singulant and amplitude function arising from
the $\sigma=-\lambda+a\mathrm{i}$ singularities, and $\chi_{2}$ and $\Phi_{2}$
as those arising from the $\sigma=\lambda+a\mathrm{i}$ singularities. The
first of (31), $\bar{y}_{1}$, is the contribution switched on as we pass from
left to right across the Stokes lines associated with the singular points
$\sigma=-\lambda+a\mathrm{i}$. The second, $\bar{y}_{2}$, is the contribution
switched on from left to right by the Stokes lines associated with the
$\sigma=\lambda+a\mathrm{i}$ singular point. Note that the specified
contributions in (31) are from the $a=1$ contribution, and the unspecified
complex-conjugate components are from that with $a=-1$.
We now simplify each of the expressions given in equation (31) by substituting
for the amplitude functions $\Phi_{1}$ and $\Phi_{2}$, which satisfy the same
equation as that found previously in (16). The only difference will be the
constants of integration, which we denote by $\Lambda_{1}$ and $\Lambda_{2}$.
This yields
$\left.\begin{aligned}
\bar{\phi}_{1}&\sim-\frac{4\pi\lvert\Lambda_{1}\rvert}{F^{2\alpha}\phi_{0}^{\prime}}\exp{\left(-\frac{\text{Re}[\chi_{1}]}{F^{2}}\right)}\cos{\left(\int_{0}^{\sigma}\frac{\phi_{1}^{\prime}(t)}{[\phi_{0}^{\prime}(t)]^{3}}\mathrm{d}t+\arg{[\Lambda_{1}]}-\frac{\text{Im}[\chi_{1}]}{F^{2}}\right)},\\\
\bar{\phi}_{2}&\sim-\frac{4\pi\lvert\Lambda_{2}\rvert}{F^{2\alpha}\phi_{0}^{\prime}}\exp{\left(-\frac{\text{Re}[\chi_{2}]}{F^{2}}\right)}\cos{\left(\int_{0}^{\sigma}\frac{\phi_{1}^{\prime}(t)}{[\phi_{0}^{\prime}(t)]^{3}}\mathrm{d}t+\arg{[\Lambda_{2}]}-\frac{\text{Im}[\chi_{2}]}{F^{2}}\right)}.\end{aligned}\quad\right\\}$
(32)
Since $\lvert\Lambda_{1}\rvert=\lvert\Lambda_{2}\rvert$ and
$\text{Re}[\chi_{1}]=\text{Re}[\chi_{2}]$, the prefactors multiplying each of
the cosine functions in (32) are identical, and the condition for them to
cancel, $\bar{y}_{1}+\bar{y}_{2}=0$, yields
$\displaystyle\cos{\left(\int_{0}^{\sigma}\frac{\phi_{1}^{\prime}(t)}{[\phi_{0}^{\prime}(t)]^{3}}\mathrm{d}t+\frac{\arg{[\Lambda_{1}]}+\arg{[\Lambda_{2}]}}{2}-\frac{\text{Im}[\chi_{1}+\chi_{2}]}{2F^{2}}\right)}$
(33)
$\displaystyle\qquad\qquad\qquad\qquad\times\cos{\left(\frac{\arg{[\Lambda_{1}]}-\arg{[\Lambda_{2}]}}{2}-\frac{\text{Im}[\chi_{1}-\chi_{2}]}{2F^{2}}\right)}=0.$
Note that since $\chi_{1}$ and $\chi_{2}$ satisfy the same differential
equation, $\chi^{\prime}=a\mathrm{i}(\phi_{0}^{\prime})^{-2}$, originally
derived in §3.2, the only difference between them are their constants of
integration. Therefore $\text{Im}[\chi_{1}+\chi_{2}]$ will be a function of
$\sigma$, and $\text{Im}[\chi_{1}-\chi_{2}]$ will be constant. Thus, only the
second cosine component of (33) is capable of satisfying the identity for
$\text{Re}[\sigma]>\lambda$. Since this cosine function is zero when the
argument equals $\pm\pi/2$, $\pm 3\pi/2$, and so forth, we find
$F_{k}=\sqrt{\frac{\text{Im}[\chi_{1}-\chi_{2}]}{\arg{[\Lambda_{1}]}-\arg{[\Lambda_{2}]}+\pi(2k+1)}}.$
(34)
for $k=0,1,2,\ldots$, and so forth. Equation (34) yields the discrete values
of the Froude number, $F_{k}$, for which the waves are confined to lie between
the two submerged vortices. All that remains is to evaluate
$\text{Im}[\chi_{1}-\chi_{2}]$, $\arg{[\Lambda_{1}]}$, and
$\arg{[\Lambda_{2}]}$. Each of these singulants are found by integrating
$\chi^{\prime}=\mathrm{i}(\phi_{0}^{\prime})^{-2}$, where $\phi_{0}^{\prime}$
is specified in equation (30), from the corresponding singular point. We may
decompose each singulant into a real-valued integral along the Stokes line,
and an imaginary-valued integral along the free-surface. Thus,
$\text{Im}[\chi]$ is an integral along the free-surface,
$\text{Im}[\sigma]=0$, from the intersection of the Stokes line to $\sigma$.
This yields
$\text{Im}[\chi_{1}(\sigma)-\chi_{2}(\sigma)]=\int_{-\lambda}^{\lambda}\bigg{[}1+\frac{\Gamma_{\text{c}}}{\pi}\bigg{(}\frac{1}{1+(t+\lambda)^{2}}+\frac{1}{1+(t-\lambda)^{2}}\bigg{)}\bigg{]}^{-2}\mathrm{d}t.$
(35)
In the numerical results of §4.2, the integral in (35) is evaluated with a
symbolic programming language. Note that the Stokes lines depicted in figure 4
are not truly vertical, and are slightly curved such that they intersect the
free surface at the points $-\lambda^{*}$ and $\lambda^{*}$. Thus, the range
of integration in (35) should actually lie between
$-\lambda^{*}<t<\lambda^{*}$; however since $\lambda^{*}$ is very close in
value to $\lambda$ (for $\lambda=8$ and $\Gamma_{c}=0.3$, $\lambda^{*}\approx
7.99998$), this subtlety has been ignored.
Comparisons between the analytical prediction of $F_{k}$ from (34) and
numerical results are performed in §4.2.
## 4 Numerical results
We begin in §4.1 by verifying with numerical results our analytical
predictions for the exponentially-small scaling as $F\to 0$ for the case of a
single vortex. This is given by the singulant function, $\chi$, from (15), and
comparisons are made for a range of values of the vorticity,
$\Gamma_{\text{c}}$. The analytical predictions of the Froude numbers for
trapped waves between two point vortices, given in (34), are then compared to
numerical predictions in §4.2.
A detailed description of the numerical method used is given by Forbes (1985),
which we will briefly summarise here.
1. (i)
The real-valued domain, $s$, is truncated to lie between the values of
$s_{\text{L}}$ and $s_{\text{R}}$. $N$ discretisation points are used, such
that the numerical domain is given by
$s_{k}=s_{\text{L}}+(k-1)(s_{\text{R}}-s_{\text{L}})/(N-1)$ for $1\leq k\leq
N$. The unknown solution is taken to be $y^{\prime}(s)$, which we define at
each gridpoint by $y^{\prime}_{k}=y^{\prime}(s_{k})$. The radiation conditions
are imposed by enforcing $y_{1}=0$, $y_{1}^{\prime}=0$, $x_{1}^{\prime}=1$,
$\phi_{1}^{\prime}=1$, $x_{1}=s_{\text{l}}$, and $\phi_{1}=s_{\text{L}}$, and
the initial guess for $y_{k}^{\prime}$ is either zero or a previously computed
solution.
2. (ii)
Since we assume that $y^{\prime}_{k}$ is known at the next gridpoint, the
arclength relation (2b) yields $x^{\prime}_{k}$. Trapezoidal-rule integration
then determines values for $x_{k}$ and $y_{k}$, which we use to find
$\phi_{k}^{\prime}$ from Bernoulli’s equation (2a). This process is repeated
for $k=2$ to $k=N$ to find function values at every gridpoint.
3. (iii)
The boundary-integral equation (2c) is evaluated at each gridpoint with the
known values of $x_{k}$, $y_{k}$, $\phi^{\prime}_{k}$, $x^{\prime}_{k}$, and
$y^{\prime}$. To avoid the singularity associated with the principal-valued
integral $\mathcal{I}[x,y,\phi^{\prime}]$, each unknown that is not a function
of the integration variable, $t$, is instead evaluated between gridpoints by
interpolation.
4. (iv)
This yields $N-1$ nonlinear equations from evaluating the boundary-integral
equation between each gridpoint, $(s_{k}+s_{k+1})/2$, which is closed by the
$N-1$ unknowns $y^{\prime}_{k}$ for $k=2$ to $k=N$. Solutions are found by
minimising the residual through Newton iteration. For the trapped waves
studied in §4.2, we impose an additional constraint of symmetry about $s=0$ in
the real-valued solution, $y(s)$, such that the Froude number, $F$, is
determined as an eigenvalue.
### 4.1 Waves generated by a single vortex
Figure 5: The exponentially-small dependence of the wave amplitude is shown
(dots) for numerical results for seven different values of
$\Gamma_{\text{c}}=\\{0.1,0.15,0.2,0.25,0.3,0.35,0.4\\}$. Solid lines
represent the analytical gradient found from the real part of $\chi$ in
equation (15). The behaviour of this gradient for different values of the
vortex strength $\Gamma_{\text{c}}$ is shown in figure 6.
For the numerical results presented in this section, we have used $N=2000$
grid points, and a domain specified by $s_{\text{L}}=-40$ and
$s_{\text{R}}=40$. In computing numerical solutions for a wide range of Froude
numbers, and the values of
$\Gamma_{\text{c}}=\\{0.1,0.15,0.2,0.25,0.3,0.35,0.4\\}$, the exponentially-
small scaling as $F\to 0$ of the high-frequency waves present for $s>0$ may be
measured. This is shown in the semilog plot of figure 5. We see that these
lines, each of which represents solutions with a different value of
$\Gamma_{\text{c}}$, are straight and thus the amplitude of these ripples is
exponentially small as $F\to 0$. The gradient of each of these lines is
expected to closely match the exponential scaling predicted analytically,
given by the singulant $\chi$. Along the free surface, this is given by
$\text{Re}[\chi]$ from equation (15) which takes constant values.
Figure 6: The analytical prediction for $\text{Re}[\chi]$ along the free
surface $\text{Im}[\sigma]=0$ from equation (15) is shown against the
vorticity $\Gamma_{\text{c}}$ (line). The numerical predictions, corresponding
to the slopes of the semilog plot in figure 5, are shown circled.
In figure 6, this analytical prediction is compared to the numerical values
from figure 5, and good agreement is observed. Note that there are small
instabilities present in the numerical solution which decay when the truncated
domain is extended; upon which we expect the numerical results to tend towards
the analytical prediction shown in figure 6.
Figure 7: For $F=0.45$ and $\Gamma_{\text{c}}=0.4$, a numerical solution
(dashed) is compared to an analytical solution (line) determined in §3.
Comparison between a numerical and asymptotic solution profile is shown in
figure 7 for $F=0.45$ and $\Gamma_{\text{c}}$. The numerical solution is
determined by the scheme detailed at the beginning of §4, with $N=2000$
discretisation points in the arclength, $-40\leq s\leq 40$. The asymptotic
solution plots $x(s)=x_{0}(s)+F^{2}x_{1}(s)+\bar{x}(s)$ against
$y(s)=y_{0}(s)+F^{2}y_{1}(s)+\bar{y}(s)$. These early order solutions,
$x_{0}$, $x_{1}$, $y_{0}$, and $y_{1}$ are specified in equations (8) and (9).
The exponentially-small components, $\bar{x}$ and $\bar{y}$, are implemented
from expression (25). This requires knowledge of the singulant, $\chi$, given
in (15), the amplitude functions $Y=a\mathrm{i}\Phi/\phi_{0}^{\prime}$ and
$X=-y_{1}^{\prime}Y$ determined from $\Phi$ in (17), and the Stokes
multiplier, $\mathcal{S}$, given in (28). A real-valued asymptotic solution is
obtained through evaluating the sums
$\bar{x}\rvert_{a=1}+\bar{x}\rvert_{a=-1}$ and
$\bar{y}\rvert_{a=1}+\bar{y}\rvert_{a=-1}$ on the real-valued domain,
$\sigma=s$, for $\text{Im}[\sigma]=0$. Note that in the determination of the
constant $\Lambda$, its magnitude, $\lvert\Lambda\rvert$, been fitted to equal
that found from the corresponding numerical solution, and its argument
(corresponding to a phase shift of the resultant wave) is determined from
relation (19) as $\text{arg}{[\Lambda]}=a\pi/2$.
### 4.2 Trapped gravity waves between two vortices
We considered the case of two submerged point vortices analytically in §3.5.
When each vortex had the same nondimensional circulation, $\Gamma_{\text{c}}$,
and depth equal to unity, trapped waves were seen to occur for certain
discrete values of the Froude number, $F_{k}$. In this section, we compare the
analytical prediction for $F_{k}$ from (34) with numerical results. These
trapped numerical solutions are found with the method detailed at the
beginning of §4. In imposing the additional constraint of symmetry to
eliminate waves downstream of the vortices, the special Froude number,
$F_{k}$, is determined as an eigenvalue. These results were performed for
$N=4000$ grid points, a domain between $s_{\text{L}}=-60$ and
$s_{\text{R}}=60$, and horizontal vortex placement specified as $\lambda=8$.
In figure 8, we plot the tail amplitude (for $s>\lambda$) of the asymptotic
solutions for the values of $0.3<F<0.5$, $\Gamma_{\text{c}}=0.3$, and
$\lambda=8$. This amplitude is equal to zero at the values of $F_{k}$ from
equation (34). The figure also contains additional markers denoted by (a),
where $F=0.3383$, and (b), where $F=0.4270$. This corresponds to the figure 9
where we compare numerical solutions obtained in this section, and asymptotic
solutions from §3 for those given values of $F$. The fit is excellent and the
corresponding curves are nearly visually indistinguishable at the scale of the
graphic.
Figure 8: The amplitude of oscillations present for $s>\lambda$ in the
asymptotic solutions is shown against the Froude number, $F$. Here,
$\Gamma_{\text{c}}=0.3$ and $\lambda=8$. This amplitude is equal to zero at
the locations $F_{k}$ derived in equation (34). The two points marked (a) and
(b) correspond to the profiles shown in figure 9. Figure 9: Two different
trapped wave solutions are shown for $\Gamma_{\text{c}}=0.3$ and $\lambda=8$
corresponding to (a) $F=0.3383$ and (b) $F=0.4270$. Asymptotic solutions
(solid line) are compared to numerical solutions (dashed) for (a) $k=22$ and
(b) $k=14$. In each inset, the two curves are nearly indistinguishable to
visual accuracy.
Finally, in figure 10, we compare the values of $F_{k}$ obtained analytically
and numerically. The straight lines are the analytical prediction from (34),
and dots represent the numerical values for $F_{k}$.
Figure 10: Values of the Froude number, $F_{k}$, for which the waves are
trapped between each submerged vortex are shown. The numerical results of §4.2
are shown circled, and the analytical results from equation (34) are shown
with lines. Here, $\lambda=8$, and for the numerical solutions $N=4000$,
$s_{\text{L}}=-60$, and $s_{\text{R}}=60$.
## 5 Conclusion
We have shown, through both numerical and analytical investigations, that the
waves generated by submerged point vortices are exponentially small in the
low-speed limit of $F\to 0$. Furthermore, when two submerged vortices are
considered, oscillatory waves vanish downstream for certain values of the
Froude number, $F$. Through the techniques of exponential asymptotics, we have
demonstrated how these values may be derived. Their prediction relies on the
understanding of singularities in the analytically continued domain that
generate a divergent asymptotic expansion. The remainder to this series is
exponentially small as $F\to 0$, and the study of the associated Stokes
phenomenon yields discrete values of $F$ for which the waves are trapped
between each vortex.
## 6 Discussion
The work presented here forms a basis for a number of interesting extensions
involving exponentially-small water waves with gravity, capillarity, and/or
vorticity providing singular perturbative effects.
First, it should be remarked that the classical exponential-asymptotics
theories by _e.g._ Chapman & Vanden-Broeck (2002, 2006) for capillary- and
gravity-driven surface waves produced in flows over topographies rely upon the
existence of closed-form conformal maps. In such problems, the governing
equations for the free-surface can be written in terms of a single complex-
valued unknown (_e.g._ the complex velocity), with the velocity potential
serving as the independent variable. This includes situations such as flows
past polygonal boundaries (related to the availability of the Schwarz-
Christoffel mapping). The arclength formulation we have used in this work
provides a more general setting for wave-structure interactions with arbitrary
bodies, including for instance, flows past smoothed bodies specified in
$(x,y)$-coordinates. Here, we have demonstrated that the exponential
asymptotics can be generalised to such formulations. We expect that many of
the interesting wave-structure interactions studied by _e.g._ Holmes et al.
(2013) (flow past a symmetric bottom topography) and Hocking et al. (2013)
(flow past a semi-ellipse), can be attacked using the technology we have
developed here.
Secondly, the phenomena of trapped waves is an interesting one. The
exponential asymptotics interpretation, whereby waves switched-on at one
location (the Stokes line intersection) must be switched-off at another,
provides an intuitive explanation for how trapped waves form in singularly
perturbative limits. The context, in our problem, relates to vortices fixed
within the fluid for modelling submerged obstructions, such as the submerged
cylinders studied numerically by Tuck & Scullen (1998). However, trapped waves
have been detected numerically in other geometries including submerged bumps
(Hocking et al., 2013), a semi-ellipse (Holmes et al., 2013), a trigonometric
profile (Dias & Vanden-Broeck, 2004), spikes (Binder et al., 2005), and a
rectangular bump (Lustri et al., 2012). We expect that the ‘selection
mechanism’ that produces the countably infinite set of values (34) is a kind
of universality in eigenvalue problems (cf. Chapman et al. 2022 for further
discussion and examples).
Finally, we note that in this paper, the forcing mechanism producing the waves
was via the complex-plane singularities associated with the point
vortices—then, we found that the waves were singularly perturbed due to the
inertial term in Bernoulli’s equation, thus producing exponentially small
waves, scaling as $\exp(-\text{const.}/F^{2})$. Recently, analytical solutions
have been developed for pure-vorticity-driven water waves, notably in the
works by Crowdy & Nelson (2010); Crowdy & Roenby (2014); Crowdy (2022). In
essence, we believe these solutions can serve as leading-order approximations
in the regime of small surface-tension; it might be expected that
exponentially-small parasitic ripples then exist on the surface of such
vorticity-driven profiles. This would then be similar to the work of Shelton
et al. (2021); Shelton & Trinh (2022) for parasitic capillary ripples on steep
gravity waves. Numerical and analytical work on such class of problems is
ongoing.
Acknowledgements. We are grateful for many stimulating and motivating
discussions that took place during the recent LMS-Bath symposium “New
Directions in Water Waves” held at the University of Bath in July 2022. PHT is
supported by the Engineering and Physical Sciences Research Council
[EP/V012479/1].
Declaration of interests. The authors report no conflict of interest.
## References
* Binder et al. (2005) Binder, B. J., Vanden-Broeck, J.-M. & Dias, F. 2005 Forced solitary waves and fronts past submerged obstacles. Chaos 15 (3), 037106.
* Chapman et al. (2022) Chapman, S. J., Dallaston, M. C., Kalliadasis, S., Trinh, P. H. & Witelski, T. P. 2022 The role of exponential asymptotics and complex singularities in transitions and branch merging of nonlinear dynamics. In review. .
* Chapman & Vanden-Broeck (2002) Chapman, S. J. & Vanden-Broeck, J.-M. 2002 Exponential asymptotics and capillary waves. SIAM J. Appl. Math. 62 (6), 1872–1898.
* Chapman & Vanden-Broeck (2006) Chapman, S. J. & Vanden-Broeck, J.-M. 2006 Exponential asymptotics and gravity waves. J. Fluid Mech. 567, 299–326.
* Crowdy (2022) Crowdy, D. G. 2022 Exact solutions for steadily travelling water waves with submerged point vortices. In review .
* Crowdy & Nelson (2010) Crowdy, D. G. & Nelson, R. 2010 Steady interaction of a vortex street with a shear flow. Phys. Fluids 22 (9), 096601\.
* Crowdy & Roenby (2014) Crowdy, D. G. & Roenby, J. 2014 Hollow vortices, capillary water waves and double quadrature domains. Fluid Dyn. Res. 46 (3), 031424.
* Dias & Vanden-Broeck (2004) Dias, F. & Vanden-Broeck, J.-M. 2004 Trapped waves between submerged obstacles. J. Fluid Mech. 509, 93–102.
* Dingle (1973) Dingle, R. B. 1973 Asymptotic Expansions: Their Derivation and Interpretation. Academic Press, London.
* Forbes (1985) Forbes, L. K. 1985 On the effects of non-linearity in free-surface flow about a submerged point vortex. J. Eng. Math. 19 (2), 139–155.
* Gazdar (1973) Gazdar, A. S. 1973 Generation of waves of small amplitude by an obstacle placed on the bottom of a running stream. J. Phys. Soc. Japan 34 (2), 530–538.
* Haziot et al. (2022) Haziot, S., Hur, V., Strauss, W., Toland, J., Wahlén, Erik, Walsh, S. & Wheeler, M. 2022 Traveling water waves—the ebb and flow of two centuries. Quart. Appl. Math. 80 (2), 317–401.
* Hocking et al. (2013) Hocking, G. C., Holmes, R. J. & Forbes, L. K. 2013 A note on waveless subcritical flow past a submerged semi-ellipse. J. Eng. Math. 81 (1), 1–8.
* Holmes et al. (2013) Holmes, R. J., Hocking, G. C., Forbes, L. K. & Baillard, N. Y. 2013 Waveless subcritical flow past symmetric bottom topography. Eur. J. Appl. Math. 24 (2), 213–230.
* Lustri et al. (2012) Lustri, C. J., McCue, S. W. & Binder, B. J. 2012 Free surface flow past topography: a beyond-all-orders approach. Euro. J. Appl. Math. 23 (4), 441–467.
* Miksis et al. (1981) Miksis, M., Vanden-Broeck, J.-M. & Keller, J. B. 1981 Axisymmetric bubble or drop in a uniform flow. J. Fluid Mech. 108, 89–100.
* Shelton et al. (2021) Shelton, J., Milewski, P. & Trinh, P. H. 2021 On the structure of steady parasitic gravity-capillary waves in the small surface tension limit. J. Fluid Mech. 922.
* Shelton & Trinh (2022) Shelton, J. & Trinh, P.H. 2022 Exponential asymptotics for steady parasitic capillary ripples on steep gravity waves. J. Fluid Mech. 939.
* Tuck & Scullen (1998) Tuck, E. O. & Scullen, D. C. 1998 Tandem submerged cylinders each subject to zero drag. J. Fluid Mech. 364, 211–220.
* Vanden-Broeck & Tuck (1985) Vanden-Broeck, J.-M. & Tuck, E. O. 1985 Waveless free-surface pressure distributions. J. Ship Res. 29 (03), 151–158.
* Xie & Tanveer (2002) Xie, X. & Tanveer, S. 2002 Analyticity and nonexistence of classical steady Hele-Shaw fingers. Commun. Pur. Appl. Math. 56 (3), 353–402.
## Appendix A Inner analysis at the singularities $\sigma=\pm\mathrm{i}$
In order to determine the constant of integration of the amplitude function
$\Phi(\sigma)$ from equation (16), knowledge of the inner solutions at the
singularities $\sigma=\mathrm{i}$ and $\sigma=-\mathrm{i}$, is required. In
this section, we study the inner boundary layer at both of these locations,
for which matching with the inner limit of the outer solutions determines the
constant of integration.
First, we note that in the outer region, $\sigma=O(1)$, the asymptotic series
first reorder whenever
$\phi_{0}^{\prime}(\sigma)\sim F^{2}\phi_{1}^{\prime}(\sigma),\qquad
y_{1}(\sigma)\sim F^{2}y_{2}(\sigma),\qquad x_{2}(\sigma)\sim
F^{2}x_{3}(\sigma).$ (36)
In substituting for the early orders of the asymptotic solutions from
equations (8), (9), and (11), we see that each of (36) reorder in a boundary
layer of the same width, given by $\sigma-a\mathrm{i}=O(F^{2/3})$. We thus
introduce the inner variable, $\hat{\sigma}$, by the relation
$\sigma-a\mathrm{i}=\hat{\sigma}F^{2/3},$ (37)
for which $\hat{\sigma}=O(1)$ in the inner region. Since the asymptotic series
each reorder near the two locations of $\sigma=\mathrm{i}$ and
$\sigma=-\mathrm{i}$, we have again used the notation $a=\pm 1$ to distinguish
between these two cases.
Next, to determine the form of the inner solutions, we take the inner limit of
the outer series expansions for $\phi^{\prime}$, $x$, and $y$, by substituting
for the inner variable $\hat{\sigma}$ defined in (36) and expanding as $F\to
0$. This yields
$\phi^{\prime}\sim\frac{1}{F^{2/3}}\bigg{[}-\frac{a\mathrm{i}\Gamma_{\text{c}}}{2\pi}\frac{1}{\hat{\sigma}}+\cdots\bigg{]},\quad
y\sim
F^{2/3}\bigg{[}\frac{\Gamma_{\text{c}}^{2}}{8\pi^{2}}\frac{1}{\hat{\sigma}^{2}}+\cdots\bigg{]},\quad
x\sim a\mathrm{i}+F^{2/3}\bigg{[}\hat{\sigma}+\cdots\bigg{]},$ (38)
where the omitted terms, represented by ($\cdots$), are from the inner limit
of lower order terms of the outer asymptotic expansion. For instance, the next
term in the inner limit of $\phi^{\prime}$ is of $O(\hat{\sigma}^{-4})$. The
form of the inner limits in (38) motivates our definition of the inner
solutions, $\hat{\phi}(\hat{\sigma})$, $\hat{y}(\hat{\sigma})$, and
$\hat{x}(\hat{\sigma})$, through the equations
$\phi^{\prime}=-\frac{a\mathrm{i}\Gamma_{\text{c}}}{2\pi
F^{2/3}}\frac{\hat{\phi}(\hat{\sigma})}{\hat{\sigma}},\qquad
y=\frac{\Gamma_{\text{c}}^{2}F^{2/3}}{8\pi^{2}}\frac{\hat{y}(\hat{\sigma})}{\hat{\sigma}^{2}},\qquad
x=a\mathrm{i}+\hat{\sigma}F^{2/3}\hat{x}(\hat{\sigma}).$ (39)
The form of the inner variables introduced in (39) ensures that the first term
in the series expansion for their outer limit will be equal to unity.
Furthermore, based on the form of the inner limit of the singulant, $\chi$,
from equation (14),
$\chi\sim-\frac{4a\mathrm{i}\pi^{2}}{3\Gamma_{\text{c}}^{2}}\hat{\sigma}^{3}F^{2},$
(40)
the outer limit of the inner solutions will be a series expansion in inverse
powers of $-4a\mathrm{i}\pi^{2}\hat{\sigma}^{3}/(3\Gamma_{\text{c}})$. We thus
introduce the variable $z$, defined by
$z=-\frac{4a\mathrm{i}\pi^{2}}{3\Gamma^{2}_{\text{c}}}\hat{\sigma}^{3},$ (41)
to ensure that these series expansions are in inverse powers of $z$ alone.
### A.1 Inner equation
The leading order inner equations, as $F\to 0$, may now be derived by
substituting (39) into the outer equations (6a)-(6c), yielding
$\displaystyle\hat{y}-\hat{\phi}^{2}=0,$ (42a)
$\displaystyle\Big{(}\hat{x}+3z\hat{x}^{\prime}\Big{)}^{2}-\bigg{(}\frac{1}{3z}\hat{y}-\frac{1}{2}\hat{y}^{\prime}\bigg{)}^{2}=1,$
(42b)
$\displaystyle\hat{\phi}\bigg{(}\hat{x}-\frac{1}{6z}\hat{y}\bigg{)}\bigg{(}\hat{x}+3z\hat{x}^{\prime}-\frac{1}{3z}\hat{y}+\frac{1}{2}\hat{y}^{\prime}\bigg{)}=1.$
(42c)
The inner solutions, $\hat{\phi}(z)$, $\hat{y}(z)$, and $\hat{x}(z)$, will
satisfy equations (42a)-(42c). Rather than solve these inner equations
exactly, knowledge of the inner solutions is only required under the outer
limit of $z\to\infty$ in order to match with the inner limit of the outer
solutions to determine their divergent form. Thus, we will consider the
following series expansions for these inner unknowns,
$\hat{\phi}(z)=\sum_{n=0}^{\infty}\frac{\hat{\phi}_{n}}{z^{n}},\qquad\hat{y}(z)=\sum_{n=0}^{\infty}\frac{\hat{y}_{n}}{z^{n}},\qquad\hat{x}(z)=\sum_{n=0}^{\infty}\frac{\hat{x}_{n}}{z^{n}},$
(43)
which hold as $z\to\infty$.
At leading order as $z\to\infty$ we have, by the definition on the inner
solutions in equation (39),
$\hat{\phi}_{0}=1,\qquad\hat{y}_{0}=1,\qquad\hat{x}_{0}=1.$ (44)
Determination of $\hat{\phi}_{n}$, $\hat{y}_{n}$, and $\hat{x}_{n}$, as
$n\to\infty$, requires the evaluation of a recurrence relation, which is now
given.
Firstly, substitution of expansions (44) into the inner equation (42b) yields
$\displaystyle\hat{x}_{1}=0,\qquad 2(1-3n)\hat{x}_{n}$
$\displaystyle=\frac{1}{36}\sum_{m=0}^{n-2}(2+3m)(2n-3m-4)\hat{y}_{m}\hat{y}_{n-m-2}$
(45a)
$\displaystyle~{}~{}~{}+\sum_{m=1}^{n-1}(1-3m)(3n-3m-1)\hat{x}_{m}\hat{x}_{n-m},\qquad\text{for
$~{}n\geq 2$}.$ Next, we substitute the same expansions into the inner
equation (42c), yielding
$\displaystyle\hat{\phi}_{1}=\frac{1}{2},\qquad\hat{\phi}_{n}$
$\displaystyle=\sum_{m=2}^{n}\sum_{q=1}^{m-1}\frac{(3q-1)\hat{\phi}_{n-m}}{36}\Big{(}6\hat{x}_{q}-\hat{y}_{q-1}\Big{)}\Big{(}6\hat{x}_{m-q}-\hat{y}_{m-q-1}\Big{)}$
(45b)
$\displaystyle~{}~{}~{}-\sum_{m=1}^{n}\frac{\hat{\phi}_{n-m}}{6}\Big{(}6(2-3m)\hat{x}_{m}+(3m-2)\hat{y}_{m-1}\Big{)},\qquad\text{for
$~{}n\geq 2$}.$ Lastly, a recurrence relation for $\hat{y}_{n}$ is found from
equation (42a) to be
$\hat{y}_{1}=1,\qquad\hat{y}_{n}=\sum_{m=0}^{n}\hat{\phi}_{m}\hat{\phi}_{n-m},\qquad\text{for
$~{}n\geq 2$}.$ (45c)
Assuming that $\hat{\phi}_{n-1}$, $\hat{y}_{n-1}$, and $\hat{x}_{n-1}$ are
known, $\hat{x}_{n}$ can be determined from equation (45a), which then yields
a value for $\hat{\phi}_{n}$ from equation (45b). Lastly, $\hat{y}_{n}$ is
found by evaluating equation (45c).
### A.2 Matching and determination of the constant $\Lambda$
We now match the outer limit of the inner solution, $\hat{\phi}$, with the
inner limit of the outer solution, $\phi^{\prime}$. In writing the outer limit
of the inner solution in outer variables, we have
$\phi^{\prime}=\frac{-a\mathrm{i}\Gamma_{\text{c}}}{2\pi}\sum_{n=0}^{\infty}\frac{F^{2n}\hat{\phi}_{n}}{\Big{(}-\frac{4a\mathrm{i}\pi^{2}}{3\Gamma_{\text{c}}^{2}}\Big{)}^{n}(\sigma-a\mathrm{i})^{3n+1}},$
(46)
and for the inner limit of the outer solution,
$\displaystyle\phi^{\prime}=\sum_{n=0}^{\infty}F^{2n}\phi_{n}^{\prime}$
$\displaystyle\sim\sum_{n=0}^{\infty}-F^{2n}\chi^{\prime}\Phi\frac{\Gamma(n+\alpha+1)}{\chi^{n+\alpha+1}}$
(47)
$\displaystyle\sim\sum_{n=0}^{\infty}-\frac{4\pi^{2}\Lambda}{\Gamma_{\text{c}}^{2}(-a\mathrm{i})^{1/2}}\mathrm{e}^{\mathcal{P}(a\mathrm{i})}\frac{F^{2n}\Gamma{(n+\alpha+1)}}{\Big{(}-\frac{4a\mathrm{i}\pi^{2}}{3\Gamma_{\text{c}}^{2}}\Big{)}^{n+\alpha+1}(\sigma-a\mathrm{i})^{3n+3\alpha-1/2}}.$
In the above, the inner limit of the amplitude function $\Phi$ from equation
(17) has been taken by defining
$\mathcal{P}(\sigma)=\int_{0}^{\sigma}\bigg{[}\frac{a\mathrm{i}\phi_{1}^{\prime}(t)}{[\phi_{0}^{\prime}(t)]^{3}}-\frac{3}{2(t-a\mathrm{i})}\bigg{]}\mathrm{d}t,$
(48)
such that $\mathcal{P}(\sigma)=O(1)$ as $\sigma\to a\mathrm{i}$. Matching (46)
with (47) determines the constant, $\Lambda$, as
$\Lambda=-\frac{\Gamma_{\text{c}}(-a\mathrm{i})^{1/2}\mathrm{e}^{-\mathcal{P}(a\mathrm{i})}}{6\pi}\bigg{(}-\frac{4a\mathrm{i}\pi^{2}}{3\Gamma_{\text{c}}^{2}}\bigg{)}^{\alpha}\lim_{n\to\infty}\bigg{(}\frac{\hat{\phi}_{n}}{\Gamma(n+\alpha+1)}\bigg{)}.$
(49)
|
# LensLeech: On-Lens Interaction for Arbitrary Camera Devices
Christopher Getschmann<EMAIL_ADDRESS>0000-0002-0174-5974 Aalborg
UniversityAalborgDenmark and Florian Echtler<EMAIL_ADDRESS>Aalborg
UniversityAalborgDenmark
(2023)
###### Abstract.
Cameras provide a vast amount of information at high rates and are part of
many specialized or general-purpose devices. This versatility makes them
suitable for many interaction scenarios, yet they are constrained by geometry
and require objects to keep a minimum distance for focusing. We present the
LensLeech, a soft silicone cylinder that can be placed directly on or above
lenses. The clear body itself acts as a lens to focus a marker pattern from
its surface into the camera it sits on. This allows us to detect rotation,
translation, and deformation-based gestures such as pressing or squeezing the
soft silicone. We discuss design requirements, describe fabrication processes,
and report on the limitations of such on-lens widgets. To demonstrate the
versatility of LensLeeches, we built prototypes to show application examples
for wearable cameras, smartphones, and interchangeable-lens cameras, extending
existing devices by providing both optical input and output for new
functionality.
Mobile Interfaces, Elastomer Sensors, Optical Widgets
††copyright: acmcopyright††journalyear: 2023††doi:
XXXXXXX.XXXXXXX††conference: Eighteenth International Conference on Tangible,
Embedded, and Embodied Interaction; February 11–14, 2024; Cork,
Ireland††price: 15.00††isbn: 978-1-4503-XXXX-X/18/06††ccs: Human-centered
computing Interaction devices††ccs: Human-centered computing Ubiquitous and
mobile devices††ccs: Hardware Emerging interfaces Figure 1. a) The soft
silicone attachment with an integrated lens can be placed directly on and
above cameras. b) A marker point pattern is focused by the silicone lens into
the camera. c) Arbitrary devices with cameras, such as a smartphone can track
the position, rotation, and deformation of the silicone attachment on the
camera to sense input.
## 1\. Introduction
There is an increasing number of mobile devices that make use of cameras as
primary or additional sensors. At the same time, physical input has become a
scarce feature on modern, highly-integrated devices. Many of these camera
devices are limited in their input channels and the interaction techniques
they can offer due to trade-offs and design decisions to make them smaller,
more robust, or less expensive. Very small devices such as action cameras or
wearables have either very few buttons or small touch screens, requiring to
delegate even basic input tasks to paired smartphones or suffer from the fat
finger problem (Siek et al., 2005). Larger, interchangeable-lens cameras do
provide both larger touchscreens and more buttons but could benefit from
additional input options such as back-of-device interaction as well (Baudisch
and Chu, 2009). Smartphones and tablets with capacitive touchscreens offer no
physical feedback, have reachability issues (Wobbrock et al., 2008), and
require visual confirmation for input (Buxton et al., 1985). While all of
these devices could benefit from additional physical input, they have in
common that they include a powerful sensor: a camera. Yet, any camera requires
a minimal focal distance to provide focused images necessary to process rich
user input (Xiao et al., 2013; Yamada et al., 2018).
We present the LensLeech, a soft silicone attachment that can be placed
directly on or above the front element of camera lenses. The silicone is
optically clear and deformable so forces applied by fingers, hands, or
arbitrary objects can be detected visually. By using the lower surface of the
silicone body as a close-focus lens, a marker pattern on the opposite surface
is always in focus regardless of the minimum focal distance of the camera
lens.
With the LensLeech we can transform an unused or idle camera lens into a
button, a knob, or a d-pad widget (and reverse it in seconds). These widgets
provide physical feedback (when deformed), can be operated with gloves, and
are robust, versatile, and inexpensive. This allows to add knobs and d-pads to
small wearable cameras to change settings in situ, make lens caps for large
cameras touch-sensitive, or introduce novel optical attachments for
smartphones.
In summary, we contribute:
* •
a tangible deformation sensor to create buttons, knobs, and d-pads, combining
soft body, optical elements, and sensing pattern in a single object
* •
discussions on the design and fabrication of on-lens widgets
* •
an image processing pipeline for analyzing position, rotation, and deformation
* •
application examples for integration with new and existing devices
Many research approaches aim at providing novel functionality with new or
existing sensors for future devices, often built on the assumption or
requirement of a possible miniaturization and integration of this external
sensing hardware into a new device with a new form factor. However, we
explicitly aim at retrofitting existing and well-proven interaction techniques
to sensors that make them available both to legacy devices today as well as
new ones in the future. This could help to extend the lifetime of devices in
circulation by improving their usability and reducing incentives to update to
newer hardware prematurely. A user study (with regard to the input
capabilities of the widgets) is not presented as these input modalities are
well understood and can be directly applied to this new form factor.
The remainder of this paper is organized as follows: related work is discussed
with a broad overview of vision-based elastomer sensors and on-lens/around-
lens interaction techniques, then we explain our concept of soft silicone
attachments for on-lens interaction sensing. The image processing pipeline and
fabrication procedure are summarized subsequently. We built upon that by
presenting a set of scenarios and prototypes created to show real-world
applications. Finally, we discuss the limitations of using soft silicone
attachments for on-lens interaction and conclude with specific directions for
future work.
## 2\. Related Work
Relevant to the presented work are both optical deformation sensors primarily
developed for robotic applications as well as human-computer interaction
techniques and prototypes that gather input from the space on and around
camera lenses.
### 2.1. Optical Elastomer Sensors
Optical deformation sensing of soft materials is performed either by measuring
light altered by the surface or by detecting displacement of high-contrast
markers, on the surface or encapsulated in the material. Surface deformation
measurements have been proposed based on total internal reflection (Hiraishi
et al., 1988), Lambertian reflection (Johnson and Adelson, 2009; Watanabe et
al., 2014; Dong et al., 2017; Donlon et al., 2018; Taylor et al., 2021; Wang
et al., 2021) and polarization (Sato et al., 2009). The most common type of
sensor, the GelSight family, makes use of Lambertian reflection by coating the
clear elastomer with a reflective membrane. Multispectral illumination from
below allows to derive deformation depth and thus a detailed 2.5d geometry of
the reflective surface. For marker-based sensing, high-contrast points are
painted on the clear surface of the elastomer (Wang et al., 2021; Taylor et
al., 2021; Dong et al., 2017), on the interior of an opaque hull for TacTip
sensors (Winstone et al., 2012; Ward-Cherrier et al., 2018) or colored balls
are directly encapsulated in the soft material (Kamiyama et al., 2004;
Sferrazza and D’Andrea, 2019; Yamaguchi and Atkeson, 2017).
These sensors have been used extensively for tactile sensing in robotic
applications, mounting the sensor on the end effector to measure gripping
force and detect slipping. For this, the sensor assembly is designed as a
monolithic unit consisting of camera sensor, lens, and elastomer block. While
mirrors (Donlon et al., 2018; Wang et al., 2021) and fisheye lenses (Taylor et
al., 2021) have been used to shorten optical paths to create more compact
grippers these sensors are still of considerable size and rely on a tight
integration of all components, making them incompatible with arbitrary
cameras. Modular approaches offer only exchangeable elastomers while still
using a specialized camera (Lambeta et al., 2020). Additionally, all gel-based
sensors with the exception of the sensor by Obinata et al. (Obinata et al.,
2007) and Fingervision (Yamaguchi and Atkeson, 2017) block environmental light
and require white, RGB or ultraviolet illumination by integrated LEDs. For a
detailed overview refer to the reviews by Shimonomura (Shimonomura, 2019) and
Abad et al. (Abad and Ranasinghe, 2020).
In the domain of human-computer interaction, elastomer sensors have been used
for interactive surfaces (Follmer et al., 2011), clay-like projection displays
(Punpongsanon et al., 2013) and tangibles on tabletops (Weiss et al., 2009;
Hennecke et al., 2011) to support novel interaction techniques.
### 2.2. On-Lens/Around-Lens Interaction
Placing a fingertip directly on a smartphone camera lens has been proposed as
an interaction technique in LensGestures (Xiao et al., 2013). The unfocused
environmental light passing through a finger’s tissue is used to approximate
finger positions and recognize gestures. CamTrackPoint (Yamada et al., 2018)
improves on this concept by providing tactile feedback. A spring-actuated
plastic ring is integrated with a smartphone case directly over the lens for
the finger to rest on. The thin ring blocks light with a sharp transition to
black and provides a higher precision compared to tracking the blurred finger.
A proof-of-concept for more complex on-lens input techniques is presented by
Watanabe et al. (Watanabe et al., 2014): soft and optically clear toys with a
reflective surface coating are placed on the camera while a neural network is
trained to recognize deformation/gestures from internal reflections observed
through a hole in the bottom. This represents the simplest and most basic on-
lens widget: unfocused, untagged, unpowered and depending on natural
illumination, but very easy to manufacture and not obstructing the camera when
not in use.
Interaction in the space around lenses requires mirrors to both shorten the
optical path and redirect light. Clipwidgets (Visschedijk et al., 2022) makes
use of a conical mirror in a bulky smartphone case to read the state of
physical widgets such as buttons and sliders. Similar approaches have been
presented for back-of-device interaction concepts with smartphones (Wong et
al., 2016; Matsushima et al., 2017; Kitade and Yamada, 2019). Without relying
on physical input objects Handsee (Yu et al., 2019) utilizes a prism to track
hands touching and floating above a smartphone display while Surroundsee (Yang
et al., 2013) tracks objects in the whole room with a circular 360-degree
mirror above the smartphone camera.
Similar techniques have been used without mirrors or lenses in the context of
tangibles with silicone feet for pressure sensing (Weiss et al., 2009),
deformation sensing on small wearables (Weigel and Steimle, 2017), and surface
position sensing with fibers (Wimmer, 2010). Other work that is related to the
presented concept is Bokode (Mohan et al., 2009), a marker made of a lenslet
and microfilm which magnifies a grid of 2D barcodes into the defocused lens of
a camera and Sauron (Savage et al., 2013), a design tool to integrate cameras
in hollow objects that read the state of mechanical input elements.
While physical input similar to the LensLeech can be achieved on smartphones
in particular by simply redirecting electrodes of the capacitive touchscreen
(Yu et al., 2011; Schmitz et al., 2017; Matsushima et al., 2017), the
LensLeech is not limited to touchscreens and can be applied across a range of
devices, see the application examples in section 5.
Since on-lens interaction concepts such as CamTrackPoint and LensGestures make
use of unfocused light, they are limited in their expressiveness due to the
low amount of information available. While they are suitable for smartphones
and their scratch-resistant camera assemblies, these concepts translate poorly
to interchangeable lens cameras or action cams with lens front elements often
using coatings sensitive to scratches or prints from fingertips. This is one
of the fundamental issues we intend to address with our generalizable
approach.
## 3\. The design of on-lens widgets
Figure 2. Illustrative ray diagram of the combined optical system. The field
of view inside the silicone (dashed line) depends on the field of view of the
camera, the position and diameter of the entrance pupil, as well as the
distance between silicone and camera lens.
We propose that any physical attachment enabling on or around-lens interaction
with both existing and future devices should—ideally—adhere to these basic
design considerations:
* •
safe to use near or on optical components and providing credible reassurance
to the user about this. This is a prerequisite for user acceptance.
* •
non-invasive, requiring no hardware modifications of the host device or its
camera. This ensures compatibility with existing devices that benefit most
from optical attachments.
* •
passive and unpowered, requiring only ambient illumination (if possible) to
reduce size and complexity.
* •
universal; compatible with arbitrary camera/lens combinations across a wide
range of device types.
Elastomer sensors in general fulfill the first and most important of these
requirements by virtue of their nature: they are soft. However, existing
sensors fall short in most or all other points.
As discussed, these sensors combine camera and elastomer in a permanent
assembly with a fixed position and rotation, limiting the way objects can
interact with them. Additionally, they make use of known sensor and lens
combinations to allow camera calibration and optimizations of sensor geometry
(for example by backprojecting through a calibrated lens to find optimal
marker point placements). This makes these sensors more precise and reliable
but prevents them from being used with arbitrary lenses and cameras. Finally,
most sensors require constant internal illumination. Reflective membranes
(GelSight) and rubber skins (TacTip) are blocking ambient light to avoid
interference. Only sensors relying solely on point patterns (Obinata et al.,
2007; Yamaguchi and Atkeson, 2017) can tolerate ambient illumination.
We propose an elastomer sensor design suitable for interaction sensing. The
LensLeech is a tangible soft input device that resembles the gel part of an
elastomer sensor. Our all-silicone design combines a lens, compliant body, and
a colored marker pattern in a single unit (see fig. 2). This addresses all
design requirements at the cost of reduced reliability and precision compared
to elastomer sensor assemblies that are designed to measure precise gripping
forces on robot actuators.
The small form factor of the LensLeech attachment (33mm diameter, 25.5mm
height) makes it easy to grip it with two fingers and place it on a camera. By
using the lower surface of the clear silicone body as a lens, light reflected
by the deformation-sensing pattern on the surface is collimated and can be
focused on the sensor at any distance from the camera. This makes it possible
to place the silicone foot of the LensLeech directly on or slightly above the
front element of a wide range of lenses. The combined optical system of
sensor, camera lens, silicone lens, and deformation sensing pattern is limited
by the field of view of the camera, its entrance pupil, and the distance to
the silicone attachment. This is discussed in more detail in section 6
(Limitations).
Marker Pattern
When choosing a marker pattern for deformation sensing, we need to take into
account that a positive lens required to move the focal point to the surface
of the silicone body will introduce a strong magnification effect. This
amplifies any defects or irregularities in the pattern and requires the
fabrication of very small features. The most precise and reliable method is
the deposition of single droplets of silicone paint. This makes a point
pattern the preferred choice.
A point pattern is used by other optical tactile sensors such as TacTip and
GelSight as well, however, these sensors are fixed assemblies that can compute
deviations from a static reference frame. This does not apply to a silicone
sensor that can be moved and rotated freely, thus a method to align the
currently visible region of interest with the overall marker grid is required.
A common method for identifying sections of point grids are two-dimensional
DeBruijn sequences. These are sequences that contain every subsequence of a
defined size at most once. Printed as microdots on paper these have been used
for position-tracking with digital pens (Petterson and Edso, 2003) (encoding
bits as a displacement from a regular grid) and tangibles (Schüsselbauer et
al., 2021) (encoding bits as black and white). However, unlike a rigid piece
of paper which allows displacement coding of dots, the soft silicone is easily
bent or compressed and requires coding by color or contrast.
While a hexagonal arrangement of points offers the densest packing it is
incompatible with a 2D-DeBruijn sequence. Hence, we computed a DeBruijn-like
pattern with 7-point hexagons instead of 3x3 matrices using a brute-force
approach. Each overlapping hexagonal sliding window in the pattern is unique
in the given rotation (see fig. 3). An optimal pattern does contain only
hexagons that are unique in all rotations which is simplifying pattern
matching during image processing, however, this requires a minimum of three
colors at a suitable pattern size. Higher robustness to adverse lighting
conditions and a less error-prone fabrication with only two different paints
is the reason why a less-than-optimal two-color pattern is preferable.
Our hexagon patterns consist of 127 points and require 91 unique sliding
windows. This is sufficient to cover the visible region of the silicone
attachment even when placed on a wide-angle camera. While other sensors such
as TacTip and GelSight Wedge use the same or similar number of points, only a
subset of points is visible at a time for our application due to the
magnification of the silicone lens. If a unique center hexagon is enforced
during pattern generation up to 28 patterns can be discerned. This allows to
map different silicone attachments to specific input modalities.
Figure 3. Each hexagonal sliding window appears only once. Some sliding
windows are unique in all six orientations, some can be found in multiple
locations when rotated.
Image Processing
The DeBruijn-like point pattern is color-coded in blue and green to offer a
high contrast across the range of human skin tones. Coincidentally, the
fingertips have fewer variations due to smaller differences in the skin tones
of palms overall. The detection and classification of the points is the first
step of the image processing pipeline. Background removal is performed by
thresholding in the HSV color space. The diffuse top surface of the silicone
body improves this step considerably without blocking any ambient light. From
these point candidates, colors are extracted and classified by thresholding
the two classes in the hue component of the HSV colorspace using Otsu’s method
(Otsu, 1979). This is robust to errors in white balance caused by tinted
ambient illumination or fingertips and computationally less expensive than
other classification methods. Robustness is especially important since most
auto white balance algorithms overshoot for several dozen frames when a finger
is placed on the point pattern.
For pattern matching each detected point is grouped with its 6 closest
neighbors and all 6 rotation variants are checked against a lookup table.
Correct rotation is assumed when the highest number of matches is found
between neighboring sliding windows in the camera image and ground truth
pattern. Given an optimal pattern, only one rotation would result in a match
(in the absence of any errors), yet this computationally-expensive step is
necessary to limit the pattern to only two colors. This makes the pipeline’s
processing speed highly dependent on the number of detected points. On a
laptop computer (2.3 GHz 8-Core Intel i9) 34 frames per second are processed
when 39 points are visible (30% of the pattern) and 21 FPS when 69 points
(70%) are visible. Smartphone performance numbers are not reported since the
Android implementation performs segmentation on-device but outsources the
pattern-matching step of the pipeline to a server.
Figure 4. The four types of input: a) Pressing on the silicone body b) lateral
pushing in any direction c) rotation on the optical axis d) squeezing the
silicone.
The input gestures are derived directly from the matched point pattern (see
fig. 4). A press on the top is recognized by detecting locally increased
distances between neighboring points, pushing sideways by computing the
centroid of all detected points, rotation by Kabsch‘s algorithm (Kabsch,
1976), and squeeze by a global change of point distances along the squeeze
axis. The gesture detection relies on algorithm implementations in SciPy
(Virtanen et al., 2020), while processing of image data is done using OpenCV
(Bradski, 2000).
Before discussing how these input types inform examples for real-world
application, the fabrication process of both silicone body and color-coding
pattern is described briefly.
## 4\. Fabrication
Figure 5. Cross section of the mold. The curved surfaces are ground and
polished each with a precision steel ball of the required curvature. The
optical surface is polished to a 2-micron finish and the diffuse surface to 40
microns. All three sections of the mold are aligned with metal dowel pins (not
pictured). Liquid silicone is poured through a horizontal channel in the
3d-printed plastic part.
The clear silicone body is created by mixing, degassing, and pouring liquid
silicone (Trollfactory Type 19) into a mold and letting it cure. The mold
itself requires two precisely manufactured features. The lower cavity is an
optical surface (sufficiently smooth to refract light for imaging
applications) to create the spherical convex lens of 7.5mm radius for
focusing. The curved top surface of 30mm radius diffuses light. Both surfaces
are CNC-milled from acrylic before being ground and polished. For this, the
spherical surface of the acrylic part is coated with lapping paste and pressed
against a rotating steel ball of matching radius (widely available as high-
precision replacement parts for large ball bearings). After polishing the
acrylic plates are fastened to a 3d-printed center part to complete the mold
(see fig. 5). Once cured and de-molded the point pattern is applied to the
clear silicone body with two 3d-stencils milled from acrylic (one per color).
The stencil is fabricated by drilling a duplicate of the mold top part with a
circuit board drill (1.0mm) to create channels. The soft body is pressed into
the matching cavity of the stencil from below and the pigmented silicone can
be poured on the channels (see fig. 6) before removing the remaining air from
the channels in a vacuum chamber. The stencil guarantees correct placement and
uniform point size. Only silicone itself bonds reliably to cured silicone
parts, thus uncured silicone mixed with color pigments is the most suitable
paint. The main issues in this process are ensuring that the high-viscosity
silicone reliably fills the channels and avoiding oversaturation of the
silicone with pigments in a silicone oil solution, which may inhibit the
curing process. A mixture (by weight) of Smooth-On’s Psycho Paint silicone
with 15 percent dry UV-reactive pigment powder and 25 percent solvent
(toluene) to lower the viscosity worked best. After 24 hours the pigmented
silicone binds reliably to the optically-clear silicone body, creating a point
pattern on the surface that is flexible and wear-resistant.
Figure 6. Cross section of the stenciling fixture. The silicone body is
pressed upwards against the curved surface to create a seal. Once locked in
place by clamps, the liquid pigmented silicone is poured into the recess at
the top of the stencil and makes its way through the micro-drilled channels.
When the stencil is lifted a small domed blob of partially-cured silicone
paint remains on the surface. The stencil and fixture for the silicone body
are aligned with metal dowel pins to ensure precise placement for each
consecutive stencil and color.
The choice of lens curvature during mold production is a trade-off. A mold for
a lens with a stronger curvature is more demanding in fabrication but the
shorter focal length allows to reduce the height of the silicone body. At the
same time, it decreases the depth of field and the field of view, allowing to
track a lower number of points in the pattern. Additionally, interacting with
the LensLeech deforms both top surface and lens. A strong press on the top
will reduce the height of the body by several millimeters depending on the
hardness of the silicone. If the height of the silicone body does not match
the focal length the light will not exit the system collimated, resulting in a
pattern that would be out of focus. In reality, this is rarely an issue since
autofocus cameras can compensate for this, fixed-focus cameras often have a
sufficient depth of field, and the image processing pipeline is robust to low
levels of blurring.
A paraxial approximation of the focal length can be obtained using the
lensmaker’s equation. Only the refraction of the first surface is relevant for
the LensLeech geometry, so a thin, plano-convex lens in air
($d=0,R_{2}=\infty$) can be assumed:
$\frac{1}{f}=(n-1)\left(\frac{1}{R_{1}}-\frac{1}{R_{2}}+\frac{(n-1)d}{nR_{1}R_{2}}\right)=\frac{n}{R_{1}}-\frac{1}{R_{1}}$
We have chosen a radius of $R_{1}=7.5mm$ for the lens surface and assume that
Trollfactory Type 19 has a refractive index of $n=1.41$, similar to other
platinum-cure silicones. This would result in a total focal length of 18.29mm.
To account for the deformation during interaction and the strong curvature of
the lens we increase the height of the silicone body by a factor of $1.3$ to
25mm. While the LensLeech should be able to touch the glass surface of the
camera lens, the silicone lens surface in the center requires an air gap to
refract light. Thus we extend the foot around the lens by 1.0mm to account for
deformations (cross section can be seen in fig. 2). This provided the most
reliable results during testing for high forces when pressing and squeezing
while still keeping the point pattern well within the depth of field of most
cameras when not deformed.
While fabrication is the primary challenge to making close-focus silicone lens
attachments, a thorough description would go beyond the scope of this paper.
Please refer to the companion
repository111https://github.com/volzotan/LensLeech for detailed information
about the fabrication process.
## 5\. Application Examples
Tactile on-lens input can be used in a variety of scenarios for devices with
cameras in many sizes. We present two application examples that show how on-
lens interaction can be utilized to make input on both large and small cameras
more convenient and transfer well-established tangible interaction techniques
to smartphones.
Figure 7. a) Cross section: the silicone attachment can be extended with a
3d-printed shoe matched to the specific device so it slides over the
protruding wide-angle lens of an action camera. b) The LensLeech could be used
as a rotating knob, press to confirm, squeeze to cancel.
Interactive Lens Caps for Digital Cameras
Small action cameras offer a very limited number of buttons, a tiny display
(if any), and only optionally, a touch interface on this display. While there
are techniques to facilitate touch input on very small displays (Baudisch and
Chu, 2009), it may be cumbersome. Adjusting settings is often performed via a
companion app on a smartphone which is paired with the wearable camera.
However, there are scenarios in which the phone is unavailable and direct
interaction with the device itself may be favorable. These can be casual,
everyday situations like wearing gloves or very specific use cases such as
interacting with a camera enclosed in a waterproof housing while swimming or
diving. By adopting a lens cap or protective storage case that integrates a
LensLeech (see fig. 7), we can add tangible controls such as a rotation knob
or a d-pad, and extend the number of buttons on the device for easier
navigation through nested menus. However, note that in the case of underwater
usage when the LensLeech is pressed against a waterproof housing, a different
silicone lens curvature will be required due to the refractive index of water.
Figure 8. a) The silicone attachment is placed in a 3d-printed lens cap with a
spring-loaded mechanism to allow lateral movement and rotation. b) The lens
cap on an interchangeable-lens camera.
This concept extends to larger cameras as well. Many digital consumer cameras
are infamous for convoluted menus and poor usability in general. Browsing
recorded videos and photos, changing settings in nested menus, and entering
credentials to set up wireless connections requires prolonged attention and
interaction while the sensor itself is not in use during these tasks. By
integrating a silicone attachment into a lens cap we can leverage the unused
hardware without interfering with the primary use case of the camera. Rotating
or pushing the silicone sideways could be used to traverse large lists of
settings or captured footage (see fig. 8). While the soft silicone can touch
lens coatings without damaging them, the lens cap in this case prevents direct
contact and may provide reassurance to the user for these very expensive
lenses.
Hybrid Viewfinders for Smartphones
While the LensLeech provides optical input to camera-based devices, it can be
combined with other components that offer optical output as well. This allows
to create complex passive optical add-ons to existing devices. The hybrid
viewfinder slides over the top section of a smartphone covering the front lens
and a portion of the display. By adding a beamsplitter prism to a camera
viewfinder, a section of the covered display can be reflected into the
viewfinder’s optical path (see fig. 9a) to create a hybrid optical/electronic
viewfinder for smartphone photography. By integrating the LensLeech into the
viewfinder attachment, the front camera can be used for input while the rear
camera takes images and optionally provides data for the viewfinder overlay
(see fig. 9c). Rotating the LensLeech changes the data overlay and pressing it
triggers image capture. This allows optical input and output with no hardware
modifications, transforming a smartphone into a modern rangefinder-style
camera.
Figure 9. a) Cross section of the hybrid viewfinder. The beamsplitter overlays
the light emitted by a section of the smartphone screen (blue) over the
viewfinder image of the world (yellow), while the silicone attachment rests on
the front camera (red). b) The finder can be slid over the top section of the
smartphone. The silicone attachment provides rich tangible input to control
settings and take a photo without visual confirmation as a touchscreen would
require. c) View through the finder showing an overlay of the selection menu
and a digital spirit level.
## 6\. Evaluation & Limitations
Figure 10. Mean number of identified points in relation to ambient
illumination strength. Error bars specify standard deviation. Monochromatic or
environmental light with a strong color tint will result in an undetectable
point pattern regardless of illumination strength, this is the main cause for
outliers in the plot. Note: the maximum number of points visible to the camera
varies across devices depending on pupil size and field of view.
When tested with artificially generated images (rendered images of the
deformation point pattern with a pinhole aperture instead of a lens for
focusing) the rotational error is negligible at an average of 0.03 degrees.
More relevant and considerably more challenging is the real-world performance
under low-light conditions and ambient light with color casts. As an
evaluation setup, three cameras with an attached LensLeech (Pixel 3a
smartphone, Sony A6000 + Sony 20mm 2.8 digital still camera, Raspberry Pi V1
embedded camera) were placed in complete darkness facing a display showing a
subset (201 images, three per category) of the MIT indoor scene recognition
dataset (Quattoni and Torralba, 2009). These were artificially darkened and
brightened to simulate a low-light environment (resulting in a total of 1206
images). The illuminance of each scene was measured at the surface of the
LensLeech with a TSL2591 ambient light sensor. Detection performance is
depending on the combination of sensor, lens, and environment, but in general,
it can be observed that above 150 lux reliable operation can be expected (see
fig.10). Indoor lighting conditions usually exceed 150 lux while 300-500 lux
are recommended for office work (ISO 8995-1:2002, 2002).
Figure 11. Three lenses with an identical field of view of 84° but increasing
entrance pupil diameters: a) Google Pixel 3a front camera b) Sony SEL-P1650
lens (16mm focal length) c) Sigma 16mm 1.4 DC DN (16mm focal length). For all
three images, the LensLeech is resting directly on the front element of the
camera lens.
The main limitation when using any optical attachment on lenses is the
entrance pupil diameter and its distance from the first surface of the lens.
The entrance pupil is a virtual opening within the lens barrel through which
all entering light rays pass. Size and position within lenses can vary across
lens designs, even when an image of a distant object taken with different
lenses would look identical (see fig. 11). The silicone attachment (in the
size as presented) works well on small and medium-sized lenses but requires a
different geometry on very large lenses such as professional photography or
videography lenses with large front elements and entrance pupils for better
low-light performance. As a rule of thumb: if the image of the aperture seen
through the front element of the lens is considerably larger than the silicone
lens (12mm in diameter) the number of visible points is strongly reduced. In
general, a lower bound of 19 points is required to reliably recognize input
gestures through the soft widget. Additional limiting factors on the optical
system are the field of view of the lens and the curvature of the first glass
element of the lens. A camera with a narrow field of view will reduce the
number of visible points, similar to a large entrance pupil. This makes the
presented concept more suitable for medium to wide-angle systems such as
webcams, smart home devices, smartphones, and wearable cameras. If the
LensLeech is used with hard attachments (such as the lens cap) it does not sit
directly on the glass and a strong lens curvature is not an issue.
## 7\. Discussion
Compared to other on-lens interaction concepts such as CamTrackPoint (Yamada
et al., 2018) and LensGestures (Xiao et al., 2013)) that process unfocused
light, the LensLeech is less limited in the amount of information it provides
but it requires ambient light as well. While the LensLeech performs well in
most situations, LEDs (such as smartphone flashlights and autofocus-assist
lights of still cameras) or screens of devices can be used to provide
additional artificial illumination. This can be seen in the hybrid viewfinder:
a section of the covered display illuminates the point pattern from below.
Depending on the device and application scenario this might not be a viable
option. For usage within a predefined space, near-ultraviolet flood lights can
be installed to brighten the UV-reactive pigments in the point pattern (see
fig. 12) with little interference to the brightness of the environment.
In general, an attachment solely made from silicone is simple, robust, and—to
an extent—expendable. Material cost per piece is about 2 USD/EUR when
fabricated in small quantities. This makes the LensLeech comparable to other
inexpensive attachments for mobile devices that extend I/O capabilities, such
as Google Cardboard (Cardboard, 2014) or Nintendo Labo (Labo, 2018). Similar
to the limited lifetime of corrugated cardboard, a LensLeech and its point
pattern may eventually suffer from wear and tear after extended usage.
A possible negative perception of letting an object touch the front element of
the lens is not an issue when used on smartphone lenses. The application
examples show that in other scenarios it makes sense to use device-dependent
additions such as a rigid shoe on protruding lens barrels for action cameras
or lens caps on interchangeable-lens cameras.
Figure 12. Near-UV flood illumination can considerably increase the brightness
and contrast of the UV-reactive pigments in the point pattern while it only
marginally brightens the environment. a) No ultraviolet illumination b) Single
365nm-wavelength light source.
## 8\. Future Work
A LensLeech is uniformly made from a single silicone formula with consistent
Shore hardness throughout the whole body. By making use of a two-stage mold,
the lower part of the body containing the lens could be molded separately with
a harder type of clear silicone, resulting in a lower deformation of the lens
when compressed. Also by integrating air-filled cavities and compliant
elements in multi-stage molds, tactile feedback can be provided, resulting in
a sensation when a certain amount of force is applied.
The limitation of large entrance pupil sizes can be circumvented by replacing
the single silicone lens with a grid of smaller lenses. This requires a
different fabrication technique for the mold and a point pattern that is
aligned with camera lens angle and microlens position. This limits the
silicone attachment compatibility to only a single lens, yet this may not be
an issue for applications such as model-specific lens caps.
While only a single type of LensLeech is presented, the concept is versatile.
With additional illumination, microlens arrays would allow emulating a small
touchscreen on the top surface while using an angled surface geometry makes
sensing fingerprints possible. In the near future, the emergence of cameras
under displays in smartphones would allow the use of silicone attachments as
tangible input and output devices in tabletop-like scenarios. The display can
be used both for illuminating the LensLeech to sense in dark spaces as well as
to display output in or on the body itself by refracting and redirecting the
light.
## 9\. Conclusion
We presented the LensLeech, a soft silicone attachment that allows to sense
pushing, pressing, rotating, and squeezing when placed directly on or above
lenses of arbitrary cameras. This makes it possible to add tangible input
methods to a wide range of existing and new devices, especially small action
or lifelogging cameras and smartphones. We have shown application examples
ranging from small, body-worn devices to lens caps for large cameras and
complex smartphone attachments. While the attachments are limited in their
compatibility mainly by lens geometry, the low-light performance allows them
to be used with only ambient illumination without any need for hardware
modifications on a wide range of existing devices. This simple and inexpensive
approach opens up an interaction space on lenses for rich input that was
previously inaccessible with a range of further applications in the (soft)
robotics domain.
## Reproduction Note
The example applications, source code, CAD models of molds and fixtures, a
detailed description of the fabrication process, and data/scripts for
generating plots are available publicly:
https://github.com/volzotan/LensLeech
###### Acknowledgements.
This work was funded by the Deutsche Forschungsgemeinschaft (DFG, German
Research Foundation) through project EC437/1-1.
## References
* (1)
* Abad and Ranasinghe (2020) Alexander C. Abad and Anuradha Ranasinghe. 2020. Visuotactile Sensors With Emphasis on GelSight Sensor: A Review. _IEEE Sensors Journal_ 20, 14 (July 2020), 7628–7638. https://doi.org/10.1109/JSEN.2020.2979662
* Baudisch and Chu (2009) Patrick Baudisch and Gerry Chu. 2009. Back-of-Device Interaction Allows Creating Very Small Touch Devices. In _Proceedings of the SIGCHI Conference on Human Factors in Computing Systems_ _(CHI ’09)_. Association for Computing Machinery, New York, NY, USA, 1923–1932. https://doi.org/10.1145/1518701.1518995
* Bradski (2000) Gary Bradski. 2000\. The OpenCV Library. _Dr. Dobb’s Journal of Software Tools_ (2000).
* Buxton et al. (1985) William Buxton, Ralph Hill, and Peter Rowley. 1985\. Issues and Techniques in Touch-Sensitive Tablet Input. In _Proceedings of the 12th Annual Conference on Computer Graphics and Interactive Techniques_ _(SIGGRAPH ’85)_. Association for Computing Machinery, New York, NY, USA, 215–224. https://doi.org/10.1145/325334.325239
* Cardboard (2014) Google Cardboard. 2014\. Website. Retrieved Apr 5, 2023 from https://arvr.google.com/cardboard/.
* Dong et al. (2017) Siyuan Dong, Wenzhen Yuan, and Edward H. Adelson. 2017\. Improved GelSight Tactile Sensor for Measuring Geometry and Slip. In _2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS)_. 137–144. https://doi.org/10.1109/IROS.2017.8202149
* Donlon et al. (2018) Elliott Donlon, Siyuan Dong, Melody Liu, Jianhua Li, Edward Adelson, and Alberto Rodriguez. 2018\. GelSlim: A High-Resolution, Compact, Robust, and Calibrated Tactile-sensing Finger. In _2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS)_. 1927–1934. https://doi.org/10.1109/IROS.2018.8593661
* Follmer et al. (2011) Sean Follmer, Micah Johnson, Edward Adelson, and Hiroshi Ishii. 2011. deForm: An Interactive Malleable Surface for Capturing 2.5D Arbitrary Objects, Tools and Touch. In _Proceedings of the 24th Annual ACM Symposium on User Interface Software and Technology_ _(UIST ’11)_. Association for Computing Machinery, New York, NY, USA, 527–536. https://doi.org/10.1145/2047196.2047265
* Hennecke et al. (2011) Fabian Hennecke, Franz Berwein, and Andreas Butz. 2011\. Optical Pressure Sensing for Tangible User Interfaces. In _Proceedings of the ACM International Conference on Interactive Tabletops and Surfaces_ _(ITS ’11)_. Association for Computing Machinery, New York, NY, USA, 45–48. https://doi.org/10.1145/2076354.2076362
* Hiraishi et al. (1988) H. Hiraishi, N. Suzuki, M. Kaneko, and Kazuo Tanie. 1988\. An Object Profile Detection by a High Resolution Tactile Sensor Using an Optical Conductive Plate. In _Proceedings.14 Annual Conference of Industrial Electronics Society_ , Vol. 4. 982–987. https://doi.org/10.1109/IECON.1988.666278
* ISO 8995-1:2002 (2002) ISO 8995-1:2002 2002. _Lighting of work places – Indoor: Part 1_. Standard. International Organization for Standardization, Geneva, CH.
* Johnson and Adelson (2009) Micah K. Johnson and Edward H. Adelson. 2009. Retrographic Sensing for the Measurement of Surface Texture and Shape. In _2009 IEEE Conference on Computer Vision and Pattern Recognition_. 1070–1077. https://doi.org/10.1109/CVPR.2009.5206534
* Kabsch (1976) Wolfgang Kabsch. 1976\. A Solution for the Best Rotation to Relate Two Sets of Vectors. _Acta Crystallographica Section A: Crystal Physics, Diffraction, Theoretical and General Crystallography_ 32, 5 (Sept. 1976), 922–923. https://doi.org/10.1107/S0567739476001873
* Kamiyama et al. (2004) Kazuto Kamiyama, Hiroyuki Kajimoto, Naoki Kawakami, and Susumu Tachi. 2004. Evaluation of a Vision-Based Tactile Sensor. In _IEEE International Conference on Robotics and Automation, 2004. Proceedings. ICRA ’04. 2004_ , Vol. 2. 1542–1547 Vol.2. https://doi.org/10.1109/ROBOT.2004.1308043
* Kitade and Yamada (2019) Takuya Kitade and Wataru Yamada. 2019. Prismodule: Modular UI for Smartphones Using Internal Reflection. In _The Adjunct Publication of the 32nd Annual ACM Symposium on User Interface Software and Technology_ _(UIST ’19)_. Association for Computing Machinery, New York, NY, USA, 119–121. https://doi.org/10.1145/3332167.3356892
* Labo (2018) Nintendo Labo. 2018\. Website. Retrieved Apr 5, 2023 from https://web.archive.org/web/20180117220613/https://labo.nintendo.com/.
* Lambeta et al. (2020) Mike Lambeta, Po-Wei Chou, Stephen Tian, Brian Yang, Benjamin Maloon, Victoria Rose Most, Dave Stroud, Raymond Santos, Ahmad Byagowi, Gregg Kammerer, Dinesh Jayaraman, and Roberto Calandra. 2020. DIGIT: A Novel Design for a Low-Cost Compact High-Resolution Tactile Sensor With Application to In-Hand Manipulation. _IEEE Robotics and Automation Letters_ 5, 3 (July 2020), 3838–3845. https://doi.org/10.1109/LRA.2020.2977257
* Matsushima et al. (2017) Nobutaka Matsushima, Wataru Yamada, and Hiroyuki Manabe. 2017\. Attaching Objects to Smartphones Back Side for a Modular Interface. In _Adjunct Publication of the 30th Annual ACM Symposium on User Interface Software and Technology_ _(UIST ’17)_. Association for Computing Machinery, New York, NY, USA, 51–52. https://doi.org/10.1145/3131785.3131810
* Mohan et al. (2009) Ankit Mohan, Grace Woo, Shinsaku Hiura, Quinn Smithwick, and Ramesh Raskar. 2009. Bokode: Imperceptible Visual Tags for Camera Based Interaction from a Distance. In _ACM SIGGRAPH 2009 Papers_ _(SIGGRAPH ’09)_. Association for Computing Machinery, New York, NY, USA, 1–8. https://doi.org/10.1145/1576246.1531404
* Obinata et al. (2007) Goro Obinata, Ashish Dutta, Norinao Watanabe, and Nobuhiko Moriyam. 2007. Vision Based Tactile Sensor Using Transparent Elastic Fingertip for Dexterous Handling. In _Mobile Robots: Perception & Navigation_, Sascha Kolski (Ed.). Pro Literatur Verlag, Germany / ARS, Austria. https://doi.org/10.5772/4771
* Otsu (1979) Nobuyuki Otsu. 1979\. A Threshold Selection Method from Gray-Level Histograms. _IEEE Transactions on Systems, Man, and Cybernetics_ 9, 1 (Jan. 1979), 62–66. https://doi.org/10.1109/TSMC.1979.4310076
* Petterson and Edso (2003) Mats Petter Petterson and Tomas Edso. 2003. Patent US-6548768-B1: Determination of a Position Code.
* Punpongsanon et al. (2013) Parinya Punpongsanon, Daisuke Iwai, and Kosuke Sato. 2013\. DeforMe: Projection-Based Visualization of Deformable Surfaces Using Invisible Textures. In _SIGGRAPH Asia 2013 Emerging Technologies_ _(SA ’13)_. Association for Computing Machinery, New York, NY, USA, 1–3. https://doi.org/10.1145/2542284.2542292
* Quattoni and Torralba (2009) Ariadna Quattoni and Antonio Torralba. 2009. Recognizing Indoor Scenes. In _2009 IEEE Conference on Computer Vision and Pattern Recognition_. 413–420. https://doi.org/10.1109/CVPR.2009.5206537
* Sato et al. (2009) Toshiki Sato, Haruko Mamiya, Hideki Koike, and Kentaro Fukuchi. 2009. PhotoelasticTouch: Transparent Rubbery Tangible Interface Using an LCD and Photoelasticity. In _Proceedings of the 22nd Annual ACM Symposium on User Interface Software and Technology_ _(UIST ’09)_. Association for Computing Machinery, New York, NY, USA, 43–50. https://doi.org/10.1145/1622176.1622185
* Savage et al. (2013) Valkyrie Savage, Colin Chang, and Björn Hartmann. 2013\. Sauron: Embedded Single-Camera Sensing of Printed Physical User Interfaces. In _Proceedings of the 26th Annual ACM Symposium on User Interface Software and Technology_ _(UIST ’13)_. Association for Computing Machinery, New York, NY, USA, 447–456. https://doi.org/10.1145/2501988.2501992
* Schmitz et al. (2017) Martin Schmitz, Jürgen Steimle, Jochen Huber, Niloofar Dezfuli, and Max Mühlhäuser. 2017\. Flexibles: Deformation-Aware 3D-Printed Tangibles for Capacitive Touchscreens. In _Proceedings of the 2017 CHI Conference on Human Factors in Computing Systems_ _(CHI ’17)_. Association for Computing Machinery, New York, NY, USA, 1001–1014. https://doi.org/10.1145/3025453.3025663
* Schüsselbauer et al. (2021) Dennis Schüsselbauer, Andreas Schmid, and Raphael Wimmer. 2021\. Dothraki: Tracking Tangibles Atop Tabletops Through De-Bruijn Tori. In _Proceedings of the Fifteenth International Conference on Tangible, Embedded, and Embodied Interaction_ _(TEI ’21)_. Association for Computing Machinery, New York, NY, USA, 1–10. https://doi.org/10.1145/3430524.3440656
* Sferrazza and D’Andrea (2019) Carmelo Sferrazza and Raffaello D’Andrea. 2019. Design, Motivation and Evaluation of a Full-Resolution Optical Tactile Sensor. _Sensors_ 19, 4 (Jan. 2019), 928\. https://doi.org/10.3390/s19040928
* Shimonomura (2019) Kazuhiro Shimonomura. 2019\. Tactile Image Sensors Employing Camera: A Review. _Sensors_ 19, 18 (Jan. 2019), 3933\. https://doi.org/10.3390/s19183933
* Siek et al. (2005) Katie A. Siek, Yvonne Rogers, and Kay H. Connelly. 2005\. Fat Finger Worries: How Older and Younger Users Physically Interact with PDAs. In _Human-Computer Interaction - INTERACT 2005_ _(Lecture Notes in Computer Science)_ , Maria Francesca Costabile and Fabio Paternò (Eds.). Springer, Berlin, Heidelberg, 267–280. https://doi.org/10.1007/11555261_24
* Taylor et al. (2021) Ian Taylor, Siyuan Dong, and Alberto Rodriguez. 2021. GelSlim3.0: High-Resolution Measurement of Shape, Force and Slip in a Compact Tactile-Sensing Finger. _arXiv:2103.12269 [cs]_ (March 2021). arXiv:2103.12269 [cs]
* Virtanen et al. (2020) Pauli Virtanen, Ralf Gommers, Travis E. Oliphant, Matt Haberland, Tyler Reddy, David Cournapeau, Evgeni Burovski, Pearu Peterson, Warren Weckesser, Jonathan Bright, Stéfan J. van der Walt, Matthew Brett, Joshua Wilson, K. Jarrod Millman, Nikolay Mayorov, Andrew R. J. Nelson, Eric Jones, Robert Kern, Eric Larson, C J Carey, İlhan Polat, Yu Feng, Eric W. Moore, Jake VanderPlas, Denis Laxalde, Josef Perktold, Robert Cimrman, Ian Henriksen, E. A. Quintero, Charles R. Harris, Anne M. Archibald, Antônio H. Ribeiro, Fabian Pedregosa, Paul van Mulbregt, and SciPy 1.0 Contributors. 2020. SciPy 1.0: Fundamental Algorithms for Scientific Computing in Python. _Nature Methods_ 17 (2020), 261–272. https://doi.org/10.1038/s41592-019-0686-2
* Visschedijk et al. (2022) Aaron Visschedijk, Hyunyoung Kim, Carlos Tejada, and Daniel Ashbrook. 2022. ClipWidgets: 3D-printed Modular Tangible UI Extensions for Smartphones. In _Sixteenth International Conference on Tangible, Embedded, and Embodied Interaction_ _(TEI ’22)_. Association for Computing Machinery, New York, NY, USA, 1–11. https://doi.org/10.1145/3490149.3501314
* Wang et al. (2021) Shaoxiong Wang, Yu She, Branden Romero, and Edward Adelson. 2021\. GelSight Wedge: Measuring High-Resolution 3D Contact Geometry with a Compact Robot Finger. In _2021 IEEE International Conference on Robotics and Automation (ICRA)_. 6468–6475. https://doi.org/10.1109/ICRA48506.2021.9560783
* Ward-Cherrier et al. (2018) Benjamin Ward-Cherrier, Nicholas Pestell, Luke Cramphorn, Benjamin Winstone, Maria Elena Giannaccini, Jonathan Rossiter, and Nathan F. Lepora. 2018. The TacTip Family: Soft Optical Tactile Sensors with 3D-Printed Biomimetic Morphologies. _Soft Robotics_ 5, 2 (April 2018), 216–227. https://doi.org/10.1089/soro.2017.0052
* Watanabe et al. (2014) Chihiro Watanabe, Alvaro Cassinelli, Yoshihiro Watanabe, and Masatoshi Ishikawa. 2014. Generic Method for Crafting Deformable Interfaces to Physically Augment Smartphones. In _CHI ’14 Extended Abstracts on Human Factors in Computing Systems_ _(CHI EA ’14)_. Association for Computing Machinery, New York, NY, USA, 1309–1314. https://doi.org/10.1145/2559206.2581307
* Weigel and Steimle (2017) Martin Weigel and Jürgen Steimle. 2017. DeformWear: Deformation Input on Tiny Wearable Devices. _Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies_ 1, 2 (June 2017), 28:1–28:23. https://doi.org/10.1145/3090093
* Weiss et al. (2009) Malte Weiss, Julie Wagner, Yvonne Jansen, Roger Jennings, Ramsin Khoshabeh, James D. Hollan, and Jan Borchers. 2009. SLAP Widgets: Bridging the Gap between Virtual and Physical Controls on Tabletops. In _Proceedings of the SIGCHI Conference on Human Factors in Computing Systems_ _(CHI ’09)_. Association for Computing Machinery, New York, NY, USA, 481–490. https://doi.org/10.1145/1518701.1518779
* Wimmer (2010) Raphael Wimmer. 2010\. FlyEye: Grasp-Sensitive Surfaces Using Optical Fiber. In _Proceedings of the Fourth International Conference on Tangible, Embedded, and Embodied Interaction_ _(TEI ’10)_. Association for Computing Machinery, New York, NY, USA, 245–248. https://doi.org/10.1145/1709886.1709935
* Winstone et al. (2012) Benjamin Winstone, Gareth Griffiths, Chris Melhuish, Tony Pipe, and Jonathan Rossiter. 2012. TACTIP — Tactile Fingertip Device, Challenges in Reduction of Size to Ready for Robot Hand Integration. In _2012 IEEE International Conference on Robotics and Biomimetics (ROBIO)_. 160–166. https://doi.org/10.1109/ROBIO.2012.6490960
* Wobbrock et al. (2008) Jacob O. Wobbrock, Brad A. Myers, and Htet Htet Aung. 2008\. The Performance of Hand Postures in Front- and Back-of-Device Interaction for Mobile Computing. _International Journal of Human-Computer Studies_ 66, 12 (Dec. 2008), 857–875. https://doi.org/10.1016/j.ijhcs.2008.03.004
* Wong et al. (2016) Pui Chung Wong, Hongbo Fu, and Kening Zhu. 2016. Back-Mirror: Back-of-Device One-Handed Interaction on Smartphones. In _SIGGRAPH ASIA 2016 Mobile Graphics and Interactive Applications_ _(SA ’16)_. Association for Computing Machinery, New York, NY, USA, 1–2. https://doi.org/10.1145/2999508.2999512
* Xiao et al. (2013) Xiang Xiao, Teng Han, and Jingtao Wang. 2013. LensGesture: Augmenting Mobile Interactions with Back-of-Device Finger Gestures. In _Proceedings of the 15th ACM on International Conference on Multimodal Interaction_ _(ICMI ’13)_. Association for Computing Machinery, New York, NY, USA, 287–294. https://doi.org/10.1145/2522848.2522850
* Yamada et al. (2018) Wataru Yamada, Hiroyuki Manabe, and Daizo Ikeda. 2018\. CamTrackPoint: Camera-Based Pointing Stick Using Transmitted Light through Finger. In _Proceedings of the 31st Annual ACM Symposium on User Interface Software and Technology_ _(UIST ’18)_. Association for Computing Machinery, New York, NY, USA, 313–320. https://doi.org/10.1145/3242587.3242641
* Yamaguchi and Atkeson (2017) Akihiko Yamaguchi and Christopher G. Atkeson. 2017. Implementing Tactile Behaviors Using FingerVision. In _2017 IEEE-RAS 17th International Conference on Humanoid Robotics (Humanoids)_. 241–248. https://doi.org/10.1109/HUMANOIDS.2017.8246881
* Yang et al. (2013) Xing-Dong Yang, Khalad Hasan, Neil Bruce, and Pourang Irani. 2013\. Surround-See: Enabling Peripheral Vision on Smartphones during Active Use. In _Proceedings of the 26th Annual ACM Symposium on User Interface Software and Technology_ _(UIST ’13)_. Association for Computing Machinery, New York, NY, USA, 291–300. https://doi.org/10.1145/2501988.2502049
* Yu et al. (2019) Chun Yu, Xiaoying Wei, Shubh Vachher, Yue Qin, Chen Liang, Yueting Weng, Yizheng Gu, and Yuanchun Shi. 2019\. HandSee: Enabling Full Hand Interaction on Smartphone with Front Camera-based Stereo Vision. In _Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems_ _(CHI ’19)_. Association for Computing Machinery, New York, NY, USA, 1–13. https://doi.org/10.1145/3290605.3300935
* Yu et al. (2011) Neng-Hao Yu, Sung-Sheng Tsai, I-Chun Hsiao, Dian-Je Tsai, Meng-Han Lee, Mike Y. Chen, and Yi-Ping Hung. 2011. Clip-on Gadgets: Expanding Multi-Touch Interaction Area with Unpowered Tactile Controls. In _Proceedings of the 24th Annual ACM Symposium on User Interface Software and Technology_ _(UIST ’11)_. Association for Computing Machinery, New York, NY, USA, 367–372. https://doi.org/10.1145/2047196.2047243
|
# Operators Affiliated to Banach Lattice Properties and the Enveloping Norms
###### Abstract
Several recent papers were devoted to various modifications of limited,
Grothendieck, and Dunford–Pettis operators, etc., through involving the Banach
lattice structure. In the present paper, it is shown that many of these
operators appear as operators affiliated to well known properties of Banach
lattices, like the disjoint (dual) Schur property, the disjoint Grothendieck
property, the property (d), and the sequential w∗-continuity of the lattice
operations. It is proved that the spaces consisting of regularly versions of
the above operators are all Banach spaces. The domination problem for such
operators is investigated.
Eduard Emelyanov, Svetlana Gorokhova
Keywords: Banach lattice, affiliated operators, enveloping norm, domination
problem
MSC2020: 46B25, 46B42, 46B50, 47B60
## 1 Preliminaries
Throughout the paper, vector spaces are real; operators are linear and
bounded; letters $X$, $Y$ stands for Banach spaces; and $E$, $F$ for Banach
lattices. We denote by $B_{X}$ the closed unit ball of $X$; by $\text{\rm
L}(X,Y)$ the space of all bounded operators from $X$ to $Y$; and by $E_{+}$
the positive cone of $E$. An operator $T:E\to F$ is called regular if
$T=T_{1}-T_{2}$ for some $T_{1},T_{2}\in\text{\rm L}_{+}(E,F)$. We denote by
$\text{\rm L}_{r}(E,F)$ ($\text{\rm L}_{ob}(E,F)$, $\text{\rm L}_{oc}(E,F)$)
the space of all regular (o-bounded, o-continuous) operators from $E$ to $F$.
### 1.1
Recall that a bounded $A\subseteq X$ is said to be a limited set (resp. a DP-
set) if each w∗-null (resp. w-null) sequence in $X^{\prime}$ is uniformly null
on $A$. Similarly, a bounded $A\subseteq E$ is called an a-limited set (resp.
an a-DP-set) if each disjoint w∗-null (resp. disjoint w-null) sequence in
$E^{\prime}$ is uniformly null on $A$ (cf. [6, 7, 13, 17]). Each relatively
compact set is limited, each limited set is an a-limited DP-set, and each DP-
set is an a-DP-set.
###### Assertion 1.1.1.
(cf. [11]) Let $A\subseteq X$ be limited. Then$:$
1. (i)
Every sequence in $A$ has a w-Cauchy subsequence.
2. (ii)
If $X$ is either separable or else reflexive, then $A$ is relatively compact.
3. (iii)
If $\ell^{1}$ does not embed in $X$, then $A$ is relatively w-compact.
The following technical fact (cf. [4, Prop.1.2.1]) is useful.
###### Assertion 1.1.2.
Let $A\subseteq X$ and $B\subseteq X^{\prime}$ be nonempty. Then$:$
1. (i)
A sequence $(f_{n})$ in $X^{\prime}$ is uniformly null on $A$ iff
$f_{n}(a_{n})\to 0$ for each sequence $(a_{n})$ in $A$.
2. (ii)
A sequence $(x_{n})$ in $X$ is uniformly null on $B$ iff $b_{n}(x_{n})\to 0$
for each sequence $(b_{n})$ in $B$.
A bounded $B\subseteq X^{\prime}$ (resp. $B\subseteq E^{\prime}$) is called an
L-set (resp. an a-L-set) if each w-null sequence in $X$ (resp. each disjoint
w-null sequence in $E$) is uniformly null on $B$ (cf. [24]). The next fact
follows from Assertion 1.1.2.
###### Assertion 1.1.3.
A bounded subset $A$ of $X$ is
1. (i)
limited iff $f_{n}(a_{n})\to 0$ for all w∗-null $(f_{n})$ in $X^{\prime}$ and
all $(a_{n})$ in $A$;
2. (ii)
a DP-set iff $f_{n}(a_{n})\to 0$ for all w-null $(f_{n})$ in $X^{\prime}$ and
all $(a_{n})$ in $A$.
A bounded subset $B$ of $X^{\prime}$ is
1. (iii)
an L-set iff $b_{n}(x_{n})\to 0$ for all $(b_{n})$ in $B$ and all w-null
$(x_{n})$ in $X$.
A bounded subset $A$ of $E$ is
1. (iv)
a-limited iff $f_{n}(a_{n})\to 0$ for all disjoint w∗-null $(f_{n})$ in
$E^{\prime}$ and all $(a_{n})$ in $A$;
2. (v)
an a-DP-set iff $f_{n}(a_{n})\to 0$ for all disjoint w-null $(f_{n})$ in
$E^{\prime}$ and all $(a_{n})$ in $A$.
A bounded subset $B$ of $E^{\prime}$ is
1. (vi)
an a-L-set iff $b_{n}(x_{n})\to 0$ for all $(b_{n})$ in $B$ and all disjoint
w-null $(x_{n})$ in $E$.
### 1.2
Let us recall the following properties of Banach spaces and describe operators
affiliated to these properties.
###### Definition 1.2.1.
A Banach space $X$ is said to possess:
1. a)
the Schur property (briefly, $X\in\text{\rm(SP)}$) if each w-null sequence in
$X$ is norm null;
2. b)
the Grothendieck property (briefly, $X\in\text{\rm(GP)}$) if each w∗-null
sequence in $X^{\prime}$ is w-null;
3. c)
the Dunford–Pettis property (briefly, $X\in\text{\rm(DPP)}$) if
$f_{n}(x_{n})\to 0$ for each w-null $(f_{n})$ in $X^{\prime}$ and each w-null
$(x_{n})$ in $X$;
4. d)
the Gelfand–Phillips property (briefly, $X\in\text{\rm(GPP)}$) if each limited
subset of $X$ is relatively compact (cf. [24, p.424]).
5. e)
the Bourgain–Diestel property (briefly, $X\in(\text{\rm BDP})$) if each
limited subset of $X$ is relatively w-compact [22].
Dedekind complete AM-spaces with a strong order unit belong to (GP), for a
comprehensive rescent source on the Grothendieck property see [25]. All
separable and all reflexive Banach spaces belong to (GPP) [11]. A Dedekind
$\sigma$-complete Banach lattice $E$ belongs $(\text{\rm GPP})$ iff $E$ has
o-continuous norm [12]. In particular, $c_{0},\ell^{1}\in(\text{\rm GPP})$,
yet $\ell^{\infty}\not\in(\text{\rm GPP})$. Clearly,
$\text{\rm(GPP)}\Rightarrow\text{\rm(BDP)}$. By [11], $X\in\text{\rm(BDP)}$
whenever $X$ contains no copy of $\ell^{1}$. Applying redistribution (as in
[2]) between the domain and range to the properties of Definition 1.2.1, we
obtain the following list of the affiliated operators.
###### Definition 1.2.2.
An operator $T:X\to Y$ is called:
1. a)
an [SP]-operator if $(Tx_{n})$ is norm null for each w-null $(x_{n})$ in $X$;
2. b)
a [GP]-operator if $(T^{\prime}f_{n})$ is w-null in $X^{\prime}$ for each
w∗-null $(f_{n})$ in $Y^{\prime}$;
3. c)
a [DPP]-operator if $f_{n}(Tx_{n})\to 0$ for each w-null $(f_{n})$ in
$Y^{\prime}$ and each w-null $(x_{n})$ in $X$;
4. d)
a [GPP]-operator if $T$ carries limited sets onto relatively compact sets;
5. e)
a [BDP]-operator if $T$ carries limited sets onto relatively w-compact sets.
Note that [SP]-operators coincide with Dunford–Pettis operators,
[GP]-operators coincide with Grothendieck operators, whereas [DPP]-operators
agree with weak Dunford–Pettis operators of [1, p.349].
###### Definition 1.2.3.
Let ${\cal P}$ be a class of operators between Banach spaces. A Banach space
$X$ is said to be affiliated with ${\cal P}$ if $I_{X}\in{\cal P}$. In this
case we write $X\in({\cal P})$.
It should be clear that if $(P)$ is one of the five properties mentioned in
Definition 1.2.1, then $X\in(P)$ iff $X$ affiliated with $[P]$-operators;
symbolically $([(P)])=(P)$. It is worth noticing that the reflexivity of
Banach spaces is affiliated with w-compact operators and vice versa, whereas
the finite dimensionality is affiliated with compact operators and vice versa.
### 1.3
We recall the following classes of operators.
###### Definition 1.3.1.
An operator
1. a)
$T:X\to F$ is called almost Grothendieck (shortly, $T$ is a-G) if $T^{\prime}$
takes disjoint $\text{\rm w}^{\ast}$-null sequences of $F^{\prime}$ to w-null
sequences of $X^{\prime}$ [23, Def.3.1].
2. b)
$T:X\to F$ is called almost limited (shortly, $T$ is Lm) if $T(B_{X})$ is
a-limited; i.e., $T^{\prime}$ takes disjoint $\text{\rm w}^{\ast}$-null
sequences of $F^{\prime}$ to norm null sequences of $X^{\prime}$ [19].
3. c)
$T:E\to Y$ is called almost Dunford–Pettis (shortly, $T$ is a-DP) if $T$ takes
disjoint w-null sequences to norm null ones [35].
4. d)
$T:E\to Y$ is called almost weak Dunford–Pettis (shortly, $T$ is a-wDP) if
$f_{n}(Tx_{n})\to 0$ whenever $(f_{n})$ is w-null in $Y^{\prime}$ and
$(x_{n})$ is disjoint w-null in $E$ [4, Def.5.3.1b)].
5. e)
$T:E\to Y$ is called o-limited (shortly, $T$ is o-Lm) if $T[0,x]$ is limited
for all $x\in E_{+}$; i.e., $(T^{\prime}f_{n})$ is uniformly null on all order
intervals $[0,x]\subseteq E_{+}$ for each $\text{\rm w}^{\ast}$-null $(f_{n})$
of $Y^{\prime}$ [27].
6. f)
$T:E\to F$ is called almost o-limited (shortly, $T$ is a-o-Lm) if $T[0,x]$ is
a-limited for all $x\in E_{+}$; i.e., $(T^{\prime}f_{n})$ is uniformly null on
all order intervals $[0,x]\subseteq E_{+}$ for each disjoint $\text{\rm
w}^{\ast}$-null $(f_{n})$ of $F^{\prime}$ [28, Def.3.1].
Clearly: $\text{\rm a-Lm}(X,F)\subseteq\text{\rm a-G}(X,F)$; $\text{\rm
a-DP}(E,Y)\subseteq\text{\rm a-wDP}(E,Y)$; $\text{\rm
Lm}(E,Y)\subseteq\text{\rm o-Lm}(E,Y)$; and $\text{\rm
o-Lm}(E,F)\subseteq\text{\rm a-o-Lm}(E,F)$.
Let ${\cal P}\subseteq\text{\rm L}(E,F)$. We call elements of ${\cal P}$ by
${\cal P}$-operators and denote by ${\cal P}(E,F):={\cal P}$ the set of all
${\cal P}$-operators in $\text{\rm L}(E,F)$. The ${\cal P}$-operators satisfy
the domination property if $S\in{\cal P}$ whenever $0\leq S\leq T\in{\cal P}$.
An operator $T\in\text{\rm L}(E,F)$ is said to be ${\cal P}$-dominated if $\pm
T\leq U$ for some $U\in{\cal P}$.
### 1.4 Enveloping norms on spaces of regularly ${\cal P}$-operators.
Regularly ${\cal P}$-operators were introduced in [3, 21] and the enveloping
norms in [4, 21]. Here we recall basic results. By [34, Prop.1.3.6],
$\text{\rm L}_{r}(E,F)$ is a Banach space under the regular norm
$\|T\|_{r}:=\inf\\{\|S\|:\pm T\leq S\in\text{\rm L}(E,F)\\}$. Moreover,
$\|T\|_{r}=\inf\\{\|S\|:S\in\text{\rm L}(E,F),|Tx|\leq S|x|\ \forall x\in
E\\}\geq\|T\|$ for every $T\in\text{\rm L}_{r}(E,F)$. If $F$ is Dedekind
complete, then $(\text{\rm L}_{r}(E,F),\|\cdot\|_{r})$ is a Banach lattice and
$\|T\|_{r}=\|~{}|T|~{}\|$ for every $T\in\text{\rm L}_{r}(E,F)$. The following
definition was introduced in [21, Def.2] (cf. also [3, Def.1.5.1]).
###### Definition 1.4.1.
Let ${\cal P}\subseteq\text{\rm L}(E,F)$. An operator $T:E\to F$ is called a
regularly ${\cal P}$-operator (shortly, an r-${\cal P}$-operator), if
$T=T_{1}-T_{2}$ with $T_{1},T_{2}\in{\cal P}\cap\text{\rm L}_{+}(E,F)$. We
denote by: ${\cal P}_{r}(E,F)$ the set of all regular operators in ${\cal
P}(E,F)$; and by $\text{\rm r-}{\cal P}(E,F)$ the set of all regularly ${\cal
P}$-operators in $\text{\rm L}(E,F)$.
###### Assertion 1.4.2.
([3, Prop.1.5.2]) Let ${\cal P}\subseteq\text{\rm L}(E,F)$, ${\cal P}\pm{\cal
P}\subseteq{\cal P}\neq\emptyset$, and $T\in\text{\rm L}(E,F)$. Then the
following holds.
1. (i)
$T$ is an r-${\cal P}$-operator iff $T$ is a ${\cal P}$-dominated ${\cal
P}$-operator.
2. (ii)
Suppose ${\cal P}$-operators satisfy the domination property and the modulus
$|T|$ exists in $\text{\rm L}(E,F)$. Then $T$ is an r-${\cal P}$-operator iff
$|T|\in{\cal P}$.
The replacement of $\text{\rm L}(E,F)$ in the definition of the regular norm
by an arbitrary subspace ${\cal P}\subseteq\text{\rm L}(E,F)$:
$\|T\|_{\text{\rm r-}{\cal P}}:=\inf\\{\|S\|:\pm T\leq S\in{\cal P}\\}\ \ \ \
(T\in\text{\rm r-}{\cal P}(E,F))$ (1)
gives the so-called enveloping norm on $\text{\rm r-}{\cal P}(E,F)$ [4].
Furthermore
$\|T\|_{\text{\rm r-}{\cal P}}=\inf\\{\|S\|:S\in{\cal P}\ \&\ (\forall x\in
E)\ |Tx|\leq S|x|\\}\ \ \ (T\in\text{\rm r-}{\cal P}(E,F))$ (2)
by [4, Lm.2.2.1], and if ${\cal P}_{1}$ is a subspace of ${\cal P}$ then
$\|T\|_{\text{\rm r-}{\cal P}_{1}}\geq\|T\|_{\text{\rm r-}{\cal
P}}\geq\|T\|_{r}\geq\|T\|\ \ \ \ \ (\forall\ T\in\text{\rm r-}{\cal
P}_{1}(E,F)).$ (3)
###### Assertion 1.4.3.
([4, Thm.2.3.1]) Let ${\cal P}$ be a subspace of $\text{\rm L}(E,F)$ closed in
the operator norm. Then $\text{\rm r-}{\cal P}(E,F)$ is a Banach space under
the enveloping norm.
Let ${\cal P}\subseteq\text{\rm L}(E,F)$, and denote ${\cal
P^{\prime}}:=\\{T^{\prime}:T\in{\cal P}\\}\subseteq\text{\rm
L}(F^{\prime},E^{\prime})$. Clearly, $\text{\rm r-}{\cal
P^{\prime}}(F^{\prime},E^{\prime})=(\text{\rm r-}{\cal P}(E,F))^{\prime}$.
Since $\|S^{\prime}\|=\|S\|$, it follows from (1)
$\|T^{\prime}\|_{\text{\rm r-}{\cal P^{\prime}}}=\inf\\{\|S^{\prime}\|:\pm
T^{\prime}\leq S^{\prime}\in{\cal P^{\prime}}\\}=\inf\\{\|S\|:\pm T\leq
S\in{\cal P}\\}=\|T\|_{\text{\rm r-}{\cal P}}.$
If ${\cal P}\subseteq\text{\rm L}(E,F)$ is closed in the operator norm then
${\cal P^{\prime}}\subseteq\text{\rm L}(F^{\prime},E^{\prime})$ is also closed
in the operator norm. So, the next fact follows from Assertion 1.4.3.
###### Corollary 1.4.4.
Let ${\cal P}$ be a subspace of $\text{\rm L}(E,F)$ closed in the operator
norm. Then $\text{\rm r-}{\cal P^{\prime}}(F^{\prime},E^{\prime})$ is a Banach
space under the enveloping norm.
### 1.5
In Section 2, we introduce the main definitions and discuss basic properties
of affiliated operators, especially related to enveloping norms. Section 3 is
devoted to domination results for affiliated operators, under the
consideration, with special emphasize on the property (d) and on sequential
w-continuity of lattice operations in Banach lattices. For further unexplained
terminology and notations, we refer to [1, 2, 3, 4, 7, 14, 15, 34, 36, 37,
38].
## 2 Affiliated operators and enveloping norms
Several recent papers were devoted to various modifications of limited,
Grothendieck, L- and M-weakly compact, and Dunford–Pettis operators, through
involving the structure of Banach lattices (see, e.g. [3, 4, 6, 7, 10, 13, 17,
18, 19, 20, 23, 31, 26, 28, 29, 30, 33, 37], Definition 1.3.1). In this
section we show that many of these operators appear as operators affiliated to
well known properties of Banach lattices like the disjoint (dual) Schur
property, the disjoint Grothendieck property, the property (d), and the
sequential w∗-continuity of the lattice operations. In continuation of [4] we
shortly discuss the enveloping norms correspondent to these affiliated
operators.
### 2.1
Recall that $E$ (resp. $E^{\prime}$) has sequentially w-continuous (resp.
sequentially w∗-continuous) lattice operations if $(|x_{n}|)$ is w-null (resp.
w∗-null) for each w-null $(x_{n})$ in $E$ (resp. for each w∗-null $(x_{n})$ in
$E^{\prime}$).
###### Assertion 2.1.1.
(see [27, Prop.3.1]) The following are equivalent.
1. (i)
$E^{\prime}$ has sequentially $\text{\rm w}^{\ast}$-continuous lattice
operations.
2. (ii)
Each order interval in $E$ is limited.
In particular, the dual $E^{\prime}$ of each discrete Banach lattice $E$ with
order continuous norm has sequentially w∗-continuous lattice operations [37,
Prop.1.1], [27, Cor.3.2]. Under the disjointness assumption on a sequence in
$E$ we have the following fact.
###### Assertion 2.1.2.
(cf. [1, Thm.4.34]) For every disjoint w-null $(x_{n})$ in $E$, the sequence
$(|x_{n}|)$ is also w-null.
This is no longer true for w∗-convergence (e.g. the sequence
$f_{n}:=e_{2n}-e_{2n+1}$ is disjoint $\text{\rm w}^{\ast}$-null in
$c^{\prime}$ yet $|f_{n}|({\mathbb{1}}_{\mathbb{N}})\equiv 2\not\to 0$ [13,
Ex.2.1]). We recall the following properties of Banach lattices.
###### Definition 2.1.3.
A Banach lattice $E$ has:
1. a)
the positive Schur property (briefly, $E\in\text{\rm(PSP)}$) if each w-null
sequence in $E_{+}$ is norm null (cf. [36]);
2. b)
the positive disjoint Schur property (briefly, $E\in\text{\rm(PDSP)}$) if each
disjoint w-null sequence in $E_{+}$ is norm null;
3. c)
the disjoint Schur property (briefly, $E\in\text{\rm(DSP)}$) if each disjoint
w-null sequence in $E$ is norm null;
4. d)
the dual positive Schur property (briefly, $E\in\text{\rm(DPSP)}$) if each
w∗-null sequence in $E^{\prime}_{+}$ is norm null [7, Def.3.3];
5. e)
the dual disjoint Schur property (briefly, $E\in\text{\rm(DDSP)}$) if each
disjoint w∗-null sequence in $E^{\prime}$ is norm null [32, Def.3.2];
6. f)
the positive Grothendieck property (briefly, $E\in\text{\rm(PGP)}$) if each
w∗-null sequence in $E^{\prime}_{+}$ is w-null (cf. [37, p.760]);
7. g)
the disjoint Grothendieck property (briefly, $E\in\text{\rm(DGP)}$) if each
disjoint w∗-null sequence in $E^{\prime}$ is w-null (cf. [3, Def.2.1.3]);
8. h)
the (swl)-property (briefly, $E\in\text{\rm(swl)}$) if $(|x_{n}|)$ is w-null
for each w-null sequence $(x_{n})$ in $E$;
9. i)
the (sw∗l)-property (briefly, $E\in\text{\rm(\text{\rm sw}${}^{\ast}$l)}$) if
$(|f_{n}|)$ is w∗-null for each w∗-null sequence $(f_{n})$ in $E^{\prime}$;
10. j)
the property (d) (briefly, $E\in\text{\rm(d)}$) if $(|f_{n}|)$ is $\text{\rm
w}^{\ast}$-null for each disjoint $\text{\rm w}^{\ast}$-null sequence
$(f_{n})$ in $E^{\prime}$ [17, 37];
11. k)
the bi-sequence property (briefly, $E\in\text{\rm(bi-sP)}$) if
$f_{n}(x_{n})\to 0$ for each w∗-null $(f_{n})$ in $E^{\prime}_{+}$ and each
disjoint w-null $(x_{n})$ in $E$ [7, Def.3.1];
12. l)
the strong GP-property (briefly, $E\in\text{\rm(s-GPP)}$) if each almost
limited subset of $E$ is relatively compact;
13. m)
the strong BD-property (briefly, $E\in\text{\rm(s-BDP)}$) if each almost
limited subset of $E$ is relatively w-compact.
It is well known that $\text{\rm(PSP)}=\text{\rm(PDSP)}=\text{\rm(DSP)}$.
Indeed, $\text{\rm(PSP)}\subseteq\text{\rm(PDSP)}$ holds trivially;
$\text{\rm(PDSP)}\subseteq\text{\rm(DSP)}$ is due to Assertion 2.1.2; and, for
$\text{\rm(DSP)}\subseteq\text{\rm(PSP)}$ see [36, p.16]. We include a short
proof of the following fact.
###### Assertion 2.1.4.
([7, Thm.4.2], [37, Prop.2.4]) Let $E$ be a Banach lattice. The following are
equivalent$:$
1. (i)
$E\in\text{\rm(bi-sP)}$;
2. (ii)
$E\in\text{\rm(Pbi-sP)}$, in the sense that if $f_{n}(x_{n})\to 0$ for each
w∗-null $(f_{n})$ in $E^{\prime}_{+}$ and each disjoint w-null $(x_{n})$ in
$E_{+}$.
3. (iii)
every $\text{\rm w}^{\ast}$-null sequence $(f_{n})$ in $E^{\prime}_{+}$ is
uniformly null on each disjoint w-null $(x_{n})$ in $E_{+}$.
###### Proof.
The implication i)$\Longrightarrow$ii) is obvious, whereas
ii)$\Longrightarrow$iii) follows from Proposition 1.1.2 i).
iii)$\Longrightarrow$i) Let $(f_{n})$ be w∗-null in $E^{\prime}_{+}$ and
$(x_{n})$ be disjoint w-null in $E$. By Assertion 2.1.2, $(x_{n}^{\pm})$ are
both disjoint w-null in $E_{+}$. Then $(f_{n})$ is uniformly null on both
$(x_{n}^{\pm})$, and hence on $(x_{n})=(x_{n}^{+})-(x_{n}^{-})$. By
Proposition 1.1.2 i), $f_{n}(x_{n})\to 0$, as desired. ∎
### 2.2
Applying the redistribution between the domain and range as in Definition
1.2.2 to properties of Definition 2.1.3, we obtain the correspondent
affiliated operators.
###### Definition 2.2.1.
An operator $T:E\to Y$ is called:
1. a)
a [PSP]-operator if $\|Tx_{n}\|\to 0$ for each w-null $(x_{n})$ in $E_{+}$;
2. b)
a [PDSP]-operator if $\|Tx_{n}\|\to 0$ for each disjoint w-null $(x_{n})$ in
$E_{+}$;
3. c)
a [DSP]-operator if $\|Tx_{n}\|\to 0$ for each disjoint w-null $(x_{n})$ in
$E$;
4. d)
an [s-GPP]-operator if $T$ carries almost limited subsets of $E$ onto
relatively compact subsets of $Y$;
5. e)
an [s-BDP]-operator if $T$ carries almost limited subsets of $E$ onto
relatively w-compact subsets of $Y$.
Clearly,
$\text{\rm[s-GPP]}(E,Y)\subseteq\text{\rm[GPP]}(E,Y)\bigcap\text{\rm[s-BDP]}(E,Y)\
\ \text{\rm and}$ (4) $\text{\rm[s-BDP]}(E,Y)\subseteq\text{\rm[BDP]}(E,Y).$
(5)
[DSP]-operators coincide with the almost Dunford–Pettis operators, and hence,
by [6, Thm.2.2],
$\text{\rm[PSP]}(E,Y)=\text{\rm[PDSP]}(E,Y)=\text{\rm[DSP]}(E,Y).$ (6)
###### Definition 2.2.2.
An operator $T:X\to F$ is called:
1. a)
a [DPSP]-operator if $\|T^{\prime}f_{n}\|\to 0$ for each w∗-null $(f_{n})$ in
$F_{+}^{\prime}$;
2. b)
a [DDSP]-operator if $\|T^{\prime}f_{n}\|\to 0$ for each disjoint w∗-null
$(f_{n})$ in $F^{\prime}$;
3. c)
a [PGP]-operator if $(T^{\prime}f_{n})$ is w-null for each w∗-null $(f_{n})$
in $F_{+}^{\prime}$;
4. d)
a [DGP]-operator if $(T^{\prime}f_{n})$ is w-null for each disjoint w∗-null
$(f_{n})$ in $F^{\prime}$;
5. e)
an [swl]-operator if $(|Tx_{n}|)$ is w-null for each w-null $(x_{n})$ in $X$.
[DDSP]-operators coincide with the almost limited operators, whereas
[DGP]-operators agree with the almost Grothendieck operators.
###### Proposition 2.2.3.
$(\text{\rm[DPSP]}(X,F))^{\prime}\cup(\text{\rm[DDSP]}(X,F))^{\prime}\subseteq\text{\rm[PSP]}(F^{\prime},X^{\prime})$.
###### Proof.
Let $(f_{n})$ be disjoint w-null in $F^{\prime}_{+}$. Then $(f_{n})$ is
disjoint w∗-null in $F^{\prime}_{+}$. If $T\in\text{\rm[DPSP]}(X,F)$ or
$T\in\text{\rm[DDSP]}(X,F)$ then in both cases $\|T^{\prime}f_{n}\|\to 0$.
Thus $T^{\prime}\in\text{\rm[PDSP]}(F^{\prime},X^{\prime})$, and hence
$T^{\prime}\in\text{\rm[PSP]}(F^{\prime},X^{\prime})$ by (6). ∎
###### Definition 2.2.4.
An operator $T:E\to F$ is called:
1. a)
a [dswl]-operator if $(|Tx_{n}|)$ is w-null for each disjoint w-null
$(x_{n})$;
2. b)
an $\text{\rm[sw}^{\ast}\text{\rm l]}$-operator if $(|T^{\prime}f_{n}|)$ is
w∗-null for each w∗-null $(f_{n})$ in $F^{\prime}$;
3. c)
a [d]-operator if $(|T^{\prime}f_{n}|)$ is w∗-null for each disjoint w∗-null
$(f_{n})$ in $F^{\prime}$;
4. d)
a [bi-sP]-operator if $f_{n}(Tx_{n})\to 0$ for each w∗-null $(f_{n})$ in
$F^{\prime}_{+}$ and each disjoint w-null $(x_{n})$ in $E$;
5. e)
a [Pbi-sP]-operator if $f_{n}(Tx_{n})\to 0$ for each w∗-null $(f_{n})$ in
$F^{\prime}_{+}$ and each disjoint w-null $(x_{n})$ in $E_{+}$.
###### Proposition 2.2.5.
For a Banach lattice $F$ the following are hold.
1. i)
$F\in\text{\rm(d)}$ iff $\text{\rm r-[d]}(E,F)=\text{\rm L}_{r}(E,F)$ for
every $E$.
2. ii)
$F^{\prime}$ has sequentially $\text{\rm w}^{\ast}$-continuous lattice
operations iff
$\text{\rm[sw}^{\ast}\text{\rm l]}(E,F)=\text{\rm L}_{r}(E,F)$ for every $E$.
###### Proof.
i) For the necessity, let $E$ be a Banach lattice. It is enough to prove
$\text{\rm L}_{+}(E,F)\subseteq\text{\rm[d]}(E,F)$. So, let $0\leq T:E\to F$
and $(f_{n})$ be disjoint $\text{\rm w}^{\ast}$-null in $F^{\prime}$. Since
$F\in\text{\rm(d)}$ then $(|f_{n}|)$ is $\text{\rm w}^{\ast}$-null, and then
$(T^{\prime}|f_{n}|)$ is $\text{\rm w}^{\ast}$-null in $E^{\prime}$. It
follows from $|T^{\prime}f_{n}|\leq T^{\prime}|f_{n}|$ that
$(|T^{\prime}f_{n}|)$ is $\text{\rm w}^{\ast}$-null, and hence
$T\in\text{\rm[d]}(E,F)$. The sufficiency is immediate since
$I_{F}\in\text{\rm[d]}(F,F)$ implies $F\in\text{\rm(d)}$.
ii) Just remove the disjointness condition on $(f_{n})$ in the proof of i). ∎
The next proposition shows that [Pbi-sP]-operators agree with [bi-
sP]-opeators.
###### Proposition 2.2.6.
$\text{\rm[bi-sP]}(E,F)=\text{\rm[Pbi-sP]}(E,F)$.
###### Proof.
Clearly, $\text{\rm[bi-sP]}(E,F)\subseteq\text{\rm[Pbi-sP]}(E,F)$. Let
$T\in\text{\rm[Pbi-sP]}(E,F)$, $(f_{n})$ be w∗-null in $F^{\prime}_{+}$, and
$(x_{n})$ be disjoint w-null in $E$. By Assertion 2.1.2, $(|x_{n}|)$ is
disjoint w-null in $E$. Since $T\in\text{\rm[Pbi-sP]}(E,F)$,
$f_{n}(T|x_{n}|)\to 0$. It follows from $|f_{n}(Tx_{n})|\leq f_{n}(T|x_{n}|)$
that $f_{n}(Tx_{n})\to 0$, and hence $T\in\text{\rm[bi-sP]}(E,F)$. ∎
###### Proposition 2.2.7.
Let $T\in\text{\rm L}(E,F)$. The following holds.
1. i)
$T$ is a [d]-operator iff $T$ is almost o-limited.
2. ii)
$T$ is an $\text{\rm[sw}^{\ast}\text{\rm l]}$-operator iff $T$ is o-limited.
###### Proof.
i) For the necessity, let $T\in\text{\rm[d]}(E,F)$. Suppose $x\in E_{+}$ and
$(f_{n})$ is disjoint $\text{\rm w}^{\ast}$-null in $F^{\prime}$. By the
assumption, $(|T^{\prime}f_{n}|)$ is w∗-null, and hence $|T^{\prime}f_{n}|x\to
0$. By the Riesz–Kantorovich formula,
$|T^{\prime}f_{n}|x=\sup\\{|(T^{\prime}f_{n})y|:|y|\leq x\\}\to 0$, and hence
$(T^{\prime}f_{n})$ is uniformly null on each $[0,x]$. Thus $T\in\text{\rm
a-o-Lm}(E,F)$.
For the sufficiency, let $T\in\text{\rm a-o-Lm}(E,F)$. Suppose $(f_{n})$ is
disjoint $\text{\rm w}^{\ast}$-null in $F^{\prime}$. In order to prove
$T\in\text{[d]}(E,F)$, we need to show that
$(|T^{\prime}f_{n}|)\stackrel{{\scriptstyle\text{\rm w}^{\ast}}}{{\to}}0$. It
is enough to prove that $|T^{\prime}f_{n}|x\to 0$ for each $x\in E_{+}$. Let
$x\in E_{+}$. By the assumption, $\sup\\{|(T^{\prime}f_{n})y|:|y|\leq x\\}\to
0$. Therefore, the Riesz–Kantorovich formula implies $|T^{\prime}f_{n}|x\to
0$, and hence $T\in\text{[d]}(E,F)$.
ii) Just remove the disjointness condition on $(f_{n})$ in the proof of i). ∎
### 2.3
Affiliated operators from the previous subsection form vector spaces, which
are complete under the operator norm; the details are included in the next
lemma.
###### Lemma 2.3.1.
The following sets of affiliated operators are vector spaces which are
complete in the operator norm.
1. i)
$\text{\rm[PSP]}(E,Y)$.
2. ii)
$\text{\rm[DPSP]}(X,F)$ and $\text{\rm[DDSP]}(X,F)$.
3. iii)
$\text{\rm[PGP]}(X,F)$ and $\text{\rm[DGP]}(X,F)$.
4. iv)
$\text{\rm[swl]}(X,F)$ and $\text{\rm[dswl]}(E,F)$.
5. v)
$\text{\rm[sw}^{\ast}\text{\rm l]}(E,F)$ and $\text{\rm[d]}(E,F)$.
6. vi)
$\text{\rm[bi-sP]}(E,F)$.
7. vii)
$\text{\rm[GPP]}(X,Y)$ and $\text{\rm[s-GPP]}(E,Y)$.
8. viii)
$\text{\rm[BDP]}(X,Y)$ and $\text{\rm[s-BDP]}(E,Y)$.
###### Proof.
We skip trivial checking that all sets of affiliated operators in the lemma
are vector spaces. It remains to show that each space of affiliated operators
under the consideration is a closed in the operator norm subspace of the
correspondent space of all linear operators. As arguments here are
straightforward and standard, we present them in the basic cases.
1. i)
Let $\text{\rm[PSP]}(E,Y)\ni
T_{k}\stackrel{{\scriptstyle\|\cdot\|}}{{\to}}T\in\text{\rm L}(E,Y)$. Let
$(x_{n})$ be w-null in $E_{+}$. We need to show $\|Tx_{n}\|\to 0$. Let
$\varepsilon>0$. Pick some $k\in\mathbb{N}$ with $\|T-T_{k}\|\leq\varepsilon$.
Since $T_{k}\in\text{\rm PSP}(E,Y)$, there exists $n_{0}$ such that
$\|T_{k}x_{n}\|\leq\varepsilon$ for $n\geq n_{0}$. Take $M\in\mathbb{R}$
satisfying $\|x_{n}\|\leq M$ for all $n\in\mathbb{N}$. Since
$\|Tx_{n}\|=\|(T-T_{k})x_{n}+T_{k}x_{n}\|\leq\|T-T_{k}\|\cdot\|x_{n}\|+\|T_{k}x_{n}\|\leq\varepsilon(M+1)$
for $n\geq n_{0}$, and since $\varepsilon>0$ is arbitrary, $\|Tx_{n}\|\to 0$.
2. ii)
As the case of [DDSP](X,F) is similar, we confine ourselves to considering
[DPSP](X,F).
Let $\text{\rm[DPSP]}(X,F)\ni
T_{k}\stackrel{{\scriptstyle\|\cdot\|}}{{\to}}T\in\text{\rm L}(X,F)$, and let
$(f_{n})$ be $\text{\rm w}^{\ast}$-null in $F_{+}^{\prime}$. In order to show
$(T^{\prime}f_{n})$ is norm null, let $\varepsilon>0$ and pick $k$ with
$\|T^{\prime}-T^{\prime}_{k}\|\leq\varepsilon$. Since
$T_{k}\in\text{\rm[DPSP]}(X,F)$, there exists $n_{0}$ with
$\|T_{k}^{\prime}f_{n}\|\leq\varepsilon$ for all $n\geq n_{0}$. As $(f_{n})$
is $\text{\rm w}^{\ast}$-null, there exists $M\in\mathbb{R}$ satisfying
$\|f_{n}\|\leq M$ for all $n\in\mathbb{N}$. Since
$\|T^{\prime}f_{n}\|\leq\|T^{\prime}f_{n}-T^{\prime}_{k}f_{n}\|+\|T^{\prime}_{k}f_{n}\|\leq\|T^{\prime}_{k}-T^{\prime}\|\|f_{n}\|+\varepsilon\leq\varepsilon(M+1)$
for $n\geq n_{0}$. It follows $\|T^{\prime}f_{n}\|\to 0$, as desired.
3. iii)
As the case of [DGP](X,F) is similar, we consider [PGP](X,F) only.
Let $\text{\rm[PGP]}(X,F)\ni
T_{k}\stackrel{{\scriptstyle\|\cdot\|}}{{\to}}T\in\text{\rm L}(X,F)$, and let
$(f_{n})$ be $\text{\rm w}^{\ast}$-null in $F_{+}^{\prime}$. In order to show
that $(T^{\prime}f_{n})$ is w-null, pick a $g\in F^{\prime\prime}$, and let
$\varepsilon>0$. Fix any $k$ with
$\|T^{\prime}-T^{\prime}_{k}\|\leq\varepsilon$. Since
$T_{k}\in\text{\rm[PGP]}(X,F)$, there exists $n_{0}$ with
$|g(T_{k}^{\prime}f_{n})|\leq\varepsilon$ for all $n\geq n_{0}$. Let
$M\in\mathbb{R}$ be such $\|f_{n}\|\leq M$ for all $n\in\mathbb{N}$. Because
of
$|g(T^{\prime}f_{n})|\leq|g(T^{\prime}f_{n}-T_{k}^{\prime}f_{n})|+|g(T_{k}^{\prime}f_{n})|\leq$
$\|g\|\|T^{\prime}-T_{k}^{\prime}\|\|f_{n}\|+\varepsilon\leq(\|g\|M+1)\varepsilon$
for $n\geq n_{0}$, and since $\varepsilon>0$ is arbitrary, it follows
$g(T^{\prime}f_{n})\to 0$. Since $g\in F^{\prime\prime}$ is arbitrary,
$T\in\text{\rm[PGP]}(X,F)$.
4. iv)
We consider [swl](X,F) only. The case of [dswl](E,F) is similar.
Let $\text{\rm[swl]}(X,F)\ni
T_{k}\stackrel{{\scriptstyle\|\cdot\|}}{{\to}}T\in\text{\rm L}(X,F)$ and let
$(x_{n})$ be w-null in $X$. We need to show
$|Tx_{n}|\stackrel{{\scriptstyle\text{\rm w}}}{{\to}}0$ in $F$. Let $f\in
F^{\prime}$. There exists an $M\in{\mathbb{R}}$ with $\|x_{n}\|\leq M$ for all
$n\in{\mathbb{N}}$. Take some $\varepsilon>0$ and pick $k\in{\mathbb{N}}$ with
$\|T-T_{k}\|\leq\varepsilon$. Choose $n_{0}$ such that
$|f|(|T_{k}x_{n}|)\leq\varepsilon$ for all $n\geq n_{0}$. Then
$|f(|Tx_{n}|)|\leq|f|(|(T-T_{k})x_{n}+T_{k}x_{n}|)\leq$
$\|f\|\cdot\|T-T_{k}\|\cdot M+|f|(|T_{k}x_{n}|)\leq\varepsilon(\|f\|M+1).$
Since $\varepsilon>0$ is arbitrary, $f(|Tx_{n}|)\to 0$; and, since $f\in
F^{\prime}$ is arbitrary, $|Tx_{n}|\stackrel{{\scriptstyle\text{\rm
w}}}{{\to}}0$.
5. v)
We consider $\text{\rm[d]}(E,F)$ only. The case of
$\text{\rm[sw}^{\ast}\text{\rm l]}(E,F)$ is similar.
Let $\text{\rm[d]}(E,F)\ni
T_{k}\stackrel{{\scriptstyle\|\cdot\|}}{{\to}}T\in\text{\rm L}(E,F)$, and let
$(f_{n})$ be disjoint w∗-null in $F^{\prime}$. We need to show
$(|T^{\prime}f_{n}|)\stackrel{{\scriptstyle\text{\rm w}^{\ast}}}{{\to}}0$. It
is enough to prove that $|T^{\prime}f_{n}|x\to 0$ for each $x\in E_{+}$. Let
$x\in E_{+}$ and $\varepsilon>0$. Pick $k\in\mathbb{N}$ with
$\|T^{\prime}-T_{k}^{\prime}\|\leq\varepsilon$. By the assumption,
$|T_{k}^{\prime}f_{n}|x\to 0$. So, let $n_{0}\in\mathbb{N}$ be such that
$|T_{k}^{\prime}f_{n}|x\leq\varepsilon$ whenever $n\geq n_{0}$. As $(f_{n})$
is $\text{\rm w}^{\ast}$-null, there exists $M\in\mathbb{R}$ with
$\|f_{n}\|\leq M$ for all $n\in\mathbb{N}$. By the Riesz–Kantorovich formula,
for $n\geq n_{0}$,
$|T^{\prime}f_{n}|x=\sup\\{|(T^{\prime}f_{n})y|:|y|\leq x\\}\leq$
$\sup\\{|((T^{\prime}-T_{k}^{\prime})f_{n})y|:|y|\leq
x\\}+\sup\\{|(T_{k}^{\prime}f_{n})y|:|y|\leq x\\}\leq$
$\sup\\{\|T^{\prime}-T_{k}^{\prime}\|\cdot\|f_{n}\|\cdot\|y\|:|y|\leq
x\\}+|T_{k}^{\prime}f_{n}|x\leq\varepsilon(M\|x\|+1).$
Therefore $|T^{\prime}f_{n}|x\to 0$, and hence $T\in\text{[d]}(E,F)$.
6. vi)
Let $\text{\rm[bi-sP]}(E,F)\ni
T_{k}\stackrel{{\scriptstyle\|\cdot\|}}{{\to}}T\in\text{\rm L}(E,F)$. Let
$(f_{n})$ be w∗-null in $F^{\prime}_{+}$ and let $(x_{n})$ be disjoint w-null
in $E$. We need to show $f_{n}(Tx_{n})\to 0$. Pick $M\in{\mathbb{R}}$ such
that $\|f_{n}\|\leq M$ and $\|x_{n}\|\leq M$ for all $n\in{\mathbb{N}}$. Take
some $\varepsilon>0$. Pick $k\in{\mathbb{N}}$ with
$\|T-T_{k}\|\leq\varepsilon$. Since $T_{k}\in\text{\rm[bi-sP]}(E,F)$, there
exists $n_{0}\in{\mathbb{N}}$ such that $|f_{n}(T_{k}x_{n})|\leq\varepsilon$
for $n\geq n_{0}$. Then
$|f_{n}(Tx_{n})|\leq|f_{n}((T-T_{k})x_{n})|+|f_{n}(T_{k}x_{n})|\leq$
$\|f_{n}\|\cdot\|T-T_{k}\|\cdot\|x_{n}\|+\varepsilon\leq(M^{2}+1)\varepsilon\
\ \ (\forall n\geq n_{0}).$
Since $\varepsilon>0$ is arbitrary, $f_{n}(Tx_{n})\to 0$.
7. vii)
As the case of $\text{\rm[GPP]}(X,Y)$ is similar, we consider
$\text{\rm[s-GPP]}(E,Y)$ only.
Let $\text{\rm[s-GPP]}(E,Y)\ni
T_{k}\stackrel{{\scriptstyle\|\cdot\|}}{{\to}}T\in\text{\rm L}(E,Y)$, and let
$A\subseteq E$ be a-limited. We need to show that $T(A)$ is relatively
compact. Since a-limited sets are bounded, there exists $M\in{\mathbb{R}}$
with $\|x\|\leq M$ for all $x\in A$. Choose $\varepsilon>0$ and pick a
$k\in{\mathbb{N}}$ such that $\|T-T_{k}\|\leq\varepsilon$. Then
$Tx=T_{k}x+(T-T_{k})x\in
T_{k}(A)+\|T-T_{k}\|\cdot\|x\|B_{Y}=T_{k}(A)+\varepsilon M\cdot B_{Y}$
for all $x\in A$, and hence $T(A)\subseteq T_{k}(A)+\varepsilon M\cdot B_{Y}$.
By the assumption, $T_{k}(A)$ is relatively compact. Since $\varepsilon>0$ is
arbitrary, $T(A)$ is totally bounded and hence is relatively compact, as
desired.
8. viii)
As the case of $\text{\rm[BDP]}(X,Y)$ is similar, we consider
$\text{\rm[s-BDP]}(E,Y)$ only.
Let $\text{\rm[s-BDP]}(E,Y)\ni
T_{k}\stackrel{{\scriptstyle\|\cdot\|}}{{\to}}T\in\text{\rm L}(E,Y)$, and let
$A\subseteq E$ be a-limited. We need to show that $T(A)$ is relatively
w-compact. Since a-limited sets are bounded, there exists $M\in{\mathbb{R}}$
such that $\|x\|\leq M$ for all $x\in A$. Take $\varepsilon>0$ and pick any
$k\in{\mathbb{N}}$ with $\|T-T_{k}\|\leq\varepsilon$. Then $T(A)\subseteq
T_{k}(A)+\varepsilon M\cdot B_{Y}$, as above in vii). By the assumption,
$T_{k}(A)$ is relatively w-compact. Since $\varepsilon>0$ is arbitrary, $T(A)$
is relatively w-compact by the Grothendieck result [1, Thm.3.44].
∎
The next result follows from Theorem 1.4.3 and Lemma 2.3.1.
###### Theorem 2.3.2.
Let $E$ ad $F$ be a Banach lattices. Then $\text{\rm r-[PSP]}(E,F)$,
$\text{\rm r-[DPSP]}(E,F)$, $\text{\rm r-[DDSP]}(E,F)$, $\text{\rm
r-[PGP]}(E,F)$, $\text{\rm r-[DGP]}(E,F)$, $\text{\rm r-[swl]}(E,F)$,
$\text{\rm r-[dswl]}(E,F)$, $\text{\rm r-[sw}^{\ast}\text{\rm l]}(E,F)$,
$\text{\rm r-[d]}(E,F)$, $\text{\rm r-[bi-sP]}(E,F)$, $\text{\rm
r-[GPP]}(E,F)$,
$\text{\rm r-[s-GPP]}(E,F)$, $\text{\rm r-[BDP]}(E,F)$, and $\text{\rm
r-[s-BDP]}(E,F)$ are all Banach spaces, each under its own enveloping norm.
## 3 Domination for affiliated operators
Here we gather domination results for defined above affiliated operators. Some
of them already appeared in the literature, the others seem new.
### 3.1
The [s-GPP]-operators do not satisfy the domination property in the strong
sense that even an operator which is dominated by a rank one operator need not
to be a [GPP]-operator.
###### Example 3.1.1.
(cf. [1, Ex.5.30]) Define operators $T,S:L^{1}[0,1]\to\ell^{\infty}$ by
$T(f):=(\int_{0}^{1}f(t)dt)_{k=1}^{\infty}$, and
$S(f):=(\int_{0}^{1}f(t)r_{k}^{+}(t)dt)_{k=1}^{\infty}$, where $r_{k}$ are the
Rademacher functions on $[0,1]$. Then $T$ is a rank one operator, and hence
$T\in\text{\rm[s-GPP]}(L^{1}[0,1],\ell^{\infty})$. Moreover, $0\leq S\leq T$,
yet $S$ is not a [GPP]-operator. To see this, consider the sequence of the
Rademacher functions $(r_{n})$ in $[0,\mathbb{1}]\subseteq L^{1}[0,1]$, which
is an a-limited subset of $L^{1}[0,1]$, e.g. by Proposition 3.2.1. The
sequence $(Sr_{n})=(\frac{1}{2}e_{n})$, where $e_{n}$ are the n-th unite
vectors in $\ell^{\infty}$, has no norm convergent subsequences, and hence
$S\not\in\text{\rm[GPP]}(L^{1}[0,1],\ell^{\infty})$.
We do not know whether or not the operator $S$ in Example 3.1.1 is a
[BDP]-operator.
### 3.2
It turn out that the property (d) and the sequential w-continuity ($\text{\rm
w}^{\ast}$-continuity) of lattice operations play an important role for the
domination property. Firstly, we include some related elementary facts.
###### Proposition 3.2.1.
The following are equivalent.
1. i)
$E\in\text{\rm(d)}$.
2. ii)
Each order interval in $E$ is a-limited.
###### Proof.
i)$\Longrightarrow$ii) It suffices to show that intervals $[-a,a]$ are
a-limited. Let $a\in E_{+}$, and let $(f_{n})$ be disjoint w∗-null in
$E^{\prime}$. We need to show that $(f_{n})$ is uniformly null on $[-a,a]$. By
Assertion 1.1.2, it is enough to show that $f_{n}(a_{n})\to 0$ for each
sequence $(a_{n})$ in $[-a,a]$. So, let $(a_{n})$ be in $[-a,a]$. Since
$E\in\text{\rm(d)}$ then $(|f_{n}|)$ is w∗-null in $E^{\prime}_{+}$, and hence
$f_{n}(a)\to 0$. It follows from $-f_{n}(a)\leq f_{n}(a_{n})\leq f_{n}(a)$ for
all $n\in\mathbb{N}$ that $f_{n}(a_{n})\to 0$. By Assertion 1.1.2, $(f_{n})$
is uniformly null on $[-a,a]$, as desired.
ii)$\Longrightarrow$i) Let $(f_{n})$ be disjoint w∗-null in $E^{\prime}$. We
need to show that $(|f_{n}|)$ is w∗-null. Pick an $a\in E_{+}$. By the
assumption, $(f_{n})$ is uniformly null on $[-a,a]$, and in view of the
Riesz–Kantorovich formula, $|f_{n}|a=\sup\limits_{y\in[-a,a]}|f_{n}(y)|\to 0$.
Since $a\in E_{+}$ is arbitrary, $(|f_{n}|)$ is w∗-null, as desired. ∎
The proof of the following result of [27] consists in removing the
disjointness condition in the proof of Proposition 3.2.1.
###### Assertion 3.2.2.
The following are equivalent.
1. (i)
$E^{\prime}$ has sequentially $\text{\rm w}^{\ast}$-continuous lattice
operations.
2. (ii)
Each order interval in $E$ is limited.
Here we gather several (partially positive) domination results.
###### Theorem 3.2.3.
Let $E$ and $F$ be Banach lattices. The following spaces of operators satisfy
the domination property.
1. i)
$\text{\rm[PSP]}(E,F)$.
2. ii)
$\text{\rm[DPSP]}(E,F)$.
3. iii)
$\text{\rm[DDSP]}(E,F)$, under the assumption $F\in\text{\rm(d)}$.
4. iv)
$\text{\rm[PGP]}(E,F)$.
5. v)
$\text{\rm[DGP]}(E,F)$, under the assumption $F\in\text{\rm(d)}$.
6. vi)
$\text{\rm[dswl]}(E,F)$.
7. vii)
$\text{\rm[swl]}(E,F)$, under the assumption that $E$ has sequentially
w-continuous lattice operations.
8. viii)
$\text{\rm[sw}^{\ast}\text{\rm l]}(E,F)$, under the assumption that
$F^{\prime}$ has sequentially $\text{\rm w}^{\ast}$-continuous lattice
operations.
9. ix)
$\text{\rm[d]}(E,F)$, under the assumption $F\in\text{\rm(d)}$.
10. x)
$\text{\rm[bi-sP]}(E,F)$.
###### Proof.
As above, we restrict ourselves to basic cases.
1. i)
Let $0\leq S\leq T\in\text{\rm[PSP]}(E,F)$ and let $(x_{n})$ be w-null in
$E_{+}$. Since $T\in\text{\rm[PSP]}(E,F)$ then $\|Tx_{n}\|\to 0$. It follows
from $0\leq Sx_{n}\leq Tx_{n}$ that $\|Sx_{n}\|\to 0$, and hence
$S\in\text{\rm[PSP]}(E,F)$.
2. ii)
Let $0\leq S\leq T\in\text{\rm[DPSP]}(E,F)$ and let $(f_{n})$ be w∗-null in
$F^{\prime}_{+}$. Since $T\in\text{\rm[DPSP]}(E,F)$ then
$\|T^{\prime}f_{n}\|\to 0$. It follows from $0\leq S^{\prime}\leq T^{\prime}$
that $0\leq S^{\prime}f_{n}\leq T^{\prime}f_{n}$, and hence
$\|S^{\prime}f_{n}\|\to 0$. Thus, $S\in\text{\rm[DPSP]}(E,F)$.
3. iii)
As [DDSP]-operators agree with almost limited operators, we refer for the
proof to [17, Cor.3].
4. iv)
Let $0\leq S\leq T\in\text{\rm[PGP]}(E,F)$, and $(f_{n})$ be w∗-null in
$F_{+}^{\prime}$. In order to prove $S\in\text{\rm[PGP]}(E,F)$, it suffices to
prove $g(S^{\prime}f_{n})\to 0$ for all $g\in E^{\prime}_{+}$. Let $g\in
E^{\prime}_{+}$. Since $T\in\text{\rm[PGP]}(E,F)$, $g(T^{\prime}f_{n})\to 0$.
It follows from $0\leq g(S^{\prime}f_{n})\leq g(T^{\prime}f_{n})$ that
$g(S^{\prime}f_{n})\to 0$, as desired.
5. v)
As [DGP]-operators agree with almost Grothendieck operators, we refer for the
proof to [23, Prop.3.7].
6. vi)
Let $0\leq S\leq T\in\text{\rm[dswl]}(E,F)$, and let $(x_{n})$ be disjoint
w-null in $E$. In order to prove $S\in\text{\rm[dswl]}(E,F)$, it suffices to
prove $f(|Sx_{n}|)\to 0$ for all $f\in F^{\prime}_{+}$. So, let $f\in
F^{\prime}_{+}$. By Assertion 2.1.2, $(|x_{n}|)$ is w-null. Since
$T\in\text{\rm[dswl]}(E,F)$ then $(T|x_{n}|)=(|T(|x_{n}|)|)$ is w-null, and
hence $f(T|x_{n}|)\to 0$. It follows from $|Sx_{n}|\leq S|x_{n}|\leq T|x_{n}|$
that $f(|Sx_{n}|)\to 0$ as desired.
7. vii)
Let $0\leq S\leq T\in\text{\rm[swl]}(E,F)$, and $(x_{n})$ be w-null in $E$. It
suffices to prove $f(|Sx_{n}|)\to 0$ for all $f\in F^{\prime}_{+}$. Let $f\in
F^{\prime}_{+}$. By the assumption, $(|x_{n}|)$ is w-null. Since
$T\in\text{\rm[swl]}(E,F)$, $f(T|x_{n}|)=f(|Tx_{n}|)\to 0$. In view of
$|Sx_{n}|\leq S|x_{n}|\leq T|x_{n}|$, $f(|Sx_{n}|)\to 0$, and hence
$S\in\text{\rm[swl]}(E,F)$.
8. viii)
It follows from Proposition 2.2.5 ii).
9. ix)
It follows from Proposition 2.2.5 i).
10. x)
Let $0\leq S\leq T\in\text{\rm[bi-sP]}(E,F)$. Let $(f_{n})$ be w∗-null in
$F^{\prime}_{+}$, and let $(x_{n})$ be disjoint w-null in $E$. In order to
prove $S\in\text{\rm[bi-sP]}(E,F)$, it suffices to prove $f_{n}(Sx_{n})\to 0$.
By Assertion 2.1.2, $(|x_{n}|)$ is disjoint w-null in $E$, and, since
$T\in\text{\rm[bi-sP]}(E,F)$, then $f_{n}(T|x_{n}|)\to 0$. It follows from
$|f_{n}(Sx_{n})|\leq f_{n}(S|x_{n}|)\leq f_{n}(T|x_{n}|)$ that
$f_{n}(Sx_{n})\to 0$, and hence $S\in\text{\rm[bi-sP]}(E,F)$.
∎
In view of [1, Thm.4.31], the next fact follows form Theorem 3.2.3 vii).
###### Corollary 3.2.4.
Let $E$ be an AM-space, and let $0\leq S\leq T\in\text{\rm[swl]}(E,F)$. Then
$S\in\text{\rm[swl]}(E,F)$.
## References
* [1] C. D. Aliprantis, O. Burkinshaw, Positive Operators. Springer, Dordrecht, 2006.
* [2] S. Alpay, E. Emelyanov, S. Gorokhova, $o\tau$-Continuous, Lebesgue, KB, and Levi Operators Between Vector Lattices and Topological Vector Spaces. Results Math 77, 117 (2022), 25pp.
* [3] S. Alpay, E. Emelyanov, S. Gorokhova, Generalizations of L- and M-weakly compact operators. https://arxiv.org/abs/2206.02718v3
* [4] S. Alpay, E. Emelyanov, S. Gorokhova, Regularly limited, Grothendieck, and Dunford–Pettis operators. https://arxiv.org/abs/2208.02303v3
* [5] B. Aqzzouz, K. Bouras, (L) sets and almost (L) sets in Banach lattices. Quaest. Math. 36, no.1 (2013), 107–118.
* [6] B. Aqzzouz, A. Elbour, Some characterizations of almost Dunford–Pettis operators and applications. Positivity 15 (2011), 369–380.
* [7] B. Aqzzouz, A. Elbour, A. W. Wickstead, Positive almost Dunford–Pettis operators and their duality. Positivity 15 (2011), 185–197.
* [8] B. Aqzzouz, J. H’Michane, The class of b-AM-compact operators. Quaest. Math. 36 (2013), 309–319.
* [9] K. Bouras, Almost Dunford–Pettis sets in Banach lattices. Rend. Circ. Mat. Palermo 62 (2013), 227–236.
* [10] K. Bouras, D. Lhaimer, M. Moussa, On the class of almost L-weakly and almost M-weakly compact operators. Positivity 22 (2018), 1433–1443.
* [11] J. Bourgain, J. Diestel, Limited operators and strict cosingularity. Math. Nachr. 119 (1984), 55–58.
* [12] A. V. Buhvalov, Locally convex spaces that are generated by weakly compact sets. Vestnik Leningrad Univ. 7 Mat. Mekh. Astronom. Vyp.2 (1973), 11–17.
* [13] J. X. Chen, Z. L. Chen, G. X. Jia, Almost limited sets in Banach attices. J. Math. Anal. Appl. 412 (2014), 547–553.
* [14] J. Diestel, Sequences and series in Banach spaces. Graduate Texts in Math. 92. Berlin, Heidelberg, New York: Springer, 1984.
* [15] J. Diestel, J. J. Uhl, Vector measures. Mathematical Surveys No. 15, American Math. Soc., 1977.
* [16] P. G. Dodds, D. H. Fremlin, Compact operators in Banach lattices. Isr. J. Math. 34 (1979), 287–320.
* [17] A. Elbour, Some characterizations of almost limited operators. Positivity 21 (2017), 865–874.
* [18] A. Elbour, F. Afkir, M. Sabiri, Some properties of almost L-weakly and almost M-weakly compact operators. Positivity 24 (2020), 141–149.
* [19] A. Elbour, N. Machrafi, M. Moussa, Weak Compactness of Almost Limited Operators. Journal of Function Spaces, (2014), Article ID 263159, 4 pages.
* [20] K. El Fahri, A. El Kaddouri, M. Moussa, Some results on limited operators. Acta Univ. Apulensis Math. Inform. 43 (2015), 125–135.
* [21] E. Emelyanov, Algebras of Lebesgue and KB regular operators on Banach lattices. https://arxiv.org/abs/2203.08326v3
* [22] G. Emmanuele, The (BD) property in $L_{1}(\mu,E)$. Indiana Univ. Math. J. 36, (1987), 229–230.
* [23] P. Galindo, V. C. C. Miranda, Grothendieck-type subsets of Banach lattices. J. Math. Anal. Appl. 506 (2022)
* [24] I. Ghenciu, The weak Gelfand–Phillips property in spaces of compact operators. Comment. Math. Univ. Carol. 58 (2017), 35–47.
* [25] M. González, T. Kania, Grothendieck spaces: the landscape and perspectives. Japan. J. Math. 16 (2021), 247–313.
* [26] J. H’Michane, K. El Fahri, On the domination of limited and order Dunford–Pettis operators. Ann. Math. Que. 39 (2015), 169–176.
* [27] A. El Kaddouri, Mohammed Moussa, About the class of ordered limited operators. Acta Univ. Carolinae 54 (2013), 37–43.
* [28] A. El Kaddouri, K. El Fahri, J. H’Michane, M. Moussa, The class of order-almost limited operators on Banach lattices. Acta Math. Univ. Comenian. (N.S.) 84 (2015), 97–102.
* [29] D. Lhaimer, K. Bouras, M. Moussa, On the class of b-L-weakly and order M-weakly compact operators. Math. Bohem. 145, no. 3 (2020), 255–264.
* [30] D. Lhaimer, K. Bouras, M. Moussa, On the class of order $L$\- and order $M$-weakly compact operators. Positivity 25 (2021), 1569–1578.
* [31] M. L. Lourenco, V. C. C. Miranda, The property (d) and the almost limited completely continuous operators. https://arxiv.org/pdf/2011.02890v3
* [32] N. Machrafi, A. Elbour, M. Moussa, Some characterizations of almost limited sets and applications. https://arxiv.org/abs/1312.2770v2
* [33] N. Machrafi, K. El Fahri, M. Moussa, B. Altin, A note on weak almost limited operators. Hacet. J. Math. Stat. 48(3) (2019) 759–770.
* [34] P. Meyer-Nieberg, Banach Lattices. Universitext, Springer-Verlag, Berlin 1991.
* [35] J. A. Sanchez, Operators on Banach lattices. Ph. D. Thesis, Complutense University, Madrid, 1985.
* [36] W. Wnuk, Banach spaces with properties of the Schur type – a survey. Confer. Sem. Mat. Univ. Bari. 29 (1993), 1–25.
* [37] W. Wnuk, On the dual positive Schur property in Banach lattices. Positivity 17 (2013), 759–773.
* [38] A. C. Zaanen, Riesz spaces II. Amsterdam, The Netherlands: North-Holland Publishing, 1983.
|
# Transport Properties in Gapped Bilayer Graphene
N. Benlakhouy<EMAIL_ADDRESS>Laboratory of Theoretical Physics,
Faculty of Sciences, Chouaïb Doukkali University, PO Box 20, 24000 El Jadida,
Morocco A. El Mouhafid<EMAIL_ADDRESS>Laboratory of Theoretical
Physics, Faculty of Sciences, Chouaïb Doukkali University, PO Box 20, 24000 El
Jadida, Morocco A. Jellal<EMAIL_ADDRESS>Laboratory of Theoretical
Physics, Faculty of Sciences, Chouaïb Doukkali University, PO Box 20, 24000 El
Jadida, Morocco Canadian Quantum Research Center, 204-3002 32 Ave Vernon,
BC V1T 2L7, Canada
###### Abstract
We investigate transport properties through a rectangular potential barrier in
AB-stacked bilayer graphene (AB-BLG) gapped by dielectric layers. Using the
Dirac-like Hamiltonian with a transfer matrix approach we obtain transmission
and reflection probabilities as well as the associated conductance. For two-
band model and at normal incidence, we find extra resonances appearing in
transmission compared to biased AB-BLG, which are Fabry-Pérot resonance type.
Now by taking into account the inter-layer bias, we show that both of
transmission and anti-Klein tunneling are diminished. Regarding four band
model, we find that the gap suppresses transmission in an energy range by
showing some behaviors look like ”Mexican hats”. We examine the total
conductance and show that it is affected by the gap compared to AA-stacked
bilayer graphene. In addition, we find that the suppression in conductance is
more important than that for biased AB-BLG.
###### pacs:
73.22.Pr, 72.80.Vp, 73.63.-b
## I Introduction
The experimental realization of monolayer graphene (MLG) in 2004 by Novoselov
and Geim [1] opened up a new field in physics. Such material has attractive
electronic, optical, thermal, and mechanical properties. In particular, the
observation of Klein tunneling [3, 2], anomalous quantum Hall effect [1, 4],
and optical transparency [5]. This makes graphene a good platform for
nanoscale adaptor applications [6]. Bilayer graphene (BLG) is a system formed
by two stacked sheets of graphene. Besides that, there are two distinct kinds
of stacking: AB-BLG or AB-(Bernal) [7], and AA-BLG. AB-BLG has a parabolic
dispersion relation with four bands where two of them touch at zero energy,
whereas the other two bands split together by the interlayer hopping parameter
$\gamma_{1}\approx 0.4$ eV [8]. This structure is much more stable and its
high-quality samples are developed and studied theoretically and
experimentally [9, 10, 11, 12, 13]. AA-BLG has a linear energy gapless
spectrum with two Dirac cones switched in energy by the quantity
$\gamma_{1}\approx 0.2$ eV [14], and because of this AA-BLG attained enormous
theoretical interest [15, 16, 17, 18, 19, 20]. Such a structure is expected to
be metastable, just lately, stable samples were discovered [21, 22, 23, 24].
The AB-BLG may have clearly defined benefits than MLG, due to greater
possibilities for balancing their physical properties. For reference: quantum
Hall effect [9, 25], spin-orbit coupling and transverse electric field [26],
transmission probability in presence of electric and magnetic static fields
[27, 13], and quantum dots [28].
Experimentally, the evidence of Klein tunneling in MLG was confirmed [29, 3,
30, 31], which means that there is no electron confinement, and then a gap
must be created to overcome this issue. In fact, many methods of induction a
band gap in MLG have been elaborated such as substrates [32, 33, 34, 35, 36,
37, 38, 39] and doping with impurities [40, 41]. Regarding AB-BLG, band gap
can be realized by applying an external electric field [42, 9] or induced by
using dielectric materials like hexagonal boron nitride (h-BN) or SiC [44]. To
this end, it is theoretically showed that quantum spin Hall phase can be
identified in gapped AB-BLG even when the Rashba interaction approached zero
[44].
The introduction of an inter-layer bias to AB-BLG opens a gap in the energy
spectrum and has a major effect on electronic properties [29]. Here, we
analyze the effects of a biased AB-BLG gapped by dielectric layers to show the
impact of band gap on transport properties. In both layers of AB-BLG, band gap
is the same allowing to open a gap. Using transfer matrix method together with
current density, we calculate transmission and reflection probabilities as
well as corresponding conductance. At low-energy, $E<\gamma_{1}$, and in
presence of the band gap $\Delta_{0}$ we find that Fabry-Pérot resonances [48]
strongly appear in the transmission. Now by including also the inter-layer
bias $\delta$, we show that the total transmission and anti-Klein tunneling
significantly diminished. For energies exceeding the inter-layer coupling
$\gamma_{1}$, $E>\gamma_{1}$, we obtain a new mode of propagating giving rise
to the four transmission channels. In this case, $\Delta_{0}$ suppresses the
transmission in the energy range
$V_{0}-(\Delta_{0}+\delta)<E<V_{0}+(\Delta_{0}+\delta)$, and shows some
behaviors that look like “Mexican hats”. Finally we find that the resulting
conductance in gapped AB-BLG gets modified compared to gapped AA-BLG.
Moreover, we find that the suppression in conductance is more important than
that for biased AB-BLG [29] because the energy range for a null conductance
increases as long as $\Delta_{0}$ increase and also the number of peaks get
reduced.
The paper is organized as follows. In Sec II we construct our theoretical
model describing biased and gapped AB-BLG giving rise to four band energies.
In Sec III we explain in detail the formalism used in calculating transmission
and reflection probabilities together with conductance. In Sec IV we
numerically analyze our results and give different discussions with published
works on the topic. Finally, in Sec. V we summarize our main conclusions.
## II Theoretical model
Figure 1: The parameters of a rectangular barrier structure.
In the AB-stacked bilayer graphene the atom $B_{1}$ of the top layer is placed
directly below the atom $A_{2}$ of the bottom layer with van der Waals inter-
layer coupling parameter $\gamma_{1}$, while $A_{1}$ and $B_{2}$ do not lie
directly below or above each other. Based on [29, 44] we consider a biased and
gapped AB-BLG described by the following Hamiltonian near the point K
$\mathcal{H}=\begin{pmatrix}V_{0}+\vartheta_{1}&v_{F}\pi^{{\dagger}}&-v_{4}\pi^{{\dagger}}&v_{3}\pi\\\
v_{F}\pi&V_{0}+\vartheta_{2}&\gamma_{1}&-v_{4}\pi^{{\dagger}}\\\
-v_{4}\pi&\gamma_{1}&V_{0}-\vartheta_{2}&v_{F}\pi^{{\dagger}}\\\
v_{3}\pi^{{\dagger}}&-v_{4}\pi&v_{F}\pi&V_{0}-\vartheta_{1}\\\ \end{pmatrix}$
(1)
where $v_{F}=\frac{\gamma_{0}}{\hbar}\frac{3a}{2}\approx 10^{6}$ m/s is the
Fermi velocity of electrons in each graphene layer, $a=0.142$ nm is the
distance between adjacent carbon atoms,
$v_{3,4}=\frac{v_{F}\gamma_{3,4}}{\gamma_{0}}$ represent the coupling between
the layers, $\pi=p_{x}+ip_{y},\pi^{{\dagger}}=p_{x}-ip_{y}$ are the in-plan
momenta and its conjugate with $p_{x,y}=-i\hbar\partial_{x,y}$,
$\gamma_{1}\approx 0.4$ eV is the interlayer coupling term. The electrostatic
potential $V_{0}$ of width $d$ (Fig. 1) can be varied on the $i$-th layer
using top and back gates on the sample. $\vartheta_{1}=\delta+\Delta_{0}$,
$\vartheta_{2}=\delta-\Delta_{0}$ with $\delta$ corresponds to an externally
induced inter-layer potential difference, and $\Delta_{0}$ is the band gap.
The skew parameters, $\gamma_{3}\approx 0.315$ eV and $\gamma_{4}\approx
0.044$ eV have negligible effect on the band structure at high energy [25,
45]. Recently, it was shown that even at low energy these parameters have also
negligible effect on the transmission [29], hence we neglect them in our
calculations.
Under the above approximation and for a barrier potential configuration as
depicted in Fig. 1, the Hamiltonian (1) can be written as
$H=\left(\begin{array}[]{cccc}V_{0}+\vartheta_{1}&\nu_{F}\pi^{{\dagger}}&0&0\\\
\nu_{F}\pi&V_{0}+\vartheta_{2}&\gamma_{1}&0\\\
0&\gamma_{1}&V_{0}-\vartheta_{2}&\nu_{F}\pi^{{\dagger}}\\\
0&0&\nu_{F}\pi&V_{0}-\vartheta_{1}\\\ \end{array}\right)$ (2)
By considering the length scale $l=\hbar v_{F}/\gamma_{1}$, which represents
the inter-layer coupling length $l=1.64$ nm, we define the dimensionless
quantities: $x\equiv x/l$ and $k_{y}\equiv lk_{y}$ together with
$\delta\equiv\frac{\delta}{\gamma_{1}}$,
$\Delta_{0}\equiv\frac{\Delta_{0}}{\gamma_{1}}$,
$E\equiv\frac{E}{\gamma_{1}}$, $V_{0}\equiv\frac{V_{0}}{\gamma_{1}}$. The
eigenstates of Eq. (2) are four-components spinors
$\psi(x,y)=[{\psi}_{A_{1}},{\psi}_{B_{1}},{\psi}_{A_{2}},{\psi}_{B_{2}}]^{{\dagger}}$,
here ${\dagger}$ denotes the transpose of the row vector. As a consequence of
the transnational invariance along the $y$-direction, we have $[H,p_{y}]=0$,
and then we decompose the spinor as
$\psi(x,y)=e^{ik_{y}y}\left[\phi_{A_{1}}(x),\phi_{B_{1}}(x),\phi_{A_{2}}(x),\phi_{B_{2}}(x)\right]^{T}$
(3)
We solve the time-independent Schrödinger equation $H\psi=E\psi$ to obtain a
general solution in the region II and then require $V_{0}=\delta=\Delta_{0}=0$
to derive the solutions in the regions I and III. Indeed, by substituting Eq.
(2) and Eq. (3) we get four related differential equations
$\displaystyle-i(\partial_{x}+k_{y})\phi_{B_{1}}$ $\displaystyle=$
$\displaystyle\varepsilon_{1}\phi_{A_{1}}$ (4a)
$\displaystyle-i(\partial_{x}-k_{y})\phi_{A_{1}}$ $\displaystyle=$
$\displaystyle\varepsilon_{2}\phi_{B_{1}}-\phi_{A_{2}}$ (4b)
$\displaystyle-i(\partial_{x}+k_{y})\phi_{B_{2}}$ $\displaystyle=$
$\displaystyle\varepsilon_{3}\phi_{A_{2}}-\phi_{B_{1}}$ (4c)
$\displaystyle-i(\partial_{x}-k_{y})\phi_{A_{2}}$ $\displaystyle=$
$\displaystyle\varepsilon_{4}\phi_{B_{2}}$ (4d)
where we have set $\varepsilon_{1}=\varepsilon-\vartheta_{1}$,
$\varepsilon_{2}=\varepsilon-\vartheta_{2}$,
$\varepsilon_{3}=\varepsilon+\vartheta_{2}$,
$\varepsilon_{4}=\varepsilon+\vartheta_{1}$ and $\varepsilon=E-V_{0}$. We
solve Eq. (4a) for $\phi_{A_{1}}$, Eq. (4d) for $\phi_{B_{2}}$ and substitute
the results in Eqs. (4b,4c). This process yields
$\displaystyle(\partial_{x}^{2}-k_{y}^{2}+\varepsilon_{1}\varepsilon_{2})\phi_{B_{1}}$
$\displaystyle=$ $\displaystyle\varepsilon_{1}\phi_{A_{2}}$ (5a)
$\displaystyle(\partial_{x}^{2}-k_{y}^{2}+\varepsilon_{3}\varepsilon_{4})\phi_{A_{2}}$
$\displaystyle=$ $\displaystyle\varepsilon_{4}\phi_{B_{1}}$ (5b)
Then for constant parameters, the energy bands are solution of the following
equation
$\left[-k^{2}+\varepsilon_{1}\varepsilon_{2}\right]\left[-k^{2}+\varepsilon_{3}\varepsilon_{4}\right]-\varepsilon_{1}\varepsilon_{4}=0$
(6)
such that $k=\sqrt{k_{x}^{2}+k_{y}^{2}}$ and the four possible wave vectors
are given by
$k^{s}_{x}=\sqrt{-k_{y}^{2}+\varepsilon^{2}+\delta^{2}-\Delta_{0}^{2}\pm\sqrt{\varepsilon^{2}(1+4\delta^{2})-(\delta+\Delta_{0})^{2}}}$
(7)
where $s=\pm$ defines the modes of propagation, which will be discussed in
numerical section. Therefore, the four energy bands can be derived as
$\displaystyle\varepsilon^{s}_{\pm}=s\sqrt{k^{2}+\delta^{2}+\Delta_{0}^{2}+\frac{1}{2}\pm\sqrt{k^{2}\left(1+4\delta^{2}\right)+\left(\frac{1}{2}-2\delta\Delta_{0}\right)^{2}}}$
(8)
At this level, we have some comments in order. Indeed, firstly by taking
$\delta=0$, (8) reduces
$\varepsilon^{s}_{\pm}|_{\delta=0}=s\sqrt{k^{2}+\Delta_{0}^{2}+\frac{1}{2}\pm\sqrt{k^{2}+\frac{1}{4}}}$
(9)
Secondly for the case $\Delta_{0}=0$, we end up with Ben et al. result [29]
$\displaystyle\varepsilon^{s}_{\pm}|_{\Delta_{0}=0}=s\sqrt{k^{2}+\delta^{2}+\frac{1}{2}\pm\sqrt{k^{2}\left(1+4\delta^{2}\right)+\frac{1}{4}}}$
(10)
Now by comparing (9) and (10), we clearly notice that both quantities $\delta$
and $\Delta_{0}$ are inducing different gaps in the energy spectrum. Certainly
this difference will affect the transmission probabilities (Figs. 3, 4) as
well as conductance (Fig. 7).
Figure 2: Energy spectrum of AB-stacked graphene bilayer inside (solid curves)
and outside (dashed curves) the barrier. Here blue (brown) curves correspond
to $k^{+}(k^{-})$ propagating modes for biased and gapped $(V_{0}\neq
0,\delta\neq 0,\Delta_{0}\neq 0)$ systems. $\delta^{\prime}=\delta+\Delta_{0}$
and $\gamma_{1}^{\prime}=\sqrt{\gamma_{1}^{2}+(\delta-\Delta_{0})^{2}}$.
It is known that the perfect AB-BLG has a parabolic dispersion relation with
four bands, of which two touch each other at $k=0$. In Fig. 2 we show the
energy bands as a function of the momentum $k_{y}$, for the biased and gapped
AB-BLG. We observe that when the AB-BLG is subjected to a gap $\Delta_{0}$ and
an inter-layer bias $\delta$ the two bands are switched and placed at
$V_{0}\pm\sqrt{\gamma_{1}^{2}+(\delta-\Delta_{0})^{2}}$, and the touching
bands are shifted by $2\delta^{\prime}=2(\delta+\Delta_{0})$. One should
notice that there are two cases related to whether the wave vector
$k^{s}_{0}=\sqrt{-k_{y}^{2}+\varepsilon^{2}\pm\varepsilon}$ is real or
imaginary. Indeed for $E<\gamma_{1}$, just $k^{+}_{0}$ is real, and for that
reason, the propagation is only possible for $k^{+}_{0}$ mode. However when
$E>\gamma_{1}$, both $k^{\pm}_{0}$ are real which presenting a new propagation
mode.
As concerning the eigenspinors in regions II, we show that the solution of
Eqs. (5) is a plane wave generated by
$\phi_{B_{1}}^{2}=a_{1}e^{ik_{x}^{+}x}+a_{2}e^{-ik_{x}^{+}x}+a_{3}e^{ik_{x}^{-}x}+a_{4}e^{-ik_{x}^{-}x}$
(11)
where $a_{n}$ are coefficients of normalization, with $n=1,\cdots,4$. The
remaining components of the eigenspinors can be obtained as
$\displaystyle\phi_{A_{1}}^{2}$
$\displaystyle=a_{1}\Lambda^{+}_{+}e^{ik_{x}^{+}x}+a_{2}\Lambda^{+}_{-}e^{-ik_{x}^{+}x}+a_{3}\Lambda^{-}_{+}e^{ik_{x}^{-}x}+a_{4}\Lambda^{-}_{-}e^{-ik_{x}^{-}x}$
(12) $\displaystyle\phi_{A_{2}}^{2}$
$\displaystyle=a_{1}\rho^{+}e^{ik_{x}^{+}x}+a_{2}\rho^{+}e^{-ik_{x}^{+}x}+a_{3}\rho^{-}e^{ik_{x}^{-}x}+a_{4}\rho^{-}e^{-ik_{x}^{-}x}$
(13) $\displaystyle\phi_{B_{2}}^{2}$
$\displaystyle=a_{1}\chi^{+}_{+}\rho^{+}e^{ik_{x}^{+}x}+a_{2}\chi^{+}_{-}\rho^{+}e^{-ik_{x}^{+}x}+a_{3}\chi^{-}_{+}\rho^{-}e^{ik_{x}^{-}x}+a_{4}\chi^{-}_{-}\rho^{-}e^{-ik_{x}^{-}x}$
(14)
where we have introduced the quantities $\Lambda^{\pm}_{\pm}=\frac{-ik_{y}\pm
k_{x}^{\pm}}{\varepsilon-\vartheta_{1}}$,
$\rho^{\pm}=\frac{(\epsilon-\vartheta_{1})(\epsilon-\vartheta_{2})-k_{y}^{2}-(k_{x}^{\pm})^{2}}{\epsilon-\vartheta_{1}}$,
$\chi^{\pm}_{\pm}=\frac{ik_{y}\pm k_{x}^{\pm}}{\varepsilon+\vartheta_{1}}$. In
matrix notation, the general solution of our system in region II can be
written as
$\psi_{2}(x,y)=\mathcal{G}_{2}\cdot\mathcal{M}_{2}(x)\cdot\mathcal{C}_{2}\
e^{ik_{y}y}$ (15)
where the four-component vector ${\cal{C}}_{2}$ represents the coefficients
$a_{n}$ expressing the relative weights of the different traveling modes,
which have to be set according to the propagating region [29]. The matrices
$\mathcal{M}_{2}(x)$ and $\mathcal{G}_{2}$ are given by
$\mathcal{G}_{2}=\begin{pmatrix}1&1&1&1\\\
\Lambda^{+}_{-}&\Lambda^{+}_{+}&\Lambda^{-}_{+}&\Lambda^{-}_{-}\\\
\rho^{+}&\rho^{+}&\rho^{-}&\rho^{-}\\\
\chi^{+}_{+}\rho^{+}&\chi^{+}_{-}\rho^{+}&\chi^{-}_{+}\rho^{-}&\chi^{-}_{-}\rho^{-}\\\
\end{pmatrix},\qquad\mathcal{M}_{2}(x)=\begin{pmatrix}e^{ik_{x}^{+}x}&0&0&0\\\
0&e^{-ik_{x}^{+}x}&0&0\\\ 0&0&e^{ik_{x}^{-}x}&0\\\ 0&0&0&e^{-ik_{x}^{-}x}\\\
\end{pmatrix},\qquad\mathcal{C}_{2}=\begin{pmatrix}a_{1}\\\ a_{2}\\\ a_{3}\\\
a_{4}\\\ \end{pmatrix}$ (16)
As claimed before, to get solutions in the other regions we have to set
$V_{0}=\delta=\Delta_{0}=0$. Then the eigenspinors in region I is
$\displaystyle\phi_{A_{1}}^{1}$
$\displaystyle=\delta_{s,1}e^{ik^{+}_{0}x}+r^{s}_{+}e^{-ik^{+}_{0}x}+\delta_{s,-1}e^{ik^{-}_{0}x}+r^{s}_{-}e^{-ik^{-}_{0}x}$
(17) $\displaystyle\phi_{B_{1}}^{1}$
$\displaystyle=\delta_{s,1}\Lambda^{+}_{-}e^{ik^{+}_{0}x}+r^{s}_{+}\Lambda^{+}_{+}e^{-ik^{+}_{0}x}+\delta_{s,-1}\Lambda^{-}_{+}e^{ik^{-}_{0}x}+r^{s}_{-}\Lambda^{-}_{-}e^{-ik^{-}_{0}x}$
(18) $\displaystyle\phi_{A_{2}}^{1}$
$\displaystyle=\delta_{s,1}\rho^{+}e^{ik^{+}_{0}x}+r^{s}_{+}\rho^{+}e^{-ik^{+}_{0}x}+\delta_{s,-1}\rho^{-}e^{ik^{-}_{0}x}+r^{s}_{-}\rho^{-}e^{-ik^{-}_{0}x}$
(19) $\displaystyle\phi_{B_{2}}^{1}$
$\displaystyle=\delta_{s,1}\chi^{+}_{+}\rho^{+}e^{ik^{+}_{0}x}+r^{s}_{+}\rho^{+}\chi^{+}_{-}e^{-ik^{+}_{0}x}+\delta_{s,-1}\rho^{-}\chi^{-}_{+}e^{ik^{-}_{0}x}+r^{s}_{-}\rho^{-}\chi^{-}_{-}e^{-ik^{-}_{0}x}$
(20)
and in the region III reads as
$\displaystyle\phi_{A_{1}}^{3}$
$\displaystyle=t^{s}_{+}e^{ik^{+}_{0}x}+t^{s}_{-}e^{ik^{-}_{0}x}$ (21)
$\displaystyle\phi_{B_{1}}^{3}$
$\displaystyle=t^{s}_{+}\Lambda^{+}_{-}e^{ik^{+}_{0}x}+t^{s}_{-}\Lambda^{-}_{+}e^{ik^{-}_{0}x}$
(22) $\displaystyle\phi_{A_{2}}^{3}$
$\displaystyle=t^{s}_{+}\rho^{+}e^{ik^{+}_{0}x}+t^{s}_{-}\rho^{-}e^{ik^{-}_{0}x}$
(23) $\displaystyle\phi_{B_{2}}^{3}$
$\displaystyle=t^{s}_{+}\chi^{+}_{+}\rho^{+}e^{ik^{+}_{0}x}+t^{s}_{-}\chi^{-}_{+}\rho_{-}e^{ik^{-}_{0}x}$
(24)
Since the potential is zero in regions I and III, we have the relation
$\mathcal{G}_{1}\cdot\mathcal{M}_{1}(x)=\mathcal{G}_{3}\cdot\mathcal{M}_{3}(x)$.
We will see how the above results will be used to determine different physical
quantities. Specifically, we focus on the transmission and reflection
probabilities as well as the conductance.
## III Transmission probability and conductance
To determine the transmission and reflection probabilities, we impose the
appropriate boundary conditions in the context of the transfer matrix approach
[46, 47]. Continuity of the spinors at interfaces gives the components of the
vectors
$\mathcal{C}_{1}^{s}=\begin{pmatrix}\delta_{s,1}\\\ r_{+}^{s}\\\
\delta_{s,-1}\\\ r_{-}^{s}\\\
\end{pmatrix},\qquad\mathcal{C}_{3}^{s}=\begin{pmatrix}t_{+}^{s}\\\ 0\\\
t_{-}^{s}\\\ 0\\\ \end{pmatrix}$ (25)
where $\delta_{s,\pm}$ is the Kronecker symbol. The continuity at $x=0$ and
$x=d$ can be written in a matrix notation as
$\displaystyle\mathcal{G}_{1}\cdot\mathcal{M}_{1}(0)\cdot\mathcal{C}_{1}^{s}=\mathcal{G}_{2}\cdot\mathcal{M}_{2}(0)\cdot\mathcal{C}_{2}$
(26)
$\displaystyle\mathcal{G}_{2}\cdot\mathcal{M}_{2}(d)\cdot\mathcal{C}_{2}=\mathcal{G}_{3}\cdot\mathcal{M}_{3}(d)\cdot\mathcal{C}_{3}^{s}$
(27)
Using the transfer matrix method together with the relation
$\mathcal{G}_{1}\cdot\mathcal{M}_{1}(x)=\mathcal{G}_{3}\cdot\mathcal{M}_{3}(x)$
we can connect $\mathcal{C}_{1}^{s}$ with $\mathcal{C}_{3}^{s}$ through the
matrix $\mathcal{N}$
$\mathcal{C}_{1}^{s}=\mathcal{N}\cdot\mathcal{C}_{3}^{s}$ (28)
where
$\mathcal{N}=\mathcal{G}_{1}^{-1}\cdot\mathcal{G}_{2}\cdot\mathcal{M}_{2}^{-1}(d)\cdot\mathcal{G}_{2}^{-1}\cdot\mathcal{G}_{1}\cdot\mathcal{M}_{1}(d)$
(29)
Consequently, the transmission and reflection coefficients can be derived from
$\left(\begin{array}[]{cccc}t_{+}^{s}\\\ r_{+}^{s}\\\ t_{-}^{s}\\\
r_{-}^{s}\\\
\end{array}\right)=\left(\begin{array}[]{cccc}\mathcal{N}_{11}&0&\mathcal{N}_{13}&0\\\
\mathcal{N}_{21}&-1&\mathcal{N}_{23}&0\\\
\mathcal{N}_{31}&0&\mathcal{N}_{33}&0\\\
\mathcal{N}_{41}&0&\mathcal{N}_{43}&-1\\\
\end{array}\right)^{-1}\left(\begin{array}[]{cccc}\delta_{s,1}\\\ 0\\\
\delta_{s,-1}\\\ 0\\\ \end{array}\right)$ (30)
where $\mathcal{N}_{ij}$ are the matrix elements of $\mathcal{N}$. Then, after
some algebras, we obtain the transmission and reflection coefficients
$\displaystyle t_{+}^{s}$
$\displaystyle=\frac{\delta_{s,-1}\mathcal{N}_{13}-\delta_{s,1}\mathcal{N}_{33}}{\mathcal{N}_{13}\mathcal{N}_{31}-\mathcal{N}_{11}\mathcal{N}_{33}},\qquad
t_{-}^{s}=\frac{-\delta_{s,-1}\mathcal{N}_{11}+\delta_{s,1}\mathcal{N}_{31}}{\mathcal{N}_{13}\mathcal{N}_{31}-\mathcal{N}_{11}\mathcal{N}_{33}}$
(31) $\displaystyle r_{+}^{s}$
$\displaystyle=\mathcal{N}_{21}t_{+}^{s}+\mathcal{N}_{23}t_{-}^{s},\qquad
r_{-}^{s}=\mathcal{N}_{41}t_{+}^{s}+\mathcal{N}_{43}t_{-}^{s}$ (32)
To calculate the transmission and reflection probabilities, we have to take
into account the change in velocity of the waves when they are scattered into
a different propagation mode. For this, it is convenient to use the current
density $\bm{J}$
$\bm{J}=v_{F}\bm{\psi}^{\dagger}\begin{pmatrix}\sigma_{x}&0\\\ 0&\sigma_{x}\\\
\end{pmatrix}\bm{\psi}$ (33)
where $\sigma_{x}$ is the Pauli matrix. Then Eq. (33) gives the incident
$\bm{J}_{x}^{\text{inc}}$, reflected $\bm{J}_{x}^{\text{ref}}$ and transmitted
$\bm{J}_{x}^{\text{tra}}$ current densities. Finally the transmission $T$ and
reflection $R$ probabilities are
$T^{s}_{\pm}=\frac{k^{\pm}_{0}}{k^{s}_{0}}|t^{s}_{\pm}|^{2},\qquad
R^{s}_{\pm}=\frac{k^{\pm}_{0}}{k^{s}_{0}}|r^{s}_{\pm}|^{2}$ (34)
To preserve the probability of current, $T$ and $R$ are normalized as
$\sum_{i,j}\left(T^{j}_{i}+R^{j}_{i}\right)=1$ (35)
where the index $i=\pm$ points to the arriving mode, when the index $j=\pm$
points to the exiting mode. For example in the case of channel $k^{+}$, gives
$T^{+}_{+}+T^{-}_{+}+R^{+}_{+}+R^{-}_{+}=1$. As already mentioned, for
$E>\gamma_{1}$ we have two modes of propagation ($k^{+}_{0},k^{-}_{0}$)
leading to four transmissions $T^{s}_{\pm}$ and four reflections $R^{s}_{\pm}$
channels, through the four conduction bands. For sufficiently enough low
energy or in the two-band model, $E<\gamma_{1}$, the two modes lead to one
transmission $T$ channel and one reflection $R$ channel.
From the transmission probabilities, we can calculate the conductance $G$, at
zero temperature, using the Landauer-Büttiker formula
$G(E)=G_{0}\frac{L_{y}}{2\pi}\int_{-\infty}^{\infty}dk_{y}\sum_{i,j=\pm}T_{i}^{j}\left(E,k_{y}\right)$
(36)
with $L_{y}$ the length of the sample in the $y$-direction, and
$G_{0}=4e^{2}/h$. The factor $4$ comes from the valley and spin degeneracies
in graphene. In order to get the total conductance of the system, we need to
sum over all the transmission channels
$G_{T}=\sum_{i,j}G^{j}_{i}$ (37)
## IV NUMERICAL RESULTS AND DISCUSSION
In this section, we numerically analyze and discuss our main results. First,
we evaluate the transmission probability in the two-band model at normal
incidence (i.e. $k_{y}=0$). To understand our system more effectively in Fig.
3, we present the effect of the band gap $\Delta_{0}$ on the transmission as a
function of the incident energy $E$ and the width $d$ of the barrier. In the
(left panel), we plot the energy dependence of the transmission probability
for a barrier of width $d=10$ nm, $d=25$ nm, and $d=100$ nm for biased
$\delta=0$ and unbiased system $\delta\neq 0$ with band gap $\Delta_{0}$. For
$\Delta_{0}\neq 0$, we observe appearance of resonances in the transmission
probability for the energy range $E<V_{0}-\delta^{\prime}$,
$\delta^{\prime}=\delta+\Delta_{0}$, which can be attributed to the finite
size of the AB-BLG as well as the presence of charge carriers with different
chirality. These phenomena are known as Fabry-Pérot resonances [48]. For the
energy range $V_{0}-\delta^{\prime}<E<V_{0}+\delta^{\prime}$, there is a bowl
(window) of zero transmission for $d=100$ nm in contrary for $d=10$ nm and
$d=25$ nm the transmission is not zero. However, for
$E>V_{0}+\delta^{\prime}$, the transmission still looks like Ben et al.
results [29]. Note that the transmission of width $d=100$ nm, shows anti-Klien
tunneling, which is a direct consequence of the pseudospin conservation in the
system. In the (right panel), we plot the width dependence of the transmission
probability for the incident energies $E=\frac{1}{5}V_{0}$,
$E=\frac{2}{5}V_{0}$ and $E=\frac{8}{5}V_{0}$. It is clearly seen that for
$E=\frac{1}{5}V_{0}$ and $E=\frac{2}{5}V_{0}$ with $\delta_{0}=0$,
$\Delta_{0}=0.01\gamma_{1}$, resonance peaks show up (see upper panel), which
are absent for the case $\Delta_{0}=0$ [29]. In the middle and bottom panel,
by taking into account the effect of a finite bias $\delta=0.01\gamma_{1}$, we
observe a decrease of resonance in the transmission probability, and more
precisely when $\Delta_{0}$ is greater than $\delta$.
Figure 3: (Color online) The transmission probability at normal incidence
through a barrier of height $V_{0}=0.05\gamma_{1}$ with
$\Delta_{0}=0.01\gamma_{1}$ and $\delta=0$ (for upper panel),
$\Delta_{0}=\delta=0.01\gamma_{1}$ (for middle panel) and
$\Delta_{0}=0.03\gamma_{1}$ and $\delta=0.01\gamma_{1}$ (for bottom panel).
(Left panel): The energy dependence of the transmission probability for
barrier widths $d=10$ nm (blue), $d=25$ nm (red), and $d=100$ nm (green).
(Right panel): The width dependence of the transmission probability for
incident energies $E=\frac{1}{5}V_{0}$ (blue), $E=\frac{2}{5}V_{0}$ (red) and
$E=\frac{8}{5}V_{0}$ (green).
To investigate the effect of band gap, for energy greater than the interlayer
hopping parameter, $E>\gamma_{1}$, in Fig. 4 we show the transmission and
reflection channels as a function of the incident energy $E$ and transverse
wave vector $k_{y}$ for potential height $V_{0}=\frac{3}{2}\gamma_{1}$ and
width $d=25$ nm. The superimposed dashed curves indicate different propagating
modes inside and outside the barriers. For ungapped and unbiased AB-BLG
(pristine AB-BLG), Ben et al. [29] showed that all channels are symmetric with
respect to normal incidence, $k_{y}=0$, i.e. $T^{+}_{-}=T^{-}_{+}$ and
$R^{+}_{-}=R^{-}_{+}$. This is due to the valley equivalence, namely the
transmission probabilities of electrons moving in the opposite direction
(scattering from $k^{+}$ to $k^{-}$ in the vicinity of the first valley, and
scattering from $k^{-}$ to $k^{+}$ in the vicinity of the second valley) are
the same. Now as for our case by introducing a gap $\Delta_{0}=0.3\gamma_{1}$,
with a null inter-layer bias, $\delta=0$, we observe that the transmissions
are completely suppressed in the energy range
$V_{0}-\Delta_{0}<E<V_{0}+\Delta_{0}$ due to the absence of traveling modes.
In $T^{+}_{+}$ channel and for energies smaller than $V_{0}-\gamma_{1}$, we
find that the resonances are decreased and Klein tunneling get less
incandescent than that seen in [29]. We notice that there is asymmetric in the
transmission channels with respect to normal incidence,
$T^{+}_{-}(k_{y})=T^{-}_{+}(-k_{y})$, but reflection channels still showing
symmetric behavior, $R^{+}_{-}(k_{y})=R^{-}_{+}(k_{y})$, because the incident
electrons back again in an electron state [29]. This is not the case for
gapped AA-BLG, whereas $T^{+}_{-}$ and $T^{-}_{+}$ channels preserve the
momentum symmetry [43]. In addition, there is a significant distinction for
all reflection channels, $R^{s}_{\pm}$, between gapped AB-BLG and biased AB-
BLG. Indeed, in our case we observe that the scales of $R^{s}_{\pm}$ get
reduced inside the barrier. It is remarkably seen that our transmission
channels, $T^{s}_{\pm}$, showed some bowels in the energy spectrum instead of
“Mexican hats” as have been see in [29]. This show that $\Delta_{0}$ can be
used to control the transmission behavior in AB-BLG.
Figure 4: (Color online) Density plot of transmission and reflection
probabilities as a function of the incident energy $E$ and transverse wave
vector $k_{y}$, through a potential barrier of height $V_{0}=1.5\gamma_{1}$
and width $d=25$ nm and band gap $\Delta_{0}=0.3\gamma_{1}$ with $\delta=0$.
The dashed white and black lines represent the band inside and outside the
barrier, respectively.
In Fig. 5 we show the density plot of the transmission and reflection
channels, for biased and gapped systems, $\delta={0.3}\gamma_{1}$,
$\Delta_{0}=0.3\gamma_{1}$. The transmission is completely suppressed in the
energy range $V_{0}-\delta^{\prime}<E<V_{0}+\delta^{\prime}$,
$\delta^{\prime}=\Delta_{0}+\delta$ . We notice that the symmetric inter-layer
sublattice equivalence is also broken in this case as seen in Fig. 4. We
recall that such symmetry broken can be achieved by taking either $\delta\neq
0$ or $\Delta_{0}\neq 0$, which means that there is violation of invariance
under the exchange $k_{y}\longrightarrow-k_{y}$ as noted in [29, 49] for AB-
BLG, in contrast to the AA-BLG [50]. Therefore, the transmission and
reflection probabilities are not symmetric with respect to normal incidence as
seen in Fig. 5.
Figure 5: (Color online) The same as in Fig. 4, but now for the band gap
$\Delta_{0}=0.3\gamma_{1}$ with $\delta=0.3\gamma_{1}$. The dashed white and
black lines represent the band inside and outside the barrier, respectively.
Fig. 6 presents the same plot as in Fig. 5 except that we choose a band gap
$\Delta_{0}=0.5\gamma_{1}$ greater than inter-layer bias
$\delta=0.3\gamma_{1}$. In this situation, we notice a significant difference
in the transmission and reflection channels. Indeed, we observe that Klein
tunneling becomes less than that see for the case
$\Delta_{0}=\delta=0.3\gamma_{1}$ in Fig. 5. In addition, it is clearly seen
that some resonances disappear for the energy range $E<V_{0}-\delta^{\prime}$.
Moreover, we find that the energy bands are pushed and showed some behaviors
look like “Mexican hats”, which are more clear than those see in Fig. 5. These
results are similar to those obtained in [51], by analyzing the transmission
probabilities for a system composed of two single layer-AB bilayer-two single
layer (2SL-AB-2SL) of graphene subjected to strong gate potential. In summary,
we observe that all transmissions for $\delta\neq 0$ and $\Delta_{0}\neq 0$
are weak compared to the biased AB-BLG [29], or gapped AB-BLG (Fig. 4) cases.
Figure 6: (Color online) The same as in Fig. 4, but now for the band gap
$\Delta_{0}=0.5\gamma_{1}$ with $\delta=0.3\gamma_{1}$. The dashed white and
black lines represent the band inside and outside the barrier, respectively.
In Figs. 7 we plot the energy dependence of the corresponding conductance for
different values of the band gap and an inter-layer bias
$\delta=0.3\gamma_{1}$. The band gap $\Delta_{0}=0.3\gamma_{1}$ contributed by
opening a gap in the energy spectrum of AB-BLG at $V_{0}\pm\Delta_{0}$, and
this of course reflected on the conductance as shown in Fig. 7(a). The
resonances that are clear in the transmission probability show up as peaks,
and the total conductance $G_{\text{Tot}}$ has a convex form. For low energies
we have $G_{\text{Tot}}=G^{+}_{+}$ meaning that the propagation is only via
$k^{+}$ mode, while $k^{-}$ mode is cloaked in this regime until
$E>V_{0}+\Delta_{0}$. $G^{-}_{-}$ starts conducting by making an appearance as
a rapid increase in the total conductance. Furthermore,
$G^{+}_{-}=G^{-}_{+}=0$ since $T^{+}_{-}=T^{-}_{+}=0$ at low energy but at
$E=\gamma_{1}$ both modes are coupled and $G^{+}_{-}$, $G^{+}_{-}$ start
conducting that is why $G_{\text{Tot}}\neq G^{+}_{+}$. However the band gap
does not break the equivalence in the scattered channels of the conductance
such that $G^{-}_{+}=G^{+}_{-}$ still equivalent for all energy ranges (see
Fig. 7(a)), in contrast to the case of the double barriers [52]. By comparing
our results with those of the biased AB-BLG [29], we observe that some
shoulders of the peaks are removed and the contribution of the transmission
channels on the total conductance are not much more pronounced as a result of
the gap $\Delta_{0}$ induced by dielectric layers. This confirms that our
$\Delta_{0}$ has a significant impact on the transport properties and differs
from that induced by bias in AB-BLG [29]. Instead of contrast, the total
conductance of a gapped AA-BLG is approximately unchanged even though the band
gap has a significant impact on the intracone transport [43]. Now we involve
both of parameters by presenting Figs 7(b) and 7(c) corresponding,
respectively, to $\Delta_{0}=\delta=0.3\gamma_{1}$, and
$\Delta_{0}=0.5\gamma_{1},\delta=0.3\gamma_{1}$. As expected we observe large
suppression of the conductance in the energy range
$V_{0}-\delta^{\prime}<E<V_{0}+\delta^{\prime}$, and hence some peaks are
removed with a decrease of the total conductance $G_{\text{Tot}}$.
Figure 7: (Color online): Conductance as a function of the incident energy for
biased and gapped AB-BLG with potential height $V_{0}=1.5\ \gamma_{1}$ and
width $d=25$ nm. (a): $\Delta_{0}=0.3\gamma_{1}$, $\delta=0$. (b):
$\Delta_{0}=0.3\gamma_{1}$, $\delta=0.3\gamma_{1}$, (c):
$\Delta_{0}=0.5\gamma_{1}$, $\delta=0.3\gamma_{1}$. The solid curves
correspond to the total conductance and the dashed curves correspond to
different contributions of the four transmission channels.
## V Summary and conclusion
We have theoretically investigated the transport properties through
rectangular potential barriers of biased AB-BLG gapped by dielectric layers.
By solving Dirac equation, the four band energies are obtained to be dependent
on the band gap $\Delta_{0}$ together with the inter-layer bias $\delta$.
Subsequently, using transfer matrix method we have evaluated the corresponding
transmission, reflection probabilities, and conductance. In particular, we
have analyzed the transmission probability in the two-band model at normal
incidence, (i.e $k_{y}=0$), firstly in the presence of $\Delta_{0}$ and
secondly by taking into account $\Delta_{0}$ and $\delta$. As a result, we
have observed that the presence of $\Delta_{0}$ induces extra resonances
appearing in transmission profiles. However by adding $\delta$, we have
observed that the transmission decreased more and anti-Klein tunneling in AB-
BLG is no longer preserved.
Furthermore, we have obtained a new mode of propagation for energies exceeding
the inter-layer coupling $\gamma_{1}$. In this case, we have showed that the
band gap $\Delta_{0}$ breaks the inter-layer sublattice equivalence with
respect to $k_{y}=0$. Such asymmetry is apparent in the scattered transmission
where it depends on the incident mode. The corresponding conductance does not
incorporate this asymmetric, and the locations of their peaks are changed in
the presence of $\Delta_{0}$ compared to $\delta$ [29].
## VI Acknowledgments
The generous support provided by the Saudi Center for Theoretical Physics
(SCTP) is highly appreciated by all authors. A.J. thanks Dr. Michael Vogl for
fruitful discussion.
## References
* [1] K. S. Novoselov, A. K. Geim, S. V. Morozov, D. Jiang, Y. Zhang, S. V. Dubonos, I. V. Grigorieva and A. A. Firsov, Science 306, 666 (2004).
* [2] O. Klein, Z. Phys. 53, 157 (1929).
* [3] M. I. Katsnelson, K. S. Novoselov, and A. K. Geim, Nat. Phys. 2, 620 (2006).
* [4] Y. B. Zhang, Y. W. Tan, H. L. Stormer, and P. Kim, Nature 438, 201 (2005).
* [5] R. Nair, P. Blake, A. Grigorenko, K. Novoselov, T. Booth, T. Stauber, N. Peres, and A. Geim, Science 320, 1308 (2008).
* [6] A. H. Castro Neto, F. Guinea, N. M. R. Peres, K. S. Novoselov, and A. K. Geim, Rev. Mod. Phys. 81, 109 (2009).
* [7] J. D. Bernal, Proc. R. Soc. A 106, 749 (1924).
* [8] Z. Q. Li, E. A. Henriksen, Z. Jiang, Z. Hao, M. C. Martin, P. Kim, H. L. Stormer, and D. N. Basov, Phys. Rev. Lett. 102, 037403 (2009).
* [9] E. McCann and V. I. Fal’ko, Phys. Rev. Lett. 96, 086805 (2006).
* [10] A. Rozhkov, A. Sboychakov, A. Rakhmanov, and F. Nori, Phys. Rep. 648, 1 (2016).
* [11] T. Ohta, A. Bostwick, T. Seyller, K. Horn, and E. Rotenberg, Science 313, 951 (2006).
* [12] M. O. Goerbig, Rev. Mod. Phys. 83, 1193 (2011).
* [13] I. Redouani, A. Jellal, and H. Bahlouli, J. Low Temp. Phys. 181, 197 (2015).
* [14] I. Lobato and B. Partoens, Phys. Rev. B 83, 165429 (2011).
* [15] A. L. Rakhmanov, A. V. Rozhkov, A. O. Sboychakov, and F. Nori, Phys. Rev. Lett. 109, 206801 (2012).
* [16] Y. Mohammadi and B. A. Nia, Solid State Commun. 201, 76 (2015).
* [17] R.-B. Chen, Y.-H. Chiu, and M.-F. Lin, Carbon 54, 268 (2013).
* [18] C.-W. Chiu, S.-C. Chen, Y.-C. Huang, F.-L. Shyu, and M.-F. Lin, Appl. Phys. Lett. 103, 041907 (2013).
* [19] I. Redouani and A. Jellal, Mater. Res. Express 3, 065005 (2016).
* [20] Y. Zahidi, I. Redouani, and A. Jellal, Physica E 71, 259 (2016).
* [21] J.-K. Lee, S.-C. Lee, J.-P. Ahn, S.-C. Kim, J. I. B. Wilson, and P. John, J. Chem. Phys. 129, 234709 (2008).
* [22] J. Borysiuk, J. Soltys, and J. Piechota, J. Appl. Phys. 109, 093523 (2011).
* [23] P. L. de Andres, R. Ramírez, and J. A. Vergés, Phys. Rev. B 77, 045403 (2008).
* [24] Z. Liu, K. Suenaga, P. J. F. Harris, and S. Iijima, Phys. Rev. Lett. 102, 015501 (2009).
* [25] E. McCann, Phys. Rev. B 74, 161403(R) (2006).
* [26] S. Konschuh, M. Gmitra, D. Kochan, and J. Fabian, Phys. Rev. B 85, 115423 (2012).
* [27] A. Jellal, I. Redouani and H. Bahlouli, Physica E 72, 149 (2015).
* [28] G. Giavaras and F. Nori, Phys. Rev. B 83, 165427 (2011).
* [29] B. Van Duppen and F. M. Peeters, Phys. Rev. B 87, 205427 (2013).
* [30] A. F. Young and P. Kim, Nat. Phys. 5, 222 (2009).
* [31] N. Stander, B. Huard, and D. Goldhaber-Gordon, Phys. Rev. Lett. 102, 026807 (2009).
* [32] W.-X. Wang, L.-J. Yin, J.-B. Qiao, T. Cai, S.-Y. Li, R.-F. Dou, J.-C. Nie, X. Wu, and L. He, Phys. Rev. B 92, 165420 (2015).
* [33] P. San-Jose, A. Gutiérrez-Rubio, M. Sturla, and F. Guinea, Phys. Rev. B 90, 075428 (2014).
* [34] M. Kindermann, B. Uchoa, and D. L. Miller, Phys. Rev. B 86, 115415 (2012).
* [35] J. C. W. Song, A. V. Shytov, and L. S. Levitov, Phys. Rev. Lett. 111, 266801 (2013).
* [36] J. Jung, A. M. DaSilva, A. H. MacDonald, and S. Adam, Nat. Commun. 6, 6308 (2015).
* [37] M. S. Nevius, M. Conrad, F. Wang, A. Celis, M. N. Nair, A. Taleb-Ibrahimi, A. Tejeda, and E. H. Conrad, Phys. Rev. Lett. 115, 136802 (2015).
* [38] M. Zarenia, O. Leenaerts, B. Partoens, and F. M. Peeters, Phys. Rev. B 86, 085451 (2012).
* [39] B. Uchoa, V. N. Kotov, and M. Kindermann, Phys. Rev. B 91, 121412(R) (2015).
* [40] S.Y. Zhou, D.A. Siegel, A.V. Fedorov, A. Lanzara, Phys. Rev. Lett. 101, 086402 (2008).
* [41] R. N. Costa Filho, G. A. Farias, and F. M. Peeters, Phys. Rev. B 76, 193409 (2007).
* [42] Y. Zhang, T-T. Tang, C. Girit, Z. Hao, M.C Martin, A. Zettl, M. F Crommie, Y R. Shen, and F. Wang, Nat. 459 820 (2009).
* [43] H. M. Abdullah and H. Bahlouli, J. Comput. Sci. 26, 135 (2018).
* [44] X. Zhai and G. Jin, Phys. Rev. B 93, 205427 (2016).
* [45] E. McCann, D.S.L. Abergel, and V.I. Fal’ko, Solid State Communications 143, 110 (2007).
* [46] M. Barbier, P. Vasilopoulos, and F. M. Peeters, Phys. Rev. B 82, 235408 (2010).
* [47] Michaël Barbier, P. Vasilopoulos, F. M. Peeters, and J. M. Pereira, Jr, Phys. Rev. B 79, 155402 (2009).
* [48] I. Snyman and C. W. J. Beenakker, Phys. Rev. B 75, 045322 (2007).
* [49] J. Nilsson, A. H. Castro Neto, F. Guinea, and N. M. R. Peres, Phys. Rev. B 76, 165416 (2007).
* [50] H. M. Abdullah, M. A. Ezzi, and H. Bahlouli, J. App. Phy. 124, 204303 (2018).
* [51] H. M. Abdullah, B. Van Duppen, M. Zarenia , H. Bahlouli, and F. M. Peeters, J. Phys.: Condens. Matter 29, 425303 (2017).
* [52] H. M. Abdullah, A. El Mouhafid, H. Bahlouli, and A. Jellal, Mater. Res. Express 4, 025009 (2017).
|
# Thermal Casimir effect for a Dirac field on flat space with a nontrivial
circular boundary condition
Joás Venâncio<EMAIL_ADDRESS>Departamento de Física, Universidade
Federal de Pernambuco, Recife, Pernambuco 50740-560, Brazil. Lameque Filho
<EMAIL_ADDRESS>Departamento de Física, Universidade Federal da
Paraíba, João Pessoa, Caixa Postal 5008, Brazil. Herondy Mota
<EMAIL_ADDRESS>Departamento de Física, Universidade Federal da Paraíba,
João Pessoa, Caixa Postal 5008, Brazil. Azadeh Mohammadi
<EMAIL_ADDRESS>Departamento de Física, Universidade Federal de
Pernambuco, Recife, Pernambuco 50740-560, Brazil.
###### Abstract
This work investigates the thermal Casimir effect associated with a massive
spinor field defined on a four-dimensional flat space with a circularly
compactified spatial dimension whose periodicity is oriented along a vector in
$xy$-plane. We employ the generalized zeta function method to establish a
finite definition for the vacuum free energy density. This definition
conveniently separates into the zero-temperature Casimir energy density and
additional terms accounting for temperature corrections. The structure of
existing divergences is analyzed from the asymptotic behavior of the spinor
heat kernel function and removed in the renormalization by subtracting scheme.
The only non-null heat coefficient is the one associated with the Euclidean
divergence. We also address the need for a finite renormalization to treat the
ambiguity in the zeta function regularization prescription associated with
this Euclidean heat kernel coefficient and ensure that the renormalization
procedure is unique. The high- and low-temperature asymptotic limits are also
explored. In particular, we explicitly show that free energy density lacks a
classical limit at high temperatures, and the entropy density agrees with the
Nernst heat theorem at low temperatures.
Casimir effect. Spinor field. Zeta function. Heat kernel. Boundary condition
## I Introduction
The Casimir effect is a fascinating quantum phenomenon initially proposed by
H. Casimir in 1948 Casimir . In its standard form, such an effect establishes
that two parallel, electrically neutral conducting plates in close proximity
experience an attractive force inversely proportional to the fourth power of
the distance between them. This attraction arises from alterations in vacuum
fluctuations of the electromagnetic field. Since this force between the plates
is extremely weak, the Casimir effect was initially perceived as a theoretical
curiosity. M. Sparnaay conducted the pioneering experimental attempt, however
with low precision, to detect this effect in 1958 Sparnaay . It was only
confirmed decades after by several high-accuracy experiments Bressi2002 ;
Lamoreaux ; Lamoreauxx ; Mohideen . Since then, spurred by the progress in
theories of particles and fields, the Casimir effect has been investigated in
increasingly complicated configurations, not only due to its theoretical and
mathematical aspects but also due to the countless technological applications
arising from the macroscopic manifestation of a fully quantum effect Bordag ;
Klimchitskaya ; Mostepanenko ; Ford1975 ; Dowker 1976 ; Dowker1978 ;
Appelquist1983 ; Hosotani1983 ; Brevik2002 ; Zhang2015 ; Henke2018 ;
Bradonjic2009 ; Peng2018 ; Pawlowski2013 ; Gambassi2009 ; Machta2012 ;
Milton2019 . A thorough review concerning the Casimir effect is presented in
Refs. Klimchitskaya2009 ; Milton2001 .
Although originally associated with the electromagnetic field, the Casimir
effect is not an exclusive feature of this particular field. Other fields, for
instance, scalar and spinor fields, and gauge fields (Abelian and non-
Abelian), can exhibit analogous phenomena under nontrivial boundary conditions
Stokes2015 ; Farina2006 ; Muniz2018 ; Cunha2016 ; Mobassem2014 ; Bytsenko2005
; Pereira2017 ; Photon2001 ; Chernodub2018 ; Edery2007 . However, among the
vast literature concerning the Casimir effect, the majority of the
investigations have been focused on scalar fields. The reason for this is not
conceptual but, most likely, the more significant technical complexity
involved in the formalism needed to treat spinor fields, for instance.
Spinor fields play an important role in many branches of physics since they
represent fermion fields. Additionally, they carry the fundamental
representation of the orthogonal group, making spinors the building block of
all other representations of this group. In this sense, spinors are the most
fundamental entities of a space endowed with a metric Cartan1966 ; Benn1987 ;
JoasBook2019 . In particular, studying vacuum energy associated with the
quantized version of these fields sets a scenario for which the physics
involved is quite rich.
The presence of divergencies is an inherent feature of vacuum energy when
calculated with the quantum field theory (QFT) techniques. Knowing how to deal
with them is challenging in general. This special concern has resulted in the
development of regularization and renormalization techniques in mathematical
physics, which can be applied to remove the divergences associated with the
calculations involved in the Casimir effect Oikonomou2010 ; Cheng2010 ;
Cavalcanti2004 ; Elizalde2008 . This study concentrates explicitly on a robust
and elegant regularization method employing the generalized zeta function.
This function is constructed from the eigenvalues of a differential operator,
which governs the quantum field dynamics Elizalde1995 ; Hawking1977 ;
Elizalde2012 . The divergencies are typically introduced in the partition
function in QFT by the determinant of the operator, which is an infinite
product over all eigenvalues, and encoded into the generalized zeta function
Elizalde1994 . Once we obtain the partition function, the canonical ensemble
establishes the formal connection with thermodynamics. It facilitates the
calculation of free energy, which allows for considering temperature
corrections to the vacuum energy. Basil1978 ; Plunien1986 ; Kulikov1988 ;
Maluf2020 . The structure of the existing divergences in these calculations
typically involves examining the asymptotic behavior of the two-point heat
kernel function associated with the relevant operator, as considered in M.
Kac’s seminal paper Kac1966 and further explored in Elizalde1994 ; Bordag2000
; Vassilevich2003 ; Kirstein2010 . This zeta function investigation
predominantly focuses on Laplace-type operators associated with scalar fields,
with comparatively less emphasis on Dirac-type operators associated with
spinor fields Branson1992A ; Branson1992B .
One potential explanation for this disparity is the requirement for the
considered operator, which governs the propagation of the quantum field under
specified boundary conditions, to be self-adjoint. The self-adjointness is
necessary for the construction of zeta and heat kernel functions. The most
common boundary conditions, widely used in the Casimir effect for the scalar
field, are Dirichlet and Neumann ones. However, these conditions do not
directly extend to spinor fields due to the first-order nature of Dirac
operators. Instead, the bag model boundary conditions first presented in Refs.
Chodos1974 ; Johnson1975 make the Dirac operator formally self-adjoint. This
was also investigated in Ref. Arrizabalaga2017 recently. In particular, the
Casimir effect for spinor fields under bag model boundary conditions has been
addressed in Ref. Mamayev1980 and for Majorana spinor fields with temperature
corrections in Refs. Oikonomou2010 ; Cheng2010 ; Erdas2011 ; Elizalde2012Maj .
Alternative methods to maintain the self-adjoint nature of the Dirac operator
have also been explored. For example, the Casimir effect involving spinor
fields confined by a spherical boundary has been examined in Refs. Bender1976
; Elizalde1998 using the zeta function method. This approach was recently
extended to include a spherically symmetric $\delta$-function potential in
Ref. Fucci2023 . Furthermore, Elko fields, which are spinor fields satisfying
a Klein-Gordon-like equation, allow for the imposition of boundary conditions
similar to those used for scalar fields. The finite temperature Casimir effect
for Elko spinor fields in a field theory at a Lifshitz fixed point is
discussed in Refs. Pereira2017 ; Pereira2019 ; Maluf2020 .
Boundary conditions play a pivotal role in the exploration of the Casimir
effect. Interestingly, it is possible to induce boundary conditions through
identification conditions in spaces with nontrivial topology, thereby
eliminating the need for material boundaries. Such topologies induce boundary
conditions on the quantum fields that distort the corresponding vacuum
fluctuations, such as a material boundary does, producing a Casimir-like
effect Klimchitskaya ; Milton2001 . The Casimir effect for different types of
fields and boundary conditions in spaces with nontrivial topology has been
addressed in Refs. Mostepanenko2011 ; HerondyJunior2015 ; Mohammadi2022 ;
Herondy2023 ; Farias2020 ; Xin-zhou ; Zhai ; Li ; Xin2011 .
In the present work, we have delved into the thermal Casimir effect using the
generalized zeta function approach for a massive spinor field defined on a
four-dimensional flat space with a circularly compactified spatial dimension,
whose periodicity is oriented not along a coordinate axis as usual, but along
a vector in the ${xy\text{-plane}}$, dubbed compact vector. This space
introduces a topological constraint that imposes a spatial anti-periodic
boundary condition along the compact vector on the spinor field. Up to a
coordinate origin redefinition, this condition is referred to as the anti-
helix condition in Ref. Xiang2011 , where the authors investigated the zero-
temperature Casimir effect for spinor fields induced by the helix topology.
However, to our knowledge, a study that adds thermal effects induced by this
topology in the spinor field context has not appeared in the literature. The
calculations conducted in this study not only extend the findings from Ref.
Xiang2011 to finite temperature but also revisit the results from Ref.
Bellucci2009 in a limiting case. Additionally, our study serves as a spinor
extension of the thermal Casimir effect studied in Ref. Giulia2021 , which
focused on scalar fields subjected to a helix boundary condition.
The structure of this paper is organized as follows. Section II provides a
general expression for the partition function associated with a massive Dirac
field defined on a space endowed with a flat Euclidean metric, in the path
integral representation. In Section III, we outline the mathematical framework
employed to compute the vacuum-free energy using the generalized zeta function
method. This method involves imposing an anti-periodic condition on the Dirac
field in imaginary time $t$ and analyzing existing divergences based on the
asymptotic behavior of the spinor heat kernel. In particular, we discuss the
presence of ambiguities in the zeta function regularization due to nonzero
heat kernel coefficients and the necessity of requiring vacuum energy to
renormalize to zero for large masses. In Section IV, we derive the spinor heat
kernel two-point function and the Casimir energy density, incorporating
temperature corrections induced by the anti-periodic boundary condition along
the compact vector. We also analyze the low- and high-temperature asymptotic
limits. Finally, Section V provides a summary of the paper, highlighting the
distinctions between the spinor and scalar cases. Throughout this paper, we
adopt the natural units where $c=\hbar=k_{B}=1$.
## II Path integrals
To illustrate the use of the generalized zeta function method in quantum field
theory (QFT), we revisit some known underlying facts. In the path integral
formulation, the one-loop partition function associated with a complex matter
field $\Psi$ (and its conjugate $\bar{\Psi}$) can be obtained from the
following source-free generating functional
$Z=\int\mathcal{D}\bar{\Psi}\mathcal{D}\Psi\,e^{i\mathcal{S}(\Psi,\bar{\Psi})},$
(1)
where $\mathcal{D}\bar{\Psi}\mathcal{D}\Psi$ stands for the integration
measure over the field space, whose dynamics is described by the action
${\mathcal{S}}(\Psi,\bar{\Psi})$. Such representation provides a
straightforward method for introducing temperature into QFT. This can be
achieved by defining a Euclidean action $\mathcal{S}_{E}(\Psi,\bar{\Psi})$
through a rotation in the complex plane, known as Wick rotation, with the
fields satisfying periodic (for scalar fields) or anti-periodic (for spinor
fields) conditions in imaginary time with period $\beta$. In this Euclidean
approach to QFT, $Z$ is the one-loop partition function for a canonical
ensemble at the temperature $T=\beta^{-1}$.
### II.1 Spinor fields
We can start with the path integral for spinor fields. Let
$\\{\boldsymbol{e}_{a}\\}~{}(a=1,2,3,\ldots,N)$ be an orthonormal frame field
that spans $M=(\mathbb{R}^{N},\boldsymbol{g})$, a $N$-dimensional space
endowed with a flat Euclidean metric $\boldsymbol{g}$ whose components with
respect to basis $\\{\boldsymbol{e}_{a}\\}$ are
$\boldsymbol{g}(\boldsymbol{e}_{a},\boldsymbol{e}_{b})=\delta_{ab},$ (2)
where $\delta_{ab}$ is the Kronecker delta. That is, the space can be covered
by cartesian coordinates ${\\{x^{a}\\}=\\{t,x,y,\dots,z\\}}$ such that the
line element on $M$ is given by
$\displaystyle ds^{2}=dt^{2}+dx^{2}+dy^{2}+\cdots+dz^{2}.$ (3)
The imaginary time coordinate $t$ is compactified into a finite length equal
to the inverse of temperature $\beta$, so that $M$ is closed in the
$t$-direction. This is equivalent to consider spinor fields on
$M=\mathbb{S}^{1}\times\mathbb{R}^{N-1}$ satisfying anti-periodic boundary
conditions. The associated action has the form
$\mathcal{S}_{E}(\Psi,\bar{\Psi})=\int_{0}^{\beta}dt\int
d^{N-1}x\,\sqrt{g}\,\bar{\Psi}(x)\not{D}(m)\Psi(x),$ (4)
where $g$ is the metric determinant and $\not{D}(m)$ is the standard skew-
adjoint Dirac operator, $\not{D}=\gamma^{a}\partial_{a}$, in the presence of a
mass term
$\not{D}(m)=\gamma^{a}\partial_{a}+m,\quad a=1,2,3,\ldots,N.$ (5)
The frame $\\{\boldsymbol{e}_{a}\\}$ can be faithfully represented by the
Dirac matrices $\\{\gamma_{a}\\}$ that generate the Clifford algebra over $M$
$\gamma_{a}\gamma_{b}+\gamma_{b}\gamma_{a}=2\,\boldsymbol{g}(\boldsymbol{e}_{a},\boldsymbol{e}_{b})\mathbb{1}.$
(6)
In Euclidean signature, the Dirac matrices defined above are Hermitian,
denoted by ${\gamma_{a}^{\dagger}=\gamma_{a}}$, and the conjugate spinor
$\bar{\Psi}$ is simply the Hermitian conjugate of $\Psi$, written as
${\bar{\Psi}=\Psi^{\dagger}}$. Since the dimension of the spinor space in $N$
dimensions is $[N/2]$ (the floor of the number $N/2$), $\\{\gamma_{a}\\}$ and
$\mathbb{1}$ stand for $2^{[D/2]}\times 2^{[D/2]}$ matrices. In four
dimensions ($N=4$), for instance, they are $4\times 4$ matrices. The spectral
theory of general first-order differential operator of Dirac type can be found
in Refs. Branson1992A ; Branson1992B .
Our goal is to solve the integral (1). To accomplish this, we can expand the
spinor fields $\Psi$ and $\bar{\Psi}$ in terms of four-component complete
orthonormal sets of Dirac spinors $\psi_{j}$:
$\displaystyle\Psi(x)=\sum_{j}\Psi_{j}\psi_{j}(x),$ (7)
$\displaystyle\bar{\Psi}(x)=\sum_{j}\bar{\Psi}_{j}\psi^{\dagger}_{j}(x),$ (8)
The coefficients $\Psi_{j}$ and $\bar{\Psi}_{j}$ are independent Grassmannian
variables, and the index $j$ labels the field modes. The spinors $\psi_{j}$
are eigenfunctions of $\not{D}$ with eigenvalues determined by the equation
$\not{D}\psi_{j}=i\lambda_{j}\psi_{j},\quad\forall~{}\lambda_{j}\in\mathbb{R},$
(9)
and satisfy the following orthonormality and completeness relations
$\displaystyle\int
d^{N}x\,\sqrt{g}\,\psi^{\dagger}_{j}(x)\psi_{k}(x)=\delta_{jk},$ (10)
$\displaystyle\sum_{j}\psi_{j}(x)\psi^{\dagger}_{j}(x^{\prime})=\delta(x-x^{\prime})\,\mathbb{1},$
(11)
where $\delta(x-x^{\prime})$ is Dirac delta-function in the Euclidean
coordinates $\\{x,x^{\prime}\\}$. Taking into account the orthonormality
property (10) and the field expansions (7) and (8), the action (4) can be put
into the diagonal form
$\displaystyle\mathcal{S}_{E}(\Psi,\bar{\Psi})$
$\displaystyle=\sum_{j}\lambda_{j}(m)\bar{\Psi}_{j}{\Psi}_{j},$ (12)
where $\lambda(m)$ is given by:
$\displaystyle\lambda_{j}(m)=i\lambda_{j}+m.$ (13)
Now, under the decompositions (7) and (8), the anti-periodic functional
integral over the fields can be written in terms of $\Psi_{j}$ and
$\bar{\Psi}_{j}$ as
$\displaystyle\int_{\text{anti-
periodic}}\mathcal{D}\bar{\Psi}\mathcal{D}\Psi=\int\prod_{j}\frac{1}{\mu}\,d\bar{\Psi}_{j}d\Psi_{j},$
(14)
in which an arbitrary scale parameter $\mu$ has been introduced. An
interesting discussion on the meaning of $\mu$ can be consulted in
Elizalde1990 . By using the fact that the integration rules for Grassmannian
degrees of freedom are
$\int d\Psi_{j}=0\quad\text{and}\quad\int\Psi_{j}d\Psi_{j}=0,$ (15)
which must be equally satisfied by $\bar{\Psi}_{j}$, we are eventually led to
the following result
$\displaystyle{\,}\int
d\bar{\Psi}_{j}d\Psi_{j}e^{-\lambda_{j}(m)\bar{\Psi}_{j}\Psi_{j}}$
$\displaystyle=$ $\displaystyle\int
d\bar{\Psi}_{j}d\Psi_{j}\left[1-\lambda_{j}(m)\right]\bar{\Psi}_{j}\Psi_{j}=\lambda_{j}(m).$
(16)
The exponential series’ quadratic and higher-order powers vanish identically
due to Grassmannian anticommutative properties. Assuming that Eqs. (7)-(II.1)
hold, the path integral (1) over the Grassmann-valued Dirac spinors $\Psi$ and
$\bar{\Psi}$ gives the one-loop functional determinant of the operator
$\not{D}(m)$ with a positive exponent, as follows
$\displaystyle Z$
$\displaystyle=\int\prod_{j}\frac{1}{\mu}\,d\bar{\Psi}_{j}d\Psi_{j}\,e^{-\sum_{j}\lambda_{j}(m)\bar{\Psi}_{j}\Psi_{j}}$
$\displaystyle=\prod_{j}\frac{\lambda_{j}(m)}{\mu}=\text{det}\left[\frac{\not{D}(m)}{\mu}\right].$
(17)
Note that the above functional determinant is divergent because of infinite
product over the eigenvalues. This divergence indicates a need for some
regularization procedure. In this paper, we will adopt a powerful and elegant
regularization technique that utilizes the so-called generalized zeta
function, the zeta function of an operator.
## III Generalized zeta function
Let $L$ be a positive-definite self-adjoint second-order elliptic differential
operator, i.e. the eigenvalues $\lambda_{j}$ of $L$ are real and non-negative.
The zeta function associated with the operator $L$ is defined as
$\zeta_{L}(z)=\sum_{j}\lambda_{j}^{-z},$ (18)
where the sum over $j$ means the sum over the spectrum of $L$. In $N$
dimensions, the serie (18) will converge for ${\text{Re}(z)>N/2}$ and can be
analytically continued for the other values of $z$ Seeley1967 . In particular,
it is regular at $z=0$.
Now, we can use the zeta function above to provide a regularized version of
the ill-defined product of all eigenvalues. Taking the exponential of the
derivative of the zeta function with respect to $z$, evaluated at $z=0$, the
zeta-function regularized determinant can be defined by the relation
$\text{ln~{}det}L=\sum_{j}\text{ln}\lambda_{j}:=-\zeta_{L}^{\prime}(0),$ (19)
where $\zeta_{L}^{\prime}(z)$ stands for the derivative of $\zeta_{L}(z)$ with
respect to $z$. The definition (19) is well defined because the zeta function
is regular at $z=0$, and encodes all divergences present in the sum
$\sum_{j}\text{ln}\lambda_{j}$.
Defined previously as a series over the eigenvalues of an operator, the zeta
function admits also an integral representation by making a Mellin transform,
that is
$\displaystyle\zeta_{L}(z)=\frac{1}{\Gamma(z)}\int_{0}^{\infty}d\tau\,\tau^{z-1}K_{L}(\tau),$
(20)
where $K_{L}(\tau)$ is a spectral function called global heat kernel, defined
as
$\displaystyle K_{L}(\tau)=\text{Tr}\left(e^{-\tau L}\right).$ (21)
with Tr standing for the trace operation. In the case of the operator
$\not{D}^{2}(m)$ which is a $2^{[N/2]}\times 2^{[N/2]}$ matrix in the spinor
indices, Tr should be understood with an extra factor $2^{[N/2]}$ included.
Besides that, being $\lambda_{j}$ the eigenvalues of the operator $L$, we can
rewrite ${\text{Eq.}~{}\eqref{heat kernel trace}}$ as
$\displaystyle K_{L}(\tau)=\sum_{j}e^{-\tau\lambda_{j}},$ (22)
which diverges for $\tau\rightarrow 0$. In general, the structure of the
divergences present in the zeta function can be accessed from the asymptotic
behavior of the heat kernel for small $\tau$. For $\tau\rightarrow 0$, the
heat kernel admits the following expansion
$\displaystyle
K_{L}(\tau)\sim\frac{1}{(4\pi\tau)^{N/2}}\sum_{p}c_{p}(L)\,\tau^{p},~{}p=0,\frac{1}{2},1,\frac{3}{2},\ldots,$
(23)
where $c_{p}:=c_{p}(L)$ are the heat kernel coefficients. To review many of
the basic properties of the heat kernel method in QFT, including some
historical remarks, we refer to Vassilevich2003 ; Vassilevich2011 ; Gilkey1995
.
It is now possible to obtain a link between the one-loop partition function
and the generalized zeta function. Using the cyclic property of the trace and
the fact that the hermitian chiral matrix $\gamma^{5}$ (denoted this way
independently of the dimension) anticommutes with all Dirac matrices
$\gamma^{a}$, we can show the important property
$\displaystyle\text{Tr ln}\left[\not{D}(m)\right]$ $\displaystyle=\text{Tr
ln}\left[\gamma^{5}\not{D}(m)\gamma^{5}\right]=\text{Tr
ln}\left[\not{D}^{\dagger}(m)\right]$
$\displaystyle=\frac{1}{2}\left\\{\text{Tr ln}\left[\not{D}(m)\right]+\text{Tr
ln}\left[\not{D}^{\dagger}(m)\right]\right\\}$
$\displaystyle=\frac{1}{2}\text{Tr ln}\left[\not{D}^{2}(m)\right],$ (24)
where $\not{D}^{2}(m)$ is the negative of the spinor Laplacian on $M$,
$\not{D}^{2}=\gamma^{a}\gamma^{b}\partial_{a}\partial_{b}$, in the presence of
the mass
$\displaystyle\not{D}^{2}(m)=-\not{D}^{2}+m^{2}.$ (25)
Note in particular that the spinors $\psi_{j}$ are eigenfunctions of
$\not{D}^{2}(m)$ with non-negative eigenvalues
$\displaystyle\not{D}^{2}(m)\psi_{j}=\left(\lambda_{j}^{2}+m^{2}\right)\psi_{j}.$
(26)
Employing the identity
$\text{det}L=e^{\text{Tr~{}ln}L},$ (27)
one can derive from Eq. (24) the important relation
$\text{ln}\,\text{det}\left[\not{D}(m)\right]=\frac{1}{2}\,\text{ln}\,\text{det}\left[\not{D}^{2}(m)\right],$
(28)
establishing the massive extension of the relation between the determinant of
the Dirac operator and the square root of the determinant of its associated
Laplace-type operator. From Eqs. (18), (19) and (28), the zeta-function
regularization allows us to write the one-loop partition function as follows
Bytsenko1992
$\text{ln}Z=-\frac{1}{2}\left[\zeta_{\not{D}^{2}(m)}^{\prime}(0)+\text{ln}(\mu^{2})\zeta_{\not{D}^{2}(m)}(0)\right],$
(29)
which has the same structure as the scalar case, up to a global sign
Hawking1977 ; Giulia2021 . This is expected since we are working with the zeta
function associated with operator $\not{D}^{2}(m)$, which is of Laplace type,
instead of $\not{D}(m)$.
With the expression (29), one can obtain the free energy $F$, defined as
Landau1980
$F=-\frac{1}{\beta}\,\text{ln}Z,$ (30)
which is needed for the derivation of the Casimir energy at finite
temperature. A thermodynamics quantities closely related to the free energy is
the entropy, defined as
$\displaystyle S$ $\displaystyle=-\frac{\partial F}{\partial T},$ (31)
which, as we will see later, satisfies the third law of thermodynamics (the
Nernst heat theorem).
Although the zeta-function method encodes all divergences present in the
functional determinant, the structure of these divergences, however, plays a
central role in the renormalization procedure. Let us now utilize this
mathematical machinery to discuss a generic case of the Casimir energy
associated with the spinor field in four dimensions. To achieve our purpose,
it is convenient to decompose the time dependence of the spinor field in the
Fourier basis, namely
$\psi_{j}(x)=e^{-i\omega_{n}t}\chi_{\ell}(\boldsymbol{r}),$ (32)
stemming from the fact that $\partial_{t}$ is an obvious Killing vector field
of our metric, where $\ell$ is a generic index denoting the spatial quantum
modes of the field. Imposing the anti-periodic condition in the imaginary time
$t$ on the spinor field,
$\displaystyle\psi_{j}(t,\boldsymbol{r})=-\psi_{j}(t+\beta,\boldsymbol{r}),$
(33)
one can prove that the allowed frequencies must have the form
$\displaystyle\omega_{n}=\frac{2\pi}{\beta}\left(n+\frac{1}{2}\right),\quad\forall~{}n\in\mathbb{Z}.$
(34)
The condition (33) corresponds to compacting the imaginary-time dimension $t$
into a circumference of length $\beta$. This amounts to considering spinor
fields defined over a four-dimensional space with topology of the type
$\mathbb{S}^{1}\times\mathbb{R}^{3}$, where periodicity represented by
$\mathbb{S}^{1}$ is oriented at the $t$-direction. In Refs. Kulikov1989 ;
Ahmadi2005 ; Joas2017 , spinor fields are worked out in several spaces whose
topology is formed from the direct products.
Because of the time decomposition (32), it is particularly useful to write the
operator $\not{D}^{2}(m)$ as
$\not{D}^{2}(m)=L_{1}+\not{\nabla}^{2}(m),$ (35)
where $L_{1}$ and $\not{\nabla}^{2}(m)$ are defined as follows
$\displaystyle
L_{1}=-\partial^{2}_{t}\quad\text{and}\quad\not{\nabla}^{2}(m)=-\partial^{i}\partial_{i}+m^{2}.$
(36)
$\not{\nabla}^{2}(m)$ is an elliptic, self-adjoint, second-order differential
spinor operator defined on the spatial part of $M$. The generalized zeta
function method associated with the scalar operators defined on spaces with
different conditions can be found in Refs. Hawking1977 ; Giulia2021 . The
trace of the operator $\not{D}^{2}(m)$ can also be split into temporal and
spatial parts through the trace property
$\displaystyle\text{Tr}\left[e^{-\tau\not{D}^{2}(m)}\right]=4\,\text{Tr}\left(e^{-\tau
L_{1}}\right)\text{Tr}\left[e^{-\tau\not{\nabla}^{2}(m)}\right],$ (37)
where the multiplicative factor $4$ is due to the spinor nature of
$\not{D}^{2}(m)$. The eigenvalues of $L_{1}=-\partial_{t}^{2}$ can be obtained
from Eq. (34), so we have that
$\displaystyle\text{Tr}\left(e^{-\tau
L_{1}}\right)=\sum_{n=-\infty}^{\infty}e^{-\tau\frac{4\pi^{2}}{\beta^{2}}\left(n+\frac{1}{2}\right)^{2}}.$
(38)
Defining the constant parameters ${a=4\pi^{2}\tau/\beta^{2}},b=n$ and $c=1/2$,
and using the Jacobi inversion identity Kirstein2010 ,
$\displaystyle\sum_{n=-\infty}^{\infty}e^{-a\left(b+c\right)^{2}}=\sqrt{\dfrac{\pi}{a}}\sum_{n=-\infty}^{\infty}e^{-\frac{\pi^{2}}{a}b^{2}-2\pi
ibc},$ (39)
we can rewrite Eq. (38) as follows:
$\displaystyle\text{Tr}\left(e^{-\tau
L_{1}}\right)=\dfrac{\beta}{\sqrt{4\pi\tau}}\left[1+2\sum_{n=1}^{\infty}\cos(\pi
n)e^{-\frac{\beta^{2}}{4\tau}n^{2}}\right],$ (40)
in which the first term inside the brackets represents the ${n=0}$ term in the
series. Summing up these results, one eventually obtains the integral
representation of the zeta function $\zeta_{\not{D}^{2}(m)}$ associated with
$\not{D}^{2}(m)$, which is a Laplace type operator defined in a flat space
with a metric of Euclidean signature and acts on a spinor field in thermal
equilibrium at finite temperature ${T=\beta^{-1}}$, satisfying anti-
periodicity conditions. It follows from (37), (40), and (20) that the zeta
function $\zeta_{\not{D}^{2}(m)}$ can be put in the form
$\displaystyle{~{}}\zeta_{\not{D}^{2}(m)}(z)=\frac{\beta}{\sqrt{4\pi}}\left[\frac{Z_{1}(z)}{\Gamma(z)}+Z_{2}(z,\beta)\right],$
(41)
with
$\displaystyle Z_{1}(z)$
$\displaystyle=\,\Gamma\left(z-1/2\right)\zeta_{\not{\nabla}^{2}(m)}\left(z-1/2\right),$
(42) $\displaystyle Z_{2}(z,\beta)$
$\displaystyle=\frac{2}{\Gamma(z)}\sum_{n=1}^{\infty}\cos(\pi n)\times$
$\displaystyle\times\int_{0}^{\infty}d\tau\,\tau^{z-\frac{3}{2}}e^{-\frac{\beta^{2}}{4\tau}n^{2}}K_{\not{\nabla}^{2}(m)}(\tau).$
(43)
where $\zeta_{\not{\nabla}^{2}(m)}$ and $K_{\not{\nabla}^{2}(m)}$ are the zeta
function and the global heat kernel associated with the spinor operator
$\not{\nabla}^{2}(m)$.
Once the zeta function is obtained, we should compute the vacuum-free energy,
Eq. (30), which may have divergent parts. In order to analyze such
divergences, it is convenient to perform the small-$\tau$ asymptotics
expansion of the heat kernel Bordag2000 :
$\displaystyle K_{\not{\nabla}^{2}(m)}(\tau)\sim\frac{e^{-\tau
m^{2}}}{(4\pi\tau)^{3/2}}\sum_{p=0,1/2,1,\ldots}c_{p}\,\tau^{p},$ (44)
where $c_{p}=c_{p}(\not{\nabla}^{2})$ are the heat kernel coefficients
associated with the massless operator $\not{\nabla}^{2}$. Now, using the
integral representation of $\zeta_{\not{\nabla}^{2}(m)}$, Eq. (20), we can use
this asymptotic behavior of $K_{\not{\nabla}^{2}(m)}$ to write the function
$Z_{1}(z)$ as
$\displaystyle Z_{1}(z)$
$\displaystyle=\int_{0}^{\infty}d\tau\,\tau^{z-3/2}K_{\not{\nabla}^{2}(m)}(\tau)$
$\displaystyle=\frac{1}{(4\pi)^{3/2}}\sum_{p=0,1/2,1,\ldots}\frac{c_{p}\,\Gamma\left(z+p-2\right)}{\left(m^{2}\right)^{z+p-2}},$
(45)
which has simple poles located at
$\displaystyle z+p-2=-\kappa,\quad\forall~{}\kappa\in\mathbb{N},$ (46)
since the gamma function diverges only at non-positive integers, with the
corresponding residues containing non-negative mass exponents
$\displaystyle\text{Res}\left(Z_{1}(z),-\kappa\right)=\frac{(-1)^{\kappa}\,c_{2-\kappa-z}\,m^{2\kappa}}{(4\pi)^{3/2}\kappa!}.$
(47)
As we are only interested in the limit $z\rightarrow 0$, the constraint (46)
translates into considering the series (III) up to order $p\leq 2$, to be
consistent with the poles at $\kappa=0,1,2$. In particular, this means that
the terms in the series with semi-integer $p$ have no poles, the divergent
contributions come from the dominant coefficients $c_{0},c_{1}$ and $c_{2}$,
with $c_{0}$ and $c_{1}$ multiplied by non-negative mass exponents. However,
these divergent contributions are canceled out by the pole in $\Gamma(z)$ in
the denominator of $\zeta_{\not{D}^{2}(m)}(z)$. Indeed, near $z=0$
$\displaystyle\frac{1}{\Gamma(z)}=z+\gamma_{E}z^{2}+\mathcal{O}(z^{3}),$ (48)
where $\gamma_{E}$ is the Euler constant. In particular, this implies that
$Z_{2}(0,\beta)=0$, since the remaining integral in Eq. (43) is finite at
$z=0$. Thus,
$\displaystyle{~{}}\zeta_{\not{D}^{2}(m)}(0)=\frac{\beta}{16\pi^{2}}\,c_{2}(m),$
(49)
where
$\displaystyle
c_{2}(m)=\sum_{\kappa=0}^{2}\frac{(-1)^{\kappa}}{\kappa!}c_{2-\kappa}m^{2\kappa}=\frac{m^{4}c_{0}}{2}-m^{2}c_{1}+c_{2}.$
(50)
In order to obtain the expression for $\zeta^{\prime}_{\not{D}^{2}(m)}(0)$, we
should note that while $Z_{2}(z,\beta)$ and its first derivative with respect
to $z$ are finite at $z=0$, $Z_{1}(z)$ has a pole at $z=0$ coming from the
pole of ${\zeta_{\not{\nabla}^{2}(m)}\left(z-1/2\right)}$ at this point, with
residue
$\displaystyle\text{Res}(\zeta_{\not{\nabla}^{2}(m)}\left(z-1/2\right),z=0)=-\frac{c_{2}(m)}{16\pi^{2}}.$
(51)
So, separating off this pole contribution and taking the derivative of
$\zeta_{\not{D}^{2}(m)}(z)$ with respect to $z$, after some algebra, leads to
the relation for the regularized (reg) free energy for the Dirac field as
follows
$\displaystyle
F(\beta,m,\mu)=-\frac{1}{2}\,\text{FP}[\zeta_{\not{\nabla}^{2}(m)}(-1/2)]$
$\displaystyle+\frac{c_{2}(m)}{16\pi^{2}}\left\\{\text{ln}\left(\mu^{2}\right)+2[1-\text{ln}(2)]\right\\}$
$\displaystyle+\frac{1}{\sqrt{4\pi}}\sum_{n=1}^{\infty}\cos(\pi
n)\int_{0}^{\infty}d\tau\,\tau^{-3/2}e^{-\frac{\beta^{2}}{4\tau}n^{2}}K_{\not{\nabla}^{2}(m)}(\tau),$
(52)
where $\text{FP}[\zeta_{\not{\nabla}^{2}(m)}(-1/2)]$ stands for the finite
part of $\zeta_{\not{\nabla}^{2}(m)}(-1/2)$. This result corresponds to the
massive spinor counterpart of the one obtained by Kirstein in Ref.
Kirstein2010 for the massless scalar field, in which the only nonvanishing
heat kernel coefficient is $c_{2}$. Finally, taking the limit
$\beta\rightarrow\infty$, we obtain the following expression for zero-
temperature free energy associated with the massive spinor field
$\displaystyle E(m,\mu)$
$\displaystyle=\underset{\beta\rightarrow\infty}{\text{lim}}F(\beta,m,\mu)$
$\displaystyle=-\frac{1}{2}\,\text{FP}[\zeta_{\not{\nabla}^{2}(m)}(-1/2)]+\frac{c_{2}(m)}{16\pi^{2}}\,\text{ln}(\tilde{\mu}^{2}),$
(53)
where the rescaled parameter $\tilde{\mu}=\mu e/2$ has been employed. The
$\beta$-dependent remaining term in Eq. (III) is the temperature correction to
the free energy given by
$\displaystyle\Delta F(\beta,m)$
$\displaystyle=\frac{1}{\sqrt{4\pi}}\sum_{n=1}^{\infty}\cos(\pi n)\times$
$\displaystyle\times\int_{0}^{\infty}d\tau\,\tau^{-3/2}e^{-\frac{\beta^{2}}{4\tau}n^{2}}K_{\not{\nabla}^{2}(m)}(\tau).$
(54)
At this stage, it is worth noting that there remains no singularity when
$z\rightarrow 0$. So, the spinor free energy is finite. However, when the heat
kernel coefficient $c_{2}(m)$ is nonvanishing, the zeta function
regularization prescription becomes ambiguous due to its natural dependence on
the arbitrary parameter $\mu$, which has been rescaled without loss of
generality. This scale freedom when $c_{2}(m)\neq 0$ is also responsible for
the so-called conformal anomaly Vassilevich2011 ; Vassilevich2003 . It is
worth mentioning that in the massless case ($m=0$), all information concerning
free energy ambiguity is contained in the $c_{2}$ coefficient so that such
ambiguity is present only if $c_{2}\neq 0$.
To ensure the uniqueness of the renormalization process, such ambiguity can be
removed by the subtraction of the contribution arising from the heat kernel
coefficients $c_{p}$ with $p\leq 2$. After performing this finite
renormalization, the remaining part can be expressed as the sum of the zero-
temperature Casimir energy $E_{\text{cas}}(m)$ plus the temperature correction
$\Delta F(\beta,m)$
$\displaystyle F(\beta,m)=E_{\text{cas}}(m)+\Delta F(\beta,m),$ (55)
where the Casimir energy at zero temperature $E_{\text{cas}}(m)$ is as follows
$\displaystyle
E_{\text{cas}}(m)=-\frac{1}{2}\,\text{FP}[\zeta_{\not{\nabla}^{2}(m)}(-1/2)].$
(56)
Note that as the Casimir energy exhibits a mass dependence of the type
$e^{-\tau m^{2}}$, serving as a convergence factor in the integral
representation of $\zeta_{\not{\nabla}^{2}(m)}(-1/2)$, it must vanish when the
mass tends to infinity. This is due to the fact that there can be no quantum
fluctuations at this limit.
In the same fashion, one can use the small-$\tau$ heat kernel expansion, which
can also be seen as a large-$m$ expansion, to fix the ambiguity problem
uniquely. Since the heat kernel coefficient $c_{2}(m)$ increases with non-
negative powers of the mass, one must require that $E(m,\mu)$ should be
renormalized to zero for large $m$ Bordag2000
$\displaystyle\underset{m\rightarrow\infty}{\text{lim}}E(m,\mu)\rightarrow 0,$
(57)
removing all the dependence on the scale factor $\mu$ of the Casimir energy.
It is worth pointing out that when $c_{2}(m)$ is identically null, i.e., when
the FP prescription is redundant, the finite renormalization is unnecessary
because the ambiguity is naturally removed, and hence the scale freedom is
broken.
So far, the zeta function method has been utilized to obtain a generic
expression for the vacuum free energy associated with a massive spinor field
defined on a four-dimensional flat space endowed with a Euclidean metric. In
particular, given the anti-periodic condition of spinor fields in imaginary
time, we have been able to find a constraint on the eigenvalues of $L_{1}$.
From now on, we shall consider a space with a circularly compactified
dimension that imposes an anti-periodic boundary condition along a vector on
the spinor field. By imposing this spatial condition, we will explicitly
obtain the restrictions that the eigenvalues of $\not{\nabla}^{2}(m)$ must
obey, hence evaluating the spinor vacuum free energy.
## IV Spinor field in a nontrivial compactified space
This section aims to find an analytical expression for the zero-temperature
Casimir energy and its corresponding temperature corrections induced by a
topological constraint simulating a boundary condition imposed on the spinor
field along a vector in plane. To accomplish this, we will adopt the heat
kernel approach to zeta-function regularization.
Consider the space $M=(\mathbb{R}^{4},\boldsymbol{g})$, where $\boldsymbol{g}$
is a positive-definite symmetric metric whose components are given by Eq. (2),
namely $\delta_{ab}$, so that the line element on $M$ takes the form
$\displaystyle ds^{2}=dt^{2}+dx^{2}+dy^{2}+dz^{2},$ (58)
where $\\{x^{a}\\}=\\{t,x,y,z\\}$ are cartesian coordinates. We recall that
the coordinate $t$ is compactified into a circumference length $\beta$ as
discussed in Sec. II, equivalent to equipping $M$ with the topology
$\mathbb{S}^{1}(\text{time})\times\mathbb{R}^{3}$.
Here we consider the space $\mathbb{R}^{3}$ with a circularly compactified
dimension, where the periodicity represented by the circle $\mathbb{S}^{1}$ is
oriented in the direction of a vector $\boldsymbol{L}\in\mathbb{R}^{2}$ given
by
$\displaystyle\boldsymbol{L}=a\,\boldsymbol{e}_{x}-b\,\boldsymbol{e}_{y},\quad~{}a,b\in\mathbb{R},$
(59)
referred to here as compact vector. $\boldsymbol{e}_{x}$ and
$\boldsymbol{e}_{y}$ denote the unit vectors along the directions $x$ and $y$,
respectively, and the parameters $a$ and $b$ constant displacements. In
particular, the compact dimension size is determined by the vector length
$\displaystyle L(a,b)=|\boldsymbol{L}|=\sqrt{a^{2}+b^{2}}.$ (60)
Although not along a coordinate axis as usual, the compactification in a
$\mathbb{S}^{1}$ topology along $\boldsymbol{L}$ is quite natural since $L$ is
a homogeneous function of degree $1$, that is ${L(na,nb)=n\,L(a,b)}$ for all
non-null integer $n$. Choosing a suitable frame field can recover the usual
$\mathbb{S}^{1}$ topology, as we will see later.
Along the compact vector, the spinor field is assumed to satisfy the following
boundary condition
$\displaystyle\psi_{j}(t,\boldsymbol{r})=-\psi_{j}(t,\boldsymbol{r}+\boldsymbol{L}),$
(61)
similar to the temporal anti-periodicity condition, Eq. (33). In fact, when
${b=0}$ and ${a\neq 0}$, the spinor field satisfies a spatial anti-periodicity
condition, ${\psi_{j}(t,x,y,z)=-\psi_{j}(t,x+a,y,z)}$, induced by the compact
subspace of the coordinate $x$. The condition (61) means that the spinor field
undergoes a sign change after traveling a distance $a$ in the $x$-direction
and $b$ in the $y$-direction and returns to its initial value after traveling
distances $2a$ and $2b$, namely
${\psi_{j}(t,\boldsymbol{r})=\psi_{j}(t,\boldsymbol{r}+2\boldsymbol{L})}$. In
particular, through a coordinate origin redefinition, without changing the
orientation of the axes, one can equally write the condition (61) as
${\psi_{j}(t,x+a,y,z)=-\psi_{j}(t,x,y+b,z)}$. In Ref. Xiang2011 , this latter
condition was investigated in the helix-like topology context and called the
anti-helix condition, with $a$ and $b$ labeling the circumference length and
pitch of the helix, respectively.
An ansatz for the massive spinor field in the geometry of the space $M$ was
given in Eq. (32), namely
${\psi_{j}(t,\boldsymbol{r})=e^{-i\omega_{n}t}\chi_{\ell}(\boldsymbol{r})}$,
with the spatial part $\chi_{\ell}(\boldsymbol{r})$ satisfying the eigenvalue
equation
$\displaystyle\not{\nabla}^{2}(m)\chi_{\ell}(\boldsymbol{r})=\left(\lambda_{\ell}^{2}+m^{2}\right)\chi_{\ell}(\boldsymbol{r}).$
(62)
The eigenfunctions $\chi_{\ell}$ of the above equation have the form
$\displaystyle\chi_{\ell}(\boldsymbol{r})=\mathcal{N}\,e^{i\,\boldsymbol{k}\cdot\boldsymbol{r}}\,u_{s}(\boldsymbol{k}),$
(63)
with $\mathcal{N}$ being a normalization constant and $u_{s}(\boldsymbol{k})$
being four-component spinors whose explicit form is unnecessary for our
purposes. There are four spinors for each choice of momentum $\boldsymbol{k}$,
two of which have positive energy and two with negative energy Xiang2011 .
We are interested in obtaining the finite temperature Casimir energy under the
influence of the boundary condition (61), which imposes the following non-
trivial relation for the momentum along the compact vector
$\displaystyle\boldsymbol{k}\cdot\boldsymbol{L}=k_{x}a-k_{y}b=2\pi\left(n+\frac{1}{2}\right),\quad\forall~{}n\in\mathbb{Z}.$
(64)
This means that the label $\ell$ in the spinor field (63) should be understood
as the set of quantum numbers ${\\{\ell\\}=\\{n,k_{y},k_{z},s\\}}$, since
$k_{x}$ can be eliminated employing Eq. (64). In particular, the sum over
$\ell$ becomes
$\displaystyle\sum_{\ell}\rightarrow\sum_{n}\int dk_{y}\int dk_{z}\sum_{s}.$
(65)
Thus, utilizing the identification mentioned above in the completeness
relation (11) for the spinor field $\chi_{\ell}$ obeying the boundary
condition (61), we are left with the normalization constant
$\displaystyle\mathcal{N}=\frac{1}{2\pi\sqrt{a}}.$ (66)
Given the spinor fields (63), one can determine the eigenvalues in Eq. (62),
allowing for the construction of the spinor heat kernel. Assuming that the
requirement (64) holds, the corresponding eigenvalues are found to be
$\displaystyle\lambda_{\ell}^{2}$
$\displaystyle=k_{x}^{2}+k_{y}^{2}+k_{z}^{2}$
$\displaystyle=\left[\frac{2\pi}{a}\left(n+\frac{1}{2}\right)+\frac{b}{a}\,k_{y}\right]^{2}+k_{y}^{2}+k_{z}^{2}.$
(67)
It is worth mentioning that with an appropriate choice of frame field, it is
possible to align the compactification on $\mathbb{S}^{1}$ along one of the
coordinate axes. In the momentum space, this can be achieved by defining for
instance
$\displaystyle k_{y}=\frac{a}{L}\left[k_{Y}+\frac{2\pi
b}{a\,L}\left(n+\frac{1}{2}\right)\right].$ (68)
This transformation leads to the eigenvalues (IV) to be written as
$\displaystyle\lambda_{\ell}^{2}=\left[\frac{2\pi}{L}\left(n+\frac{1}{2}\right)\right]^{2}+k_{Y}^{2}+k_{z}^{2}.$
(69)
These eigenvalues stem in particular from the spatial anti-periodic boundary
condition ${\psi_{j}(t,X,Y,z)=-\psi_{j}(t,X+L,Y,z)}$ induced by the usual
topology ${\mathbb{S}^{1}}(\text{space})\times\mathbb{R}^{2}$, whereby the
coordinate $X$ is compactified into a circumference length $L$. In fact, along
this compact dimension, the latter condition produces the discrete momentum
${k_{X}=2\pi(n+1/2)/L}$. In particular, this means that in the limiting case
${b=0}$, our results recover the ones presented in Ref. Bellucci2009 for a
specific case and include temperature corrections. In this study, the authors
investigated the Casimir effect for spinor fields in toroidally compactified
spaces, including general phases in the boundary condition along the compact
dimensions.
Building upon the previous results, we can introduce the heat kernel approach
to obtain a zeta-function analytical expression for a spinor field defined on
$M$ with the eigenvalues (IV). Instead of the global heat kernel, it is more
appropriate to utilize the local heat kernel. The reason is that the heat
kernel carries information concerning the space where the field is defined,
making it particularly valuable when focusing on the the influence of
topological constraint imposed by the boundary conditions on the thermal
vacuum fluctuations.
### IV.1 Spinor heat kernel and Casimir energy density
The spinor heat kernel
$K_{\not{\nabla}^{2}(m)}(\boldsymbol{r},\boldsymbol{r}^{\prime},\tau)$ is a
two-point function locally defined as solutions of the heat conduction
equation
$\left[\frac{\partial}{\partial\tau}+\not{\nabla}^{2}(m)\right]\,K_{\not{\nabla}^{2}(m)}(\boldsymbol{r},\boldsymbol{r}^{\prime},\tau)=0\quad\text{for}~{}\tau>0,$
(70)
supplemented with the initial condition
$\lim_{\tau\rightarrow
0}K_{\not{\nabla}^{2}(m)}(\boldsymbol{r},\boldsymbol{r}^{\prime},\tau)=\delta(\boldsymbol{r}-\boldsymbol{r}^{\prime})\,\mathbb{1}.$
(71)
The operator $\not{\nabla}^{2}(m)$ is taken to act on the first argument of
$K_{\not{\nabla}^{2}(m)}$. Similar to $\not{\nabla}^{2}(m)$,
$K_{\not{\nabla}^{2}(m)}$ is represented by a $4\times 4$ matrix.
Taking into account Eq. (26), the solutions of Eq. (70) can be expressed in
terms of the eigenvalues and eigenfunctions of $\not{\nabla}^{2}(m)$
$K_{\not{\nabla}^{2}(m)}(\boldsymbol{r},\boldsymbol{r}^{\prime},\tau)=e^{-\tau
m^{2}}\sum_{\ell}e^{-\lambda_{\ell}^{2}\tau}\chi_{\ell}(\boldsymbol{r})\chi_{\ell}^{\dagger}(\boldsymbol{r}^{\prime}).$
(72)
One can verify that the above expression provides a solution to Eq. (70), as
well as satisfying the initial condition (71) since the spinor field obeys Eq.
(11).
Inserting the spinor solution (63) along with the normalization constant (66)
and eigenvalues (IV) into spinor heat kernel (72), it follows the expression
$\displaystyle
K_{\not{\nabla}^{2}(m)}(\boldsymbol{r},\boldsymbol{r}^{\prime},\tau)$
$\displaystyle=\dfrac{e^{-\tau m^{2}}}{4\pi^{2}a}$
$\displaystyle\times\sum_{n}e^{-\frac{4\pi^{2}\tau}{a^{2}}\left(n+\frac{1}{2}\right)^{2}+\frac{2\pi
i}{a}\left(n+\frac{1}{2}\right)\Delta x}$ $\displaystyle\times\int
dk_{y}\,e^{-\tau\frac{L^{2}}{a^{2}}k_{y}^{2}+\left[i\Delta v-\frac{4\pi
b\tau}{a^{2}}\left(n+\frac{1}{2}\right)\right]k_{y}}$ $\displaystyle\times\int
dk_{z}e^{-\tau k_{z}^{2}+ik_{z}\Delta z}\,\mathbb{1},$ (73)
where
$\displaystyle\Delta v=\frac{b}{a}\,\Delta x+\Delta y.$ (74)
We can write Eq. (73) in a more compact form. To perform this, let us define
the complex parameters $w$ and $q$ as follows
$\displaystyle w=\frac{b\Delta v}{L^{2}}-\frac{\Delta
x}{a}-\frac{q}{2}\quad\text{and}\quad q=\frac{4\pi i\tau}{L^{2}},$ (75)
and introduce the following Jacobi function defined in terms of the parameters
$w$ and $q$ as Elizalde1994
$\displaystyle\theta_{3}(w,q)=\sum_{n=-\infty}^{\infty}e^{i\pi qn^{2}-2\pi
iwn}.$ (76)
Evaluating the integrals over the independent momenta $k_{y}$ and $k_{z}$ in
Eq. (73), we end up with the following relation between the spinor heat kernel
and the Jacobi function
$\displaystyle
K_{\not{\nabla}^{2}(m)}(\boldsymbol{r},\boldsymbol{r}^{\prime},\tau)=\frac{e^{-\frac{\left|\boldsymbol{r}-\boldsymbol{r}^{\prime}\right|^{2}}{4\tau}-\tau
m^{2}}}{\left(4\pi\tau\right)^{3/2}}\frac{\sqrt{-iq}}{e^{-i\pi\frac{\omega^{2}}{q}}}\,\theta_{3}(w,p)\,\mathbb{1}.$
(77)
Since we are interested in the contributions coming from the topology for the
thermal vacuum fluctuations, it is convenient to separate the Euclidean part
of the heat kernel, which should not depend on the topology parameters. This
can be done by rewriting the $\theta_{3}$ Jacobi function utilizing the
following identity
$\displaystyle\theta_{3}(w,q)=\frac{1}{\sqrt{-iq}}\,e^{-i\pi\frac{\omega^{2}}{q}}\,\theta_{3}\left(\frac{w}{p},-\frac{1}{q}\right).$
(78)
Employing this identity, leads to the following expression
$\displaystyle
K_{\not{\nabla}^{2}(m)}(\boldsymbol{r},\boldsymbol{r}^{\prime},\tau)=K_{\not{\text{E}}}(\boldsymbol{r},\boldsymbol{r}^{\prime},\tau)\sum_{n=-\infty}^{\infty}e^{-\frac{L^{2}}{4\tau}n^{2}+i\pi
n},$ (79)
with
$\displaystyle
K_{\not{\text{E}}}(\boldsymbol{r},\boldsymbol{r}^{\prime},\tau)=\frac{1}{\left(4\pi\tau\right)^{3/2}}\,e^{-\frac{\left|\boldsymbol{r}-\boldsymbol{r}^{\prime}\right|^{2}}{4\tau}-\tau
m^{2}}\,\mathbb{1}.$ (80)
where $K_{\not{\text{E}}}$ is the spinor version of the well-known Euclidean
heat kernel associated with the massive scalar Laplacian operator defined on
the flat space $\mathbb{R}^{3}$ Vassilevich2003 . Note that
$K_{\not{\text{E}}}$ is identified with the term $n=0$ in the series.
Let us now explore the heat kernel properties at the coincidence limit
$\boldsymbol{r}^{\prime}\to\boldsymbol{r}$ in Eq. (79) which results in
$\displaystyle K_{\not{\nabla}^{2}(m)}(\boldsymbol{r},\boldsymbol{r},\tau)$
$\displaystyle=K_{\not{\text{E}}}(\boldsymbol{r},\boldsymbol{r},\tau)$
$\displaystyle+\frac{e^{-\tau
m^{2}}}{(4\pi\tau)^{3/2}}\sum_{n=1}^{\infty}2\cos(\pi
n)e^{-\frac{L^{2}}{4\tau}n^{2}}\,\mathbb{1}.$ (81)
The first term on the right-hand side corresponds to the Euclidean heat
kernel, while the second term encodes information about the space topology, as
can be seen from its dependence on parameter $L$. For small $\tau$, the heat
kernel admits an expansion in powers of $\tau$, with coefficients reflecting
the space configuration. In our case, by evaluating the above series at small
$\tau$, one can see that all terms are exponentially small except for the one
associated with the Euclidean heat kernel contribution (${n=0}$). Thus, the
spinor heat kernel $K_{\not{\nabla}^{2}(m)}(\boldsymbol{r})$ on Euclidean
geometry with a circular compactification along $\boldsymbol{L}$ exhibits an
asymptotic behavior similar to the one considered in Eq. (44) with only one
non-vanishing heat kernel coefficient
$\displaystyle
K_{\not{\nabla}^{2}(m)}(\boldsymbol{r},\boldsymbol{r},\tau)\sim\frac{e^{-\tau
m^{2}}}{\left(4\pi\tau\right)^{3/2}}\sum_{p}c_{p}(\boldsymbol{r})\tau^{p}+\mathcal{O}(e^{-1/\tau}),$
(82)
where the local heat kernel coefficients $c_{p}(\boldsymbol{r})$ are given by
$\displaystyle c_{p}(\boldsymbol{r})=\delta_{0p}\,\mathbb{1}\quad\forall~{}p.$
(83)
$\mathcal{O}(e^{-1/\tau})$ stands for those terms going to zero faster than
any positive power of $\tau$ and, therefore, can be neglected. In contrast
with the global case, the local heat kernel coefficients carry spinor indices,
hence $4\times 4$ matrices. Note that $K_{\not{\nabla}^{2}(m)}(\tau)$ can be
obtained from
$K_{\not{\nabla}^{2}(m)}(\boldsymbol{r},\boldsymbol{r}^{\prime},\tau)$
performing an integral in the whole space
$\displaystyle K_{\not{\nabla}^{2}(m)}(\tau)=\int
d^{3}\boldsymbol{r}\,\sqrt{g^{(3)}}\,\text{tr}\left[K_{\not{\nabla}^{2}(m)}(\boldsymbol{r},\boldsymbol{r},\tau)\right],$
(84)
where $g^{(3)}$ is the spatial part of the metric determinant, and the trace
operation tr is taken over the spinor indices only. Thus, the global heat
kernel coefficient $c_{0}$ is found to be
$\displaystyle c_{0}$ $\displaystyle=4\,V_{3},$ (85)
where $V_{3}$ is the volume of the $3$-dimensional base space of $M$. As
discussed in Sec. III, for nonvanishing heat kernel coefficients
$c_{p}\,(p\leq 2)$, the zeta function is not finite, and the renormalization
procedure is not unique. In fact, although the vacuum energy is finite due to
the FP prescription introduced in the Casimir energy, the coefficient $c_{0}$
gives origin to the terms in the vacuum energy, which increase with non-
negative powers of the mass, besides the logarithmic dependence on the scale
factor $\mu$. To ensure a unique renormalization procedure and obtain an
unambiguous spinor vacuum free energy, all contributions associated with
$c_{0}$ should be disregarded, thereby renormalizing the energy to zero for
large masses.
After performing the finite renormalization, we can proceed with the
analytical calculation of the spinor vacuum free energy. First, we should note
that even though we are working in the local regime, the two-point function
$K_{\not{\nabla}^{2}(m)}(\boldsymbol{r},\boldsymbol{r},\tau)$ is coordinate-
independent. Therefore, considering that global quantities can be derived from
local ones by integrating over the space coordinates, the local version of the
spinor vacuum free energy differs from the global version by a volume element
and retains the same form as Eq. (55), namely
$\displaystyle\mathcal{F}(\beta,m)=\mathcal{E}_{\text{cas}}(m)+\Delta\mathcal{F}(\beta,m),$
(86)
where $\mathcal{E}_{\text{cas}}(m)$ is then the Casimir energy density at zero
temperature
$\displaystyle\mathcal{E}_{\text{cas}}(m)=-\frac{1}{2}\,\text{FP}[\zeta_{\not{\nabla}^{2}(m)}(\boldsymbol{r},-1/2)],$
(87)
and $\Delta\mathcal{F}(\beta,m)$ is the temperature correction with the form
$\displaystyle\Delta$
$\displaystyle\mathcal{F}(\beta,m)=\frac{1}{\sqrt{4\pi}}\sum_{n=1}^{\infty}\cos(\pi
n)\times$
$\displaystyle\times\int_{0}^{\infty}d\tau\,\tau^{-3/2}e^{-\frac{\beta^{2}}{4\tau}n^{2}}\text{tr}\left[K_{\not{\nabla}^{2}(m)}(\boldsymbol{r},\boldsymbol{r},\tau)\right].$
(88)
Here, the trace operation tr is taken over the spinor indices only and the
local zeta function is defined in terms of
$K_{\not{\nabla}^{2}(m)}(\boldsymbol{r},\boldsymbol{r},\tau)$ giving rise to
$\displaystyle\zeta_{\not{\nabla}^{2}(m)}$
$\displaystyle(\boldsymbol{r},z-1/2)=\frac{1}{\Gamma(z-1/2)}$
$\displaystyle\times\int_{0}^{\infty}d\tau\,\tau^{z-3/2}\,\text{tr}\left[K_{\not{\nabla}^{2}(m)}(\boldsymbol{r},\boldsymbol{r},\tau)\right].$
(89)
Then, inserting the spinor heat kernel (IV.1) into Eq. (IV.1), we conclude
from Eq. (87) that the renormalized expression for the zero temperature
Casimir energy density associated with a spinor field of mass $m$ depends on
the topology parameter $L$ according to the relation
$\displaystyle\mathcal{E}_{\text{cas}}(m,L)=\frac{2}{\pi^{2}L^{4}}\sum_{n=1}^{\infty}\frac{\cos(\pi
n)}{n^{2}}(mL)^{2}\,K_{2}(nmL),$ (90)
where $K_{2}(z)$ is the MacDonald function. Note that the FP prescription
removed the divergent contribution provided by the Euclidean heat kernel. The
above result is exactly the one shown in Ref. Xiang2011 obtained in a
different approach than the one presented here for both massive and massless
spinors. In particular, the massless one can be obtained by making use of the
following limit
$\displaystyle\underset{z\rightarrow
0}{\text{lim}}\,z^{2}K_{2}(nz)=\frac{2}{n^{2}}.$ (91)
In fact, by separating the even and odd terms in the series (90), and using
the above equation, one can promptly verify that the following massless limit
holds
$\displaystyle\mathcal{E}_{\text{cas}}(L)$
$\displaystyle=\underset{m\rightarrow
0}{\text{lim}}\mathcal{E}_{\text{cas}}(m,L)$
$\displaystyle=\frac{1}{4\pi^{2}L^{4}}\left[\zeta(4)-\zeta\left(4,\frac{1}{2}\right)\right],$
(92)
where $\zeta(z)$ is the standard Riemann zeta function and $\zeta(z,w)$ is the
Hurwitz zeta function defined for $\text{Re}(z)>1$ and $w\neq 0,-1,-2,\ldots$,
in the form Elizalde1994
$\displaystyle\zeta(z,w)=\sum_{n=0}^{\infty}(n+w)^{-z}.$ (93)
Using the relation
$\displaystyle\zeta\left(z,\frac{1}{2}\right)=(2^{z}-1)\,\zeta(z),$ (94)
along with the fact that ${\zeta(4)=\pi^{4}/90}$, we are left with the
expression for the Casimir energy density, at zero temperature, associated
with a massless spinor field
$\displaystyle\mathcal{E}_{\text{cas}}(L)=-\frac{7}{2}\,\frac{\pi^{2}}{90\,L^{4}},$
(95)
It depends only on topology parameter, in complete agreement with the massless
case obtained in Ref. Xiang2011 . In particular, its value is also $7/2$ times
the result found in the massless scalar case under periodic boundary
conditions along the compact vector $\boldsymbol{L}$ Zhai ; Giulia2021 . It is
worth mentioning that this case is unambiguous since the heat kernel
coefficient $c_{2}$ is identically zero, so the renormalization procedure is
unnecessary.
If one is interested in the limit $mL\gg 1$, then it is legitimate to consider
Mcdonald’s function behavior at ${\text{large}~{}z}$
$\displaystyle
K_{2}(z)\simeq\left(\frac{\pi}{2z}\right)^{1/2}e^{-z}\quad\text{for}\quad|\text{arg}(z)|<\pi/2.$
(96)
In this limiting case, one can see that the Casimir energy density decays
exponentially with the mass of the field
$\displaystyle\mathcal{E}_{\text{cas}}\left(m\gg\frac{1}{L}\right)=-\frac{2m^{2}}{\pi^{2}L^{2}}\sqrt{\frac{\pi}{2mL}}\,e^{-mL},$
(97)
as expected, since an infinitely heavy field should not present quantum
fluctuations and hence should not produce Casimir energy Bordag2000 . In Ref.
Maluf2020 , a similar analysis is carried out for the Casimir energy for a
real scalar field and the Elko neutral spinor field in a field theory at a
Lifshitz fixed point.
### IV.2 Finite-temperature corrections
Let us now investigate the temperature correction,
$\Delta\mathcal{F}(\beta,m)$, to the vacuum energy densities. Inserting the
heat kernel (IV.1) into $\Delta\mathcal{F}(\beta,m)$ defined in Eq. (IV.1)
leads to the following analytical expression for the temperature correction
associated with the massive spinor field, in terms of a double sum
$\displaystyle\Delta\mathcal{F}(\beta,m,L)$
$\displaystyle=\Delta\mathcal{F}_{\text{E}}(\beta,m)$
$\displaystyle+\frac{4m^{4}}{\pi^{2}}\sum_{n=1}^{\infty}\sum_{p=1}^{\infty}\cos(\pi
n)\cos(\pi p)$ $\displaystyle\times
f_{2}\left(m\beta\sqrt{p^{2}+\frac{L^{2}}{\beta^{2}}n^{2}}\right).$ (98)
For notational simplicity, we introduced the function $f_{\nu}(z)$ related to
the Mcdonald function $K_{\nu}(z)$ as follows
$\displaystyle f_{\nu}(z)=\frac{K_{\nu}(z)}{z^{\nu}}.$ (99)
The term $\Delta\mathcal{F}_{\text{E}}(\beta,m)$ is the contribution coming
from the Euclidean heat kernel and thus does not depend on the parameter $L$.
It has the following form
$\displaystyle\Delta\mathcal{F}_{\text{E}}(\beta,m)=\frac{2m^{4}}{\pi^{2}}\sum_{n=1}^{\infty}\cos(\pi
n)\,f_{2}(m\beta n).$ (100)
In particular, from Eq. (91), we conclude that the temperature correction term
in the massless limit is as follows
$\displaystyle\Delta\mathcal{F}_{\text{E}}\left(\beta\right)=-\frac{7}{2}\frac{\pi^{2}}{90}\,T^{4},$
(101)
the standard black body radiation energy density associated with the massless
spinor field. As we have seen, this contribution is directly related to the
non-null coefficient $c_{0}$. In more general spaces with nontrivial topology,
however, there may be temperature corrections to the above Stefan-Boltzmann
law, proportional to $T^{4}$, coming from heat kernel coefficients associated
with spacetime topology. These coefficients vanish in the limit of infinite
space Herondy2023 ; Basil1978 .
Since the Casimir effect is a purely quantum phenomenon, the above term should
not dominate in the high-temperature limit. Although not divergent, this
quantum term should be subtracted in the renormalization procedure to obtain a
correct classical contribution in this limit. By doing so, we end up with the
renormalized version of the free energy (86)
$\displaystyle\mathcal{F}_{\text{ren}}(\beta,m,L)=\frac{2m^{4}}{\pi^{2}}\sum_{n=1}^{\infty}\cos(\pi
n)f_{2}\left(nmL\right)$
$\displaystyle+\frac{4m^{4}}{\pi^{2}}\sum_{n=1}^{\infty}\sum_{p=1}^{\infty}\cos(\pi
n)\cos(\pi
p)f_{2}\left(m\beta\sqrt{p^{2}+\frac{L^{2}}{\beta^{2}}n^{2}}\right).$ (102)
In particular, by using the limit (91), the above expression yields the
massless contribution
$\displaystyle\mathcal{F}_{\text{ren}}(\beta,L)=\mathcal{E}_{\text{cas}}(L)+\frac{8}{\pi^{2}\beta^{4}}\sum_{n=1}^{\infty}\sum_{p=1}^{\infty}\frac{\cos(\pi
n)\cos(\pi p)}{\left(p^{2}+\frac{L^{2}}{\beta^{2}}n^{2}\right)^{2}},$ (103)
where $\mathcal{E}_{\text{cas}}(L)$ is the Casimir energy density associated
with massless spinor field at zero temperature, Eq. (95). The presence of the
double sum is convenient if one is interested in obtaining the low- and high-
temperature asymptotic limits. Although the final result is the same,
performing the sum in $p$ first is more straightforward for obtaining the
high-temperature limit, while performing the sum in $n$ first is less
complicated for obtaining the low-temperature limit. The final result for the
free energy is equivalent. Choosing which one first is a simple question of
convenience to attain our purposes.
Figure 1: Plot of the ratio
$R=\mathcal{F}_{\text{ren}}/\mathcal{E}_{\text{cas}}$ in terms of $a\,T$ for
several values of the parameter $\gamma$. $R$ decreases with $aT$ and tends to
zero when $aT$ goes to infinity.
To conduct our analysis, let us rewrite $L$ as ${L=a\sqrt{\gamma}}$ by
convenience, where $\gamma=1+(b/a)^{2}$. In Fig. 1, we have plotted the ratio
$R$ of the renormalized free energy density,
$\mathcal{F}_{\text{ren}}(\beta,L)$, to the Casimir energy density,
$\mathcal{E}_{\text{cas}}(L)$, varying with $aT$ for different values of the
parameter $\gamma$.
In each case, the plot shows the ratio
$R=\mathcal{F}_{\text{ren}}/\mathcal{E}_{\text{cas}}$ going to $1$ as $T$
approaches zero, as we should expect, and decaying to zero as $T$ approaches
infinity. In particular, this decay becomes more pronounced as the parameter
$\gamma$ increases, as illustrated by the curve for $\gamma=9$. The curves
associated with $\gamma>2$, which decay to zero faster, correspond to the case
where $b$ is greater than $a$, whereas the curve with $\gamma=2$ illustrates
the particular case $b=a$. In the limiting case when $\gamma\simeq 1$ ($b\ll
a$), the system exhibits a structure known as a quantum spring, as discussed
by Feng2010 in the context of the scalar Casimir effect.
In what follows, we will analyze the asymptotic limits of temperature
corrections to the massless free energy density above.
#### IV.2.1 High-temperature limit
Let us analyze the high-temperature limit, ${\beta\ll L}$, of the final
expression (103). In this case, it is more appropriate to perform the
summation in $p$ first, namely
$\displaystyle\frac{1}{\beta^{4}}\sum_{p=1}^{\infty}$
$\displaystyle\frac{\cos(\pi
p)}{\left(p^{2}+\frac{L^{2}}{\beta^{2}}n^{2}\right)^{2}}=-\frac{1}{2\,L^{4}}\frac{1}{n^{4}}$
$\displaystyle+\frac{T^{2}}{2L^{2}}\frac{1}{n^{2}}\,\text{csch}(n\pi L\,T)$
$\displaystyle\times\left[\coth(n\pi L\,T)+\frac{1}{n\pi L\,T}\right].$ (104)
Inserting this summation into Eq. (103), we are eventually led to the
following expression
$\displaystyle\mathcal{F}_{\text{ren}}\left(\beta,L\right)$
$\displaystyle=\frac{4\,T^{2}}{L^{2}}\sum_{n=1}^{\infty}\frac{\cos(\pi
n)}{n^{2}}\text{csch}(n\pi L\,T)$ $\displaystyle\times\left[\coth(n\pi
L\,T)+\frac{1}{n\pi L\,T}\right].$ (105)
Note that the first term on the right-hand side in Eq. (IV.2.1), after
performing the summation in $n$, gives rise to Casimir energy density
associated with massless spinor field (95) but with the opposite sign.
Therefore, the effect of this latter term is entirely compensated by the
corresponding one in the free energy density (103). Such a natural
cancellation between the massless Casimir energy density and its corresponding
temperature correction is not unusual. It is an intrinsic characteristic of
temperature corrections at the high-temperature limit Plunien1986 .
Now, our task is to evaluate the series (IV.2.1) at a high-temperature limit,
${\beta\ll L}$ (or equivalently ${L\,T\gg 1}$). Through an asymptotic
expansion up to terms of order $\mathcal{O}(e^{-\pi L\,T})$, we arrive at the
finite-temperature
$\displaystyle\ \mathcal{F}_{\text{ren}}\left(\beta\ll
L\right)\simeq-\frac{8\,T}{L^{3}}\left(1+\pi L\,T\right)\,e^{-\pi L\,T},$
(106)
which is exponentially suppressed at high $T$ and converges to zero as
${T\rightarrow\infty}$, in accordance with Fig. 1. This behavior is expected
for a spinor field since, differently from the scalar field Giulia2021 , it
lacks a temperature correction term that is linearly dependent on $T$. We
emphasize the need for the free energy density to undergo a finite
renormalization by subtracting from it the blackbody radiation contribution,
proportional to $T^{4}$, to obtain the correct classical limit, a free energy
density renormalized to zero at very high temperatures. Ref. Mostepanenko2011
found a similar result for temperature corrections associated with the spinor
field in the closed Friedmann cosmological model.
With the renormalized free energy density now available, we can obtain an
analytical expression for the renormalized entropy density. Employing the
relation (31), we have
$\displaystyle\mathcal{S}_{\text{ren}}(\beta,L)$
$\displaystyle=\frac{4\pi\,T^{2}}{L^{2}}\sum_{n=1}^{\infty}\frac{\cos(\pi
n)}{n}\text{csch}(n\pi L\,T)$ $\displaystyle\times\left[1-\frac{1}{(n\pi
L\,T)^{2}}+2\,\text{csch}^{2}(n\pi L\,T)\right.$
$\displaystyle-\left.\frac{1}{n\pi L\,T}\coth(n\pi L\,T)\right].$ (107)
The corresponding asymptotic expansion in the high-temperature regime,
${L\,T\gg 1}$, decays exponentially with the temperature $T$
$\displaystyle\mathcal{S}_{\text{ren}}\left(\beta\ll
L\right)\simeq-\frac{8}{\pi L^{3}}\left[1+\pi L\,T-(\pi
L\,T)^{2}\right]\,e^{-\pi L\,T}.$ (108)
Note that the lack of a classical term proportional to $T$ in the free energy
density results in the Casimir entropy density approaching zero at very high
temperatures, which differs from the scalar case where it is dominated by a
constant term Giulia2021 .
#### IV.2.2 Low-temperature limit
Let us now consider the asymptotic expansion of the expression (103) in the
low-temperature regime, where ${\beta\gg L}$, or equivalently ${L\,T\ll 1}$.
To accomplish this, as previously mentioned, we shall perform the summation in
$n$ first, providing
$\displaystyle\frac{1}{\beta^{4}}\sum_{n=1}^{\infty}$
$\displaystyle\frac{\cos(\pi
p)}{\left(p^{2}+\frac{L^{2}}{\beta^{2}}n^{2}\right)^{2}}=-\frac{T^{4}}{2\,\pi^{2}}\frac{1}{p^{4}}$
$\displaystyle+\frac{4\,T^{2}}{L^{2}}\frac{1}{p^{2}}\,\text{csch}\left(\frac{p\pi}{L\,T}\right)$
$\displaystyle\times\left[\frac{L\,T}{p\pi}+\coth\left(\frac{p\pi}{L\,T}\right)\right].$
(109)
Substituting it back in Eq. (103), we get
$\displaystyle\mathcal{F}_{\text{ren}}\left(\beta,L\right)$
$\displaystyle\simeq-\frac{7}{2}\,\frac{\pi^{2}}{90\,L^{4}}+\frac{7}{2}\frac{\pi^{2}}{90}\,T^{4}$
$\displaystyle+\frac{4\,T^{2}}{L^{2}}\sum_{p=1}^{\infty}\frac{\cos(\pi
p)}{p^{2}}\,\text{csch}\left(\frac{p\pi}{L\,T}\right)$
$\displaystyle\times\left[\frac{L\,T}{p\pi}+\coth\left(\frac{p\pi}{L\,T}\right)\right],$
(110)
which at the low-temperature regime up to terms of the order
$\mathcal{O}(e^{-\frac{\pi}{L\,T}})$, presents the following free energy
density asymptotic behavior
$\displaystyle\mathcal{F}_{\text{ren}}\left(\beta\gg L\right)$
$\displaystyle\simeq-\frac{7}{2}\,\frac{\pi^{2}}{90\,L^{4}}+\frac{7}{2}\frac{\pi^{2}}{90}\,T^{4}$
$\displaystyle-\frac{8\,T^{2}}{L^{2}}\left(1+\frac{L\,T}{\pi}\right)\,e^{-\frac{\pi}{L\,T}}.$
(111)
Let us make a few remarks comparing our findings and those reported in
Giulia2021 for the low-temperature behavior of the free energy associated
with the scalar field under helix topology. Apart from the additional $T^{3}$
term for temperature correction observed in the scalar case, they differ by
constant multiplicative factors that naturally arise because of the spinor
degrees of freedom. Note that at small $T$, the above asymptotic expansion is
dominated by the first term, the massless Casimir energy density at zero
temperature, Eq. (95), as expected Mostepanenko2011 ; Herondy2023 ; Basil1978
.
The entropy density can be obtained by inserting Eq. (IV.2.2) into Eq. (31),
providing
$\displaystyle\mathcal{S}_{\text{ren}}(\beta,L)$
$\displaystyle=-\frac{7\pi^{2}}{45}\,T^{3}$
$\displaystyle-\frac{4\pi}{L^{3}}\sum_{p=1}^{\infty}\frac{\cos(\pi
p)}{p}\,\text{csch}\left(\frac{p\pi}{L\,T}\right)$
$\displaystyle\times\left[1+3\left(\frac{L\,T}{p\pi}\right)^{2}+2\,\text{csch}^{2}\left(\frac{p\pi}{L\,T}\right)\right.$
$\displaystyle-\left.\frac{3L\,T}{p\pi}\coth\left(\frac{p\pi}{L\,T}\right)\right].$
(112)
Its corresponding asymptotic expansion in the low-temperature limit, where
${L\,T\ll 1}$, is found to be
$\displaystyle\mathcal{S}_{\text{ren}}\left(\beta\gg L\right)$
$\displaystyle\simeq\frac{7\pi^{2}}{45}\,T^{3}$
$\displaystyle+\frac{8\pi}{L^{3}}\left[1+\frac{3L\,T}{\pi}+3\left(\frac{L\,T}{\pi}\right)^{2}\right]\,e^{-\frac{\pi}{L\,T}}.$
(113)
As expected, the above expression tends to zero as the temperature approaches
zero. It implies that the entropy density for a massless spinor field
satisfying an anti-periodic condition along the compact vector satisfies the
third law of thermodynamics (the Nernst heat theorem) Landau1980 .
## V Conclusion
In the present work, we have investigated the thermal Casimir effect
associated with a massive spinor field defined on a four-dimensional flat
space with a circularly compactified dimension. The periodicity represented by
$\mathbb{S}^{1}$ is oriented not along a coordinate axis as usual, but along a
vector $\boldsymbol{L}$ belonging to the $xy$-plane, Eq. (59). This geometry
introduces a topological constraint inducing a spatial anti-periodic boundary
condition on the spinor field, Eq. (61), which modifies the vacuum
fluctuations, producing the Casimir effect. Imposing this boundary condition
led to the discrete eigenvalues for the momentum along vector
$\boldsymbol{L}$, Eq. (64), allowing for determining explicitly the
eigenvalues (IV). They are used to construct the generalized zeta function for
the spinor field and thus remove the formal divergences involved in the
Casimir effect.
These divergencies were introduced by the Dirac operator determinant in the
partition function originating from the infinite product over eigenvalues, Eq.
(II.1). This divergence was encoded into the generalized zeta function
employing the important relation connecting it with the partition function,
Eq. (29). It was analyzed from the asymptotic behavior of the spinor heat
kernel function, Eq. (44), and removed in the renormalization scheme by
subtraction of the divergent contribution associated with non-null heat kernel
coefficients. A rather peculiar aspect of the zeta function regularization
prescription is related to the existence of ambiguities. Such ambiguities
appear whenever the mass-dependent $c_{2}(m)$ heat kernel coefficient is
nonvanishing, Eq. (50), due to natural dependence on parameter $\mu$, Eq.
(III). For the geometry presented here, $c_{0}$ was the only non-null heat
kernel coefficient, Eq. (85), associated with the Euclidean heat kernel
contribution, Eqs. (80) and (100). In order to derive physically meaningful
expressions, all contributions associated with $c_{0}$ were dropped to ensure
that the renormalization procedure is unique and thus obtain an unambiguous
spinor vacuum free energy. Besides that, since $c_{0}$ is multiplied by mass
with a positive exponent, we adopt an additional requirement that vacuum
energy should be renormalized to zero for large masses.
We outline all the mathematical machinery required for computing the vacuum-
free energy density, starting with the construction of the partition function
for the spinor field through Euclidean path integrals. In this Euclidean
approach, we find closed and analytical expressions for the vacuum free energy
density associated with the spinor field in thermal equilibrium at finite
temperature $T=\beta^{-1}$, satisfying anti-periodic conditions in the
imaginary time $t$ and along vector $\boldsymbol{L}$. This energy density can
be expressed as a summation of the zero-temperature Casimir energy density,
Eq. (90), and temperature correction terms, Eq. (IV.2), which generalize the
results presented in Refs. Bellucci2009 ; Xiang2011 . We also analyzed the
high- and low-temperature asymptotic limits, which agree entirely with the
curves shown in Fig. 1. The ratio of the renormalized free energy density to
the Casimir energy density goes to $1$ as $T$ approaches zero and decays to
zero as $T$ approaches infinity. At high temperatures, in particular, we have
shown that the $c_{0}$ coefficient gives rise to the Stefan-Boltzmann law,
proportional to $T^{4}$. Although not divergent, this quantum term was
subtracted in the renormalization procedure to obtain a correct classical
contribution in this limit. Also, the free energy density does not possess a
classical limit at high temperatures. Except for this classical limit, all our
results for spinor fields differ from the ones for scalar fields by constant
multiplicative factors that naturally arise because of the spinor degrees of
freedom. Finally, our analysis confirms that the entropy density agrees with
the Nernst heat theorem.
## VI Acknowledgments
J. V. would like to thank Fundação de Amparo a Ciência e Tecnologia do Estado
de Pernambuco (FACEPE), for their partial financial support. A. M.
acknowledges financial support from the Brazilian agencies Conselho Nacional
de Desenvolvimento Científico e Tecnológico (CNPq), grant no. 309368/2020-0
and Coordenação de Aperfeiçoamento de Pessoal de Nível Superior (CAPES). H. M.
is partially supported by CNPq under grant no. 308049/2023-3. L. F.
acknowledges support from CAPES, financial code 001.
## References
* (1) H. B. G. Casimir, On the Attraction Between Two Perfectly Conducting Plates, Indag. Math. 10, 261 (1948).
* (2) M. J. Sparnaay, Measurements of attractive forces between flat plates, Physica 24, 751 (1958).
* (3) G. Bressi, G. Carugno, R. Onofrio, and G. Ruoso, Measurement of the Casimir force between parallel metallic surfaces, Phys. Rev. Lett., 88, 041804 (2002).
* (4) S. K. Lamoreaux, Erratum: Demonstration of the casimir force in the 0.6 to 6 µm range, Phys. Rev. Lett. 81, 5475 (1998).
* (5) S. K. Lamoreaux, Demonstration of the Casimir force in the 0.6 to 6 micrometers range, Phys. Rev. Lett. 78, 5 (1997). Erratum: Phys.Rev.Lett. 81, 5475 (1998).
* (6) U. Mohideen and A. Roy, Precision measurement of the Casimir force from 0.1 to 0.9 micrometers, Phys. Rev. Lett. 81 4549 (1998).
* (7) M. Bordag, U. Mohideen, and V. M. Mostepanenko, New developments in the Casimir effect, Phys. Rept. 353, 1 (2001).
* (8) V. M. Mostepanenko and N. N. Trunov, The Casimir effect and its applications (Clarendon Press, Oxford, New York, 1997).
* (9) L. H. Ford, Quantum vacuum energy in general relativity, Phys. Rev. D 11, 3370 (1975).
* (10) J. S. Dowker and R. Critchley, Covariant Casimir calculations, J. Phys. A 9, 535 (1976).
* (11) J. S. Dowker, Thermal properties of Green’s functions in Rindler, de Sitter, and Schwarzschild spaces, Phys. Rev. D 18, 1856 (1978).
* (12) T. Appelquist and A. Chodos, Quantum Effects in Kaluza-Klein Theories, Phys. Rev. Lett. 50, 141 (1983).
* (13) Y. Hosotani, Dynamical gauge symmetry breaking as the Casimir effect, Phys. Lett. B 129, 193 (1983).
* (14) I. Brevik, K.A. Milton, S.D. Odintsov, Entropy Bounds in $R\times S^{3}$ Geometries, Ann. Phys. 302, 120 (2002).
* (15) A. Zhang, Thermal Casimir Effect in Kerr Space-time, Nuclear Physcis B 898, 2020, (2015).
* (16) C. Henke, Quantum vacuum energy in general relativity, Eur. Phys. J. C 78, 126 (2018).
* (17) K. Bradonjic, J. Swain, A. Widom, and Y. Srivastava, The casimir effect in biology: The role of molecular quantum electrodynamics in linear aggregations of red blood cells, Journal of Physics: Conference Series 161, 012035 (2009).
* (18) Peng Liu and Ji-Huan He, Geometric potential: An explanation of nanofibers wettability, Thermal Science 22, 33 (2018).
* (19) P. Pawlowski and P. Zielenkiewicz, The quantum casimir effect may be a universal force organizing the bilayer structure of the cell membrane, The Journal of membrane biology 246, 383 (2013).
* (20) A. Gambassi, The Casimir effect: From quantum to critical fluctuations, J. Phys. Conf. Ser. 161, 012037 (2009).
* (21) B. B. Machta, S. L. Veatch, and J. P. Sethna, Critical casimir forces in cellular membranes, Phys. Rev. Lett. 109, 138101 (2012).
* (22) G. L. Klimchitskaya, U. Mohideen and V. M. Mostepanenko, The Casimir force between real materials: Experiment and theory, Rev. Mod. Phys. 81, 1827 (2009).
* (23) K. Milton and I. Brevik, Casimir Physics Applications, Symmetry 11, 201 (2019).
* (24) K. A. Milton, The Casimir Effect: Physical Manifestation of Zero-Point Energy (World Scientific, Singapore, 2001).
* (25) M. Bordag, G. L. Klimchitskaya, U. Mohideen, and V. M. Mostepanenko, Advances in the Casimir effect (Oxford University, Press, Oxford, 2009), Vol. 145.
* (26) A. Stokes and R. Bennett, The Casimir effect for fields with arbitrary spin, Annals Phys. 360, 246 (2015).
* (27) C. Farina, The Casimir Effect: Some Aspects. Brazilian Journal of Physics 36, 1137 (2006).
* (28) C. R. Muniz, M. O. Tahim, M. S. Cunha and H. S. Vieira, On the global Casimir effect in the Schwarzschild spacetime, JCAP 1801, 006 (2018).
* (29) M. S. Cunha, C. R. Muniz, H. R. Christiansen, V. B. Bezerra, Relativistic Landau levels in the rotating cosmic string spacetime, Eur. Phys. J. C 76, 512 (2016).
* (30) S. Mobassem, Casimir effect for massive scalar field, Mod. Phys. Lett. A 29, 1450160 (2014).
* (31) S. H. Pereira, J. M. Hoff da Silva and R. dos Santos, Casimir effect for Elko fields, Mod. Phys. Lett. A 32, 1730016 (2017).
* (32) A. A. Bytsenko, M. E. X. Guimarães and V. S. Mendes, Casimir Effect for Gauge Fields in Spaces with Negative Constant Curvature, Eur. Phys. J. C 39, 249 (2005).
* (33) M. N. Chernodub, V. A. Goy, A. V. Molochkov and Ha Huu Nguyen, Casimir effect in Yang-Mills theory, Phys. Rev. Lett. 121, 191601 (2018)
* (34) A. Edery, Casimir piston for massless scalar fields in three dimensions, Phys. Rev. D 75 (2007) 105012, [hep-th/0610173].
* (35) E. Ponton and E. Poppitz, Casimir Energy and Radius Stabilization in Five and Six Dimensional Orbifolds, JHEP 0106, 019 (2001).
* (36) E. Cartan, The Theory of Spinors, Dover (1966).
* (37) I. Benn and R. Tucker, An Introduction to spinors and geometry with applications in physics, Adam Hilger, Bristol (1987).
* (38) J. Venâncio, The spinorial formalism. Lambert Academic Publishing, Germany (2019).
* (39) V. K. Oikonomou and N. D. Tracas, Slab Bag Fermionic Casimir effect, Chiral Boundaries and Vector Boson-Majorana Fermion Pistons, Int. J. Mod. Phys. A 25, 5935 (2010).
* (40) H. Cheng, Casimir effect for parallel plates involving massless Majorana fermions at finite temperature, Phys. Rev. D 82, 045005 (2010).
* (41) R. M. Cavalcanti, Casimir force on a piston, Phys. Rev. D 69, 065015 (2004).
* (42) E. Elizalde, Zeta function methods and quantum fluctuations, J. Phys. A 41, 304040 (2008).
* (43) E. Elizalde, Ten Physical Applications of Spectral Zeta Functions (Springer, New York, 1995).
* (44) S. W. Hawking, Zeta function regularization of path integrals in curved spacetime, Commun. math. Phys. 55, 133 (1977).
* (45) E. Elizalde, Zeta function regularization in Casimir effect calculations and J.S. Dowker’s contribution, Int. J. Mod. Phys. A 27, 1260005 (2012).
* (46) E. Elizalde, S. D. Odintsov, A. Romeo, A. A. Bytsenko, and S. Zerbini, Zeta regularization techniques with applications, World Scientific (1994).
* (47) M. Basil Altaie and J. S. Dowker, Spinor fields in an Einstein universe: Finite-temperature effects, Phys. Rev. D 18, 3557 (1978).
* (48) G. Plunien, B. Muller and W. Greiner, The Casimir effect, Phys. Rept. 134, 87 (1986).
* (49) I. K. Kulikov, P. I. Pronin. Finite temperature contributions to the renormalized energy-momentum tensor for an arbitrary curved space-time. Czech J Phys 38, 121 (1988).
* (50) R. V. Maluf, D. M. Dantas and C. A. S. Almeida, The Casimir effect for the scalar and Elko fields in a Lifshitz-like field theory, Eur. Phys. J. C 80, 442 (2020).
* (51) M. Kac, Can one hear the shape of a drum? Amer. Math. Monthly 73, 1 (1966).
* (52) M. Bordag, Ground state energy for massive fields and renormalization, Commun. Mod. Phys., Part D 1, 347 (2000).
* (53) D. V. Vassilevich, Heat kernel expansion: user’s manual, Phys. Rept. 388, 279 (2003).
* (54) K. Kirstein, Basic zeta functions and some applications in physics, 2010. arXiv:1005.2389
* (55) T. Branson and P. Gilkey, Residues of the eta function for an operator of Dirac type, J. Funct. Anal. 108, 47 (1992).
* (56) T. Branson and P. Gilkey, Residues of the eta function for an operator of Dirac type with local boundary condtitons, Differential Geom. Appl. 2, 249 (1992).
* (57) A. Chodos, R. L. Jaffe, K. Johnson, C. B. Thorn, and V. F. Weisskopf, New extended model of hadrons, Phys. Rev. D 9, 3471 (1974).
* (58) K. Johnson, The M.I.T. bag model, Acta Phys. Polon. B 6, 865 (1975).
* (59) N. Arrizabalaga, L. Le Treust and N. Raymond, On the MIT bag model in the non-relativistic limit, Commun. Math. Phys. 354, 641 (2017).
* (60) S. G. Mamayev and N. N. Trunov, Vacuum expectation values of the energy-momentum tensor of quantized fields on manifolds with different topologies and geometries. III, Sov. Phys. J. 23, 551, (1980).
* (61) A. Erdas, Finite temperature Casimir effect for massless Majorana fermions in a magnetic field, Phys. Rev. D 83, 025005 (2011).
* (62) E. Elizalde et al., The Casimir energy of a massive fermionic field confined in a $(d+1)$-dimensional slab-bag, International Journal of Modern Physics A 18, 1761 (2012).
* (63) E. Elizalde, M. Bordag and K. Kirsten, Casimir energy for a massive fermionic quantum field with a spherical boundary, J. Phys. A: Math. Gen. 31, 1743 ( 1998).
* (64) C. M Bender and P. Hays, Zero-point energy of fields in a finite volume, Phys. Rev. D 14, 2622 (1976).
* (65) G. Fucci and C. Romaniega, Casimir energy for spinor fields with $\delta$-shell potentials, J. Phys. A: Math. Theor. 56, 265201 (2023).
* (66) S. H. Pereira and R. S. Costa, Partition function for a mass dimension one fermionic field and the dark matter halo of galaxies, Mod. Phys. Lett. A 34, 1950126 (2019).
* (67) Xiang-hua Zhai, Xin-zhou Li, Chao-Jun Feng, Casimir effect with a helix torus boundary condition, Mod. Phys. Lett. A 26, 1953 (2011).
* (68) K. E. L. de Farias and H. F. Santana Mota, Quantum vacuum fluctuation effects in a quasi-periodically identified conical spacetime, Phys. Lett. B 807 (2020) 135612, [arXiv:2005.03815].
* (69) Chao-Jun Feng and Xin-zhou Li, Quantum Spring from the Casimir Effect, Phys. Lett. B 691 (2010) 167–172, [arXiv:1007.2026].
* (70) Xiang-hua Zhai, Xin-zhou Li, Chao-Jun Feng, The Casimir force of Quantum Spring in the (D+1)-dimensional spacetime, Mod. Phys. Lett. A 26 (2011) 669–679, [arXiv:1008.3020].
* (71) Chao-Jun Feng and Xin-zhou Li, Quantum Spring, Int. J. Mod. Phys. Conf. Ser. 7 (2012) 165–173, [arXiv:1205.4475].
* (72) H. F. Mota and V. B. Bezerra, Topological thermal Casimir effect for spinor and electromagnetic fields, Phys. Rev. D 92, 124039 (2015).
* (73) K. E. L. de Farias, A. Mohammadi, and H. F. Santana Mota, Thermal Casimir effect in a classical liquid in a quasi-periodically identified conical spacetime, Phys. Rev. D 105, 085024 (2022).
* (74) H. Mota, Vacuum energy, temperature corrections and heat kernel coefficients in $(D+1)$-dimensional spacetimes with nontrivial topology, 2023. arXiv:2312.01909
* (75) V. B. Bezerra, V. M. Mostepanenko, H. F. Mota, and C. Romero, Thermal Casimir effect for neutrino and electromagnetic fields in the closed Friedmann cosmological model, Phys. Rev. D 84, 104025 (2011).
* (76) N. D. Birrell and L. H. Ford, Renormalization of Self-Interacting Scalar Field Theories in a Nonsimply Connected Spacetime. Physical Review D 22, 330 (1980).
* (77) Xiang-hua Zhai, Xin-zhou Li and Chao-Jun Feng, Fermionic Casimir effect with helix boundary condition, Eur. Phys. J. C 71, 1654 (2011).
* (78) S. Bellucci, A.A. Saharian, Fermionic Casimir effect for parallel plates in the presence of compact dimensions with applications to nanotubes, Phys. Rev. D 80, 105003 (2009).
* (79) G. Aleixo, H. F. Santana Mota, Thermal Casimir effect for the scalar field in flat spacetime under a helix boundary condition, Phys. Rev. D 104, 045012 (2021).
* (80) Chao-Jun Feng, Xin-Zhou Li, Xiang-Hua Zhai, Casimir Effect under Quasi-Periodic Boundary Condition Inspired by Nanotubes, Mod. Phys. Lett A 29, 1450004 (2014
* (81) B. Liu, C. L. Yuan, H. L. Hu, et al. Dynamically actuated soft heliconical architecture via frequency of electric fields. Nat Commun 13, 2712 (2022).
* (82) I. Greenfeld, I. Kellersztein and H. D. Wagner, Nested helicoids in biological microstructures. Nat Commun 11, 224 (2020).
* (83) E. Elizalde and A. Romeo, Heat-kernel approach to zeta function regularization of the Casimir effect for domains with curved boundaries, Int. J. Mod. Phys. A 5, 1653, (1990).
* (84) R.T. Seeley, Complex powers of an elliptic operator, Amer. Math. Soc. Proc. Symp. Pure Math. 10, 288 (1967).
* (85) D. V. Vassilevich, Operators, Geometry and Quanta, Springer (2011).
* (86) P. Gilkey, Invariance Theory, the Heat Equation, and the Atiya-Singer Index Theorem. CRC Press, Boca Raton, FL, 1995.
* (87) I. K. Kulikov and P. I. Pronin, Topology and chiral symmetry breaking in four-fermion interaction, Acta Phys. Polon. B 20, 713 (1989).
* (88) N. Ahmadi and M. Nouri-Zonoz, Massive spinor fields in flat spacetimes with non-trivial topology, Phys. Rev. D 71, 104012 (2005).
* (89) J. Venâncio and C. Batista, Separability of the Dirac equation on backgrounds that are the direct product of bidimensional spaces, Physical Review D 95, 084022 (2017).
* (90) S. S. Gousheh, A. Mohammadi, and L. Shahkarami, Casimir energy for a coupled fermion-kink system and its stability, Phys. Rev. D 87, 045017 (2013).
* (91) A. Mohammadi, E. R. Bezerra de Mello and A. A. Saharian, Induced fermionic currents in de Sitter spacetime in the presence of a compactified cosmic string, Class. Quantum Grav. 32, 135002 (2015).
* (92) E. A. F. Bragança, E. R. Bezerra de Mello, and A. Mohammadi, Induced fermionic vacuum polarization in a de Sitter spacetime with a compactified cosmic string, Phys. Rev. D 101, 045019 (2020).
* (93) A. Bytsenko, L. Vanzo, S. Zerbini, Zeta-function regularization approach to finite temperature effects in Kaluza-Klein space-times, Mod. Phys. Lett. A 7, 2669 (1992).
* (94) S. Zerbini, Spinor fields on $2+1$ topologically nontrivial spacetime, Letters in Mathematical Physics 27, 19 (1993).
* (95) L. D. Landau and E. M. Lifshitz, Statistical Physics, Part I. Pergamon Press, Oxford, 1980.
* (96) C.J. Feng, X.Z. Li, Quantum spring from the Casimir effect, Phys. Lett. B 691, 167 (2010).
* (97) M. Abramowitz and I. A. Stegun, Handbook of Mathematical Functions with Formulas, Graphs, and Mathematical Tables (Dover, New York, 1972).
|
# Fruit Picker Activity Recognition with Wearable Sensors and Machine Learning
††thanks: This work is a product of a collaboration with Grow Logic, who
provided the dataset, photographs, and the context to the problem. The work
was supported by the Systems program in Agriculture & Food, CSIRO.
Joel Janek Dabrowski∗ Data61, CSIRO
Brisbane, Australia
<EMAIL_ADDRESS>Ashfaqur Rahman Data61, CSIRO
Hobart, Australia
<EMAIL_ADDRESS>
###### Abstract
In this paper we present a novel application of detecting fruit picker
activities based on time series data generated from wearable sensors. During
harvesting, fruit pickers pick fruit into wearable bags and empty these bags
into harvesting bins located in the orchard. Once full, these bins are quickly
transported to a cooled pack house to improve the shelf life of picked fruits.
For farmers and managers, the knowledge of when a picker bag is emptied is
important for managing harvesting bins more effectively to minimise the time
the picked fruit is left out in the heat (resulting in reduced shelf life). We
propose a means to detect these bag-emptying events using human activity
recognition with wearable sensors and machine learning methods. We develop a
semi-supervised approach to labelling the data. A feature-based machine
learning ensemble model and a deep recurrent convolutional neural network are
developed and tested on a real-world dataset. When compared, the neural
network achieves 86% detection accuracy.
###### Index Terms:
human activity recognition, convolutional neural network, recurrent neural
network, deep learning, agriculture, time-series
## I Introduction
It is estimated that up to 25% of all fruit and vegetable produce go to waste
before ever leaving the farm, costing Australian farmers $\$2.84$ billion
annually [1]. Waste may be caused by pests, disease, weather events, or damage
inflicted during harvesting. In harvesting, fruit are picked and placed in
harvesting bins located throughout the paddocks or orchards, and transported
to a packing house for processing prior to market distribution. The picked
fruit are spoiled when they are left in the harvesting bins exposed to heat
for extended periods of time. Farmers and managers must thus ensure that the
pickers fill the bins quickly and that the bins are transported to the cooled
packing house as soon as they are full.
In general, the machine learning paradigm provides promising approaches to
address challenges in harvesting, however examples of such applications are
scarce in the literature [2]. This study contributes to this gap in the
literature and provides a means to monitor fruit pickers and the filling of
the harvesting bins to allow for more effective management of the harvesting
process.
SensorBinsBag Figure 1: Top: Photograph of a fruit picker, a picker bag,
sensor, and harvesting bin. The bin is mounted on a trailer, which is
typically pulled by a tractor. There are typically many bins scattered across
the orchard. Bottom: Photograph of bag-emptying events where two fruit pickers
empty the bags into a bin.
As illustrated in Figure 1, fruit pickers are equipped with a fruit picker
bag. Fruit are picked and placed into these bags, and once the bag is full,
the fruit is emptied into a harvesting bin. This is referred to as a bag-
emptying event. Detecting bag-emptying events provides the means to
simultaneously monitor the fruit picker productivity and bin levels. The bag-
emptying event rate is a direct measurement of the fruit picker productivity.
Furthermore, given the known capacity of a picker bag and the knowledge of
which bin the fruit is emptied into, bag-emptying events provide a measure of
the filling rate of the bins. With this knowledge, farmers are able to more
effectively manage pickers and bins.
Given recent success in the application of deep and machine learning
approaches to human activity recognition [3], we propose a novel application
of these approaches to detect bag-emptying events using wearable sensors. A
wearable sensor is placed on a picker’s bag strap as illustrated in Figure 1
to measure the picker’s movements and their proximity to the harvest bins.
Given the sensor measurements, bag-emptying events are detected using machine
learning models. We consider two models: a feature-based machine learning
ensemble classifier and a deep Recurrent Convolutional Neural Network (RCNN).
The models are compared on a dataset collected during avocado harvesting.
In addition to bag-emptying event detection, we address a data labelling
problem where the duration of bag-emptying events are unknown. For this, the
K-means clustering algorithm is used to perform semi-supervised labelling of
the data. Approximate times of when the events occurred and several features
of the data are used to learn the duration of bag-emptying events.
To our knowledge, fruit picker activity recognition based on time series data
from wearable sensors is a novel application. The main contributions of this
study include: (1) we use a wearable accelerometer sensor for measuring fruit
picker activity, (2) we used machine learning to understand fruit picker
productivity, and (3) we develop a new methodology that uses a combination of
semi-supervised labelling and supervised learning to detect bag-emptying
events. The key advantages of our approach is it does not interfere with the
harvesting process and it is autonomous.
This article begins with a discussion on related work in Section II. The
dataset and data labelling approaches are discussed in Sections III and IV
respectively. The models are described in Section V and our methodology is
provided in Section VI. Results are presented in Section VII and the article
is concluded in Section VIII.
## II Related Work
### II-A Machine Learning in Agriculture
Machine learning has been applied to various agricultural problems and several
literature surveys have been produced [2, 4, 5]. Most applications relate to
crop management and include yield prediction, disease detection, weed
detection, crop quality prediction, and water management. Literature on the
application of machine learning approaches specifically to harvesting and the
harvesting process are scarce.
### II-B Fruit Picker Productivity
In relation to fruit picker activity recognition, several studies have
considered tracking pickers for yield mapping, e.g., see [6, 7, 8, 9, 10].
Tracking pickers can be challenging due to Global Positioning System (GPS)
signal losses through foliage. Various alternatives to GPS have been thus been
proposed [11, 12]. However, even accurate picker tracking does not necessarily
provide any direct information on the picker’s activity. In our work, we are
directly measuring the fruit picker’s activity, which to our knowledge has not
been considered in the literature before.
A more direct approach to measuring picker productivity is to have a fixed
platform located at the bin containing some form of digital scale and a device
to identify the picker (such as RFID or a bar code scanner) [13, 14, 15, 16].
The picker identifies them self, weighs the picked fruit, and releases the
weighed fruit into the bin. Drawbacks of this approach include the additional
time that is required for weighing bags and the additional supervisors
required to ensure that the weighing processes is being conducted correctly.
Our approach does not interfere with the harvesting process and does not
require additional supervisors. Furthermore, our approach considers time-
series data rather than data at a single point in time.
### II-C Human Activity Recognition and Machine Learning
Detecting fruit picker bag-emptying events can be considered as a Human
Activity Recognition (HAR) problem. Generally, HAR involves using some form of
classifier to predict an activity given data from a sensor that directly
monitors human movement, such as wearable sensors. Surveys on HAR using
wearable sensors [17] deep learning approaches to HAR [3] have been conducted.
Applications with wearable sensors in agriculture include human–robot
interaction [18] and the assessment of vibration risk with agricultural
machinery [19]. To our knowledge HAR has not been applied to monitor fruit
pickers.
A wide range of wearable sensors exist [17] including: accelerometers, global
positioning systems (GPS), radio frequency identification (RFID),
environmental sensors (such as temperature), and physiological sensors (such
as heart rate monitors). A survey has been conducted on sensor positioning on
the body [20]. Sensor positions may include the waist, arms, wrist, ankle, and
the torso. The chest is suggested to be the preferred location for medium-
level activities such as walking and house-work. The actions in such
activities are similar to that of fruit picking. The chest was thus chosen as
the location for this study.
Many HAR models begin by processing the accelerometer data using a sliding
window [17]. Various features are extracted from the data in the window as it
is slid across the dataset. Features may include mean, standard deviation,
minimum, maximum, energy, main frequency component, root mean square of the
derivative, and correlation between axes [20, 21]. The features are fed into a
classifier, which classifies the activity type. Various classifiers such as
decision trees, neural networks, Bayesian models, Markov models, and
classifier ensembles have may be considered [17]. In this study, such a
feature-based ensemble classifier is compared with a deep neural network
model.
Feature selection and design can be a tedious task that typically requires
domain knowledge. Deep learning algorithms are often designed to be end-to-end
methods that take the raw input data and output a prediction. Features are
learned within the multiple layers of network. A survey of various deep
learning architectures that have been applied to sensor-based activity
detection problems has been conducted [22]. These architectures include the
convolutional neural network (CNN) and the recurrent neural network (RNN).
The CNN is a neural network that applies a convolution operation on the data
presented to its inputs [23]. CNNs exploit local interactions in the data and
provide scale invariant features [24]. In wearable sensor HAR problems,
several CNN based models have been proposed [25, 26, 24].
The RNN is a neural network that is designed for sequential applications [23,
27]. It contains a neural network that is replicated over time where the
replications are sequentially connected. Like the CNN, the RNN has been also
applied to several HAR problems [28, 29, 30].
RNN and CNN models have been compared on HAR tasks [31]. It is found that RNNs
perform well for activities where long term dependency is required, such as
opening a door. CNNs perform well when long-term dependencies are not required
such as gait analysis. Given that RNNs and CNNs each have their own
advantages, combining the two architectures may provide a more widely
applicable and more powerful model. The combination of the architectures forms
a recurrent convolutional neural network (RCNN) and provide a promising
framework for HAR [32, 33], and is an approach we consider in this study.
## III Dataset
### III-A Sensors and Data Collection
The Haltian Nexus Prototype sensors were used in the wearable sensor. These
sensors comprise a ST LIS2DH accelerometer, a Nordic nRF52832 Wirepas radio
gateway module, and a data logger. Accelerometer readings for 3 axes were
logged at a frequency of 50Hz with a measurement range of $\pm$4 G and
sensitivity of 8 mG. The radio logged its Received Signal Strength Indicator
(RSSI) value with respect to a Haltian Thingsee POD2 Prototype node every
second. The POD2 node was located at the picker’s bin.
### III-B Avocado Farm Trial Dataset
A trial was conducted on an avocado farm. The data was acquired for two
different pickers over several hours. A sensor was attached to each pickers
bag strap and positioned on the picker’s chest as illustrated in Figure 1(a).
Bag-emptying event times were manually recorded by a human observer. The
dataset comprises 580986 samples of data with 64 bag-emptying events.
Each of the 3 data streams from the 3-axis accelerometer are filtered with a
bandpass filter. The high-cut frequency is set to reduce aliasing. The low-cut
frequency is set to remove any offsets caused by gravitational effects. The
three filtered accelerometer data streams and the RSSI data stream are
combined to form a dataset of four data streams, which are scaled to a range
of $[0,1]$.
### III-C Dataset Balancing
The dataset is imbalanced where only $28\%$ of the samples are associated with
bag-emptying events. This imbalance can create an undesirable bias in the
classifiers. Resampling is used to form a balanced dataset. Sequences of
samples are extracted to ensure the sequential nature of the data is
maintained. To introduce some form of randomness in the sampling, the length
of the extracted sequence is selected according to a normal distribution. The
mean length $\mu$ and variance $\sigma$ of the bag-emptying event sequences
are calculated from the sequence length of the manual label bag-emptying
events. For each bag-emptying event sequence, only the
$n\in\mathcal{N}(\mu,\sigma)$ preceding non-bag-emptying event samples are
preserved. All remaining non-bag-emptying event samples are removed. The
result is a dataset with sequence lengths that are similarly distributed
between classes.
## IV Bag-Emptying Event Labelling
The supervised machine learning algorithms require labels of the bag-emptying
events for training. Although the times that the bag-emptying events occurred
were recorded, the duration of the events were not recorded. Furthermore,
owing to human error, the recoded bag-emptying times are only considered to be
approximations. The bag-emptying event times are thus required to be refined
and the bag-emptying event durations are required to be determined. For this,
we consider two approaches: (1) a manual labelling approach using expert
knowledge and (2) a semi-supervised approach based on K-means clustering.
### IV-A Manual Labelling of Bag-Emptying Events
To manually determine the bag-emptying event times and durations, the
following bag-drop process is noted: (1) A pickers gait changes under the
strain of a full bag as they walk from the trees to the bin; (2) the picker
lifts the bag into the bin; (3) a flap at the bottom of the bag is opened to
release the fruit (e.g. see Figure 1); (4) the bag is shaken and tugged to
empty it; and (5) once the bag is empty, it is removed from the bin and the
bottom flap is reattached.
The scaled accelerometer and RSSI data surrounding a bag-emptying event are
illustrated in Figure 2. The plot indicates a change in the dynamics as the
picker transitions from normal picking activity to the bag-emptying event. The
RSSI increases as the picker approaches the bin and the accelerometer signal
level increases due to gait change. A spike in the accelerometer data occurs
as the picker lifts the bag into the bin (this spike is not evident in every
bag-emptying event). The accelerometer signal level remains high as the bag is
shaken, removed and reassembled. The RSSI and the accelerometer signal
decrease as the picker returns to fruit picking activities. Based on these
observations the bag-emptying times were manually refined and the duration of
the bag-emptying events were defined. The average bag-emptying event duration
is 50 seconds with the emptying of the bag typically lasting between 10 and 20
seconds.
Figure 2: Plot of the sensor data with a bag-emptying event. Sensor data is
plotted in blue. The manually defined bag-emptying event is plotted in orange
with a value of 1 indicating a bad drop event.
### IV-B Semi-supervised labelling of Bag-Emptying Events
The K-means clustering algorithm is used to perform semi-supervised learning
of the bag-emptying event labels111The K-means algorithm is selected due to
its computational efficiency, however more complex clustering algorithms could
also be considered.. The approach is to pre-define the duration of bag-
emptying events around the logged bag-emptying times. The K-means algorithm is
then used to refine the duration by clustering samples according to
statistical features of bag-emptying event and non-bag-emptying event data.
Data sample labels are initialised by defining the duration of all bag-
emptying events to be 1200 samples in length, with 500 samples before and 700
samples after the manually logged bag-drop event time. Samples within this
window are associated with bag-emptying events and samples outside this window
are associated with normal fruit picking activity.
A 256-sample sliding window is shifted sample-by-sample over the data
sequences. For each shift, a set of statistical features are extracted to
produce a feature vector associated with each sample. This feature vector
comprises the mean, standard deviation, minimum, maximum, and standard
deviation of the derivative for the associated sample. Additionally, to
capture the sequential structure of the data, an indicator is included to
specify whether the neighbouring samples labels are associated with bag-
emptying events.
K-means algorithm is initialised by grouping the feature vectors into two
clusters according to the initial bag-emptying event labels. The mean values
of each cluster form the initial values for the K-means clustering algorithm.
These clusters are refined using the K-means algorithm, which corresponds to
refining the labels of each data sample. The algorithm was typically run over
10 iterations, where the number of iterations can affect how much the clusters
change from the predefined settings.
The K-means clustering algorithm can produce false positives and false
negatives, where false positives are spurious bag-emptying events predicted
far from the logged time and false negatives are spurious non-bag-emptying
events predicted near logged bag-emptying event times.
The results of the K-means algorithm are thus filtered using the intersection
and union operators of mathematical set theory. Predicted bag-emptying event
sequences are compared with the predefined bag-emptying event sequence as
illustrated in Figure 3. To reduce false positives an intersection operator is
used. If none of the samples in the predicted sequence overlap with the
predefined sequence, the prediction is rejected. As illustrated in the bottom
plot of Figure 3, the predictions at the beginning and end of the sequence are
removed as they do not overlap with the predefined bag-emptying event
sequence. To reduce false negatives, a union operation is applied between the
predicted and predefined sequences. As illustrated in the bottom plot of
Figure 3, the false negative is removed. Note that this operator is only
applied to join predicted bag-emptying events. It is not permitted to extend
the predicted bag-emptying event duration. This operation can however result
in an unreasonably long bag-emptying event sequence if it joins two long
sequences of positive samples. If a predicted sequence is longer than twice
the predefined bag-emptying event sequence, it is rejected and the predefined
bag-emptying event sequence is used instead.
Figure 3: Plots of the process of semi-supervised bag-emptying event
labelling. The predefined labels, the K-means clustering results, and the
filtered K-means clustering results are plotted in the top, middle and bottom
figures respectively.
The overall algorithm for the semi-supervised labelling approach is presented
in Algorithm 1.
Algorithm 1 Semi-supervised Labelling of bag-emptying events using the K-means
algorithm.
0: The dataset $X$, a vector of bag-emptying event start times
$t^{\text{start}}$, a vector of bag-emptying event end times $t^{\text{end}}$,
and the dataset labels $Y$.
1: Define the window size $q=256$
2: Define an empty set of dataset labels $\hat{Y}=\varnothing$
3: for each bag-emptying event, $i$ do
4: Extract dataset sequence surrounding the $i^{\text{th}}$ event
$A=X_{t^{\text{end}}_{i-1}:t^{\text{start}}_{i+1}}$
5: Extract the labels associated with $A$
$B=Y_{t^{\text{end}}_{i-1}:t^{\text{start}}_{i+1}}$
6: for each index $j$ of sample in $A$ do
7: Extract the window of data associated with sample $j$ $W=A_{j:j+q}$
8: Compute the feature vector for sample $j$ $f_{j}=\text{features}(W)$
9: Set the class of sample $j$ according to the window
$c_{j}=\begin{cases}0&\frac{1}{q}\sum_{k=1}^{q}B_{j+k}<0.5\\\
1&\frac{1}{q}\sum_{k=1}^{q}B_{j+k}\geq 0.5\end{cases}$
10: end for
11: Update the classes using the K-means algorithm
$c\leftarrow\text{kmeans}(f,c)$
12: Filter the labels using mathematical set theory
$c\leftarrow\text{filter}(c,B)$
13: Append the filtered labels $c$ to the new label set $\hat{Y}$.
14: end for
15: return The updated labels $\hat{Y}$.
## V Models
Detecting a bag-emptying event from the wearable sensor data is a challenging
task. As described in Section IV, the events comprises various sub-activities
involved in the bag-emptying event. Furthermore, the bag-drop signals may vary
according to pickers and the environment. The models are required to handle
these variations.
### V-A Feature Based Ensemble Model
RSSIAcc. ZAcc. YAcc. XFeature vector (std., energy, RMS($dx/dt$),
RMS($d^{2}x/dt^{2}$), mean($dx/dt$), mean($d^{2}x/dt^{2}$), min, max)Naive
BayesANN (512)$\sum$$60\%$$40\%$Input dataFeature vectorEnsembleClass Figure
4: Architecture of the ensemble model. A feature vector is assembled from a
window of data. A naive Bayes and an ANN perform a classification given the
feature vector. The outputs of these classifiers are weighted and summed to
determine the class.
A traditional feature-based ensemble model is applied for detecting bag-
emptying events. The model is illustrated in Figure 4. A 256 sample sliding
window with zero overlap is applied to the data. The following features are
extracted in each window: standard deviation, energy, the RMS first and second
derivative, mean first and second derivative, minimum value, and maximum value
[21, 20].
The features are provided as inputs to an ensemble classifier comprising a
Gaussian Naive Bayes classifier and a neural network. These are two commonly
used classifiers in HAR [17] and are sufficiently different from each other to
provide diversity in the ensemble. The neural network comprises a single
hidden layer with 512 neurons with hyperbolic tangent activation functions.
The ADAM algorithm [34] is used to train the neural network. The parameters of
the Gaussian naive Bayes classifier are estimated using maximum likelihood.
Both classifiers output a prediction in the form of a probability of a bag-
emptying event. These predictions are combined in the ensemble through a
weighted summation. The Naive Bayes classifier is weighted with 60% of the
vote (which was determined through cross validation).
### V-B Recurrent Convolutional Neural Network (RCNN)
The architecture of the RCNN used in this study is illustrated in Figure 5 and
is based on the models presented in [32] and [33]. For each data stream, a set
of samples are extracted using a 256 sample sliding window with zero overlap.
The windows of samples from the 4 sensor streams are combined into a tensor,
where each stream represents channel for the input of the CNN portion of the
model. A key advantage of the RCNN over the ensemble model is that it is an
end-to-end model and does not require feature engineering.
The CNN portion of the RCNN performs feature extraction and the RNN portion of
the RCNN models temporal dynamics over the sample windows. The RCNN outputs
the class of the input window sample. The rectified linear unit (ReLU) is used
as the activation function in hidden layers, the filter and layer sizes in the
network were determined through trial and error, and the ADAM algorithm [34]
is used for training.
RSSIAcc. ZAcc. YAcc. XConv. 1 (filters = 32, kernel size = 32)Pool 1 (pool
size = 8, stride = 4)Conv. 2 (filters = 16, kernel size = 16)Pool 2 (pool size
= 4, stride = 2)Dense (512 neurons)Input tensorCNNRNNOutputs Figure 5:
Architecture of the RCNN model for a single sequence sample (based on [32] and
[33]). The input contains a tensor comprising a window of samples from each
data stream. The CNN comprises two convolutional layers, two pooling layers,
and a densely (fully) connected neural network. The CNN output is passed to an
LSTM cell of the RNN. The LSTM cell output is passed to a sigmoidal neuron
which outputs the class of the input window sample. The LSTM models temporal
dynamics over sample windows.
## VI Methodology
Both the models use a sliding window with zero overlap. The window of data is
presented to the classifier. The set of data in the window is classified to
belong either to a bag-emptying event or a non-bag-drop event. Each data
sample within the window is associated with this predicted class. During
training, a window may be slid to a position where it partially falls within a
bag-emptying event. The ground-truth class label for the window is calculated
as the average true class of all the samples in the window.
To validate models and results, a 6-fold cross validation test is performed.
The dataset is split into 6 equal data segments. Each data segment is a
continuous time series of 96831 samples. The model is trained on 5 of the data
segments and tested on the remaining segment. This is repeated such that the
model is tested on each data segment of the dataset.
Accuracy, precision, recall, and F-score are used to measure the performance
of the classifiers presented in this study. Accuracy describes the ratio of
the number of correct classifications to the total number of classified
samples. Precision describes the ratio of correct classifications to the total
number of classifications made for the particular class. It considers the
number of incorrectly predicted bag-emptying events and thus provides a
measure of the classifier quality. Recall describes the ratio of correct
classifications to the total number of items which truly belong to the
predicted class. It considers the number of bag-emptying event samples that
were missed. Recall thus provides a measure of the probability of correctly
classifying the bag-emptying event. Finally, the F-score is defined as the
harmonic mean between the precision and recall. It provides a measure to
describe both the precision and recall together.
## VII Results
### VII-A Comparison of Labelling Approaches
The median value of the accuracy, precision, recall, and F-score across the
6-fold cross validations for the predefined labels, manually defined labels,
and the learned labels are provided in Table I.
TABLE I: Feature-based ensemble model and RCNN model median value results for the predefined labels, manually defined labels, and the learned labels. Model | Measure | Predefined | Manual | Learned
---|---|---|---|---
Ensemble | Accuracy | 74% | 80% | 80%
Precision | 81% | 89% | 83%
Recall | 63% | 66% | 71%
F-score | 71% | 77% | 77%
RCNN | Accuracy | 79% | 76% | 86%
Precision | 86% | 80% | 89%
Recall | 71% | 72% | 83%
F-score | 77% | 75% | 84%
The ensemble model produces the poorest results with the predefined labels.
Improved performance is obtained with the manually and learned labels. The
performance for the manually and learned labels are similar. This is a key
result as it validates the proposed approach to learning the dataset labels.
The manual labels produce a higher precision than the learned labels. However,
the learned labels provide a higher recall than the manual labels. The recall
is considered to a more important measure as it relates to the probability of
detecting bag-emptying events.
(a) Ensemble model. (b) RCNN model.
Figure 6: Box whisker plot of the cross validation results.
The best results for the RCNN are obtained with the learned labels. This
reinforces the validation of the semi-supervised label learning approach. The
RCNN produces results that are much higher than those produced by the ensemble
model. The ensemble model’s precision for the manual labels matches the
capability of the RCNN. This however is achieved at the cost of a low recall
value of 66%. Along with the high precision value, the RCNN produces a recall
of 83%. These comparisons are reiterated by the 7% difference in F-score
results between the two models.
### VII-B Results with the Learned Labels
A box whisker plot of the results over the 6-fold cross validation test are
presented in Figure 6. The RCNN performs better than the ensemble model for
all performance measures. The precision box of the RCNN reaches high values.
However, the box is the largest of all boxes in the diagram. A larger box
indicates that there is more variability in the precision results. The quality
of the model is however still high considering that the range of the box
remains above 80%. The ensemble model’s recall box is large with whiskers that
extend to low values. This indicates a high level of uncertainty in the
ensemble models recall results. The recall box of the RCNN is narrow
indicating high certainty in the RCNN recall results. The range of the box
remains above 80% indicating good prediction ability of the RCNN model. Both
models produce narrow F-score boxes. The range of the RCNN F-score box is
higher than the ensemble model indicating superior performance overall.
A plot of the predictions and data for one of the cross validation folds is
presented in Figure 7. The learned labels correspond well with the changes in
the accelerometer and RSSI data. The predictions of the ensemble model are
sporadic resulting in several false positives and false negatives. The RCNN
model predictions are smoother over time. The RCNN however misses the third
bag-emptying event. This is possibly due to the RSSI level unexpectedly
dropping to a minimum during this bag-emptying event. Unlike the RCNN, the
ensemble model is able to detect the third bag-emptying event. This seems to
indicate that the RCNN relies more on RSSI data and the ensemble model relies
more on accelerometer data. Note that the last bag-emptying event is not
missed by the RCNN. It is detected in the following cross validation fold.
(a) Ensemble model. (b) RCNN model.
Figure 7: bag-emptying event predictions and dataset for the convolutional
recurrent neural network. The blue curves plot the dataset. The orange curve
the top figure plots the model predictions. A value of 1 indicates a bag-
emptying event.
The plots of the predictions for the remaining cross validation folds are
illustrated in Figure 8. The ensemble model results are more sporadic and the
predictions are more confident. The RCNN is less confident and the predictions
are smoother over time. This is preferred when false positives have a high
risk.
(a) (b)
(c) (d)
(e) (f)
(g) (h)
(i) (j)
Figure 8: Bag-emptying event predictions for the second to sixth cross
validation folds along the rows. Plots of the ensemble model are presented in
the left column and plots of the RCNN model are presented in the right column.
The horizontal axis is the sample number. The vertical axis is the probability
of a bag-emptying event. The blue curve plots the manual bag-emptying events
and orange curve plots the predicted bag-emptying events.
## VIII Discussion and Conclusion
In this study, we present a novel application on measuring fruit picker
productivity. The picker productivity is measured by detecting bag-emptying
events from wearable sensor data. A traditional feature-based ensemble model
and a deep convolutional recurrent neural network are applied to predict bag-
emptying events from the wearable sensor data. Furthermore, a semi-supervised
method for learning the bag-emptying event labels is presented.
Results indicate that both models are able to successfully detect the bag-
emptying events. The RCNN model is more accurate than the ensemble model but
it is less confident, which results in 3 of the 64 bag-emptying events being
missed. The ensemble model has at least one positive detection within each
bag-emptying event suggesting that all bag-emptying events. As the ensemble
model does not directly model any temporal relationship between samples, its
predictions are noisy resulting in multiple detections for a single bag-
emptying event. The RCNN models temporal dynamics to produces smooth
predictions over time and more accurate predictions of bag-emptying event
durations.
In future work, the RCNN could be improved by increasing its capacity and
training it on more data. The capacity of the RCNN can be increased by
introducing more CNN filters and by increasing the CNN depth. Such
improvements may provide the model with the capability to learn more advanced
features. The RNN can be improved by adding multiple layers and by using bi-
directional RNNs. More complex models however generally require more data.
Collecting more data is thus a priority for future work. Other than improving
models, the data captured by the sensors provide information relating to other
problems such as health and safety. For example, the accelerometer sensor can
be used to detect excessive bag weight or falls.
## References
* [1] L. Fedunik-Hofman, “Food waste: preventing a multi-billion dollar problem,” https://www.science.org.au/curious/earth-environment/food-waste-preventing-multi-billion-dollar-problem, 2023\.
* [2] M. Altalak, M. Ammad uddin, A. Alajmi, and A. Rizg, “Smart agriculture applications using deep learning technologies: A survey,” _Applied Sciences_ , vol. 12, no. 12, 2022.
* [3] S. Zhang, Y. Li, S. Zhang, F. Shahabi, S. Xia, Y. Deng, and N. Alshurafa, “Deep learning in human activity recognition with wearable sensors: A review on advances,” _Sensors_ , vol. 22, no. 4, p. 1476, 2022.
* [4] L. Benos, A. C. Tagarakis, G. Dolias, R. Berruto, D. Kateris, and D. Bochtis, “Machine learning in agriculture: A comprehensive updated review,” _Sensors_ , vol. 21, no. 11, 2021.
* [5] R. Sharma, S. S. Kamble, A. Gunasekaran, V. Kumar, and A. Kumar, “A systematic literature review on machine learning applications for sustainable agriculture supply chain performance,” _Computers & Operations Research_, vol. 119, p. 104926, 2020.
* [6] F. K. Anjom, S. G. Vougioukas, and D. C. Slaughter, “Development and application of a strawberry yield-monitoring picking cart,” _Computers and Electronics in Agriculture_ , vol. 155, pp. 400 – 411, 2018.
* [7] J. Whitney, Q. Ling, W. Miller, and T. A. Wheaton, “A dgps yield monitoring system for florida citrus,” _Applied Engineering in Agriculture_ , vol. 17, no. 2, p. 115, 2001.
* [8] D. Thomas, C. Perry, G. Vellidis, J. Durrence, L. Kutz, C. Kvien, B. Boydell, and T. Hamrita, “Development and implementation of a load cell yield monitor for peanut,” _Applied Engineering in Agriculture_ , vol. 15, no. 3, p. 211, 1999.
* [9] G. Pelletier and S. K. Upadhyaya, “Development of a tomato load/yield monitor,” _Computers and Electronics in Agriculture_ , vol. 23, no. 2, pp. 103 – 117, 1999.
* [10] W. Miller and J. Whitney, “Evaluation of weighing systems for citrus yield monitoring,” _Applied Engineering in Agriculture_ , vol. 15, no. 6, p. 609, 1999.
* [11] Y. Ampatzidis, S. Vougioukas, and M. Whiting, “A wearable module for recording worker position in orchards,” _Computers and Electronics in Agriculture_ , vol. 78, no. 2, pp. 222 – 230, 2011.
* [12] R. Arikapudi, S. G. Vougioukas, F. Jiménez-Jiménez, and F. K. Anjom, “Estimation of fruit locations in orchard tree canopies using radio signal ranging and trilateration,” _Computers and Electronics in Agriculture_ , vol. 125, pp. 160 – 172, 2016.
* [13] Y. Ampatzidis, L. Tan, R. Haley, and M. D. Whiting, “Cloud-based harvest management information system for hand-harvested specialty crops,” _Computers and Electronics in Agriculture_ , vol. 122, pp. 161 – 167, 2016\.
* [14] Y. G. Ampatzidis, M. D. Whiting, B. Liu, P. A. Scharf, and F. J. Pierce, “Portable weighing system for monitoring picker efficiency during manual harvest of sweet cherry,” _Precision Agriculture_ , vol. 14, no. 2, pp. 162–171, Apr 2013.
* [15] Y. G. Ampatzidis, M. D. Whiting, P. A. Scharf, and Q. Zhang, “Development and evaluation of a novel system for monitoring harvest labor efficiency,” _Computers and Electronics in Agriculture_ , vol. 88, pp. 85 – 94, 2012.
* [16] Y. Ampatzidis and S. Vougioukas, “Field experiments for evaluating the incorporation of rfid and barcode registration and digital weighing technologies in manual fruit harvesting,” _Computers and Electronics in Agriculture_ , vol. 66, no. 2, pp. 166 – 172, 2009.
* [17] O. D. Lara and M. A. Labrador, “A survey on human activity recognition using wearable sensors,” _IEEE Communications Surveys Tutorials_ , vol. 15, no. 3, pp. 1192–1209, Third 2013.
* [18] A. Anagnostis, L. Benos, D. Tsaopoulos, A. Tagarakis, N. Tsolakis, and D. Bochtis, “Human activity recognition through recurrent neural networks for human–robot interaction in agriculture,” _Applied Sciences_ , vol. 11, no. 5, 2021.
* [19] G. Aiello, P. Catania, M. Vallone, and M. Venticinque, “Worker safety in agriculture 4.0: A new approach for mapping operator’s vibration risk through machine learning activity recognition,” _Computers and Electronics in Agriculture_ , vol. 193, p. 106637, 2022.
* [20] L. Atallah, B. Lo, R. King, and G. Yang, “Sensor positioning for activity recognition using wearable accelerometers,” _IEEE Transactions on Biomedical Circuits and Systems_ , vol. 5, no. 4, pp. 320–329, Aug 2011.
* [21] N. Ravi, N. Dandekar, P. Mysore, and M. L. Littman, “Activity recognition from accelerometer data,” in _Proceedings of the 17th Conference on Innovative Applications of Artificial Intelligence - Volume 3_ , ser. IAAI’05. AAAI Press, 2005, pp. 1541–1546.
* [22] J. Wang, Y. Chen, S. Hao, X. Peng, and L. Hu, “Deep learning for sensor-based activity recognition: A survey,” _Pattern Recognition Letters_ , 2018.
* [23] I. Goodfellow, Y. Bengio, and A. Courville, _Deep learning_. MIT press, 2016.
* [24] M. Zeng, L. T. Nguyen, B. Yu, O. J. Mengshoel, J. Zhu, P. Wu, and J. Zhang, “Convolutional neural networks for human activity recognition using mobile sensors,” in _6th International Conference on Mobile Computing, Applications and Services_ , Nov 2014, pp. 197–205.
* [25] S.-M. Lee, S. M. Yoon, and H. Cho, “Human activity recognition from accelerometer data using convolutional neural network,” in _2017 IEEE International Conference on Big Data and Smart Computing (BigComp)_ , Feb 2017, pp. 131–134.
* [26] J. B. Yang, M. N. Nguyen, P. P. San, X. L. Li, and S. Krishnaswamy, “Deep convolutional neural networks on multichannel time series for human activity recognition,” in _Proceedings of the 24th International Conference on Artificial Intelligence_ , ser. IJCAI’15. AAAI Press, 2015, pp. 3995–4001.
* [27] A. Graves, _Supervised Sequence Labelling with Recurrent Neural Networks_. Berlin, Heidelberg: Springer Berlin Heidelberg, 2012.
* [28] A. Murad and J.-Y. Pyun, “Deep recurrent neural networks for human activity recognition,” _Sensors_ , vol. 17, no. 11, 2017.
* [29] M. Inoue, S. Inoue, and T. Nishida, “Deep recurrent neural network for mobile human activity recognition with high throughput,” _Artificial Life and Robotics_ , vol. 23, no. 2, pp. 173–185, Jun 2018.
* [30] Y. Guan and T. Plötz, “Ensembles of deep lstm learners for activity recognition using wearables,” _Proc. ACM Interact. Mob. Wearable Ubiquitous Technol._ , vol. 1, no. 2, pp. 11:1–11:28, Jun. 2017.
* [31] N. Y. Hammerla, S. Halloran, and T. Plötz, “Deep, convolutional, and recurrent models for human activity recognition using wearables,” in _Proceedings of the Twenty-Fifth International Joint Conference on Artificial Intelligence_ , ser. IJCAI’16. AAAI Press, 2016, pp. 1533–1540.
* [32] F. J. Ordóñez and D. Roggen, “Deep convolutional and lstm recurrent neural networks for multimodal wearable activity recognition,” _Sensors_ , vol. 16, no. 1, 2016.
* [33] S. Yao, S. Hu, Y. Zhao, A. Zhang, and T. Abdelzaher, “Deepsense: A unified deep learning framework for time-series mobile sensing data processing,” in _Proceedings of the 26th International Conference on World Wide Web_ , ser. WWW ’17. Republic and Canton of Geneva, Switzerland: International World Wide Web Conferences Steering Committee, 2017, pp. 351–360.
* [34] D. P. Kingma and J. Ba, “Adam: A method for stochastic optimization,” _arXiv preprint arXiv:1412.6980_ , 2014.
|
$\displaystyle\frac{1}{1+\frac{1}{\sigma_{t-1}^{2}}\mathcal{D}_{\mathcal{F}}^{2}(X_{t-1};X_{[t-2]},\sigma_{[t-2]})}\sup_{f_{1},f_{2}\in\mathcal{F}}\frac{(f_{1}(X_{t})-f_{2}(X_{t}))^{2}}{\sum_{s=1}^{t-2}\frac{1}{\sigma_{s}^{2}}(f_{1}(X_{s})-f_{2}(X_{s}))^{2}+\lambda}$
$\displaystyle={}$
$\displaystyle\frac{1}{1+\frac{1}{\sigma_{t-1}^{2}}\mathcal{D}_{\mathcal{F}}^{2}(X_{t-1};X_{[t-2]},\sigma_{[t-2]})}\mathcal{D}_{\mathcal{F}}^{2}(X_{t};X_{[t-2]},\sigma_{[t-2]})\geq\dots$
$\displaystyle\geq{}$
$\displaystyle\frac{1}{\prod_{s=t_{0}+1}^{t-1}(1+\frac{1}{\sigma_{s}^{2}}\mathcal{D}_{\mathcal{F}}^{2}(X_{s};X_{[s-1]},\sigma_{[s-1]}))}\mathcal{D}_{\mathcal{F}}^{2}(X_{t};X_{[t_{0}]},\sigma_{[t_{0}]}),$
where the first inequality holds due to
$\displaystyle\sum_{s=1}^{t-1}\frac{1}{\sigma_{s}^{2}}(f_{1}(X_{s})-f_{2}(X_{s}))^{2}$
$\displaystyle={}$
$\displaystyle\sum_{s=1}^{t-2}\frac{1}{\sigma_{s}^{2}}(f_{1}(X_{s})-f_{2}(X_{s}))^{2}+\frac{1}{\sigma_{t-1}^{2}}(f_{1}(X_{t-1})-f_{2}(X_{t-1}))^{2}$
$\displaystyle\leq{}$
$\displaystyle\sum_{s=1}^{t-2}\frac{1}{\sigma_{s}^{2}}(f_{1}(X_{s})-f_{2}(X_{s}))^{2}+\frac{1}{\sigma_{t-1}^{2}}\mathcal{D}_{\mathcal{F}}^{2}(X_{t-1};X_{[t-2]},\sigma_{[t-2]})\sum_{s=1}^{t-2}\frac{1}{\sigma_{s}^{2}}(f_{1}(X_{s})-f_{2}(X_{s}))^{2}$
$\displaystyle={}$
$\displaystyle\left(1+\frac{1}{\sigma_{t-1}^{2}}\mathcal{D}_{\mathcal{F}}^{2}(X_{t-1};X_{[t-2]},\sigma_{[t-2]})\right)\sum_{s=1}^{t-2}\frac{1}{\sigma_{s}^{2}}(f_{1}(X_{s})-f_{2}(X_{s}))^{2}.$
Thus, we have
$\displaystyle\mathcal{D}_{\mathcal{F}}^{2}(X_{t};X_{[t_{0}]},\sigma_{[t_{0}]})$
$\displaystyle\leq\prod_{s=t_{0}+1}^{t-1}\left(1+\frac{1}{\sigma_{s}^{2}}\mathcal{D}_{\mathcal{F}}^{2}(X_{s};X_{[s-1]},\sigma_{[s-1]})\right)\mathcal{D}_{\mathcal{F}}^{2}(X_{t};X_{[t-1]},\sigma_{[t-1]})$
$\displaystyle\leq\exp\left\\{\sum_{s=t_{0}+1}^{t-1}\frac{1}{\sigma_{s}^{2}}\mathcal{D}_{\mathcal{F}}^{2}(X_{s};X_{[s-1]},\sigma_{[s-1]})\right\\}\mathcal{D}_{\mathcal{F}}^{2}(X_{t};X_{[t-1]},\sigma_{[t-1]}),$
where the second inequality holds due to the inequality $1+x\leq\exp\\{x\\}$.
∎
###### Lemma H.5.
Let $\\{\sigma_{t},\beta_{t}\\}_{t\geq 1}$ be a sequence of non-negative
numbers, $\sigma_{\mathrm{min}},\gamma,\lambda>0$, $\\{X_{t}\\}_{t\geq
1}\subset\mathcal{X}$ and $\\{\bar{\sigma}_{k}\\}_{k\geq 1}$ be recursively
defined:
$\bar{\sigma}_{t}^{2}=\max\\{\sigma_{t}^{2},\sigma_{\mathrm{min}}^{2},\gamma^{2}\mathcal{D}_{\mathcal{F}}(X_{t};X_{[t-1]},\sigma_{[t-1]})\\}.$
Then we have
$\sum_{t=1}^{T}\min\\{1,\beta_{t}\mathcal{D}_{\mathcal{F}}(X_{t};X_{[t-1]},\sigma_{[t-1]})\\}\leq\dim_{\mathcal{F}}+\max_{t\in[T]}\beta_{t}\cdot\gamma^{2}\dim_{\mathcal{F}}+\sqrt{\dim_{\mathcal{F}}}\cdot\sqrt{\sum_{t=1}^{T}\beta_{t}^{2}(\sigma_{t}^{2}+\sigma_{\mathrm{min}}^{2})},$
where $\mathcal{D}_{\mathcal{F}}$ and
$\dim_{\mathcal{F}}=\dim_{\mathcal{F}}(\sigma_{\mathrm{min}},T)$ are in
Definition 3.2.
###### Proof.
We decompose $[T]$ as the union of three disjoint sets
$\mathcal{J}_{1},\mathcal{J}_{2},\mathcal{J}_{3}$:
$\displaystyle\mathcal{J}_{1}$
$\displaystyle=\left\\{t\in[T]\Big{|}\frac{1}{\bar{\sigma}_{t}}\mathcal{D}_{\mathcal{F}}(X_{t};X_{[t-1]},\sigma_{[t-1]})>1\right\\},$
$\displaystyle\mathcal{J}_{2}$
$\displaystyle=\left\\{t\in[T]\Big{|}\frac{1}{\bar{\sigma}_{t}}\mathcal{D}_{\mathcal{F}}(X_{t};X_{[t-1]},\sigma_{[t-1]})\leq
1,\bar{\sigma}_{t}\in\\{\sigma_{t},\sigma_{\mathrm{min}}\\}\right\\},$
$\displaystyle\mathcal{J}_{3}$
$\displaystyle=\left\\{t\in[T]\Big{|}\frac{1}{\bar{\sigma}_{t}}\mathcal{D}_{\mathcal{F}}(X_{t};X_{[t-1]},\sigma_{[t-1]})\leq
1,\bar{\sigma}_{t}=\gamma\sqrt{\mathcal{D}_{\mathcal{F}}(X_{t};X_{[t-1]},\sigma_{[t-1]})}\right\\}.$
For the summation over $\mathcal{J}_{1}$, we have
$\sum_{t\in\mathcal{J}_{1}}\min\\{1,\beta_{t}\mathcal{D}_{\mathcal{F}}(X_{t};X_{[t-1]},\sigma_{[t-1]})\\}\leq|\mathcal{J}_{1}|\leq\sum_{t\in\mathcal{J}_{1}}\min\left\\{1,\frac{1}{\bar{\sigma}_{t}^{2}}\mathcal{D}_{\mathcal{F}}^{2}(X_{t};X_{[t-1]},\sigma_{[t-1]})\right\\}\leq\dim_{\mathcal{F}}.$
Next, for the summation over $\mathcal{J}_{2}$, we have
$\displaystyle\sum_{t\in\mathcal{J}_{2}}\min\\{1,\beta_{t}\mathcal{D}_{\mathcal{F}}(X_{t};X_{[t-1]},\sigma_{[t-1]})\\}$
$\displaystyle\leq{}$
$\displaystyle\sum_{t\in\mathcal{J}_{2}}\beta_{t}\bar{\sigma}_{t}\cdot\frac{1}{\bar{\sigma}_{t}}\mathcal{D}_{\mathcal{F}}(X_{t};X_{[t-1]},\sigma_{[t-1]})\\}$
$\displaystyle\leq{}$
$\displaystyle\sum_{t=1}^{T}\beta_{t}\max\\{\sigma_{t},\sigma_{\mathrm{min}}\\}\frac{1}{\bar{\sigma}_{t}}\mathcal{D}_{\mathcal{F}}(X_{t};X_{[t-1]},\sigma_{[t-1]})\\}$
$\displaystyle\overset{(a)}{\leq}{}$
$\displaystyle\sqrt{\sum_{t=1}^{T}\beta_{t}^{2}(\sigma_{t}^{2}+\sigma_{\mathrm{min}}^{2})}\cdot\sqrt{\sum_{t=1}^{T}\min\left\\{1,\frac{1}{\bar{\sigma}_{t}^{2}}\mathcal{D}_{\mathcal{F}}^{2}(X_{t};X_{[t-1]},\sigma_{[t-1]})\right\\}}$
$\displaystyle\leq{}$
$\displaystyle\sqrt{\dim_{\mathcal{F}}}\cdot\sqrt{\sum_{t=1}^{T}\beta_{t}^{2}(\sigma_{t}^{2}+\sigma_{\mathrm{min}}^{2})},$
where $(a)$ holds due to Cauchy-Schwartz inequality and
$\frac{1}{\bar{\sigma}_{t}}\mathcal{D}_{\mathcal{F}}(X_{t};X_{[t-1]},\sigma_{[t-1]})\leq
1$. Then, for the summation over $\mathcal{J}_{3}$, we have
$\displaystyle\sum_{t\in\mathcal{J}_{3}}\min\\{1,\beta_{t}\mathcal{D}_{\mathcal{F}}(X_{t};X_{[t-1]},\sigma_{[t-1]})\\}$
$\displaystyle\leq{}$
$\displaystyle\sum_{t\in\mathcal{J}_{3}}\beta_{t}\bar{\sigma}_{t}\cdot\frac{1}{\bar{\sigma}_{t}}\mathcal{D}_{\mathcal{F}}(X_{t};X_{[t-1]},\sigma_{[t-1]})$
$\displaystyle\overset{(a)}{\leq}{}$
$\displaystyle\max_{t\in[T]}\beta_{t}\cdot\gamma^{2}\sum_{t=1}^{T}\min\left\\{1,\frac{1}{\bar{\sigma}_{t}^{2}}\mathcal{D}_{\mathcal{F}}^{2}(X_{t};X_{[t-1]},\sigma_{[t-1]})\right\\}$
$\displaystyle\leq{}$
$\displaystyle\max_{t\in[T]}\beta_{t}\cdot\gamma^{2}\dim_{\mathcal{F}},$
where $(a)$ holds due to
$\bar{\sigma}_{t}=\gamma^{2}\frac{1}{\bar{\sigma}_{t}}\mathcal{D}_{\mathcal{F}}(X_{t};X_{[t-1]},\sigma_{[t-1]})$
and
$\frac{1}{\bar{\sigma}_{t}}\mathcal{D}_{\mathcal{F}}(X_{t};X_{[t-1]},\sigma_{[t-1]})\leq
1$. Finally, putting pieces together finishes the proof. ∎
###### Lemma H.6 (Modified from Lemma 2 in Zhang et al. 2021a).
Let $\lambda_{1},\lambda_{2},\lambda_{4}>0$, $\lambda_{3}\geq 1$ and
$i^{\prime}=\lceil\log_{2}\lambda_{1}\rceil$. Let
$a_{0},a_{1},a_{2},\dots,a_{i^{\prime}}$ be non-negative reals such that
$a_{i}\leq\lambda_{1}$ for any $0\leq i\leq i^{\prime}$, and
$a_{i}\leq\lambda_{2}\sqrt{a_{i+1}+2^{i+1}\cdot\lambda_{3}}+\lambda_{4}$ for
any $0\leq i<i^{\prime}$. Then we have
$\displaystyle a_{0}$
$\displaystyle\leq\max\left\\{\left(\lambda_{2}+\sqrt{\lambda_{2}^{2}+\lambda_{4}}\right)^{2},\lambda_{2}\sqrt{4\lambda_{3}}+\lambda_{4}\right\\}\leq\lambda_{2}\sqrt{4\lambda_{3}}+4\lambda_{2}^{2}+3\lambda_{4},$
$\displaystyle a_{1}$
$\displaystyle\leq\max\left\\{\left(\lambda_{2}+\sqrt{\lambda_{2}^{2}+\lambda_{4}}\right)^{2},\lambda_{2}\sqrt{8\lambda_{3}}+\lambda_{4}\right\\}\leq\lambda_{2}\sqrt{8\lambda_{3}}+4\lambda_{2}^{2}+3\lambda_{4}.$
|
# (N)$\text{NLO}+\text{NLL}^{\prime}$ accurate predictions for plain and
groomed 1-jettiness in neutral current DIS
Max Knobbe111Email<EMAIL_ADDRESS>Institut für Theoretische
Physik, Georg-August-Universität Göttingen, Friedrich-Hund-Platz 1, 37077
Göttingen, Germany Daniel Reichelt222Email<EMAIL_ADDRESS>Institute for Particle Physics Phenomenology, Department of Physics, Durham
University, Durham DH1 3LE, United Kingdom Steffen Schumann555Email:
<EMAIL_ADDRESS>Institut für Theoretische Physik,
Georg-August-Universität Göttingen, Friedrich-Hund-Platz 1, 37077 Göttingen,
Germany
###### Abstract
The possibility to reanalyse data taken by the HERA experiments offers the
chance to study modern QCD jet and event-shape observables in deep-inelastic
scattering. To address this, we compute resummed and matched predictions for
the 1-jettiness distribution in neutral current DIS with and without grooming
the hadronic final state using the soft-drop technique. Our theoretical
predictions also account for non-perturbative corrections from hadronisation
through parton-to-hadron level transfer matrices extracted from dedicated
Monte Carlo simulations with S HERPA . To estimate parameter uncertainties in
particular for the beam-fragmentation modelling we derive a family of replica
tunes to data from the HERA experiments. While NNLO QCD normalisation
corrections to the NLO+NLL’ prediction are numerically small, hadronisation
corrections turn out to be quite sizeable. However, soft-drop grooming
significantly reduces the impact of non-perturbative contributions. We
supplement our study with hadron-level predictions from S HERPA based on the
matching of NLO QCD matrix elements with the parton shower. Good agreement
between the predictions from the two calculational methods is observed.
###### Contents
1. 1 Introduction
2. 2 Phase space and observable definition
3. 3 DIS Monte Carlo simulations with S HERPA
1. 3.1 MEPS@NLO predictions for DIS
2. 3.2 Tuning the beam fragmentation model against HERA data
4. 4 (N)$\text{NLO}+\text{NLL}^{\prime}$ resummation for 1-jettiness in DIS
1. 4.1 NLL resummation in the C AESAR approach
2. 4.2 Grooming in DIS
3. 4.3 Calculational tools and setup
5. 5 Results for (groomed) 1-jettiness in DIS
6. 6 Conclusions
7. A Tuning details
## 1 Introduction
Event shape observables offer great potential for detailed studies of the
intriguing dynamics of Quantum Chromodynamics (QCD), thereby providing insight
into various strong interaction phenomena. For example, they offer sensitivity
to the strong coupling constant $\alpha_{S}$, the colour charges of the QCD
quanta, and parton density functions, when considering hadronic initial state
particles. Predictions for event shape distributions can be obtained from
fixed-order perturbation theory, all-orders resummation of logarithmically
enhanced contributions, as well as detailed particle-level simulations as
provided by Monte Carlo event generators. Accordingly, they form a rather
unique testbed for a variety of theoretical approaches, ranging from cutting-
edge multi-loop calculations to detailed aspects in the modelling of the non-
perturbative parton-to-hadron transition.
Event shapes have played a central role in the QCD measurement program of past
$e^{+}e^{-}$ collider experiments, see for instance [1, 2, 3, 4, 5]. Also at
hadron–hadron machines they are considered in studies of hadronic final
states. Possibly even more prominently, closely related jet-substructure
observables have attracted enormous attention and sparked the development of
modern grooming and tagging techniques, see Ref. [6] for a recent review. Also
in deep-inelastic lepton–nucleon scattering experiments several event shape
variables have been measured [7, 8, 9, 10, 11, 12]. However, the LEP and HERA
experiments phased out in the years 2000 and 2007, respectively, such that
later breakthroughs in calculational methods and modern observable definitions
have not yet been fully exploited.
Their complementarity and partially reduced complexity when compared to
present day LHC measurements, make the LEP and HERA data a real treasure for
additional tests of our theoretical understanding and simulation capabilities.
In the past years a small number of re-analyses of the LEP data have been
published, see for instance [13, 14, 15, 16]. Furthermore, there are efforts
to provide open data sets that can directly be used by the entire community
[17, 18].
To open the treasure chest of their large data set for modern QCD studies the
HERA H1 collaboration has recently started to publish a series of new,
fascinating measurements that allow one to confront contemporary state-of-the-
art predictions with precise DIS data. Besides their relevance for
benchmarking our present day tools, such analyses build an important stepping
stone towards future electron–hadron colliders like the EIC at BNL [19, 20] or
the LHeC at CERN [21, 22].
We here compile predictions for the 1-jettiness event shape in the Breit frame
[23], that is equivalent to the well known thrust variable [24], for the HERA
kinematics, _i.e._ lepton–proton collisions at $\sqrt{s}=319\;\text{GeV}$.
Furthermore, we consider grooming of the hadronic final states based on the
soft-drop method prior to the observable evaluation. We derive differential
distributions for groomed and ungroomed $\tau^{b}_{1}$ differential in the
photon virtuality $Q^{2}\in[150,20000]\;\text{GeV}^{2}$, and the events
inelasticity $y\in[0.05,0.94]$. We perform Monte Carlo simulations with the S
HERPA generator based on next-to-leading-order (NLO) matrix elements for the
one- and two-jet final states matched to the parton shower and hadronised
using S HERPA ’s new cluster fragmentation model [25]. To estimate the
hadronisation modelling uncertainties in particular related to the beam
remnant fragmentation we derive a set of replica tunes [26] to a selection of
DIS measurements from the H1 and ZEUS experiments.
Furthermore, we compute resummed predictions at next-to-leading-logarithmic
(NLL) accuracy in the observable value based on the implementation of the C
AESAR resummation formalism [27] in the S HERPA framework [28]. These get
matched to the NNLO QCD result for the inclusive DIS process and the NLO
matrix elements for the two-jet channel. For the NNLO QCD corrections we rely
on an implementation in S HERPA presented in [29]. This results in
predictions of $\text{NLO}+\text{NLL}^{\prime}$ accuracy for the actual event-
shape distributions, while we achieve NNLO precision for the total event rate.
In consequence, we refer to our predictions as being
(N)$\text{NLO}+\text{NLL}^{\prime}$ accurate. To account for non-perturbative
corrections we derive parton-to-hadron level transfer matrices differential in
the event shape variables that we extract from particle level simulations with
S HERPA [30], thereby also accounting for the cluster-model parameter
uncertainties through the set of replica tunes to HERA data.
Our calculations are targeted on an upcoming measurement by the H1 experiment,
for that preliminary results have recently been presented [31, 32]. Results
based on simulations with S HERPA in a similar fiducial phase space have been
compared to data from jet-substructure observables in neutral current DIS in
[33]. Our study extends earlier work on the simulation of DIS events with S
HERPA [34]. Furthermore, this is the first time we include NNLO QCD correction
in resummation calculations with S HERPA .
The article is organised as follows: in Sec. 2 we introduce the considered
observables and define the fiducial phase space used in our study of the
hadronic final states produced in $ep$ collisions at HERA. In Sec. 3 we
describe the setup used to simulate DIS events with S HERPA as well as the
tuning of its beam-fragmentation parameters. In Sec. 4 we present our
framework to compile (N)$\text{NLO}+\text{NLL}^{\prime}$ predictions, based on
the implementation of the C AESAR formalism in S HERPA . Here, we also
present our approach to treat non-perturbative corrections based on transfer
matrices extracted from MC simulations, see Sec. 4.1. We present our final
(N)$\text{NLO}+\text{NLL}^{\prime}+\text{NP}$ results in Sec. 5, alongside
with MC predictions from S HERPA . We compile our conclusions and give an
outlook in Sec. 6.
## 2 Phase space and observable definition
We consider deep-inelastic scattering (DIS) of leptons with momentum $p$ of
off protons with momentum $P$ at HERA energies, _i.e._
$E_{l}=27.6\;\text{GeV}$ and $E_{p}=920\;\text{GeV}$, resulting in a centre-
of-mass energy of $\sqrt{s}=319\;\text{GeV}$. Denoting the outgoing lepton
momentum as $p^{\prime}$, we define the momentum difference, at LO carried by
the virtual photon, as
$q=p-p^{\prime}\equiv(0,0,0,-Q)~{},$ (1)
where the last equivalence defines the Breit frame, which we will assume
whenever frame-specific formulae are given. We also introduce the usual
Bjorken variable $x_{B}$ and inelasticity $y$
$\displaystyle x_{B}$ $\displaystyle=\frac{Q^{2}}{2P\cdot q}\,,$ (2)
$\displaystyle y$ $\displaystyle=\frac{P\cdot q}{P\cdot p}~{}.$ (3)
We consider events with $150<Q^{2}/\text{GeV}^{2}<2\cdot 10^{4}$ and
$0.05<y<0.94$. No other cuts are applied, but we have studied 1-jettiness in
smaller bins of $Q^{2}$ and $y$, and will only discuss a selection of results
here***Results over the full range of $Q^{2},y$ in several bins of both
variables are available upon request..
We take into account all final state particles apart from the outgoing lepton
for the calculation of event-shape variables. We study a well known
observable, referred to as thrust $\tau_{Q}$ [24] or alternatively 1-jettiness
$\tau^{b}_{1}$ [23]. Several equivalent definitions exist in the literature.
For concreteness we define it by dividing the event into a current hemisphere
$\mathcal{H}_{C}$ and a beam hemisphere $\mathcal{H}_{B}$. Working in the
Breit frame, we can introduce two reference vectors
$n_{\pm}=(1,0,0,\pm 1)$ (4)
and denote the hemispheres according to the final state particles momentum
fractions along those,
$\displaystyle\mathcal{H}_{C}=\\{p_{i}:p_{i}\cdot n_{+}>p_{i}\cdot
n_{-}\\}\quad\text{and}\quad\mathcal{H}_{B}=\\{p_{i}:p_{i}\cdot
n_{+}<p_{i}\cdot n_{-}\\}~{}.$ (5)
We can now define thrust as the sum of the longitudinal momentum components of
all particles in the current hemisphere. As we prefer to work with an
observable that vanishes in the soft limit, _i.e._ the limit where all final
state partons apart from the struck quark have arbitrarily small momenta, we
ultimately use
$\tau=1-\frac{2}{Q}\sum_{p_{i}\in\mathcal{H}_{C}}p_{i}^{z}\,.$ (6)
Despite this definition only summing over one of the hemispheres, thrust,
_i.e._ 1-jettiness, is actually sensitive to emissions anywhere in the event,
and indeed is a global event shape in the sense of _e.g._ [27]. Note this
statement depends on the precise definition, including the normalisation
factor here given by $Q/2$, that differs in the thrust variant we use for
tuning in the following.
In addition we study 1-jettiness calculated based on events that have been
groomed of soft wide-angle radiation. Soft-drop grooming was first introduced
in [35] as a jet substructure technique, including as a special case the
modified Mass Drop Tagger [36, 37]. It has since been generalised and applied
also to jets at lepton colliders [38, 18] and event shapes at both lepton [38,
39] and hadron [40] colliders. A version applicable to DIS was proposed in
[41], based on the C ENTAURO jet algorithm [42], that accounts for the
forward-backward asymmetry when considering the Breit frame. This sequential
cluster algorithm is based on the distance measure between particles with
momenta $p_{i},p_{j}$
$\displaystyle d_{ij}$
$\displaystyle=(\Delta\bar{z}_{ij})^{2}+2\bar{z}_{i}\bar{z}_{j}(1-\cos\Delta\phi_{ij})\,,$
(7) $\displaystyle\text{with}\;\;\bar{z}_{i}$
$\displaystyle=2\sqrt{1+\frac{q\cdot p_{i}}{x_{B}P\cdot
p_{i}}}\quad\text{and}\quad~{}\Delta\bar{z}_{ij}=\bar{z}_{i}-\bar{z}_{j}\,.$
(8)
Note that [42] discusses more general functional forms of the distance
measure, while we concentrate here on the definition given in [41]. As in all
other soft-drop grooming methods the objects of interest, in this case the
full event, are first clustered according to this sequential algorithm, and
then the reverse clustering history is considered. The last cluster step is
undone, and the softness of the softer of the two branches is evaluated. For
the DIS case, [41] suggests to use
$z_{i}=\frac{P\cdot p_{i}}{P\cdot q}$ (9)
as a measure for softness. The formal soft-drop criterion then reads
$\frac{\min[z_{i},z_{j}]}{z_{i}+z_{j}}>z_{\text{cut}}~{},$ (10)
with $z_{\text{cut}}$ the grooming parameter. If this is satisfied, _i.e._
both branches are classified as hard, the algorithm terminates. Otherwise the
softer branch (with smaller $z$) is dropped, and the procedure is repeated
with the harder branch. This iteration stops when either Eq. (10) is
satisfied, or there is only one particle left in the hard branch such that no
further unclustering is possible.
We finally recalculate 1-jettiness, using Eq. (6) but restricting the sum to
particles in the current hemisphere that have not been dropped during
grooming, thereby considering variable values for $z_{\text{cut}}$.
## 3 DIS Monte Carlo simulations with S HERPA
We derive hadron-level predictions for the DIS event shapes using a pre-
release version of S HERPA -3.0 [43], that will supersede the current S
HERPA -2.2 series [44]. This major release features extended physics-modelling
capabilities, including, for example, the automated evaluation of electroweak
(EW) corrections at the one-loop order [45, 46, 47] or in the Sudakov
approximation [48, 49], a complete reimplementation of the cluster
hadronisation model [25], as well as an improved user interface based on Yaml
[50]. To analyse our simulated event samples we employ the R IVET analysis
package [51]. For jet clustering we use the C ENTAURO plugin [42] within the
F AST J ET framework [52].
### 3.1 MEPS@NLO predictions for DIS
The basics of simulating DIS processes by merging parton-shower evolved
higher-multiplicity tree-level matrix elements within the S HERPA framework
have been presented in [34]. We here lift this to next-to-leading order (NLO)
accurate QCD matrix elements. To this end, we consider the massless single and
dijet production channels in neutral current DIS at NLO, and three- and four-
jets at leading order (LO), _i.e._
$e^{-}p\to e^{-}+1,2\,j\,@\,\text{NLO}+3,4\,j\,@\,\text{LO},$ (11)
where we consider $u,d,s$ quarks to be massless and add additional LO
processes for the remaining massive quarks. The massless and massive channels
get matched to the S HERPA Catani–Seymour dipole shower [53] and merged
according to the MEPS@NLO [54] and MEPS@LO [55] truncated shower formalism,
respectively. The contributing one-loop amplitudes are obtained from O PEN L
OOPS [56], that employs the C OLLIER library [57] for the evaluation of
tensor and scalar integrals. All tree-level matrix elements are provided by C
OMIX [58], and PDFs are obtained from LHAPDF [59].
To determine the perturbative scales entering the calculation, the final
states of the multi-parton final states get clustered to a two-to-two core
process [55]. For the reconstructed core the factorisation, renormalisation,
and parton shower starting scale are set to
$\displaystyle\mu_{\text{F}}=\mu_{\text{R}}=\mu_{\text{Q}}:=\mu_{\text{core}}\,.$
(12)
For jet-associated DIS three configurations need to be distinguished [34]:
1. (i)
virtual photon exchange, _i.e._ $ej\to ej$, where
$\mu_{\text{core}}^{2}=Q^{2}$,
2. (ii)
interaction of the virtual photon with a QCD parton, _i.e._ $\gamma^{*}j\to
j_{1}j_{2}$, with $\mu^{2}_{\text{core}}=m_{\perp,1}m_{\perp,2}$ defined as
the product of the two jet transverse masses
$m_{\perp,i}=\sqrt{m^{2}_{i}+p_{\perp,i}^{2}}$ relative to the beam axis,
3. (iii)
and pure QCD channels, _i.e._ $jj\to jj$, where
$\mu^{2}_{\text{core}}=-\frac{1}{\sqrt{2}}\left(s^{-1}+t^{-1}+u^{-1}\right)^{-1}$
is a scaled harmonic mean of the Mandelstam variables $s,t,u$.
Beyond the core process, the arguments of the strong-coupling factors are
determined by the clustering algorithm [55]. The merging-scale parameter,
separating the different jet-multiplicity contributions, is dynamically set to
$Q_{\text{cut}}=\frac{\bar{Q}_{\text{cut}}}{\sqrt{1+\bar{Q}^{2}_{\text{cut}}/Q^{2}}}\,,\quad\text{using}\quad\bar{Q}_{\text{cut}}=5\,\text{GeV}\,.$
(13)
As parton density functions we use the NNLO PDF4LHC21_40_pdfas set [60] with
$\alpha_{S}(M^{2}_{Z})$=0.118.
To estimate perturbative uncertainties, we consider 7-point variations of the
factorisation ($\mu_{F}$) and renormalisation ($\mu_{R}$) scales in the matrix
element and the parton shower that get evaluated on-the-fly [61], _i.e._
$\\{(\tfrac{1}{2}\mu_{\text{R}},\tfrac{1}{2}\mu_{\text{F}}),(\tfrac{1}{2}\mu_{\text{R}},\mu_{\text{F}}),(\mu_{\text{R}},\tfrac{1}{2}\mu_{\text{F}}),(\mu_{\text{R}},\mu_{\text{F}}),(\mu_{\text{R}},2\mu_{\text{F}}),(2\mu_{\text{R}},\mu_{\text{F}}),(2\mu_{\text{R}},2\mu_{\text{F}})\\}\,.$
(14)
The resummation scale $\mu_{Q}$ we keep fixed. The final uncertainty estimate
is derived by forming an envelope of all variations.
### 3.2 Tuning the beam fragmentation model against HERA data
Ref. [25] presented a new cluster fragmentation model for S HERPA that will
be used in S HERPA -3, superseding the old cluster model described in [62],
that was used in the S HERPA -1.X [63] and S HERPA -2.X [44] released. A
particular feature of the new implementation is a specific treatment of the
fragmentation of hadronic clusters that contain beam remnant particles. To
calibrate the corresponding model parameters we performed dedicated tunes
using HERA data for hadronic final state observables in neutral current DIS.
Broadly speaking, a cluster hadronisation simulation features two basic
components, a cluster-formation and a cluster-decay model [64, 65]. Based on
the pre-confinement property of QCD [66], finite mass colour neutral mesonic
and baryonic clusters can be formed from the final state of a parton shower
evolution of a hard scattering event. These primary clusters are then subject
to an iterative fission process that ultimately results in the transition to
known hadronic resonances, whose decays can be treated by a dedicated package.
Both elements of the hadronisation model introduce sets of parameters that
need to be carefully adjusted by comparing model predictions and measurements
for suitable observables, a process commonly known as tuning.
In Ref. [26] the free model parameters were calibrated against hadronic
observables measured in electron–positron annihilation experiments. However,
in leptonic collisions the beam fragmentation modelling is not probed and the
corresponding parameters remained unconstrained. This affects in particular
the parametrisation of the decay of clusters that contain a remnant particle
of an incident hadron, _e.g._ a (anti-)quark and (anti-)diquark from the
break-up of the incoming proton in DIS. We consider the two-body decay of a
beam cluster with flavours $f_{1}$ and $\bar{f}_{2}$, where a (di)quark-
flavour pair $f\bar{f}$ is drawn from the vacuum, resulting in
${\cal{C}}[f_{1}\bar{f}_{2}]\to{\cal{C}}_{1}[f_{1}\bar{f}]\;{\cal{C}}_{2}[f\bar{f}_{2}]\,.$
(15)
To fix the kinematics of the two-body decay in the rest frame of ${\cal{C}}$,
the absolute value of the transverse momentum of the decay products
${\cal{C}}_{1}$ and ${\cal{C}}_{2}$ is selected according to a Gaussian
distribution ${\cal{N}}(0,k^{2}_{T,0}/2)$ that is truncated at the parton-
shower cut-off $p_{T,\text{min}}$, _i.e._
${\cal{P}}(k_{T})\propto\exp\left(-k_{T}^{2}/k_{T,0}^{2}\right)\Theta(p^{2}_{T,\text{min}}-k_{T}^{2})\,.$
(16)
The parameter $k_{T,0}$ is thereby considered as independent of the incident
cluster type. The direction of the two-component $\vec{k}_{T}$ is picked
uniformly in the transverse plane, with $f_{1}$ and $\bar{f}_{2}$ pointing
along the positive and negative $z$-axis, respectively. This leaves one to fix
the longitudinal momentum fractions $z^{(1),(2)}$ with respect to the light-
like vectors $n^{\mu}_{\pm}=(1,0,0,\pm 1)$. For the case of a beam-remnant
cluster, still working in its rest frame, these are distributed according to
${\cal{P}}(z)\propto
z^{\alpha_{B}}(1-z)^{\beta_{B}}\cdot\exp\left\\{-\gamma_{B}\frac{1}{z}\left(\frac{k^{2}_{T}+(m_{f_{1}}+m_{\bar{f}_{2}})^{2}}{k^{2}_{T,0}}\right)\right\\}\,.$
(17)
Note the similarity to the symmetric Lund string fragmentation function [67].
This results in the four-momenta of the decay products being given by
$\displaystyle p^{\mu}_{{\cal{C}}_{1}}$ $\displaystyle=$
$\displaystyle\frac{m_{{\cal{C}}}}{2}\left(z^{(1)}n^{\mu}_{+}+(1-z^{(2)})n^{\mu}_{-}\right)+k^{\mu}_{T}\,,$
(18) $\displaystyle p^{\mu}_{{\cal{C}}_{2}}$ $\displaystyle=$
$\displaystyle\frac{m_{{\cal{C}}}}{2}\left((1-z^{(1)})n^{\mu}_{+}+z^{(2)}n^{\mu}_{-}\right)-k^{\mu}_{T}\,.$
(19)
According to Eq. (17) the relevant free parameters specifically steering the
decays of beam clusters are $\alpha_{B}$, $\beta_{B}$, and $\gamma_{B}$. To
calibrate those we performed dedicated tunes based on a variety of hadronic
observables measured by the HERA experiments H1 and ZEUS. The remaining
hadronisation parameters are set according to the LEP data tune described in
Ref. [26].
We employ the A PPRENTICE tuning tool [68], with reference data for DIS
analyses at centre of mass energies of $\sqrt{s}=$300\text{\,}\mathrm{GeV}$$,
_i.e_. lepton energies of $27.5\text{\,}\mathrm{GeV}$ and proton energies of
$820\text{\,}\mathrm{GeV}$. The tuning requires an initial set of Monte Carlo
runs, that are then used to generate a polynomial, bin-wise approximation of
the Monte Carlo response with respect to changes in the hadronisation-model
parameters. The predictions for the grid points are generated using the
calculational setup described in Sec. 3.1.
The selection of observables considered for the tuning includes classic
variables sensitive to hadronisation. In particular, we use event-shape
distributions like thrust and jet broadening [9], energy flows and charged
particle spectra [69, 70] and multiplicities [71, 72], as well as quark
fragmentation functions [73, 74]. Further details on the used analyses and
observables are provided in App. A.
Given we consider model parameters newly introduced that have not been tuned
before, we have little prior knowledge about their preferred values and thus
need to start out with rather wide parameter ranges. To narrow these down, we
make an initial pass to get a rough idea of the relevant regions. The
corresponding ranges are outlined in Tab. 1. For a second run we restrict the
tuning ranges using the results of the exploration run, resulting in an
iterative procedure to further narrow down the considered parameter intervals.
The initial run, with largely unconstrained parameter values also serves the
purpose of filtering out the most sensitive observables from the considered
analyses. Observables or observable regions that remain unchanged under the
variation of the tuning parameters are not suited for the following tunes and
therefore dropped.
Similar to the procedure described in Ref. [26], we generate a set of
equivalent tunes that only differ by the Monte Carlo runs used to construct
the polynomial approximations as described above. The tunes are thus fully
equivalent and can be used to estimate the non-perturbative model-parameter
uncertainties as illustrated in Fig. 1 for a selection of data from the HERA
experiments. We call these alternative parameter sets replica tunes. To
reflect the uncertainty associated with the three beam-fragmentation
parameters we here consider seven such replicas, _cf._ Tab. 1 for the
resulting uncertainty variations.
parameter | parameter tag | tuning range | central tune | uncertainty variation
---|---|---|---|---
$\alpha_{B}$ | ALPHA_B | [-1, 20] | 14.2 | [13.9, 14.8]
$\beta_{B}$ | BETA_B | [0.5, 4] | 1.59 | [1.14, 1.60]
$\gamma_{B}$ | GAMMA_B | [1, 20] | 8.11 | [8.06, 9.47]
Table 1: A HADIC++ model parameters considered in the tuning. Quoted are the
initial parameter interval, the obtained central-tune value, and uncertainty
ranges extracted from 7 replica tunes.
Figure 1: S HERPA predictions for the hadronisation tune, for observables
measured by the H1 and ZEUS experiments at
$\sqrt{s}=$296\text{\,}\mathrm{GeV}$$. Shown is the transverse energy flow
(left) [69], thrust $\tau^{\prime}$ (center) [9] and the charged particle
multiplicity $n_{\mathrm{ch}}$ (right) [71]. Note, the statistical
uncertainties of the simulated data is small compared to the non-perturbative
tuning uncertainties indicated by the blue band.
## 4 (N)$\text{NLO}+\text{NLL}^{\prime}$ resummation for 1-jettiness in DIS
The 1-jettiness observable considered here is equivalent to thrust in DIS,
which has originally been resummed at NLL accuracy in [24, 75]. The more
general $n$-jettiness [76, 77] was suggested for lepton–hadron collisions in
[78], and has been resummed to NNLL accuracy [79]. For 1-jettiness, analytic
fixed order results at LO have been presented in [80], and the NLL calculation
has been matched to fixed order at NLO accuracy in [81]. The resummed
calculations in this formalism for event shapes in DIS were extended to N3LL
in [82]. Grooming for DIS has first been suggested in [41] based on jets
defined with the C ENTAURO jet algorithm [42]. The same Ref. [41] also
provided NNLL results for both 1-jettiness and jet mass after soft drop
grooming. Non-perturbative corrections have there been modelled through a two-
parameter shape function [83, 84]. To our knowledge there are no published
results studying these observables including matching to fixed order or using
a fixed order calculation alone.
### 4.1 NLL resummation in the C AESAR approach
To perform the NLL resummation of logarithms $L$ of event shapes in DIS we use
the implementation of the C AESAR formalism [27] available in the S HERPA
framework [28, 85]. For a recursive infrared and collinear (rIRC) safe
observable, the cumulative cross section for observable values up to
$v=\exp(-L)$ can be expressed to all orders, in general as a sum over partonic
channels $\delta$, as follows:
$\begin{split}\Sigma_{\mathrm{res}}(v)&=\sum_{\delta}\Sigma_{\mathrm{res}}^{\delta}(v)\,,\,\,\text{with}\\\
\Sigma_{\mathrm{res}}^{\delta}(v)&=\int
d\mathcal{B_{\delta}}\frac{\mathop{d\sigma_{\delta}}}{\mathop{d\mathcal{B_{\delta}}}}\exp\left[-\sum_{l\in\delta}R_{l}^{\mathcal{B_{\delta}}}(L)\right]\mathcal{P}^{\mathcal{B}_{\delta}}(L)\mathcal{S}^{\mathcal{B_{\delta}}}(L)\mathcal{F}^{\mathcal{B_{\delta}}}(L)\mathcal{H}^{\delta}(\mathcal{B_{\delta}})\,,\end{split}$
(20)
where $\frac{\mathop{d\sigma_{\delta}}}{\mathop{d\mathcal{B_{\delta}}}}$ is
the fully differential Born cross section for channel $\delta$ and
$\mathcal{H}$ implements the kinematic cuts applied to the Born phase space
$\mathcal{B}$. For a 2-jet observable like thrust in DIS, there is only one
relevant partonic Born channel, corresponding to an incoming and an outgoing
quark. This also implies that the soft function $\mathcal{S}$, which
implements colour evolution, is trivial in our case. Further, since we are
dealing with an additive observable, the multiple emission function
$\mathcal{F}$ is simply given by
$\mathcal{F}(L)=e^{-\gamma_{E}R^{\prime}}/\Gamma(1+R^{\prime})$, with
$R^{\prime}(L)=\partial R/\partial L$ and $R(L)=\sum_{l\in\delta}R_{l}(L)$.
The collinear radiators $R_{l}$ for the hard legs $l$ were computed in [27]
for a general observable $V$ scaling for the emission of a soft-gluon of
relative transverse momentum $k_{t}^{(l)}$ and relative rapidity $\eta^{(l)}$
with respect to leg $l$ as
$V(k)=\left(\frac{k_{t}^{\left(l\right)}}{\mu_{Q}}\right)^{a}e^{-b_{l}\eta^{\left(l\right)}}d_{l}\left(\mu_{Q}\right)g_{l}\left(\phi\right)\,.$
(21)
For the case of 1-jettiness we are focusing on in this publication, we have
$a=b_{l}=1$, and fixing $\mu_{Q}^{2}=Q^{2}$ also $d_{l}g_{l}=1$ since there is
no dependence on the azimuthal angle $\phi$. The precise form of the logarithm
can be varied according to
$L\to\ln\left[\frac{x_{L}}{v}-x_{L}+1\right]\to\ln\frac{x_{L}}{v}\quad\text{as}\quad
v\to 0\,,$ (22)
to estimated the impact of sub-leading logarithms, while leaving the
distribution at the kinematic endpoint $v\sim 1$ unchanged. Note this implies
an additional contribution to $R_{l}(L)$ to restore NLL accuracy.
The PDF factor $\mathcal{P}$, in our study applicable only to the hadronic
beam, is here given by
$\mathcal{P}=\frac{f_{q}(x,e^{-2L/(a+b)}\mu_{F}^{2})}{f_{q}(x,\mu_{F}^{2})}\,,$
(23)
corrects for the true initial-state collinear scale. We thereby account for
the full DGLAP evolution by calculating a simple ratio. For the purpose of
matching to a fixed order calculation, we also need the expansion of the ratio
to a given order in $\alpha_{\text{s}}$. We generally follow the approach of
[27] to implement the expansion of a leading order approximation. This of
course introduces additional effects beyond our considered logarithmic
accuracy. We argue it is safe to ignore those, given the generally small
numerical size of these contributions as seen for example in [28]. We here for
the first time apply the C AESAR implementation in S HERPA to an observable
that is sensitive to the PDF ratio (note this only applies to the ungroomed
version of thrust) and at the same time match to the NLO calculation for the
differential distribution and the NNLO result for the inclusive DIS process.
We hence need to take care of the expansion to one order higher. Following
[27], the numerator of Eq. (23) can to NLL accuracy be written and expanded in
powers of $\alpha_{\text{s}}$ as
$\displaystyle\mathbf{f}(x,e^{-2L/(a+b)}\mu_{F}^{2})$
$\displaystyle=\exp\left[-T\left(\frac{L}{a+b}\right)\mathbf{P}\otimes\right]\mathbf{f}(x,\mu_{F}^{2})$
$\displaystyle\sim
1-\left(T^{(1)}\left(\frac{L}{a+b}\right)+T^{(2)}\left(\frac{L}{a+b}\right)\right)\mathbf{P}\otimes\mathbf{f}(x,\mu_{F}^{2})$
$\displaystyle\phantom{=1}+\frac{1}{2}\left(T^{(1)}\left(\frac{L}{a+b}\right)\right)^{2}\mathbf{P}\otimes\mathbf{P}\otimes\mathbf{f}(x,\mu_{F}^{2})+\mathcal{O}\left(\alpha_{\text{s}}^{3}\right),$
(24)
where $T^{(i)}$ denotes the $i$th term obtained by expanding the integrated
strong coupling
$T(L)=-\frac{1}{\pi\beta_{0}}\ln(1-2\alpha_{\text{s}}\beta_{0}L)$ (25)
in powers of $\alpha_{\text{s}}$. The bold-faced symbols represent matrices
(of splitting functions, $\mathbf{P}$) and vectors
($\mathbf{f}=(f_{u},f_{d},f_{s},\dots)$) in flavour space, and the convolution
is given by
$\mathbf{P}\otimes\mathbf{f}(x,\mu_{F}^{2})=\int_{x}^{1}\frac{dz}{z}\mathbf{P}\left(\frac{x}{z}\right)\mathbf{f}(z,\mu_{F}^{2})\,.$
(26)
New terms at $\mathcal{O}(\alpha_{\text{s}}^{2})$ hence originate from the
higher order expansion of $T$, mixed terms with other parts of the resummation
multiplying the leading order expansion, and the convolution of two splitting
functions with the PDF in the last line of Eq. (24). The last one is the only
one that requires a non-trivial implementation. We use the expressions from
[86] for convoluted splitting functions, and solve the final integral for the
convolution with the PDF through Monte Carlo integration, as done at leading
order.
We match our resummed calculation in the multiplicative matching scheme along
the lines of [85], which we briefly recap here. The matching to fixed order is
done at the level of cumulative distributions $\Sigma(v)$. Note that we have
dropped the label for the partonic channel since in our case there is a single
one only. We expand the inclusive cross section $\sigma_{\text{fo}}$ as well
as the fixed-order and resummed cumulative distributions, $\Sigma_{\text{fo}}$
and $\Sigma_{\text{res}}$ in series of $\alpha_{\text{s}}$:
$\displaystyle\sigma_{\text{fo}}$
$\displaystyle=\sigma^{(0)}+\sigma^{(1)}_{\text{fo}}+\sigma^{(2)}_{\text{fo}}+\dots\,,$
(27) $\displaystyle\Sigma_{\text{fo}}(v)$
$\displaystyle=\sigma^{(0)}+\Sigma^{(1)}_{\text{fo}}(v)+\Sigma^{(2)}_{\text{fo}}(v)+\dots\,,$
(28) $\displaystyle\Sigma_{\text{res}}(v)$
$\displaystyle=\sigma^{(0)}+\Sigma^{(1)}_{\text{res}}(v)+\Sigma^{(2)}_{\text{res}}(v)+\dots\,,$
(29)
where the number in parentheses indicates the respective order in
$\alpha_{\text{s}}$, and $\sigma^{(0)}$ denotes the Born-level cross section.
Our final matched expression for the cumulative distribution, with the
dependencies on the observable value suppressed, reads:
$\Sigma_{\text{matched}}=\Sigma_{\text{res}}\left(1+\frac{\Sigma^{(1)}_{\text{fo}}-\Sigma^{(1)}_{\text{res}}}{\sigma^{(0)}}+\frac{\Sigma^{(2)}_{\text{fo}}-\Sigma^{(2)}_{\text{res}}}{\sigma^{(0)}}-\frac{\Sigma^{(1)}_{\text{res}}}{\sigma^{(0)}}\frac{\Sigma^{(1)}_{\text{fo}}-\Sigma^{(1)}_{\text{res}}}{\sigma^{(0)}}\right)\,.$
(30)
Note that, compared to our earlier works, we use $\Sigma^{(2)}$ directly, thus
reproducing the inclusive cross section to one order higher, _i.e._ NNLO, what
requires the calculation of $\sigma^{(2)}_{\text{fo}}$. Importantly, the
resummed NLL result $\Sigma_{\text{res}}$ is multiplied by
$\frac{\Sigma^{(1)}_{\text{fo}}-\Sigma^{(1)}_{\text{res}}}{\sigma^{(0)}}\to\frac{\alpha_{\text{s}}}{2\pi}C_{1}\quad\text{as}\quad
v\to 0\,.$ (31)
We refer to the distribution that includes all NLL terms in
$\Sigma_{\text{res}}$ and additionally the coefficient $C_{1}$ as
$\text{NLL}^{\prime}$ accurate. Through this matching procedure we achieve a
formal accuracy of $\text{NLO}+\text{NLL}^{\prime}$ for the differential
distribution and NNLO for the inclusive event rate, referred to as
(N)$\text{NLO}+\text{NLL}^{\prime}$ in what follows.
In addition to the perturbative contribution described above, there is a
significant non-perturbative component to the distribution of event shapes,
that we necessarily need to take into account in order to accurately describe
actual collider data. While it has been shown in various circumstances that
soft-drop grooming reduces the impact of hadronisation corrections, see for
example [83, 38, 39, 40, 87, 30], it is typically still necessary to account
for a remaining small non-perturbative contribution. We here adopt the
approach of [30] to extract transfer matrices from Monte Carlo simulations.
Transfer matrices are defined as
$\mathcal{T}_{hp}=\frac{\int\mathop{dP}\frac{\mathop{d\sigma}}{\mathop{dP}}\Theta_{p}\left(P\right)\Theta_{h}\left(H(P)\right)}{\int\mathop{dP}\frac{\mathop{d\sigma}}{\mathop{dP}}\Theta_{p}\left(P\right)}\,,$
(32)
with
$\displaystyle\Theta_{p}\left(P\right)$
$\displaystyle=\prod_{i=1}^{m}\theta(V_{i}(P)-v^{\text{min}}_{p,i})\theta(v^{\text{max}}_{p,i}-V_{i}(P))\,,$
(33) $\displaystyle\Theta_{h}\left(H(P)\right)$
$\displaystyle=\prod_{i=1}^{m}\theta\left(V_{i}\left(H(P)\right)-v^{\text{min}}_{h,i}\right)\theta\left(v^{\text{max}}_{h,i}-V_{i}\left(H(P)\right)\right)\,,$
(34)
for a transition between the parton level phase space $P$ and the
corresponding hadron level configuration $H(P)$, characterised by a set of
observables $V_{i}$ that can be calculated on both of them. For our purpose,
we assume that the requirements on the DIS kinematics, _cf._ Sec. 2,
sufficiently fix the remaining degrees of freedom other than 1-jettiness
$\tau$. This first of all means that we do not allow non-perturbative
corrections to change the underlying Born kinematics, _i.e._ the $Q^{2},\,y$
bin, in contrast to, for example, measurements performed on jets with a
potentially affected transverse momentum spectrum. On the other hand this
implicitly assumes that it is valid to average the corrections for all
configurations with a common 1-jettiness value. Hence, we are only concerned
with events migrating between different bins in $\tau$ within a given
$Q^{2},\,y$ bin. The transfer matrices as defined above can readily be
extracted from the S HERPA event generator by analysing the different stages
of the events evolution, _i.e._ after parton showering but before
hadronisation and thereafter. For practical details of our event generation
setup see Sec. 3. Our final results are then calculated from the resummed and
matched parton level bins $\Delta\sigma_{p}^{\text{PL}}$ as
$\mathop{\Delta\sigma_{h}^{\text{HL}}}=\sum_{p}\mathcal{T}_{hp}\mathop{\Delta\sigma_{p}^{\text{PL}}}\,.$
(35)
### 4.2 Grooming in DIS
The framework described above has already been employed to obtain resummed
predictions for soft-drop thrust in lepton–lepton collisions at
$\text{NLO}+\text{NLL}^{\prime}$ precision [39], for soft-drop groomed
hadronic event shapes [40] and groomed jet substructure observables at the LHC
[87, 88, 30]. The extensions made in [40] to accommodate the phase space
constraints implied by soft-drop grooming, with general parameters
$z_{\text{cut}}$ and $\beta$, are directly applicable here. Note that [41]
does not define a $\beta\neq 0$ version of grooming in DIS, and we make no
attempt here to extend it.
The applicability of the results from [40] to DIS event shapes relies on two
statements. First, within the current hemisphere the phase space constraints
to radiation in the soft and collinear limits correspond to the case of final
state radiation in general hadronic collisions. Second, in the beam hemisphere
any soft and collinear radiation is groomed away. Accordingly, we can treat
radiation in $\mathcal{H}_{B}$ equivalent to the initial state radiation case
in [40], even if the precise shape of the phase space boundary is different,
but such difference does not enter at NLL accuracy. We analyse the behaviour
of the C ENTAURO algorithm and the associated soft-drop grooming variant in
the language of the C AESAR framework in the following to illustrate this.
Recall that we are working in the Breit frame. At NLL accuracy, we have to
take into account ensembles of soft particles, well separated in rapidity,
around a Born configuration consisting of the proton momentum
$P^{\mu}=\frac{Q}{2x_{B}}n^{\mu}_{+}$ (36)
and the outgoing struck quark in $n_{-}$ direction. The virtual photon carries
momentum
$q=\frac{Q}{2}(n_{-}-n_{+})~{}.$ (37)
We parameterise the momenta of additional soft gluons as
$k_{i}^{\mu}=k_{t}^{i}\left(\frac{e^{\eta_{i}}}{2}n_{-}^{\mu}+\frac{e^{-\eta_{i}}}{2}n_{+}^{\mu}+n_{\perp}^{\mu}\right)\,,$
(38)
where $n_{\perp}$ is a transverse unit vector perpendicular to $n_{+}$ and
$n_{-}$. The variable introduced in the C ENTAURO algorithm, _cf._ Eq. (8),
can be written using the phase space variables $\eta_{i}$, $k_{t}^{i}$ as
$\bar{z}_{i}=2e^{-\eta_{i}}~{},$ (39)
such that the expression for the distance measure, _cf._ Eq. (7), becomes
$d_{ij}=4\left(e^{-2\eta_{i}}+e^{-2\eta_{j}}+2e^{-(\eta_{i}+\eta_{j})}\cos\Delta\phi_{ij}\right)\sim
4e^{-2\eta_{i}}\,,$ (40)
where we have identified the behaviour for strong ordering in $\eta$,
$\eta_{i}\ll\eta_{j}$. In this limit, the algorithm builds up a single jet
containing the hard quark by adding the next remaining gluon that is most
collinear to this jet. The last clustering will add the gluon most collinear
to the beam direction to the jet. If all gluons are separated in rapidity well
enough, there are no other clusters to be taken care of.
From this discussion it is clear that all comparisons of scales during soft
drop will be between a soft gluon and a jet containing the hard quark. At Born
level, the four-momentum of the jet will be approximately that of the quark,
and the gluon will be the softer of the two. With this in mind the hardness
measure for soft drop for soft momentum $k_{i}$ can be written as
$z_{i}\sim\frac{k_{t}^{i}}{Q}e^{\eta_{i}}~{}.$ (41)
Within the current hemisphere, the phase space restriction, on an emission
that passes the soft-drop criterion, is given by
$\frac{k_{t}e^{\eta}}{Q}>z_{\text{cut}}~{},$ (42)
which precisely matches the one given in [40] for $\beta=0$ (see Sec. 3.4
point (iv), and note that the hard quark has energy $Q/2$ in the Breit frame).
Note that particles outside of the current hemisphere will enter in Eq. (42)
with negative rapidity $\eta$. They will hence be groomed away unless they are
at very high $k_{t}$, only causing logarithms of $z_{\text{cut}}$. We note
again that the precise shape of the phase space boundary is different from
what is given in [40] for initial states. The main point is however that only
logarithms of $z_{\text{cut}}$ are produced, which we ignore noting again that
we work in the limit $v\ll z_{\text{cut}}$.
### 4.3 Calculational tools and setup
As already stated, the resummation calculation for 1-jettiness is accomplished
with the C AESAR plugin to S HERPA that hooks into the event generation
framework†††Note, during the course of this work the plugin has been ported to
the S HERPA -3.0 release series.. S HERPA thereby provides all the process
management, and gives access to the C OMIX matrix element generator [58], as
well as phase-space integration and event-analysis functionalities. We make
use of S HERPA ’s interface to L HA P DF [59] and use the PDF4LHC_40_pdfas
PDF set, as we do for the parton-shower simulations outlined in the previous
section. The value of the strong coupling is set accordingly, _i.e._
$\alpha_{S}(M^{2}_{Z})=0.118$. The S HERPA framework is also used to compile
all the required higher-order tree-level and one-loop calculations. For the
NLO QCD computations we use the S HERPA implementation of the Catani–Seymour
dipole subtraction [89] and the interfaces to the R ECOLA [90, 91] and O PEN
L OOPS [92] one-loop amplitude generators. The calculation of NNLO accurate
predictions for DIS has been automated in S HERPA in [29], and we use it to
compute cross sections $\sigma^{(2)}_{\text{fo}}$ at order
$\alpha_{\text{s}}^{2}$ differential in $Q^{2}$ and $y$ to achieve overall
NNLO accuracy for inclusive cross sections. This corresponds to an accuracy of
the distribution differential in thrust at NLO, and we refer to the combined
accuracy of our fixed order predictions including cross sections as (N)NLO.
The plugin implements the building blocks of the C AESAR master formula Eq.
(20), along with the necessary expansion in $\alpha_{s}$ used in the matching
with fixed-order calculations. The building blocks are evaluated fully
differentially for each Born-level configuration $\mathcal{B}_{\delta}$ of a
given momentum configuration. Jet clustering and grooming functionalities are
accessed through the interface of S HERPA to F AST J ET [52]. Non-
perturbative corrections are extracted from dedicated runs of the S HERPA
generator using the identical setup described in Sec. 3, thereby employing the
functionality of the R IVET analysis tool to provide access to intermediate
evolution stages through the H EP MC event record [93].
## 5 Results for (groomed) 1-jettiness in DIS
Having outlined our calculational techniques for describing hadronic final
state observables in neutral current DIS, we can finally present our numerical
results for the 1-jettiness event shape. We begin by discussing selected
results for the ungroomed case. We have compiled predictions for a wide range
of $Q^{2}$ values, _i.e._ $Q^{2}\in[150,20000]\;\text{GeV}^{2}$. Furthermore,
we consider the production cross section differential in the events
inelasticity, thereby covering the region $y\in[0.05,0.94]$. For brevity, we
here focus on three kinematic regions corresponding to medium values of
$y\in[0.4,0.7]$ and rather low ($Q^{2}\in[150,200]\,\text{GeV}^{2}$), medium
($Q^{2}\in[440,700]\,\text{GeV}^{2}$) and high
($Q^{2}\in[3500,8000]\,\text{GeV}^{2}$) photon virtuality.
Along with the central predictions we show error bands indicating the
perturbative uncertainty obtained from 7-point variations of
$\mu_{R},\mu_{F}$, in both the shower and the semi-analytic calculation, and
in addition a variation of $x_{L}=0.5,2$ in the latter, _cf._ Eq. (22).
Furthermore, we include an uncertainty estimate related to the tuning of beam-
fragmentation parameters based on replica tunes, see Sec. 3.2. Generally, this
contribution is found to be rather small compared to the perturbative
uncertainties. We observe the overall uncertainties for the NLO QCD matrix
element plus parton-shower simulations and the resummation predictions to be
of similar sizes.
We first analyse the behaviour of the $\text{NLO}+\text{NLL}^{\prime}$
resummation calculation upon inclusion of the NNLO normalisation correction
and non-perturbative effects. To this end we compile in Fig. 2 corresponding
predictions for the three kinematic regions specified before. From the lower
panels, showing the ratio to the respective $\text{NLO}+\text{NLL}^{\prime}$
result, it can be read off, that correcting the normalisation to NNLO accuracy
has a rather small impact. The differential cross section receives a small
negative correction, of at most a few percent at small $\tau$ in the lower
$Q^{2}$ region. Note, however, that even the smallest $Q^{2}$ values in this
analysis remain sizeable compared to the overall range accessible for the HERA
experiments. Somewhat more significant is the reduction in the perturbative
uncertainties when going from NLO to NNLO, in particular for the bulk of the
distributions, _i.e._ low values of 1-jettiness.
Next, we consider the inclusion of non-perturbative corrections based on the
transfer-matrix approach described in Sec. 4.1. As clearly visible in Fig. 2
these significantly alter the shape of the distributions, introducing a
sizeable shift towards larger 1-jettiness values. In particular for the low
and medium $Q^{2}$ region the first bin gets almost entirely depopulated. In
contrast, for values of $\tau\approx 0.1\dots 0.2$ corrections can reach up to
$+100\%$. The effect of hadronisation corrections is less pronounced at higher
$Q^{2}$. We furthermore note, that the non-perturbative corrections through
the bin migration via transfer matrices partially compensate the dependence of
the perturbative calculation on scale variations and in particular of
$\mu_{R}$.
Figure 2: Distributions of ungroomed 1-jettiness in selected $y-Q^{2}$ bins,
at different stages of the calculation, at $\text{NLO}+\text{NLL}^{\prime}$
accuracy, including the normalisation at NNLO
((N)$\text{NLO}+\text{NLL}^{\prime}$) accuracy, and including non-perturbative
corrections. All results correspond to DIS kinematics with $y\in[0.4,0.7]$ and
the plots represent from left to right regions of
$Q^{2}/\text{GeV}^{2}\in[150,200]$, $[440,700]$, and $[3500,8000]$,
respectively. The lower panels present the ratio to the plain
$\text{NLO}+\text{NLL}^{\prime}$ result.
We close this first discussion of the resummed predictions for ungroomed
1-jettiness by pointing to the distinct peak at $\tau\approx 1$ for the low
and medium $Q^{2}$ distributions, emerging after a significant decline of the
differential cross section from lower to larger observable values. For the
given observable definition the configuration $\tau=1$ can be attributed to
events with an empty current hemisphere ${\mathcal{H}}_{C}$ [80]. Such
configurations first appear when considering the NLO real-emission correction
to the DIS process, when both final state partons feature negative
longitudinal momenta in the Breit frame, such that 1-jettiness defaults to 1,
see Eq. (6). We here account for these configurations through matching to the
exact NLO QCD result for $\tau$, _i.e._ including the full
${\cal{O}}(\alpha_{S})$ corrections to the two-parton channel. It can be
observed, that hadronisation corrections reduce the amount of $\tau\approx 1$
events, what can be expected, as the fragmentation of partons originally in
the beam hemisphere might spill over hadrons in the current hemisphere.
We now turn to the presentation of the hadron level results from MEPS@NLO
simulations with S HERPA as outlined in Sec. 3 and compare those to the
(N)$\text{NLO}+\text{NLL}^{\prime}+\text{NP}$ predictions. In Fig. 3 we
compare the respective results for the three considered kinematic regions. We
observe an overall fair agreement between the matrix element improved shower
simulations at hadron level obtained from S HERPA and the resummed and
matched calculation at (N)$\text{NLO}+\text{NLL}^{\prime}+\text{NP}$ accuracy,
corrected for non-perturbative effects. In general the merged prediction
features a somewhat harder spectrum, _i.e._ favours somewhat larger observable
values. This might also be attributed to the inclusion of the exact tree-level
three- and four-jet matrix elements, see Eq. (11). These contributions feature
LO scale dependence and are thus the source for the somewhat enlarged
theoretical uncertainties in the shower simulation towards larger values of
$\tau$. However, the regions of small 1-jettiness agree within uncertainties
for all three kinematic regions, up until the peak of the respective
distribution. Towards the kinematic endpoint, the two approaches tend to agree
again, with both calculations predicting very similar cross sections for
events with $\tau\sim 1$.
Figure 3: Distributions of 1-jettiness in selected $y-Q^{2}$ bins, _i.e._
$y\in[0.4,0.7]$ and, from left to right, $Q^{2}/\text{GeV}^{2}\in[150,200]$,
$[440,700]$, and $[3500,8000]$, respectively. Shown are hadron level MEPS@NLO
predictions from S HERPA and results at
(N)$\text{NLO}+\text{NLL}^{\prime}+\text{NP}$ accuracy. The lower panels
present the ratio to the MEPS@NLO result.
Besides the plain 1-jettiness event shape we here also consider the effect of
soft-drop grooming the hadronic final state. In Fig. 4 we show resummed
predictions for groomed 1-jettiness, referred to as $\tau^{\text{SD}}$ in what
follows, integrated over the full $Q^{2}$ range, _i.e._
$Q^{2}\in[150,20000]\;\text{GeV}^{2}$, and the inelasticity region
$y\in[0.2,0.7]$. We compiled predictions for three commonly considered values
of $z_{\text{cut}}$, namely $z_{\text{cut}}=0.05,0.1,0.2$, thereby always
assuming the angular grooming parameter $\beta=0$. As seen for the ungroomed
case, we note rather small effects of the NNLO normalisation corrections
compared to the $\text{NLO}+\text{NLL}^{\prime}$ calculation. Also the
systematic uncertainties hardly change from NLO to NNLO. However, the size of
the non-perturbative corrections is significantly reduced relative to the
ungroomed case, staying below $50\%$ and being largely flat over a wide range
of $\tau^{\text{SD}}$, apart from very low values of 1-jettiness and at the
endpoint $\tau^{\text{SD}}\sim 1$. This confirms the potential of soft-drop
grooming to mitigate hadronisation effects for event shape observables also in
DIS, seen before in $e^{+}e^{-}$ [38, 39] and $pp$ collisions [40].
Figure 4: Distributions of groomed 1-jettiness, at different stages of the
calculation, at $\text{NLO}+\text{NLL}^{\prime}$ accuracy, including the
normalisation at NNLO ((N)$\text{NLO}+\text{NLL}^{\prime}$) accuracy, and
including non-perturbative corrections. From left to right the plots represent
predictions for the grooming parameter $z_{\text{cut}}=0.05,0.1,0.2$,
respectively. The lower panels present the ratio to the plain
$\text{NLO}+\text{NLL}^{\prime}$ result.
The comparison of the (N)$\text{NLO}+\text{NLL}^{\prime}+\text{NP}$ results
with hadron level simulations at MEPS@NLO accuracy is presented in Fig. 5. For
all the $z_{\text{cut}}$ values, we observe good agreement between our S
HERPA simulation and the resummation calculation somewhat better than for the
ungroomed case. In all three cases, the
(N)$\text{NLO}+\text{NLL}^{\prime}+\text{NP}$ calculation predicts a larger
cross section in the $\tau\sim 1$ bin, although still compatible within the
uncertainty of the event generator for $z_{\text{cut}}=0.05$ and the combined
uncertainty for both calculations for $z_{\text{cut}}=0.1$. Apart from this
last bin, for these two $z_{\text{cut}}$ values the resummation calculation is
consistently below the S HERPA simulation. In the case of
$z_{\text{cut}}=0.05$, this happens flat over the full spectrum
$\tau^{\mathrm{SD}}<1$, while for increasing $z_{\text{cut}}$ a slight shape
develops, with the (N)$\text{NLO}+\text{NLL}^{\prime}+\text{NP}$ cross section
decreasing faster for $\tau^{\mathrm{SD}}<z_{\text{cut}}$ than what is seen in
the Monte Carlo simulation.
Figure 5: Distributions of groomed 1-jettiness. Shown are hadron level
MEPS@NLO predictions from S HERPA and results at
(N)$\text{NLO}+\text{NLL}^{\prime}+\text{NP}$ accuracy. From left to right the
plots represent predictions for the grooming parameter
$z_{\text{cut}}=0.05,0.1,0.2$, respectively. The lower panels present the
ratio to the MEPS@NLO result.
It will be interesting to compare the
(N)$\text{NLO}+\text{NLL}^{\prime}+\text{NP}$ predictions and the S HERPA
MEPS@NLO simulations with the data of upcoming measurements by the H1
experiment. This will shed light on the found deviations between the two sets
of predictions and possibly guide the development of yet improved theoretical
predictions, _e.g._ through the inclusion of next-to-next-to-leading
logarithmic corrections.
## 6 Conclusions
We presented the calculation of theoretical predictions for the 1-jettiness
event shape in neutral current DIS at HERA energies. The here considered
1-jettiness observable, evaluated in the Breit frame, is equivalent to the
well-known thrust variable that has been widely studied at lepton and hadron
colliders. Besides plain 1-jettiness we also considered its variant after
soft-drop grooming the hadronic final state using different values of the
grooming parameter $z_{\text{cut}}$. We consider the triple-differential cross
section in the observable, momentum transfer $Q^{2}$, and the events
inelasticity $y$.
Based on the C AESAR formalism we derive NLL accurate results matched to the
exact NLO QCD matrix element for the two-jet DIS matrix element. Furthermore,
we include the exact NNLO QCD corrections to the inclusive DIS process,
thereby achieving full NNLO accuracy for the integrated observable
distribution. We furthermore correct our results of
(N)$\text{NLO}+\text{NLL}^{\prime}$ accuracy for non-perturbative
hadronisation effects through a transfer matrix that takes into account
migration in the observable value when going from parton to hadron level. The
corresponding corrections have been extracted from Monte Carlo simulations at
MEPS@NLO accuracy with the S HERPA generator. To this end, we have performed
tunes of the beam-fragmentation parameters of S HERPA ’s new cluster
fragmentation model against data from the H1 and ZEUS experiments. We thereby
also derived replica tunes that account for the parametric uncertainties.
For plain 1-jettiness we have shown results for three kinematic regions,
corresponding to medium inelasticity $y$ and ranges of rather low, medium, and
high $Q^{2}$ values. While the impact of the NNLO contributions is found to be
very small, hadronisation corrections significantly sculpt the differential
distributions, pushing events from lower to larger 1-jettiness values. When
comparing the hadronisation corrected (N)$\text{NLO}+\text{NLL}^{\prime}$
predictions with hadron level predictions from S HERPA good agreement is
found, with larger deviations dominantly in the region $0.2<\tau<0.6$. Quite
good agreement is found regarding events at the endpoint of the distribution,
_i.e._ $\tau\simeq 1$. For the low and medium $Q^{2}$ regions the distribution
here develops a significant peak, that can be attributed to events with an
empty current hemisphere.
For the soft-drop groomed variant of 1-jettiness we have shown predictions for
three values of $z_{\text{cut}}$, integrated over a wide range of $Q^{2}$,
_i.e._ $Q^{2}\in[150,20000]\;\text{GeV}^{2}$, and $y\in[0.2,0.7]$. For all
values of $z_{\text{cut}}$ non-perturbative corrections to the resummed
predictions get significantly reduced, when comparing to the ungroomed case.
Furthermore, an improved agreement with the hadron level predictions from S
HERPA is found.
It will be exciting to confront the two types of predictions with actual data
from the HERA collider that are currently being analysed by the H1 experiment.
We can expect that in particular for the ungroomed 1-jettiness observable data
should be able to discriminate between the two predictions. This will motivate
and guide the development and advancement of the theoretical predictions. For
DIS parton shower simulations there are recent developments towards the
inclusion of NNLO QCD corrections [29] and to achieve formal NLL accuracy [94,
95, 96, 97]. This would allow to match the precision of the analytic
predictions we presented in this study. Improving the analytic calculation
might require the inclusion of higher-logarithmic corrections or improved
means to account for non-perturbative corrections. Furthermore, a detailed
analysis of systematic differences between analytic NLL resummation and shower
algorithms implementing unitarity and momentum conservation along the lines of
[98] might help to pin down the origin of the observed differences.
## Acknowledgements
We would like to thank Daniel Britzger and Henry Klest for triggering us to
dive into DIS event shapes and a very fruitful communication. We furthermore
thank Johannes Hessler and Vinicius Mikuni for discussions. We are indebted to
Stefan Höche for assistance with the NNLO corrections and we are grateful to
Frank Krauss for help with S HERPA ’s new beam fragmentation model.
MK and SS acknowledge support from BMBF (05H21MGCAB) and funding by the
Deutsche Forschungsgemeinschaft (DFG, German Research Foundation) - project
number 456104544 and 510810461. DR is supported by the STFC IPPP grant
(ST/T001011/1).
## Appendix A Tuning details
We here collate more detailed information on the tuning of the A HADIC++
beam-fragmentation parameters. The R IVET analyses and considered observable
measurements by the H1 and ZEUS HERA experiments used for the tuning are
summarised in Tab. 2.
R IVET Analysis name [reference] | Observables | Virtuality range [$\text{GeV}^{2}$]
---|---|---
H1_2006_I699835 [9] | thrust, jet broadening | $Q^{2}\in[256,400]$
H1_1994_S2919893 [70] | transverse energy flow | $Q^{2}\in[10,100]$
| energy-energy correlation | $Q^{2}\in[10,100]$
H1_1995_I394793 [73] | quark fragmentation functions | $Q^{2}>100$
H1_1996_I422230 [72] | charged multiplicity distributions | $Q^{2}\in[10,1000]$
H1_1996_I424463 [99] | charged particle spectra | $Q^{2}\in[5,50]$
H1_1997_I445116 [74] | quark fragmentation functions | $Q^{2}\in[100,8000]$
| charged hadron energy spectra | $Q^{2}\in[100,8000]$
H1_2000_S4129130 [69] | transverse energy flow | $Q^{2}\in[10,2200]$
Table 2: R IVET analysis tags, observables and corresponding photon
virtuality ranges used for the tuning.
## References
* [1] D. Decamp et al., The ALEPH collaboration, _Measurement of the strong coupling constant alpha-s from global event shape variables of hadronic Z decays_ , Phys. Lett. B 255 (1991), 623–633
* [2] M. Z. Akrawy et al., The OPAL collaboration, _A Measurement of Global Event Shape Distributions in the Hadronic Decays of the $Z^{0}$_, Z. Phys. C 47 (1990), 505–522
* [3] O. Adrian et al., The L3 collaboration, _Determination of alpha-s from hadronic event shapes measured on the Z0 resonance_ , Phys. Lett. B 284 (1992), 471–481
* [4] P. Abreu et al., The DELPHI collaboration, _Energy dependence of event shapes and of alpha(s) at LEP-2_ , Phys. Lett. B 456 (1999), 322–340
* [5] J. Abdallah et al., The DELPHI collaboration, _A Study of the energy evolution of event shape distributions and their means with the DELPHI detector at LEP_ , Eur. Phys. J. C 29 (2003), 285–312, [hep-ex/0307048]
* [6] S. Marzani, G. Soyez and M. Spannowsky, _Looking inside jets: an introduction to jet substructure and boosted-object phenomenology_ , vol. 958, Springer, 2019
* [7] C. Adloff et al., The H1 collaboration, _Measurement of event shape variables in deep inelastic e p scattering_ , Phys. Lett. B 406 (1997), 256–270, [hep-ex/9706002]
* [8] C. Adloff et al., The H1 collaboration, _Investigation of power corrections to event shape variables measured in deep inelastic scattering_ , Eur. Phys. J. C 14 (2000), 255–269, [hep-ex/9912052], [Erratum: Eur.Phys.J.C 18, 417–419 (2000)]
* [9] A. Aktas et al., The H1 collaboration, _Measurement of event shape variables in deep-inelastic scattering at HERA_ , Eur. Phys. J. C 46 (2006), 343–356, [hep-ex/0512014]
* [10] J. Breitweg et al., The ZEUS collaboration, _Event shape analysis of deep inelastic scattering events with a large rapidity gap at HERA_ , Phys. Lett. B 421 (1998), 368–384, [hep-ex/9710027]
* [11] S. Chekanov et al., The ZEUS collaboration, _Measurement of event shapes in deep inelastic scattering at HERA_ , Eur. Phys. J. C 27 (2003), 531–545, [hep-ex/0211040]
* [12] S. Chekanov et al., The ZEUS collaboration, _Event shapes in deep inelastic scattering at HERA_ , Nucl. Phys. B 767 (2007), 1–28, [hep-ex/0604032]
* [13] J. Schieck, S. Bethke, S. Kluth, C. Pahl and Z. Trocsanyi, The JADE collaboration, _Measurement of the strong coupling $alpha_{S}$ from the three-jet rate in $e^{+}e^{-}$ \- annihilation using JADE data_, Eur. Phys. J. C 73 (2013), no. 3, 2332, [arXiv:1205.3714 [hep-ex]]
* [14] G. Abbiendi et al., The ALEPH, DELPHI, L3, OPAL, LEP collaboration, _Search for Charged Higgs bosons: Combined Results Using LEP Data_ , Eur. Phys. J. C 73 (2013), 2463, [arXiv:1301.6065 [hep-ex]]
* [15] J. Abdallah et al., The DELPHI collaboration, _Measurement of the electron structure function F $\frac{e}{2}$ at LEP energies_, Phys. Lett. B 737 (2014), 39–47
* [16] N. Fischer, S. Gieseke, S. Kluth, S. Plätzer and P. Skands, The OPAL collaboration, _Measurement of observables sensitive to coherence effects in hadronic Z decays with the OPAL detector at LEP_ , Eur. Phys. J. C 75 (2015), no. 12, 571, [arXiv:1505.01636 [hep-ex]]
* [17] A. Badea, A. Baty, P. Chang, G. M. Innocenti, M. Maggi, C. Mcginn, M. Peters, T.-A. Sheng, J. Thaler and Y.-J. Lee, _Measurements of two-particle correlations in $e^{+}e^{-}$ collisions at 91 GeV with ALEPH archived data_, Phys. Rev. Lett. 123 (2019), no. 21, 212002, [arXiv:1906.00489 [hep-ex]]
* [18] Y. Chen et al., _Jet energy spectrum and substructure in e +e- collisions at 91.2 GeV with ALEPH Archived Data_, JHEP 06 (2022), 008, [arXiv:2111.09914 [hep-ex]]
* [19] A. Accardi et al., _Electron Ion Collider: The Next QCD Frontier: Understanding the glue that binds us all_ , Eur. Phys. J. A 52 (2016), no. 9, 268, [arXiv:1212.1701 [nucl-ex]]
* [20] R. Abdul Khalek et al., _Science Requirements and Detector Concepts for the Electron-Ion Collider: EIC Yellow Report_ , Nucl. Phys. A 1026 (2022), 122447, [arXiv:2103.05419 [physics.ins-det]]
* [21] J. L. Abelleira Fernandez et al., The LHeC Study Group collaboration, _A Large Hadron Electron Collider at CERN: Report on the Physics and Design Concepts for Machine and Detector_ , J. Phys. G 39 (2012), 075001, [arXiv:1206.2913 [physics.acc-ph]]
* [22] P. Agostini et al., The LHeC, FCC-he Study Group collaboration, _The Large Hadron–Electron Collider at the HL-LHC_ , J. Phys. G 48 (2021), no. 11, 110501, [arXiv:2007.14491 [hep-ex]]
* [23] D. Kang, C. Lee and I. W. Stewart, _Using 1-Jettiness to Measure 2 Jets in DIS 3 Ways_ , Phys. Rev. D 88 (2013), 054004, [arXiv:1303.6952 [hep-ph]]
* [24] V. Antonelli, M. Dasgupta and G. P. Salam, _Resummation of thrust distributions in DIS_ , JHEP 02 (2000), 001, [hep-ph/9912488]
* [25] G. S. Chahal and F. Krauss, _Cluster Hadronisation in Sherpa_ , SciPost Phys. 13 (2022), no. 2, 019, [arXiv:2203.11385 [hep-ph]]
* [26] M. Knobbe, F. Krauss, D. Reichelt and S. Schumann, _Measuring Hadronic Higgs Boson Branching Ratios at Future Lepton Colliders_ , arXiv:2306.03682 [hep-ph]
* [27] A. Banfi, G. P. Salam and G. Zanderighi, _Principles of general final-state resummation and automated implementation_ , JHEP 03 (2005), 073, [hep-ph/0407286]
* [28] E. Gerwick, S. Höche, S. Marzani and S. Schumann, _Soft evolution of multi-jet final states_ , JHEP 02 (2015), 106, [arXiv:1411.7325 [hep-ph]]
* [29] S. Höche, S. Kuttimalai and Y. Li, _Hadronic Final States in DIS at NNLO QCD with Parton Showers_ , Phys. Rev. D 98 (2018), no. 11, 114013, [arXiv:1809.04192 [hep-ph]]
* [30] D. Reichelt, S. Caletti, O. Fedkevych, S. Marzani, S. Schumann and G. Soyez, _Phenomenology of jet angularities at the LHC_ , JHEP 03 (2022), 131, [arXiv:2112.09545 [hep-ph]]
* [31] J. Hessler, D. Britzger and S. Lee, The H1 collaboration, _Measurement of 1-jettiness in deep-inelastic scattering at HERA_ , PoS EPS-HEP2021 (2022), 367, [arXiv:2111.11364 [hep-ex]]
* [32] J. Hessler, _Measurement of the 1-jettiness Event Shape Observable in Deep-inelastic Electron-Proton Scattering_ , Master’s thesis, Munich, Tech. U., 2021
* [33] V. Andreev et al., The H1 collaboration, _Unbinned Deep Learning Jet Substructure Measurement in High $Q^{2}$ ep collisions at HERA_, arXiv:2303.13620 [hep-ex]
* [34] T. Carli, T. Gehrmann and S. Höche, _Hadronic final states in deep-inelastic scattering with Sherpa_ , Eur. Phys. J. C 67 (2010), 73–97, [arXiv:0912.3715 [hep-ph]]
* [35] A. J. Larkoski, S. Marzani, G. Soyez and J. Thaler, _Soft Drop_ , JHEP 05 (2014), 146, [arXiv:1402.2657 [hep-ph]]
* [36] J. M. Butterworth, A. R. Davison, M. Rubin and G. P. Salam, _Jet substructure as a new Higgs search channel at the LHC_ , Phys. Rev. Lett. 100 (2008), 242001, [arXiv:0802.2470 [hep-ph]]
* [37] M. Dasgupta, A. Fregoso, S. Marzani and G. P. Salam, _Towards an understanding of jet substructure_ , JHEP 1309 (2013), 029, [arXiv:1307.0007 [hep-ph]]
* [38] J. Baron, S. Marzani and V. Theeuwes, _Soft-Drop Thrust_ , JHEP 08 (2018), 105, [arXiv:1803.04719 [hep-ph]], [Erratum: JHEP 05, 056 (2019)]
* [39] S. Marzani, D. Reichelt, S. Schumann, G. Soyez and V. Theeuwes, _Fitting the Strong Coupling Constant with Soft-Drop Thrust_ , JHEP 11 (2019), 179, [arXiv:1906.10504 [hep-ph]]
* [40] J. Baron, D. Reichelt, S. Schumann, N. Schwanemann and V. Theeuwes, _Soft-drop grooming for hadronic event shapes_ , JHEP 07 (2021), 142, [arXiv:2012.09574 [hep-ph]]
* [41] Y. Makris, _Revisiting the role of grooming in DIS_ , Phys. Rev. D 103 (2021), no. 5, 054005, [arXiv:2101.02708 [hep-ph]]
* [42] M. Arratia, Y. Makris, D. Neill, F. Ringer and N. Sato, _Asymmetric jet clustering in deep-inelastic scattering_ , Phys. Rev. D 104 (2021), no. 3, 034005, [arXiv:2006.10751 [hep-ph]]
* [43] _The S HERPA -3.0.beta code can be obtained from: https://sherpa-team.gitlab.io/changelog.html_
* [44] E. Bothmann et al., The Sherpa collaboration, _Event Generation with Sherpa 2.2_ , SciPost Phys. 7 (2019), no. 3, 034, [arXiv:1905.09127 [hep-ph]]
* [45] M. Schönherr, _An automated subtraction of NLO EW infrared divergences_ , Eur. Phys. J. C 78 (2018), no. 2, 119, [arXiv:1712.07975 [hep-ph]]
* [46] M. Schönherr, _Next-to-leading order electroweak corrections to off-shell WWW production at the LHC_ , JHEP 07 (2018), 076, [arXiv:1806.00307 [hep-ph]]
* [47] S. Bräuer, A. Denner, M. Pellen, M. Schönherr and S. Schumann, _Fixed-order and merged parton-shower predictions for WW and WWj production at the LHC including NLO QCD and EW corrections_ , JHEP 10 (2020), 159, [arXiv:2005.12128 [hep-ph]]
* [48] E. Bothmann and D. Napoletano, _Automated evaluation of electroweak Sudakov logarithms in Sherpa_ , Eur. Phys. J. C 80 (2020), no. 11, 1024, [arXiv:2006.14635 [hep-ph]]
* [49] E. Bothmann, D. Napoletano, M. Schönherr, S. Schumann and S. L. Villani, _Higher-order EW corrections in ZZ and ZZj production at the LHC_ , JHEP 06 (2022), 064, [arXiv:2111.13453 [hep-ph]]
* [50] https://yaml.org/
* [51] C. Bierlich et al., _Robust Independent Validation of Experiment and Theory: Rivet version 3_ , SciPost Phys. 8 (2020), 026, [arXiv:1912.05451 [hep-ph]]
* [52] M. Cacciari, G. P. Salam and G. Soyez, _FastJet User Manual_ , Eur. Phys. J. C72 (2012), 1896, [arXiv:1111.6097 [hep-ph]]
* [53] S. Schumann and F. Krauss, _A Parton shower algorithm based on Catani-Seymour dipole factorisation_ , JHEP 03 (2008), 038, [arXiv:0709.1027 [hep-ph]]
* [54] S. Höche, F. Krauss, M. Schönherr and F. Siegert, _QCD matrix elements + parton showers: The NLO case_ , JHEP 04 (2013), 027, [arXiv:1207.5030 [hep-ph]]
* [55] S. Höche, F. Krauss, S. Schumann and F. Siegert, _QCD matrix elements and truncated showers_ , JHEP 05 (2009), 053, [arXiv:0903.1219 [hep-ph]]
* [56] F. Buccioni, J.-N. Lang, J. M. Lindert, P. Maierhöfer, S. Pozzorini, H. Zhang and M. F. Zoller, _OpenLoops 2_ , Eur. Phys. J. C 79 (2019), no. 10, 866, [arXiv:1907.13071 [hep-ph]]
* [57] A. Denner, S. Dittmaier and L. Hofer, _Collier: a fortran-based Complex One-Loop LIbrary in Extended Regularizations_ , Comput. Phys. Commun. 212 (2017), 220–238, [arXiv:1604.06792 [hep-ph]]
* [58] T. Gleisberg and S. Höche, _Comix, a new matrix element generator_ , JHEP 12 (2008), 039, [arXiv:0808.3674 [hep-ph]]
* [59] A. Buckley, J. Ferrando, S. Lloyd, K. Nordström, B. Page, M. Rüfenacht, M. Schönherr and G. Watt, _LHAPDF6: parton density access in the LHC precision era_ , Eur. Phys. J. C 75 (2015), 132, [arXiv:1412.7420 [hep-ph]]
* [60] R. D. Ball et al., The PDF4LHC Working Group collaboration, _The PDF4LHC21 combination of global PDF fits for the LHC Run III_ , J. Phys. G 49 (2022), no. 8, 080501, [arXiv:2203.05506 [hep-ph]]
* [61] E. Bothmann, M. Schönherr and S. Schumann, _Reweighting QCD matrix-element and parton-shower calculations_ , Eur. Phys. J. C 76 (2016), no. 11, 590, [arXiv:1606.08753 [hep-ph]]
* [62] J.-C. Winter, F. Krauss and G. Soff, _A Modified cluster hadronization model_ , Eur. Phys. J. C 36 (2004), 381–395, [hep-ph/0311085]
* [63] T. Gleisberg, S. Höche, F. Krauss, M. Schönherr, S. Schumann, F. Siegert and J. Winter, _Event generation with SHERPA 1.1_ , JHEP 02 (2009), 007, [arXiv:0811.4622 [hep-ph]]
* [64] G. Marchesini and B. R. Webber, _Simulation of QCD Jets Including Soft Gluon Interference_ , Nucl. Phys. B 238 (1984), 1–29
* [65] B. R. Webber, _A QCD Model for Jet Fragmentation Including Soft Gluon Interference_ , Nucl. Phys. B 238 (1984), 492–528
* [66] D. Amati and G. Veneziano, _Preconfinement as a Property of Perturbative QCD_ , Phys. Lett. B 83 (1979), 87–92
* [67] B. Andersson, G. Gustafson, G. Ingelman and T. Sjostrand, _Parton Fragmentation and String Dynamics_ , Phys. Rept. 97 (1983), 31–145
* [68] M. Krishnamoorthy, H. Schulz, X. Ju, W. Wang, S. Leyffer, Z. Marshall, S. Mrenna, J. Müller and J. B. Kowalkowski, _Apprentice for Event Generator Tuning_ , EPJ Web Conf. 251 (2021), 03060, [arXiv:2103.05748 [hep-ex]]
* [69] C. Adloff et al., The H1 collaboration, _Measurements of transverse energy flow in deep inelastic scattering at HERA_ , Eur. Phys. J. C 12 (2000), 595–607, [hep-ex/9907027]
* [70] I. Abt et al., The H1 collaboration, _Energy flow and charged particle spectrum in deep inelastic scattering at HERA_ , Z. Phys. C 63 (1994), 377–390
* [71] M. Derrick et al., The ZEUS collaboration, _Measurement of multiplicity and momentum spectra in the current fragmentation region of the Breit frame at HERA_ , Z. Phys. C 67 (1995), 93–108, [hep-ex/9501012]
* [72] S. Aid et al., The H1 collaboration, _Charged particle multiplicities in deep inelastic scattering at HERA_ , Z. Phys. C 72 (1996), 573–592, [hep-ex/9608011]
* [73] S. Aid et al., The H1 collaboration, _A Study of the fragmentation of quarks in $e^{-}p$ collisions at HERA_, Nucl. Phys. B 445 (1995), 3–21, [hep-ex/9505003]
* [74] C. Adloff et al., The H1 collaboration, _Evolution of e p fragmentation and multiplicity distributions in the Breit frame_ , Nucl. Phys. B 504 (1997), 3–23, [hep-ex/9707005]
* [75] M. Dasgupta and G. P. Salam, _Resummed event shape variables in DIS_ , JHEP 08 (2002), 032, [hep-ph/0208073]
* [76] I. W. Stewart, F. J. Tackmann and W. J. Waalewijn, _N-Jettiness: An Inclusive Event Shape to Veto Jets_ , Phys. Rev. Lett. 105 (2010), 092002, [arXiv:1004.2489 [hep-ph]]
* [77] T. T. Jouttenus, I. W. Stewart, F. J. Tackmann and W. J. Waalewijn, _The Soft Function for Exclusive N-Jet Production at Hadron Colliders_ , Phys. Rev. D 83 (2011), 114030, [arXiv:1102.4344 [hep-ph]]
* [78] Z.-B. Kang, S. Mantry and J.-W. Qiu, _N-Jettiness as a Probe of Nuclear Dynamics_ , Phys. Rev. D 86 (2012), 114011, [arXiv:1204.5469 [hep-ph]]
* [79] Z.-B. Kang, X. Liu, S. Mantry and J.-W. Qiu, _Probing nuclear dynamics in jet production with a global event shape_ , Phys. Rev. D 88 (2013), 074020, [arXiv:1303.3063 [hep-ph]]
* [80] D. Kang, C. Lee and I. W. Stewart, _Analytic calculation of 1-jettiness in DIS at $\mathcal{O}\left({\alpha}_{s}\right)$_, JHEP 11 (2014), 132, [arXiv:1407.6706 [hep-ph]]
* [81] Z.-B. Kang, X. Liu and S. Mantry, _1-jettiness DIS event shape: NNLL+NLO results_ , Phys. Rev. D 90 (2014), no. 1, 014041, [arXiv:1312.0301 [hep-ph]]
* [82] D. Kang, C. Lee and I. W. Stewart, _DIS Event Shape at N3LL_ , PoS DIS2015 (2015), 142
* [83] C. Frye, A. J. Larkoski, M. D. Schwartz and K. Yan, _Factorization for groomed jet substructure beyond the next-to-leading logarithm_ , JHEP 07 (2016), 064, [arXiv:1603.09338 [hep-ph]]
* [84] A. H. Hoang, S. Mantry, A. Pathak and I. W. Stewart, _Nonperturbative Corrections to Soft Drop Jet Mass_ , JHEP 12 (2019), 002, [arXiv:1906.11843 [hep-ph]]
* [85] N. Baberuxki, C. T. Preuss, D. Reichelt and S. Schumann, _Resummed predictions for jet-resolution scales in multijet production in e +e- annihilation_, JHEP 04 (2020), 112, [arXiv:1912.09396 [hep-ph]]
* [86] M. Ritzmann and W. J. Waalewijn, _Fragmentation in Jets at NNLO_ , Phys. Rev. D 90 (2014), no. 5, 054029, [arXiv:1407.3272 [hep-ph]]
* [87] S. Caletti, O. Fedkevych, S. Marzani, D. Reichelt, S. Schumann, G. Soyez and V. Theeuwes, _Jet angularities in Z+jet production at the LHC_ , JHEP 07 (2021), 076, [arXiv:2104.06920 [hep-ph]]
* [88] S. Caletti, O. Fedkevych, S. Marzani and D. Reichelt, _Tagging the initial-state gluon_ , Eur. Phys. J. C 81 (2021), no. 9, 844, [arXiv:2108.10024 [hep-ph]]
* [89] T. Gleisberg and F. Krauss, _Automating dipole subtraction for QCD NLO calculations_ , Eur. Phys. J. C 53 (2008), 501–523, [arXiv:0709.2881 [hep-ph]]
* [90] S. Actis, A. Denner, L. Hofer, J.-N. Lang, A. Scharf and S. Uccirati, _RECOLA: REcursive Computation of One-Loop Amplitudes_ , Comput. Phys. Commun. 214 (2017), 140–173, [arXiv:1605.01090 [hep-ph]]
* [91] B. Biedermann, S. Bräuer, A. Denner, M. Pellen, S. Schumann and J. M. Thompson, _Automation of NLO QCD and EW corrections with Sherpa and Recola_ , Eur. Phys. J. C 77 (2017), 492, [arXiv:1704.05783 [hep-ph]]
* [92] F. Cascioli, P. Maierhofer and S. Pozzorini, _Scattering Amplitudes with Open Loops_ , Phys. Rev. Lett. 108 (2012), 111601, [arXiv:1111.5206 [hep-ph]]
* [93] M. Dobbs and J. B. Hansen, _The HepMC C++ Monte Carlo event record for High Energy Physics_ , Comput. Phys. Commun. 134 (2001), 41–46
* [94] M. Dasgupta, F. A. Dreyer, K. Hamilton, P. F. Monni, G. P. Salam and G. Soyez, _Parton showers beyond leading logarithmic accuracy_ , Phys. Rev. Lett. 125 (2020), no. 5, 052002, [arXiv:2002.11114 [hep-ph]]
* [95] J. R. Forshaw, J. Holguin and S. Plätzer, _Building a consistent parton shower_ , JHEP 09 (2020), 014, [arXiv:2003.06400 [hep-ph]]
* [96] F. Herren, S. Höche, F. Krauss, D. Reichelt and M. Schoenherr, _A new approach to color-coherent parton evolution_ , arXiv:2208.06057 [hep-ph]
* [97] M. van Beekveld and S. Ferrario Ravasio, _Next-to-leading-logarithmic PanScales showers for Deep Inelastic Scattering and Vector Boson Fusion_ , arXiv:2305.08645 [hep-ph]
* [98] S. Höche, D. Reichelt and F. Siegert, _Momentum conservation and unitarity in parton showers and NLL resummation_ , JHEP 01 (2018), 118, [arXiv:1711.03497 [hep-ph]]
* [99] C. Adloff et al., The H1 collaboration, _Measurement of charged particle transverse momentum spectra in deep inelastic scattering_ , Nucl. Phys. B 485 (1997), 3–24, [hep-ex/9610006]
|
School of Physics and Astronomy, University of Minnesota Twin Cities,
Minneapolis, Minnesota 55455, USA<EMAIL_ADDRESS>
# Evaluation of the mean excitation energies of gaseous and liquid argon
M. Strait
###### Abstract
Current and future experiments need to know the stopping power of liquid
argon. It is used directly in calibration, where commonly the minimum-ionizing
portion of muon tracks is used as a standard candle. Similarly, muon range is
used as a measure of muon energy. More broadly, the stopping power figures
into the simulation of all charged particles, and so uncertainty propagates
widely throughout data analysis of all sorts. The main parameter that controls
stopping power is the mean excitation energy, or I-value. Direct experimental
information for argon’s I-value come primarily from measurements of gaseous
argon, with a very limited amount of information from solid argon, and none
from liquid argon. A powerful source of indirect information is also available
from oscillator strength distribution calculations. We perform a new
calculation and find that from oscillator strength information alone, the
I-value of gaseous argon is $(187\pm 5)$ eV. In combination with the direct
measurements and other calculations, we recommend $(187\pm 4)$ eV for gaseous
argon. For liquid argon, we evaluate the difference in central value and
uncertainty incurred by the difference of phase and recommend $(197\pm 7)$ eV.
All uncertainties are given to 68% C.L.
Interaction of radiation with matter, Neutrino detectors 2212.06286
Evaluation of the mean excitation energy of liquid argon
M. Strait<EMAIL_ADDRESS>School of Physics and Astronomy, University of
Minnesota Twin Cities, Minneapolis, Minnesota 55455, USA
## 1 Introduction
The International Commission on Radiation Units and Measurements (ICRU)
recommended values and uncertainties for the mean excitation energy of gaseous
argon twice, once in 1984 [1] and again in 2016 [2]. The first of these was in
Report 37 which gives an evaluated value of $(188\pm 10$) eV. The uncertainty
is not at 68% C.L., but is a “[figure] of merit, arrived at by subjective
judgments, and with a meaning that is not easily defined.” The report further
explains that errors are given at roughly 90% C.L., and that one “could
convert them to ‘standard deviations’ by multiplying them by a factor of about
one half.” These two statements indicate difference confidence levels.
However, ICRU-37 derives from NBSIR 82-2550 [3], which gives only the first of
these uncertainty interpretations, so this note will consider ICRU-37
uncertainties to be at 90% C.L.; this also agrees with ICRU-90’s comments on
ICRU-37. The 68% C.L. uncertainty is therefore 6 eV.
ICRU-37 uses four experimental results for stopping power and range [4, 5, 6,
7] in their evaluation. Other methods of estimating I-values are cited, such
as the semi-empirical oscillator strength distribution calculated in Ref. [8],
but they are not used in the evaluation.
ICRU Report 90 [2] updates the I-value of gaseous argon to $(187\pm 3)$ eV,
where the uncertainty now means roughly the usual one standard deviation (“an
interval having a confidence of approximately 67%”). The only new experimental
result used is Ref. [9], which does not add much. The decrease in the
recommended uncertainty is almost entirely from inclusion of oscillator
strength distribution calculations.
This note re-evaluates the sources of information on gaseous argon’s I-value.
First, direct experimental evidence from stopping power and range measurements
are reviewed in section 2. Second, the state of oscillator strength
distribution calculations is reviewed, and a new calculation performed, in
section 3. Other indirect methods of estimating the I-value are also reviewed
in this section. In section 4 all of the information on the I-value of gaseous
argon is combined into a new recommended value and uncertainty. In the
following section 5, information on phase effects are reviewed and we evaluate
an I-value for liquid argon. Finally, the major implications of the
recommended I-value for liquid argon for an experiment such as DUNE [10] are
given in section 6.
## 2 Gaseous argon experiments
### 2.1 Brolley & Ribe 1955
Brolley & Ribe [4] measured the stopping power of deuterons in argon gas
relative to air. Deuterons with initial energy 10.05 MeV were sent through a
cell filled with argon, and the pressure in the cell was adjusted until a
downstream NaI(Tl) crystal registered a drop of 25% of the initial energy.
This is a stopping power measurement that samples the $dE/dx$ only for fast
projectiles. It therefore suffers less than a range measurement from
uncertainties related to slow particles, primarily in the difficulty of
evaluating shell corrections.
The authors do not directly report an I-value for argon, but rather quote an
“absolute stopping cross section $dE/dx$” of $(3.72\pm 0.08)\times
10^{-15}\,\mathrm{eV\,cm}^{2}$ for argon gas at 219.2 mmHg, i.e.
$56.1\,\mathrm{MeV\,g^{-1}\,cm^{2}}$. From this, the ICRU inferred an I-value
of $(190\pm 15)$ eV. The uncertainty in the cross section appears to have been
evaluated from a combination of the degree of consistency between runs at the
same argon pressure, 1.4%, and the uncertainty of the initial beam energy,
1.3%.
The authors are not clear about the confidence level of their uncertainties. A
reanalysis shows that their $\pm 0.08\times 10^{-15}\,\mathrm{eV\,cm}^{2}$
directly corresponds to ICRU’s $\pm 15$ eV. Given that the three uncertainties
quoted by the ICRU for Refs. [4, 5, 6] are $\pm 15$ eV, $\pm 7$ eV and $\pm
10$ eV, that these measurements have consistent central values, and that the
final ICRU recommendation is $10$ eV at 90% C.L. (“subjective”), it appears
that the ICRU has assumed that Brolley & Ribe’s $\pm 0.08\times
10^{-15}\,\mathrm{eV\,cm}^{2}$ is at roughly 68% C.L. This is the most
conservative likely interpretation (i.e. the other sensible choices are 90% or
95% confidence, and these would indicate smaller uncertainties), so the same
will be assumed in this note’s evaluation.
### 2.2 Martin & Northcliffe 1962
Martin & Northcliffe [5] measured energy loss of few-MeV alpha particles in
gaseous argon and report an I-value of $(190\pm 17)$ eV. The ICRU-37 table
shows this as $(190\pm 7)$ eV instead, as does predecessor document NBSIR
82-2550. As Ref. [5] unambiguously says “$\pm 17\,$eV” twice, this appears to
be a simple mistake, not a re-evaluation. With this misreading, Martin &
Northcliffe would be the best experimental result, while the correct
uncertainty makes it subdominant.
The authors give a detailed list of sources of uncertainty as well as a
discussion of theoretical difficulties in calculating corrections at low
energy needed to obtain the I-value for argon. The confidence level of their
result would seem to be subjective — “obtained by estimating the maximum and
minimum slopes consistent with the errors displayed.” This note’s evaluation
treats the uncertainty as being at 68% C.L., but this may be very
conservative. Like the previous experiment, this is a stopping power
measurement and is relatively less vulnerable to uncertainties related to very
slow particles.
### 2.3 Hanke & Bichsel 1970
In another stopping power measurement, Hanke & Bichsel [6] used alpha
particles from radioactive decay to measure the I-value of gaseous argon. The
authors quote 182 eV and 167 eV as their results, for two evaluations of shell
corrections. ICRU-37 uses a re-evaluation of $(188\pm 10)$ eV. Hanke & Bichsel
provides the most information on gaseous argon’s I-value (given that we have
used the correct uncertainty of 17 eV from Ref. [5]). From context, the ICRU
quoted uncertainty is probably meant to be at 68% C.L., but unfortunately it
is not clear how it was obtained.
### 2.4 Besenbacher et al 1979
Ref. [7] reports on a measurement of stopping power for protons in the range
40 keV to 1 MeV and alpha particles in the range 100 keV to 2.4 MeV. The
I-value of 194 eV is quoted with no error in their Table II. ICRU has not
evaluated an uncertainty either, nor is it clear how it would be done. At such
low energies, shell corrections and other complications are very important.
For the purpose of this note’s evaluation, it is assumed that this experiment
is a factor of several less precise than those that report uncertainties, and
an error of $\pm 30$ eV has been assigned. The final result below is
insensitive to the precise value of this error; so long as it is several times
larger than the uncertainty of the more reliable inputs, this note’s
evaluation of the central value and uncertainty are both unchanged to within
the precision displayed.
### 2.5 Baumgart et al 1983
Ref. [9] (not used in ICRU-37) reports on a measurement of the stopping power
of argon to protons of between 60 and 800 keV. Similarly to the previous
experiment, a value of 190 eV is given, but no uncertainty is quoted, and it
is not clear how one would be extracted. Shell corrections are, again, a major
concern. An error of $\pm 30$ eV has again been assigned for this note’s
evaluation.
## 3 Calculations for gaseous argon
### 3.1 Oscillator strength distribution
ICRU-37 says that the most reliable calculations of I-value come from semi-
empirical dipole oscillator strength distributions (OSD), i.e. the
photoabsorption cross section as a function of energy. ICRU-37 does not use
OSD calculations as part of its evaluation of the recommended I-value for
argon, but ICRU-90 does, with much of the reason for its small recommended
uncertainty of 3 eV being its adopted uncertainty of 2 eV for the calculation
of Kamakura et al 2006 [11], who reported 191 eV. However, neither this
reference, nor the older calculations from Kumar & Meath 1985 [12] and
Eggarter 1975 [8], report an uncertainty themselves. ICRU-90 explains its own
by saying that it is “based on those quoted for similar results.”
This is on very shaky ground. Similar results would only have similar
uncertainties if the uncertainties in the underlying photon cross section data
were similar between the various materials. But the underlying data for
Kamakura come from the review of Berkowitz 2002 [13], who cautions that
“Information on the oscillation strengths [of argon] is still rather limited”,
a warning that does not appear for similar cases (e.g. O, $\mathrm{O_{2}}$, N,
$\mathrm{N}_{2}$, Ne). Moreover, ICRU-90 averages several OSD calculations,
then expands all the errors such that the reduced $\chi^{2}$ is unity. But the
calculations are not independent, being based on mostly the same underlying
data, so they cannot be validly combined in this way.
The calculation of Kamakura is based on the recommended oscillator strengths
from Berkowitz. Here we will repeat the calculation, adding more recent
experimental data, and estimating uncertainties. The result is $(187\pm 5)$
eV, and is found by
$\log
I=\left.\left(\sum_{n}f_{n}\log(E_{n})+\int_{\mathrm{IP}}^{\infty}\frac{df}{dE}\log(E)dE\right)\middle/S(0)\right.,$
where
$S(0)=\sum_{n}f_{n}+\int_{\mathrm{IP}}^{\infty}\frac{df}{dE}dE,$
and in each case the sum is over discrete states and the integral is over the
continuum from the ionization potential to infinity. $E$ is the incoming
photon energy and $f$ is the oscillator strength, i.e.
$f=\frac{2\epsilon_{0}m_{e}c}{\pi e^{2}\hbar}\sigma,$
where $\sigma$ is the photoabsorption cross section, $\epsilon_{0}$ is the
permittivity of free space, $m_{e}$ is the mass of the electron, $c$ is the
speed of light, and $e$ is the elementary charge.
By the Thomas-Reiche-Kuhn sum rule, $S(0)=Z$, i.e. 18 for argon, up to
relativistic and multipole effects that are expected to be negligible for
low-$Z$ elements. We find $S(0)$ to be $17.8\pm 0.8$, in good agreement with
the sum rule. No explicit correction to the oscillator strengths to make
$S(0)$ equal to 18 is made, as the I-value is invariant under a uniform
scaling of oscillator strengths, and we do not impose additional constraints
that would motivate any non-uniform scaling.
The following discussion examines energy ranges from lowest to highest. Our
method for adopting central values for each part of the spectrum will closely
follow the choices and methods of Berkowitz’s review, updated with newer data.
Table 1: Oscillator strengths for discrete transitions. The rightmost column indicates the primary source of information. Uncertainties are judgments made here based on consistency between several measurements or as stated by the original experimental group [14][15]. Level | Energy (eV) | $f$ | Group
---|---|---|---
4s | 11.62 | $0.0580\pm 0.0034$ | Gibson and Risley
4s′ | 11.83 | $0.2214\pm 0.0068$ | Gibson and Risley
5s | 14.09 | $0.026\pm 0.003$ | Chan
3$\mathrm{\bar{d}}$ | 14.15 | $0.090\pm 0.009$ | Chan
5s′ | 14.26 | $0.012\pm 0.001$ | Chan
3d′ | 14.30 | $0.106\pm 0.011$ | Chan
3d | 13.86 | $0.00110\pm 0.00011$ | Chan
4d | 14.71 | $0.0019\pm 0.0002$ | Chan
6s | 14.85 | $0.0144\pm 0.0014$ | Chan
4$\mathrm{\bar{d}}$ | 14.86 | $0.0484\pm 0.0048$ | Chan
4d′ | 15.00 | $0.0209\pm 0.0021$ | Chan
6s′ | 15.02 | $0.0221\pm 0.0022$ | Chan
5d | 15.12 | $0.0041\pm 0.0008$ | Chan
7s | 15.19 | $0.0139\pm 0.0014$ | Chan
5$\mathrm{\bar{d}}$ | 15.19 | $0.043\pm 0.010$ | Chan
Others | 15.5 | $0.18\pm 0.02$ | Berkowitz
Table 2: Contributions to the uncertainty on the I-value in the OSD calculation, by energy range. Groups of ranges treated as fully correlated are shown without separating horizontal lines and with subtotals. The total uncertainty is found by adding each group in quadrature. Energy | Uncertainty (eV)
---|---
4s, 4s′ (Gibson and Risley) | 0.3
Other discrete below the IP (Chan) | 1.5
Unresolved discrete below the IP (Berkowitz) | 0.2
Window resonances 26–29 eV (Madden, Berrah) | 0.1
1s$\rightarrow$4pm, 3202.3 eV (Deslattes) | 00.01
15.7596–15.9371 (Berkowitz) | 0.2
15.9371–29.3295 (Samson, Carlson) | 2.7
29.3295–48.0 (Samson, Carlson) | 0.8
Subtotal | 3.6
48.0–79.3 (Watson, Samson, Suzuki) | 0.1
79.3–243 (Watson, Samson, Suzuki, Henke) | 0.1
Subtotal | 0.2
243–250 (Suzuki) | 00.02
243–336 (Suzuki, Chan) | 0.6
336–500 (Suzuki) | 1.1
500–929 (Suzuki) | 0.9
Subtotal | 2.5
929–3202 (Wuilleumier, Zheng, Henke, Suzuki) | 0.8
3206–10k (Wuilleumier, Zheng, Millar, McCrary) | 1.4
Subtotal | 2.2
10k–100k (Chantler) | 0.3
First, the discrete spectrum. The values listed in table 1 were chosen. The
uncertainties adopted are primarily those stated by the experimental groups.
As an exception, the uncertainties for the 5d and 5$\mathrm{\bar{d}}$ levels
are expanded by a factor of two because Berkowitz finds the values
suspiciously large. Within each of the two experimental groups, Chan et al
[15] and Gibson & Risley [14], we conservatively take the uncertainties to be
fully correlated. The final entry in the table is Berkowitz’s estimate for all
unresolved discrete transitions close to the ionization potential. Two
uncertainties are assigned to Berkowitz’s estimate, a 10% uncertainty fully
correlated to the discrete transitions attributed to Chan, since Berkowitz’s
number is an extrapolation based primarily on Chan’s measurements, and an
uncorrelated 10% uncertainty to cover the extrapolation procedure. Despite the
conservatism used in these decisions, the overall uncertainty to the I-value
from transitions below the ionization potential, shown in table 2, is only
$\pm 1.5$ eV, which is a minor contributor to the total.
[width=0.69]osd-figure3.pdf
Figure 1: Oscillator strength distribution for gaseous argon, 15–48 eV,
showing data of Samson 1966 [16], Carlson et al 1973 [17], Samson et al 1991
[18], and Samson & Stolte 2002 [19]. This note’s evaluation is shown as the
solid red line and Berkowitz 2002’s evaluation [13] in dot-dashed red. The
shaded energy range from 26.6 to 29.2 eV is handled specially; see the text.
The vertical line at 29.3295 eV is the boundary between polynomial fits used
in the evaluations. The bottom pane, on this plot and the following plots,
shows the fractional differences between the present evaluation and the
various data and other evaluations.
For the narrow energy range between the $\mathrm{{}^{2}P_{3/2}}$ and
$\mathrm{{}^{2}P_{1/2}}$ ionization potentials, Berkowitz assumes a constant
cross section of 20.75 Mb. We accept this and assign a 20% uncertainty to this
cross section to cover the data of Samson 1966 [16] as shown in figure 1. This
section does not contribute significantly to the I-value or its uncertainty.
Following Berkowitz, the continuum is partitioned into several energy ranges,
and the oscillator strength data fit to polynomials within these ranges. These
are of the form
$\mathrm{\frac{df}{dE}=\frac{eV}{Ry}}\sum_{i=2}^{7}a_{i}y^{i},$
where $\mathrm{eV/Ry}=1/13.606$, $y=15.9371\,\mathrm{eV}/E$. In Berkowitz’s
evaluation, a 4-term polynomial is used, i.e. $a_{6}=a_{7}=0$. Here, we use a
6-term polynomial for the first range and a 4-term polynomial for all the
rest.
Table 3: Coefficients for the piecewise polynomial fit to various energy ranges, used in the OSD calculation. Energy (eV) | $a_{2}$ | $a_{3}$ | $a_{4}$ | $a_{5}$ | $a_{6}$ | $a_{7}$
---|---|---|---|---|---|---
15.9371–29.3295 | $-74.283\,0$ | $\phantom{+}386.182$ | $-494.182$ | $-29.413\,5$ | $402.690$ | $-187.177$
29.3295–48.0 | $\phantom{+}122.781$ | $-890.881$ | $2\,080.83$ | $-1\,516.569$ | — | —
48.0–79.3 | $\phantom{+}16.428\,0$ | $-66.036\,0$ | $-10.122\,0$ | $\phantom{+}203.586$ | — | —
79.3–243 | $\phantom{+}10.457\,5$ | $\phantom{+}27.350\,8$ | $-512.203$ | $1\,122.41$ | — | —
336–500 | $\phantom{+}27.194\,2$ | $7\,158.62$ | $-134\,219$ | $\phantom{+}729\,590$ | — | —
500–929 | $-36.618\,3$ | $\phantom{+}13\,081.3$ | $-296\,263$ | $1\,967\,090$ | — | —
929–3202 | $\phantom{+}31.255\,2$ | $\phantom{+}280.033$ | $\phantom{+}732\,348$ | $-28\,474\,100$ | — | —
3206–10k | $-43.245\,8$ | $\phantom{+}207\,562$ | $-39\,331\,400$ | $3\,787\,040\,000$ | — | —
For the first range, 15.9371 to 29.3295 eV, the data of Samson & Stolte 2002
[19] has become available since Berkowitz’s review. See figure 1. This
motivates refitting, and we have found that a substantially better fit results
from adding two additional terms to Berkowitz’s polynomial. The coefficients
are shown in table 3. The region of window resonances from 26.6 to 29.2 eV is
excluded from the fit; these are treated as separate discrete transitions
below. Samson 1966 and Samson 2002 are used in the fit, but the Samson 1966
data from 29 to 30 eV is excluded from the fit for both this energy range and
the following range, as it is in substantial conflict with other measurements.
The data of Carlson et al 1973 [17] is not directly used, but is displayed to
illustrate the consistency and to guide the choice of adopted uncertainty.
Samson 1966 does not state an uncertainty except to suppose that the overall
error on the sum of all continuous oscillator strengths is about 5%. Samson
2002 says that uncertainties range from 1 to 3% but unfortunately does not
specify which energies have the lower or higher uncertainties. Carlson states
no uncertainties. The older two measurements are largely within 3% of Samson
2002 and within 3% of the fit, with the largest difference between the fit and
Samson 1966 being 6% around 26 eV and the largest difference for Carlson being
4% around 19 eV. We adopt 3% as the uncertainty for this range. Here and in
the following, the number given indicates the uncertainty on the overall
normalization of the range.
The window resonances evident in figure 1 are included as discrete transitions
with total oscillator strength of $-0.055$, following Berkowitz. Given that
two experimental groups, Madden et al 1969 [20] and Berrah et al 1996 [21]
have results within 10% of each other, we assign a 10% uncertainty. This is a
tiny contributor to the overall uncertainty on the I-value, and so it is not
examined more closely.
In the following range, 29.3295–48.0 eV, we performed the fit on the data of
Samson 1966, excluding 29–30 eV, Carlson 1973, excluding $>$37 eV where
evidently background dominates, Samson 1991, and Samson 2002. We note that the
very low point for 43.8 eV from Samson 1966 is a probable typo — the original
data table gives 23 cm-1, but 32 cm-1 would be much more consistent with both
the surrounding points and later experimental results. Since a 6-term
polynomial gave very similar results to a 4-term polynomial, the latter was
used; this reasoning holds for each of the following energy ranges as well.
Nearly the same result is obtained if Samson 2002 is fit alone. Given the
less-good agreement between experiments in this region, a 4% uncertainty is
adopted. We note the rather large fractional difference between our evaluation
and that of Berkowitz towards the end of this range. Our evaluation is clearly
a better fit to all of the data, but in any case the _absolute_ difference is
very small since there is little oscillator strength between 45 and 48 eV.
Since both the 15.9371–29.3295 eV and 29.3295–48.0 eV ranges are dominated by
the work of Samson, their uncertainties are treated as fully correlated. The
15.9371–48 eV range is the dominant contributor of uncertainty to the I-value,
giving $\pm 3.6$ eV.
[width=0.69]osd-figure4.pdf
Figure 2: Oscillator strength distribution for gaseous argon, 48–230 eV,
showing the data of Samson 1966 [16], Henke et al 1967 [22], Watson 1972 [23],
Samson et al 1991 [18], Samson & Stolte 2002 [19], and Suzuki & Saito 2005
[24]. The evaluation of Berkowitz 2002 [13] is also shown. The conventions are
the same as for figure 1. Note suppressed zero.
For 48.0 eV to 79.3 eV, we fit a 4-term polynomial to the data of Watson 1972
[23], Samson 2002 and Suzuki & Saito 2005 [24]. See figure 2. Watson states 3%
uncertainty. Suzuki says that uncertainties are between 0.05% and 2%, and
within 1% for most energies, but gives no information as to which energies
have which uncertainties. Accordingly, we assign 2% uncertainties in the fit.
The three experiments have good agreement, with only one point of Watson lying
more than 2% from the fit. With this robust confirmation from different
groups, we adopt a 2% uncertainty for this range.
From 79.3 to 243 eV, the data of the same three groups is fit, plus that of
Henke et al 1967 [22]. This latter data is treated as having 5% uncertainties,
despite the very small errors reported by the authors in this unpublished
report. There is somewhat more disagreement between the various results in
this range, and so we adopt a 3% uncertainty. As nearly all the information
comes from the same groups as the previous energy range, the two are treated
as fully correlated.
[width=0.69]osd-figure5.pdf
Figure 3: Oscillator strength distribution for gaseous argon, 230–500 eV,
showing the data of Henke et al 1967 [22], Chan et al 1992 [15], and Suzuki &
Saito 2005 [24], as well as the evaluations of Henke et al 1993 [25] and
Berkowitz 2002 [13]. Between 230 and 336 eV, Berkowitz uses the data of Chan
1992 without a functional form.
From 243 to 250 eV, the L-edge, we directly use the data of Suzuki 2005 shown
in figure 3. From 250 to 336 eV, we directly average Suzuki 2005 with Chan
1992, without any functional form, scaled by their stated uncertainties.
Although Chan’s data covers the entire range from the ionization potential to
500 eV, we do not display it for the L-edge because with a resolution of 1 eV,
it lacks the ability to resolve the structure of the edge. (We have not
displayed it for lower energies because it would add little information.)
From 336 to 500 eV, we display Chan’s data, but do not use it. Above 336 eV,
it rises steadily away from the work of other groups, and, as Berkowitz
states, its use in this region would give too much contribution to the sum
rule. We fit the data of Suzuki to a polynomial, as above. From 243 to 500 eV,
we assign an uncertainty of 5%, given that the evaluation rests heavily on a
single group’s data, Suzuki’s, and the main second source, Chan, does not
agree very well, even in the 250–336 eV range. Two points from Henke 1967 do
agree well with Suzuki, but this is quite sparse. We also display on the plot
the Henke et al 1993 [25] evaluation for comparison only.
[width=0.69]osd-figure6.pdf
Figure 4: Oscillator strength distribution for gaseous argon, 500–10,000 eV,
with data of Wuilleumier 1965 [26], Henke et al 1967 [22], McCrary et al 1970
[27], Millar & Greening 1974 [28], Suzuki & Saito 2005 [24], and Zheng et al
2006 [29], and the evaluations of Henke et al 1993 [25], Chantler 1995 [30],
and Berkowitz 2002 [13]. Chantler 1993 is divided into the total cross
section, which is what attenuation experiments measure, and the
photoabsorption cross section, which is of interest for the OSD calculation.
The difference is used to correct the attenuation measurements of McCrary and
Millar.
From 500 to 929 eV, we again fit a polynomial to Suzuki’s data. See figure 4.
In this range there are four points by Henke 1967 which agree very well, and
we choose an uncertainty of 3%, fully correlated with the 243–500 eV range.
For this range and the following, The evaluation of Chantler et al 1995 [30]
is displayed along with Henke’s evaluation, for comparison.
From 929 to 3202 eV, just before the K-edge, we fit the data of Wuilleumier
1965 [26], Henke 1967, Suzuki 2005, and Zheng et al 2006 [29] to a polynomial.
Given the level of agreement between these several groups, we assign a 3%
uncertainty, uncorrelated with the previous regions.
Between 3202 and 3206 eV, we treat the oscillator strength as consisting of
the discrete 1s$\rightarrow$4pm resonance, measured by Deslattes et al 1983
[31], with a strength of 0.0022 and a 10% uncertainty. This is for
completeness only, as it has a negligible impact on the I-value or its
uncertainty.
From 3206 eV to 10 keV, we fit the data of Wuilleumier 1965, Zheng 2005,
Millar & Greening 1974 [28] and McCrary et al 1970 [27]. (Here we depart from
direct use of Berkowitz’s energy divisions by using a single fit for this
range, which Berkowitz divides into two sections.) For the higher energy data
of Millar and McCrary, we use the evaluations of Chantler to subtract the non-
photoabsorption portion of the cross section. Again an uncertainty of 3% is
assigned based on the agreement between groups. Given the overlap of the
experimental groups between the 3206 eV to 10 keV range and the 929 to 3202 eV
range, we treat these two ranges as fully correlated. This choice is not
crucial; a less conservative choice to consider them uncorrelated would lower
the uncertainty on the I-value by only 0.2 eV.
[width=0.69]osd-figure7.pdf
Figure 5: Oscillator strength distribution for gaseous argon from $10^{4}$ to
$10^{5}$ eV. The data of McCrary et al 1970 [27], and Millar & Greening 1974
[28] are shown, along with the evaluations of Henke 1993 [25] and Chantler
[30]. Subtraction to obtain oscillator strength from attenuation measurements
is done as in figure 4. The Chantler evaluation is used directly in the
present evaluation.
From 10 to 100 keV, we follow Berkowitz and directly use Chantler’s
evaluation. We assign a 3% uncertainty based on agreement with the data of
Millar and McCrary, being somewhat conservative because of the subtraction
procedure necessary to isolate the photoabsorption component of attenuation at
high energies. See figure 5.
Above 100 keV, we are unaware of any photoabsorption data, and we follow
Berkowitz, using the formula of Bethe & Salpeter [32], directly evaluating it
up to 1 GeV and then integrating the asymptotic form from there to infinity.
Because the range from 100 keV to infinity gives almost no contribution to the
I-value, changing it by only 0.1 eV if it is neglected entirely, no
uncertainty is assigned.
As stated above, the result for the I-value from combining all of these energy
ranges is $(187\pm 5)$ eV. There are several energy ranges each contributing
significantly to the uncertainty on the I-value. From the ionization potential
to 48 eV has the biggest contribution because of the large amount of
oscillator strength present. The range from the L-edge to 929 eV contributes
near the second most because of its fairly large oscillator strength and
larger fractional uncertainties. Similarly, the range just above the K-edge is
a significant contributor to the uncertainty. Discrete transitions below the
IP are next most important given their substantial contribution to the total
oscillator strength and 10% uncertainties. In contrast, the range 48–243 eV
has a very small contribution because of the fairly small oscillator strengths
coupled with low uncertainties, and above 10 keV, the contribution to the
uncertainty is small because the oscillator strength is small.
As a byproduct of this calculation, we can estimate the quantities $I(-1)$ and
$I(1)$, where $I(p)$ is defined as:
$\log
I(p)=\left.\left(\sum_{n}f_{n}E_{n}^{p}\log(E_{n})+\int_{\mathrm{IP}}^{\infty}\frac{df}{dE}E^{p}\log(E)dE\right)\middle/S(p)\right.,$
and
$S(p)=\sum_{n}f_{n}E_{n}^{p}+\int_{\mathrm{IP}}^{\infty}E^{p}\frac{df}{dE}dE,$
such that the I-value that is the focus of this note is $I(0)$.
We find $I(-1)=26.51\pm 0.28$. For $I(-1)$, the lower energies are more
important. The uncertainty comes almost entirely from discrete transitions
below the IP and from the region just above the L-edge, in equal proportions.
The result for $I(1)$ is $(3600\pm 80)\,\mathrm{eV}^{2}$. For $I(1)$, higher
energies are more important. Nearly all of the uncertainty comes from the
region above the L-edge ($\pm 60\,\mathrm{eV}^{2}$) and from the 10-100 keV
range where we use the evaluation of Chantler ($\pm 40\,\mathrm{eV}^{2}$).
This assumes an uncertainty below $\sim$10% above 100 keV where we use the
formula of Bethe and Salpeter. Unlike for $I(0)$, the region above 100 keV is
not negligible for $I(1)$; neglecting it reduces the result by
$200\,\mathrm{eV}^{2}$. For this reason, we do not venture an estimate for
$I(2)$ since it is even more sensitive to high energies and we are unaware of
any photoabsorption data above 100 keV where the majority of the uncertainty
is most likely to lie.
#### 3.1.1 Comparison with other results
In 2010, Kumar and Thakkar gave another estimate of argon’s I-value, 186.3 eV
with an estimated $\pm 2$% ($\pm 3.7\,$eV) uncertainty [33]. There are three
major differences between their evaluation and ours. First, their method uses
the sum rule and molar refractivity data as constraints. We choose to use
oscillator strengths alone, without constraints from molar refractivity data.
Second, for any given energy interval, they choose a single data set or
evaluated set of oscillator strengths for their fit. This has the effect of
discarding modern oscillator strength data in many energy ranges. For instance
from 319.9 eV to 100 keV, the 1973 evaluation of Veigele [34] is used even
though data from Millar 1974, Suzuki 2005 and Zheng 2006 all exist in this
range. In contrast, we include all modern data, and in each energy range use a
fit to all data judged as reliable.
Third, we display the underlying uncertainties assigned to each part of the
oscillator strength distribution and the effect that each has towards the
final uncertainty on the I-value. It is not clear how to trace Kumar and
Thakkar estimated uncertainty on the I-value back to the underlying data.
Despite all these differences, the present evaluation arrives at a very
similar result for the I-value. For $I(1)$, Kumar and Thakkar find
$3620\,\mathrm{eV}^{2}$ $\pm 3$% ($\pm 110\,\mathrm{eV}^{2}$), also very
similar to our result. For $I(-1)$, they find 26.53 $\pm 1$% ($\pm 0.27$),
essentially identical to our result.
### 3.2 Periodic trends
As an independent method of estimating the I-value, an interpolation can be
done from nearby elements. First, using the I-values for aluminum, silicon and
calcium, a simple interpolation on a plot of Z vs. I/Z can be made. The result
is 189 eV. I-values certainly do not lie on a smooth curve, so this result
cannot be taken too literally. Nevertheless, it would be surprising, given the
clear periodic trends, if argon’s I-value lay much above 200 eV or much below
170 eV, even in the abscense of other methods of evaluation.
[width=0.64]calci.pdf
Figure 6: Reproduction of ICRU-37 Fig 3.2 for $Z=1$–38. Note suppressed zero.
The vertical dashed line indicates argon’s atomic number.
A more sophisticated treatment is given in eq. 4.1 of ICRU-37,
$I_{\mathrm{int}}(Z)=\frac{I_{c}(Z)}{Z_{2}-Z_{1}}\left[\frac{I(Z_{1})}{I_{c}(Z_{1})}(Z_{2}-Z)+\frac{I(Z_{2})}{I_{c}(Z_{2})}(Z-Z_{1})\right],$
where $Z$ is the atomic number of the element whose I-value is to be
interpolated, $Z_{1}$ and $Z_{2}$ are the atomic numbers of the next lower and
next higher element with experimentally determined I-values, $I_{c}$ indicates
a calculated I-value and a bare $I$ indicates an experimental I-value. ICRU’s
choice of calculated I-values for gasses are the set from Chu & Powers 1972
[35], and for solids those from Ziegler 1980 [36], displayed in ICRU-37 Fig
3.2 and reproduced here in figure 6. Another set of calculated I-values are
those from Dehmer et al 1975 [37] and Inokuti et al 1981 [38], where the
former covers $Z=1$–18 and the latter 19–38; these are also shown in figure 6.
Although ICRU does not calculate an interpolated value for argon, as it has
experimental data, the result is 194 eV, given that $Z_{1}=14$ (silicon),
$I(Z_{1})=173$ eV, $I_{c}(Z_{1})=123.5$ eV, $Z_{2}=20$ (calcium),
$I(Z_{2})=191$ eV, $I_{c}(Z_{2})=147.9$ eV, and $I_{c}(Z)=146.3$ eV. This
procedure accounts for the phase of the substance. If argon is treated as a
solid, then $I_{c}(Z)=147.9$ eV instead, and the result is 196 eV.
The uncertainties on the experimental data of silicon and calcium (the nearest
elements with experimental data on each side of argon) contribute 3 eV to the
uncertainty of this interpolation. The choice of theoretical input is much
more important. If, instead of the ICRU recommendation, Dehmer and Inokuti are
used (the unused curve of their Fig. 3.2), the result shifts 20 eV upwards to
214 eV. These inputs are assumed to be worse, but still represent a reasonable
calculation, and so we may qualitatively take the theory uncertainty to be
somewhat smaller than the difference. Therefore, this evaluation includes the
I-value interpolated from periodic trends as $(194\pm 12)$ eV.
### 3.3 Hartree-Fock wave functions
Only two calculations for argon listed in ICRU-90 come with uncertainties
stated by the original authors. One has such large uncertainties as to be
irrelevant. The other is Bell et al 1972 [39], which uses a method involving
Hartree-Fock wave functions. Since it does not share underlying data with the
OSD calculations, it can be included separately in this note’s evaluation.
Bell’s result, as given by the ICRU table, is $(174\pm 3.5)$ eV. What Bell
actually says is 12.8 Ry, that “the predictions of different representations
[agreed] to within 2% in all cases,” and that a “full error analysis […] is
beyond the scope of the present paper.” Bell goes on to compare the
calculation for helium with a more sophisticated treatment, finding that the
difference is 6% and “Errors in $I$ arising from the Hartree-Fock
approximation are probably similar for the other atoms considered here”. Since
both the 2% and 6% errors are relevant, an uncertainty of 11 eV is more
correct, and is used in the evaluation in this note.
## 4 Evaluation for gaseous argon
[width=0.64]argonIpdf.pdf
Figure 7: Evaluation of the mean excitation energy of gaseous argon, using
experimental results [4, 5, 6, 9, 7], interpolation from periodic trends, and
calculations [39, 11]. The ICRU-90 recommendation is shown for comparison,
interpreting their central value and uncertainty as a Gaussian PDF. All curves
except for the one for ICRU-90 share a normalization.
The five direct experimental results are combined with the three indirect
methods to produce this note’s evaluation (see figure 7). Each underlying
result is represented as a PDF which is the sum of a Gaussian plus a uniform
distribution. The Gaussian’s mean and standard deviation are the quoted
central value and uncertainty of the given result. The uniform distribution
has the range 130–240 eV and a normalization representing a subjective
judgment about how likely the result is to be incorrect (through any means,
e.g., unaccounted for systematic error, incorrect calculation, typographic
error, etc.), typically 5–10%. This reflects the tendency of older papers (and
sometimes newer papers) to have results that are incompatible with each other
at many times the stated errors. The Gaussian is assigned the remainder of the
normalization.
The underlying PDFs are multiplied together to produce the evaluated PDF,
which is integrated from the mean to find the evaluated uncertainty. The
present evaluation for gaseous argon is $(187\pm 4)$ eV.
Since the method just described unfortunately must include subjective
estimates of the correctness of past experiments and calculations, a second
method was used to check how robust the result is. In this method, it is
assumed that exactly one of the input experiments or calculations is
incorrect. The average is taken with Gaussian distributions only, but with
each input experiment or calculation dropped in turn, producing several
results. A weighted average is then produced from these results with the
weights set by the uncertainty of the experiment or calculation that was
excluded. The result of this alternate procedure is $(187\pm 5)$ eV, which
does not differ significantly from the main procedure. If our own OSD
calculation is exempted from this procedure, the result is $(187\pm 4)$ eV,
identical to the main result.
Our answer is ultimately very similar to ICRU-90’s, with the same central
value and an uncertainty just 1 eV larger. It has, however, been arrived at
through a substantially different process. Our uncertainty is larger than
ICRU-90’s for several reasons: First, the ICRU misread the Martin &
Northcliffe’s error as 7 eV rather than 17 eV. Second, the present note’s OSD
evaluation is $(187\pm 5)$ eV, which has a larger uncertainty than ICRU
assigned to the several OSD calculations used in their evaluation. Finally,
ICRU-90 misunderstood the uncertainty of the Bell calculation, understating it
by a factor of three.
This note’s evaluation for gaseous argon is dominated by the OSD calculation
presented above. If it is excluded from the average, the result is instead
$(187\pm 6)$ eV. If only direct experimental evidence from range and stopping
power measurements is used, the value $(189\pm 8)$ eV. It can be seen that
there is good agreement among the several methods used to estimate the
I-value.
## 5 Evaluation for liquid argon
[width=0.64]phaseinterp.pdf
Figure 8: Effect of phase on I-value. The points with error bars are each
discussed in the text. For compounds water and the hydrocarbons, $Z$ is a
weighted average of the constituent elements. The atomic number of argon is
shown as a vertical dotted line. A smooth curve is drawn to give an estimate
of the effect for argon.
Liquid argon is not just a very dense gaseous argon; there is binding energy
associated with the phase change. Naively, if electrons are bound more
strongly, the stopping power should decrease. Data on this effect is limited,
but instructive. This effect is predicted to be large for strongly bound
systems such as metals, and smaller for molecular substances. Examples of the
latter:
* •
Hydrogen: ICRU-37 (Table 5.7) recommends 19.2 eV for gaseous $\mathrm{H_{2}}$
and 21.8 eV for liquid hydrogen (14% higher).
* •
Water: The ICRU-90 recommended value for liquid water is $(78\pm 2)$ eV.
ICRU-37 gives $(71.6\pm 2)$ eV for water vapor and $(75\pm 3)$ eV for liquid
water. The addendum for ICRU-73 [40] gives 69.1 eV for water vapor and 78 eV
for liquid water. ICRU-90 cites this without recommending a value for water
vapor. The phase effect on the I-value ranges from 5% to 13% depending on the
numbers chosen. For the purposes of this evaluation, the phase effect is
considered to be $(9\pm 4)$%.
* •
Nitrogen: ICRU-37 recommends $(82\pm 2)$ eV for gaseous $\mathrm{N}_{2}$ and
$(90.5\pm 2.6)$ eV for liquid nitrogen. After converting their 90% C.L. errors
to standard ones, this gives an increase of $(10\pm 3)$%.
* •
Oxygen: ICRU-37 recommends $(95\pm 2)$ eV for gaseous $\mathrm{O}_{2}$ and
$(104.3\pm 2.6)$ eV for liquid oxygen ($(10\pm 3)$%).
* •
n-propane, n-pentane, n-hexane, n-heptane: ICRU-37 recommends I-values
differing by 10% or 11% between the liquid and gas phases. All of the gaseous
measurements were performed by the same group, as were all of the liquid
measurements, so the uncertainties are considered fully correlated. Averaging
them gives $(11\pm 5)$%.
* •
Bromine: ICRU-37 recommends 343 eV for gaseous bromine and 357 eV for
condensed bromine, although both are interpolations from adjacent elements.
Taking into account the uncertainties of this procedure as was done for argon
in section 3.2, this is considered to be a change of $(4\pm 6)$%.
* •
Iodine: ICRU-37 recommends 474 eV for gaseous iodine and 491 eV for condensed
iodine. Using the same procedure as for bromine, an evaluation of $(4\pm 6)$%
is obtained.
None of these are noble gasses like argon, but several have similar boiling
points and are likewise non-polar, which implies that the relevant binding
energies are similar. The effect decreases with atomic number, as expected,
since a smaller fraction of the electrons participate in chemical binding. By
fitting a smooth curve to the changes as a function of $Z$ (see figure 8), we
would expect around a 7% increase in the I-value of argon. An uncertainty of
3% is estimated, based on the mean of the uncertainties for water, nitrogen
and oxygen, the closest three experimental data points.
There is a limited amount of experimental evidence for argon itself. As
summarized in ICRU-49 [41], two groups have studied the difference of stopping
power between solid and gaseous argon, but unfortunately only with alpha
particles below 3 MeV. Chu et al 1978 [42] found a 5–10% decrease in stopping
power for solid argon below 1.0 MeV, and none for 1–2 MeV. Besenbacher et al
1981 [43] found no phase effect to within their 3% uncertainty for 0.5–3.0 MeV
alphas, implying less than a 5% change in the mean excitation energy.
Moreover, Besenbacher 1981 points out that their measurements of solid argon
are compatible with Chu 1978; the apparent difference comes from the use of
different values for the gaseous stopping power. The situation remains a
little confused, but the result from Besenbacher alone still allows for a
change of the I-value of up to 9 eV. This is surprisingly small compared to
what’s expected from the results from the non-polar molecules listed above. It
may be that results from slow alpha particles do not reliably translate to
results for fast protons.
Solid argon likely has a very similar stopping power to liquid argon, as both
are condensed phases. Unfortunately, to the author’s knowledge, there are no
substances for which the I-value has been measured in both solid and liquid
phases. This lack of experimental evidence motivates conservatism in the error
assignments.
Since data from other molecules suggests an increase in the I-value of $(7\pm
3)$%, while data from solid argon suggests $(0\pm 5)\%$, it seems reasonable
to take the weighted average of these and arrive at an estimated I-value
increase of $(5\pm 3)$% for argon in a condensed phase. This gives $(197\pm
7)$ eV as the present evaluation for the mean excitation energy of liquid
argon.
## 6 Implications
Geant4 [44], by default, uses the (older) ICRU-37 central value for gaseous
argon, 188 eV, regardless of phase. In a liquid argon neutrino detector, the
two most important quantities that the I-value feed into are muon $dE/dx$ at
minimum ionization and muon range. The former is used for calorimetric
calibration and the latter to measure muon energy.
At muon minimum ionizing, around 270 MeV, changing the I-value from 188 eV to
197 eV decreases $dE/dx$ by 0.3%. This shift increases towards lower energy
and is 0.5% at 100 MeV. The effect on calibration is dependent on the details
of the procedure, but is perhaps 0.4%. Muon range at 1 GeV is increased by
0.3%. As with $dE/dx$, this shift increases as muon energy decreases, and is
0.6% at 100 MeV.
As examples of two other quantities that affect simulation and reconstruction,
the effect on proton $dE/dx$ and range is also calculated. Proton $dE/dx$ is
decreased by 0.7% at 50 MeV and 0.6% at 400 MeV. Across these energies, range
increases by between 0.6% and 0.7% with the largest change around 100 MeV.
Since the change in the I-value is 9 eV and the recommended uncertainty is 7
eV, the uncertainties on all these ranges and stopping powers are nearly as
large as the shifts quoted above. Notably, the effect of a change in I-value
on calorimetric energy calibration with muon $dE/dx$ and the muon energy
reconstruction with range have the same sign: the whole neutrino event energy
is overestimated by using gaseous argon’s I-value. Nor is there any
cancellation of this kind of uncertainty from use of a liquid argon near
detector.
Energy reconstruction has many uncertainties besides those from the I-value,
but the author’s estimate is that the I-value dominates muon energy estimation
uncertainty below 1 GeV. Most of the information about $\nu_{\mu}$ energy
comes from the muon, and $\Delta m^{2}_{32}$ is directly proportional to the
measured energies of oscillation minima and maxima. All other oscillation
parameters rest, if not as directly, on energy reconstruction as well. The
I-value is, therefore, a critical parameter for DUNE and other liquid argon
detectors.
To conclude, this note recommends a new value and uncertainty for the mean
excitation energies of gaseous and liquid argon, $(187\pm 4)$ eV and $(197\pm
7)$ eV, respectively. The central value for liquid argon is significantly
higher than that most recently recommended by the ICRU for gaseous argon, and
the uncertainty is substantially larger. While this recommendation is believed
to be a useful improvement, it is notable that it rests strongly on an
indirect calculation based on oscillator strength distributions. Direct
experimental evidence is also used, but none of the inputs are clean. Three of
the five experiments lack any original statement of uncertainty, while the
remaining two give numerical uncertainties, but without confidence levels. Of
these latter two, only one directly states an I-value. With so much freedom of
interpretation, another evaluation could easily arrive at a different result.
This fact should further motivate experimental work to improve the
uncertainty.
## 7 Acknowledgments
This work was supported by the U.S. Department of Energy. Thanks also to
Andrew Furmanski, Tom Junk, and Abigail Waldron for helpful discussions.
## References
* [1] International Commission on Radiation Units and Measurements, _Report 37: Stopping powers for electrons and positrons_ (1984).
* [2] International Commission on Radiation Units and Measurements, _Report 90: Key data for Ionizing-radiation dosimetry: measurement standards and applications_ (2016).
* [3] M.J. Berger and S.M. Seltzer, _National Bureau of Standards Internal Report 82-2550: Stopping powers and ranges of electrons and protons_ (1982).
* [4] J.E. Brolley and F.L. Ribe, _Energy loss by 8.86-Mev deuterons and 4.43-Mev protons_ , _Phys. Rev._ 98 (1955) 1112.
* [5] F.W. Martin and L.C. Northcliffe, _Energy loss and effective charge of He, C, and Ar ions below 10 MeV/amu in gases_ , _Phys. Rev._ 128 (1962) 1166.
* [6] C.C. Hanke and H. Bichsel, _Precision energy loss measurements for natural alpha particles in argon_ , _K. Danske Vidensk. Selsk., Mat.-Fys. Medd._ 38 (1970) .
* [7] F. Besenbacher, H.H. Andersen, P. Hvelplund and H. Knudsen, _Stopping power of swift hydrogen and helium ions in gasses_ , _K. Danske Vidensk. Selsk., Mat.-Fys. Medd._ 40 (1979) .
* [8] E. Eggarter, _Comprehensive optical and collision data for radiation action, II. Ar_ , _J. Chem. Phys._ 62 (1975) 883.
* [9] H. Baumgart, W. Arnold, H. Berg, E. Huttel and G. Clausnitzer, _Proton stopping power in various gasses_ , _Nucl. Inst. Meth._ 204 (1983) 597.
* [10] DUNE collaboration, _Deep Underground Neutrino Experiment (DUNE), far detector technical design report, volume I: Introduction to DUNE_ , _JINST_ 15 (2020) T08008 [2002.02967].
* [11] S. Kamakura, N. Sakamoto, H. Ogawa, H. Tsuchida and M. Inokuti, _Mean excitation energies for stopping power of atoms and molecules evaluated from oscillator-strength spectra_ , _J. Appl. Phys._ 100 (2006) .
* [12] A. Kumar and W.J. Meath, _Integrated dipole oscillator strength and dipole properties for Ne, Ar, Kr, Xe, HF, HCl, and HBr_ , _Canadian Journal of Chemistry_ 63 (1985) 1616.
* [13] J. Berkowitz, _Atomic and Molecular Photoabsorption: Absolute Total Cross Sections_ , Academic, San Diego (2002).
* [14] N.D. Gibson and J.S. Risley, _Absolute measurements of optical oscillator strengths of noble-gas resonance lines_ , _Phys. Rev. A_ 52 (1995) 4451.
* [15] W.F. Chan, G. Cooper, X. Guo, G.R. Burton and C.E. Brion, _Absolute optical oscillator strengths for the electronic excitation of atoms at high resolution. III. the photoabsorption of argon, krypton, and xenon_ , _Phys. Rev. A_ 46 (1992) 149.
* [16] J.A.R. Samson, _The measurement of the photoionization cross sections of the atomic gasses_ , vol. 2, Elsevier Science and Technology (1966).
* [17] R.W. Carlson, D.L. Judge, M. Ogawa and L.C. Lee, _Photoabsorption cross section of argon in the 180–700-angstrom wavelength region_ , _Appl. Opt._ 12 (1973) 409.
* [18] J.A.R. Samson, L. Lyn, G.N. Haddad and G.C. Angel, _Recent progress on the measurement of absolute atomic photoionization cross sections_ , _J. Phys. IV France_ 1 (1991) C1.
* [19] J.A.R. Samson and W.C. Stolte, _Precision measurements of the total photoionization cross-sections of He, Ne, Ar, Kr, and Xe_ , _Journal of Electron Spectroscopy and Related Phenomena_ 123 (2002) 265.
* [20] R.P. Madden, D.L. Ederer and K. Codling, _Resonances in the photo-ionization continuum of Ar i (20-150 ev)_, _Phys. Rev._ 177 (1969) 136.
* [21] N. Berrah, B. Langer, J. Bozek, T.W. Gorczyca, O. Hemmers, D.W. Lindle et al., _Angular-distribution parameters and R-matrix calculations of Ar resonances_ , _Journal of Physics B: Atomic, Molecular and Optical Physics_ 29 (1996) 5351\.
* [22] B.L. Henke, R.L. Elgin, R.E. Lent and R.B. Ledingham, “X-ray absorption in the 2-to-200 A region.” Air Force Office of Scientific Research report 67-1254, 1967.
* [23] W.S. Watson, _Photoionization of helium, neon and argon in the 60-230 eV photon energy range_ , _Journal of Physics B: Atomic and Molecular Physics_ 5 (1972) 2292.
* [24] I.H. Suzuki and N. Saito, _Total photoabsorption cross-section of Ar in the sub-keV energy region_ , _Radiation physics and chemistry (Oxford, England : 1993)_ 73 (2005) 1.
* [25] B.L. Henke, E.M. Gullikson and J.C. Davis, _X-ray interactions: Photoabsorption, scattering, transmission, and reflection at E = 50-30,000 eV, Z = 1-92_ , _Atomic Data and Nuclear Data Tables_ 54 (1993) 181.
* [26] F. Wuilleumier, _Analyse continue du spectre d’absorption de l’argon et du néon entre 2 et 8 Å_ , _J. Phys. France_ 26 (1965) 776.
* [27] J.H. McCrary, L.D. Looney, C.P. Constanten and H.F. Atwater, _Attenuation coefficients of gases for 4.5-145-keV photons_ , _Phys. Rev. A_ 2 (1970) 2489.
* [28] R.H. Millar and J.R. Greening, _Experimental x-ray mass attenuation coefficients for materials of low atomic number in the energy range 4 to 25 keV_ , _Journal of Physics B: Atomic and Molecular Physics_ 7 (1974) 2332.
* [29] L. Zheng, M. Cui, Y. Zhao, J. Zhao and K. Chen, _Total photoionization cross-sections of Ar and Xe in the energy range of 2.1–6.0keV_ , _Journal of Electron Spectroscopy and Related Phenomena_ 152 (2006) 143.
* [30] C.T. Chantler, _Theoretical form factor, attenuation, and scattering tabulation for Z=1–92 from E=1–10 eV to E=0.4–1.0 MeV_ , _Journal of Physical and Chemical Reference Data_ 24 (1995) 71 [https://pubs.aip.org/aip/jpr/article-pdf/24/1/71/9766843/71_1_online.pdf].
* [31] R.D. Deslattes, R.E. LaVilla, P.L. Cowan and A. Henins, _Threshold studies of a multivacancy process in the ${K}\beta$ region of argon_, _Phys. Rev. A_ 27 (1983) 923.
* [32] H.A. Bethe and E.E. Salpeter, _Quantum mechanics of one- and two-electron atoms_ , Springer-Verlag, Berlin (1957).
* [33] A. Kumar and A.J. Thakkar, _Dipole oscillator strength distributions with improved high-energy behavior: Dipole sum rules and dispersion coefficients for Ne, Ar, Kr, and Xe revisited_ , _The Journal of Chemical Physics_ 132 (2010) 074301.
* [34] W.J. Veigele, _Photon cross sections from 0.1 keV to 1 MeV for elements Z = 1 to Z = 94_ , _Atomic Data and Nuclear Data Tables_ 5 (1973) 51.
* [35] W.K. Chu and D. Powers, _Calculation of mean excitation energy for all elements_ , _Physics Letters A_ 40 (1972) 23.
* [36] J.F. Ziegler, _Handbook of stopping cross-sections for energetic ions in all elements_ (Jan, 1980), 10.1016/B978-0-08-021607-2.50001-5.
* [37] J.L. Dehmer, M. Inokuti and R.P. Saxon, _Systematics of moments of dipole oscillator-strength distributions for atoms of the first and second row_ , _Phys. Rev. A_ 12 (1975) 102.
* [38] M. Inokuti, J.L. Dehmer, T. Baer and J.D. Hanson, _Oscillator-strength moments, stopping powers, and total inelastic-scattering cross sections of all atoms through strontium_ , _Phys. Rev. A_ 23 (1981) 95.
* [39] R.J. Bell, D.R.B. Bish and P.E. Gill, _Separate subshell contributions to the stopping power of rare gases_ , _J. Phys. B_ 5 (1972) .
* [40] P. Sigmund, A. Schinner and H. Paul, _Errata and Addenda: ICRU Report 73 (Stopping of Ions Heavier than Helium)_ , 2009\.
* [41] International Commission on Radiation Units and Measurements, _Report 49: Stopping Powers and Ranges for Protons and Alpha Particles_ (1993).
* [42] W.K. Chu, M. Braun, J.A. Davies, N. Matsunami and D.A. Thompson, _Energy loss of He ions in solidified gases_ , _Nuclear Instruments and Methods_ 149 (1978) 115.
* [43] F. Besenbacher, J. Bøttiger, O. Graversen, J.L. Hansen and H. Sørensen, _Stopping power of solid argon for helium ions_ , _Nuclear Instruments and Methods in Physics Research_ 188 (1981) 657.
* [44] S. Agostinelli et al., _Geant4 — a simulation toolkit_ , _Nuclear Instruments and Methods in Physics Research Section A: Accelerators, Spectrometers, Detectors and Associated Equipment_ 506 (2003) 250\.
|
§ CONCLUSION
Within this paper, we proposed an online communication-efficient distributed changepoint detection method, and it can achieve similar performance as an idealistic setting but save many transmission costs. Numerically, we show that the local threshold and window size have an impact on the performance of our algorithm, and there is a trade-off in choosing a local threshold and window size. In application, we recommend choosing a large local threshold in general cases. But when the change is extremely small, the choice of the local threshold depends on the communication and storage budgets. If the communication budget is much more limited, choosing a large threshold with a large window size is sensible. If the storage cost is much more expensive, choosing a small threshold with small window size will approximately achieve the idealistic performance.
The violation of independent assumptions will negatively affect the power of our proposed method. We tried to solve this problem by inflating thresholds or estimating the long-run variance. Both ways can, to some extent, improve our algorithm when the auto-correlation problem is not severe. However, both approaches fail to detect changes in highly auto-correlated data. Therefore, one of the future research directions is how to detect change within highly auto-correlated data in real-time.
Yang gratefully acknowledges the financial support of the EPSRC via the STOR-i Centre for Doctoral Training (EP/S022252/1). This research was also supported by EPSRC grant EP/R004935/1 (Eckley) together with financial support from BT Research (Eckley, Fearnhead, Yang). The authors are also grateful to Lawrence Bardwell who played a key role in inspiring this work, and Dave Yearling (BT) for several helpful conversations that helped shape this research.
[1]
Adams, R. P. and MacKay, D. J. (2007).
Bayesian online changepoint detection.
arXiv preprint arXiv:0710.3742.
[2]
Alrashdi, I., Alqazzaz, A., Aloufi, E., Alharthi, R., Zohdy, M., and Ming, H. (2019).
Ad-iot: Anomaly detection of iot cyberattacks in smart city using machine learning.
In 2019 IEEE 9th Annual Computing and Communication Workshop and Conference (CCWC), pages 0305–0310.
[3]
Appadwedula, S., Veeravalli, V. V., and Jones, D. L. (2005).
Energy-efficient detection in sensor networks.
IEEE Journal on Selected Areas in Communications, 23(4):693–702.
[4]
Aue, A., Horváth, L., Kühn, M., and Steinebach, J. (2012).
On the reaction time of moving sum detectors.
Journal of Statistical Planning and Inference, 142(8):2271 – 2288.
[5]
Banerjee, T. and Veeravalli, V. V. (2015).
Data-efficient quickest change detection in sensor networks.
IEEE Transactions on Signal Processing, 63(14):3727–3735.
[6]
Chan, H. P. (2017).
Optimal sequential detection in multi-stream data.
The Annals of Statistics, 45(6):2736–2763.
[7]
Chen, Y., Wang, T., and Samworth, R. J. (2022).
High-dimensional, multiscale online changepoint detection.
Journal of the Royal Statistical Society Series B: Statistical Methodology, 84(1):234–266.
[8]
Cho, H. and Fryzlewicz, P. (2014).
Multiple-Change-Point Detection for High Dimensional Time Series via Sparsified Binary Segmentation.
Journal of the Royal Statistical Society Series B: Statistical Methodology, 77:475–507.
[9]
Enikeeva, F. and Harchaoui, Z. (2019).
High-dimensional change-point detection under sparse alternatives.
The Annals of Statistics, 47(4):2051–2079.
[10]
Fisch, A. T. M., Eckley, I. A., and Fearnhead, P. (2022).
A linear time method for the detection of collective and point anomalies.
Statistical Analysis and Data Mining: The ASA Data Science Journal, 15(4):494–508.
[11]
Gallant, A. R. (2009).
Nonlinear statistical models.
John Wiley & Sons.
[12]
Gösmann, J., Stoehr, C., Heiny, J., and Dette, H. (2022).
Sequential change point detection in high dimensional time series.
Electronic Journal of Statistics, 16(1):3608–3671.
[13]
Horváth, L. and Hušková, M. (2012).
Change-point detection in panel data.
Journal of Time Series Analysis, 33(4):631–648.
[14]
Horváth, L., Kühn, M., and Steinebach, J. (2008).
On the performance of the fluctuation test for structural change.
Sequential Analysis, 27(2):126–140.
[15]
Kengne, W. and Ngongo, I. S. (2022).
Inference for nonstationary time series of counts with application to change-point problems.
Annals of the Institute of Statistical Mathematics, 74(4):801–835.
[16]
Kiefer, N. M. and Vogelsang, T. J. (2002a).
Heteroskedasticity-autocorrelation robust standard errors using the Bartlett kernel without truncation.
Econometrica, 70(5):2093–2095.
[17]
Kiefer, N. M. and Vogelsang, T. J. (2002b).
Heteroskedasticity-autocorrelation robust testing using bandwidth equal to sample size.
Econometric Theory, 18(6):1350–1366.
[18]
Kirch, C. and Kamgaing, J. T. (2015).
On the use of estimating functions in monitoring time series for change points.
Journal of Statistical Planning and Inference, 161:25–49.
[19]
Kirch, C. and Weber, S. (2018).
Modified sequential change point procedures based on estimating functions.
Electronic Journal of Statistics, 12(1):1579–1613.
[20]
Kovács, S., Bühlmann, P., Li, H., and Munk, A. (2023).
Seeded binary segmentation: a general methodology for fast and optimal changepoint detection.
Biometrika, 110(1):249–256.
[21]
Leisch, F., Hornik, K., and Kuan, C.-M. (2000).
Monitoring structural changes with the generalized fluctuation test.
Econometric Theory, 16(6):835–854.
[22]
Liu, K., Zhang, R., and Mei, Y. (2019).
Scalable SUM-shrinkage schemes for distributed monitoring large-scale data streams.
Statistica Sinica, 29:1–22.
[23]
Mei, Y. (2005).
Information bounds and quickest change detection in decentralized decision systems.
IEEE Transactions on Information Theory, 51(7):2669–2681.
[24]
Mei, Y. (2010).
Efficient scalable schemes for monitoring a large number of data streams.
Biometrika, 97(2):419–433.
[25]
Mei, Y. (2011).
Quickest detection in censoring sensor networks.
In 2011 IEEE International Symposium on Information Theory Proceedings, pages 2148–2152.
[26]
Newey, W. K. and West, K. D. (1986).
A simple, positive semi-definite, heteroskedasticity and autocorrelationconsistent covariance matrix.
National Bureau of Economic Research Cambridge, Mass., USA.
[27]
Pinto, G. and Castor, F. (2017).
Energy efficiency: A new concern for application software developers.
Communications of the ACM, 60(12):68–75.
[28]
Rago, C., Willett, P., and Bar-Shalom, Y. (1996).
Censoring sensors: a low-communication-rate scheme for distributed detection.
IEEE Transactions on Aerospace and Electronic Systems, 32(2):554–568.
[29]
Romano, G., Eckley, I. A., Fearnhead, P., and Rigaill, G. (2023).
Fast online changepoint detection via functional pruning cusum statistics.
Journal of Machine Learning Research, 24:1–36.
[30]
Tartakovsky, A., Nikiforov, I., and Basseville, M. (2014).
Sequential analysis: Hypothesis testing and changepoint detection.
CRC press.
[31]
Tartakovsky, A. G. and Kim, H. (2006).
Performance of certain decentralized distributed change detection procedures.
In 2006 9th International Conference on Information Fusion, pages 1–8. IEEE.
[32]
Tartakovsky, A. G. and Veeravalli, V. V. (2002).
An efficient sequential procedure for detecting changes in multichannel and distributed systems.
In Proceedings of the Fifth International Conference on Information Fusion. FUSION 2002. (IEEE Cat.No.02EX5997), volume 1, pages 41–48 vol.1.
[33]
Truong, C., Oudre, L., and Vayatis, N. (2020).
Selective review of offline change point detection methods.
Signal Processing, 167:107299.
[34]
Tveten, M., Eckley, I. A., and Fearnhead, P. (2022).
Scalable change-point and anomaly detection in cross-correlated data with an application to condition monitoring.
The Annals of Applied Statistics, 16(2):721–743.
[35]
Varghese, B., Wang, N., Barbhuiya, S., Kilpatrick, P., and Nikolopoulos, D. S. (2016).
Challenges and opportunities in edge computing.
In 2016 IEEE International Conference on Smart Cloud (SmartCloud), pages 20–26.
[36]
Veeravalli, V. V. (2001).
Decentralized quickest change detection.
IEEE Transactions on Information Theory, 47(4):1657–1665.
[37]
Wang, T. and Samworth, R. J. (2018).
High dimensional change point estimation via sparse projection.
Journal of the Royal Statistical Society Series B: Statistical Methodology, 80(1):57–83.
[38]
Ward, K., Dilillo, G., Eckley, I., and Fearnhead, P. (2024).
Poisson-FOCuS: An efficient online method for detecting count bursts with application to gamma ray burst detection.
Journal of the American Statistical Association, page (to appear).
[39]
Weber, S. M. (2017).
Change-Point Procedures for Multivariate Dependent Data.
PhD thesis, Karlsruher Institut für Technologie (KIT).
[40]
White, H. and Domowitz, I. (1984).
Nonlinear regression with dependent observations.
Econometrica: Journal of the Econometric Society, pages 143–161.
[41]
Wu, H., Hu, J., Sun, J., and Sun, D. (2019).
Edge computing in an IoT base station system: Reprogramming and real-time tasks.
Complexity, 2019:4027638:1–4027638:10.
[42]
Xie, Y. and Siegmund, D. (2013).
Sequential multi-sensor change-point detection.
The Annals of Statistics, 41(2):670–692.
[43]
Yau, C. Y., Sze Him Isaac, L., and Ng, W. L. (2017).
Sequential change-point detection in time series models based on pairwise likelihood.
Statistica Sinica, 27.
[44]
Yu, Y., Madrid Padilla, O. H., Wang, D., and Rinaldo, A. (2023).
A note on online change point detection.
Sequential Analysis, 42(4):438–471.
[45]
Zeileis, A., Leisch, F., Kleiber, C., and Hornik, K. (2005).
Monitoring structural change in dynamic econometric models.
Journal of Applied Econometrics, 20(1):99–121.
[46]
Zhang, R. and Mei, Y. (2018).
Asymptotic statistical properties of communication-efficient quickest detection schemes in sensor networks.
Sequential Analysis, 37(3):375–396.
|
# Fast Methods for Computing Photometric Variability of Eccentric Binaries:
Boosting, Lensing, and Variable Accretion
Daniel J. D’Orazio Niels Bohr International Academy, Niels Bohr Institute,
Blegdamsvej 17, 2100 Copenhagen, Denmark<EMAIL_ADDRESS>Paul C.
Duffell Department of Physics and Astronomy, Purdue University, 525
Northwestern Avenue, West Lafayette, IN 47907-2036, USA<EMAIL_ADDRESS>Christopher Tiede Niels Bohr International Academy, Niels Bohr Institute,
Blegdamsvej 17, 2100 Copenhagen, Denmark<EMAIL_ADDRESS>
###### Abstract
We analyze accretion-rate time series for equal-mass binaries in co-planar
gaseous disks spanning a continuous range of orbital eccentricities up to
$0.8$, for both prograde and retrograde systems. The dominant variability
timescales match that of previous investigations; the binary orbital period is
dominant for prograde binaries with $e\gtrsim 0.1$, with a $5\times$ longer
“lump” period taking over for $e\lesssim 0.1$. This lump period fades and
drops from $5\times$ to $4.5\times$ the binary period as $e$ approaches $0.1$,
where it vanishes. For retrograde orbits, the binary orbital period dominates
at $e\lesssim 0.55$ and is accompanied by a $2\times$ longer-timescale
periodicity at higher eccentricities. The shape of the accretion-rate time
series varies with binary eccentricity. For prograde systems, the orientation
of an eccentric disk causes periodic trading of accretion between the binary
components in a ratio that we report as a function of binary eccentricity. We
present a publicly available tool, binlite, that can rapidly ($\lesssim 0.01$
sec) generate templates for the accretion-rate time series, onto either binary
component, for choice of binary eccentricity below $0.8$. As an example use-
case, we build lightcurve models where the accretion rate through the
circumbinary disk and onto each binary component sets contributions to the
emitted specific flux. We combine these rest-frame, accretion-variability
lightcurves with observer-dependent Doppler boosting and binary self-lensing.
This allows a flexible approach to generating lightcurves over a wide range of
binary and observer parameter space. We envision binlite as the access point
to a living database that will be updated with state-of-the-art hydrodynamical
calculations as they advance.
keywords
## 1 Introduction
The binary-gas-disk interaction arises in a multitude of astrophysical
environments. It is important for binary orbital evolution (see, e.g., Lai &
Muñoz, 2022): from sculpting planetary system architectures (e.g., Ward, 1997;
Nelson, 2018), to impacting stellar binary demographics (e.g., Valli et al.,
2024), to facilitating mergers of stellar-mass (e.g., Stone et al., 2017) and
supermassive (e.g., Begelman et al., 1980) black hole binaries (SMBHBs).
Gas disks also offer a way to observe such systems in the electromagnetic (EM)
spectrum through accretion. In the realm of stellar binary+disk systems, radio
and millimeter-wavelength observations have revealed a variety of disks
feeding and forming young stellar binaries as well as planetary systems (e.g.,
Tobin et al., 2016; Alves et al., 2019; Czekala et al., 2021; Zurlo et al.,
2023). In the time domain, photometric variability associated with periodic
accretion onto stellar binaries has been observed in multiple systems (e.g.,
Tofflemire et al., 2017a, b), providing further data on the disk+binary
interaction.
Resolving disks around SMBHBs is more difficult than in the stellar case due
to a lack of known sub-parsec-separation SMBHBs, uncertainty in the emission
structure surrounding the accreting SMBHB at the relevant wavelengths, and the
high spatial resolution needed for imaging of these distant sources (see
D’Orazio & Loeb, 2018, 2019). Similar to the stellar analogue, however,
accretion onto SMBHBs could be observable via bright, periodically modulated
EM emission. While there is no definitive evidence for the sub-parsec-
separation SMBHBs that will merge within the age of the Universe, accretion
rates onto these systems can be as high as for single SMBHs (e.g., Farris et
al., 2012; D’Orazio et al., 2013), suggesting that they could be a sub-
population of the quasars. In addition to being bright, such a binary-quasar
population could be identified by its periodic imprint on quasar lightcurves,
on $\sim$year or shorter timescales (Haiman et al., 2009; Kelley et al., 2021;
Xin & Haiman, 2021; Haiman et al., 2023). Such periodicity can arise from the
binary’s modulation of the accretion rate (e.g., Hayasaki & Mineshige, 2008;
MacFadyen & Milosavljević, 2008; D’Orazio et al., 2013; Farris et al., 2014),
or observer-dependent relativistic effects due to the binary orbit (D’Orazio
et al., 2015; D’Orazio & Di Stefano, 2018; Hu et al., 2020). Both offer a way
to identify such systems in photometric time-domain data, with multiple
searches having identified $\sim 250$ candidates to date (see D’Orazio &
Charisi, 2023, and references therein).
Time-domain searches require predictions for periodic signatures of binary
accretion and also characterization of the intrinsic variability noise (e.g.,
Vaughan et al., 2016; Zhu & Thrane, 2020, for the SMBHB case). Here we make a
step towards the former, by characterising the variable accretion rates of
eccentric binaries embedded in circumbinary disks (CBDs). We present an
analysis of accretion variability measured from 2D isothermal numerical
hydrodynamical calculations of gas disks accreting onto equal mass binaries,
for a continuous range of binary eccentricities $e\leq 0.8$ and for both
prograde D’Orazio & Duffell (2021, hereafter DD21) and retrograde
configurations of the binary and disk angular momentum Tiede & D’Orazio (2024,
hereafter TD23). We use this data to build a publicly available tool named
binlite that can rapidly generate accretion-rate time series data, for any
binary eccentricity in the simulated range via Fourier decompositions of the
simulation data.
Section 2 describes our methods while Sections 3.1, 3.2, and 3.3 present the
results of our periodicity analysis, accretion-rate time series
reconstruction, and calculation of preferential accretion rates. As an example
use-case and to demonstrate the wide range of periodic lightcurves that can
arise from accreting eccentric binaries, Section 4 presents a method for
generating light-curves of accreting black hole binaries, in a chosen
observing band, while including the observer-dependent relativistic effects of
Doppler boosting and gravitational self-lensing for multiple observer viewing
angles.
We envision this tool and the data it is built from as a starting point from
which further sophistication in numerical models and post-processing can be
added with the goal of generating a publicly available, living-lightcurve
database for modelling, interpreting, and searching for emission from
accreting binary systems. We discuss these future prospects and current
limitations in Section 5.
## 2 Methods
Throughout we consider a binary of total mass $M$, with equal mass components
(described by mass ratio $q\equiv M_{2}/M_{1}=1$), orbital eccentricity $e$,
semi-major axis $a$, and orbital angular frequency $\Omega_{b}$. A locally
isothermal, circumbinary disk accretes onto the binary and is modelled with
viscous hydrodynamics in the two dimensions in the plane of the binary orbit.
In this case the disk is characterized by the disk aspect ratio in vertical
hydrostatic equilibrium, $h$, which describes the relative importance of
pressure forces, and the kinematic coefficient of viscosity $\nu$. Because of
the simplified physics, one can scale results to any value of $M$ or $a$,
which amounts to choosing an orbital timescale via $\Omega_{b}$. Throughout we
scale the accretion rate by the equivalent steady-state value for a single
mass, $\dot{M}_{0}=3\pi\Sigma_{0}\nu$, for arbitrary surface-density scale
$\Sigma_{0}$. We consider both prograde and retrograde configurations of the
binary orbit with respect to the CBD. In both cases the equations of
hydrodynamics are solved using the moving-mesh code DISCO (Duffell, 2016).
#### Accretion From a Prograde Circumbinary Disk
The accretion-rate time-series data for prograde disks around eccentric
binaries is taken directly from the output of 2D numerical viscous
hydrodynamical calculations described in DD21. These calculations assume a
locally isothermal equation of state, which keeps the aspect-ratio a constant
value of $h=0.1$ in the circumbinary disk, and a spatially constant
coefficient of kinematic viscosity $\nu=10^{-3}a^{2}\Omega_{b}$. Specifically,
we utilize the main calculations in DD21, which evolve the binary and disk for
25,000 binary orbits, with the first 500 orbits relaxing the disk around a
binary on a circular orbit, and the following 20,000 binary orbits sweeping
the binary eccentricity linearly from $e=0$ to $e=0.9$. Here we utilize the
accretion rates measured via a sink prescription (Eq. 3 of DD21) onto each
component of the binary, as a function of time.
#### Accretion From a Retrograde Circumbinary Disk
The accretion-rate time-series data for retrograde disks around eccentric
binaries is taken directly from the output of 2D numerical viscous
hydrodynamical calculations described in TD23. These calculations assume the
same disk and binary parameters as the prograde case except that the binary
eccentricity is swept linearly from $e=0.0-0.8$ over a timescale of 10,500
orbits, with 500 orbits to relax the disk around a binary with a circular
orbit. We note that the accretion sinks in TD23 have half the characteristic
sink size and are implemented to be “torque free”, compared to the standard
sink implementation in DD21 (see Dempsey et al., 2020; Dittmann & Ryan, 2021,
for further clarificaiton on these sink types).
### 2.1 Accretion-Rate Variability Timescales
We first compute the dominant variability timescales as a function of binary
orbital eccentricity. We follow Duffell et al. (2020); TD23 and compute a 2D
periodogram of the accretion-rate time series by taking the norm of the
quantity,
$\mathcal{P}(e,\omega)=\frac{1}{\sqrt{2\pi\sigma^{2}_{\mathcal{P}}}}\int^{t(e_{f})}_{t(e_{0})}{\mathrm{e}^{-\frac{1}{2}\frac{\left(t(e)-\tau\right)^{2}}{\sigma^{2}_{\mathcal{P}}}}\dot{M}(\tau)\mathrm{e}^{-i\omega\tau}d\tau},$
(1)
which picks out the power in Fourier components with frequency $\omega$ in a
window of the accretion-rate time series centered on time $t(e)$ and of
characteristic width $2\sqrt{2\log 2}\sigma_{\mathcal{P}}$. For our choice of
$\sigma_{\mathcal{P}}=30(2\pi\Omega^{-1}_{b})$, this corresponds to a small
window of $\sim 70$ orbits in eccentricity centered around any $e$ in the time
series.
$\begin{array}[]{cc}\includegraphics[scale={0.6}]{fin_fig/edot2D_DoDuf21sweep_tround_adjst_Nouts600xNPs600}&\includegraphics[scale={0.6}]{fin_fig/edot2D_RetroDISCO_tround_adjst_Nouts300xNPs300_nmn0p1_nmx5p5}\end{array}$
Figure 1: 2D periodograms of the total accretion rate from prograde (left)
and retrograde (right) circumbinary disks onto binaries with orbital
eccentricity $e$ ranging from 0.0 to 0.8. The y-axis indicates the timescale
in units of orbital periods. Black indicates regions where there is no power,
while purple and yellow regions have increasingly more power. Both panels are
normalized to the same color scale.
### 2.2 Fourier Reconstruction
Our primary goal is to generate accretion-rate time series, which are periodic
over the binary orbital period, for any chosen value of the orbital
eccentricity. This is possible given the continuous sweep of our solutions
through binary parameter space. We start with the total accretion rate onto
the binary $\dot{M}(t)$ and its two components $\dot{M}_{1}(t)$ and
$\dot{M}_{2}(t)$ computed over the entire 25,000 (10,500) orbit sweep from the
numerical calculation of DD21 (TD23). Because these calculations carried out a
linear sweep in eccentricity with time, we also have a linear relation between
the time $t$ and the orbital eccentricity, $t(e)$.
The accretion-rate time series at a given orbital eccentricity, onto either
component of the binary, is reconstructed with a Fourier series,
$\displaystyle\dot{M}_{\rm{Rec}}(t,e)=\alpha_{0}$ $\displaystyle+$
$\displaystyle\sum_{n}\alpha_{n}(e)\cos{\left(n\Omega t\right)}$ (2)
$\displaystyle+$ $\displaystyle\sum_{n}\beta_{n}(e)\sin{\left(n\Omega
t\right)},$
for chosen fundamental frequency $\Omega$ and its integer multiples. That is,
a simple, yet accurate reconstruction is possible when the primary power is
concentrated at integer multiples of one frequency. A natural choice is the
binary frequency $\Omega_{b}$, and we show in Section 3.1 that this is indeed
the best choice except for a few regions of parameter space where a lower
frequency dominates, but still at approximate integer multiples of
$\Omega_{b}$.
The Fourier amplitudes are computed for the nearly continuous range of orbital
eccentricities by convolving the accretion-rate time series with a Gaussian
centered around a chosen eccentricity $e$ that arises at time $t(e)$ in the
eccentricity sweep,
$\displaystyle\alpha_{0}(e)$ $\displaystyle=$
$\displaystyle\frac{1}{\sqrt{2\pi\sigma^{2}}}\int^{t(e_{f})}_{t(e_{0})}\dot{M}(t^{\prime})e^{-\frac{1}{2}\left[\frac{\left(t^{\prime}-t(e)\right)^{2}}{\sigma^{2}}\right]}\
dt^{\prime}$ (3) $\displaystyle\alpha_{n}(e)$ $\displaystyle=$
$\displaystyle\frac{2}{\sqrt{2\pi\sigma^{2}}}\int^{t(e_{f})}_{t(e_{0})}\dot{M}(t^{\prime})e^{-\frac{1}{2}\left[\frac{\left(t^{\prime}-t(e)\right)^{2}}{\sigma^{2}}\right]}\cos\left({n\Omega
t^{\prime}}\right)\ dt^{\prime}$ $\displaystyle\beta_{n}(e)$ $\displaystyle=$
$\displaystyle\frac{2}{\sqrt{2\pi\sigma^{2}}}\int^{t(e_{f})}_{t(e_{0})}\dot{M}(t^{\prime})e^{-\frac{1}{2}\left[\frac{\left(t^{\prime}-t(e)\right)^{2}}{\sigma^{2}}\right]}\sin\left({n\Omega
t^{\prime}}\right)\ dt^{\prime},$
where $\sigma$ chooses the window width of $\dot{M}$ times-series data over
which to construct the Fourier series. In practice, we calculate
reconstructions up to $e=0.81$ for prograde systems and $e=0.791$ for
retrograde systems. Throughout we choose $\sigma=10(2\pi\Omega^{-1}_{b})$,
which corresponds to a Gaussian full width at half maximum of $\approx 23.5$
orbits.
Tracking accretion rates onto both components of an equal-mass-ratio binary is
necessary because, for prograde disks, some binary eccentricities excite disk
eccentricities that allow the accretion-rate to periodically favor one binary
component over the other (Dunhill et al., 2015; Muñoz & Lai, 2016; Siwek et
al., 2023b). The simulations of DD21 find that for $0.0\leq e\leq 0.18$ and
$e\geq 0.38$ the cavity is eccentric and precesses on super-orbital timescales
($\mathcal{O}(10^{2})$ orbital periods). This causes the accretion to favor
one binary component for approximately one half of the precession period of
the eccentric cavity. When binary and cavity eccentricity vectors pass through
a perpendicular configuration, the accretion-rate ratio quickly swaps to favor
the other component for the other half of the cavity precession period. Hence,
when the circumbinary cavity is eccentric and precessing, for prograde disks
around binaries with eccentricities in the range $0.0\leq e\leq 0.18$ and
$e\geq 0.38$, there are three possible accretion states: one where the primary
dominates accretion, one where the secondary dominates accretion, and one
shorter-lived stated where the two share the accretion rate as they swap
between the first two states. In these cases the accretion-rate ratio averages
to unity when taken over a disk precession period.
DD21 finds that prograde disks around binaries with eccentricities in the
range $0.18\leq e\leq 0.38$ are much more symmetric around the origin and
either do not precess (due to lack of disk eccentricity or to locking with the
binary eccentricity vector) or have much longer precession periods than for
higher or lower binary eccentricities. Similar observations were made from the
numerical calculations of Miranda et al. (2017) and Siwek et al. (2023b).
Siwek et al. (2023b) classify regions of parameter space where the disk
eccentricity is “locked” to the binary eccentricity, finding such a state for
$e=0.2$, $q=1$, and otherwise finding precessing states for
$e=0.0,0.4,0.6,0.8$, in agreement with DD21. Here we find that even in the
locked regime, the small asymmetry of the disk can still cause the accretion
rate to be spilt unequally between the binary components, but with a different
nature than for the precessing solutions. While the eccentric binary-disk
dynamics are worth understanding further in this regime, for the purposes of
this study, we note that even in this “symmetric" non-precessing disk state,
asymmetries arise that cause unequal accretion rates onto the binary
components (See Section 4).
For retrograde systems, where persistent disk eccentricities are not excited,
the accretion rates are always split evenly between the binary components.
$\begin{array}[]{cc}\includegraphics[scale={0.35}]{fin_fig/Progr_All_Mdot_recon_ecc0.0999619_Nfourier29}&\includegraphics[scale={0.35}]{fin_fig/Progr_All_Mdot_recon_ecc0.199592_Nfourier29}\vspace{-23pt}\\\
\includegraphics[scale={0.35}]{fin_fig/Progr_All_Mdot_recon_ecc0.299222_Nfourier29}&\includegraphics[scale={0.35}]{fin_fig/Progr_All_Mdot_recon_ecc0.399682_Nfourier29}\vspace{-23pt}\\\
\includegraphics[scale={0.35}]{fin_fig/Progr_All_Mdot_recon_ecc0.499312_Nfourier29}&\includegraphics[scale={0.35}]{fin_fig/Progr_All_Mdot_recon_ecc0.599773_Nfourier29}\\\
\includegraphics[scale={0.35}]{fin_fig/Progr_All_Mdot_recon_ecc0.699403_Nfourier29}&\includegraphics[scale={0.35}]{fin_fig/Progr_All_Mdot_recon_ecc0.799863_Nfourier29}\end{array}$
Figure 2: Match of accretion-rate templates constructed with our code binlite
(dashed curves), compared to the simulation output (solid curves). The total
accretion rate onto the binary is plotted in black, while the breakdown of
accretion onto each binary component is represented by the coral and teal
curves. Accretion rates are plotted in units of the steady-state value
$\dot{M}_{0}$. The time series is displayed over four binary orbits comprising
the center of a $\sim 20$ orbit segment from which the Fourier reconstruction
is built. Deviations between the reconstruction and simulation output for
$e=0.5,0.6,0.7$ derive from inter-orbital variability (peak-to-peak
variations) whose presence is apparent from the noisy region in the top-right
portion of the corresponding periodogram in Figure 1. The binary components
have equal masses but the eccentric cavity causes preferential accretion that
could equivalently be favoring either black hole. Hence, labeling of coral and
teal curves can be interchanged.
## 3 Results
### 3.1 Periodicity Analysis
#### Prograde Periodogram
The left panel of Figure 1 shows $|\mathcal{P}(e,\omega)|$ for prograde
accretion computed via Eq. (1) over a grid of $600\times 600$ values of
$\omega$ and $e$. Bright colors denote significant power while dark colors
denote lack of power at that corresponding point in parameter space – the
y-axis indicates the periodicity timescale in units of the binary orbital
period, and the x-axis indicates binary eccentricity.
For $e\lesssim 0.1$ we find that, in agreement with previous works (e.g.,
Muñoz & Lai, 2016; Miranda et al., 2017; Zrake et al., 2021), the power is
concentrated at the orbital period and its harmonics, but is dominant in a
small range of timescales centered around five binary orbital periods. This
corresponds to the timescale for the cavity “lump” to circulate and
periodically alter the feeding rate to the binary (MacFadyen & Milosavljević,
2008; D’Orazio et al., 2013).
For $0.05\lesssim e\lesssim 0.1$, power in the “lump-timescale" splits into
branches centered on the five-binary-orbit feature. This branching can be seen
most easily in the higher frequency ($1/2.5P^{-1}_{b}$) harmonic of the lump
timescale on the left side of the left panel of Figure 1. The dominant lump-
branch drops in frequency from $1/5P^{-1}_{b}$ to $1/4.5P^{-1}_{b}$ as $e$
increases from $0.05\rightarrow 0.1$. At $e=0.1$ the lump branches vanish and
are replaced by power that is almost entirely focused at the orbital timescale
and its higher harmonics. Orbital period timescales then dominate for
$e\gtrsim 0.1$. At higher orbital eccentricities, particularly for
$0.5\lesssim e\lesssim 0.8$, the power is noisy at lower frequencies. This
arises due to less stability in the accretion-rate time series from orbit to
orbit. We demonstrate this further in the next section.
Our primary takeaway for the purpose of generating reconstructions of the
accretion-rate time series is that the orbital period is a dominant feature
for all eccentricities $e\geq 0.1$, hence we choose $\Omega_{b}$ as our
fundamental frequency for the Fourier reconstruction, Eqs. (2) and (3). For
$e<0.1$, we choose $\Omega_{b}/5$ as the fundamental frequency, but with more
terms in the reconstruction.
#### Retrograde Periodogram
The right panel of Figure 1 shows $|\mathcal{P}(e,\omega)|$ for retrograde
accretion computed via Eq. (1) over a grid of $300\times 300$ values of
$\omega$ and $e$. A different version of this is also published in TD23 (over
a different range of timescales). Our main purpose for showing it here is to
emphasize that the orbital timescale periodicity is strong for all
eccentricities. However, in the retrograde case, a strong, two-times-orbital
periodicity arises for $e\sim 0.55$. Hence, for retrograde systems we choose
the fundamental Fourier reconstruction frequency to be $\Omega_{b}/2$. Note
that the retrograde periodogram is much less noisy than its prograde
counterpart, indicating steadier accretion rate-times series, even for high
eccentricities.
### 3.2 Accretion-Rate Time Series
#### Prograde Time Series
Figure 2 presents example accretion-rate time series for prograde binaries
with eight different values of orbital eccentricity. The solid lines show the
accretion rates measured directly from the numerical calculations while the
dashed lines are the accretion templates built from our Fourier reconstruction
(Eqs. (2) and (3)) using $\Omega=\Omega_{b}$ and a total of 30 Fourier
components. Accretion rates onto each component are denoted in coral and teal,
while the total is plotted in black. Vertical dotted lines denote the time of
pericenter while vertical dot-dashed lines denote apocenter. Note that even
though we have included enough Fourier components to capture sharp features in
the time series (e.g., the bottom right panel of Figure 2), small deviations
between the reconstructions (dashed) and the simulation (solid) are apparent
for $e\gtrsim 0.5$. This is due to the inter-orbit variability which manifests
as the noisy upper-right region in the prograde periodogram (left panel) of
Figure 1, i.e., at some eccentricities the accretion rate is less steady from
one orbit to the next, affecting our reconstructions which are built from an
average over $\sim 20$ orbits.
Figure 3 demonstrates a reconstruction for a prograde binary with $e=0.01$,
where the $\omega\sim\Omega_{b}/5$ periodicity of the cavity lump dominates.
In this case the Fourier reconstruction uses $\Omega=\Omega_{b}/5$ and $60$
Fourier components. Here, the more complex nature of the variability is
apparent in the less exact match of reconstruction and direct simulation (see
again the more complex structure in the $e<0.01$ portion of the left
periodogram in Figure 1). Despite this, the reconstruction captures the main
qualitative features of the time series, including crucially, the periodicity
at both $\Omega_{b}$ and $\Omega_{b}/5$, and reliable reconstruction of the
contribution of each component accretion rate to the total.
$\begin{array}[]{c}\includegraphics[scale={0.35}]{fin_fig/Progr_All_Mdot_recon_ecc0.00946463_Nfourier59}\end{array}$
Figure 3: The same as Figure 2, but for a prograde, $e=0.01$ system, where
variability at $\Omega_{b}$ and $\Omega_{b}/5$ co-exist.
#### Retrograde Time Series
Figure 4 shows the reconstructed total, primary, and secondary accretion rates
for retrograde binary-disk systems using $\Omega=\Omega_{b}/2$ and 30 terms in
the Fourier reconstruction. The reconstructed and simulated cases capture both
orbital and twice-orbital periodicity very well and result in nearly identical
reconstructed vs. simulated curves. This is due to the much more steady nature
of retrograde disk solutions across the parameter space. The total retrograde
accretion rates are described further in TD23, while here we additionally show
the component accretion rates. These are nearly identical to each other, as is
expected for equal-mass binaries when no other asymmetry arises, i.e., the
eccentric disk of the prograde case.
$\begin{array}[]{cc}\includegraphics[scale={0.35}]{fin_fig/Retro_All_Mdot_recon_ecc0.00948017_Nfourier29}&\includegraphics[scale={0.35}]{fin_fig/Retro_All_Mdot_recon_ecc0.099293_Nfourier29}\\\
\includegraphics[scale={0.35}]{fin_fig/Retro_All_Mdot_recon_ecc0.199916_Nfourier29}\vspace{-23pt}&\includegraphics[scale={0.35}]{fin_fig/Retro_All_Mdot_recon_ecc0.299708_Nfourier29}\\\
\includegraphics[scale={0.35}]{fin_fig/Retro_All_Mdot_recon_ecc0.3995_Nfourier29}\vspace{-23pt}&\includegraphics[scale={0.35}]{fin_fig/Retro_All_Mdot_recon_ecc0.499292_Nfourier29}\\\
\includegraphics[scale={0.35}]{fin_fig/Retro_All_Mdot_recon_ecc0.599916_Nfourier29}&\includegraphics[scale={0.35}]{fin_fig/Retro_All_Mdot_recon_ecc0.699708_Nfourier29}\end{array}$
Figure 4: The same as Figure 2 but for retrograde, eccentric binaries. Here
the accretion rates onto the primary (coral) and secondary (teal) are nearly
identical because there is no precessing, circumbinary cavity.
### 3.3 Accretion-Rate Ratio $Q$
Over long enough timescales ($\mathcal{O}(10^{2}-10^{3})$ binary orbits), the
accretion rates onto equal-mass binaries will average to unity (e.g., Siwek et
al., 2023a)111Though this may not be the case when binary eccentricity and
disk eccentricity vectors are locked relative to each other, as could be the
case for our $e\sim 0.2$ results – this requires further investigation..
However, when computing lightcurves, the relevant quantity is the accretion-
rate ratio between the binary components over timescales spanning orbital
periods. To quantify this, we define the ratio
$Q=\left<\dot{M}_{2}/\dot{M}_{1}\right>$, averaged over an integer number of
orbits for which the eccentric accretion-rate imbalance operates (much shorter
than a disk precession timescale).
The ratio of component accretion rates, $Q$ is always unity for retrograde,
equal-mass binaries (see Figure 4) but is a function of eccentricity for
prograde binaries. In Figure 5 we present $Q$ measured from the prograde
simulations as a function of binary eccentricity. The grey x’s represent 1000
values drawn from the reconstructed accretion rates averaged over 5 orbits,
sampled evenly in $e$. Because preferential accretion trades between primary
and secondary, we present the minimum of $Q$ and $Q^{-1}$, with both values
being valid choices when modelling equal-mass binaries considered here. When
plotted this way, the extreme values of the ratio of accretion rates can be
approximated by a simple function of eccentricity, inspired by the ratio of
pericenter and apocenter distances of the binary components,
$\displaystyle Q_{\rm{min}}$ $\displaystyle\approx$
$\displaystyle\frac{1-\mathcal{P}(e)e}{1+\mathcal{A}(e)e}$ (4)
$\displaystyle\mathcal{P}(e)$ $\displaystyle=$ $\displaystyle 2-e^{2}-2e^{3}$
$\displaystyle\mathcal{A}(e)$ $\displaystyle=$ $\displaystyle 2+e^{2},$
which is drawn as the solid blue line in Figure 5.
For most binary eccentricities the accretion-rate periodically switches back
and forth between favoring each of the binary components. Hence, the ratio of
accretion rates used to compute lightcurves can take values between
$Q_{\rm{min}}$ and $Q^{-1}_{\rm{min}}$. As can be seen from the density of
grey x’s in Figure 5, the binary spends more time accreting at some ratios
than others: $Q=1$ is sparsely sampled because this value is encountered
during the relatively rapid stage where the accretion rate switches from
favoring one component to favoring the other. The black points and associated
error bars show the average accretion ratio from a set of constant-
eccentricity verification runs (detailed in DD21) and the black triangles
illustrate the smallest value over the run. Those values of $Q$ that lie
within the error bars from these constant-eccentricity runs provide a good
indicator for where the binary spends most of its time accreting.
For intermediate values of eccentricity, the disk is more symmetric around the
origin and is either slowly precessing or not precessing at all. This results
in regions with narrow spreads of accretion-rate ratio $Q$, seen as the
clustering of grey x’s between $0.018\lesssim e\lesssim 0.38$ in Figure 5.
While the spread of $Q$ values is much smaller in this range than outside of
it, $Q$ is not continuous in $e$ here and jumps through four different states
of monotonically varying $Q(e)$, while also experiencing islands of
oscillating disk solutions near the transitions between these four states.
This behavior is likely due to locking of the angle between disk and binary
eccentricity and would be worth studying further for elucidating the
binary+disk dynamics in this regime and its relation to observability via,
e.g., accretion variability.
$\begin{array}[]{c}\includegraphics[scale={0.5}]{fin_fig/Qofecc_csteccs_Ne1000}\end{array}$
Figure 5: The ratio $Q\equiv\left<\dot{M}_{2}/\dot{M}_{1}\right>$ of
accretion rates onto each binary component, averaged over a duration shorter
than the disk precession frequency. The grey x’s are measured from our
reconstruction while the blue line is an analytic approximation for the
extreme values. As a check of the values derived from the eccentricity sweep,
the black points show results measured from simulations with fixed
eccentricities – black points with error bars represent the average and
standard deviation of $\rm{min}\left(Q,Q^{-1}\right)$ over a precession period
of the disk , and the triangles denote the smallest values over the same
range.
## 4 Application: Construction of Boosted and Lensed Lightcurves
As an example-use case and to demonstrate the wide range of lightcurve shapes
that can arise from accreting, eccentric binaries, we develop a simple
procedure for converting the accretion-rate time series of the previous
section to a rest-frame flux. Primarily for application to accreting black
hole binaries, we then convert this to an observer-dependent flux by including
relativistic orbital Doppler boosting of emission emanating from the minidisks
around each binary component (D’Orazio et al., 2015), as well as binary self-
lensing in the point mass, point-source limit (D’Orazio & Di Stefano, 2018; Hu
et al., 2020).
Doppler boosting and binary lensing have been put forth as mechanisms for
causing unique periodic variability in accreting SMBHB systems (D’Orazio et
al., 2015; D’Orazio & Di Stefano, 2018; Hu et al., 2020; Davelaar & Haiman,
2022; Ingram et al., 2021; D’Orazio & Charisi, 2023; Major Krauth et al.,
2023). Modeling these signatures consistently requires knowledge of the
fraction of light coming from each of the minidisks and the CBD. It should
also be combined with the intrinsic variability of the source in order to
understand systems where both hydrodynamical variability as well as observer
dependent effects are jointly operating.222A combination of Doppler boosting
and hydrodynamic variability near merger is simulated in Tang et al. (2018).
Hence, the reconstructed accretion-rate time series of the previous section
allow us to significantly build upon toy models for the variability of
accreting SMBHBs.
We model the emission from the accreting binary in a given frequency band as a
constant specific flux from the circumbinary disk, $F^{0}_{\nu,\mathrm{CBD}}$,
plus time-dependent emission from the primary and secondary minidisks. We
model the minidisk emission as a constant, average flux times a time-dependent
function $F^{0}_{\nu,1}p(t)$ and $F^{0}_{\nu,2}s(t)$, where
$p(t)\equiv\dot{M}_{1}/\left<\dot{M}_{1}+\dot{M}_{2}\right>$ and
$s(t)\equiv\dot{M}_{2}/\left<\dot{M}_{1}+\dot{M}_{2}\right>$ are the
reconstructed time-variable accretion rates computed with binlite (Section
3.2) and normalized by the average total accretion rate at that eccentricity.
#### Lightcurve Generation
With boosting and lensing taken into account the total observed flux is
$\displaystyle
F_{\nu}=F^{0}_{\nu,1}\mathcal{D}_{1}\mathcal{M}_{1}p(t)+F^{0}_{\nu,2}\mathcal{D}_{2}\mathcal{M}_{2}s(t)+F^{0}_{\nu,\mathrm{CBD}},$
(5)
where $\mathcal{D}_{i}\equiv D^{3-\alpha}_{i}$ is the time-dependent Doppler-
boost magnification for Doppler factor $D_{i}$ and frequency-dependent log-
spectral slope $\alpha$, in the observing band (assumed here to be the same
for all disk components), while $\mathcal{M}_{i}$ is the time-dependent
lensing magnification for the specified binary component (see D’Orazio & Di
Stefano, 2018; Hu et al., 2020). Defining
$\displaystyle F^{0}_{\nu,\mathrm{Tot}}$ $\displaystyle\equiv$ $\displaystyle
F^{0}_{\nu,1}+F^{0}_{\nu,2}+F^{0}_{\nu,\mathrm{CBD}},$
$\displaystyle\chi_{\nu,1}$ $\displaystyle\equiv$ $\displaystyle
F^{0}_{\nu,1}/F^{0}_{\nu,\mathrm{Tot}},$ $\displaystyle\chi_{\nu,2}$
$\displaystyle\equiv$ $\displaystyle F^{0}_{\nu,2}/F^{0}_{\nu,\mathrm{Tot}},$
(6)
we can write the observed in-band flux, normalized to the total average (rest
frame) flux as,
$\displaystyle\frac{F_{\nu}}{\left<F^{0}_{\nu,\mathrm{Tot}}\right>}$
$\displaystyle=$
$\displaystyle\chi_{\nu,1}\mathcal{D}_{1}\mathcal{M}_{1}p(t)+\chi_{\nu,2}\mathcal{D}_{2}\mathcal{M}_{2}s(t)$
(7) $\displaystyle+$ $\displaystyle(1-\chi_{\nu,1}-\chi_{\nu,2}),$
which reduces to
$\displaystyle\frac{F_{\nu}}{\left<F^{0}_{\nu,\mathrm{Tot}}\right>}=(1-\chi_{\nu,2})\mathcal{D}_{1}\mathcal{M}_{1}p(t)+\chi_{\nu,2}\mathcal{D}_{2}\mathcal{M}_{2}s(t),$
(8)
for a simplified case where the minidisks are assumed to outshine the CBD.
Hence, we need to know the relative fluxes, in a specified frequency band,
from each minidisk and the circumbinary disk. This requires knowing the disk
spectra, which depends on physics of radiative energy balance in the accretion
flow not captured by our isothermal simulations. For simplicity, we treat the
circumbinary disk and both minidisks as separate components and approximate
the spectra of each component with composite black-body spectra of the
optically thick alpha-disk solutions. In this case, the spectrum is set by
each disk temperature profile, $T(r)\propto(M\dot{M}/r^{3})^{1/4}$, which, for
equal-mass binary components differs between minidisks only through the
accretion rate, and by an extra mass factor $2^{1/4}$ for the circumbinary
disk surrounding the total binary mass.
This allows us to compute spectra of each of the disk components by choosing a
total accretion rate through the CBD, a binary mass ratio, and using the split
in accretion rates onto the binary components measured from the reconstructed
accretion rates of Section 3. Over long enough timescales, the average
accretion rate through the circumbinary disk is split evenly for equal mass
binary components. However, as discussed in Sections 3.2 and 3.3, for prograde
eccentric binaries, this balance can be shifted back and forth between
components over periods of $\mathcal{O}(100)$ binary orbits as the eccentric
disk precesses with respect to the binary argument of pericenter. To quantify
this, we use the accretion-rate ratio explored in Section 3.3 and Figure 5,
$Q\equiv\left<\dot{M}_{2}/\dot{M}_{1}\right>$. We require that the minidisks
are fed by a circumbinary disk with total mass-accretion rate $\dot{M}_{\rm
CBD}$. Then the measured $Q$ and choice of $\dot{M}_{\rm CBD}$ specify the
system:
$\displaystyle\dot{M}_{1}$ $\displaystyle=$ $\displaystyle\dot{M}_{\rm
CBD}\left(1+Q\right)^{-1};\quad\dot{M}_{2}=Q\dot{M}_{1},$ (9)
so that, evaluated at the same radius,
$\displaystyle T_{\rm CBD}=2^{1/4}(1+Q)^{1/4}T_{1};\quad T_{2}=Q^{1/4}T_{1},$
(10)
where we have assumed that the binary components have equal masses in the
second line. The average fluxes from each disk component are,
$\displaystyle F^{0}_{\nu,1}$ $\displaystyle=$
$\displaystyle\frac{2\pi\cos{I}}{d^{2}}\int^{r_{o,1}}_{r_{i,1}}B_{\nu}\left[T_{i,1}\left(\frac{r}{r_{i,1}}\right)^{-3/4}\right]\
rdr$ (11) $\displaystyle F^{0}_{\nu,2}$ $\displaystyle=$
$\displaystyle\frac{2\pi\cos{I}}{d^{2}}\int^{r_{o,2}}_{r_{i,2}}B_{\nu}\left[Q^{1/4}T_{i,1}\left(\frac{r}{r_{i,1}}\right)^{-3/4}\right]\
rdr$ $\displaystyle F^{0}_{\nu,\mathrm{CBD}}$ $\displaystyle=$
$\displaystyle\frac{2\pi\cos{I}}{d^{2}}\int^{r_{o,\mathrm{CBD}}}_{r_{i,\mathrm{CBD}}}B_{\nu}\left[2^{1/4}(1+Q)^{1/4}T_{i,1}\left(\frac{r}{r_{i,1}}\right)^{-3/4}\right]\
rdr,$
for a source at distance $d$ and a common disk-inclination angle $I$. For the
examples here, we assume that the CBD extends from $r_{i,\mathrm{CBD}}=2a$ to
$r_{o,\mathrm{CBD}}=100a$ and that the minidisks extend from the Schwarzschild
inner-most stable circular orbit (ISCO) of the black hole, e.g.,
$r_{i,1}=6GM_{1}/c^{2}$, to the tidal truncation radius, $r_{0}\approx
0.27a=0.27(\Omega_{b})^{-2/3}(GM)^{1/3}$ (Roedig et al., 2014). The quantity
$T_{i,1}$ is the temperature in the primary minidisk at $r_{i,1}$. We
emphasise that in this model, Eqs. (11) set the average flux scale for each
disk component while time variability comes from the binlite accretion-rate
time series of Section 3.2.
#### Example Lightcurves
We compute example lightcurves in the V-band (optical) for different binary
viewing angles and eccentricities in Figures 6 and 7. For these examples we
choose binary parameters $M=2\times 10^{9}{M_{\odot}}$, $P=1$yr, place the
source at a luminosity distance of $1.5$ Gpc ($z\approx 0.29$), and prescribe
a total accretion rate onto the binary of $10\%$ of the Eddington rate, with
$10\%$ accretion efficiency. To set the amplitude of Doppler-boost variability
we choose a spectral index in the observing band of $\alpha=-1$ (see, e.g.,
Charisi et al., 2018) and keep $\cos{I}=1$ fixed for easy comparison
throughout. For the accretion-rate ratio we use Eq. (4). For an eccentricity
of $e=0.4$, this results in a value of $Q=0.169$, and V-band flux ratios of
$\chi_{V,1}=0.308$ and $\chi_{V,2}=0.160$. Hence, in this example, the
minidisks are contributing $\approx 53\%$ of the total V-band flux. Note that
this relative contribution can be a strong function of observing band, binary
masses, and accretion rate, ranging from 0 to 1. We compute the flux in the
V-band, $F_{V}(t)$ by multiplying Eq. (7) by the sum of Eqs. (11) evaluated at
$\nu=5.5\times 10^{14}$ Hz, and then computing an approximate V-band apparent
magnitude $m_{V}=-2.5\log_{10}\left[F_{V}(t)/F_{V,0}\right]$, using the V-band
zero-point flux of $F_{V,0}=3630.22$ Jy.
$\begin{array}[]{cc}\includegraphics[scale={0.35}]{fin_fig/LCmagV_vs_t_e0.40_omega0.00_chi10.307801_chi20.160371_Mbin9.30103_Pbin1yr_fEddCBD0.1_t00.00P}&\includegraphics[scale={0.35}]{fin_fig/LCmagV_vs_t_e0.40_omega90.00_chi10.307801_chi20.160371_Mbin9.30103_Pbin1yr_fEddCBD0.1_t00.00P}\\\
\includegraphics[scale={0.35}]{fin_fig/LCmagV_vs_t_e0.40_omega180.00_chi10.307801_chi20.160371_Mbin9.30103_Pbin1yr_fEddCBD0.1_t00.00P}&\includegraphics[scale={0.35}]{fin_fig/LCmagV_vs_t_e0.40_omega270.00_chi10.307801_chi20.160371_Mbin9.30103_Pbin1yr_fEddCBD0.1_t00.00P}\end{array}$
Figure 6: V-band apparent magnitude lightcurves for prograde systems, (see
Section 4). Each panel is drawn for a different azimuthal viewing angle of the
binary orbit relative to the argument of pericenter, $\varpi$, and each
differently colored line is for a different binary inclination to the line of
sight. The dashed line represents a face-on binary and so exhibits purely
accretion-rate-induced flux variability (compare to Fig. 2). The strongest
Doppler and lensing effects arise for the red line, drawn for a binary
inclined close to the line of sight, $I=85^{\circ}$. Each panel assumes a
binary with $M=2\times 10^{9}{M_{\odot}}$ and $P=1$yr, at a distance of $1.5$
Gpc, and accreting at $10\%$ of the Eddington rate.
$\begin{array}[]{cc}\includegraphics[scale={0.35}]{fin_fig/LCmagV_vs_t_e0.20_omega0.00_chi10.278139_chi20.206308_Mbin9.30103_Pbin1yr_fEddCBD0.1_t00.00P}&\includegraphics[scale={0.35}]{fin_fig/LCmagV_vs_t_e0.40_omega0.00_chi10.307801_chi20.160371_Mbin9.30103_Pbin1yr_fEddCBD0.1_t00.00P}\\\
\includegraphics[scale={0.35}]{fin_fig/LCmagV_vs_t_e0.60_omega0.00_chi10.317951_chi20.141664_Mbin9.30103_Pbin1yr_fEddCBD0.1_t00.00P}&\includegraphics[scale={0.35}]{fin_fig/LCmagV_vs_t_e0.80_omega0.00_chi10.29837_chi20.176344_Mbin9.30103_Pbin1yr_fEddCBD0.1_t00.00P}\end{array}$
Figure 7: The same as Figure 6 except each panel is for a different binary
orbital eccentricity and the same viewing azimuth. Note that, as in Figure 6,
viewing angles $\varpi$ and $I$ affect only the Doppler boost and lensing
signatures while the binary eccentricity affects both accretion variability
and Doppler+lensing magnifications.
Figures 6 and 7 show that a wide range of lightcurve morphologies arise when
allowing viewing angle and eccentricity to vary. It is significant that each
lightcurve is uniquely fixed by binary and observer parameters. Specifically,
the Doppler+lensing-induced and accretion-rate-induced features cannot be
shifted in phase independently of one another. This is because both accretion-
rate and the Doppler+lensing variability have features fixed to specific
values of the binary phase. The peaks and troughs of the Doppler boost
modulation (as well as its shape) and the lensing flares occur at unique
values of the binary phase for a given eccentricity and observing angle, while
accretion-rate variability for eccentric orbits encodes the binary phase via
accretion-rate peaks that occur near pericenter for prograde orbits and
between pericenter and apocenter for retrograde orbits (see Figure 2).
In contrast, for a near-circular-orbit binary, the peak of orbital timescale
variability is related to the passage of a binary component by the near-side
of the lopsided circumbinary cavity (e.g., D’Orazio et al., 2013) and so
depends on the relative orientation of the cavity and not the binary phase
with respect to the observer’s line of sight. Put another way, a lightcurve
exhibiting hydrodynamic and Doppler+lensing variability for a near-circular
orbit binary would only allow identification up to an undetermined orientation
of the circumbinary disk cavity on the sky. For an eccentric binary this free
parameter is eliminated, and the lightcurve model is fully specified by binary
parameters and the observer’s viewing angle.
That the lightcurves in Figures 6 and 7 are unique to the chosen binary
parameters and observer angles offers considerable constraining power compared
to the circular-orbit case. It also makes this signature more difficult to
duplicate via non-SMBHB drivers of variability.
Finally, we note that for equal-mass binaries it is often assumed that the
orbital Doppler effect is nullified or greatly diminished since both black-
hole minidisks are assumed to be emitting at the same luminosity, and via the
two-body problem, will have opposite line-of-sight velocities (e.g., D’Orazio
et al., 2015). However, because the accretion rate can be split unequally
between the components of eccentric binaries (Fig. 5), the Doppler boost can
still cause significant modulations for equal mass binaries when the orbital
eccentricity is non-zero.
## 5 Discussion and Conclusion
We have analysed the variability (Fig. 1) and relative magnitudes (Fig. 5) of
accretion rates measured from viscous hydrodynamical simulations of disks
accreting onto equal-mass, eccentric binaries in both prograde and retrograde
configurations. With the goal of generating lightcurve models to facilitate
searches for accreting binaries, we developed a tool, named binlite, which can
rapidly generate accretion-rate time series at any eccentricity in our
continuous sweep of simulations ($e\leq 0.8$). We then post-processed these
accretion-rate time series to generate simple models for lightcurves at
optical wavelengths, including also observer-dependent effects of orbital
Doppler boosting and gravitational self-lensing.
It is important to note that the details of the accretion rates presented here
will likely vary with different included physics, sink prescriptions, and
numerical methods for solving the equations of hydrodynamics. However, while
these simulations correspond to a simplest non-trivial inclusion of 2D,
viscous, isothermal hydrodynamics, they capture some robust features that lead
to accretion-rate periodicities observed over a wide range of calculations
that include different physics: 3D (e.g., Moody et al., 2019), self-gravity
(Franchini et al., 2024b), magneto-hydrodynamics (MHD, e.g., Shi et al., 2012;
Shi & Krolik, 2015), General Relativity (e.g., Noble et al., 2021), non-
isothermal equations of state (e.g., Westernacher-Schneider et al., 2022; Wang
et al., 2023), for fixed and live binaries (e.g., Franchini et al., 2023), and
are robust over a wide range of numerical techniques (Duffell et al., 2024).
Hence, while exact shapes of accretion-rate times series will depend on the
physical parameters and numerical methods employed, the accretion-rate times
series available through binlite can give insight into the types of accretion
variability expected and aid in building templates with which to search for
such signatures.
Furthermore, the simple lightcurve models presented in Section 4, could be
expanded and adapted to numerous situations. More complex spectra that take
into account different accretion flow properties could be added to this
picture, e.g., emission characteristics of radiatively inefficient accretion
flows, (see the Methods Section of D’Orazio et al., 2015), or those tailored
to proto-planetary disks (e.g., Zhu, 2015). Timescales for the disk spectra to
respond to the changing accretion rate could also be taken into account. For
example, the lightcurve generation procedure presented here could be modified
by smoothing the reconstructed accretion rates in time with a smoothing kernel
set by a buffering timescale due to, e.g., photon diffusion. Beyond this, the
fluid properties of the disk (via post-processing or inclusion of radiative
cooling terms in the energy equation) can be used to generate mock spectra, as
has been done for a much smaller parameter space in a number of works using
viscous hydrodynamics (e.g., Farris et al., 2015; Tang et al., 2018;
Westernacher-Schneider et al., 2022; Krauth et al., 2023; Franchini et al.,
2024a; Cocchiararo et al., 2024), as well as general relativistic MHD (e.g.
d’Ascoli et al., 2018; Combi et al., 2021; Gutiérrez et al., 2022; Avara et
al., 2023).
In addition to advancing lightcurve models with the accretion-rate time series
investigated here, the accretion rates accessible with this tool should also
be updated with the newest, and a wider range of, simulation results.
Utilising both simple and fast simulations, which will expand available data
to a wider range of parameter space (e.g., a wider range of binary and disk
parameters for the types of simulations analysed here (D’Orazio et al., 2016;
Tiede et al., 2020; Dittmann & Ryan, 2022, 2023)), and also simulations
including more physics that can improve accuracy in smaller portions of
parameter space. We plan to add such improvements over time from our own
calculations and also from the wider community.
## 6 Public availability : binlite
We have developed a simple Python package for rapidly generating periodic
accretion rate time series and associated flux series at any eccentricity in
our continuous sweep simulations.
binlite is available in the Python Package Index, and it can be installed with
⬇
python -m pip install binlite
and imported locally as
⬇
import binlite as blt
It contains two main modules
⬇
blt.accretion
blt.flux
for generating variability series of the mass accretion onto the binary and
the flux at a given frequency (under the assumptions detailed in Section 4)
respectively. The source code and more detailed documentation are also
available at github.com/nbia-gwastro/binlite.
D.J.D. received funding from the European Union’s Horizon 2020 research and
innovation programme under Marie Sklodowska-Curie grant agreement No.
101029157. D.J.D. and C.T. acknowledge support from the Danish Independent
Research Fund through Sapere Aude Starting Grant No. 121587. P.D. acknowledges
support from the National Science Foundation under grant AAG-2206299.
## References
* Alves et al. (2019) Alves, F. O., Caselli, P., Girart, J. M., et al. 2019, Science, 366, 90
* Avara et al. (2023) Avara, M. J., Krolik, J. H., Campanelli, M., et al. 2023, arXiv e-prints, arXiv:2305.18538
* Begelman et al. (1980) Begelman, M. C., Blandford, R. D., & Rees, M. J. 1980, Nature, 287, 307
* Charisi et al. (2018) Charisi, M., Haiman, Z., Schiminovich, D., & D’Orazio, D. J. 2018, MNRAS, 476, 4617
* Cocchiararo et al. (2024) Cocchiararo, F., Franchini, A., Lupi, A., & Sesana, A. 2024, arXiv e-prints, arXiv:2402.05175
* Combi et al. (2021) Combi, L., Lopez Armengol, F. G., Campanelli, M., et al. 2021, arXiv e-prints, arXiv:2109.01307
* Czekala et al. (2021) Czekala, I., Ribas, Á., Cuello, N., et al. 2021, ApJ, 912, 6
* d’Ascoli et al. (2018) d’Ascoli, S., Noble, S. C., Bowen, D. B., et al. 2018, ApJ, 865, 140
* Davelaar & Haiman (2022) Davelaar, J., & Haiman, Z. 2022, PRD, 105, 103010
* Dempsey et al. (2020) Dempsey, A. M., Muñoz, D., & Lithwick, Y. 2020, ApJL, 892, L29
* Dittmann & Ryan (2021) Dittmann, A. J., & Ryan, G. 2021, ApJ, 921, 71
* Dittmann & Ryan (2022) —. 2022, MNRAS, 513, 6158
* Dittmann & Ryan (2023) —. 2023, arXiv e-prints, arXiv:2310.07758
* D’Orazio & Charisi (2023) D’Orazio, D. J., & Charisi, M. 2023, arXiv e-prints, arXiv:2310.16896
* D’Orazio & Di Stefano (2018) D’Orazio, D. J., & Di Stefano, R. 2018, MNRAS, 474, 2975
* D’Orazio & Duffell (2021) D’Orazio, D. J., & Duffell, P. C. 2021, ApJL, 914, L21
* D’Orazio et al. (2016) D’Orazio, D. J., Haiman, Z., Duffell, P., MacFadyen, A., & Farris, B. 2016, Monthly Notices of the Royal Astronomical Society, 459, 2379
* D’Orazio et al. (2013) D’Orazio, D. J., Haiman, Z., & MacFadyen, A. 2013, MNRAS, 436, 2997
* D’Orazio et al. (2015) D’Orazio, D. J., Haiman, Z., & Schiminovich, D. 2015, Nature, 525, 351
* D’Orazio & Loeb (2018) D’Orazio, D. J., & Loeb, A. 2018, ApJ, 863, 185
* D’Orazio & Loeb (2019) —. 2019, PRD, 100, 103016
* Duffell (2016) Duffell, P. C. 2016, ArXiv e-prints, arXiv:1605.03577
* Duffell et al. (2020) Duffell, P. C., D’Orazio, D., Derdzinski, A., et al. 2020, ApJ, 901, 25
* Duffell et al. (2024) Duffell, P. C., Dittmann, A. J., D’Orazio, D. J., et al. 2024, arXiv e-prints, arXiv:2402.13039
* Dunhill et al. (2015) Dunhill, A. C., Cuadra, J., & Dougados, C. 2015, MNRAS, 448, 3545
* Farris et al. (2014) Farris, B. D., Duffell, P., MacFadyen, A. I., & Haiman, Z. 2014, ApJ, 783, 134
* Farris et al. (2015) —. 2015, MNRAS, 446, L36
* Farris et al. (2012) Farris, B. D., Gold, R., Paschalidis, V., Etienne, Z. B., & Shapiro, S. L. 2012, Physical Review Letters, 109, 221102
* Franchini et al. (2024a) Franchini, A., Bonetti, M., Lupi, A., & Sesana, A. 2024a, arXiv e-prints, arXiv:2401.10331
* Franchini et al. (2023) Franchini, A., Lupi, A., Sesana, A., & Haiman, Z. 2023, MNRAS, 522, 1569
* Franchini et al. (2024b) Franchini, A., Prato, A., Longarini, C., & Sesana, A. 2024b, arXiv e-prints, arXiv:2402.00938
* Gutiérrez et al. (2022) Gutiérrez, E. M., Combi, L., Noble, S. C., et al. 2022, ApJ, 928, 137
* Haiman et al. (2009) Haiman, Z., Kocsis, B., & Menou, K. 2009, ApJ, 700, 1952
* Haiman et al. (2023) Haiman, Z., Xin, C., Bogdanović, T., et al. 2023, arXiv e-prints, arXiv:2306.14990
* Hayasaki & Mineshige (2008) Hayasaki, K., & Mineshige, S. 2008, ORIGIN OF MATTER AND EVOLUTION OF GALAXIES: The 10th International Symposium on Origin of Matter and Evolution of Galaxies: From the Dawn of Universe to the Formation of Solar System. AIP Conference Proceedings, 1016, 406
* Hu et al. (2020) Hu, B. X., D’Orazio, D. J., Haiman, Z., et al. 2020, MNRAS, 495, 4061
* Ingram et al. (2021) Ingram, A., Motta, S. E., Aigrain, S., & Karastergiou, A. 2021, MNRAS, 503, 1703
* Kelley et al. (2021) Kelley, L. Z., D’Orazio, D. J., & Di Stefano, R. 2021, MNRAS, 508, 2524
* Krauth et al. (2023) Krauth, L. M., Davelaar, J., Haiman, Z., et al. 2023, MNRAS, 526, 5441
* Lai & Muñoz (2022) Lai, D., & Muñoz, D. J. 2022, arXiv e-prints, arXiv:2211.00028
* MacFadyen & Milosavljević (2008) MacFadyen, A. I., & Milosavljević, M. 2008, ApJ, 672, 83
* Major Krauth et al. (2023) Major Krauth, L., Davelaar, J., Haiman, Z., et al. 2023, arXiv e-prints, arXiv:2310.19766
* Miranda et al. (2017) Miranda, R., Muñoz, D. J., & Lai, D. 2017, MNRAS, 466, 1170
* Moody et al. (2019) Moody, M. S. L., Shi, J.-M., & Stone, J. M. 2019, The Astrophysical Journal, 875, 66
* Muñoz & Lai (2016) Muñoz, D. J., & Lai, D. 2016, ApJ, 827, 43
* Nelson (2018) Nelson, R. P. 2018, in Handbook of Exoplanets, ed. H. J. Deeg & J. A. Belmonte, 139
* Noble et al. (2021) Noble, S. C., Krolik, J. H., Campanelli, M., et al. 2021, ApJ, 922, 175
* Roedig et al. (2014) Roedig, C., Krolik, J. H., & Miller, M. C. 2014, ApJ, 785, 115
* Shi & Krolik (2015) Shi, J.-M., & Krolik, J. H. 2015, ApJ, 807, 131
* Shi et al. (2012) Shi, J.-M., Krolik, J. H., Lubow, S. H., & Hawley, J. F. 2012, ApJ, 749, 118
* Siwek et al. (2023a) Siwek, M., Weinberger, R., & Hernquist, L. 2023a, MNRAS, 522, 2707
* Siwek et al. (2023b) Siwek, M., Weinberger, R., Muñoz, D. J., & Hernquist, L. 2023b, MNRAS, 518, 5059
* Stone et al. (2017) Stone, N. C., Metzger, B. D., & Haiman, Z. 2017, MNRAS, 464, 946
* Tang et al. (2018) Tang, Y., Haiman, Z., & MacFadyen, A. 2018, MNRAS, 476, 2249
* Tiede & D’Orazio (2024) Tiede, C., & D’Orazio, D. J. 2024, MNRAS, 527, 6021
* Tiede et al. (2020) Tiede, C., Zrake, J., MacFadyen, A., & Haiman, Z. 2020, ApJ, 900, 43
* Tobin et al. (2016) Tobin, J. J., Kratter, K. M., Persson, M. V., et al. 2016, Nature, 538, 483
* Tofflemire et al. (2017a) Tofflemire, B. M., Mathieu, R. D., Ardila, D. R., et al. 2017a, ApJ, 835, 8
* Tofflemire et al. (2017b) Tofflemire, B. M., Mathieu, R. D., Herczeg, G. J., Akeson, R. L., & Ciardi, D. R. 2017b, ApJL, 842, L12
* Valli et al. (2024) Valli, R., Tiede, C., Vigna-Gómez, A., et al. 2024, arXiv e-prints, arXiv:2401.17355
* Vaughan et al. (2016) Vaughan, S., Uttley, P., Markowitz, A. G., et al. 2016, MNRAS, 461, 3145
* Wang et al. (2023) Wang, H.-Y., Bai, X.-N., Lai, D., & Lin, D. N. C. 2023, MNRAS, 526, 3570
* Ward (1997) Ward, W. R. 1997, Icarus, 126, 261
* Westernacher-Schneider et al. (2022) Westernacher-Schneider, J. R., Zrake, J., MacFadyen, A., & Haiman, Z. 2022, PRD, 106, 103010
* Xin & Haiman (2021) Xin, C., & Haiman, Z. 2021, MNRAS, 506, 2408
* Zhu & Thrane (2020) Zhu, X.-J., & Thrane, E. 2020, ApJ, 900, 117
* Zhu (2015) Zhu, Z. 2015, ApJ, 799, 16
* Zrake et al. (2021) Zrake, J., Tiede, C., MacFadyen, A., & Haiman, Z. 2021, ApJL, 909, L13
* Zurlo et al. (2023) Zurlo, A., Gratton, R., Pérez, S., & Cieza, L. 2023, European Physical Journal Plus, 138, 411
|
# Periods of the Long-Term Variability of the Blazar 0716+714 and
Their Inter-Correlations in a Helical Jet Model
Marina S. Butuzova Crimean Astrophysical Observatory of RAS, 298409, Nauchny,
Russia
###### Abstract
Various quasi-periods for the long-term variability of the radio emission,
optical emission, and structural position angle of the inner part of the
parsec-scale jet in the blazar 0716+714 have been detected. The relationships
between these quasi-periods are interpreted assuming that the variability
arises due to helical structure of the jet, which is preserved from regions
near the jet base to at least 1 milliarcsecond from the core observed in radio
interferometric observations. The radiating jet components should display
radial motions with Lorentz factors of $\approx 3$, and decelerate with
distance from the jet base. The best agreement with the data is given in the
case of non-radial motions of these components with a constant physical speed.
It is also shown that the helical shape of the jet strongly influences
correlations both between fluxes observed in different spectral ranges and
between the flux and position angle of the inner part of the parsec-scale jet.
blazar, S5 0716+714, helical jet
Astronomy Reports, Volume 62, Issue 10, pp.654-663, 2018, DOI:
10.1134/S1063772918100037
## 1 Introduction
Blazars are a class of Active Galactic Nuclei (AGN) whose relativistic parsec-
scale jets are oriented close to the line of sight. Therefore, the flux
density emitted by the jet is enhanced by relativistic effects in the
observer’s frame, so that it dominates the emission of other parts of the AGN.
This may be able to explain the fact that no lines are observed in the
spectrum of the blazar 0716+714. Indirect estimates of its redshift have
yielded $z\approx 0.31$ (Bychkova et al., 2006; Nilsson et al., 2008),
$0.2315<z<0.3407$ (Danforth et al., 2013), and $z\geq 0.52$ (Sbarufatti et
al., 2005). This object is variable over the entire electromagnetic spectrum,
on both short and long time scales (see, e.g., Wagner et al., 1996; Raiteri et
al., 2003; Wu et al., 2007; Poon et al., 2009; Gorshkov et al., 2011; Volvach
et al., 2012; Rani et al., 2013; Liao et al., 2014; Bychkova et al., 2015).
0716+714 has also been observed with Very Long Baseline Interferometry (VLBI)
in the framework of a 2 cm survey using the Very Long Baseline Array (VLBA),
and monitored as part of the MOJAVE project (see, e.g., Bach et al., 2005;
Lister et al., 2013; Pushkarev et al., 2017). The bright compact VLBI core
visible in these maps is the region of the jet where the medium becomes
optical thin to radiation at the given wavelength. For more than 150 sources
(including 0716+714), the position of the VLBI core has been found to shift
closer to the jet base with increasing frequency (Pushkarev et al., 2012),
interpreted as an effect of synchrotron self-absorption in the jet (Marcaide &
Shapiro, 1984; Lobanov, 1998; Kovalev et al., 2008). That is, the magnetic-
field strength and density of the radiating particles decrease with distance
from the jet base, so that the medium becomes optically thin to radiation at
increasingly lower frequencies. Synchrotron self-absorption could also be
responsible for the observed delays between flares observed at different radio
frequencies on single dishes (see, e.g., Kudryavtseva et al., 2011; Agarwal et
al., 2017). Without interpretation by synchrotron selfabsorption a number of
authors have found time delays between variability at different radio
frequencies and in different spectral ranges.
For example, Raiteri et al. (2003) found that the delay between the
variability of 0716+714 at 22$-$23 GHz and 15 GHz is 6$-$9 days, between
22$-$23 and 8 GHz is 22$\pm$2${}^{\text{d}}$, and between 15 and 5 GHz is
53$\pm$2${}^{\text{d}}$. No reliable correlation between the optical and 15
GHz fluxes has been found. However, Rani et al. (2013) found a correlation
between the V and 230 GHz fluxes with a time delay of $\approx 65^{\text{d}}$.
The delay between the optical and gamma-ray ranges is $\approx 1.4^{\text{d}}$
(Larionov et al., 2013). The millimeter wave length variability lags the
gamma-ray variability by $82\pm 32^{\text{d}}$ (Rani et al., 2014). Wu et al.
(2007) and Poon et al. (2009) attempted to determine the delays between the
variability of 0716+714 in different optical bands. If there are such delays,
they are less than the time resolution of those observations.
Thus, the flux variability at low frequencies is delayed compared to the
variability at higher frequencies. Analyses of long-term series of
observations reveal different quasi-periods111Here and below, quasi-periods
refer to variability periods detected in certain time intervals at a specified
confidence level using specialized methods such as discrete correlation
functions (Edelson & Krolik, 1988), structure functions (Simonetti et al.,
1985), and others. for the long-term optical ($\approx 3.3$ yrs (Raiteri et
al., 2003)) and radio (5.5$-$6 yrs (Raiteri et al., 2003; Bychkova et al.,
2015; Liu et al., 2012) variability, and also for variations in the position
angle of the inner jet PA${}_{\text{in}}$ ($\approx 11$ yrs Lister et al.
(2013)).
The variations of PA${}_{\text{in}}$ can be explained in a natural way if the
jet has a helical shape (Bach et al., 2005; Lister et al., 2013), which also
often provides the simplest interpretation of a variety of observed properties
of AGNs. For example, in the case of 0716+714, variations in the spectral
energy distribution (Ostorero et al., 2001) and the kinematics of features in
the parsec-scale jet (Butuzova, 2018) have both been explained in this way.
The helical jet shape also gives rise to periodic variations in the viewing
angle of the radiating regions, leading to corresponding variations of the
Doppler factor, which should be manifest as longterm periodicity of the flux
variations. The differences in the long-term periods for the radio
variability, optical variability, and PA${}_{\text{in}}$ variations can be
explained by either an overall deceleration of the speed of the jet or non-
radial motions of the radiating regions in the jet. The latter is most
probable for 0716+714 (see Section 2). Section 3 presents interpretations of
the following results. The first is the opposite results obtained in searches
for correlations between the radio and optical flux variations using the same
statistical method (Raiteri et al., 2003; Rani et al., 2013) The second is the
fact that time intervals when there is a strong positive correlation between
PA${}_{\text{in}}$ and the gamma-ray flux alternate with intervals in which
there is a strong negative correlation between these quantities (Rani et al.,
2014). A discussion of the obtained results and our conclusions are presented
in Section 4.
## 2 Relationship between the long-term variability periods for various
observed quantities
Analysis of many-year radio and optical light curves of the blazar 0716+714
indicate the presence of quasi-periods in the long-term variability. Optical
data for 1994$-$2001 indicated a period of about 3.3 yrs (Raiteri et al.,
2003). This period was never confirmed, possibly because other studies
considered data obtained over shorter time intervals. For example, Rani et al.
(2013) did not detect any reliable periods for the long-term variability in
their analysis of data for 2007$-$2010\. It is difficult to detect long-term
periodicity in the optical due to the superposition of short-term flares on
the longer-term trend. In the radio, where the flare component is weaker,
variability periods of 5.6$-$6 yrs at 14.5 and 15 GHz for 1978$-$2001 (Raiteri
et al., 2003), 5.5$-$6 yrs at 22 GHz for 1992.7$-$2001.2 (Bach et al., 2005),
and $5.8\pm 0.4$ yrs at 15 GHz for data after 2001 (Liu et al., 2012) have
been found.
Data from more than 30 years of observations carried out on telescopes of the
Crimean Astrophysical Observatory (CrAO), the Metsahovi Radio Observatory, and
the University of Michigan Radio Astronomy Observatory at frequencies from 4.8
to 36.8 GHz also indicate the presence of a period of $\approx 8$ yrs,
together with shorter periods (Bychkova et al., 2015). Quasi-periodicity is
also present in the variations of the position angle of the inner parsec-scale
jet of 0716+714. Analysis of data for 26 epochs of observations (from 1992.7)
at 2.9, 8.4, 15.3, and 22.2 GHz led to the detection of variations in the
position angle of the jet lying within 1 mas from the VLBI core with a period
of $7.4\pm 1.5$ yrs and an amplitude of $3.5^{\circ}$ (Bach et al., 2005).
Observations at 15 GHz obtained from 1994.5$-$2011.5 displayed a period for
the PA${}_{\text{in}}$ variations of 10.9 yrs and an amplitude of $11^{\circ}$
(in this case, for the mean position angle of all jet features at distances
$0.15-1$ mas from the core, weighted according to their flux densities)
(Lister et al., 2013).
Therefore, the lack of agreement between the periods for the radio and optical
variability seems to suggest that the brightness variations of the blazar
0716+714 in these spectral ranges are associated with physical processes
occurring in its jet. On the other hand, the periodic variations of the
position angle of the inner jet testify to periodic variations in the jet
direction, leading to variations in the spectral flux density, since
$F_{\nu}\propto\nu^{-\alpha}\delta^{s+\alpha},$ (1)
where $\nu$ is the observing frequency, $\alpha$ the spectral index, $s=3$ if
the depth of the radiating region can be neglected ($s=2$ otherwise), and the
Doppler factor is
$\delta=\delta\left(\theta,\beta\right)=\sqrt{1-\beta^{2}}\left(1-\beta\cos\theta\right)^{-1}.$
(2)
Here, $\theta$ is the angle between the velocity vector for a jet component
and the line of sight at the given time, and $\beta$ is the physical speed of
the radiating feature in units of the speed of light $c$. If $\beta$ is
constant, periodic variations of $\theta$ will give rise to long-term
variability of the radio and optical flux densities and of PA${}_{\text{in}}$
with the same period. However, the periods for these three types of variations
are different. We will elucidate the origins of this contradiction under the
hypotehsis that the jet is helical in shape. We will explore this using the
schematic representation of a jet comprised of individual components forming a
helical line on the surface of an notional cone (Butuzova, 2018). This
corresponds to results of recent studies based on stacked VLBI images for
individual sources (Pushkarev et al., 2017), which indicate that, for many
sources, including 0716+714, the jet features on scales from hundreds to
thousands of parsecs are located inside a cone. We take jet components to be
individual radiating regions of the jet that become observable when they reach
distances from the VLBI core of $\lesssim 0.1$ mas. For our subsequent
arguments, it is not important whether these components are regions of
enhanced particle density or shocks where electrons are accelerated and
subsequently injected into the surrounding space. The position of a component
on the surface of this cone can be described by an azimuth angle $\varphi$
measured along a circular arc formed by a planar cross section of the cone
perpendicular to its axis. (A detailed schematic of the jet and its
geometrical parameters are described by Butuzova (2018)). The coordinate
origin for $\varphi$ was taken to be the point located in the plane of the
line of sight and the cone axis, on the far side of the cone relative to the
observer.
Taking into account synchrotron self-absorption, we assumed that the medium
becomes transparent to the optical radiation of a jet component when it
reaches circle 1 (Fig. 1) formed by the cross section of the cone by a plane
orthogonal to the cone axis at a distance $r_{1}$ from its apex. Continuing
from the active nucleus, the jet component reaches the analogously formed
circle 2 at the distance $r_{2}$. In this region, the medium becomes
transparent to the radio emission of the jet formed in its VLBI core (at the
given frequency). Moving farther, the component reaches circle 3 at a distance
$r_{3}$ from the cone apex, where it is manifest on VLBI maps as the closest
component to the core. We took this to be the distance at which
PA${}_{\text{in}}$ is measured. For each of these circles, we introduced a
notional point moving such that it coincides with the position of the jet
component intersecting the corresponding circle at a given time. As follows
from formulas (1), (2), the periods of the optical and radio variability will
then be equal to the period of rotation of notional points around circles 1
and 2, respectively. Figure 1 shows that the period for variations of
PA${}_{\text{in}}$ will similarly be equal to the period of rotation of a
notional point around circle 3 (see also Butuzova, 2018, Fig. 2). In order for
the helical structure of the jet to be preserved over a long time, we assumed
that the speed of the components was constant, or at least that this speed
varies with distance from the active nucleus in the same way for all
components.
Figure 1: Schematic of the arrangement of regions in the jet in which the
observed optical and radio emission arise and the region in which
PA${}_{\text{in}}$ becomes measureable (the regions where a jet component
crosses the circles 1, 2, and 3, respectively). The helical jet (without
details of its components) is presented at a specified time by the thick bold
curve. The thin curve shows the position of part of the jet in the vicinity of
circle 3 at some subsequent time. We can see that the point where the jet
crosses circle 3 shifts with time (similarly for circles 1 and 2). Over the
variability periods for the optical emission, radio emission or
PA${}_{\text{in}}$, these points undergo a full revolution about circles 1, 2,
and 3, respectively.
### 2.1 Radial Motion of the Jet Components
We will first consider the case when the jet components move outward along the
generating cone (so-called radial, or ballistic motion, see Fig. 2). Without
loss of generality, we can assume that the difference in the azimuth angles of
each of two successive components is some value $\varphi_{d}$ in radians. We
denote $\Delta t^{\prime}_{1}$ to be the time interval in the comoving frame
between the times when any two successive components cross circle 1. Over this
time, a notional point moving along circle 1 describes an arc with length
$\varphi_{d}$. For simplicity, we assumed that $2\pi$ is a multiple of
$\varphi_{d}$. In this case, some number $n$ of components cross circle 1 over
the rotation period of the notional point. Since the interval between two
events in the observer’s frame is smaller than the interval in the source rest
frame by a factor $\delta$,
$\Delta t^{\prime}=\delta\Delta t,$ (3)
the variability period for the optical emission in the observer’s frame will
be
$P_{1}=\sum_{j=1}^{n}\frac{\Delta
t^{\prime}_{1}}{\delta\left(\theta\left(\varphi\right)\right)}\approx\frac{n\Delta
t^{\prime}_{1}}{\delta\left(\theta_{0},\beta\right)}.$ (4)
The right-hand side of (4) was obtained as follows. The angle $\theta$ of the
components crossing circle 1 over the period varies from $\theta_{0}-\xi$ to
$\theta_{0}+\xi$, where $\theta_{0}$ is the angle between the cone axis and
the line of sight. According to formula (2), the Doppler factor varies
cyclically in some interval. Deviations of the Doppler factor from its mean
value can be neglected, since their magnitudes are not large, due to the
smallness of the angle $\xi$, and their sum over the period is zero. For our
further estimates, we took the mean Doppler factor to be
$\delta\left(\theta_{0},\beta\right)$.
Figure 2: Schematic for interpreting the variability periods in the case of
ballistic motion of the jet features. Part of the helical jet in the vicinity
of the circle 1 is shown by the bold curve. The filled squares represent
several jet components. The trajectories of each of the components lie along
the generator of the cone (dashed curves). The hollow squares show the
positions where these components will intersect the circles 1 and 2. The arrow
indicates the direction of motion of a notional point along the circle 2
(similarly for the circle 1).
Moving farther, the component crosses circle 2. The variability period for the
radio emission $P_{2}$ can be written similarly to (4), but with the subscript
“1” replaced with “2”. Due to the character of the motion (Fig. 2), the
azimuthal angle of each jet component does not vary with time. The number of
components crossing circles 1 and 2 during the variation period $\varphi$ is
also constant. The time intervals between the moments of intersection by two
successive components of circles 1 and 2 are equal $\left(\Delta
t^{\prime}_{1}=\Delta t^{\prime}_{2}\right)$. In the absence of deceleration
of the components, we should have $P_{2}=P_{1}$. Since this is not observed,
we supposed that the speed of the components crossing circle 2 was
$\beta_{2}=a\beta_{1}$. Here, $0<a<1$ and $a>\beta_{2}$ (otherwise,
$\beta_{1}>1$). It follows from (4) that the ratio of the variability periods
in the optical and radio will be
$\frac{P_{1}}{P_{2}}=\frac{\delta\left(\theta_{0},\beta_{2}\right)}{\delta\left(\theta_{0},\beta_{1}\right)}.$
(5)
We used (2) and (5) to write the equation
$\frac{1-\beta_{1}\cos\theta_{0}}{1-a\beta_{1}\cos\theta_{0}}\sqrt{\frac{1-a^{2}\beta_{1}^{2}}{1-\beta_{1}^{2}}}=\frac{P_{1}}{P_{2}},$
(6)
which was solved numerically for $a$ for values $\beta_{1}=0.3-0.9999$
(corresponding to Lorentz factors $\Gamma=1-70$) with $P_{1}=3.3$ yrs and
$P_{2}=5.8$ yrs.
We found that one root is always greater than one. The other root is negative
when $\beta_{1}<0.52$ and satisfies our conditions when $\beta_{1}\geqslant
0.52$. When $a\approx 0.962$ and $\beta_{1}=0.9948$, the maximum value
$\beta_{2}\approx 0.957$ is reached, corresponding to $\Gamma=3.5$. Solving
(6) for $P_{2}$ and the variation period for the position angle of the inner
jet for this interval of $\beta_{1}$ values yields the maximum value
$\beta_{2}=0.949$ (for $a\approx 0.954$ and $\beta_{1}=0.995$), which
corresponds to $\Gamma=3.1$. Thus, agreement of the variability periods
observed at difference distances from the jet base can be achieved when
$\Gamma\sim 3-4$. This does not agree with the values $\Gamma\sim 10-20$
inferred from observations of superluminal motions of jet components in
0716+714 (see, e.g., Bach et al., 2005; Nesci et al., 2005; Pushkarev et al.,
2009).
### 2.2 Non-Radial Motion of the Jet Components
Let us now consider a helical jet whose components move non-ballistically,
i.e., at some angle to a radial trajectory. We denote $p$ to be the pitch
angle (angle between the generating cone and the velocity vector of a jet
component), and $\psi$ to be the angle between the tangent to the helix of the
jet and the generating cone at a given point (Fig. 3). If $p=\psi$, the jet
will appear stationary in space, and will always cross circles 1, 2, and 3 at
the same points. In this case, there should be no periodic variability, since
the angle $\theta$, and consequently the Doppler factor, do not change in the
regions responsible for the observed quantities. If $p\neq\psi$, we will
observe the jet helix rotating about its axis. Due to the conical geometry of
the jet, the variations of the azimuth angle $\varphi$ decrease with
increasing $r$. However, we are interested in variations of $\varphi$ at the
constant distances $r_{i}$ from the cone apex to the circles $i$ corresponding
to the regions making the main contributions to the optical ($i=1$) and radio
($i=2$) emission, and to the region where PA${}_{\text{in}}$ can be measured
($i=3$) (Fig. 1).
Figure 3: Schematic of a helical jet with non-ballistic motion of its
components (shown by the filled squares). Part of the helical path of the jet
located near circle 1 is shown (bold curve). The thin curves show the position
of this part of the jet and the components at some later time, with $\psi\neq
p$. The triangles on circle 1 show the places where the jet components cross
circle 1. The arrows show the direction of motion of a notional point along
circle 1.
Variations of the azimuthal angle of the part of the jet reaching a given
distance $r_{i}$ from the cone apex can be found from the schematic presented
in Fig. 4, under the condition that $\beta c\,dtr_{i}$:
$d\varphi\approx\frac{\beta c\,dt\,\sin(\psi-p)}{r_{i}\sin\xi\cos\psi},$ (7)
where $\xi$ is the opening angle of the cone ($\xi=1^{\circ}$, Butuzova,
2018). Since the angular frequency of a notional point moving along circle $i$
is $\omega_{i}=d\varphi/dt$, we find that the ratio of the periods of two
observable quantities $i$ and $k$ is equal to the ratio of the distances from
the cone apex to the region of the jet where these quantities are measured:
$\frac{P_{i}}{P_{k}}=\frac{r_{i}}{r_{k}}.$ (8)
Figure 4: Schematic of part of the jet (bold curves) at a distance $r_{i}$
from the cone apex at the initial time (1) and after a time dt (2).
Substituting various pairs of the known variability periods for 0716+714 into
(8) yields the three independent relations
$\begin{split}r_{2}=&1.76\,r_{1},\\\ r_{3}=&3.30\,r_{1},\\\
r_{3}=&1.88\,r_{2}.\end{split}$ (9)
It follows from the last two equations of (9) that $r_{2}/r_{1}\approx 1.76$,
which is equal to the directly inferred ratio $r_{2}/r_{1}$ in the first
equation of (9). Thus, the observed periods for the long-term variability in
the ratio, optical and inner-jet position angle PA${}_{\text{in}}$ show good
consistency. This supports a picture with non-radial motion of the jet
features with $p\neq\psi$ and an absence of deceleration at the distances from
the active nucleus considered here.
Let us suppose that PA${}_{\text{in}}$ is measured at a specified distance
from the VLBI core at 15 GHz, equal to 0.15 mas. Then,
$r_{3}-r_{2}=0.15$(mas)$/\sin\theta_{0}$. Using the third equation of (9) and
$\theta_{0}=5.3^{\circ}$ (Butuzova, 2018), we obtain the distance $r_{2}=1.84$
mas. In a $\Lambda$CDM model with $H_{0}=71$ km s-1Mpc-1, $\Omega_{m}=0.27$,
and $\Omega_{\Lambda}=0.73$ (Komatsu et al., 2009) and adopting a redshift of
$z=0.3$ for 0716+714, the physical distance from the cone apex to the position
of the VLBI core is 8.1 pc. This is consistent with the distance of the VLBI
core from the black hole of 6.68 pc at 15.4 GHz determined by Pushkarev et al.
(2012) using these same cosmological parameters. This provides additional
support for our picture of the jet. Continuing our reasoning using (9), we
find that the distance between the jet apex and the region where the optical
emission becomes observable is 4.6 pc. Due to the small delay in the
variability (at the limit of the time resolution of high frequency data of
Rani et al., 2013; Larionov et al., 2013), we infer that the gamma-ray and
optical emission is formed in the same region, or at least in closely spaced
regions. Rani et al. (2014) estimated that the gamma-ray emission arises from
a region located $3.8\pm 1.9$ pc closer to the black hole relative to the VLBI
core (observed at 43 and 86 GHz), also consistent with our results.
## 3 Correlation between flux and inner-jet position angle
Assuming that the long-term variability of the blazar 0716+714 is due to
periodic variations in the direction of motion of the jet components, we
expect there should be a relationship between the measured spectral flux
density $F_{\nu}$ and the inner-jet position angle. Such relationships between
the gamma-ray flux $F_{\gamma}$ and both the flux from the VlBI core at 43 and
86 GHz and PA${}_{\text{in}}$ have been investigated for 0716+714 by Rani et
al. (2014). They found that time intervals with a strong positive correlation
between $F_{\gamma}$ and PA${}_{\text{in}}$ alternate with intervals in which
there is a strong negative correlation between these two quantities.
This result was explained by Rani et al. (2014) by the fact that, in a curved
(possibly helical) jet, the regions responsible for the observed quantities
are located at different distances from the active nucleus, and therefore have
different viewing angles $\theta$. If the $\theta$ values for two regions are
roughly the same, there will be a strong positive correlation between the
corresponding observed quantities. If the $\theta$ values are different, there
may be a strong negative correlation. However, for the helical jet we are
considering here, there is no direct relationship between $F_{\nu}$ and
PA${}_{\text{in}}$, which can be explained as follows. According to Butuzova
(2018) (Eq. (1)), the position angle is
$\text{PA}_{\text{in}}=\text{PA}_{0}+\Delta\text{PA}=\text{PA}_{0}+\frac{\sin\xi\sin\varphi}{\cos\xi\sin\theta_{0}+\sin\xi\cos\theta_{0}\cos\varphi}$
(10)
(PA0 is the mean value of PA${}_{\text{in}}$) and the angle between the
component velocity and the line of sight depend on the geometrical parameters
of the cone and the position of the component relative to the cone axis and
the line of sight (i.e., on the angle $\varphi$). We can use these equations,
where the periodicity appears only due to variations in the azimuthal angle,
to model the observed correlation between the flux and PA${}_{\text{in}}$.
Let us first consider the case of radialmotion of the components. We find from
(1) with the substitution of (2), in which (see Butuzova, 2018, Eq. (5))
$\theta=\theta_{b}=\arccos\left(\cos\xi\cos\theta_{0}-\sin\xi\sin\theta_{0}\cos\varphi\right),$
that the extrema of the function $F_{\nu}$ occur at values $\varphi=n\pi$
(maxima for $n=1$, 3, 5, etc. and minima for $n=0$, 2, 4, etc.). That is, the
qualitative variations of $F_{\nu}$ do not depend on the choice of $\nu$,
$\alpha$, $s$, and $\beta$. The upper panel of Fig. 5 presents the variations
of $F_{\nu}$ calculated using (1) and (10) (for $\alpha=0.5$, $s=2$, and
$\beta=0.995$, which corresponds to $\Gamma=10$) and deviations of the inner-
jet position angle from its mean value $\Delta$PA (for $\theta_{0}/\xi=5.3$)
as functions of the azimuthal angle $\varphi$. The flux is normalized so as to
enable a visual comparison of its variations with the variations of
$\Delta$PA. The resulting curves were divided into several sections in
$\varphi$, such that qualitative variations in the behavior of both quantities
did not arise within each section.
Figure 5: Deviations of the position angle of the inner jet from its mean
value $\Delta$PA (solid curve) and the observed flux $F_{\nu}$ (dashed curve)
over the variation period for $\varphi$ (upper), and the correlation
coefficient $r_{p}$ between these quantities (lower) for ballistic motion of
features of a helical jet.
The corresponding formulae were used to compose datasets of $\Delta$PA and
$F_{\nu}$ values for each section, for variations of $\varphi$ in steps of
$0.5^{\circ}$, and the Pearson correlation coefficient ($r_{p}$) between these
datasets was calculated. The resulting $r_{p}$ value will have its maximum
possible value, since, in contrast to the observational data, there is no
measurement error. An alternation of intervals of strong positive and negative
correlations can be seen (Fig. 5, lower panel). Intermediate values of the
correlation coefficient are present only in short intervals. This theoretical
result is in qualitative agreement with the observations of the behavior of
the gamma-ray flux and the inner-jet position angle considered by Rani et al.
(2014). Thus, we found that, for the same flux value (corresponding to the
same viewing angle $\theta_{b}$), either a positive or negative correlation
with PA${}_{\text{in}}$ could be present in the observations.
In the case of non-radial component motions, the variations of the viewing
angle are given by Eqs. (11)-(13) from (Butuzova, 2018), which we do not
present here due to their unwieldiness. The extrema of the function $F_{\nu}$
occur at the values
$\varphi=-\arcsin\left(\frac{\sin
p}{\sqrt{\cos^{2}p\sin^{2}\xi+\sin^{2}p}}\right)+\pi n,$ (11)
where even $n$ correspond to maxima and odd $n$ to minima of the function
$F_{\nu}$. According to (11), for the pitch angle found by Butuzova (2018) of
$p=5.5^{\circ}$, the flux reaches a maximum when $\varphi\approx 100^{\circ}$.
As $p$ is decreased, the peak $F_{\nu}$ shifts toward 180∘, and the maximum
$F_{\nu}$ occurs for $\varphi\approx 90^{\circ}$ when $p>20^{\circ}$. The
upper panel of Fig. 6 plots the functions $F_{\nu}$ and $\Delta$PA for the
same parameters as in the previous case. The correlation coefficient $r_{p}$
was constructed using an analogous procedure (Fig. 6, middle panel). This
figure shows that a strong positive correlation between the inner jet position
angle and the observed flux should always be present. However, since these
quantities arise in regions located at different distances from the jet base
in our model (see Section 2), we can conclude with confidence that the
azimuthal angles of the components for which $F_{\nu}$ and $\Delta$PA are
observed at a given time differ by some amount $\Delta\varphi$.
Figure 6: Upper: variations of the inner-jet position angle relative to the
mean value $\Delta$PA (dashed curve) and of the flux $F_{\nu}$ for the case of
zero difference in the azimuthal angles of the regions responsible for the
observed quantities (curve 1) and for $\Delta\varphi\approx 79^{\circ}$ (curve
2). The central and lower panels show the correlation coefficients $r_{p}$ for
the former and latter cases, respectively.
Agreement with the results of Rani et al. (2014) requires that this difference
be $\Delta\varphi\approx 79^{\circ}$ (curve 2 for $F_{\nu}$ in the upper panel
and curve for $r_{p}$ in the lower panel in Fig. 6). On the other hand, strong
positive or negative correlations between $F_{\nu}$ and $\Delta$PA are also
possible for radial component motions if the difference in the azimuthal
angles is $\Delta\varphi\approx\pi/2$ or $\approx\pi$ (see Fig. 5).
Consequently, the correlation between the observed quantities may be
insignificant when analyzing data over long time intervals, as was found by
Raiteri et al. (2003), for example, in their analysis of the radio and optical
fluxes during 1994$-$2001\. In contrast, the data for the shorter interval
2007$-$2010 analyzed by Rani et al. (2013) revealed a correlation between the
indicated quantities at a significance level of more than 99$\%$. Further, the
difference in the azimuthal angles of the regions responsible for the observed
quantities can appreciably affect both the correlation coefficient between
PA${}_{\text{in}}$in and $F_{\nu}$ and the duration of the time interval when
a given correlation coefficient is observed. Finally, the character of the
motions of individual components of helical jet cannot be determined by
analyzing the correlation between PA${}_{\text{in}}$ and $F_{\nu}$. It is
important to note that, when investigating correlations between fluxes
observed in different spectral ranges, distinct correlation coefficients in
the different time intervals will also be present.
## 4 Discussion and conclusion
The hypothesis that AGN jets may be helical has been widely applied for
several decades to interpret various observed properties such as their
microvariability (Camenzind & Krockenberger, 1992), the shape and variations
of the spectral energy distributions for the blazars Mrk 501 (Villata &
Raiteri, 1999) and 0716+714 Ostorero et al. (2001), the long-term brightness
variability of OJ 287 (Sillanpaa et al., 1988) and 0716+714 (Nesci et al.,
2005), variations in the speeds and non-radial component motions (Rastorgueva
et al., 2009), and quasi-periodicity of variations of the inner-jet position
angle for 0716+714 (Bach et al., 2005; Lister et al., 2013). A helical jet
shape could form due to precession of the jet nozzle or the development of
(magneto)hydrodynamical instabilities. Kelvin–Helmholtz instability (Hardee,
1982) has been widely studied as a means of estimating the physical parameters
of jets and the ambient medium (e.g., for 3C 120 (Hardee, 2003) and 0836+710
(Perucho et al., 2012)). Alternatively, wavelike perturbations at the
boundaries of the observed isophotes of jets could be related to the
development of a magnetohydrodynamical analog of wind instability (Gestrin &
Kontorovich, 1986).
In this study, we have supposed that the radiating jet components form a
helical curve, without considering their physical nature. For example, the jet
components could be individual radiating parts of the jet (plasmoids or
regions of shocks) or volume elements (in the case of a spatially continuous
radiating jet). A helical shape suggests the presence of periodic variations
of the angle between the jet velocity and the line of sight at some constant
distance from the core, which should be manifest as long-term quasi-periodic
variability of the radiation flux over the entire observed range of the
electromagnetic spectrum of the blazar 0716+714. The differences in the quasi-
periods for the variations of PA${}_{\text{in}}$ (Bach et al., 2005; Lister et
al., 2013) and of the radio- (Raiteri et al., 2003; Bychkova et al., 2015;
Bach et al., 2005; Liu et al., 2012) and optical (Raiteri et al., 2003) fluxes
can most simply be explained in the jet geometry considered if the radiation
in different spectral ranges is emitted at different distances from the jet
apex. This spatial separation of regions radiating at different frequencies
can arise due to synchrotron self-absorption in the jet or the energy losses
of the radiating electrons. In both cases, the higher the frequency of the
observed emission, the closer to the jet base the region in which it is
generated. The delays in the flux variability observed at different
frequencies also testify to the action of this effect (see, e.g., Raiteri et
al., 2003; Larionov et al., 2013; Rani et al., 2014).
In this study, we have brought the variability periods in different spectral
ranges into agreement in the case of radial and non-radial motions of the
radiating components. The former case requires a low Lorentz factor for the
components (no more than 4) with overall deceleration, at least from the
region where the optical emission is formed to 0.15-0.5 mas from the core,
where the position angle of the inner jet is measured. This does not agree
with observations of features in the parsec-scale jet of 0716+714 (Bach et
al., 2005; Nesci et al., 2005; Pushkarev et al., 2009). Moreover, it was shown
by Butuzova (2018) that differences between the relative speeds of components
in the inner and outer jet observed in different years can be explained in the
framework of a helical-jet model with non-ballistic component motions, such as
are observed in the jet (Bach et al., 2005; Rastorgueva et al., 2009). It was
shown that period ratio is equal to ratio of the physical distances from the
jet apex of the regions responsible for the measured quantities. This enables
us to introduce another absolute distance scale, which will subsequently
facilitate deeper studies of the jet properties. In our picture of the jet and
the appearance of long-term variability due to geometric effects, the radio
periods found by Bychkova et al. (2015) cannot carry information about the
properties of the central engine without taking into account the non-radial
motions of the jet components.
The helical shape of a jet with spatially separated regions responsible for
the observed emission at different frequencies and region where the inner jet
position angle is measured complicates studies of correlations between the
observed quantities. It has been shown that there cannot be a constant
correlation coefficient between quantities formed in regions at different
fixed distances from the jet apex. A strong positive correlation observed in
one time interval will be replaced with a negative correlation in another.
This is due to both the different azimuthal angles of these regions and the
fact that $\varphi$ varies irregularly in the observer’s rest frame, due to
variations of $\delta$, especially for non-radial component motions (Butuzova,
2018). This agrees with certain observational facts.
For example, Rani et al. (2013) noted an alternation of intervals when
positive and negative correlations between PA${}_{\text{in}}$ and the gamma-
ray flux of the blazar 0716+714 were observed. Rani et al. (2013) also
indicate that a strong correlation between the gamma-ray and optical fluxes
was observed over roughly 500 days (Pearson correlation coefficient
$r_{p}=0.66$), while $r_{p}=0.36$ over the following $\approx$400 days. A
correlation was also found between the radio flux and PA${}_{\text{in}}$
during 1994$-$2014 ($r_{p}=0.44$) (Liu et al., 2012), while no correlation
between these quantities was found for data obtained from August 2008 through
September 2013 (Rani et al., 2014).
We can introduce some clarity into this picture only if we know the
geometrical parameters of the helical jet, the character of the motion of the
jet components,and the arrangement of the studied regions responsible for
various observed quantities relative to the plane containing the axis of the
helical jet and the line of sight, and not only relative to the line of sight,
as was supposed in the simplest case (Rani et al., 2014). We also showed in
Section 3 that it is not possible to determine the character of the component
motions in a helical jet based on the correlation between the radio flux and
the inner-jet position angle, as was done by Liu et al. (2012).
Our hypothesis that the jet of the blazar 0716+714 has the form of a helical
curve located on the surface of a cone may seem somewhat idealized. However,
this is consistent with the results of many years of VLBI observations (Lister
et al., 2013; Pushkarev et al., 2017). In addition, a helical jet with non-
radial component motions makes it possible to find agreement between estimates
of the component velocities in the VLBI jet and the viewing angle obtained in
different studies, which often differ appreciably (Butuzova, 2018). As we have
shown, such a jet can provide a simple explanation for the differences in the
observed long-term quasi-periods for different quantities, as well as the
differences in the correlations between the observed quantities in different
time intervals.
The author thanks A.B. Pushkarev for useful comments.
## References
* Agarwal et al. (2017) Agarwal, A., Mohan, P., Gupta, A. C., et al. 2017, MNRAS, 469, 813, doi: 10.1093/mnras/stx847
* Bach et al. (2005) Bach, U., Krichbaum, T. P., Ros, E., et al. 2005, A&A, 433, 815, doi: 10.1051/0004-6361:20040388
* Butuzova (2018) Butuzova, M. S. 2018, Astronomy Reports, 62, 116, doi: 10.1134/S1063772918020038
* Bychkova et al. (2006) Bychkova, V. S., Kardashev, N. S., Boldycheva, A. V., Gnedin, Y. N., & Maslennikov, K. L. 2006, Astronomy Reports, 50, 802, doi: 10.1134/S1063772906100040
* Bychkova et al. (2015) Bychkova, V. S., Vol’vach, A. E., Kardashev, N. S., et al. 2015, Astronomy Reports, 59, 851, doi: 10.1134/S1063772915080016
* Camenzind & Krockenberger (1992) Camenzind, M., & Krockenberger, M. 1992, A&A, 255, 59
* Danforth et al. (2013) Danforth, C. W., Nalewajko, K., France, K., & Keeney, B. A. 2013, ApJ, 764, 57, doi: 10.1088/0004-637X/764/1/57
* Edelson & Krolik (1988) Edelson, R. A., & Krolik, J. H. 1988, ApJ, 333, 646, doi: 10.1086/166773
* Gestrin & Kontorovich (1986) Gestrin, S. G., & Kontorovich, V. M. 1986, Soviet Astronomy Letters, 12, 220
* Gorshkov et al. (2011) Gorshkov, A. G., Ipatov, A. V., Konnikova, V. K., et al. 2011, Astronomy Reports, 55, 97, doi: 10.1134/S106377291102003X
* Hardee (1982) Hardee, P. E. 1982, ApJ, 257, 509, doi: 10.1086/160008
* Hardee (2003) —. 2003, ApJ, 597, 798, doi: 10.1086/381223
* Komatsu et al. (2009) Komatsu, E., Dunkley, J., Nolta, M. R., et al. 2009, ApJS, 180, 330, doi: 10.1088/0067-0049/180/2/330
* Kovalev et al. (2008) Kovalev, Y. Y., Lobanov, A. P., Pushkarev, A. B., & Zensus, J. A. 2008, A&A, 483, 759, doi: 10.1051/0004-6361:20078679
* Kudryavtseva et al. (2011) Kudryavtseva, N. A., Gabuzda, D. C., Aller, M. F., & Aller, H. D. 2011, MNRAS, 415, 1631, doi: 10.1111/j.1365-2966.2011.18808.x
* Larionov et al. (2013) Larionov, V. M., Jorstad, S. G., Marscher, A. P., et al. 2013, ApJ, 768, 40, doi: 10.1088/0004-637X/768/1/40
* Liao et al. (2014) Liao, N. H., Bai, J. M., Liu, H. T., et al. 2014, ApJ, 783, 83, doi: 10.1088/0004-637X/783/2/83
* Lister et al. (2013) Lister, M. L., Aller, M. F., Aller, H. D., et al. 2013, AJ, 146, 120, doi: 10.1088/0004-6256/146/5/120
* Liu et al. (2012) Liu, X., Mi, L., Liu, B., & Li, Q. 2012, Ap&SS, 342, 465, doi: 10.1007/s10509-012-1191-6
* Lobanov (1998) Lobanov, A. P. 1998, A&A, 330, 79. https://arxiv.org/abs/astro-ph/9712132
* Marcaide & Shapiro (1984) Marcaide, J. M., & Shapiro, I. I. 1984, ApJ, 276, 56, doi: 10.1086/161592
* Nesci et al. (2005) Nesci, R., Massaro, E., Rossi, C., et al. 2005, AJ, 130, 1466, doi: 10.1086/444538
* Nilsson et al. (2008) Nilsson, K., Pursimo, T., Sillanpää, A., Takalo, L. O., & Lindfors, E. 2008, A&A, 487, L29, doi: 10.1051/0004-6361:200810310
* Ostorero et al. (2001) Ostorero, L., Raiteri, C. M., Villata, M., et al. 2001, Mem. Soc. Astron. Italiana, 72, 147
* Perucho et al. (2012) Perucho, M., Kovalev, Y. Y., Lobanov, A. P., Hardee, P. E., & Agudo, I. 2012, ApJ, 749, 55, doi: 10.1088/0004-637X/749/1/55
* Poon et al. (2009) Poon, H., Fan, J. H., & Fu, J. N. 2009, ApJS, 185, 511, doi: 10.1088/0067-0049/185/2/511
* Pushkarev et al. (2012) Pushkarev, A. B., Hovatta, T., Kovalev, Y. Y., et al. 2012, A&A, 545, A113, doi: 10.1051/0004-6361/201219173
* Pushkarev et al. (2009) Pushkarev, A. B., Kovalev, Y. Y., Lister, M. L., & Savolainen, T. 2009, A&A, 507, L33, doi: 10.1051/0004-6361/200913422
* Pushkarev et al. (2017) —. 2017, MNRAS, 468, 4992, doi: 10.1093/mnras/stx854
* Raiteri et al. (2003) Raiteri, C. M., Villata, M., Tosti, G., et al. 2003, A&A, 402, 151, doi: 10.1051/0004-6361:20030256
* Rani et al. (2014) Rani, B., Krichbaum, T. P., Marscher, A. P., et al. 2014, A&A, 571, L2, doi: 10.1051/0004-6361/201424796
* Rani et al. (2013) Rani, B., Krichbaum, T. P., Fuhrmann, L., et al. 2013, A&A, 552, A11, doi: 10.1051/0004-6361/201321058
* Rastorgueva et al. (2009) Rastorgueva, E. A., Wiik, K., Savolainen, T., et al. 2009, A&A, 494, L5, doi: 10.1051/0004-6361:200811425
* Sbarufatti et al. (2005) Sbarufatti, B., Treves, A., & Falomo, R. 2005, ApJ, 635, 173, doi: 10.1086/497022
* Sillanpaa et al. (1988) Sillanpaa, A., Haarala, S., Valtonen, M. J., Sundelius, B., & Byrd, G. G. 1988, ApJ, 325, 628, doi: 10.1086/166033
* Simonetti et al. (1985) Simonetti, J. H., Cordes, J. M., & Heeschen, D. S. 1985, ApJ, 296, 46, doi: 10.1086/163418
* Villata & Raiteri (1999) Villata, M., & Raiteri, C. M. 1999, A&A, 347, 30
* Volvach et al. (2012) Volvach, A. E., Volvach, L. N., Bychkova, V. S., et al. 2012, Astronomy Reports, 56, 275, doi: 10.1134/S1063772912030079
* Wagner et al. (1996) Wagner, S. J., Witzel, A., Heidt, J., et al. 1996, AJ, 111, 2187, doi: 10.1086/117954
* Wu et al. (2007) Wu, J., Zhou, X., Ma, J., et al. 2007, AJ, 133, 1599, doi: 10.1086/511773
|
∎
11institutetext: R. Machleidt 22institutetext: Department of Physics,
University of Idaho, Moscow, ID 83844, USA
22email<EMAIL_ADDRESS>
# What is ab initio?
R. Machleidt
(Received: date / Accepted: date)
###### Abstract
Microscopic nuclear theory is based on the tenet that atomic nuclei can be
accurately described as collections of point-like nucleons interacting via
two- and many-body forces obeying nonrelativistic quantum mechanics—and the
concept of the ab initio approach is to calculate nuclei accordingly. The
forces are fixed in free-space scattering and must be accurate. We will
critically review the history of this approach from the early beginnings until
today. An analysis of current ab initio calculations reveals that some
mistakes of history are being repeated today. The ultimate goal of nuclear
theory are high-precision ab initio calculations which, as it turns out, may
be possible only at the fifths order of the chiral expansion. Thus, for its
fulfillment, nuclear theory is still facing an enormous task.
## 1 Introduction
The tenet of microscopic nuclear theory is that atomic nuclei can be
accurately described as collections of point-like nucleons interacting via
two- and many-body forces obeying nonrelativistic quantum mechanics—the forces
being fixed in free-space scattering.
The microscopic or ab initio approach to nuclear structure and reactions is
then defined as calculating the properties of nuclei in accordance with the
tenet.
It is the purpose of this note to discuss how consistent or inconsistent the
fundamental model of nuclear theory has been pursued through the history of
nuclear physics and to provide an outlook for the future.
## 2 Early history of the microscopic approach
The microscopic approach to nuclear structure is almost as old as nuclear
physics itself. Brueckner and co-workers introduced Brueckner theory as early
as 1954 [1] and performed the first semi-realistic microscopic nuclear matter
calculation in 1958 [2]. Already that same year, Brueckner discussed finite
nuclei proposing the local density approximation [3].
In the second half of the 1960’s, one of the hottest topics in nuclear
structure physics was calculating the properties of finite nuclei without
recourse through nuclear matter using Brueckner-Hartree-Fock (BHF) theory. The
Oak Ridge National Laboratory (ORNL) with its computer power played a leading
role in this effort that was guided by Thomas Davies and Michel Baranger [4,
5]. BHF (and coupled cluster) calculations of finite nuclei continued into the
early 1970s with work by the Bochum [6] and the Bonn-Jülich groups [7].
In parallel to the above developments, research on the microscopic derivation
of the shell-model effective interaction was conducted (again, applying
Brueckner theory) that had been kicked off by Kuo and Brown in 1966 [8].
Applying the nucleon-nucleon ($NN$) potentials available at the time, the BHF
approach reproduced about one half of the binding energies of closed-shell
nuclei which, in the early phase, was seen as a great success [4], but in the
long run did not satisfy demands for more quantitative predictions. Therefore,
a departure from the microscopic approach happened around 1973 as reflected
most notably in a lead-talk by Michel Baranger at the International Conference
on Nuclear Physics in Munich in 1973 [9].
The shell-model effective interaction suffered a similar fate at the
International Conference on Effective Interactions and Operators in Nuclei in
Tucson, Arizona, in 1975, organized by Bruce Barrett [10].
And so it happened that in the early 1970s, the microscopic approach was
abandoned and replaced by phenomenological effective interactions (also know
as mean-field models): the Skyme interaction [11] as revived by Vautherin and
co-workers [12, 13], the Gogny force [14, 15], and the relativistic mean-field
model of Walecka [16, 17].
Ironically, the calculations with those effective interactions continued to be
called “microscopic”, for which John Negele had provided the (debatable)
justification in his Ph.D. thesis of 1970 [18]. Before calculating finite
nuclei in the local density approximation, Negele had adjusted the
insufficient binding of nuclear matter provided by the Reid soft-core
potential [19] (11 MeV per nucleon) by hand to the presumed empirical value of
15.68 MeV making “the assumption that when higher-order corrections have been
evaluated carefully, nuclear-matter theory will indeed produce the correct
binding” [18]. Negele had many followers [20, 21, 22].
However, the true “deeper reason” for those effective interactions was much
simpler: “To get better results!” [23]. Clearly, the trends that won
popularity in the early 1970s were a setback for the fundamental research in
nuclear structure.
Nuclear structure theory at its basic level is not about fitting data to get
“good” results. Fundamental nuclear structure theory is about answering the
question:
> Do the same nuclear forces that explain free-space scattering experiments
> also explain the properties of finite nuclei and nuclear matter when applied
> in nuclear many-body theory?
One can think of many reasons why the basic tenet should be wrong. According
to the EMC effect, nucleons swell when inserted into nuclei which might affect
the force between nucleons [24]. Meson exchange in the nuclear medium may be
different than in free-space for various reasons [25, 26, 27]. The excitation
of resonances, e. g. $\Delta(1232)$ isobars, within the nucleon-nucleon
interaction process is subject to changes when happening in a nuclear medium
[28, 29, 30, 31]. And many more ideas have been advanced, like e. g., Brown-
Rho scaling [32]. In fact, in the 1970s, a popular belief was that medium
effects on the $NN$ interaction may be the solution to the problem of lacking
saturation [33].
Thus, it is a good question to ask whether medium modifications of nuclear
forces show up in a noticeable way and/or are even needed for quantitative
nuclear structure predictions. But when we re-adjust the free-space forces
arbitrarily to get “good” results, then we will never find out. Note also that
at some (high) energy and high density, the picture of point-like nucleons is
bound to break down [34]. So, the issue behind the nuclear theory tenet is:
Are the energies typically involved in conventional nuclear structure physics
low enough to treat nucleons as structure-less objects?
To come back to history: the renunciation of the truly microscopic approach
lasted about two decades (essentially the 1970s and 80s). Then, in the early
1990s, the microscopic theory was revived by the Argonne-Urbana group [35,
36]. The crucial element in those new microscopic calculations was the
inclusion of a three-nucleon force (3NF). The idea of a nuclear 3NF was not
new. In fact, it is almost as old as meson theory itself [37]. But for years
it had been considered just an academic topic, too difficult to incorporate
into actual calculations, anyhow. But the persistent failure to saturate
nuclear matter at reasonable energies and densities, as well as the the
underbinding of nuclei, finally compelled nuclear structure physicists to take
a serious look at the 3NF issue, as explained in the exemplary Comment by Ben
Day [38] based upon first test calculations by the Urbana group [39]. The 3NF
definitely improved nuclear saturation and the properties of light nuclei,
even though nothing was perfect [36].
## 3 Recent history
After the year of 2000, two changes occurred. First, the term ‘microscopic’
was increasingly replaced by the term ‘ab initio’ [40]—for reasons nobody
knows (but nothing to worry about because both mean the same). Second and more
importantly, nuclear forces based upon chiral effective field theory (EFT)
entered the picture [41, 42]. This development was of great advantage. Note
that for a microscopic approach to be truly microscopic, the free-space forces
need to be accurate. But with phenomenological or meson-theoretic forces it
was difficult to define what sufficiently accurate means, since the errors in
those theories are unknown. However, in the framework of an EFT, the
theoretical uncertainty can be determined and, thus, related with the accuracy
of the predictions. Hence, in the framework of an EFT:
> Accurate free-space forces are forces that predict experiment within the
> theoretical uncertainty of the EFT at the given order.
After 2000, it also became well established that predictive nuclear structure
must include 3NFs, besides the usual two-nucleon force (2NF) contribution.
Another advantage of chiral EFT is then that it generates 2NFs and multi-
nucleon forces simultaneously and on an equal footing. In the $\Delta$-less
theory [43, 44], 3NFs occur for the first time at next-to-next-to-leading
order (NNLO) and continue to have additional contributions in higher orders.
If an explicit $\Delta$-isobar is included in chiral EFT ($\Delta$-full theory
[45, 46, 47, 48]), then 3NF contributions start already at next-to-leading
order (NLO).
In the initial phase, the 3NFs were typically adjusted in $A=3$ and/or the
$A=4$ systems and the ab initio calculations were driven up to the oxygen
region [49]. It turned out that for $A\raisebox{-1.29167pt}{\small$\
\stackrel{{\scriptstyle\textstyle<}}{{\sim}}$ }16$ the ground-state energies
and radii are predicted about right, no matter what type of chiral or
phenomenological potentials were applied (local, nonlocal, soft, hard, etc.)
and what the details of the 3NF adjustments to few-body systems were [49, 50,
51, 52, 53, 54].
However, around the year of 2015, the picture changed, when the many-body
practitioners were able to move up to medium-mass nuclei (e. g., the calcium
or even the tin regions). Large variations of the predictions now occurred
depending on what forces were used, and cases of severe underbinding [55] as
well as of substantial overbinding [56] were observed. Ever since, the nuclear
structure community understands that accurate ab initio explanations of
intermediate and heavy nuclei is an outstanding problem.
There have been several attempts to predict the properties of medium-mass
nuclei with more accuracy. Of the various efforts, we will now list four
cases, which are representative for the status, and will denote each case with
a short label for ease of communication. We restrict ourselves to cases, where
the properties of medium-mass nuclei and nuclear matter have been calculated,
because the simultaneous description of both systems is part of the
problem.111Other interesting cases are the models by Soma et al. [57] and
Maris et al. [54] for which, however, presently no nuclear matter results are
available.
Figure 1: Upper panel: Ground-state energies per nucleon, $E/A$, of selected
closed-shell oxygen, calcium, and nickel isotopes as obtained in the “Hoppe”
case [58]. Results are shown for various chiral interactions as denoted. The
blue and orange bands give the NNLO and N3LO uncertainty estimates,
respectively. $\Lambda=450$ MeV in all cases except the green curve. Black
bars indicate experimental data. Lower panel: Same as upper panel, but for
charge radii. (Reproduced from Ref. [58] with permission.)
Figure 2: Ground-state energies per nucleon (top panel) and point-proton rms
radii (bottom panel) for selected medium-mass isotopes as obtained in the
“Hüther” case [59]. The light blue and pink bands represent the theoretical
uncertainties at NNLO and N3LO, respectively. $\Lambda=450$ MeV. Black bars
indicate the experimental data. (Figure courtesy of R. Roth)
* •
“Magic” [60, 61]: A seemingly successful interaction for the intermediate mass
region commonly denoted by “1.8/2.0(EM)” (sometimes dubbed “the Magic force”).
It is a similarity renormalization group (SRG) evolved version of the N3LO 2NF
of Ref. [42] complemented by a NNLO 3NF adjusted to the triton binding energy
and the point charge radius of 4He. With this force, the ground-state energies
all the way up to the tin isotopes are reproduced perfectly—but with charge
radii being on the smaller side [62, 63]. Nuclear matter saturation is also
reproduced reasonably well, but at a slightly too high saturation density
[60].
* •
“GO” [64, 65]: A family of $\Delta$-full NNLO potentials constructed by the
Göteborg/Oak Ridge (GO) group. The authors claim to obtain “accurate binding
energies and radii for a range of nuclei from $A=16$ to $A=132$, and provide
accurate equations of state for nuclear matter” [65].
* •
“Hoppe” [66, 58]: Recently developed soft chiral 2NFs [67] at NNLO and N3LO
complemented with 3NFs at NNLO and N3LO, respectively, to fit the triton
binding energy and nuclear matter saturation. These forces applied in in-
medium similarity renormalization group (IM-SRG [68]) calculations of finite
nuclei up to 68Ni predict underbinding and slightly too large radii [58], see
Fig. 1.
* •
“Hüther” [59]: The same 2NFs used in “Hoppe”, but with the 3NFs adjusted to
the triton and 16O ground-state energies. The interactions so obtained
reproduce accurately experimental energies and point-proton radii of nuclei up
to 78Ni [59], see Fig. 2. However, when the 2NF plus 3NF combinations of
“Hüther” are utilized in nuclear matter, then overbinding and no saturation at
realistic densities is obtained [69], see Fig. 3.
Figure 3: Energy per nucleon, $E/A$, as a function of density, $\rho$, of
symmetric nuclear matter as obtained in calculations with the 2NFs and 3NFs
consistently at NNLO [69]. In the two cases shown, the 2NF is the same, while
the 3NFs are the ones used in the calculations of finite nuclei in the “Hoppe”
and “Huether” cases as denoted. $\Lambda=450$ MeV in both cases. The error
bars show the theoretical uncertainties around saturation, which is expected
to occur in the area of the gray box.
Obviously, in some cases, there appears to be a problem with achieving
simultaneously accurate results for nuclear matter and medium-mass nuclei: In
the “Hoppe” case, nuclear matter is saturated correctly, but nuclei are
underbound; while in the “Hüther” case, nuclei are bound accurately, but
nuclear matter is overbound. Other cases seem to have solved this problem. But
are they all truly ab initio? Our assessment:
* •
“Magic”: The construction of this force includes some inconsistencies. The 2NF
is SRG evolved, while the 3NF is not. Moreover, the SRG evolved 2NF is used
like an original force with the induced 3NFs omitted. Note that ab inito also
implies that the forces are based upon some sort of theory in a consistent
way. This is here not true and, thus, this case is not ab initio.
* •
“GO”: In Ref. [70] it has been shown that the predictions by the $\Delta$-full
$NN$ potentials at NNLO constructed by the Gőteborg-Oak Ridge (GO) group [65]
are up to 40 times outside the theoretical error of chiral EFT at NNLO. So,
they fail on accuracy. The reason for their favorable reproduction of the
energies (and radii) of intermediate-mass nuclei, can be traced to incorrect
$P$-wave and $\epsilon_{1}$ mixing parameters [70]. Thus, this case is
especially far from being ab initio. It is just a repetition of the mistakes
of the early 1970s.
* •
“Hoppe”: In this case, the 2NF and 3NF forces are consistently chiral EFT
based. Moreover, the 2NFs are accurate. However, there is another accuracy
aspect that is, in general, quietly ignored [71, 72]: Are the 3NFs accurate?
The accuracy of the chiral 3NF at NNLO was thoroughly investigated in Ref.
[73] for a variety of cutoffs ranging from 400-550 MeV and large variations of
the NNLO 3NF parameters, $c_{D}$ and $c_{E}$. A typical result is shown in
Fig. 4. It is seen that the 3$N$ data are reproduced within the truncation
errors at NNLO (green bands). On the other hand, it is also clearly seen that
the theoretical uncertainties are very large. Moreover, it was found in Ref.
[73] that the cutoff dependence is weak and that the variations of the 3NF
LECs $c_{D}$ and $c_{E}$ make only small differences relative to the large
uncertainties. Thus, we can assume that the NNLO 3NFs used in “Hoppe” will
yield results that lie within the NNLO uncertainties shown in Fig. 4 by the
green bands and, consequently, the “Hoppe” 3NF is accurate. Hence, “Hoppe”
passes on all accounts and is, therefore, truly ab initio.
* •
“Hüther”: An assessment similar to “Hoppe” applies. Thus, this case is also
truly ab initio.
Figure 4: Predictions for the differential cross section, nucleon and
deuteron analyzing powers $A^{n}_{y}$ and $A^{d}_{y}$ as well as deuteron
tensor analyzing powers $A_{yy}$, $A_{xz}$, and $A_{xx}$ in elastic
nucleon–deuteron scattering at a laboratory energy of 135 MeV at NLO (yellow
bands) and NNLO (green bands). The light- (dark-) shaded bands indicate 95%
(68%) confidence levels. The dotted (dashed) lines show the results based on
the CD-Bonn $NN$ potential [74] (CD-Bonn $NN$ potential in combination with
the Tucson-Melbourne 3NF [75]). Black symbols represent the data together with
their experimental errors. (Reproduced from Ref. [73].)
The bottom line is that not all calculations, which have been published in the
literature under the label of ab initio, are really ab initio. Indeed, of the
cases we considered here, only 50% pass the test. But we need to point out
that even in the two cases we declared ab initio, there are concerns. The NNLO
predictions by Hoppe and Hüther for finite nuclei barely overlap within their
theoretical uncertainties and, for nuclear matter, they do not overlap at all.
Obviously, there are problems with the error estimates and the uncertainties
are much larger than the shown ones. The true NNLO truncation errors of the
Hoppe and Hüther calculations are probably as large as the differences between
the two predictions. In this way, the two predictions are actually consistent
with each other, in spite of their seeming discrepancy. Chiral EFT is a model-
independent theory and, thus, different calculations at the same order should
agree within truncation errors.
At N3LO the predictions differ even more. However, for current N3LO
calculations, a strong caveat is in place. As pointed out in Ref. [76], there
is a problem with the regularized 3NF at N3LO (and higher orders) in all
present nuclear structure calculations. The N3LO 3NFs currently in use are all
regularized by a multiplicative regulator applied to the 3NF expressions that
are derived from dimensional regularization. This approach leads to a
violation of chiral symmetry at N3LO and destroys the consistency between two-
and three-nucleon forces [76]. Consequently, all current calculations that
include a N3LO 3NF contain an uncontrolled error and are, therefore,
unreliable. When a consistent regularization scheme has been found, the
calculations have to be repeated. At the present time, reliable predictions
exist only at NNLO, NLO, and LO.
## 4 The future: ab initio plus precision
Figure 5: Latest ab initio predictions by the LENPIC collaboration [54]:
Ground-state energies and point-proton radii for doubly magic oxygen and
calcium isotopes obtained from the $NN$ potential of Ref. [77] complemented by
NNLO 3NFs using a cutoff of 450 MeV (left-hand panel) and of 500 MeV (right-
hand panel). The blue squares represent the predictions by complete NNLO
calculations with the blue error bands showing the chiral NNLO truncation
uncertainties at the 95% confidence level. The green and purple points and
pink error bands are based upon incomplete calculations and are to be ignored.
Black bars indicate the experimental data. (Reproduced from Ref. [54] with
permission.)
It is comforting to know that at least a few correct ab initio calculations do
exist. But these cases show that the precision at NNLO is very poor. The same
is true for the latest LENPIC calculations [54], see Fig. 5 (which we did not
include in our case study, because nuclear matter results are lacking). At
N3LO (if one day correct such calculations become available) the precision
will most likely not be substantially better.
As stated at the outset, the purpose of the ab initio approach is to test if
the tenet of nuclear theory is correct or not. Within huge errors as, e. g. in
Fig. 4, any approach may come out right. So, that is not a good basis for a
reliable test. We need more precision! This is in particular true for the 3NF
and the reproduction of the 3$N$ data, which has been thoroughly investigated
in Refs [73, 78] with the conclusion that, at N4LO, there is a chance to
achieve the desirable precision—for several reasons. The long- and
intermediate-range topologies of the 3NF at N4LO are expected to be much
larger than the corresponding ones at N3LO because, at N4LO, the subleading
$\pi NN$ seagull vertex is involved with LECs $c_{i}$, which are large [79,
80]. This will provide the 3NF at N4LO with more leverage as compared to N3LO.
Moreover, at N4LO, 13 new 3$N$ contact terms occur [81] with essentially free
parameters introducing considerable flexibility [82, 78] (see also Ref. [83]).
Worth mentioning is also that, at N4LO, the 3NF includes all 20 operators of
the most general 3NF [84]. Furthermore, the plentiful N4LO 3NF terms may also
provide what is needed to improve the status of the medium-mass nuclei and
nuclear matter.
Thus, the future of truly microscopic nuclear structure is to go for complete
N4LO calculations—a gigantic task.
## 5 Summary and outlook
To summarize, let me just reiterate the main statements.
The tenet of microscopic nuclear theory is:
> Atomic nuclei can be accurately described as collections of point-like
> nucleons interacting via two- and many-body forces obeying nonrelativistic
> quantum mechanics—the forces being fixed in free-space scattering.
And in the ab initio approach, nuclei are calculated accordingly.
We need to critically investigate if the tenet is true. To that end, we have
to answer the question:
> Do the same nuclear forces that explain free-space scattering experiments
> also explain the properties of finite nuclei and nuclear matter when applied
> in nuclear many-body theory?
Either way, the answer is of fundamental relevance. The correct answer can
only be obtained if the free-space forces are accurate, where accurate is
defined by:
> Accurate free-space forces are forces that predict experiment within the
> theoretical uncertainty of the applied EFT at the given order.
Moreover, one would also require that the applied nuclear forces are based
upon some sort of theory in a consistent way.
Without strictly adhering to these principles, the true answer to the
fundamental question will not be found. Once again, the goal is not to obtain
“good” results, but to understand whether there are non-negligible medium
effects on nuclear forces when inserted into the nuclear many-body problem.
In our community, the term ab initio is often used in a way that is too lose
and many calculations that are presented as ab initio do not pass muster. Such
calculations repeat the mistakes of history and, thus, do not move us forward.
The ultimate goal of nuclear theory should be to conduct calculations that
test the tenet with high precision. There is strong evidence that this
precision can only be achieved at N4LO of the chiral EFT expansion.
Calculations of this kind, which must also include all many-body forces at
that order, are very challenging, and the current status of ab initio
calculations is far from meeting that goal.
In this context, it should be mentioned that the uncertainties of the many-
body calculations must also be included in the error analysis. With
calculations now moving up to heavy nuclei, current many-body techniques need
to be tested critically for which bechmark calculations would be the right
tool.
The work that is left to do in microscopic nuclear theory is monumental.
###### Acknowledgements.
This work was supported in part by the U.S. Department of Energy under Grant
No. DE-FG02-03ER41270.
## References
* [1] K. A. Brueckner, C. A. Levinson, and H. M. Mahmoud. Two-Body Forces and Nuclear Saturation. 1. Central Forces. Phys. Rev., 95:217–228, 1954.
* [2] K. A. Brueckner and J. L. Gammel. Properties of Nuclear Matter. Phys. Rev., 109:1023–1039, 1958.
* [3] K. A. Brueckner, J. L. Gammel, and H. Weitzner. Theory of Finite Nuclei. Phys. Rev., 110:431–445, 1958.
* [4] M. Baranger. Recent progress in the understanding of finite nuclei from the two-nucleon interaction. In Proceedings of the International School of Physics “Enrico Fermi”, Course XL: Nuclear Structure and Nuclear Reactions, Varenna, 1967, eds. M. Jean and R. A. Ricci, pages 511–614, Academic Press, New York and London, 1969.
* [5] K. T. R. Davies, M. Baranger, R. M. Tarbutton, and T. T. S. Kuo. Brueckner-Hartree-Fock Calculations of Spherical Nuclei in an Harmonic-Oscillator Basis. Phys. Rev., 177:1519–1526, 1969.
* [6] H. Kuemmel and J. G. Zabolitzky. Fully Self-Consistent Brueckner-Hartree-Fock and Renormalized Brueckner-Hartree-Fock Calculation for He-4 and O-16. Phys. Rev. C, 7:547–552, 1973.
* [7] R. Machleidt, H. Muether, and A. Faessler. One Boson Exchange Potential and the Ground State of O-16. Nucl. Phys. A, 241:18–28, 1975.
* [8] T. Kuo and G. E. Brown. Structure of finite nuclei and the free nucleon-nucleon interaction: An application to O-18 and F-18. Nucl. Phys., 85:40–86, 1966.
* [9] M. Baranger. Effective interactions and the nucleon-nucleon force. In Proceedings of the International Conference on Nuclear Physics, Munich, 1973, Vol. 2, eds. J. de Boer and H. J. Mang, pages 93–106, North-Holland Publishing Company, Amsterdam, 1973.
* [10] Proceedings of the International Conference on Effective Interactions and Operators in Nuclei, Tucson, Arizona, USA, ed. B. R. Barrett. Lecture Notes in Physics, Vol. 40, Springer, Berlin, 1975.
* [11] T. Skyrme. The effective nuclear potential. Nucl. Phys., 9:615–634, 1959.
* [12] D. Vautherin and D. M. Brink. Hartree-Fock calculations with Skyrme’s interaction. 1. Spherical nuclei. Phys. Rev. C, 5:626–647, 1972.
* [13] D. Vautherin. Selfconsistent description of nuclei. In Proceedings of the International Conference on Nuclear Physics, Munich, 1973, Vol. 2, eds. J. de Boer and H. J. Mang, pages 107–130, North-Holland Publishing Company, Amsterdam, 1973.
* [14] D. Gogny. In Proceedings of the International Conference on Nuclear Physics, Munich, 1973, Vol. 1, eds. J. de Boer and H. J. Mang, page 48, North-Holland Publishing Company, Amsterdam, 1973.
* [15] J. Decharge and D. Gogny. Hartree-Fock-Bogolyubov calculations with the D1 effective interactions on spherical nuclei. Phys. Rev. C, 21:1568–1593, 1980.
* [16] J. D. Walecka. A Theory of highly condensed matter. Annals Phys., 83:491–529, 1974.
* [17] Brian D. Serot and John Dirk Walecka. The Relativistic Nuclear Many Body Problem. Adv. Nucl. Phys., 16:1–327, 1986.
* [18] John W. Negele. Structure of Finite Nuclei in the Local-Density Approximation. Phys. Rev. C, 1:1260–1321, 1970.
* [19] Roderick V. Reid, Jr. Local phenomenological nucleon-nucleon potentials. Annals Phys., 50:411–448, 1968.
* [20] X. Campi and D. W. Sprung. Spherical nuclei in the local density approximation. Nucl. Phys. A, 194:401–442, 1972.
* [21] G. Fái and J. Németh. Density-dependent effective interactions in finite nuclei (II). Nucl. Phys. A, 208:463–476, 1973.
* [22] R. Machleidt, K. Holinde, and J. Németh. One-boson-exchange potential and structure of finite nuclei in the local-density approximation. Nucl. Phys. A, 251:93–104, 1975.
* [23] Quote from numerous conversations the author had with nuclear structure physicists in the 1970s.
* [24] Manoj K. Banerjee. Nucleon in nuclear matter. Phys. Rev. C, 45:1359–1373, 1992.
* [25] K. Kotthoff, R. Machleidt, and D. Schutte. Meson Exchange Corrections and Properties of Nuclear Matter and Neutron Matter. Nucl. Phys. A, 264:484–492, 1976.
* [26] L. Wilets. Green’s functions method for the relativistic field theory many body problem. In Mesons in Nuclei, Vol. III, eds. M. Rho and D. Wilkinson, pages 791–837, North-Holland Publishing Company, Amsterdam, 1979.
* [27] G. E. Brown, V. Koch, and Mannque Rho. The Pion at finite temperature and density. Nucl. Phys. A, 535:701–714, 1991.
* [28] Anthony M. Green and J. A. Niskanen. The Saturating Effect of the Delta (1236) in Nuclear Matter. Nucl. Phys. A, 249:493–509, 1975.
* [29] Anthony M. Green. Nucleon Resonances in Nuclei. Rept. Prog. Phys., 39:1109–1190, 1976.
* [30] B. D. Day and F. Coester. Influence of Virtual Delta States on the Saturation Properties of Nuclear Matter. Phys. Rev. C, 13:1720–1740, 1976.
* [31] K. Holinde and R. Machleidt. Effect of the Delta (1236) Resonance on n n Scattering, Nuclear Matter and Neutron Matter. Nucl. Phys. A, 280:429–466, 1977.
* [32] G. E. Brown and Mannque Rho. Scaling effective Lagrangians in a dense medium. Phys. Rev. Lett., 66:2720–2723, 1991.
* [33] R. Machleidt. The Meson theory of nuclear forces and nuclear structure. Adv. Nucl. Phys., 19:189–376, 1989.
* [34] Omar Benhar. Testing the Paradigm of Nuclear Many-Body Theory. Particles, 6:611, 2023.
* [35] R. B. Wiringa. From deuterons to neutron stars: variations in nuclear many-body theory. Rev. Mod. Phys., 65:231–242, 1993.
* [36] B. S. Pudliner, V. R. Pandharipande, J. Carlson, and Robert B. Wiringa. Quantum Monte Carlo calculations of A $<$= 6 nuclei. Phys. Rev. Lett., 74:4396–4399, 1995.
* [37] H. Primakoff and T. Holstein. Many-Body Interactions in Atomic and Nuclear Systems. Phys. Rev., 55:1218–1234, 1939.
* [38] B. D. Day. Nuclear Saturation and Nuclear Forces. Comments Nucl. Part. Phys., 11(3):115–126, 1983.
* [39] J. Carlson, V. R. Pandharipande, and Robert B. Wiringa. Three-nucleon interaction in 3-body, 4-body, and infinite-body systems. Nucl. Phys. A, 401:59–85, 1983.
* [40] P. Navratil, J. P. Vary, and B. R. Barrett. Properties of C-12 in the ab initio nuclear shell model. Phys. Rev. Lett., 84:5728–5731, 2000.
* [41] E. Epelbaum, Walter Gloeckle, and Ulf-G. Meissner. Nuclear forces from chiral Lagrangians using the method of unitary transformation. 2. The two nucleon system. Nucl. Phys. A, 671:295–331, 2000.
* [42] D. R. Entem and R. Machleidt. Accurate charge dependent nucleon nucleon potential at fourth order of chiral perturbation theory. Phys. Rev. C, 68:041001, 2003.
* [43] R. Machleidt and D. R. Entem. Chiral effective field theory and nuclear forces. Phys. Rept., 503:1–75, 2011.
* [44] Evgeny Epelbaum, Hans-Werner Hammer, and Ulf-G. Meissner. Modern Theory of Nuclear Forces. Rev. Mod. Phys., 81:1773–1825, 2009.
* [45] C. Ordonez, L. Ray, and U. van Kolck. The Two nucleon potential from chiral Lagrangians. Phys. Rev. C, 53:2086–2105, 1996.
* [46] Norbert Kaiser, S. Gerstendorfer, and W. Weise. Peripheral NN scattering: Role of delta excitation, correlated two pion and vector meson exchange. Nucl. Phys. A, 637:395–420, 1998.
* [47] Hermann Krebs, Evgeny Epelbaum, and Ulf-G. Meissner. Nuclear forces with Delta-excitations up to next-to-next-to-leading order. I. Peripheral nucleon-nucleon waves. Eur. Phys. J. A, 32:127–137, 2007.
* [48] E. Epelbaum, H. Krebs, and Ulf-G. Meissner. Delta-excitations and the three-nucleon force. Nucl. Phys. A, 806:65–78, 2008.
* [49] Bruce R. Barrett, Petr Navratil, and James P. Vary. Ab initio no core shell model. Prog. Part. Nucl. Phys., 69:131–181, 2013.
* [50] Robert Roth, Joachim Langhammer, Angelo Calci, Sven Binder, and Petr Navratil. Similarity-Transformed Chiral NN+3N Interactions for the Ab Initio Description of 12-C and 16-O. Phys. Rev. Lett., 107:072501, 2011.
* [51] M. Piarulli et al. Light-nuclei spectra from chiral dynamics. Phys. Rev. Lett., 120(5):052503, 2018.
* [52] D. Lonardoni, S. Gandolfi, J. E. Lynn, C. Petrie, J. Carlson, K. E. Schmidt, and A. Schwenk. Auxiliary field diffusion Monte Carlo calculations of light and medium-mass nuclei with local chiral interactions. Phys. Rev. C, 97(4):044318, 2018.
* [53] P. Maris et al. Light nuclei with semilocal momentum-space regularized chiral interactions up to third order. Phys. Rev. C, 103(5):054001, 2021.
* [54] P. Maris et al. Nuclear properties with semilocal momentum-space regularized chiral interactions beyond N2LO. Phys. Rev. C, 106(6):064002, 2022.
* [55] D. Lonardoni, A. Lovato, Steven C. Pieper, and R. B. Wiringa. Variational calculation of the ground state of closed-shell nuclei up to $A=40$. Phys. Rev. C, 96(2):024326, 2017.
* [56] Sven Binder, Joachim Langhammer, Angelo Calci, and Robert Roth. Ab Initio Path to Heavy Nuclei. Phys. Lett. B, 736:119–123, 2014.
* [57] V. Somà, P. Navrátil, F. Raimondi, C. Barbieri, and T. Duguet. Novel chiral Hamiltonian and observables in light and medium-mass nuclei. Phys. Rev. C, 101(1):014318, 2020.
* [58] J. Hoppe, C. Drischler, K. Hebeler, A. Schwenk, and J. Simonis. Probing chiral interactions up to next-to-next-to-next-to-leading order in medium-mass nuclei. Phys. Rev. C, 100(2):024318, 2019.
* [59] Thomas Hüther, Klaus Vobig, Kai Hebeler, Ruprecht Machleidt, and Robert Roth. Family of Chiral Two- plus Three-Nucleon Interactions for Accurate Nuclear Structure Studies. Phys. Lett. B, 808:135651, 2020.
* [60] K. Hebeler, S. K. Bogner, R. J. Furnstahl, A. Nogga, and A. Schwenk. Improved nuclear matter calculations from chiral low-momentum interactions. Phys. Rev. C, 83:031301, 2011.
* [61] Kai Hebeler. Three-nucleon forces: Implementation and applications to atomic nuclei and dense matter. Phys. Rept., 890:1–116, 2021.
* [62] J. Simonis, S. R. Stroberg, K. Hebeler, J. D. Holt, and A. Schwenk. Saturation with chiral interactions and consequences for finite nuclei. Phys. Rev. C, 96(1):014303, 2017.
* [63] T. D. Morris, J. Simonis, S. R. Stroberg, C. Stumpf, G. Hagen, J. D. Holt, G. R. Jansen, T. Papenbrock, R. Roth, and A. Schwenk. Structure of the lightest tin isotopes. Phys. Rev. Lett., 120(15):152503, 2018.
* [64] A. Ekström, G. Hagen, T. D. Morris, T. Papenbrock, and P. D. Schwartz. $\Delta$ isobars and nuclear saturation. Phys. Rev. C, 97(2):024332, 2018.
* [65] W. G. Jiang, A. Ekström, C. Forssén, G. Hagen, G. R. Jansen, and T. Papenbrock. Accurate bulk properties of nuclei from $A=2$ to $\infty$ from potentials with $\Delta$ isobars. Phys. Rev. C, 102(5):054301, 2020.
* [66] C. Drischler, K. Hebeler, and A. Schwenk. Chiral interactions up to next-to-next-to-next-to-leading order and nuclear saturation. Phys. Rev. Lett., 122(4):042501, 2019.
* [67] D. R. Entem, R. Machleidt, and Y. Nosyk. High-quality two-nucleon potentials up to fifth order of the chiral expansion. Phys. Rev. C, 96(2):024004, 2017.
* [68] H. Hergert, S. K. Bogner, T. D. Morris, A. Schwenk, and K. Tsukiyama. The In-Medium Similarity Renormalization Group: A Novel Ab Initio Method for Nuclei. Phys. Rept., 621:165–222, 2016.
* [69] Francesca Sammarruca and Randy Millerson. Exploring the relationship between nuclear matter and finite nuclei with chiral two- and three-nucleon forces. Phys. Rev. C, 102(3):034313, 2020.
* [70] Y. Nosyk, D. R. Entem, and R. Machleidt. Nucleon-nucleon potentials from $\Delta$-full chiral effective-field-theory and implications. Phys. Rev. C, 104(5):054001, 2021.
* [71] A. Kievsky, M. Viviani, and S. Rosati. Polarization observables in p - d scattering below 30-MeV. Phys. Rev. C, 64:024002, 2001.
* [72] L. E. Marcucci, A. Kievsky, L. Girlanda, S. Rosati, and M. Viviani. N-d Elastic Scattering Using the Hyperspherical Harmonics Approach with Realistic Local and Non-Local Interactions. Phys. Rev. C, 80:034003, 2009.
* [73] E. Epelbaum et al. Towards high-order calculations of three-nucleon scattering in chiral effective field theory. Eur. Phys. J. A, 56(3):92, 2020.
* [74] R. Machleidt. The High precision, charge dependent Bonn nucleon-nucleon potential. Phys. Rev. C, 63:024001, 2001.
* [75] S. A. Coon and H. K. Han. Reworking the Tucson-Melbourne three nucleon potential. Few Body Syst., 30:131–141, 2001.
* [76] Evgeny Epelbaum, Hermann Krebs, and Patrick Reinert. High-precision nuclear forces from chiral EFT: State-of-the-art, challenges and outlook. Front. in Phys., 8:98, 2020.
* [77] P. Reinert, H. Krebs, and E. Epelbaum. Semilocal momentum-space regularized chiral two-nucleon potentials up to fifth order. Eur. Phys. J. A, 54(5):86, 2018.
* [78] H. Witała, J. Golak, and R. Skibiński. Significance of chiral three-nucleon force contact terms for understanding of elastic nucleon-deuteron scattering. Phys. Rev. C, 105(5):054004, 2022.
* [79] Hermann Krebs, A. Gasparyan, and Evgeny Epelbaum. Chiral three-nucleon force at N4LO I: Longest-range contributions. Phys. Rev. C, 85:054006, 2012.
* [80] Hermann Krebs, A. Gasparyan, and Evgeny Epelbaum. Chiral three-nucleon force at $N^{4}LO$ II: Intermediate-range contributions. Phys. Rev. C, 87(5):054007, 2013.
* [81] L. Girlanda, A. Kievsky, and M. Viviani. Subleading contributions to the three-nucleon contact interaction. Phys. Rev. C, 84(1):014001, 2011. [Erratum: Phys.Rev.C 102, 019903 (2020)].
* [82] L. Girlanda, A. Kievsky, M. Viviani, and L. E. Marcucci. Short-range three-nucleon interaction from A=3 data and its hierarchical structure. Phys. Rev. C, 99(5):054003, 2019.
* [83] L. Girlanda, E. Filandri, A. Kievsky, L. E. Marcucci, and M. Viviani. Effect of the N3LO three-nucleon contact interaction on p-d scattering observables. Phys. Rev. C, 107(6):L061001, 2023.
* [84] E. Epelbaum, A. M. Gasparyan, H. Krebs, and C. Schat. Three-nucleon force at large distances: Insights from chiral effective field theory and the large-Nc expansion. Eur. Phys. J. A, 51(3):26, 2015.
|
# Challenges to observation of many-body localization
Piotr Sierant The Abdus Salam International Center for Theoretical Physics,
Strada Costiera 11, 34151, Trieste, Italy Institute of Theoretical Physics,
Jagiellonian University in Kraków, Łojasiewicza 11, 30-348 Kraków, Poland
Jakub Zakrzewski<EMAIL_ADDRESS>Institute of Theoretical Physics,
Jagiellonian University in Kraków, Łojasiewicza 11, 30-348 Kraków, Poland
Mark Kac Complex Systems Research Center, Jagiellonian University in Krakow,
Kraków, Poland.
###### Abstract
We study time dynamics of 1D disordered Heisenberg spin-1/2 chain focusing on
a regime of large system sizes and a long time evolution. This regime is
relevant for observation of many-body localization (MBL), a phenomenon that is
expected to freeze the dynamics of the system and prevent it from reaching
thermal equilibrium. Performing extensive numerical simulations of the
imbalance, a quantity often employed in the experimental studies of MBL, we
show that the regime of a slow power-law decay of imbalance persists to
disorder strengths exceeding by at least a factor of 2 the current estimates
of the critical disorder strength for MBL. Even though we investigate time
evolution up to few thousands tunneling times, we observe no signs of the
saturation of imbalance that would suggest freezing of system dynamics and
provide a smoking gun evidence of MBL. We demonstrate that the situation is
qualitatively different when the disorder is replaced by a quasiperiodic
potential. In this case, we observe an emergence of a pattern of oscillations
of the imbalance that is stable with respect to changes in the system size.
This suggests that the dynamics of quasiperiodic systems remain fully local at
the longest time scales we reach provided that the quasiperiodic potential is
sufficiently strong. Our study identifies challenges in an unequivocal
experimental observation of the phenomenon of MBL.
## I Introduction
Generic isolated quantum many-body systems initialized in an out-of-
equilibrium state are expected to approach featureless thermal states
described by the eigenstate thermalization hypothesis [1, 2, 3]. Many-body
localization (MBL) [4, 5] has been put forward as a mechanism that prevents
the approach to equilibrium due to an interplay of interactions and strong
disorder.
The phenomenon of MBL has received a lot of attention over the last decade [6,
7, 8]. The MBL phase is characterized by presence of local integrals of motion
[9, 10, 11, 12, 13, 14, 15] that inhibit the transport [6, 16], and slow down
the spreading of the quantum entanglement [17, 18]. MBL has been investigated
numerically in disordered spin chains [19, 20, 21, 22] that map onto spinless
fermionic chains, in systems of spinful fermions [23, 24, 25, 26] or bosons
[27, 28, 29] and found in systems with random interactions [30, 31, 32] or in
various types of quasiperiodic systems [33, 34, 35]. All those investigations
were confirming the belief that MBL is a robust mechanism of ergodicity
breaking, that can be expected to occur in a wide class of local, one-
dimensional quantum many-body systems provided that a sufficiently strong
quenched disorder is present.
This belief was challenged in [36] where it was argued that MBL might not be
stable in the asymptotic sense, i.e. in the limit of an infinite time and
system size, and the observations of earlier works indicate only a presence of
an MBL regime found at a finite system size and finite times. This lead to an
intense debate about the stability of MBL [37, 38, 39] and its dynamical
properties [40, 41, 42, 43, 44]. Despite these works, it is presently unclear
whether a stable MBL phase exists much deeper in the MBL regime than it was
previously estimated [45] or whether there is no stable MBL phase at all [46].
An example of the latter scenario is provided by disordered constrained spin
chains which, despite hosting a wide non-ergodic regime at finite system sizes
[47] become ergodic in the thermodynamic limit [48].
Figure 1: Interactions induce a slow decay of the imbalance $I(t)$ that
persists to long times. This is visualized comparing results for non-
interacting ($\Delta=0$) and interacting ($\Delta=1$) systems. Data for
disordered XXZ model (1) at disorder strength $W=4$. The squares denote the
Heisenberg time $t_{H}$ that scales exponentially with system size $L$.
The double limit of infinite time and system size is the source of
difficulties in establishing the status of MBL. On one hand, one may
investigate properties of eigenstates of many-body systems, that encode the
properties of the system at infinite time. However, the eigenstates can be
found in an unbiased fashion only for relatively small system sizes $L$ (for
instance, for the usually studied spin-1/2 chains, $L\leq 24$ [49, 50]), which
does not allow for a fully controlled extrapolation of the results to the
thermodynamic limit $L\rightarrow\infty$. On the other hand, tensor network
algorithms [51, 52] such as Time Evolving Block Decimation (TEBD) [53, 54] or
Time-Dependent Variational Principle (TDVP) [55, 56, 57, 58] allow one to
study time evolution of systems comprised of hundreds or even thousands of
sites. Unfortunately, the time evolution of many-body systems can be traced
faithfully with such algorithms only up to times restricted by the growth of
the entanglement in the system. Since, in strongly disordered systems, the
entanglement entropy grows only logarithmically in time, maximal times of
several hundred tunneling times were achieved in [59, 25, 60, 61].
Nevertheless, there is no straightforward way of extrapolating these results
to the infinite time limit.
Figure 1 illustrates the difficulties in assessing whether the system is
ergodic or MBL in a quench experiment. It shows the time evolution of the so-
called imbalance $I(t)$ for a disordered XXZ spin-1/2 chain (precise
definitions are given in the following section). An ergodic system has no
memory of its initial state and the imbalance vanishes in the long-time limit:
$I(t)\stackrel{{\scriptstyle t\rightarrow\infty}}{{\rightarrow}}0$. In
contrast, the information about the initial density profile persists
indefinitely in the MBL phase in which $I(t)\stackrel{{\scriptstyle
t\rightarrow\infty}}{{\rightarrow}}I_{0}>0$. For the non-interacting system
($\Delta=0$) one clearly sees that after initial oscillations, the imbalance
saturates to a constant value. Such a behavior allows for a straightforward
experimental observation of Anderson localization in the absence of
interactions [62, 63]. The main effect of interactions is that the imbalance
decays to much longer times, as exhibited by data for $\Delta=1$. The time
scale at which $I(t)$ ceases to decay is of the order of Heisenberg time
$t_{H}$ [64] that is proportional to an inverse of the mean level spacing of
the system and hence it is exponentially large in the system size $L$. In a
consequence, the data presented in Fig. 1 allow us only to conclude that at
the considered disorder strength $W=4$, the system is in a finite time MBL
regime [45]. The value of the imbalance in the $t\rightarrow\infty$ limit is
clearly decreasing with the system size $L$ and it is impossible to determine
from the data in Fig. 1 whether in the limit $L\rightarrow\infty$,
$t\rightarrow\infty$ the system remains MBL at $W=4$ or whether the ergodicity
is restored.
The presence of MBL regime has been demonstrated in a number of numerical
works as well as in experiments with cold atoms and ions [65, 66, 67, 68, 69,
70, 71]. The aim of this work is to determine whether we can observe
unambiguous signatures of the MBL phase in the time evolution of disordered
many-body systems. To that end we perform extensive numerical simulations of
disordered XXZ spin-1/2 chain and concentrate on the time evolution of density
correlation functions.
Let us note that we, on purpose, limit our discussion to short-ranged
interactions although MBL has been addressed also for long-range (e.g. dipolar
[72, 73, 74, 75, 76], Ising-type [77, 78, 79, 80, 81] or cavity-mediated [82,
83]) interactions. Similarly we do not address the existence and properties of
localization in disorder-free potentials (such as e.g. tilted lattices) - the
subject of intensive recent studies [84, 85, 86, 87, 88, 89, 90, 91, 92, 93,
94, 95, 96, 97]. We want to concentrate on the “pure”, traditional MBL case.
The paper is structured as follows. In Sec. II we introduce the XXZ spin
chain. We provide results for small system sizes and formulate tentative
criteria for observation of MBL phase in Sec. III. Then, we verify whether
those criteria are fulfilled by dynamics of the XXZ spin chain in the regime
of large disorder strengths and system sizes in Sec. V. Subsequently, we
investigate time evolution of entanglement entropy in that regime in Sec. VI.
Finally, instead of random disorder we consider time dynamics of the system
with a quasiperiodic potential in Sec. VII. We draw our conclusions in Sec.
VIII.
## II Model and observables
In this work we concentrate on 1D XXZ spin chain with Hamiltonian given by
$H=J\sum_{i=1}^{{L-1}}\left(S^{x}_{i}S^{x}_{i+1}+S^{y}_{i}S^{y}_{i+1}+\Delta
S^{z}_{i}S^{z}_{i+1}\right)+\sum_{i=1}^{L}h_{i}S^{z}_{i}$ (1)
where $\vec{S}_{i}$ are spin-1/2 matrices, $J=1$ is fixed as the energy unit,
open boundary conditions are assumed and $h_{i}\in[-W,W]$ are independent,
uniformly distributed random variables. The Jordan-Wigner transformation
allows to map XXZ spin chain (1), to a system of interacting spinless
fermions, with the tunneling matrix element equal to $J$ and nearest-neighbor
interaction strength $\Delta$. This allows to make connection between
disordered XXZ model and optical lattice experiments (as e.g. in [65]). The
random-field XXZ spin chain has been widely studied in the MBL context, see
e.g. [98, 22, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108]. Various
estimates of the critical disorder strength $W_{C}$ for the transition to MBL
phase include: $W_{C}\approx 3.7$ [22], $W_{C}\approx 3.8$ [109],
$W_{C}\approx 4.2$ [110, 60], $W_{C}\gtrapprox 5$ [59, 111], $W_{C}\approx
5.4$ [50].
Besides the random disorder $h_{i}\in[-W,W]$, we also consider the case of
quasiperiodic (QP) potential, for which $h_{j}=W^{\mathrm{QP}}\cos(2\pi
kj+\phi)$, where $k=(\sqrt{5}-1)/2$ and $\phi$ is a random phase taken from
the uniform distribution between $[0,2\pi]$. The QP potential breaks the
translation invariance of the system playing a role similar to disorder and
leading to MBL at a critical strong amplitude of the QP potential
$W^{\mathrm{QP}}_{C}$, with various estimates ranging from
$W^{\mathrm{QP}}_{C}\approx 1.5$ [33, 112, 113, 114, 115, 116] through
$W^{\mathrm{QP}}_{C}\approx 2.4$ [117, 118], up to $W^{\mathrm{QP}}_{C}\approx
4$ [119]. Its important to note that the properties of the transition to MBL
phase in QP systems are distinct from the transition in system with random
disorder [34, 120, 121, 122].
We analyze dynamics of imbalance
$I(t)=D\sum_{i=1+l_{0}}^{L-l_{0}}\langle\psi(t)|S^{z}_{i}|\psi(t)\rangle\langle\psi|S^{z}_{i}|\psi\rangle,$
(2)
where $|\psi(t)\rangle=e^{-iHt}|\psi\rangle$, $\ket{\psi}$ is the initial
state, the constant $D$ assures that $I(0)=1$, $l_{0}>0$ diminishes the
influence of boundaries (in our calculations we take $l_{0}=2$). The results
are averaged over $n_{\mathrm{real}}$ disorder realizations. As the initial
state we take the Néel state with every second spin pointing up and every
second spin down
$|\psi\rangle=|\uparrow\downarrow\ldots\uparrow\downarrow\rangle$. In the
following Section we also take $\ket{\psi}$ as a product state of eigenstates
of $S^{z}_{i}$ operators with average energy $\bra{\psi}H\ket{\psi}$ being in
the middle $10\%$ of the spectrum of $H$ – we refer to such a choice as to a
density correlation function $C(t)$.
We note that other observables, see e.g. [102, 123], suffer from finite size
and finite time limitations similar as (2). Hence, it seems that their
behavior is always governed by the broad distributions of relaxation time
scales [124], and that is why we concentrate on the very simple observable
given by (2), which has another advantage of being directly accessible in
experiments with cold atoms [65].
To find the time evolved state $\ket{\psi(t)}$ we employ Chebyshev expansion
of the evolution operator $e^{-iHt}$ [125], which allows us to investigate
time evolution of systems of $L\leq 20$ sites up to the Heisenberg time
$t_{H}=2\pi/\overline{s}\sim e^{cL}$ (where $\overline{s}$ is the average
level spacing in the middle of the spectrum and $c$ determines the scaling of
Hilbert space dimension with system size: for spin-1/s chains $c=\ln 2$). For
larger system sizes $L=50,100,200$ we use a TDVP algorithm, with bond
dimension $\chi$, specified later in the text for each $W$ and $L$ considered.
In the latter case we focus on relatively large disorder strengths $W\geq 8$
which allows us to investigate time evolution up to a few thousand tunneling
times $J^{-1}$.
## III How to observe an MBL phase?
Numerical [126] as well as experimental [67] investigations of the imbalance
$I(t)$ indicate a presence of a wide regime of disorder strength $W$ in which
the imbalance decays according to a power-law $I(t)\sim
t^{-\overline{\beta}}$. As a criterion for a transition to MBL, the work [59]
introduced the condition that $\overline{\beta}$ vanishing within error bars
implies the onset of MBL. The problem with such a criterion is that the error
bars on $\overline{\beta}$ can be significantly reduced with increasing time
of evolution and number of disorder samples, pushing the tentative boundary of
MBL to larger and larger disorder strengths. An alternative was put forward in
[61], which used a cut-off $\overline{\beta}_{\mathrm{cut}}$ such that
$\overline{\beta}<\overline{\beta}_{\mathrm{cut}}$ implies MBL behavior. The
cut-off value of $\overline{\beta}_{\mathrm{cut}}$ was taken from a comparison
of critical disorder strength estimated from gap ratio statistics as
$W_{C}\approx 4$ for system size $L\approx 20$ and the decay rate of imbalance
at that system size.
The latter criterion also runs into problems. If we assume a simplified model
of the decay of the imbalance, in which $I(t)\sim t^{-\overline{\beta}}$ for
$t<t_{H}$, and then $I(t)=\mathrm{const}$ for $t>t_{H}$ (which is mildly
consistent with data shown in Fig. 1), then the value of the imbalance at
infinite time is: $I(\infty)=I(t_{H})=e^{-cL\overline{\beta}}$. Hence, in
order to have a finite value of imbalance in the $t\rightarrow\infty$ the
exponent governing decay of imbalance should vanish at least as
$\overline{\beta}\sim L^{-1}$. Keeping this in mind we now examine the
dynamics of the density correlation function $C(t)$ in a system of moderate
size $L\leq 20$.
Figure 2: Time evolution of density correlation function $C(t)$ in disordered
XXZ model. Panels a) and b) – $C(t)$ for various system sizes $L=10,...,L=20$
at disorder strengths $W=3,5$, data averaged over $n_{\mathrm{real}}>10^{4}$
disorder realizations. Panels c) and d) – time evolution of the flowing beta
function $\beta(t)$ that locally describes the exponent of decay of $C(t)$.
The red squares denote the Heisenberg time $t_{H}\sim e^{cL}$.
Figure 2a) shows $C(t)$ for disorder $W=3$ for which the XXZ spin chain is in
the ergodic phase. The density correlation function, as well as the imbalance
are characterized by oscillations at small times due to the coupling between
neighboring spins. Those oscillations are gradually damped with time $t$, and
the slow decay becomes the main feature of the dynamics of $I(t)$ and $C(t)$.
With an increasing system size, the power-law decay of $C(t)$ persists to
longer and longer times, not changing much beyond the Heisenberg time $t_{H}$.
The interaction induced decay of $C(t)$ is evidently getting more abrupt with
increasing $L$. The situation is, in fact quite similar for $W=5$ (see Fig.
2b) ), which, according to the majority of estimates (e.g. [22, 35, 110]) is
already in the MBL phase. While the decay of $C(t)$ is much slower than for
$W=3$, it persist to long-times and the saturation value of $C(t)$ is
decreasing with $L$.
To investigate the slow decay of $C(t)$ in more quantitative fashion, we
consider a time-dependent $\beta(t)$ function [117], that is obtained from the
fit $C(t_{1})=at_{1}^{-\beta(t)}$ in the interval $t\in[t_{1},1.5t_{1}]$. The
resulting $\beta(t)$ functions are shown in Fig. 2 c),d). For $W=3$ we observe
that at first, the decay of $C(t)$ is well described by a power-law
($\beta(t)$ is constant) and then the decay gradually slows down, stopping at
the time scale approximately order of magnitude larger than $t_{H}$. For
$W=5$, the slow down of the decay of $C(t)$ occurs at smaller times, however,
a non-vanishing $\beta(t)$ up to Heisenberg time signals further, non-
negligible decay of the density correlation function.
Results presented in this section show that the correlation functions decay up
to Heisenberg time or even longer. Moreover, comparison of results for $W=3$
and $W=5$ indicates that it is hard to propose an accurate phenomenological
model for the decay of $C(t)$. Nevertheless, building on intuitions obtained
in this section, we conclude that an unambiguous observation of MBL phase
should include at least one of the two conditions:
1. (A)
the value of the exponent $\overline{\beta}$ that is decreasing with system
size as $L^{-1}$ \- in such a case even if the power-law decay persists up to
the Heisenberg time, the imbalance is non-vanishing in the limit $t\to\infty$;
2. (B)
a decrease of value of $\beta(t)$ with time $t$ that occurs in a system size
independent fashion indicating the saturation of the imbalance at all
experimentally accessible times beyond a certain time scale.
The results for small system sizes indicate that if the dynamics of the
imbalance satisfies either the criterion A or B, the system is in an
asymptotic MBL phase. In that sense, the conditions A and B can be thought of
as conditions sufficient for the observation of MBL phase. The conditions A
and B must be verified with care and their fulfillment is not in a strict
sense a proof for a stable MBL phase: a system that satisfies either of them
could still be ergodic. For instance, one may imagine a decrease of $\beta(t)$
in time in a system size independent fashion below a certain (large from the
experimental perspective) time scale combined with an onset of a fast decay of
imbalance beyond a certain larger time scale. Nevertheless, such scenarios
seem to be ruled out by the results for small system sizes and for that reason
we treat the conditions A and B as sufficient for observation of MBL phase. At
the same time, we would like to note that neither of the conditions is a
necessary criterion for an observation of MBL phase. Other scenarios in which
the system breaks ergodicity can be envisioned. For instance, the imbalance
may behave in a non-monotonous in time manner with a non-zero infinite time
average in the large system size limit, disallowing the analysis of $I(t)$
with a power-law decay.
With those remarks in mind, we now turn to an analysis of time dynamics of
large systems in the strong-disorder, long-time regime, which seems to be the
most suitable one to find signatures of the MBL phase. The criteria A and B
will be the guiding principles of our analysis. First, however, let us briefly
consider a non-interacting system.
Figure 3: Comparison of the time evolution for noninteracting system between exact propagation and TDVP approximate algorithm ($L/2$ fermions for the system size $L=50$ at disorder strength $W=10$). Top: the imbalance, $I(t)$ (left) and the entanglement entropy in the middle of the chains, $S(t)$ (right) obtained in exact propagation (blue curves extending to larger times) and via TDVP (lighter, orange line). Bottom shows the difference between exact and TDVP results for imbalance (left) and entropy (right). Table 1: Details of numerical simulations for $W=10$ and $\Delta=0$: system size $L=50$, maximal time reached in time evolution $t_{\mathrm{max}}$, the bond dimension $\chi$ (not displayed for the exact numerical calculation), number of disorder realizations $n_{\mathrm{real}}$, and the exponent $\overline{\beta}$ obtained from the fit $I(t)\sim t^{-\overline{\beta}}$ in interval $t\in[100,t_{\mathrm{max}}]$. The error of $\overline{\beta}$ is estimated by resampling over the disorder realizations (here, as well in the rest of this manuscript). | $t_{\mathrm{max}}$ | $\chi$ | $n_{\mathrm{real}}$ | $\overline{\beta}$
---|---|---|---|---
$L$=50 | 1500 | 128 | 1000 | $(0.97\pm 1.12)\cdot 10^{-4}$
$L$=50 | 1500 | - | 1000 | $(0.96\pm 1.12)\cdot 10^{-4}$
$L$=50 | 5000 | - | 1000 | $(-0.23\pm 0.43)\cdot 10^{-4}$
## IV Non-interacting test case
We consider now the Hamiltonian (1) and set $\Delta=0$ which via Jordan-Wigner
transformation maps to a set of non-interacting spinless fermions in a random
on-site potential $h_{i}$. This model is known to be Anderson localized [127]
for an arbitrary amplitude of the disorder $W$. Since the model is non-
interacting, we calculate the time evolution of an initial state $\ket{\psi}$
in numerically exact fashion in time polynomial in system size (see Appendix
A.3). As the initial state we take the Néel state
$|\psi\rangle=|\uparrow\downarrow\ldots\uparrow\downarrow\rangle$. The
obtained time evolved imbalance $I(t)$ provides a reference for our
approximate time propagation using TDVP. Note that while in the non-
interacting case obtaining the exact solutions for arbitrary disorder
realization is a straightforward task, this is not so for TDVP – in the latter
case the algorithm keeps track of a matrix product state that belongs to the
full many-body Hilbert space in a manner similar to the interacting case.
The TDVP algorithm used is described in detail in the Appendix A. The
convergence of TDVP crucially relies on a value of the bond dimension $\chi$.
The time evolved state $\ket{\psi_{\chi}(t)}$ obtained with TDVP becomes a
better and better approximation of the exact time evolved state
$\ket{\psi(t)}$ as $\chi$ increases. However, the simulation cost increases
with the value of the bond dimension as $\chi^{3}$. Hence, one has to choose
the value of $\chi$ such that the observables of interest are converged with
the bond dimension, i.e. do not change with increase of $\chi$ so that one can
safely assume that their value approximates well the value in the exact time
evolved state $\ket{\psi(t)}$. For the interacting model (1) we present
details on the convergence of results with the bond dimension $\chi$ in
Appendix A. In the remainder of this section we compare the exact solution
$\ket{\psi(t)}$ for the non-interacting case ($\Delta=0$) with the time
evolved state obtained with TDVP.
For our test we take disorder amplitude $W=10$ and propagate the Néel state up
to time $t_{\mathrm{max}}=1500$ for 1000 disorder realizations using TDVP with
bond dimension $\chi=128$. Fig. 3a) compares the obtained imbalance $I(t)$
with the result of exact numerical solution for non-interacting model. The
exact solution and TDVP result agree very well up to $t_{\mathrm{max}}=1500$
reach in TDVP simulation. The exact imbalance typically exceeds the TDVP
result, the difference, shown in Fig. 3c) , grows in time and saturates around
$t=800$ at $2\cdot 10^{-6}$. The TDVP slightly underestimates the imbalance in
agreement with the findings of [60]. Nevertheless, both TDVP as well as the
exact results show that the exponent $\overline{\beta}$ governing the decay of
the imbalance is vanishing within the estimated error bars as shown in Tab. 1.
The vanishing $\overline{\beta}$ fulfills trivially the criterion A for
observation of localization.
We also calculated entanglement entropy $S(t)$ for a bipartition of the
lattice into subsystems $A$ and $B$ of length $L/2$:
$S(t)=-\mathrm{Tr}_{A}[\rho_{A}\ln\rho_{A}],$ (3)
where $\rho_{A}=\mathrm{Tr}_{B}\ket{\psi(t)}\bra{\psi(t)}$, $\mathrm{Tr}_{C}$
denotes trace with respect to degrees of freedom of subsystem $C$ and
$\ket{\psi(t)}$ is the state of the system. The entanglement entropy $S(t)$ is
shown in Fig. 3 b). We observe that after an initial increase, the entropy
oscillates around a constant value - similarly to the imbalance $I(t)$. As
Fig. 3 d)shows, TDVP slightly overestimates the entanglement entropy. The
ratio between the error of TDVP simulation and the value of the observable is
roughly two orders of magnitude larger than for the imbalance. Nevertheless,
the results from TDVP and the numerically exact simulation practically overlap
showing that TDVP provides a reliable information about the entanglement
entropy growth.
Encouraged by this comparison we shift towards the interacting case for which
a comparison with the exact dynamics is not possible. There, we necessarily
rely on self-consistency tests of our simulations described in Appendix A.
## V Time evolution of imbalance at strong disorder in large systems
Figure 4: Time evolution of imbalance $I(t)$ for systems of size
$L=50,100,200$ at disorder strength $W=8$, details of the simulations and fits
are given in Tab. 2. Top: the shaded lines denote $I(t)$ whereas the solid
lines denote a running overage of $I(t)$ over window $(t-25,t+25)$, dashed
lines denote power-law fits $I(t)\sim t^{-\overline{\beta}}$ in time interval
$t\in[100,1500]$. Bottom: the running beta function $\beta(t)$, dashed lines
show the value of $\overline{\beta}$, the error of $\beta(t)$ is estimated by
resampling over the disorder realizations (here, as well in the rest of this
manuscript).
Taking into account the various estimates of the critical disorder strength
$W_{C}$ for transition to MBL phase, discussed in Sec. II, we fix the disorder
amplitude at $W=8$ and $W=10$. Such disorder strengths, according to the
aforementioned estimates of $W_{C}$, are expected to lay significantly above
the transition to the MBL phase.
Table 2: Details of numerical simulations for $W=8$: system size $L$, maximal time reached in time evolution $t_{\mathrm{max}}$, the bond dimension $\chi$, number of disorder realizations $n_{\mathrm{real}}$, and the exponent $\overline{\beta}$ obtained from the fit $I(t)\sim t^{-\overline{\beta}}$ in interval $t\in[100,1500]$. The error of $\overline{\beta}$ is estimated by resampling over the disorder realizations (here, as well in the rest of this manuscript). | $t_{\mathrm{max}}$ | $\chi$ | $n_{\mathrm{real}}$ | $\overline{\beta}$
---|---|---|---|---
$L$=50 | 1500 | 128 | 4000 | $(10.03\pm 1.23)\cdot 10^{-4}$
$L$=100 | 1500 | 128 | 2000 | $(11.07\pm 0.97)\cdot 10^{-4}$
$L$=200 | 1500 | 160 | 1000 | $(11.03\pm 0.81)\cdot 10^{-4}$
The evolution of imbalance $I(t)$ for $W=8$ is shown in Fig. 4 whereas the
details of numerical simulations are shown in Tab. 2. After an initial
transient decay and oscillations that last up to $t\approx 100$, we observe a
slow but steady monotonic decrease of $I(t)$ that persists up to the largest
time $t_{\mathrm{max}}=1500$ reached in the simulation. The value of
$t_{\mathrm{max}}$ is not sufficiently large to unambiguously pin-point the
functional form of the decay of $I(t)$. Nevertheless, we observe that the
imbalance is well fitted by a power-law decay $I(t)\sim t^{-\overline{\beta}}$
in the interval $t\in[100,1500]$. The values of the exponent
$\overline{\beta}$, shown in Tab. 2, are positive confirming that the slow
decay of $I(t)$ is present (for a discussion of the stability of the value of
$\overline{\beta}$ with respect to the choice of the fitting interval see
Appendix. B). Moreover, within the estimated error bars, the values of
$\overline{\beta}$ are the same for system sizes $L=50,100,200$, indicating
clearly that the condition A for the observation of MBL phase is not met at
$W=8$.
To check whether the condition B is fulfilled, we consider the flowing beta
function $\beta(t)$ obtained from fitting $I(t_{1})=at_{1}^{-\beta(t)}$ in the
interval $t\in[t_{1},1.5t_{1}]$. The result, shown in the bottom panel of Fig.
4, indicates that the decay of the imbalance slows down considerably for
$t{\approx}150$. However, beyond that time the value of the $\beta(t)$
oscillates around the exponent $\overline{\beta}$. Therefore, we see no traces
of slowing-down of the decay of imbalance at $W=8$.
In conclusion, for $W=8$, neither the criterion A nor B is fulfilled. Hence,
we proceed to repeat our analysis for larger disorder strength $W=10$.
Figure 5: Time evolution of imbalance for $W=10$, denotation the same as in
Fig. 4. Details of the simulations and fits given in Tab. 3.
Time evolution of the imbalance $I(t)$, as well as the flowing $\beta(t)$
function are shown in Fig. 5.
Table 3: Details of numerical simulations for $W=10$, denotations the same as in Tab. 2. | $t_{\mathrm{max}}$ | $\chi$ | $n_{\mathrm{real}}$ | $\overline{\beta}$
---|---|---|---|---
$L$=50 | 1500 | 128 | 4000 | $(3.93\pm 0.82)\cdot 10^{-4}$
$L$=100 | 1500 | 128 | 2000 | $(3.60\pm 0.53)\cdot 10^{-4}$
$L$=200 | 1200 | 160 | 1000 | $(3.50\pm 0.87)\cdot 10^{-4}$
$L$=50 | 5000 | 192 | 2000 | $(3.08\pm 0.51)\cdot 10^{-4}$
While the decay of imbalance clearly slowed down considerably, as reflected by
the values of the exponent $\overline{\beta}$ shown in Tab. 3, upon the
increase of disorder strength from $W=8$ to $W=10$, the system size dependence
of $\overline{\beta}$ remains the same: the values of $\overline{\beta}$ are,
within the estimated error bars, similar for $L=50,100,200$, clearly not
satisfying the criterion A. The flowing $\beta(t)$ function, shown in the
bottom panel of Fig. 5 indicates that the decay of imbalance is relatively
fast around $t\approx 200$ and then slows down considerably at $t\approx 500$
for which the value of $\beta(t)$ is vanishing. However, around $t\approx 800$
the flowing $\beta(t)$ function acquires again the value similar to
$\overline{\beta}$ and the decay of imbalance persists and the criterion B is
not met.
To make sure that our conclusions for $W=10$ are valid, we increased the
maximal time reached in our simulations to $t_{\mathrm{max}}=5000$ for system
size $L=50$, the results are presented in Fig. 6. We indeed observe that the
slow decay of imbalance $I(t)$ persists up to the longest time achieved in our
simulation. This is exemplified by the power-law fit $I(t)\sim
t^{-\overline{\beta}}$ that accurately matches the decay of imbalance in the
whole interval $t\in[100,5000]$, with the exponent $\overline{\beta}$ close to
the values obtained for the shorter time intervals, see Tab. 3. Moreover, the
flowing $\beta(t)$ function oscillates around the value $\overline{\beta}$ in
the whole interval of available times. We see no signs of the slow down of
decay of $I(t)$, which leads us to conclude that the criterion B is not
fulfilled for $W=10$.
Figure 6: Time evolution of imbalance for $W=10$ in extended time interval,
denotation the same as in Fig. 4. Details of the simulations and fits given in
Tab. 3.
In conclusion, we found no clear signatures of the MBL phase in results
presented in this sections, even though we considered significantly larger
times and disorder strengths than in earlier studies [59, 60]. One immediate
question is whether we can go even further in the attempts to observe MBL
phase and consider larger disorder strength $W$ and bigger maximal time
$t_{\mathrm{max}}$. The factor that limits such a continuation most severely
is the slow-down of decay of $I(t)$ with $W$. In order to observe in a
statistically significant way a decay of $I(t)$ at larger $W$ the increase of
$t_{\mathrm{max}}$ should be coupled with an increase of the number of
disorder realizations $n_{\mathrm{real}}$. This considerably increases the
resources needed for such numerical simulations. The same considerations apply
to experiments with quantum many-body systems which are limited by a finite
coherence time (typically limited to at most 1000 tunneling times [96] thus
shorter than the times considered by us) as well as resources needed to
perform disorder averages.
## VI Time evolution of entanglement entropy
The time dependence of the entanglement entropy is one of the tools that may
be used to identify the existence of MBL phase. While typically in the
deconfined systems the entanglement entropy grows linearly in time when the
evolution is started from the low entanglement, e.g. separable state, in MBL
one expects a logarithmic entanglement entropy growth [128, 129]. It is,
therefore, instructive to study the entropy growth in our case in the regime
of large disorder strengths and long times probed in our numerical
simulations. Since the Hamiltonian (1) conserves the total magnetization
$\sum_{i=1}^{L}S^{z}_{i}$, the entanglement entropy $S$ of subsystem $A$
consisting of lattice sites $1,\ldots L/2$ can be written as a sum
$S(t)=S_{n}(t)+S_{c}(t)$, where $S_{n}(t)$ is the number entropy and $S_{c}$
denotes the configurational entropy [130, 131, 132, 133, 69, 82, 97]. The
number entropy is given by
$S_{n}(t)=-\sum_{n}p(n)\ln p(n),$ (4)
where $p(n)$ is the probability that $\sum_{i=1}^{L/2}S^{z}_{i}$ is equal to
$n$. (We note that $\sum_{i=1}^{L/2}S^{z}_{i}$ is proportional to the total
number of spinless fermions in subsystem $A$ after Jordan-Wigner
transformation of (1), explaining the term “number entropy”.) The
configurational entropy is given by
$S_{c}(t)=-\sum_{n}p(n)\mathrm{Tr}[\rho(n)\ln\rho(n)],$ (5)
where $\rho(n)$ is the block of the reduced density matrix in sector with
$\sum_{i=1}^{L/2}S^{z}_{i}=n$.
Figure 7: Time evolution of entanglement entropy for $L=50$ and $W=10$. Top:
configuration entanglement entropy $S_{c}(t)$ denoted by solid line, dashed
lines denote power-law and logarithmic fits $f(t)$. The inset shows the
residual $f(t)-S_{c}(t)$. Bottom: The number entanglement entropy $S_{n}(t)$
is denoted by solid line, dashed line denotes a double-logarithmic fit
$f_{2}(t)$. The inset shows the residual $f_{2}(t)-S_{n}(t)$.
Our results for the entanglement entropies $S_{n}(t)$ and $S_{c}(t)$ are shown
in Fig. 7. The configurational entropy $S_{c}(t)$ is expected to grow
logarithmically in time [128, 129] in the MBL regime. We observe that after an
initial transient at times $t\lessapprox 10$, the growth of $S_{c}(t)$ is well
described by a power-law $S_{c}(t)\propto t^{\gamma}$ with $\gamma=0.250(2)$
in the interval $t\in[10,600]$. This behavior resembles the time dynamics of
entanglement entropy observed in the ergodic regime at moderate values of
disorder $W\approx 2.5$ [126]. However, at longer times, the increase of
$S_{c}(t)$ slows down and is well fitted by $S_{c}(t)=a+b\ln t$ with
$a=-0.04437(7)$ and $b=0.02001(9)$ for $t\in[400,5000]$ in agreement with
expectations for the MBL regime. The growth of the number entropy is
significantly slower, and is very well fitted by a double logarithmic formula
$S_{n}(t)=a+b\ln\ln t$ with $a=0.1496(6)$ and $b=0.0120(3)$ in a wide regime
of times $t\in[20,5000]$. This confirms the prediction of [40, 44] for the
significantly larger system size and disorder strength than tested before.
In conclusion, the slow decay of imbalance observed in Sec. V is accompanied
by a logarithmic increase of the configurational entanglement entropy
$S_{c}(t)$ and a double logarithmic growth of the number entropy $S_{n}(t)$.
Those quantities provide a complementary to the imbalance insight into the
dynamics of the slow delocalization of the system. At the same time, they do
not allow for an observation of the MBL phase in the fashion similar to the
imbalance. For a localized system, one expects a saturation of $S_{n}(t)$
[41]. The upper limit, $S_{n}=\ln(3){\approx 1.01}$, predicted in [41] is much
higher than the values reached by a very slow double logarithmic growth of
$S_{n}(t)$ observed in Fig. 7. Note also that a very recent study, [134],
instead of such a a slow double logarithmic growth predicts a power-law
approach of ${S}_{n}$ to its asymptotic value at $t\to\infty$. This cannot be
tested for the large system sizes $(L\geq 50)$ considered by us since we are
unable to determine the asymptotic value of $\lim_{t\to\infty}S_{n}(t)$.
## VII Quasiperiodic systems
In this section we attempt at observation of MBL phase in dynamics of the
system with QP potential, defined in Sec. II. To that end we investigate the
impact of the amplitude of QP potential $W^{\mathrm{QP}}$ on time evolution of
imbalance $I(t)$.
Figure 8: Time evolution of imbalance $I(t)$ for QP potential. Top: results
for the amplitude of QP potential $W^{\mathrm{QP}}=2$ the shaded lines denote
$I(t)$ whereas the solid lines denote a running overage of $I(t)$ over window
$(t-25,t+25)$, dashed lines denote power-law fits $I(t)\sim
t^{-\overline{\beta}}$ in time interval $t\in[500,5000]$. Bottom: the same for
$W^{\mathrm{QP}}=3$. Details of simulations are given in Tab. 4
The results for $W^{\mathrm{QP}}=2,3$ are shown in Fig. 8. The behavior of
$I(t)$ is qualitatively similar to the systems with random disorder: after an
initial transient, the decay of imbalance is well fitted by a power-law
$I(t)\sim t^{-\overline{\beta}}$. The exponent $\overline{\beta}$ is clearly
increasing with system size both for $W^{\mathrm{QP}}=2$ and
$W^{\mathrm{QP}}=3$, as shown in Tab. 4, suggesting that the system
delocalizes in the thermodynamic limit at those values of $W^{\mathrm{QP}}$
and neither the criterion A nor B for observation of MBL phase is met.
Table 4: Details of numerical simulations QP potential, denotations the same as in Tab. 2. The bond dimension $\chi$ is not displayed for calculation performed with the Chebyshev expansion of the evolution operator. | $t_{\mathrm{max}}$ | $\chi$ | $n_{\mathrm{real}}$ | $\overline{\beta}$
---|---|---|---|---
$L$=12, $W^{\mathrm{QP}}$=2 | 5000 | - | $10^{6}$ | $(1.8\pm 0.2)\cdot 10^{-3}$
$L$=16, $W^{\mathrm{QP}}$=2 | 5000 | - | $10^{5}$ | $(9.5\pm 0.2)\cdot 10^{-3}$
$L$=20, $W^{\mathrm{QP}}$=2 | 5000 | - | $5\cdot 10^{4}$ | $(19.0\pm 0.1)\cdot 10^{-3}$
$L$=12, $W^{\mathrm{QP}}$=3 | 5000 | - | $10^{6}$ | $(3.3\pm 0.4)\cdot 10^{-4}$
$L$=16, $W^{\mathrm{QP}}$=3 | 5000 | - | $10^{5}$ | $(8.8\pm 0.3)\cdot 10^{-4}$
$L$=20, $W^{\mathrm{QP}}$=3 | 5000 | - | $5\cdot 10^{4}$ | $(8.9\pm 0.6)\cdot 10^{-4}$
$L$=12, $W^{\mathrm{QP}}$=4 | 5000 | - | $10^{6}$ | $(2.1\pm 0.4)\cdot 10^{-4}$
$L$=16, $W^{\mathrm{QP}}$=4 | 5000 | - | $10^{5}$ | $(2.8\pm 0.3)\cdot 10^{-4}$
$L$=50, $W^{\mathrm{QP}}$=4 | 4000 | 128 | | $(3.0\pm 1.3)\cdot 10^{-4}$
$L$=12, $W^{\mathrm{QP}}$=5 | 10000 | - | $10^{6}$ | $(0.3\pm 0.7)\cdot 10^{-4}$
$L$=16, $W^{\mathrm{QP}}$=5 | 10000 | - | $10^{5}$ | $(1.1\pm 0.8)\cdot 10^{-4}$
$L$=50, $W^{\mathrm{QP}}$=5 | 4500 | 128 | 2000 | -
$L$=100, $W^{\mathrm{QP}}$=5 | 3000 | 128 | 1000 | -
$L$=200, $W^{\mathrm{QP}}$=5 | 2500 | 128 | 600 | -
Figure 9: Time evolution of imbalance $I(t)$ for QP potential for
$W^{\mathrm{QP}}=4,5$, denotation the same as in Fig. 8. Details of the
simulations and fits given in Tab. 4.
The decay of imbalance $I(t)$ slows down considerably when the amplitude of
the QP potential is increased to $W^{\mathrm{QP}}=4$ as shown in Fig. 9. The
exponents $\overline{\beta}$ governing the power-law decay of imbalance for
$W^{\mathrm{QP}}=4$ are comparable to the exponents obtained for $W=10$ for
the random disorder. However, the behavior of the running averages of $I(t)$
(shown by the solid lines in Fig. 9) is different: we observe significant
oscillations around the fitted power-law decay. The pattern of those
oscillations is not stable with increasing the system size, $L$.
This behavior changes qualitatively for $W^{\mathrm{QP}}=5$. For this
amplitude of the QP potential we observe an emergence of a pattern of
oscillations of $I(t)$ at times $t\gtrapprox 200$ that remains the same when
the system size is increased from $L=12$ to $L=200$. This is the first case
for which we observe that the increase of the system size does not enhance its
delocalization. Instead, this result shows that the dynamics of a small system
comprised of $L=12$ sites is reproduced in the bulk of the large system of
$L=200$ sites. Such a behavior suggests that the system remains MBL in the
thermodynamic limit at $W^{\mathrm{QP}}=5$, although our approach is
inherently limited to dynamics at finite times and cannot give a definite
answer about the fate of the system at $t\rightarrow\infty$.
Two remarks are in order. Firstly, the values of the running average of $I(t)$
are not changing monotonically with $L$: the curve for $L=16$ is on the top
whereas that for $L=50$ on the bottom. This is caused by the statistical
fluctuations associated with the finite number of disorder realizations
$n_{\mathrm{real}}$ as well as by the erratic changes of $2\pi kL$ modulo
$2\pi$ with $L$ that determine the number of full periods of the QP potential
in the whole chain. Secondly, the emergent pattern of oscillations of $I(t)$
prevents us from determining whether the imbalance $I(t)$ slowly decays in
time. Performing a power-law fit in the interval $t\in[1000,10000]$ we have
found non-vanishing values of $\overline{\beta}$ as shown in Tab. 4. However,
$\overline{\beta}$ changes significantly when the interval in which the fit is
performed changes. This shows that the criteria A and B are effectively
inapplicable to the dynamics of imbalance in QP potential.
We refer the reader to Appendix C for further numerical studies of the
imbalance in QP potential where we show that the persistent oscillations tend
to decay for even larger values of the amplitude $W^{QP}$. We also show there
that the character of the oscillations depends on the parameter $k$ which
determines the quasiperiodicity of the potential by simulating the dynamics
also for $k=\sqrt{2}/2$.
Figure 10: Time evolution of entanglement entropy for QP potential with
$W^{QP}=5$. Top: configuration entanglement entropy $S_{c}(t)$ for system
sizes $L=16,20,50,200$. The inset shows the same but on log-log scale. The
dashed lines divide the time into intervals I, II, III (see text). Bottom: The
corresponding number entanglement entropy $S_{n}(t)$.
To explore the dynamics in the QP potential from a different perspective, we
calculate the number, $S_{n}(t)$, and the configurational, $S_{c}(t)$,
entropies for $W^{QP}=5$. The results are shown in Fig. 10. Rather than
observing an anticipated monotonic increase of the entanglement entropies, we
may distinguish three time intervals (A,B,C) in the time dependence of
$S_{c,n}(t)$. In the interval $I$, for $t\lesssim 4500$, we observe an
algebraic in time increase of $S_{c}(t)$ (compare the inset in the top panel
in Fig. 10 drawn in the log-log scale). This behavior is accompanied by a slow
increase of the number entropy $S_{n}(t)$ which initially follows a
logarithmic growth, saturates around $t\approx 300$ and then again seems to
follow a logarithmic growth. In the interval I, the results for the small
$L=16,20$ and large system sizes $L=50,200$ practically overlap both for
$S_{c}(t)$ and $S_{n}(t)$. This is another property suggesting the locality of
the dynamics at $W^{QP}=5$. We observe for $L=16,20$ that the behavior of
$S_{c}(t)$ changes qualitatively at larger times: $S_{c}(t)$ is approximately
constant in the region II ($4500<t<20000$) and grows logarithmically in time
in the region III ($t>20000$). Both for $S_{c}(t)$ as well as for $S_{n}(t)$
the results for $L=16$ and $L=20$ are practically overlapping in the time
intervals II and III. Unfortunately, the regimes II and III are inaccessible
in TDVP calculations for large system sizes. This prevents us from deciding
whether the initial power-law growth of $S_{c}(t)$ is continued in the large
time limit for large system sizes (leading to a slow approach towards
ergodicity) or whether the features of the entanglement growth at $L=20$ are
consistent with the behavior system for $L\rightarrow\infty$ (leading to a
stable MBL phase).
## VIII Conclusions
In this work we have addressed the problem of a possible experimental
observation of MBL. The presence of interactions gives rise to a slow dynamics
towards equilibrium in strongly disordered systems. This leads us to argue
that an observation of even a very slow decay of correlation functions in a
finite interval of time is insufficient to claim an unambiguous observation of
MBL.
For relatively small systems comprising of less than $L=20$ lattice sites, we
calculated time dynamics beyond the Heisenberg time which allowed us to
extrapolate the results to the infinite time limit. Building on intuitions
obtained in that way, we formulated the criteria A and B for an observation of
the MBL phase. The criterion A requires a slowdown of the decay of density
correlation functions as $L^{-1}$ when the system size $L$ is increased. The
criterion B demands a saturation of correlation functions beyond a certain
time scale in a system size independent manner. We would like to emphasize
that these criteria are neither sufficient nor necessary conditions to prove
that a system is MBL. Rather, we perceive the criteria A and B as hints of
whether the dynamics of a given system breaks the ergodicity or not.
Performing large scale tensor network simulations of time evolution of
disordered XXZ spin chain of up to $L=200$ sites we did not find a regime of
parameters in which the criterion A or B for observation of MBL would be
satisfied. For considered disorder strengths we always encountered the slow
but persistent decay of imbalance hinting at a slow approach of the system
towards the eventually delocalized future. This conclusion was obtained even
though we focused on the regime of disorder strengths lying significantly
above the current estimates of the critical disorder strength for transition
to MBL phase and pushed the maximal time reached in our simulations to few
thousands tunneling times. In that respect, our results are consistent with
the nonexistence of MBL phase in the thermodynamic limit, see also [36, 46].
We also revisited the dynamics of the entanglement entropy confirming the
logarithmic growth of its configurational part and the double logarithmic
increase of the number entropy in the regime of long times and large system
sizes confirming predictions of [40, 44].
Finally, we investigated the time evolution of QP systems. The dynamics of
quasiperiodic system is very much similar to random system at intermediate
values of the amplitude $W^{\mathrm{QP}}$, with a slow, power-law like decay
of imbalance. However, for a stronger QP potential, at $W^{\mathrm{QP}}=5$, we
demonstrated an emergence of a pattern of oscillations in the imbalance
$I(t)$. This pattern remains stable with the increase of the system size. This
qualitatively different behavior of the imbalance in a striking fashion shows
that the dynamics of QP systems at sufficiently large potential strengths
becomes local. While we were eventually not able to fully exclude the decay of
the imbalance in the infinite time limit, the result for QP systems appears to
be not far from being sufficient to claim an observation of MBL phase. In any
case, our results show that the asymptotic properties of transition to MBL
phase may be probed more easily in QP systems (see [122] for the analysis of
QP system from the spectral perspective).
We would like to stress that our results, especially for disordered systems,
do not exclude the existence of a stable MBL phase. Rather, they provide lower
bounds on time scales and disorder strengths required to observe the freezing
of system dynamics in the long time limit that defines the MBL phase. Those
lower bounds are relevant both for future numerical simulations of disordered
systems as well as for experiments with quantum simulators.
###### Acknowledgements.
This work would not be possible without the help of Titas Chanda who provided
us with his tensor network codes and generously helped with their
implementation. Thank you, Titas! We are also indebted to Anatoly Polkovnikov
and Dries Sels for discussions as well as to Elmer V. H. Doggen for questions
regarding our error analysis. The numerical computations have been possible
thanks to the support of PL-Grid Infrastructure. The TDVP simulations have
been performed using ITensor library (https://itensor.org). This research has
been supported by National Science Centre (Poland) under project
2019/35/B/ST2/00034 (J.Z.)
## Appendix A Tests on the numerical accuracy of the presented results
The results presented in the main text are obtained using different numerical
techniques that will be described in detail below. We also provide details of
the numerical method for used for the non-interacting system.
There are two types of errors in our results. The first is the statistical
error which arises due to fluctuation of results from one disorder realization
to another at fixed parameters of the system. The resulting errors in the
exponent $\overline{\beta}$ governing the decay of imbalance $I(t)$, as well
as in the running $\beta(t)$ function are estimated by the bootstrap
technique, i.e. by resampling over the disorder realizations. In the figures
we plot the imbalance $I(t)$ as well as a running average of the imbalance.
Importantly, however, in the fits that determine $\beta(t)$ and
$\overline{\beta}$ we always use the full data for the imbalance $I(t)$. The
second type of uncertainties are the systematic errors that might occur when
the numerical simulations are not fully converged. Those systematic errors are
particularly relevant for TDVP results. Below, we describe numerical tests
that confirm that the values $\chi$ used by us in the main text are sufficient
for the results to be converged, i.e. independent of the value of the bond
dimension $\chi$.
### A.1 Chebyshev time propagation
For small system sizes ($L\leq 20$) we use Chebyshev propagation scheme as
described in detail in [125]. In a nutshell, this approach approximates the
time evolution operator $U(\Delta t)=\exp(-iH\Delta t)$ over time period
$\Delta t$ as
$U(\Delta t)\approx\mathrm{e}^{-\mathrm{i}b\Delta t}\left(J_{0}(a\Delta
t)+2\sum_{k=1}^{N}(-i)^{k}J_{k}(a\Delta
t)T_{k}\left(\mathcal{H}\right)\right),$ (6)
where $a=(E_{\rm max}-E_{\rm min})/2$, $b=(E_{\rm max}+E_{\rm min})/2$ and
$E_{\rm min}$ ($E_{\rm max}$) is the energy of the ground state (the highest
excited eigenstate) of the Hamiltonian $H$. The Hamiltonian is rescaled to
$\mathcal{H}=\frac{1}{a}(H-b)$ so that spectrum of $\mathcal{H}$ belongs to
the $[-1,1]$ interval, $J_{k}(t)$ is the Bessel function of the order $k$ and
$T_{k}(x)$ is the Chebyshev polynomial of order $k$. The order of expansion
$N$ needed to assure convergence of the expansion (6) for a given time step
$\Delta t$ is computed in the following way. We take a random normalized state
$\ket{\psi_{R}}$, calculate the state $U(\Delta t)\ket{\psi_{R}}$ with a
certain trial order of expansion $N_{tr}$ and compute its norm. If the norm of
$U(\Delta t)\ket{\psi_{R}}$ deviates from unity by more than $10^{-13}$, we
know that $N_{tr}$ needs to be increased; otherwise $N_{tr}$ is decreased.
This allows us to perform a binary search for $N_{tr}$ in the interval
$N_{tr}\in[5,5000]$ (the upper boundary is determined by the maximal time step
$\Delta t$ and parameters of the model). The result of this binary search,
$N^{0}_{tr}$, is then incremented by $20\%$, yielding the desired order
$N=1.2N^{0}_{tr}$. We calculate the order of expansion whenever the time step
$\Delta t$ changes in our algorithm. To calculate the time evolution of an
initial state $\ket{\psi(0)}$ we repeatedly apply (6) to obtain
$\ket{\psi(\Delta t)}$, $\ket{\psi(2\Delta
t)}$,$\ldots$,$\ket{\psi(t_{\mathrm{max}})}$. We have tested this procedure
for system sizes $L\leq 16$ comparing $\ket{\psi(t_{\mathrm{max}})}$ with
state $\ket{\psi_{ED}(t_{\mathrm{max}})}=U(t_{\mathrm{max}})\ket{\psi(0)}$
evolved using time evolution operator $U(t_{\mathrm{max}})$ determined by
means of the full exact diagonalization of the Hamiltonian $H$. For
$t_{\mathrm{max}}=10^{5}$ we checked that the norm
$||\ket{\psi(t_{\mathrm{max}})}-\ket{\psi_{ED}(t_{\mathrm{max}})}||$ is
smaller than $10^{-10}$ in the whole parameter range considered in this work.
The deviation from unity of the norm of the state propagated with the
Chebyshev expansion: $1-||\ket{\psi(t_{\mathrm{max}})}||$ was smaller than
$10^{-12}$ for all system sizes considered in this work. For $L\leq 16$ the
corresponding deviations in the value of $C(t)$ function (as compared to
$\ket{\psi_{ED}(t_{\mathrm{max}})}$) were smaller than $10^{-13}$. We note
that the main advantage of the Chebyshev expansion is that it efficiently
utilizes the sparse matrix structure of the Hamiltonian of the system. This is
due to the fact that a single time propagation step $U(\Delta t)\ket{\psi(t)}$
reduces to $\mathcal{O}(N)$ matrix-vector products and a calculation of linear
combinations of vectors.
### A.2 Tensor network approaches
Chebyshev propagation scheme is not effective for larger system sizes since it
operates on the quantum states expressed as vectors in the full Hilbert space
that is exponentially large in system size. In contrast, tensor network
techniques parameterize only a fraction of the full Hilbert space, encoding
the state of the system in a matrix product state (MPS). This allows us to
investigate time evolution of systems larger than $L>25$. The tensor network
techniques were developed over the years starting from seminal works of Vidal
[53, 54] and White [135]. The link between the two approaches was illuminated
in [136]. Those schemes are known as time-dependent density matrix
renormalization group techniques (tDMRG) or time evolving block decimation
TEBD techniques. The important modification came with the variational approach
leading to algorithms based on Time Dependent Variational Principle (TDVP)
optimal for an assumed limitation of the Hilbert space [55, 56, 57, 58]. The
time evolution can be calculated effectively with tensor network approaches
only when the bond dimension $\chi$ of the MPS is sufficiently large to encode
the state of the system. This gives rise to an upper limit on the entanglement
entropy in the state of the system for a given $\chi$. This, in turn,
translates into maximal time $t_{\mathrm{max}}$ to which time evolution of the
system can be accurately simulated with TDVP/tDMRG for a given bond dimension
$\chi$. Calculations in our work rely on the fact that for disorder strengths
$W=8-10$ the spreading of entanglement in the system is very slow which allows
us to probe the time evolution at times equal to few thousand tunneling times.
The TDVP algorithm for time evolution consists of two stages. In the first
stage we use the so called 2-site TDVP which allows for an accurate estimation
of the errors. They appear mainly due to the truncation of the Hilbert space
via Schmidt decomposition between the sites. When a disregarded Schmidt weight
exceeds $10^{-12}$ the Hilbert space is enlarged so in this stage the
algorithm is practically exact until the bond dimension reaches the prescribed
value $\chi$ at a given bond. At this stage we switch (at this bond) to 1-site
TDVP algorithm, from this moment errors due to the Hilbert space truncation
start to accumulate. This is a standard, well developed strategy [60, 61]
which we follow in our work.
The detailed comparison of the performance of TEBD and TDVP algorithms for the
random-field XXZ chain, but for lower disorder amplitudes than in the present
work, was performed in in our previous work [60]. It was shown, in particular,
that the TEBD algorithm that is unconverged, i.e. the bond dimension is not
sufficiently large to follow the time evolution of the state up to the
requested time $t_{\mathrm{max}}$, spuriously indicates a stabilization of the
imbalance $I(t)$ suggesting a localization in the system. In contrast,
unconverged TDVP has a tendency to show a delocalization in the system by
overestimating the degree of decay of the imbalance $I(t)$. An analogous
behavior of TDVP was also observed in a different disorder-free models in
[58].
Figure 11: Comparison of the imbalance $I(t)$ (averaged over times
$[t-10,t+10]$) for system size $L=200$ and disorder strength $W=10$ obtained
with TEBD and TDVP algorithms. The bond dimension is fixed as $\chi=128$ and
the results are averaged over $24$ disorder realizations. The inset shows the
difference between the imbalances $I(t)$ for TDVP and TEBD propagation
schemes. Figure 12: Comparison of the imbalance $I(t)$ (averaged over times
$[t-10,t+10]$) for system size $L=50$ and disorder strength $W=8$ obtained
with TEBD and TDVP algorithms. The bond dimension is fixed as $\chi=128$ and
the results are averaged over $1000$ disorder realizations. The dashed line
shows the fitted power-law decay of $I(t)$. The inset shows the difference
between the imbalances $I(t)$ for TDVP and TEBD propagation schemes.
This motivates us to compare results for the imbalance $I(t)$ obtained with
TEBD and TDVP algorithms as shown in Fig. 11. The results are averaged over 24
disorder realizations for $L=200$ and $W=10$. We observe that the agreement
between TDVP and TEBD results is excellent indicating the convergence for
individual disorder realizations. The difference between the curves at late
times oscillates around ${4}\cdot 10^{-5}$. This small discrepancy can be
compared with the total change of the value of imbalance $\Delta I=7\cdot
10^{-4}$ in the interval $t\in[100,1200]$ for $L=200$ (cf. Fig. 5). The latter
value is more than an order of magnitude larger than the discrepancy between
TDVP and TEBD results. This suggests that the exponent of power-law decay
$\overline{\beta}=(3.50\pm 0.87)\cdot 10^{-4}$ for $L=200$ (see Tab. 3) is
accurately estimated.
While TEBD is faster “per time step” for such a large disorder amplitude
($W=10$) we must take very small time step $\Delta t=0.001$ for TEBD to obtain
converged results. The error of the approximate unitary evolution may be
estimated by a relative energy change in TEBD algorithm
$[(\braket{\psi_{TEBD}(t_{\mathrm{max}})}{H}{\psi_{TEBD}(t_{\mathrm{max}})}-E_{0})/E_{0}$
where $\ket{\psi_{TEBD}(t)}$ is the state obtained in TEBD time evolution and
$E_{0}=\braket{\psi(0)}{H}{\psi(0)}$]. It remains below $10^{-4}$ for even the
most unfavorable disorder realisation (for $\chi=128$). At the same time, the
total accumulated error, equal to sum of squares of Schmidt coefficients
disregarded in all time steps, associated with necessary truncations inherent
to TEBD is below $10^{-5}$. As shown in the following, the resulting error is
sufficiently small to obtain an accurate estimate of the exponent
$\overline{\beta}$. The required small step makes, however, the application of
TEBD scheme not practical. For large disorder amplitudes and large time
scales, it is more efficient to use TDVP. It allows us to keep the time step
at a reasonable value, $\Delta t=0.1$ (the time scale is fixed by $J=1$ in
(1)). We have checked by decreasing the time step that the chosen value leads
to accurate results. The agreement of TDVP results with the numerically exact
results for the non-interacting case, shown in Fig. 3 provides another test of
the convergence of our results with the time step.
While the comparison for $L=200$ is carried out for 24 disorder realizations
only, we supplement it with comparison for $W=8$ and $L=50$ carried out for
over a 1000 disorder realizations in Fig. 12. Again the discrepancy between
curves is small (of the order of ${4}\cdot 10^{-5}$ as for $L=200$ data in
Fig. 11. Larger number of disorder realizations allows us to extract reliably
$\beta$ values from both simulations. They agree very well with each other
indicating the agreement between both aogorithms used.
Figure 13: The imbalance $I(t)$ (averaged over times $[t-10,t+10]$) for system
size $L$ and disorder strength $W$ obtained with TDVP algorithm with bond
dimension $\chi$, dashed lines show power-law fits $I(t)\sim
t^{-\overline{\beta}}$ in time interval $t\in[100,t_{\mathrm{max}}]$ . Top
panel: $L=200$, $W=8$, results averaged over 250 disorder realizations,
$t_{\mathrm{max}}=1500$. Center panel: $L=100$, $W=10$, results averaged over
1000 disorder realizations,$t_{\mathrm{max}}=1500$. Bottom panel: $L=200$,
$W=10$, results averaged over 984 disorder realizations,
$t_{\mathrm{max}}=1200$. The insets show the difference between the imbalances
for the larger and the smaller value of $\chi$. Figure 14: Top panel:
Comparison of the imbalance $I(t)$ (averaged over times $[t-10,t+10]$) for
system size $L=50$ and disorder strength $W=10$ obtained with TDVP algorithm
with bond dimension $\chi=[50,90,128,192]$. The results are averaged over
$1000$ disorder realizations. The inset shows the difference $\Delta
I(t)=I_{\chi}(t)-I_{\chi=192}(t)$ between the imbalance obtained with TDVP
with the largest bond dimension $\chi=192$ and the imbalances obtained with
$\chi=128,90,50$. Bottom panel: the same, but data for the smaller bond
dimensions: $\chi=48,64,96,128$ obtained with TEBD algorithm. Note the
difference in the range of the horizontal axes of the two panels. Figure 15:
Top panel: Comparison of the values of the exponent $\overline{\beta}$
governing the decay of the imbalance $I(t)$ for disorder strength $W=10$ and
system size $L=50$ obtained with TDVP propagation scheme. The fitting was
performed in the interval $t\in[100,t_{\mathrm{max}}]$ and results are shown
as a function of $1/\chi$, solid lines denote second order polynomial fits in
$1/\chi$, the red point at $1/\chi=0$ shows the result
$\overline{\beta}=(3.08\pm 0.51)\cdot 10^{-4}$ from Tab. 3. Bottom panel: The
exponent $\overline{\beta}$ as a function of $1/\chi$ for TDVP and TEBD
algorithms (for $L=50$, $W=10$). The values of $\overline{\beta}$ presented in
both are extracted from data shown in Fig. 14.
The above comparison of TEBD and TDVP algorithms suggests a good convergence
of our TDVP results. To further investigate the accuracy of the TDVP scheme,
we compare results obtained for a varying bond dimension $\chi$. Fig. 13
summarizes our results. In each of the investigated cases we observe that the
curves showing the imbalance $I(t)$ practically overlap for the both bond
dimensions considered (cf. the insets in Fig. 13). We observe that the
exponents $\overline{\beta}$ govering the power-law decay of imbalance for
smaller and larger $\chi$ are consistent with each other indicating a good
convergence of the data with the bond dimension. At the same time, we observe
that the $\overline{\beta}$ slightly decreases with the increase of the bond
dimension $\chi$ in each of the analyzed cases.
This dependence is further analyzed in the top panel of Fig. 14 in which we
have supplemented the data for $\chi=128$ and $\chi=192$ with results for
smaller bond dimensions $\chi=50,90$. Interestingly, the agreement of results
for $\chi\geq 50$ up to time $t\approx 1000$ shows that already the results
for $\chi=50$ are a good estimate of the imbalance $I(t)$ in that time
interval at $W=10$. At larger times, the results for $\chi=50$ are unconverged
and show a spurious signatures of delocalization in the system consistently
with our expectations based on [60]. The bottom panel of Fig. 14 compares the
TDVP results for $\chi=192$ with the imbalance obtained with TEBD and bond
dimensions $\chi=48,64,96,128$. Contrary to the expectations from [60], we see
that TEBD also indicates weaker and weaker decay of the imbalance $I(t)$ as
the bond dimension $\chi$ is increased. The difference between the TDVP and
TEBD results for the largest $\chi$ presented is no bigger than $4\cdot
10^{-5}$ indicating that both algorithms yield consistent estimates of
$\overline{\beta}$, $\beta(t)$.
To clarify the dependence of the results on the value of $\chi$, we plot the
values of the exponent $\overline{\beta}$ as function of $1/\chi$ in the top
panel of Fig. 15. The value of the exponent $\overline{\beta}$ decreases
monotonously with the bond dimension $\chi$. The change in the value of
$\beta$ when $\chi$ increases from $50$ to $192$ is the smallest for
$t_{\mathrm{max}}=1500$ (indicating that smaller bond dimensions are needed to
get converged results for $t<1500$), and increases with the increase of
$t_{\mathrm{max}}$. Nevertheless, the extrapolations of $\overline{\beta}$
with a second order polynomial in $1/\chi$ give consistent results for all
considered values of $t_{\mathrm{max}}$. Importantly, those extrapolations are
in agreement with the result $\overline{\beta}=(3.08\pm 0.51)\cdot 10^{-4}$
from Tab. 3, confirming the convergence of our simulations with the bond
dimension. The bottom panel of Fig. 15 shows a comparison of
$\overline{\beta}$ for $t_{\mathrm{max}}=1500$ for TEBD and TDVP results. The
values of $\overline{\beta}$ are nearly independent of $\chi$ for
$\chi\gtrapprox 90$ confirming that both algorithms are very close to being
converged at those bond dimensions for $t<t_{\mathrm{max}}=1500$. Finally, the
extrapolation of those results to large $\chi$ limit yields the consistent
values of $\overline{\beta}$ for both TEBD and TDVP in line with our message
about the persistence of a slow decay of the imbalance even at the large
disorder strength $W=10$.
### A.3 Time evolution for free fermions
Here, for completeness, we provide details of the standard (see [137] and
references therein) approach to time evolution of a system of non-interacting
fermions used by us in Sec. IV. The Hamiltonian (1), upon Jordan-Wigner
transformation, becomes
$\hat{H}=2J\sum_{i=1}^{L-1}\left(\hat{c}^{{\dagger}}_{i}\hat{c}_{i+1}+\hat{c}^{{\dagger}}_{i+1}\hat{c}_{i}+\frac{\Delta}{2}\hat{n}_{i}\hat{n}_{i+1}\right)+\sum_{i=1}^{L}h_{i}\hat{n}_{i},$
(7)
where $\hat{c}^{{\dagger}}_{i}$ ($\hat{c}_{i}$) is creation (anihilation)
operator of spinless fermion at site $i$, canonical anti-commutation relation
$\\{\hat{c}_{i},\hat{c}^{{\dagger}}_{j}\\}=\delta_{ij}$ is fulfilled, and the
number operator is given as $\hat{n}_{i}=\hat{c}^{{\dagger}}_{i}\hat{c}_{i}$.
For $\Delta=0$, the model (7) becomes non-interacting. Then, it can be written
as a quadratic form of the fermionic operators
$\hat{H}=\sum_{i,j=1}^{L}h_{ij}\hat{c}^{{\dagger}}_{i}\hat{c}_{j},$ (8)
where we have introduced a $L\times L$ matrix $\mathbf{h}=(h_{ij})$. Time
dependence of the fermion anihilation operator is given by
$\hat{c}_{i}(t)=e^{i\hat{H}t}\,\hat{c}_{i}e^{-i\hat{H}t}=\sum_{j=1}^{L}(e^{-i\mathbf{h}t})_{ij}\hat{c}_{j},$
(9)
where the second equality can be obtained from the Baker–Campbell–Hausdorff
formula. Defining a $L\times L$ correlation matrix
$\mathbf{C}(t)=(\mathbf{C}(t))_{i,j}=\bra{\psi}\hat{c}^{{\dagger}}_{i}(t)\hat{c}_{i}(t)\ket{\psi},$
(10)
and using (9), we find that
$\mathbf{C}(t)=e^{i\mathbf{h}t}\mathbf{C}(0)\,e^{-i\mathbf{h}t}.$ (11)
The correlation matrix $\mathbf{C}(0)$ at $t=0$ is determined by the initial
state, and for the Néel state the only non-vanishing coefficients are
$\mathbf{C}(0)_{2k,2k}=1$ where $k=1,\ldots,L/2$. The imbalance is given by
$I(t)=D\sum_{i=1+l_{0}}^{L-l_{0}}(-1)^{i}(\mathbf{C}(t))_{ii},$ (12)
where the constant $D$ assures that $I(0)=0$. Finally, to calculate
entanglement entropy for a bipartition of the system into subsystems
consisting of sites $1,\ldots,l_{A}$ and $l_{A}+1,\ldots L$, we calculate
eigenvalues $\lambda_{i}$ of the submatrix $(\mathbf{C}(t))_{i,j=1}^{l_{A}}$
and compute the entanglement entropy as [138, 139]
$S(t)=-\sum_{i=1}^{l_{A}}[\lambda_{i}\ln(\lambda_{i})+(1-\lambda_{i})\ln(1-\lambda_{i})].$
(13)
The formulas (12) and (13) allow us to calculate the imbalance and
entanglement entropy for the XXZ spin chain with $\Delta=0$ with numerical
cost scaling as $L^{3}$.
Figure 16: The exponent $\overline{\beta}$ obtained from the fit $I(t)\sim
t^{-\overline{\beta}}$ in interval $t\in[t_{\mathrm{min}},t_{\mathrm{max}}]$
as a function of $t_{\mathrm{min}}$ for $W=8,10$ and system size $L=50,200$.
## Appendix B Stability of the power-law fits to choice of time interval
In the main text, the imbalance was fitted by a power-law decay: $I(t)\sim
t^{-\overline{\beta}}$ in interval $t\in[t_{\mathrm{min}},t_{\mathrm{max}}]$,
where $t_{\mathrm{min}}=100$ and the value of $t_{\mathrm{max}}$ was equal to
the maximal time reached in time evolution
($t_{\mathrm{max}}=1200,1500,5000$). In this appendix, we discuss the impact
of changes of $t_{\mathrm{min}}$ on the value of exponent $\overline{\beta}$.
The result is shown in Fig. 16. For $W=8$, we observe that the value
$\overline{\beta}$ remains, within the estimated error bars, constant in the
interval $t_{\mathrm{min}}=[80,200]$, justifying the choice
$t_{\mathrm{min}}=100$ made in the main text. Inclusion of times $t\lessapprox
80$ leads to an increase of $\overline{\beta}$ – consistently with the small
$t$ behavior of the imbalance shown in Fig. 4. When
$t_{\mathrm{min}}\gtrapprox 200$, the precision of estimation of
$\overline{\beta}$ decreases as the fitting interval
$[t_{\mathrm{min}},t_{\mathrm{max}}]$ gets narrower. Similar trends are
observed for $W=10$ for data with $t_{\mathrm{max}}=1500$. However, the
stability of the fit is greatly improved when $t_{\mathrm{max}}=5000$: then,
the choices of $t_{\mathrm{min}}$ from interval $[80,1000]$ lead to the values
of $\overline{\beta}$ that agree within the estimated error bars.
The value of $\overline{\beta}$ decreases approximately $3$ times when $W$ is
increased from $8$ to $10$. However, $\overline{\beta}$ clearly remains
positive within the estimated error bars, showing that the imbalance $I(t)$
indeed decays in time. Extrapolating the trend of changes in
$\overline{\beta}$, we may expect that $\overline{\beta}\approx 10^{-4}$ at
$W=12$. Assuming a similar scaling of the statistical error of
$\overline{\beta}$, already at $W=12$ we would need to either increase the
number of disorder realizations or increase $t_{\mathrm{max}}$ as compared to
their respective values at $W=8,10$ to be certain that the value of
$\overline{\beta}$ at $W=12$ is positive.
Figure 17: Persistent oscillations for QP potential. The imbalance $I(t)$
(averaged over times $[t-25,t+25]$) is shown by solid lines for various
amplitudes of QP potential $W^{QP}$, shades show the imbalance without time
averaging. The results are averaged over more than $5000$ realizations of QP
potential and the system size is fixed as $L=16$. Top panel shows results for
$k=\frac{\sqrt{5}-1}{2}$ whereas the bottom panel for $k=\frac{\sqrt{2}}{2}$.
The range of the vertical axis is the same for all subplots.
## Appendix C Oscillations of the imbalance for quasiperiodic systems
In Sec. VII of the main text, we have demonstrated an emergence of persistent
oscillations of the imbalance $I(t)$ for sufficiently strong QP potential. In
this Appendix we provide further details on this phenomenon.
Since the oscillations do not depend on the system size (at least for $L\geq
12$), we fix the system size as $L=16$ and investigate the time evolution of
the imbalance $I(t)$ varying the amplitude $W^{QP}$ of the QP potential as
well as the wave vector $k$ that determines the shape of the QP potential
(recall that $h_{j}=W^{\mathrm{QP}}\cos(2\pi kj+\phi)$). The results are shown
in Fig. 17.
By comparing the results for fixed $k=\frac{\sqrt{5}-1}{2}$, we note that the
amplitude of oscillations diminishes when $W^{QP}=5$ is increased to
$W^{QP}=8$. This could be expected as in the limit of $W^{QP}\to\infty$, the
initial Néel state becomes an eigenstate of the XXZ model. In that limit, the
oscillations are absent and the imbalance remains trivially equal to unity
throughout the time evolution. Thus, the imbalance oscillations occur only in
a limited range of amplitudes of the QP potential: $W^{QP}$ must be
sufficiently large to give rise to a very slow dynamics (unlike in Fig. 8) but
not large enough to give rise to a trivial dynamics. A similar decrease of the
oscillations of $I(t)$ upon the increase of $W^{QP}$ is visible in Fig. 17 for
$k=\frac{\sqrt{2}}{2}$.
The differences in the pattern of oscillations for $k=\frac{\sqrt{5}-1}{2}$
and $k=\frac{\sqrt{2}}{2}$ (well pronounced for the smaller values of $W^{QP}$
in Fig. 17) demonstrate that the oscillations of $I(t)$ depend in a non-
trivial fashion on the value of the constant $k$. By investigating time
dynamics of the density correlation function $C(t)$ for initial states that
are eigenstates of the $S^{z}_{i}$ operator but are different than the Néel
state, we noted that the pattern of the oscillations of $I(t)$ depends
strongly also on the initial state. In particular, the density correlation
function $C(t)$ averaged over such initial states shows no long-time
oscillations. This shows that the emergence of the pattern of oscillations is
determined by the interplay of a spatial structure of the initial state and
the constant $k$ of the QP potential.
## References
* Deutsch [1991] J. M. Deutsch, Quantum statistical mechanics in a closed system, Phys. Rev. A 43, 2046 (1991).
* Srednicki [1994] M. Srednicki, Chaos and quantum thermalization, Phys. Rev. E 50, 888 (1994).
* Rigol _et al._ [2008] M. Rigol, V. Dunjko, and M. Olshanii, Thermalization and its mechanism for generic isolated quantum systems, Nature 452, 854 EP (2008).
* Gornyi _et al._ [2005] I. V. Gornyi, A. D. Mirlin, and D. G. Polyakov, Interacting electrons in disordered wires: Anderson localization and low-$t$ transport, Phys. Rev. Lett. 95, 206603 (2005).
* Basko _et al._ [2006] D. Basko, I. Aleiner, and B. Altschuler, Metal-insulator transition in a weakly interacting many-electron system with localized single-particle states, Ann. Phys. (NY) 321, 1126 (2006).
* Nandkishore and Huse [2015] R. Nandkishore and D. A. Huse, Mny-body-localization and thermalization in quantum statistical mechanics, Ann. Rev. Cond. Mat. Phys. 6, 15 (2015).
* Alet and Laflorencie [2018] F. Alet and N. Laflorencie, Many-body localization: An introduction and selected topics, Comptes Rendus Physique 19, 498 (2018).
* Abanin _et al._ [2019] D. A. Abanin, E. Altman, I. Bloch, and M. Serbyn, Colloquium: Many-body localization, thermalization, and entanglement, Rev. Mod. Phys. 91, 021001 (2019).
* Huse _et al._ [2014] D. A. Huse, R. Nandkishore, and V. Oganesyan, Phenomenology of fully many-body-localized systems, Phys. Rev. B 90, 174202 (2014).
* Ros _et al._ [2015] V. Ros, M. Mueller, and A. Scardicchio, Integrals of motion in the many-body localized phase, Nuclear Physics B 891, 420 (2015).
* Serbyn _et al._ [2013a] M. Serbyn, Z. Papić, and D. A. Abanin, Local conservation laws and the structure of the many-body localized states, Phys. Rev. Lett. 111, 127201 (2013a).
* Imbrie [2016] J. Z. Imbrie, Diagonalization and many-body localization for a disordered quantum spin chain, Phys. Rev. Lett. 117, 027201 (2016).
* Wahl _et al._ [2017] T. B. Wahl, A. Pal, and S. H. Simon, Efficient representation of fully many-body localized systems using tensor networks, Phys. Rev. X 7, 021018 (2017).
* Mierzejewski _et al._ [2018] M. Mierzejewski, M. Kozarzewski, and P. Prelovšek, Counting local integrals of motion in disordered spinless-fermion and Hubbard chains, Phys. Rev. B 97, 064204 (2018).
* Thomson and Schiró [2018] S. J. Thomson and M. Schiró, Time evolution of many-body localized systems with the flow equation approach, Phys. Rev. B 97, 060201 (2018).
* Žnidarič _et al._ [2016] M. Žnidarič, A. Scardicchio, and V. K. Varma, Diffusive and subdiffusive spin transport in the ergodic phase of a many-body localizable system, Phys. Rev. Lett. 117, 040601 (2016).
* Serbyn _et al._ [2013b] M. Serbyn, Z. Papić, and D. A. Abanin, Universal slow growth of entanglement in interacting strongly disordered systems, Phys. Rev. Lett. 110, 260601 (2013b).
* Iemini _et al._ [2016] F. Iemini, A. Russomanno, D. Rossini, A. Scardicchio, and R. Fazio, Signatures of many-body localization in the dynamics of two-site entanglement, Physical Review B 94, 214206 (2016).
* Santos _et al._ [2004] L. F. Santos, G. Rigolin, and C. O. Escobar, Entanglement versus chaos in disordered spin chains, Phys. Rev. A 69, 042304 (2004).
* Oganesyan and Huse [2007] V. Oganesyan and D. A. Huse, Localization of interacting fermions at high temperature, Phys. Rev. B 75, 155111 (2007).
* Pal and Huse [2010] A. Pal and D. A. Huse, Many-body localization phase transition, Phys. Rev. B 82, 174411 (2010).
* Luitz _et al._ [2015] D. J. Luitz, N. Laflorencie, and F. Alet, Many-body localization edge in the random-field Heisenberg chain, Phys. Rev. B 91, 081103 (2015).
* Mondaini and Rigol [2015] R. Mondaini and M. Rigol, Many-body localization and thermalization in disordered Hubbard chains, Phys. Rev. A 92, 041601 (2015).
* Prelovšek _et al._ [2016] P. Prelovšek, O. S. Barišić, and M. Žnidarič, Absence of full many-body localization in the disordered Hubbard chain, Phys. Rev. B 94, 241104 (2016).
* Zakrzewski and Delande [2018] J. Zakrzewski and D. Delande, Spin-charge separation and many-body localization, Phys. Rev. B 98, 014203 (2018).
* Kozarzewski _et al._ [2018] M. Kozarzewski, P. Prelovšek, and M. Mierzejewski, Spin subdiffusion in the disordered Hubbard chain, Phys. Rev. Lett. 120, 246602 (2018).
* Sierant and Zakrzewski [2018] P. Sierant and J. Zakrzewski, Many-body localization of bosons in optical lattices, New Journal of Physics 20, 043032 (2018).
* Orell _et al._ [2019] T. Orell, A. A. Michailidis, M. Serbyn, and M. Silveri, Probing the many-body localization phase transition with superconducting circuits, Phys. Rev. B 100, 134504 (2019).
* Hopjan and Heidrich-Meisner [2020] M. Hopjan and F. Heidrich-Meisner, Many-body localization from a one-particle perspective in the disordered one-dimensional Bose-Hubbard model, Phys. Rev. A 101, 063617 (2020).
* Sierant _et al._ [2017] P. Sierant, D. Delande, and J. Zakrzewski, Many-body localization due to random interactions, Phys. Rev. A 95, 021601 (2017).
* Bar Lev _et al._ [2016] Y. Bar Lev, D. R. Reichman, and Y. Sagi, Many-body localization in system with a completely delocalized single-particle spectrum, Phys. Rev. B 94, 201116 (2016).
* Li _et al._ [2017] X. Li, D.-L. Deng, Y.-L. Wu, and S. Das Sarma, Statistical bubble localization with random interactions, Phys. Rev. B 95, 020201 (2017).
* Iyer _et al._ [2013] S. Iyer, V. Oganesyan, G. Refael, and D. A. Huse, Many-body localization in a quasiperiodic system, Phys. Rev. B 87, 134202 (2013).
* Khemani _et al._ [2017] V. Khemani, D. N. Sheng, and D. A. Huse, Two universality classes for the many-body localization transition, Phys. Rev. Lett. 119, 075702 (2017).
* Macé _et al._ [2019] N. Macé, N. Laflorencie, and F. Alet, Many-body localization in a quasiperiodic Fibonacci chain, SciPost Phys. 6, 50 (2019).
* Šuntajs _et al._ [2020] J. Šuntajs, J. Bonča, T. Prosen, and L. Vidmar, Quantum chaos challenges many-body localization, Phys. Rev. E 102, 062144 (2020).
* Sierant _et al._ [2020a] P. Sierant, D. Delande, and J. Zakrzewski, Thouless time analysis of anderson and many-body localization transitions, Phys. Rev. Lett. 124, 186601 (2020a).
* Abanin _et al._ [2021] D. Abanin, J. Bardarson, G. De Tomasi, S. Gopalakrishnan, V. Khemani, S. Parameswaran, F. Pollmann, A. Potter, M. Serbyn, and R. Vasseur, Distinguishing localization from chaos: Challenges in finite-size systems, Annals of Physics 427, 168415 (2021).
* Panda _et al._ [2020] R. K. Panda, A. Scardicchio, M. Schulz, S. R. Taylor, and M. Žnidarič, Can we study the many-body localisation transition?, EPL (Europhysics Letters) 128, 67003 (2020).
* Kiefer-Emmanouilidis _et al._ [2020] M. Kiefer-Emmanouilidis, R. Unanyan, M. Fleischhauer, and J. Sirker, Evidence for unbounded growth of the number entropy in many-body localized phases, Phys. Rev. Lett. 124, 243601 (2020).
* Luitz and Lev [2020] D. J. Luitz and Y. B. Lev, Absence of slow particle transport in the many-body localized phase, Phys. Rev. B 102, 100202 (2020).
* Sels and Polkovnikov [2021] D. Sels and A. Polkovnikov, Dynamical obstruction to localization in a disordered spin chain, Phys. Rev. E 104, 054105 (2021).
* Crowley and Chandran [2021] P. J. D. Crowley and A. Chandran, A constructive theory of the numerically accessible many-body localized to thermal crossover (2021), arXiv:2012.14393 [cond-mat.dis-nn] .
* Kiefer-Emmanouilidis _et al._ [2021] M. Kiefer-Emmanouilidis, R. Unanyan, M. Fleischhauer, and J. Sirker, Slow delocalization of particles in many-body localized phases, Phys. Rev. B 103, 024203 (2021).
* Morningstar _et al._ [2021] A. Morningstar, L. Colmenarez, V. Khemani, D. J. Luitz, and D. A. Huse, Avalanches and many-body resonances in many-body localized systems (2021), arXiv:2107.05642 [cond-mat.dis-nn] .
* Sels [2021] D. Sels, Markovian baths and quantum avalanches (2021), arXiv:2108.10796 [cond-mat.dis-nn] .
* Chen _et al._ [2018] C. Chen, F. Burnell, and A. Chandran, How does a locally constrained quantum system localize?, Phys. Rev. Lett. 121, 085701 (2018).
* Sierant _et al._ [2021] P. Sierant, E. G. Lazo, M. Dalmonte, A. Scardicchio, and J. Zakrzewski, Constraint-induced delocalization, Phys. Rev. Lett. 127, 126603 (2021).
* Pietracaprina _et al._ [2018] F. Pietracaprina, N. Macé, D. J. Luitz, and F. Alet, Shift-invert diagonalization of large many-body localizing spin chains, SciPost Phys. 5, 45 (2018).
* Sierant _et al._ [2020b] P. Sierant, M. Lewenstein, and J. Zakrzewski, Polynomially filtered exact diagonalization approach to many-body localization, Phys. Rev. Lett. 125, 156601 (2020b).
* Schollwoeck [2011] Schollwoeck, The density-matrix renormalization group in the age of matrix product states, Ann. Phys. (NY) 326, 96 (2011).
* Paeckel _et al._ [2019] S. Paeckel, T. Köhler, A. Swoboda, S. R. Manmana, U. Schollwöck, and C. Hubig, Time-evolution methods for matrix-product states, Annals of Physics 411, 167998 (2019).
* Vidal [2003] G. Vidal, Efficient classical simulation of slightly entangled quantum computations, Phys. Rev. Lett. 91, 147902 (2003).
* Vidal [2004] G. Vidal, Efficient simulation of one-dimensional quantum many-body systems, Phys. Rev. Lett. 93, 040502 (2004).
* Haegeman _et al._ [2011] J. Haegeman, J. I. Cirac, T. J. Osborne, I. Pižorn, H. Verschelde, and F. Verstraete, Time-dependent variational principle for quantum lattices, Phys. Rev. Lett. 107, 070601 (2011).
* Koffel _et al._ [2012] T. Koffel, M. Lewenstein, and L. Tagliacozzo, Entanglement entropy for the long-range ising chain in a transverse field, Phys. Rev. Lett. 109, 267203 (2012).
* Haegeman _et al._ [2016] J. Haegeman, C. Lubich, I. Oseledets, B. Vandereycken, and F. Verstraete, Unifying time evolution and optimization with matrix product states, Phys. Rev. B 94, 165116 (2016).
* Goto and Danshita [2019] S. Goto and I. Danshita, Performance of the time-dependent variational principle for matrix product states in the long-time evolution of a pure state, Phys. Rev. B 99, 054307 (2019).
* Doggen _et al._ [2018] E. V. H. Doggen, F. Schindler, K. S. Tikhonov, A. D. Mirlin, T. Neupert, D. G. Polyakov, and I. V. Gornyi, Many-body localization and delocalization in large quantum chains, Phys. Rev. B 98, 174202 (2018).
* Chanda _et al._ [2020a] T. Chanda, P. Sierant, and J. Zakrzewski, Time dynamics with matrix product states: Many-body localization transition of large systems revisited, Phys. Rev. B 101, 035148 (2020a).
* Chanda _et al._ [2020b] T. Chanda, P. Sierant, and J. Zakrzewski, Many-body localization transition in large quantum spin chains: The mobility edge, Phys. Rev. Research 2, 032045 (2020b).
* Billy _et al._ [2008] J. Billy, V. Josse, Z. Zuo, A. Bernard, B. Hambrecht, P. Lugan, D. Clément, L. Sanchez-Palencia, P. Bouyer, and A. Aspect, Direct observation of anderson localization of matter waves in a controlled disorder, Nature 453, 891 (2008).
* Chabé _et al._ [2008] J. Chabé, G. Lemarié, B. Grémaud, D. Delande, P. Szriftgiser, and J. C. Garreau, Experimental observation of the anderson metal-insulator transition with atomic matter waves, Physical Review Letters 101, 10.1103/physrevlett.101.255702 (2008).
* Schiulaz _et al._ [2019] M. Schiulaz, E. J. Torres-Herrera, and L. F. Santos, Thouless and relaxation time scales in many-body quantum systems, Phys. Rev. B 99, 174313 (2019).
* Schreiber _et al._ [2015] M. Schreiber, S. S. Hodgman, P. Bordia, H. P. Lüschen, M. H. Fischer, R. Vosk, E. Altman, U. Schneider, and I. Bloch, Observation of many-body localization of interacting fermions in a quasirandom optical lattice, Science 349, 842 (2015).
* Smith _et al._ [2016] J. Smith, A. Lee, P. Richerme, B. Neyenhuis, P. W. Hess, P. Hauke, M. Heyl, D. A. Huse, and C. Monroe, Many-body localization in a quantum simulator with programmable random disorder, Nature Physics 12, 907 (2016).
* Lüschen _et al._ [2017] H. P. Lüschen, P. Bordia, S. Scherg, F. Alet, E. Altman, U. Schneider, and I. Bloch, Observation of slow dynamics near the many-body localization transition in one-dimensional quasiperiodic systems, Phys. Rev. Lett. 119, 260401 (2017).
* Kohlert _et al._ [2019] T. Kohlert, S. Scherg, X. Li, H. P. Lüschen, S. Das Sarma, I. Bloch, and M. Aidelsburger, Observation of many-body localization in a one-dimensional system with a single-particle mobility edge, Phys. Rev. Lett. 122, 170403 (2019).
* Lukin _et al._ [2019] A. Lukin, M. Rispoli, R. Schittko, M. E. Tai, A. M. Kaufman, S. Choi, V. Khemani, J. Léonard, and M. Greiner, Probing entanglement in a many-body–localized system, Science 364, 256 (2019).
* Rispoli _et al._ [2019] M. Rispoli, A. Lukin, R. Schittko, S. Kim, M. E. Tai, J. Léonard, and M. Greiner, Quantum critical behaviour at the many-body localization transition, Nature 573, 385 (2019).
* Léonard _et al._ [2020] J. Léonard, M. Rispoli, A. Lukin, R. Schittko, S. Kim, J. Kwan, D. Sels, E. Demler, and M. Greiner, Signatures of bath-induced quantum avalanches in a many-body–localized system (2020), arXiv:2012.15270 [cond-mat.quant-gas] .
* Yao _et al._ [2014] N. Y. Yao, C. R. Laumann, S. Gopalakrishnan, M. Knap, M. Müller, E. A. Demler, and M. D. Lukin, Many-body localization in dipolar systems, Phys. Rev. Lett. 113, 243002 (2014).
* Pino [2014] M. Pino, Entanglement growth in many-body localized systems with long-range interactions, Phys. Rev. B 90, 174204 (2014).
* Burin [2015a] A. L. Burin, Many-body delocalization in a strongly disordered system with long-range interactions: Finite-size scaling, Phys. Rev. B 91, 094202 (2015a).
* Burin [2015b] A. L. Burin, Localization in a random xy model with long-range interactions: Intermediate case between single-particle and many-body problems, Phys. Rev. B 92, 104428 (2015b).
* Deng _et al._ [2020] X. Deng, G. Masella, G. Pupillo, and L. Santos, Universal algebraic growth of entanglement entropy in many-body localized systems with power-law interactions, Phys. Rev. Lett. 125, 010401 (2020).
* Hauke and Heyl [2015] P. Hauke and M. Heyl, Many-body localization and quantum ergodicity in disordered long-range ising models, Phys. Rev. B 92, 134204 (2015).
* Singh _et al._ [2017] R. Singh, R. Moessner, and D. Roy, Effect of long-range hopping and interactions on entanglement dynamics and many-body localization, Phys. Rev. B 95, 094205 (2017).
* Sous and Grant [2018] J. Sous and E. Grant, Possible many-body localization in a long-lived finite-temperature ultracold quasineutral molecular plasma, Phys. Rev. Lett. 120, 110601 (2018).
* Sous and Grant [2019] J. Sous and E. Grant, Many-body physics with ultracold plasmas: quenched randomness and localization, New Journal of Physics 21, 043033 (2019).
* Botzung _et al._ [2019] T. Botzung, D. Vodola, P. Naldesi, M. Müller, E. Ercolessi, and G. Pupillo, Algebraic localization from power-law couplings in disordered quantum wires, Phys. Rev. B 100, 155136 (2019).
* Sierant _et al._ [2019] P. Sierant, K. Biedroń, G. Morigi, and J. Zakrzewski, Many-body localization in presence of cavity mediated long-range interactions, SciPost Phys. 7, 8 (2019).
* Kubala _et al._ [2021] P. Kubala, P. Sierant, G. Morigi, and J. Zakrzewski, Ergodicity breaking with long-range cavity-induced quasiperiodic interactions, Phys. Rev. B 103, 174208 (2021).
* van Nieuwenburg _et al._ [2019] E. van Nieuwenburg, Y. Baum, and G. Refael, From bloch oscillations to many-body localization in clean interacting systems, Proceedings of the National Academy of Sciences 116, 9269 (2019).
* Schulz _et al._ [2019] M. Schulz, C. A. Hooley, R. Moessner, and F. Pollmann, Stark many-body localization, Phys. Rev. Lett. 122, 040606 (2019).
* Wu and Eckardt [2019] L.-N. Wu and A. Eckardt, Bath-induced decay of stark many-body localization, Phys. Rev. Lett. 123, 030602 (2019).
* Taylor _et al._ [2020] S. R. Taylor, M. Schulz, F. Pollmann, and R. Moessner, Experimental probes of stark many-body localization, Phys. Rev. B 102, 054206 (2020).
* Guardado-Sanchez _et al._ [2021] E. Guardado-Sanchez, B. M. Spar, P. Schauss, R. Belyansky, J. T. Young, P. Bienias, A. V. Gorshkov, T. Iadecola, and W. S. Bakr, Quench dynamics of a fermi gas with strong nonlocal interactions, Phys. Rev. X 11, 021036 (2021).
* Khemani _et al._ [2020] V. Khemani, M. Hermele, and R. Nandkishore, Localization from Hilbert space shattering: From theory to physical realizations, Phys. Rev. B 101, 174204 (2020).
* Doggen _et al._ [2021] E. V. H. Doggen, I. V. Gornyi, and D. G. Polyakov, Stark many-body localization: Evidence for hilbert-space shattering, Phys. Rev. B 103, L100202 (2021).
* Yao and Zakrzewski [2020] R. Yao and J. Zakrzewski, Many-body localization of bosons in an optical lattice: Dynamics in disorder-free potentials, Phys. Rev. B 102, 104203 (2020).
* Yao _et al._ [2021a] R. Yao, T. Chanda, and J. Zakrzewski, Nonergodic dynamics in disorder-free potentials, Annals of Physics 435, 168540 (2021a), special Issue on Localisation 2020\.
* Guo _et al._ [2020] A. Y. Guo, M. C. Tran, A. M. Childs, A. V. Gorshkov, and Z.-X. Gong, Signaling and scrambling with strongly long-range interactions, Phys. Rev. A 102, 010401 (2020).
* Chanda _et al._ [2020c] T. Chanda, R. Yao, and J. Zakrzewski, Coexistence of localized and extended phases: Many-body localization in a harmonic trap, Phys. Rev. Research 2, 032039 (2020c).
* Morong _et al._ [2021] W. Morong, F. Liu, P. Becker, K. S. Collins, L. Feng, A. Kyprianidis, G. Pagano, T. You, A. V. Gorshkov, and C. Monroe, Observation of stark many-body localization without disorder, Nature 599, 393 (2021).
* Scherg _et al._ [2021] S. Scherg, T. Kohlert, P. Sala, F. Pollmann, B. Hebbe Madhusudhana, I. Bloch, and M. Aidelsburger, Observing non-ergodicity due to kinetic constraints in tilted fermi-hubbard chains, Nature Communications 12, 4490 (2021).
* Yao _et al._ [2021b] R. Yao, T. Chanda, and J. Zakrzewski, Many-body localization in tilted and harmonic potentials, Phys. Rev. B 104, 014201 (2021b).
* Berkelbach and Reichman [2010] T. C. Berkelbach and D. R. Reichman, Conductivity of disordered quantum lattice models at infinite temperature: Many-body localization, Phys. Rev. B 81, 224429 (2010).
* Agarwal _et al._ [2015] K. Agarwal, S. Gopalakrishnan, M. Knap, M. Müller, and E. Demler, Anomalous diffusion and griffiths effects near the many-body localization transition, Phys. Rev. Lett. 114, 160401 (2015).
* Bera _et al._ [2015] S. Bera, H. Schomerus, F. Heidrich-Meisner, and J. H. Bardarson, Many-body localization characterized from a one-particle perspective, Phys. Rev. Lett. 115, 046603 (2015).
* Enss _et al._ [2017] T. Enss, F. Andraschko, and J. Sirker, Many-body localization in infinite chains, Phys. Rev. B 95, 045121 (2017).
* Bera _et al._ [2017a] S. Bera, G. De Tomasi, F. Weiner, and F. Evers, Density propagator for many-body localization: Finite-size effects, transient subdiffusion, and exponential decay, Phys. Rev. Lett. 118, 196801 (2017a).
* Herviou _et al._ [2019] L. Herviou, S. Bera, and J. H. Bardarson, Multiscale entanglement clusters at the many-body localization phase transition, Phys. Rev. B 99, 134205 (2019).
* Colmenarez _et al._ [2019] L. A. Colmenarez, P. A. McClarty, M. Haque, and D. J. Luitz, Statistics of correlation functions in the random heisenberg chain, SciPost Physics 7, 10.21468/scipostphys.7.5.064 (2019).
* Sierant and Zakrzewski [2019] P. Sierant and J. Zakrzewski, Level statistics across the many-body localization transition, Phys. Rev. B 99, 104205 (2019).
* Sierant and Zakrzewski [2020] P. Sierant and J. Zakrzewski, Model of level statistics for disordered interacting quantum many-body systems, Phys. Rev. B 101, 104201 (2020).
* Schiulaz _et al._ [2020] M. Schiulaz, E. J. Torres-Herrera, F. Pérez-Bernal, and L. F. Santos, Self-averaging in many-body quantum systems out of equilibrium: Chaotic systems, Phys. Rev. B 101, 174312 (2020).
* Torres-Herrera _et al._ [2020] E. J. Torres-Herrera, G. De Tomasi, M. Schiulaz, F. Pérez-Bernal, and L. F. Santos, Self-averaging in many-body quantum systems out of equilibrium: Approach to the localized phase, Phys. Rev. B 102, 094310 (2020).
* Macé _et al._ [2019] N. Macé, F. Alet, and N. Laflorencie, Multifractal scalings across the many-body localization transition, Phys. Rev. Lett. 123, 180601 (2019).
* Laflorencie _et al._ [2020] N. Laflorencie, G. Lemarié, and N. Macé, Chain breaking and kosterlitz-thouless scaling at the many-body localization transition in the random-field heisenberg spin chain, Phys. Rev. Research 2, 042033 (2020).
* Gray _et al._ [2018] J. Gray, S. Bose, and A. Bayat, Many-body localization transition: Schmidt gap, entanglement length, and scaling, Phys. Rev. B 97, 201105 (2018).
* Naldesi _et al._ [2016] P. Naldesi, E. Ercolessi, and T. Roscilde, Detecting a many-body mobility edge with quantum quenches, SciPost Phys. 1, 010 (2016).
* Setiawan _et al._ [2017] F. Setiawan, D.-L. Deng, and J. H. Pixley, Transport properties across the many-body localization transition in quasiperiodic and random systems, Phys. Rev. B 96, 104205 (2017).
* Lev _et al._ [2017] Y. B. Lev, D. M. Kennes, C. Klöckner, D. R. Reichman, and C. Karrasch, Transport in quasiperiodic interacting systems: From superdiffusion to subdiffusion, EPL (Europhysics Letters) 119, 37003 (2017).
* Bera _et al._ [2017b] S. Bera, T. Martynec, H. Schomerus, F. Heidrich-Meisner, and J. H. Bardarson, One-particle density matrix characterization of many-body localization, Annalen der Physik 529, 1600356 (2017b).
* Weidinger _et al._ [2018] S. A. Weidinger, S. Gopalakrishnan, and M. Knap, Self-consistent hartree-fock approach to many-body localization, Phys. Rev. B 98, 224205 (2018).
* Doggen and Mirlin [2019] E. V. H. Doggen and A. D. Mirlin, Many-body delocalization dynamics in long aubry-andré quasiperiodic chains, Phys. Rev. B 100, 104203 (2019).
* Weiner _et al._ [2019] F. Weiner, F. Evers, and S. Bera, Slow dynamics and strong finite-size effects in many-body localization with random and quasiperiodic potentials, Phys. Rev. B 100, 104204 (2019).
* Singh _et al._ [2021] H. Singh, B. Ware, R. Vasseur, and S. Gopalakrishnan, Local integrals of motion and the quasiperiodic many-body localization transition, Physical Review B 103, 10.1103/physrevb.103.l220201 (2021).
* Zhang and Yao [2018] S.-X. Zhang and H. Yao, Universal properties of many-body localization transitions in quasiperiodic systems, Phys. Rev. Lett. 121, 206601 (2018).
* Agrawal _et al._ [2020] U. Agrawal, S. Gopalakrishnan, and R. Vasseur, Universality and quantum criticality in quasiperiodic spin chains, Nature Communications 11, 2225 (2020).
* Aramthottil _et al._ [2021] A. S. Aramthottil, T. Chanda, P. Sierant, and J. Zakrzewski, Finite-size scaling analysis of the many-body localization transition in quasiperiodic spin chains, Phys. Rev. B 104, 214201 (2021).
* Nandy _et al._ [2021] S. Nandy, F. Evers, and S. Bera, Dephasing in strongly disordered interacting quantum wires, Phys. Rev. B 103, 085105 (2021).
* Vidmar _et al._ [2021] L. Vidmar, B. Krajewski, J. Bonča, and M. Mierzejewski, Phenomenology of spectral functions in disordered spin chains at infinite temperature, Phys. Rev. Lett. 127, 230603 (2021).
* Fehske and Schneider [2008] H. Fehske and R. Schneider, _Computational many-particle physics_ (Springer, Germany, 2008).
* Luitz _et al._ [2016] D. J. Luitz, N. Laflorencie, and F. Alet, Extended slow dynamical regime close to the many-body localization transition, Phys. Rev. B 93, 060201 (2016).
* Anderson [1958] P. W. Anderson, Absence of diffusion in certain random lattices, Phys. Rev. 109, 1492 (1958).
* Žnidarič _et al._ [2008] M. Žnidarič, T. Prosen, and P. Prelovšek, Many-body localization in the Heisenberg XXZ magnet in a random field, Phys. Rev. B 77, 064426 (2008).
* Bardarson _et al._ [2012] J. H. Bardarson, F. Pollmann, and J. E. Moore, Unbounded growth of entanglement in models of many-body localization, Phys. Rev. Lett. 109, 017202 (2012).
* Schuch _et al._ [2004a] N. Schuch, F. Verstraete, and J. I. Cirac, Nonlocal resources in the presence of superselection rules, Phys. Rev. Lett. 92, 087904 (2004a).
* Schuch _et al._ [2004b] N. Schuch, F. Verstraete, and J. I. Cirac, Quantum entanglement theory in the presence of superselection rules, Phys. Rev. A 70, 042310 (2004b).
* Donnelly [2012] W. Donnelly, Decomposition of entanglement entropy in lattice gauge theory, Phys. Rev. D 85, 085004 (2012).
* Turkeshi _et al._ [2020] X. Turkeshi, P. Ruggiero, V. Alba, and P. Calabrese, Entanglement equipartition in critical random spin chains, Phys. Rev. B 102, 014455 (2020).
* Ghosh and Žnidarič [2022] R. Ghosh and M. Žnidarič, Theory of growth of number entropy in disordered systems (2022), arXiv:2112.12987 [cond-mat.dis-nn] .
* White and Feiguin [2004] S. R. White and A. E. Feiguin, Real-time evolution using the density matrix renormalization group, Phys. Rev. Lett. 93, 076401 (2004).
* Daley _et al._ [2004] A. J. Daley, C. Kollath, U. Schollwöck, and G. Vidal, Time-dependent density-matrix renormalization-group using adaptive effective hilbert spaces, Journal of Statistical Mechanics: Theory and Experiment 2004, P04005 (2004).
* Peschel and Eisler [2009] I. Peschel and V. Eisler, Reduced density matrices and entanglement entropy in free lattice models, Journal of Physics A: Mathematical and Theoretical 42, 504003 (2009).
* Peschel [2003] I. Peschel, Calculation of reduced density matrices from correlation functions, Journal of Physics A: Mathematical and General 36, L205 (2003).
* Cheong and Henley [2004] S.-A. Cheong and C. L. Henley, Many-body density matrices for free fermions, Phys. Rev. B 69, 075111 (2004).
|
# VT-CLIP: Enhancing Vision-Language Models with Visual-guided Texts
###### Abstract
Contrastive Language-Image Pre-training (CLIP) has drawn increasing attention
recently for its transferable visual representation learning. However, due to
the semantic gap within datasets, CLIP’s pre-trained image-text alignment
becomes sub-optimal on downstream tasks, which severely harms its transferring
performance. To better adapt the cross-modality embedding space, we propose to
enhance CLIP via Visual-guided Texts, named VT-CLIP. Specifically, we guide
textual features of different categories to adaptively explore informative
regions on the image and aggregate visual features by attention mechanisms. In
this way, the texts become visual-guided, namely, more semantically correlated
with downstream images, which greatly benefits the category-wise matching
process. In few-shot settings, we evaluate our VT-CLIP on 11 well-known
classification datasets to demonstrate its effectiveness.
Index Terms— Contrastive language-image pre-training, Few-shot learning,
Image classification
## 1 Introduction
With the exploring of better network architectures, traditional deep learning
has achieved extraordinary performances over a wide range of vision tasks,
such as image classification [1], object detection [2] and so on [3]. Although
these methods are expert at specific scenarios, they lack the ability of
general vision representation and are hard to transfer to open-set
applications. Considering the wide coverage of languages, CLIP [4] proposes to
pre-train the visual representation learning contrastively with natural
language signals. Supervised by large-scale image-text pairs, CLIP extracts
both features of input images and texts by separate encoders and matches the
paired ones in the same embedding space. Such pre-trained cross-modality
alignment endows CLIP the capability of recognizing new visual concepts on
downstream tasks. Specifically, given a new dataset with images to be
recognized, one could construct the textual inputs of CLIP by the category
names, termed as prompts, and convert the classification task into a matching
problem. By this, CLIP is able to conduct zero-shot recognition in open-
vocabulary settings.
To further improve the downstream performance of CLIP, existing works
introduce different fine-tuning methods under the few-shot settings. Inspired
by prompt tuning in natural language processing [5], Context Optimization
(CoOp) [6] freezes CLIP’s pre-trained weights and adopts learnable prompts to
learn the best-fitted textual inputs other than the original hand-crafted
ones. From another perspective, CLIP-Adapter [7] appends a lightweight adapter
module [8] over the image or text encoders and is also fine-tuned with frozen
CLIP’s weights. However, they are constrained by the following limitations.
The learnable prompts in CoOp are set before the large-scale text encoder,
which is much time-consuming to back-propagate the training gradients through
such 12-layer transformer [9] for every iteration. CLIP-Adapter only conducts
feature adaption independently for each branch, and lacks cross-modal
interactions for image and language. Also, both of their learned parameters
are fixed for all images during inference, losing the adaption flexibility for
varying visual inputs.
Fig. 1: Visualization of attention map and prediction score. In our VT-CLIP,
the informative regions on the image gain more attention weights from the
texts, which improves the prediction score on ground-truth category.
In this paper, we propose VT-CLIP, which equips CLIP with Visual-guided Texts
to exert thorough adaptions of both modalities under few-shot settings.
Specifically, we introduce a visual-guided attention module to conduct feature
communication after both encoders, which views texts as the queries, and
images as keys and values. For textual branch, we input the hand-crafted
prompts with explicit semantics and utilize the encoded features for
interaction with images. For visual branch, we extract the intermediate
spatial features of images as keys and values instead of the final global
ones, which could provide more fine-grained contextual information. By
calculating the per-pixel similarities, different categories in texts could
explore informative visual regions and gather related features weighted by
their attention scores. After this, the texts are visual-guided and better
fitted for the later matching stage. As visualized in Figure 1, the textual
feature of ”airline” category focuses on the corresponding visual regions, but
other unmatched categories do not, as expected. Importantly, our visual-guided
texts are adaptive for different samples, since the attention map is
dynamically produced by the input images. We conduct extensive experiments on
11 well-known classification datasets, which fully demonstrates the
outstanding enhancement ability of visual-guided texts over CLIP.
## 2 RELATED WORK
Recently, vision-Language Models shows great potential in learning generic
visual representation with nature language supervision, which allowing zero-
shot transfer ability for various downstream classification tasks. Inspired by
the success of pre-train models [10], [11] [12] and SimVLM [13] use attention
architecture improve the performance of vision-language tasks. At present, the
recent breakthrough in vision-language learning, particularly CLIP [4] and
ALIGN [14] are driven by the noisy large-scale datasets available in the
Internet, which is 400 million image-text pair for CLIP and 1.8 billion noisy
image-text pairs for ALIGN.To fine-tune vision-Language Models on downstream
tasks like few-shot classification task, CoOp [15] propose to learn soft
prompts represented by continuous context vectors as alternative for hand-
craft prompt while CLIP-Adapter propose to adopts an additional bottleneck
layer to learn new features and performs residual style feature blending with
the original pre-trained features. Though CoOp and CLIP-Adapter achieve
significant performance in the perspective of prompt learning and feature
adapters, our VT-CLIP explores the impact of instance-level image visual
feature on refining text feature with a cross-attention module.
Prompt Learning are designed to better mine the knowledge from pre-trained
models without fine-tuning the entire model, which generate a prompting
template or function to bridge gap between the pre-training objective and
downstream tasks [5, 16, 17, 18]. Prompt engineering is an important topic in
prompt learning. Early research focus on designing hand-crafted prompts, which
generate cloze-style prompts like “fill-in-the-blank” cloze tests and benefits
a number of downstream tasks, such as sentiment analysis [16]. Recently, [17,
18] introduce gradient-based approaches which optimize continuous vectors in
the word embedding space. The limitation of prompt engineering is that hand-
crafted prompt template requires specific domain knowledge and the prompt
content learned by optimization lacks interpretability. In this paper, we
demonstrate guiding text feature with instance-level image feature through
cross-attention module is an alternative for prompt learning on large-scale
vision-language models, which is more interpretable and simpler in
architecture.
Fig. 2: Structures of VT-CLIP. Class-level text feature from CLIP text encoder
is updated by spatial visual image feature with cross attention module
## 3 METHODS
We first revisit the zero-shot CLIP in Section 3.1, and then introduce the
details of our proposed VT-CLIP in Section 3.2.
### 3.1 Zero-shot CLIP
CLIP is pre-trained to align image and text pair information through
contrastive training. CLIP contains two independent encoders for visual and
textual feature encoding. Specifically, the image encoder consists of a visual
backbone, which is ResNet [1] or ViT [19], and an attention pooling layer,
while the text encoder is a conventional Transformer Encoder . With large-
scale data traning, informative representation for both modalities are deeply
learned by CLIP’s image encoder and text encoder, thus obtaining zero-shot
classification ability. In details, for a dataset contains $K$ categories,
denoted as {$C_{1},\dots,C_{K}$}, CLIP first places all category names into
the hand-crafted template $H$ proposed by [4] to get the textual inputs, which
then are fed into the tokenizer $T$ and text encoder to obtain text features,
$T_{c}\in R^{1\times C}$. Meanwhile, the input image $I\in
R^{H\times{W\times{3}}}$, where $H$ and $W$ are the height and width of the
image respectively, is first encoded by visual backbone, termed as $VB$,
getting contextual-level spatial features $V_{s}$. After that, an attention
pooling operation is adopted to get global visual features $V_{c}$, i.e.,
$\displaystyle V_{s}=\operatorname{VB}(I),V_{s}\in R^{H_{s}\times W_{s}\times
C_{s}}$ (1) $\displaystyle V_{c}=\operatorname{Pooling}(V_{s}),V_{c}\in
R^{1\times C}$ (2) $\displaystyle
T_{c}=\operatorname{TextEn}\big{(}T([H;C_{i}])\big{)},i\in\\{1,\dots,K\\},$
(3)
where $TextEn$ denotes text encoder, and $C$ is the class-level feature
dimension. The $H_{s},W_{s},C_{s}$ are the height , width and channel
dimension for spatial feature. Via attention pooling, CLIP generates the
global visual features from spatial features which contain more local
contextual-level information. Finally, the similarity scores are calculated
as,
$\displaystyle P=\operatorname{Softmax}(V_{c}T_{c}^{T}/\tau),$ (4)
where $SoftMax(\cdot)$ and $P$ denote the softmax function and the similarity
scores for $K$ categories, and $\tau$ is a temperature parameter learned by
CLIP.
### 3.2 VT-CLIP
Different from the perspective of prompt learning , we present a new approach
to enhance vision-language model. We suppose that generic soft prompt which is
invariant to images with various content is kind of insufficient. Hence, we
propose VT-CLIP, which dynamically refine the text features using visual
spatial features. Specifically, through a visual-guided cross-attention
module, we leverage the contextual-level spatial feature, which is obtained
before pooling, to guide the text feature to adaptively explore informative
regions on the image. The learned refinement is fused with original text
features a by residual connection to preserve the robustness and
effectiveness. Following the standard architecture of the transformer decoder
blocks [9], our proposed visual-guided cross-attention module includes a self-
attention layer, a co-attention layer and a feed forward network, where
$T_{c}$ and $V_{s}$ are fed into the cross attention module, with $T_{c}$
serving as query, and $V_{s}$ as key and value, i.e.,
$\displaystyle VT_{c}=\operatorname{CrossAttn}(V_{s},V_{s},T_{c})+T_{c},$ (5)
$\displaystyle VT_{c}\in R^{1\times C}$
where $CrossAttn$ is the visual-guided cross-attention module, and $VT_{c}$
denotes the adapted text features.
Through the interaction in cross-attention, the adapted text features $VT_{c}$
become more semantically correlated with the paired image. Then, the
similarity scores are predicted via obtained $VT_{c}$, that is,
$\displaystyle P=\operatorname{Softmax}(V_{c}VT_{c}^{T}/\tau).$ (6)
During training, we freeze both visual baskbone and textual encoder, and only
optimize weights in visual-guided cross-attention module via cross-entropy
loss.
Fig. 3: Experiment Results––Main results of few-shot learning on 11 datasets.
VT-CLIP shows overall better performance over baselines across different
training shots.
## 4 EXPERIMENTS
### 4.1 Training Settings
We evaluate the performance of VT-CLIP on 11 widely-adopted image
classification datasets and follow the few-shot evaluation protocol of CLIP
[4], that is, training on $1$, $2$, $4$, $8$ and $16$ shots and testing on
full test set. During training, we adopt pre-trained visual backbone ResNet-50
from [4] with all the weights frozen, and follow the data preprocessing
protocol of CLIP, which consists of random cropping, resizing, and random
horizontal flip. Following CoOp [6], VT-CLIP is trained with batch size $32$
and learning rate $2\times 10^{-3}$ for all 11 datasets. Instead of using
learnable continuous prompts in CoOp, we adopt the same hand-crafted prompt as
CLIP. We compare VT-CLIP with three baseline works, i.e., Zero-shot CLIP [4],
CoOp [6] and CLIP-Adapter [7]. Also, in order to thoroughly demonstrate the
superiority of proposed VT-CLIP, we use the best baseline variants.
### 4.2 Performance Comparison & Analysis
The main results are presented in Figure 3. As shown in the top-left chart of
Figure 3, VT-CLIP shows outstanding average performance over three baselines,
and the accuracy gain increases as the training shots get more, which
indicates VT-CLIP serves as an effective and reliable enhancer under few-shot
settings. Also, as shown in all the twelve charts in Figure 3, our VT-CLIP
outperforms other works significantly under each shot setting on 11 datasets.
What’s more, unlike CoOp’s poor performance with little training samples, VT-
CLIP achieves more stable scores, which indicates our VT-CLIP is not sensitive
to data scale. Additionally, it is clear that VT-CLIP obtains consistently
prominent results on all the 11 datasers, which demonstrates more considerable
generalization ability than CoOp, as seen in charts of OxfordPets [20] and
Food101 [21], where CoOp falls behind even zero-shot CLIP under 16-shot
setting. As for CLIP-Adapter, VT-CLIP not only surpasses it on different
datasets, but also contains better interpretability of leveraging contextual
visual features to guide text features to be more semantically correlated to
the certain downstream task. The consistent superiority of VT-CLIP over 11
datasets fully demonstrates the effectiveness and generality of our proposed
method.
## 5 Ablation Study
In this section, we conduct several ablation studies for VT-CLIP. All
experiments below adopt the 16-shot setting.
Heads | 4 | 8 | 16 | 32
---|---|---|---|---
Caltech101 (%) | 93.06 | 93.10 | 92.37 | 92.62
DTD (%) | 64.42 | 65.72 | 64.78 | 65.43
Table 1: Head Number. Performance with different number of heads in cross
attention.
We explore the number of heads in cross attention on Caltech-
101 [22] and DTD [23]. The heads number in attention mechanism equals to the
number of heterogeneous scores computed for values indicating fitting ability
of model. As presented in Table 1, the best performance is achieved with two
heads for cross attention module. To further demonstrate the design of our
method, we conduct experiments on increase the number of cascaded attention
layers. The results are in Table 2. We observe that performance degrades as
the cascaded layers increase which indicate the complex model with multiple
cross attention layer tend to overfit the insufficient training data under few
shot scenario.
Layers | 1 | 2 | 3 | 4
---|---|---|---|---
Caltech101(%) | 93.10 | 93.06 | 92.58 | 92.29
DTD (%) | 65.72 | 64.60 | 65.60 | 64.78
Table 2: Layer Number. Performance with different number of cascaded cross
attention layers.
## 6 CONCLUSION
We propose VT-CLIP, a novel enhancement of CLIP for few-shot classification
which focuses on leveraging the contextual visual features to guide the text
features to highlight the important regions via a visual-guided cross-
attention module. In this way, the deep interaction between the image and text
branches in vision-language model is of great potential in enhancing the
model’s ability. Also, extensive experiments demonstrate that VT-CLIP
outperforms all the competitive baselines in few-shot settings on 11 widely-
used datastes. Ablation studies are conducted to further prove our design and
give a view of the extensive performance of VT-CLIP. In the future, we hope to
combine VT-CLIP with prompt learning based approaches to push the boundary
further for fine-tuning vision-language models. We will also explore the
potential of VT-CLIP on other vision or textual tasks.
## References
* [1] K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” IEEE, 2016.
* [2] N. Carion, F. Massa, G. Synnaeve, N. Usunier, A. Kirillov, and S. Zagoruyko, “End-to-end object detection with transformers,” 2020\.
* [3] Kaiming He, Georgia Gkioxari, Piotr Dollár, and Ross Girshick, “Mask r-cnn,” in Proceedings of the IEEE international conference on computer vision, 2017, pp. 2961–2969.
* [4] Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, et al., “Learning transferable visual models from natural language supervision,” in International Conference on Machine Learning. PMLR, 2021, pp. 8748–8763.
* [5] T. Shin, Y. Razeghi, Irl Logan, E. Wallace, and S. Singh, “Autoprompt: Eliciting knowledge from language models with automatically generated prompts,” 2020\.
* [6] Kaiyang Zhou, Ziwei Liu, Yu Qiao, Tao Xiang, and Chen Change Loy, “Domain generalization: A survey,” arXiv preprint arXiv:2103.02503, 2021.
* [7] Peng Gao, Shijie Geng, Renrui Zhang, Teli Ma, Rongyao Fang, Yongfeng Zhang, Hongsheng Li, and Yu Qiao, “Clip-adapter: Better vision-language models with feature adapters,” arXiv preprint arXiv:2110.04544, 2021.
* [8] Neil Houlsby, Andrei Giurgiu, Stanislaw Jastrzebski, Bruna Morrone, Quentin De Laroussilhe, Andrea Gesmundo, Mona Attariyan, and Sylvain Gelly, “Parameter-efficient transfer learning for nlp,” in International Conference on Machine Learning. PMLR, 2019, pp. 2790–2799.
* [9] Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Lukasz Kaiser, and Illia Polosukhin, “Attention is all you need,” in Advances in neural information processing systems, 2017, pp. 5998–6008.
* [10] Jacob Devlin, Ming Wei Chang, Kenton Lee, and Kristina Toutanova, “Bert: Pre-training of deep bidirectional transformers for language understanding,” 2018\.
* [11] Y. C. Chen, L. Li, L. Yu, A. E. Kholy, F. Ahmed, Z. Gan, Y. Cheng, and J. Liu, “Uniter: Learning universal image-text representations,” 2019\.
* [12] X. Li, X. Yin, C. Li, P. Zhang, X. Hu, L. Zhang, L. Wang, H. Hu, L. Dong, and F. Wei, “Oscar: Object-semantics aligned pre-training for vision-language tasks,” 2020\.
* [13] Zirui Wang, Jiahui Yu, Adams Wei Yu, Zihang Dai, and Yuan Cao, “Simvlm: Simple visual language model pretraining with weak supervision,” 2021\.
* [14] C. Jia, Y. Yang, Y. Xia, Y. T. Chen, and T. Duerig, “Scaling up visual and vision-language representation learning with noisy text supervision,” 2021\.
* [15] Kaiyang Zhou, Jingkang Yang, Chen Change Loy, and Ziwei Liu, “Learning to prompt for vision-language models,” arXiv preprint arXiv:2109.01134, 2021.
* [16] Z. Jiang, F. F. Xu, J. Araki, and G. Neubig, “How can we know what language models know?,” 2019\.
* [17] X. L. Li and P. Liang, “Prefix-tuning: Optimizing continuous prompts for generation,” 2021\.
* [18] B. Lester, R. Al-Rfou, and N. Constant, “The power of scale for parameter-efficient prompt tuning,” 2021\.
* [19] A. Dosovitskiy, L. Beyer, A. Kolesnikov, D. Weissenborn, and N. Houlsby, “An image is worth 16x16 words: Transformers for image recognition at scale,” 2020\.
* [20] Andrea Vedaldi, “Cats and dogs,” in Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2012, pp. 3498–3505.
* [21] L. Bossard, M. Guillaumin, and L. V. Gool, “Food-101 – mining discriminative components with random forests,” in Springer International Publishing, 2014.
* [22] FF Li, R. Fergus, and P. Perona, “Learning generative visual models from few training examples: An incremental bayesian approach tested on 101 object categories,” IEEE, 2004.
* [23] M. Cimpoi, S. Maji, I. Kokkinos, S. Mohamed, and A. Vedaldi, “Describing textures in the wild,” IEEE, 2013.
|
# EvoGrad: A Dynamic Take on the Winograd Schema Challenge with Human
Adversaries
###### Abstract
While Large Language Models (LLMs) excel at the Winograd Schema Challenge
(WSC), a coreference resolution task testing common-sense reasoning through
pronoun disambiguation, they struggle with instances that feature minor
alterations or rewording. To address this, we introduce EvoGrad, an open-
source platform that harnesses a human-in-the-loop approach to create a
dynamic dataset tailored to such altered WSC instances. Leveraging ChatGPT’s
capabilities, we expand our task instances from 182 to 3,691, setting a new
benchmark for diverse common-sense reasoning datasets. Additionally, we
introduce the error depth metric, assessing model stability in dynamic tasks.
Our results emphasize the challenge posed by EvoGrad: Even the best performing
LLM, GPT-3.5, achieves an accuracy of 65.0% with an average error depth of
7.2, a stark contrast to human performance of 92. 8% accuracy without
perturbation errors. This highlights ongoing model limitations and the value
of dynamic datasets in uncovering them.
Keywords: Winograd Schema Challenge, Common-sense Reasoning, Large Language
Models
EvoGrad: A Dynamic Take on the Winograd Schema Challenge with Human
Adversaries
Jing Han Sun, Ali Emami
---
University of Montreal/Mila, Brock University
Montreal, Canada, Saint Catharines, Canada
<EMAIL_ADDRESS><EMAIL_ADDRESS>
Abstract content
## 1\. Introduction
The Winograd Schema Challenge (WSC), a co-reference resolution task, was
developed to gauge the common-sense reasoning of automated systems Winograd
(1972); Levesque et al. (2011). Given subtly varying sentence pairs, the task
is to correctly associate a pronoun with a noun, as illustrated below:
Tom told Ralph, “Check,” as he moved his bishop. (Answer: Tom) Tom told Ralph,
“Check,” as he took his bishop. (Answer: Ralph)
In these examples, chess knowledge informs our interpretation of the pronoun
his—either referring to Tom or Ralph—based on the action performed, either a
move or take. While humans find such tasks intuitive, they pose a challenge
for statistical models, especially when lacking exposure to basic rules or
common knowledge. Yet, recent developments of extensive common-sense reasoning
datasets and benchmarks have allowed LLMs to achieve near-human performance on
WSC variants Brown et al. (2020); Sakaguchi et al. (2020). This impressive
accomplishment raises the question: has the WSC, seen as a definitive
alternative to the Turing Test, been definitively “defeated” Kocijan et al.
(2022)?
Figure 1: Interface of EvoGrad at https://evograd.com
At the same time, evidence suggests that even slight alterations to a WSC task
can significantly undermine a model’s performance Jia and Liang (2017);
Trichelair et al. (2018, 2019); Balasubramanian et al. (2020); Lin et al.
(2020a); Elazar et al. (2021a). This instability may reflect a discrepancy
between current supervision paradigms and the dynamic nature of common sense
acquisition. It suggests the potential value of exploring various approaches,
including the human-and-model-in-the-loop concept, as part of a broader
strategy to address these challenges Nie et al. (2020); Kiela et al. (2021);
Lu et al. (2022).
Existing datasets, often curated by select scientific communities or
crowdsourcing platforms, may also unintentionally bias models toward certain
knowledge instances or values, which may not be universally shared. This
consideration underscores the need for diverse, dynamic, and inclusive
benchmarks in the journey towards systems equipped with generalized common
sense.
Consider the chess example mentioned earlier. While the original WSC sentences
test the model’s understanding of the game’s basic rules, perturbations can
further probe deeper nuances and potential biases:
Maria told Jane, “Your move,” as she adjusted her queen. (Answer: Maria) Maria
told Jane, “Your move,” as she glanced at her clock. (Answer: Jane)
In these variations, the emphasis shifts from the action performed on a chess
piece to the broader context of a timed chess match. Slight word changes can
dramatically alter the correct answer, exposing potential model biases or gaps
in understanding. Such perturbations, especially when generated by diverse
human contributors, ensure a broader and more comprehensive test of a model’s
common-sense reasoning capabilities.
In this paper, we propose a revisit to the WSC within the framework of human-
and-model-in-the-loop. We introduce EvoGrad, an open-source, user-centric
platform dedicated to the active generation and expansion of nuanced examples
for the WSC through human-in-the-loop interactions. Our work contributes three
primary advancements:
A novel data construction mechanism: We enhance the WSC with our unique
approach to human-adversarial perturbations, combining human creativity with
the efficiency of ChatGPT. This innovative union, along with our use of
Wordnet for synonym-based variation, led to a dataset expansion from 182 to
3691 instances, setting a new standard for dynamic, diverse, and high-quality
common-sense reasoning datasets. Notably, our evaluations highlight the
challenging nature of EvoGrad, revealing significant gaps in model abilities
when compared to human benchmarks.
A new metric for model stability: In response to the instability of
transformer-based models on WSC-like tasks Abdou et al. (2020), we introduce a
metric termed error depth. This measure, derived from our data construction
process, offers a quantifiable assessment of model stability. We advocate for
its inclusion in evaluation reports alongside accuracy-based metrics, which
could discourage the development of models that achieve high scores due to
incorrect reasoning.
Online platform for user contributions: Available at https://evograd.com111All
aspects of the website remain anonymous during the submission and review
process to maintain the integrity of the user-contributed data and ensure
unbiased evaluation., our platform encourages public participation in the
continuous expansion of the dataset. Users can modify existing task instances
and observe the predictions of a chosen LLM, fostering a more participatory
and immersive data construction process (Figure 1).
$s_{0}$$per_{2}(s_{0},\text{screamed})$$s_{1(2)}$$per_{4}(s_{1(2)},\text{Melissa})$$s_{2(2,4)}$$per_{5}(s_{1(2)},\text{although})$$s_{2(2,5)}$$per_{9}(s_{0},\text{annoying})$$s_{1(9)}$$per_{7}(s_{1(9)},\text{became})$$s_{2(7,9)}$
Figure 2: Evolution figure of the sentence “Kevin yelled at Jim because he was
so upset.” up to depth level 2.
## 2\. Related Work
### 2.1. WSC-based Datasets
The Winograd Schema Challenge (WSC) Levesque et al. (2011) inspired various
datasets for pronominal coreference resolution, each tackling specific
challenges in the WSC or model evaluations. Datasets like Winogrande Sakaguchi
et al. (2020) and KnowRef Emami et al. (2019) address the WSC’s size
constraints. WinoGender Rudinger et al. (2018), WinoBias Zhao et al. (2018),
and KnowRef-60k Emami et al. (2020) focus on model biases, while WinoWhy Zhang
et al. (2020) and WinoLogic He et al. (2021) target common sense deficiencies
in models. Some research efforts enhanced the original WSC task Wang et al.
(2018); Trichelair et al. (2018); Kocijan et al. (2019); Elazar et al.
(2021a); Zahraei and Emami (2024) and utilized crowd-sourcing for task
development Isaak and Michael (2019); Sakaguchi et al. (2020). While these
static datasets each offer distinct strengths, they often introduce challenges
that necessitate prolonged research and iterations. EvoGrad, on the other
hand, adopts a dynamic framework, allowing for swift adjustments and
refinements in response to emerging challenges.
### 2.2. Dynamic Datasets
Dynamic datasets, updated over time to present new challenges, have been
developed for various tasks Zellers et al. (2019); Lin et al. (2020b).
Adversarial frameworks, as seen in Adversarial SQuAD, SWAG, HellaSWAG, CODAH
and ANLI, exemplify this approach Jia and Liang (2017); Zellers et al. (2018,
2019); Chen et al. (2019); Nie et al. (2020). Techniques such as AFLite
address biases through adversarial filtering Le Bras et al. (2020), while
other methods use continuous learning or a human-model collaborative process
Lan et al. (2017); Yang et al. (2018); Wallace et al. (2019); Dinan et al.
(2019); Nie et al. (2020); Xu et al. (2021); Kiela et al. (2021). ANLI and
Dynabench are notable for their multi-round adversarial data collection Nie et
al. (2020); Kiela et al. (2021). EvoGrad, while aligning with the dynamic
dataset philosophy, specifically targets WSC-based tasks. It merges human-and-
model collaboration, continuous learning, and domain-specific insights for
evolutionary data creation, amplifying the depth and relevance of WSC
challenges to shed light on common-sense reasoning.
### 2.3. Data Augmentation Methods in NLP
Data augmentation techniques in NLP create new examples from existing ones,
obviating the need for novel data collection Shi et al. (2021); Feng et al.
(2021). These methods include token-level manipulation, text generation
restricted, soft data enhancement, and structure-aware data augmentation Wang
and Yang (2015); Bergmanis et al. (2017); Zhang et al. (2018); Xu et al.
(2016). Our approach, mainly a token-level manipulation technique, extends
beyond the substitution of words to include the addition and removal of
tokens, allowing more significant sentence transformations Zmigrod et al.
(2019); Lu et al. (2020); Shi et al. (2018). We also measure the depth of
changes (Section 3.6) relative to the original sentence, providing insights
into model stability as a function of perturbations.
### 2.4. Large Language Models in Data Augmentation and Annotation
Large language models have emerged as effective tools for NLP data
augmentation and annotation, often exceeding the performance of crowd-workers
in terms of efficiency and cost Gilardi et al. (2023). These models have been
shown to be effective in tasks such as zero-shot gender identification and
providing explanations for implicit hate speech Kuzman et al. (2023); Huang et
al. (2023). AugGPT, for instance, outperforms traditional text augmentation
methods in few-shot learning scenarios by rephrasing sentences Dai et al.
(2023). Similarly, ChatGPT has shown potential to simplify social computing
tasks by replicating human-like annotations Zhu et al. (2023). Building on
these insights, we introduce an enhanced data augmentation method that
encompasses token substitutions, additions, and removals, aiming to address
common-sense reasoning deficiencies in the WSC and related tasks.
## 3\. EvoGrad
### 3.1. Dataset Evolution by Perturbation
We adopt an evolutionary approach to dataset expansion, initiating the process
with randomly selected instances from the original Winograd Schema Challenge
(WSC273) Levesque et al. (2011) and Winogrande Sakaguchi et al. (2020), which
are correctly resolved by all evaluated models.
Our method introduces a one-word perturbation to each sentence, effectively
mutating it via substitution. We define a perturbation function $per_{j}(s,w)$
that replaces the token at index $j$ in sentence $s$ with the token $w$.
Though primarily substitution-based, this function can also facilitate the
addition or removal of words, denoted as $per_{j}(s,w_{j}+w)$ and
$per_{j}(s,\epsilon)$ respectively, with $\epsilon$ symbolizing an empty
string.
The function is generalized as follows:
$per_{j}(s_{k(i_{1},...,i_{k})},w)=s_{(k+1)(i_{1},...,i_{j},...,i_{k+1})}\\\
j\not\in\\{i_{1},...,i_{k}\\}\text{ \& }i_{1}<...<i_{k+1}$ (1)
In this equation, $s_{k(i_{1},...,i_{k})}$ signifies the $k$th perturbation on
the base sentence $s_{0}$, wherein tokens at indices $i_{1},...,i_{k}$ have
been modified from $s_{0}$ (Equation 1). The term $k$ denotes the ‘depth’ or
generation of the sentence.
The conditions set for $j$ and indices $i_{1},...,i_{k+1}$ ensure that a depth
increment corresponds solely to the perturbation of a token distinct from
those previously perturbed (i.e.,${i_{1},...,i_{k}}$). Although repeated
modifications at the same token position are not prohibited, such sentences
maintain their original depths. This approach follows our depth
interpretation, emphasizing model stability against sentences that are
increasingly divergent from the original. This methodological choice
facilitates the systematic generation of progressively varied sentences,
thereby enriching the dataset.
The perturbation function is applied iteratively, generating a cascade of
output instances from each input instance. This process is illustrated in
Figure 2 by the sentence ‘Kevin yelled at Jim because he was so upset.’
Through several iterations of the perturbation function, we generate a wide
spectrum of sentences, each incrementally divergent from the original.
### 3.2. Scaling with ChatGPT
Beyond user contributions, we strategically employed
ChatGPT222https://chat.openai.com/ to vastly expand our dataset. We
initialized the process with 14 seed sentences (7 from WSC273 and 7 from
Winogrande-valid) and designed an elaborate prompt that enabled ChatGPT to act
as an ‘expert human annotator’. The prompts were meticulously crafted to guide
the model generation process via demonstrative examples and called for
frequent self-reflection to ensure the quality of the output. One unique
aspect of these prompts was the incorporation of a segmented generation
process, interspersed with feedback to ensure quality control and continuous
self-assessment. For each instance, we verified semantic coherence and
implemented a validation step to ensure pronouns and co-references matched
commonly accepted or typical human readings. An illustrative dialogue sample
can be found in the Appendix in section A.1.
This rigorous approach to prompt engineering culminated in the generation of
approximately 100 new instances per seed sentence. We further diversified
these generated sentences by modifying words, altering the correct antecedent,
and varying the total perturbation depth from the original sentences. This
strategy effectively harnessed the power of human creativity and the
scalability of the model to significantly expand our dataset. As a result, we
managed to augment our initial 182-instance dataset to a much more extensive
collection of 1,414 sentences, thereby facilitating a more comprehensive
evaluation of model performance on dynamic WSC tasks.
### 3.3. Scaling with Wordnet
To increase the diversity of our dataset, we utilized Wordnet Fellbaum (2010),
a lexical database, to augment the 1,414 sentences obtained from our ChatGPT
Scaling stage. This process enabled us to nearly triple our dataset size to a
final count of 3,691 sentences.
Our strategy was to introduce variability while preserving the context of the
sentence and grammatical accuracy. We achieved this by iterating over each
sentence and randomly selecting a word—excluding stop words and named
entities—for replacement. Once a word was selected, a random synonym from
Wordnet was chosen as its substitute. In cases where the chosen word was a
verb, we ensured that the replacement synonym matched the tense of the
original verb.
This approach allowed us to maintain the integrity of our original dataset
while significantly enhancing its size and complexity. The resulting sentences
provided a rich basis for model testing, aiding in the generation of a more
diverse and nuanced set of pronoun disambiguation scenarios.
Dataset | Sub | Size | Method
---|---|---|---
EvoGrad-S | - | 182 | Human (14 orig.)
EvoGrad-M | Train | 1010 | ChatGPT (1-10)
| Val | 202 | ChatGPT (11-12)
| Test | 202 | ChatGPT (13-14)
EvoGrad-L | Train | 2963 | WordNet (M Train)
| Val | 526 | WordNet (M Val)
| Test | 202 | ChatGPT (13-14)
Table 1: Summary of EvoGrad Allocation Source | Sentence | Answer | Depth
---|---|---|---
Original (WSC) | I poured water from the bottle into the cup until _ was full. | cup | 0
Human-perturbed | I poured water from the bottle into the cup because _ was empty. | cup | 2
ChatGPT-scaled | I poured water from the bottle, filling the cup until _ was empty. | bottle | 4
Wordnet-scaled | I decanted water from the feeding bottle into the cup until _ was empty. | feeding bottle | 4
Table 2: Sample instances of EvoGrad derived from an original WSC sentence,
showcasing the different methods of sentence generation and perturbation.
### 3.4. The Dataset
Table 1 outlines the construction and allocation process for our datasets,
specifically EvoGrad-small (S), EvoGrad-medium (M) and EvoGrad-large (L). The
initial dataset, EvoGrad-S, comprised 182 instances, all of which were
adaptations induced by humans from an original set of 14 sentences.
Subsequently, we generated the EvoGrad-M dataset, which was divided into three
distinct subsets: ‘train’, ‘val’, and ‘test’. These subsets were created by
perturbing the original sentences using ChatGPT, resulting in a total of 1,414
instances.
Finally, our most extensive dataset, EvoGrad-L, was constructed by augmenting
both the ‘train’ and ‘val’ subsets of EvoGrad-M using Wordnet, leading to an
overall count of 3,691 instances. The ‘test’ subset was retained from the
EvoGrad-M ‘test’ dataset and was generated through further perturbation of
EvoGrad-S sentences via ChatGPT. To illustrate the range of perturbations and
their sources, we provide sample instances in Table 2 derived from an original
WSC sentence.
### 3.5. The Platform
To foster collaborative development of EvoGrad, we have developed an
interactive platform, accessible at https://evograd.com. Here, global users
can actively contribute to the dataset’s evolution by modifying existing
sentences.
In the Build dataset page, users can select an original or perturbed sentence
from a drop-down menu labeled Original Sentence. They are then guided to input
a modified version of this sentence, replacing the target pronoun with an
underscore, in the New Sentence field. Following the Winogrande format
Sakaguchi et al. (2020), users also provide the two potential noun antecedents
in the Option 1 and Option 2 fields, specifying the correct answer.
To enhance user engagement, our platform offers immediate feedback. Users can
choose an LLM from a list - including BERT Devlin et al. (2019), RoBERTa Liu
et al. (2019), and Albert Lan et al. (2020)—and observe the model’s live
prediction. By clicking Submit, this prediction is generated, and the newly
provided data is incorporated into the dataset.
We prioritize transparency by allowing the dataset, stored as a CSV file, to
be downloaded and inspected directly from the platform. To ensure the quality
and appropriateness of the submissions, we manually validate all entries.
Users are further supported with examples and guidelines. A glimpse of the
platform’s interface is depicted in Figure 1.
Original sentence: Although she was being prosecuted, Monica was welcomed into
the sanctuary of the church by Samantha because _was a sinful criminal.
---
Perturbed Sentence | Prediction | True Label | Depth
Although she was being prosecuted, Monica was welcomed into the sanctuary of the church by Samantha because _was a guilty criminal. | Monica | Monica | 1 ✓
Although she was being prosecuted, Monica was welcomed into the sanctuary of the church by Samantha because _was a compassionate person. | Samantha | Samantha | 2 ✓
Even though she was being prosecuted, Monica was guided into the safe haven of the church by Samantha because _was a virtuous person. | Monica | Samantha | 5 ✗
While under prosecution, Monica was brought into the spiritual refuge of the church by Samantha because _was a good-natured woman. | Monica | Samantha | 6 ✗
While being prosecuted, Monica was welcomed into the church’s refuge by Samantha because _was a law-abiding person. | Monica | Samantha | 5 ✗
Table 3: Sample of perturbations constructed from Eq.1 on a Winogrande
example, with predictions corresponding to RoBERTa fine-tuned on Winogrande-
XL. The model’s incorrect predictions occur at depths 5,6 and 5, respectively,
corresponding to the number of modified tokens from the original. Therefore,
this sample of 5 perturbed instances has an average error depth (ED) of 5.333.
### 3.6. Error Depth
Given our dataset construction methodology, we propose the error depth (ED)
metric to evaluate model stability. While accuracy is a widely used metric to
gauge model performance on prediction tasks such as the WSC, it might not
effectively capture a model’s resilience against instances that progressively
deviate from the original.
There are scenarios where models predict correctly but possibly for the wrong
reasons. Sole reliance on accuracy can obscure these nuances. Ideally, a model
should demonstrate stability against token substitutions. Although, in the
context of the WSC, a token change can alter the answer label, a truly robust
model should not be overly sensitive to such modifications.
The error depth metric quantifies a model’s performance on sentences that
increasingly diverge from a correctly understood original. Specifically, the
error depth denotes the number of perturbations made to the original sentence
before the model produces its first incorrect prediction.
For clarity, let’s define the symbols:
* •
$s_{0}$: The original seed sentence.
* •
$\text{label}(s)$: The true label of sentence $s$.
* •
$\text{pred}(s)$: The model’s predicted label for sentence $s$.
* •
$n_{wrong}$: The number of incorrect predictions made by the model on
perturbed versions of the original sentence.
With these definitions, the error depth (ED) is formulated as:
$\displaystyle\overline{ED}$
$\displaystyle\overset{def}{=}\frac{1}{n_{wrong}}\sum_{k}^{n_{wrong}}k$ (2)
$\displaystyle\text{if }\text{label}(s_{0})=\text{pred}(s_{0})\text{ and }$
$\displaystyle\text{label}(s_{k(i_{1},...,i_{k})})\neq\text{pred}(s_{k(i_{1},...,i_{k})})$
Refer to Table 3 for an application of the metric to perturbations of a
sentence. In this demonstration, the model mispredicts three sentences: two
after five perturbations and one after six. Thus,
$\overline{ED}=(5+5+6)/3=5.333$. The error depth functions as an instance-
level metric, assessing a model’s stability for individual sentences.
Averaging over all instances yields $\overline{ED}$, which, when paired with
accuracy, offers a comprehensive assessment of a model’s performance on tasks
like the WSC.
### 3.7. Human Performance
Three English-proficient annotators reviewed EvoGrad-M Val and EvoGrad-L Val,
achieving mean accuracies of 95.2% and 92.8%, respectively. Importantly, they
did not exhibit an average error depth, effectively handling perturbations to
the full depth of the dataset. A high inter-annotator agreement was recorded
with a Fleiss’ Kappa of $\kappa=0.914$.
Model | Tuning | Wino-valid | EvoGrad-M-val | EvoGrad-L-val
---|---|---|---|---
BERT | EvoGrad-M | - | 60.4 (6.913) | -
EvoGrad-L | - | - | 54.9 (6.867)
Wino | 62.75 | —– (7.302) | —– (7.258)
Wino + EvoGrad-M | 63.06 | —– (7.308) | -
Wino + EvoGrad-L | 62.98 | - | —– (7.232)
RoBERTa | EvoGrad-M | - | 58.4 (6.762) | -
EvoGrad-L | - | - | 60.3 (6.727)
Wino | 76.09 | —– (6.286) | 6.393
Wino + EvoGrad-M | 76.09 | —– (6.286) | -
Wino + EvoGrad-L | 76.64 | - | 6.652
ALBERT | EvoGrad-M | - | 55.4 (6.989) | -
EvoGrad-L | - | - | 57.2 (6.853)
Wino | 64.64 | —– (7.971) | —– (7.670)
Wino + EvoGrad-M | 64.48 | —– (8.000) | -
Wino + EvoGrad-L | 64.64 | - | —– (7.694)
GPT-3* | EvoGrad-M | - | 59.41 (7.122) | -
EvoGrad-L | - | - | 56.08 (6.753)
GPT-3.5* | EvoGrad-M | - | 67.33 (7.061) | -
EvoGrad-L | - | - | 65.02 (7.245)
Table 4: Accuracy (and error depth) results of models on Winogrande-valid and
EvoGrad-val sets after training on Winogrande-XL and/or EvoGrad-train. Bold
values represent the highest accuracy and underlined values represent the
highest error depth for each model in each dataset. A single dash (-) denotes
that the model was not tuned on that specific dataset variant, hence was not
tested. Dashed (—–) values indicate that accuracy was not tested due to
potential contamination from EvoGrad’s seed examples being taken from
Winogrande, though error depth was still evaluated. Models marked with an
asterisk (*) were evaluated using few-shot learning rather than fine-tuning.
## 4\. Experiments and Results
### 4.1. Model Setup
We evaluated three primary transformer-based models that are masked language
models: BERT Devlin et al. (2019), RoBERTa Liu et al. (2019), and ALBERT Lan
et al. (2020). These models have been recognized for their strong performance
on the WSC and have led the benchmark results. Each of the models were fine-
tuned on the Winogrande-XL dataset Sakaguchi et al. (2020), which contains
approximately 40,000 task instances and is designed to reduce potential
annotation biases.
Additionally, we evaluated two left-to-right language models, specifically
GPT-3 (text-davinci-003) Brown et al. (2020) and GPT-3.5 (gpt-3.5-turbo-0613),
on the Winogrande-XL and EvoGrad datasets.
For BERT and RoBERTa, we first aimed to replicate top-performing models from
existing literature. Using Huggingface’s package Wolf et al. (2020), we
achieved validation accuracies of 62.75% for BERT-large-uncased and 76.09% for
RoBERTa-large. Although slightly below the reported accuracies in Sakaguchi et
al. (2020), variations in hyperparameter tuning may account for the
differences. A similar approach was taken for ALBERT-large-v2, with a
resulting accuracy of 64.64%.
Hyperparameters for BERT, RoBERTa, and ALBERT were selected from:
* •
Learning rates: $1e-5$, $3e-5$, $5e-5$
* •
Epochs: 3, 4, 5, 8
* •
Batch sizes: 8, 16
For training on EvoGrad-train (both medium and large versions), given its
resemblance but smaller size to Winogrande, we experimented with:
* •
Learning rates: $1e-5$, $3e-5$, $5e-5$
* •
Epochs: 1, 2, 4, 8
* •
Batch sizes: 8, 16, 32, 64
For evaluations using GPT-based models, we adopted a few-shot learning
approach. Each instance was evaluated using an instruction-based prompt
consisting of 30 random instances from the respective training set.
Figure 3: Distribution of the top three perturbations for models trained on
Winogrande + Evograd-L. From left to right: BERT, RoBERTa, and ALBERT.The
segments represent the relative frequency of each perturbation type: ‘–NN’
(noun removal), ‘+NN’ (noun addition), and either ‘–JJ’ (adjective removal) or
‘–IN’ (preposition removal).
Model | Trained on | EvoGrad-M-val | EvoGrad-L-val
---|---|---|---
BERT | EvoGrad-M | +NN (150), –NN (148), –JJ (105) | -
| EvoGrad-L | - | –NN (578), +NN (471), –JJ (342)
| Wino | –NN (92), –JJ (62), +NN (61) | –NN (365), +NN (294), –JJ (228)
| Wino + EvoGrad-M | –NN (108), +NN (90), –JJ (78) | -
| Wino + EvoGrad-L | - | –NN (373), +NN (303), –JJ (233)
RoBERTa | EvoGrad-M | –NN (170), +NN (146), –JJ (120) | -
| EvoGrad-L | - | –NN (494), +NN (416), –JJ (283)
| Wino | –NN (17), +NN (12), –JJ (11) | –NN (76), +NN (54), –JJ (41)
| Wino + EvoGrad-M | –NN (17), +NN (12), –JJ (11) | -
| Wino + EvoGrad-L | - | –NN (61), +NN (43), –IN (32)
ALBERT | EvoGrad-M | –NN (189), +NN (161), –JJ (131) | -
| EvoGrad-L | - | –NN (542), +NN (479), –JJ (316)
| Wino | –NN (92), –JJ (62), +NN (61) | –NN (272), +NN (208), –JJ (169)
| Wino + EvoGrad-M | –NN (92), –JJ (62), +NN (61) | -
| Wino + EvoGrad-L | - | –NN (294), +NN (220), –JJ (183)
GPT-3 | EvoGrad-M | –NN (173), +NN (144), –JJ (118) | -
| EvoGrad-L | - | –NN (505), +NN (448), –JJ (306)
GPT-3.5 | EvoGrad-M | –NN (161), –JJ (115), +NN(111) | -
| EvoGrad-L | - | –NN (464), +NN (364), –JJ (290)
Table 5: Top 3 perturbations and their count on incorrect predictions on
EvoGrad-val sets after fine-tuning on Winogrande-XL and EvoGrad-train.
### 4.2. Results
Our evaluation results, as shown in Tables 4 and Figure 3, offer insight into
model performance under different training conditions. We trained models
exclusively on EvoGrad-train, on Winogrande-XL (denoted as Wino), or
sequentially on both Winogrande and EvoGrad-train (denoted as Wino + EvoGrad).
This approach allowed us to understand how different training datasets
influence model robustness and stability.
Table 4 displays the models’ accuracies on the Winogrande-valid dataset
alongside their average error depth on the EvoGrad datasets. The error depth
indicates the perturbative distance at which a model starts to fail, providing
insights into model stability. While accuracy is the main metric, error depth
(shown in parentheses) gives a complementary view of model performance. Due to
the potential overlap between EvoGrad and Winogrande, we have omitted the
accuracy scores for Winogrande-trained models in EvoGrad. GPT-based models
were only evaluated on EvoGrad instances as they are evaluated through few-
shot learning.
Figure 3 visualizes the three most frequent perturbation types that lead to
incorrect predictions by the models. Each perturbation is categorized by its
effect on parts of speech. For instance, “+NN (150)” indicates a noun was
added in 150 of the incorrect predictions. A comprehensive breakdown of the
perturbation counts and their types, spanning all parts of speech observed, is
provided in Table 5.
## 5\. Discussion
#### Influence of EvoGrad on Language Model Performance
Table 4 illustrates the varied impacts of EvoGrad on Transformer models,
leading to several key insights:
* •
BERT’s improved performance post-EvoGrad training underscores its ability to
integrate the dataset’s specific perturbations effectively. This adaptability
implies that BERT may be particularly effective for tasks requiring deeper
linguistic insight or sensitivity to subtle contextual changes.
* •
RoBERTa consistently performs well both before and after training EvoGrad,
showcasing its robustness. However, its lower error depth compared to its
accuracy points to a potential trade-off between performance and stability.
This observation underscores the need to balance generalization with stability
to perturbations.
* •
The negligible change in ALBERT’s performance across various training regimes
raises questions regarding the model’s saturation point and its alignment with
the dataset. This warrants further investigation of the limits of adaptability
for certain models.
* •
While GPT-based models, especially GPT-3.5, demonstrate competitive
performance, their error depths highlight challenges related to stability.
This trend suggests that some of the newer models might prioritize
adaptability at the expense of robustness.
Figure 3 sheds light on the areas where language models are most vulnerable,
particularly in handling noun and adjective modifications. Addressing these
specific challenges is imperative for the enhancement of common-sense
reasoning in future model iterations.
#### Robustness and Adaptability to New Tasks
One of the challenges in deep learning is ensuring that the models remain
adaptable and robust when exposed to new tasks or datasets. Whether through
fine-tuning or few-shot learning, a model’s ability to incorporate new
information without significant detriment to its original capabilities is
vital. In our experiments, the transformer models exhibited this adaptability,
particularly when introduced to EvoGrad. For instance, when models were fine-
tuned on EvoGrad, their performance on the Winogrande validation set generally
improved or remained consistent (Table 4), indicating that they did not lose
their grasp of previously acquired knowledge. However, GPT-based models,
through few-shot learning, demonstrated their versatility in quickly adapting
to new tasks without the need for extensive retraining. These observations
underscore the potential of current architectures in handling evolving
datasets and tasks, highlighting their robustness in diverse learning
scenarios.
#### Evolution and Community Involvement with EvoGrad
The current rendition of EvoGrad represents only the first phase in a series
of envisioned enhancements. As the platform matures, our goal is to achieve
multiple cycles of data augmentation, model training, and fine-tuning,
striving to foster a greater social impact in the AI domain. In making EvoGrad
accessible to a diverse audience, including those new to WSC-style challenges,
we have incorporated clear prompts and guidelines, drawing inspiration from
our initial work with the 182 instances in EvoGrad-small.
Looking ahead, we are also planning to expand the platform to incorporate
other foundational NLP tasks by integrating datasets such as OntoNotes 5.0 for
Named Entity Recognition (NER) Weischedel et al. (2012), Natural Questions
(NQ) Kwiatkowski et al. (2019) for Question Answering (QA), and the SemEval
tasks for Sentiment Analysis, thereby broadening the scope and utility of
EvoGrad.
Recognizing the scale at which EvoGrad could grow, we understand the crucial
role of user-driven validation. While our dedicated team of in-house
researchers currently curates the dataset to ensure its quality, we’re eager
to transition this role to our users in the near future. This strategy not
only offloads the validation responsibility but also promises a more dynamic,
participatory, and community-centric approach to refining LLMs.
## 6\. Conclusion
In this work, we introduced EvoGrad, a dynamic platform that extends the
Winograd Schema Challenge with a human-and-model-in-the-loop methodology. The
dataset, enriched through our platform, incorporates contributions from human
experts, language model expansions, and lexical resource utilization. We also
introduced the “error depth" metric as a novel means to assess model stability
in evolving tasks. While our evaluations showed potential benefits of using
the augmented data from EvoGrad across different training regimes, the
disparity between human and machine performance on this task underlines its
complexity and the ongoing challenges in enhancing common-sense reasoning in
LLMs.
## Ethics Statement
We are presenting our publicly-accessible platform to those outside the
scientific and crowd-sourcing communities. However, our platform is still
limited to those in society who have access to a mobile device/personal
computer and internet access; a large but underrepresented group of people in
the world do not. We therefore use our platform as only a first step towards
more inclusiveness, which we open to people outside the small community of
science and crowd-sourcing, but wish to be involved in efforts that will
include the underrepresented mentioned.
We also cannot assume that everyone’s foremost priorities involve contributing
towards such endeavours as ours – many members of society are currently in
turmoils of war, famine, or even indifference or aversion towards AI, which
all may amount to their non-involvement in projects related to ours.
Accordingly, the direction towards progress is best achieved outside the
laboratory; after all, if diversity and community involvement in the
development of tasks such as ours is as correlated to positive results in AI,
our efforts as researchers should also extend towards the education, well
being, and thriving of members in society, without which our goal of a global
task is never truly realized.
## Acknowledgements
This work was supported by the Natural Sciences and Engineering Research
Council of Canada (NSERC) and by the New Frontiers in Research Fund (NFRF).
## 7\. Bibliographical References
##
* Abdou et al. (2020) Mostafa Abdou, Vinit Ravishankar, Maria Barrett, Yonatan Belinkov, Desmond Elliott, and Anders Søgaard. 2020. The sensitivity of language models and humans to Winograd schema perturbations. In _Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics_ , pages 7590–7604, Online. Association for Computational Linguistics.
* Aho and Ullman (1972) Alfred V. Aho and Jeffrey D. Ullman. 1972. _The Theory of Parsing, Translation and Compiling_ , volume 1. Prentice-Hall, Englewood Cliffs, NJ.
* American Psychological Association (1983) American Psychological Association. 1983. _Publications Manual_. American Psychological Association, Washington, DC.
* Anderson et al. (2002) David P Anderson, Jeff Cobb, Eric Korpela, Matt Lebofsky, and Dan Werthimer. 2002\. Seti@ home: an experiment in public-resource computing. _Communications of the ACM_ , 45(11):56–61.
* Ando and Zhang (2005) Rie Kubota Ando and Tong Zhang. 2005. A framework for learning predictive structures from multiple tasks and unlabeled data. _Journal of Machine Learning Research_ , 6:1817–1853.
* Andrew and Gao (2007) Galen Andrew and Jianfeng Gao. 2007. Scalable training of $L_{1}$-regularized log-linear models. In _Proceedings of the 24th International Conference on Machine Learning_ , pages 33–40.
* Balasubramanian et al. (2020) Sriram Balasubramanian, Naman Jain, Gaurav Jindal, Abhijeet Awasthi, and Sunita Sarawagi. 2020. What’s in a name? are BERT named entity representations just as good for any other name? In _Proceedings of the 5th Workshop on Representation Learning for NLP_ , pages 205–214, Online. Association for Computational Linguistics.
* Bergmanis et al. (2017) Toms Bergmanis, Katharina Kann, Hinrich Schütze, and Sharon Goldwater. 2017\. Training data augmentation for low-resource morphological inflection. In _Proceedings of the CoNLL SIGMORPHON 2017 Shared Task: Universal Morphological Reinflection_ , pages 31–39, Vancouver. Association for Computational Linguistics.
* Bird and Klein (2009) Edward Loper Bird, Steven and Ewan Klein. 2009. _Natural Language Processing with Python_. O’Reilly Media Inc.
* Brabham (2013) Daren C. Brabham. 2013. _Crowdsourcing_. MIT Press.
* Brown et al. (2020) Tom Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared D Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, Sandhini Agarwal, Ariel Herbert-Voss, Gretchen Krueger, Tom Henighan, Rewon Child, Aditya Ramesh, Daniel Ziegler, Jeffrey Wu, Clemens Winter, Chris Hesse, Mark Chen, Eric Sigler, Mateusz Litwin, Scott Gray, Benjamin Chess, Jack Clark, Christopher Berner, Sam McCandlish, Alec Radford, Ilya Sutskever, and Dario Amodei. 2020. Language models are few-shot learners. _Advances in neural information processing systems_ , 33:1877–1901.
* Chandra et al. (1981) Ashok K. Chandra, Dexter C. Kozen, and Larry J. Stockmeyer. 1981. Alternation. _Journal of the Association for Computing Machinery_ , 28(1):114–133.
* Chen et al. (2019) Michael Chen, Mike D’Arcy, Alisa Liu, Jared Fernandez, and Doug Downey. 2019. CODAH: An adversarially-authored question answering dataset for common sense. In _Proceedings of the 3rd Workshop on Evaluating Vector Space Representations for NLP_ , pages 63–69, Minneapolis, USA. Association for Computational Linguistics.
* Chen and Liu (2018) Zhiyuan Chen and Bing Liu. 2018. Lifelong machine learning. _Synthesis Lectures on Artificial Intelligence and Machine Learning_ , 12(3):1–207.
* Cooley and Tukey (1965) James W. Cooley and John W. Tukey. 1965. An algorithm for the machine calculation of complex Fourier series. _Mathematics of Computation_ , 19(90):297–301.
* Crowston (2012) Kevin Crowston. 2012. Amazon mechanical turk: A research tool for organizations and information systems scholars. In _Shaping the future of ict research. methods and approaches_ , pages 210–221. Springer.
* Dai et al. (2023) Haixing Dai, Zhengliang Liu, Wenxiong Liao, Xiaoke Huang, Zihao Wu, Lin Zhao, Wei Liu, Ninghao Liu, Sheng Li, Dajiang Zhu, et al. 2023. Chataug: Leveraging chatgpt for text data augmentation. _arXiv preprint arXiv:2302.13007_.
* Devlin et al. (2019) Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of deep bidirectional transformers for language understanding. In _Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers)_ , pages 4171–4186, Minneapolis, Minnesota. Association for Computational Linguistics.
* Dinan et al. (2019) Emily Dinan, Samuel Humeau, Bharath Chintagunta, and Jason Weston. 2019. Build it break it fix it for dialogue safety: Robustness from adversarial human attack. In _Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP)_ , pages 4537–4546, Hong Kong, China. Association for Computational Linguistics.
* Elazar et al. (2021a) Yanai Elazar, Hongming Zhang, Yoav Goldberg, and Dan Roth. 2021a. Back to square one: Artifact detection, training and commonsense disentanglement in the Winograd schema. In _Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing_ , pages 10486–10500, Online and Punta Cana, Dominican Republic. Association for Computational Linguistics.
* Elazar et al. (2021b) Yanai Elazar, Hongming Zhang, Yoav Goldberg, and Dan Roth. 2021b. Back to square one: Artifact detection, training and commonsense disentanglement in the winograd schema. In _Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing_ , pages 10486–10500.
* Emami et al. (2020) Ali Emami, Kaheer Suleman, Adam Trischler, and Jackie Chi Kit Cheung. 2020. An analysis of dataset overlap on winograd-style tasks. In _Proceedings of the 28th International Conference on Computational Linguistics_ , pages 5855–5865.
* Emami et al. (2019) Ali Emami, Paul Trichelair, Adam Trischler, Kaheer Suleman, Hannes Schulz, and Jackie Chi Kit Cheung. 2019. The KnowRef coreference corpus: Removing gender and number cues for difficult pronominal anaphora resolution. In _Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics_ , pages 3952–3961, Florence, Italy. Association for Computational Linguistics.
* Fellbaum (2010) Christiane Fellbaum. 2010. Wordnet. In _Theory and applications of ontology: computer applications_ , pages 231–243. Springer.
* Feng et al. (2021) Steven Y. Feng, Varun Gangal, Jason Wei, Sarath Chandar, Soroush Vosoughi, Teruko Mitamura, and Eduard Hovy. 2021. A survey of data augmentation approaches for NLP. In _Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021_ , pages 968–988, Online. Association for Computational Linguistics.
* French (1999) Robert M French. 1999. Catastrophic forgetting in connectionist networks. _Trends in Cognitive Sciences_ , 3(4):128–135.
* Gepperth and Hammer (2016) Alexander Gepperth and Barbara Hammer. 2016. Incremental learning algorithms and applications. In _European symposium on artificial neural networks (ESANN)_.
* Gilardi et al. (2023) Fabrizio Gilardi, Meysam Alizadeh, and Maël Kubli. 2023. Chatgpt outperforms crowd-workers for text-annotation tasks. _arXiv preprint arXiv:2303.15056_.
* Gusfield (1997) Dan Gusfield. 1997. _Algorithms on Strings, Trees and Sequences_. Cambridge University Press, Cambridge, UK.
* He et al. (2021) Weinan He, Canming Huang, Yongmei Liu, and Xiaodan Zhu. 2021. WinoLogic: A zero-shot logic-based diagnostic dataset for Winograd Schema Challenge. In _Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing_ , pages 3779–3789, Online and Punta Cana, Dominican Republic. Association for Computational Linguistics.
* Hossain and Kauranen (2015) Mokter Hossain and Ilkka Kauranen. 2015. Crowdsourcing: a comprehensive literature review. _Strategic Outsourcing: An International Journal_.
* Huang et al. (2023) Fan Huang, Haewoon Kwak, and Jisun An. 2023. Is chatgpt better than human annotators? potential and limitations of chatgpt in explaining implicit hate speech. In _Companion Proceedings of the ACM Web Conference 2023_ , pages 294–297.
* Isaak and Michael (2019) Nicos Isaak and Loizos Michael. 2019. Winoflexi: A crowdsourcing platform for the development of winograd schemas. In _AI 2019: Advances in Artificial Intelligence_ , pages 289–302, Cham. Springer International Publishing.
* Isaak and Michael (2020) Nicos Isaak and Loizos Michael. 2020. Blending nlp and machine learning for the development of winograd schemas. In _Agents and Artificial Intelligence: 12th International Conference, ICAART 2020, Valletta, Malta, February 22–24, 2020, Revised Selected Papers_ , page 188–214, Berlin, Heidelberg. Springer-Verlag.
* Jia and Liang (2017) Robin Jia and Percy Liang. 2017. Adversarial examples for evaluating reading comprehension systems. In _Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing_ , pages 2021–2031, Copenhagen, Denmark. Association for Computational Linguistics.
* Kaushik et al. (2020) Divyansh Kaushik, Eduard Hovy, and Zachary C Lipton. 2020. Learning the difference that makes a difference with counterfactually augmented data. _International Conference on Learning Representations (ICLR)_.
* Kiela et al. (2021) Douwe Kiela, Max Bartolo, Yixin Nie, Divyansh Kaushik, Atticus Geiger, Zhengxuan Wu, Bertie Vidgen, Grusha Prasad, Amanpreet Singh, Pratik Ringshia, Zhiyi Ma, Tristan Thrush, Sebastian Riedel, Zeerak Waseem, Pontus Stenetorp, Robin Jia, Mohit Bansal, Christopher Potts, and Adina Williams. 2021. Dynabench: Rethinking benchmarking in NLP. In _Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies_ , pages 4110–4124, Online. Association for Computational Linguistics.
* Kirkpatrick et al. (2017) James Kirkpatrick, Razvan Pascanu, Neil Rabinowitz, Joel Veness, Guillaume Desjardins, Andrei A. Rusu, Kieran Milan, John Quan, Tiago Ramalho, Agnieszka Grabska-Barwinska, Demis Hassabis, Claudia Clopath, Dharshan Kumaran, and Raia Hadsell. 2017. Overcoming catastrophic forgetting in neural networks. _Proceedings of the National Academy of Sciences_ , 114(13):3521–3526.
* Kocijan et al. (2019) Vid Kocijan, Ana-Maria Cretu, Oana-Maria Camburu, Yordan Yordanov, and Thomas Lukasiewicz. 2019. A surprisingly robust trick for the Winograd schema challenge. In _Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics_ , pages 4837–4842, Florence, Italy. Association for Computational Linguistics.
* Kocijan et al. (2022) Vid Kocijan, Ernest Davis, Thomas Lukasiewicz, Gary Marcus, and Leora Morgenstern. 2022. The defeat of the winograd schema challenge. _arXiv preprint arXiv:2201.02387_.
* Kratzwald and Feuerriegel (2019) Bernhard Kratzwald and Stefan Feuerriegel. 2019. Learning from on-line user feedback in neural question answering on the web. In _The World Wide Web Conference_ , WWW ’19, page 906–916, New York, NY, USA. Association for Computing Machinery.
* Kuzman et al. (2023) Taja Kuzman, Igor Mozetic, and Nikola Ljubešic. 2023. Chatgpt: Beginning of an end of manual linguistic data annotation? use case of automatic genre identification. _ArXiv, abs/2303.03953_.
* Kwiatkowski et al. (2019) Tom Kwiatkowski, Jennimaria Palomaki, Olivia Redfield, Michael Collins, Ankur Parikh, Chris Alberti, Danielle Epstein, Illia Polosukhin, Jacob Devlin, Kenton Lee, et al. 2019. Natural questions: a benchmark for question answering research. _Transactions of the Association for Computational Linguistics_ , 7:453–466.
* Lan et al. (2017) Wuwei Lan, Siyu Qiu, Hua He, and Wei Xu. 2017. A continuously growing dataset of sentential paraphrases. In _Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing_ , pages 1224–1234, Copenhagen, Denmark. Association for Computational Linguistics.
* Lan et al. (2020) Zhenzhong Lan, Mingda Chen, Sebastian Goodman, Kevin Gimpel, Piyush Sharma, and Radu Soricut. 2020. Albert: A lite bert for self-supervised learning of language representations. In _International Conference on Learning Representations_.
* Le Bras et al. (2020) Ronan Le Bras, Swabha Swayamdipta, Chandra Bhagavatula, Rowan Zellers, Matthew Peters, Ashish Sabharwal, and Yejin Choi. 2020. Adversarial filters of dataset biases. In _International Conference on Machine Learning_ , pages 1078–1088. PMLR.
* Levesque et al. (2011) Hector Levesque, Ernest Davis, and Leora Morgenstern. 2011. The winograd schema challenge. In _AAAI Spring Symposium: Logical Formalizations of Commonsense Reasoning_.
* Lin et al. (2020a) Bill Yuchen Lin, Seyeon Lee, Rahul Khanna, and Xiang Ren. 2020a. Birds have four legs?! NumerSense: Probing Numerical Commonsense Knowledge of Pre-Trained Language Models. In _Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)_ , pages 6862–6868, Online. Association for Computational Linguistics.
* Lin et al. (2020b) Bill Yuchen Lin, Wangchunshu Zhou, Ming Shen, Pei Zhou, Chandra Bhagavatula, Yejin Choi, and Xiang Ren. 2020b. CommonGen: A constrained text generation challenge for generative commonsense reasoning. In _Findings of the Association for Computational Linguistics: EMNLP 2020_ , pages 1823–1840, Online. Association for Computational Linguistics.
* Liu et al. (2019) Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. 2019. Roberta: A robustly optimized bert pretraining approach. _arXiv preprint arXiv:1907.11692_.
* Lovón-Melgarejo et al. (2021) Jesús Lovón-Melgarejo, Laure Soulier, Karen Pinel-Sauvagnat, and Lynda Tamine. 2021. Studying catastrophic forgetting in neural ranking models. In _European Conference on Information Retrieval_ , pages 375–390. Springer.
* Lu et al. (2022) Jinghui Lu, Linyi Yang, Brian Namee, and Yue Zhang. 2022. A rationale-centric framework for human-in-the-loop machine learning. In _Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)_ , pages 6986–6996, Dublin, Ireland. Association for Computational Linguistics.
* Lu et al. (2020) Kaiji Lu, Piotr Mardziel, Fangjing Wu, Preetam Amancharla, and Anupam Datta. 2020\. Gender bias in neural natural language processing. In _Logic, Language, and Security_ , pages 189–202. Springer.
* Melo et al. (2019) Gabriela Melo, Vinicius Imaizumi, and Fábio Cozman. 2019. Winograd schemas in portuguese. In _Proceedings of 16th National Meeting on Artificial and Computational Intelligence_ , pages 787–798, Porto Alegre, RS, Brasil. SBC.
* Mitchell et al. (2015) Tom Mitchell, William Cohen, Estevam Hruschka, Partha Talukdar, Justin Betteridge, Andrew Carlson, Bhavana Dalvi Mishra, Matthew Gardner, Bryan Kisiel, Jayant Krishnamurthy, Ni Lao, Kathryn Mazaitis, Thahir Mohamed, Ndapa Nakashole, Emmanouil Platanios, Alan Ritter, Mehdi Samadi, Burr Settles, Richard Wang, Derry Wijaya, Abhinav Gupta, Xinlei Chen, Abulhair Saparov, Malcolm Greaves, and Joel Welling. 2015. Never-ending learning. _Proceedings of the AAAI Conference on Artificial Intelligence_ , 29(1).
* Morgenstern et al. (2016) Leora Morgenstern, Ernest Davis, and Charles L. Ortiz. 2016. Planning, executing, and evaluating the winograd schema challenge. _AI Magazine_ , 37(1):50–54.
* Nie et al. (2020) Yixin Nie, Adina Williams, Emily Dinan, Mohit Bansal, Jason Weston, and Douwe Kiela. 2020. Adversarial NLI: A new benchmark for natural language understanding. In _Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics_ , pages 4885–4901, Online. Association for Computational Linguistics.
* Rasooli and Tetreault (2015) Mohammad Sadegh Rasooli and Joel R. Tetreault. 2015. Yara parser: A fast and accurate dependency parser. _Computing Research Repository_ , arXiv:1503.06733. Version 2.
* Rudinger et al. (2018) Rachel Rudinger, Jason Naradowsky, Brian Leonard, and Benjamin Van Durme. 2018. Gender bias in coreference resolution. In _Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 2 (Short Papers)_ , pages 8–14, New Orleans, Louisiana. Association for Computational Linguistics.
* Sakaguchi et al. (2020) Keisuke Sakaguchi, Ronan Le Bras, Chandra Bhagavatula, and Yejin Choi. 2020. Winogrande: An adversarial winograd schema challenge at scale. In _Proceedings of the AAAI Conference on Artificial Intelligence_ , volume 34, pages 8732–8740.
* Shi et al. (2021) Haoyue Shi, Karen Livescu, and Kevin Gimpel. 2021. Substructure substitution: Structured data augmentation for NLP. In _Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021_ , pages 3494–3508, Online. Association for Computational Linguistics.
* Shi et al. (2018) Haoyue Shi, Jiayuan Mao, Tete Xiao, Yuning Jiang, and Jian Sun. 2018. Learning visually-grounded semantics from contrastive adversarial samples. In _Proceedings of the 27th International Conference on Computational Linguistics_ , pages 3715–3727, Santa Fe, New Mexico, USA. Association for Computational Linguistics.
* Silver et al. (2013) Daniel L Silver, Qiang Yang, and Lianghao Li. 2013. Lifelong machine learning systems: Beyond learning algorithms. In _2013 AAAI spring symposium series_.
* Trichelair et al. (2018) Paul Trichelair, Ali Emami, Jackie Chi Kit Cheung, Adam Trischler, Kaheer Suleman, and Fernando Diaz. 2018. On the evaluation of common-sense reasoning in natural language understanding. In _Critiquing and Correcting Trends in Machine Learning NeurIPS 2018 Workshop_.
* Trichelair et al. (2019) Paul Trichelair, Ali Emami, Adam Trischler, Kaheer Suleman, and Jackie Chi Kit Cheung. 2019. How reasonable are common-sense reasoning tasks: A case-study on the Winograd schema challenge and SWAG. In _Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP)_ , pages 3382–3387, Hong Kong, China. Association for Computational Linguistics.
* Turing (1950) Alan M. Turing. 1950. Computing machinery and intelligence. _Mind_.
* Vaswani et al. (2017) Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Ł ukasz Kaiser, and Illia Polosukhin. 2017. Attention is all you need. In _Advances in Neural Information Processing Systems_ , volume 30. Curran Associates, Inc.
* Wallace et al. (2019) Eric Wallace, Pedro Rodriguez, Shi Feng, Ikuya Yamada, and Jordan Boyd-Graber. 2019\. Trick me if you can: Human-in-the-loop generation of adversarial examples for question answering. _Transactions of the Association for Computational Linguistics_ , 7:387–401.
* Wang et al. (2018) Alex Wang, Amanpreet Singh, Julian Michael, Felix Hill, Omer Levy, and Samuel Bowman. 2018. GLUE: A multi-task benchmark and analysis platform for natural language understanding. In _Proceedings of the 2018 EMNLP Workshop BlackboxNLP: Analyzing and Interpreting Neural Networks for NLP_ , pages 353–355, Brussels, Belgium. Association for Computational Linguistics.
* Wang and Yang (2015) William Yang Wang and Diyi Yang. 2015. That’s so annoying!!!: A lexical and frame-semantic embedding based data augmentation approach to automatic categorization of annoying behaviors using #petpeeve tweets. In _Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing_ , pages 2557–2563, Lisbon, Portugal. Association for Computational Linguistics.
* Weischedel et al. (2012) R Weischedel, S Pradhan, L Ramshaw, J Kaufman, M Franchini, M El-Bachouti, N Xue, M Palmer, JD Hwang, C Bonial, et al. 2012. Ontonotes release 5.0. linguistic data consortium. Technical report, Philadelphia, Technical Report.
* Winograd (1972) Terry Winograd. 1972. Understanding natural language. _Cognitive Psychology_ , 3(1):1–191.
* Wolf et al. (2020) Thomas Wolf, Lysandre Debut, Victor Sanh, Julien Chaumond, Clement Delangue, Anthony Moi, Pierric Cistac, Tim Rault, Remi Louf, Morgan Funtowicz, Joe Davison, Sam Shleifer, Patrick von Platen, Clara Ma, Yacine Jernite, Julien Plu, Canwen Xu, Teven Le Scao, Sylvain Gugger, Mariama Drame, Quentin Lhoest, and Alexander Rush. 2020. Transformers: State-of-the-art natural language processing. In _Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations_ , pages 38–45, Online. Association for Computational Linguistics.
* Xu et al. (2021) Jing Xu, Da Ju, Margaret Li, Y-Lan Boureau, Jason Weston, and Emily Dinan. 2021\. Bot-adversarial dialogue for safe conversational agents. In _Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies_ , pages 2950–2968, Online. Association for Computational Linguistics.
* Xu et al. (2016) Yan Xu, Ran Jia, Lili Mou, Ge Li, Yunchuan Chen, Yangyang Lu, and Zhi Jin. 2016\. Improved relation classification by deep recurrent neural networks with data augmentation. In _Proceedings of COLING 2016, the 26th International Conference on Computational Linguistics: Technical Papers_ , pages 1461–1470, Osaka, Japan. The COLING 2016 Organizing Committee.
* Yang et al. (2018) Zhilin Yang, Peng Qi, Saizheng Zhang, Yoshua Bengio, William Cohen, Ruslan Salakhutdinov, and Christopher D. Manning. 2018. HotpotQA: A dataset for diverse, explainable multi-hop question answering. In _Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing_ , pages 2369–2380, Brussels, Belgium. Association for Computational Linguistics.
* Zahraei and Emami (2024) Pardis Sadat Zahraei and Ali Emami. 2024. Wsc+: Enhancing the winograd schema challenge using tree-of-experts. _arXiv preprint arXiv:2401.17703_.
* Zellers et al. (2018) Rowan Zellers, Yonatan Bisk, Roy Schwartz, and Yejin Choi. 2018. SWAG: A large-scale adversarial dataset for grounded commonsense inference. In _Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing_ , pages 93–104, Brussels, Belgium. Association for Computational Linguistics.
* Zellers et al. (2019) Rowan Zellers, Ari Holtzman, Yonatan Bisk, Ali Farhadi, and Yejin Choi. 2019. HellaSwag: Can a machine really finish your sentence? In _Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics_ , pages 4791–4800, Florence, Italy. Association for Computational Linguistics.
* Zhang et al. (2020) Hongming Zhang, Xinran Zhao, and Yangqiu Song. 2020. WinoWhy: A deep diagnosis of essential commonsense knowledge for answering Winograd schema challenge. In _Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics_ , pages 5736–5745, Online. Association for Computational Linguistics.
* Zhang et al. (2018) Hongyi Zhang, Moustapha Cisse, Yann N Dauphin, and David Lopez-Paz. 2018. mixup: Beyond empirical risk minimization. In _International Conference on Learning Representations_.
* Zhang et al. (2015) Xiang Zhang, Junbo Zhao, and Yann LeCun. 2015. Character-level convolutional networks for text classification. In _Advances in Neural Information Processing Systems_ , volume 28. Curran Associates, Inc.
* Zhao et al. (2018) Jieyu Zhao, Tianlu Wang, Mark Yatskar, Vicente Ordonez, and Kai-Wei Chang. 2018\. Gender bias in coreference resolution: Evaluation and debiasing methods. In _Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 2 (Short Papers)_ , pages 15–20, New Orleans, Louisiana. Association for Computational Linguistics.
* Zhu et al. (2023) Yiming Zhu, Peixian Zhang, Ehsan-Ul Haq, Pan Hui, and Gareth Tyson. 2023. Can chatgpt reproduce human-generated labels? a study of social computing tasks. _arXiv preprint arXiv:2304.10145_.
* Zmigrod et al. (2019) Ran Zmigrod, Sabrina J. Mielke, Hanna Wallach, and Ryan Cotterell. 2019. Counterfactual data augmentation for mitigating gender stereotypes in languages with rich morphology. In _Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics_ , pages 1651–1661, Florence, Italy. Association for Computational Linguistics.
## Appendix A Appendix
### A.1. ChatGPT Dialogue During Scaling Process
Author:
You are an expert linguist, human annotator and are familiar with the Winograd
Schema Challenge. Consider the sentence: Although they ran at about the same
speed, Sue beat Sally because _had such a good start. This sentence is in the
format of a Winograd schema, where two antecedents, in this case "Sue" and
"Sally" are competing for reference in the position, _. Common sense allows us
to conclude that one of the two antecedents is most likely, in this case, Sue.
I am intending on building a dataset composed of modifications to the above
sentence. Based on this original sentence, I need you to generate 100 new
sentences that are modified by either adding, replacing, or substituting words
in the original sentence. They should be organized in of the following format,
copied from a csv file :
index sentence option1 option2 answer distance
where the original sentence has the following form: 0 Although they ran at
about the same speed, Sue beat Sally because _had such a good start. Sue Sally
1 0
Here, index refers to the index number for the sentence, sentence refers to
the sentence itself (notice the necessity of the _as the position for one of
the two antecedents), option1 refers to one antecedent, option2 refers to the
other, answer refers to the correct antecedent that belongs in _– "1"
corresponds to option1, and "2" corresponds to option2. Notice, for this
original sentence, the answer is "1" because having a good start is a likely
explanation for someone (i.e., Sue) beating another in a race. Finally,
"distance" corresponds to the number of modifications made on the original
sentence (whose distance is, by definition, 0). It is equivalent to the "token
edit distance".
I want you to create a diverse range of new entries, that vary in terms of the
modifications, the total distance, as well as the correct answer.
For example, the sentence "Although they sprinted at about the same speed, Sue
beat Sally because _had such a good start." would have an answer of 1 and
distance of 1 (because ran was changed to sprinted, but the correct answer
still remains option1, which in this case is Sue), and the sentence "Although
they sprinted at about the same speed, Sue beat Sally although _had such a
good start." would have the answer of 2 and distance of 2, because the correct
answer is now "Sally", and there were two word modifications – the second one
here was "although" instead of "because", which flips the answer. Notice that
with the right change in words, the answer can flip – I want to encourage you
to find such minimal number of changes when you’re creating new entries. You
are also allowed to change the words in the options too (e.g., Sue and Sally).
I would like the dataset to have a similar number of answer=1 entries than
answer=2 entries.
Please generate the first 10, formatted as a table here. First, you may ask
any clarification questions to make sure you understand the task.
ChatGPT:
Thank you for the instructions. Here are the first 10 entries: […]
Author:
Excellent work. Please generate 30 more, but try to keep the distance lower.
Don’t forget you can change the names too!
ChatGPT:
Sure! Here are the next 30 entries with smaller distances and changes in
names: […]
Author:
These sentences didn’t actually have the distances you claimed they had. These
should correspond to token edit distances. Please be careful when generating
sentences to actually correspond to distances – it’s okay if the distance is a
bit higher. I just want you to be accurate. You also forgot to change the
names (as possible modifications) at one point. Please generate 10 for me to
see if it’s okay.
ChatGPT:
I apologize for the inaccuracies in the distances and missing name changes in
the previous examples. […] Here are the next 10 sentences: […]
Author:
These were much better. Please generate 30 more. Pay attention to not
miscalculate the distance.
ChatGPT:
I appreciate your feedback. Here are the next 30 sentences, with the correct
token edit distances and a careful attention to detail: […]
Author:
Great job. Please generate 30 more. Again, the token distance should be
correctly calculated. For example, in your sentence:
Even though they raced at the same speed, Sue beat Sally although _had a
powerful start.
You said the distance is 3. It’s actually 5 here, because Although was dropped
and replaced with Even (1), then "though" was added (2), then raced was
replaced by ran (3), then about was dropped (4), then good was replaced by
powerful (5).
ChatGPT:
I understand your concern and I apologize for the mistake. I’ll ensure more
accuracy in calculating the token edit distance. Here are the next 30
sentences: […]
Author:
Great job. Please generate 20 more. Remember that changing the names is valid
but counts as a modification (increments distance).
ChatGPT:
Thank you for your feedback. Here are the next 20 sentences, this time
including changes in names: […]
##
* (1)
|
EXO-17-028
$HeadURL:svn+ssh://svn.cern.ch/reps/tdr2/papers/EXO-17-028/trunk/EXO-17-028.tex$
$Id:EXO-17-028.tex4665202018-06-2812:01:53Zalverson$
EXO-17-028
# Search for heavy Majorana neutrinos in same-sign dilepton channels in
proton-proton collisions at $\sqrt{s}=13\TeV$
###### Abstract
A search is performed for a heavy Majorana neutrino ($\mathrm{N}$), produced
by leptonic decay of a $\PW$ boson propagator and decaying into a $\PW$ boson
and a lepton, with the CMS detector at the LHC. The signature used in this
search consists of two same-sign leptons, in any flavor combination of
electrons and muons, and at least one jet. The data were collected during 2016
in proton-proton collisions at a center-of-mass energy of 13, corresponding to
an integrated luminosity of 35.9. The results are found to be consistent with
the expected standard model backgrounds. Upper limits are set in the mass
range between 20 and 1600in the context of a Type-I seesaw mechanism, on
$\abs{V_{\Pe\mathrm{N}}}^{2}$, $\abs{V_{\mu\mathrm{N}}}^{2}$, and
$\abs{V_{\Pe\mathrm{N}}V^{*}_{\mu\mathrm{N}}}^{2}/(\abs{V_{\Pe\mathrm{N}}}^{2}+\abs{V_{\mu\mathrm{N}}}^{2})$,
where $V_{\ell\mathrm{N}}$ is the matrix element describing the mixing of
$\mathrm{N}$ with the standard model neutrino of flavor $\ell=\Pe,\mu$. For
$\mathrm{N}$ masses between 20 and 1600, the upper limits on
$\abs{V_{\ell\mathrm{N}}}^{2}$ range between $2.3\times 10^{-5}$ and unity.
These are the most restrictive direct limits for heavy Majorana neutrino
masses above 430.
## 0.1 Introduction
The observation of neutrino oscillations [1], a mixing between several
neutrino flavors, established that at least two of the standard model (SM)
neutrinos have nonzero masses and that individual lepton number is violated.
The nonzero masses of the neutrinos are arguably the first evidence for
physics beyond the SM. Upper limits on the neutrino masses have been
established from cosmological observations [1], as well as direct
measurements, including those of tritium decays [2, 3]. The extremely small
values of these masses are difficult to explain in models that assume
neutrinos to be Dirac particles [4, 5].
The leading theoretical candidate to explain neutrino masses is the so-called
“seesaw” mechanism [6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19], in
which a new heavy Majorana neutrino $\mathrm{N}$ is postulated. In the seesaw
mechanism, the observed small neutrino masses, $m_{\nu}$, result from the
large mass of $\mathrm{N}$, with $m_{\nu}\sim
y^{2}_{\nu}v^{2}/m_{\mathrm{N}}$. Here $y_{\nu}$ is a Yukawa coupling, $v$ is
the Higgs vacuum expectation value in the SM, and $m_{\mathrm{N}}$ is the mass
of the heavy-neutrino state. One model that incorporates the seesaw mechanism,
and can be probed at the LHC, is the neutrino minimal standard model
($\nu$MSM) [20, 21, 22, 23]. In this model, the existence of new heavy
neutrinos could not only explain the very small masses of the SM neutrinos,
but also provide solutions to other problems in cosmology, such as the origin
of dark matter or the matter-antimatter asymmetry of the early universe [22,
23].
In this paper, we present the results of a search for a heavy Majorana
neutrino in the $\nu$MSM, which incorporates new heavy-neutrino states without
additional vector bosons. Searches for heavy Majorana neutrinos at hadron
colliders have been proposed by many theoretical models [24, 25, 26, 27, 28].
Numerous experiments have looked for heavy neutrinos in the mass range from
several keV to some hundred GeV, with no evidence seen, and a summary of the
limits on $\abs{V_{\ell\mathrm{N}}}^{2}$ versus $m_{\mathrm{N}}$ for these
experiments is given in Ref. [29], where $V_{\ell\mathrm{N}}$ is a matrix
element describing the mixing between the heavy neutrino and the SM neutrino
of flavor $\ell={\Pe},\;\mu$, or $\tau$. Direct searches for heavy neutrinos
have been performed at the CERN LEP collider [30, 31, 32] and, more recently,
at the CERN LHC [33, 34, 35, 36, 37]. These searches use a model-independent
phenomenological approach, assuming that $m_{\mathrm{N}}$ and
$V_{\ell\mathrm{N}}$ are free parameters.
The searches performed by the DELPHI [30] and L3 [31, 32] Collaborations at
LEP looked for the $\Pe^{+}\Pe^{-}\to\mathrm{N}\nu_{\ell}$ process, where
$\nu_{\ell}$ is any SM neutrino. For $\ell=\mu,\tau$ the limits on
$\abs{V_{\ell\mathrm{N}}}^{2}$ were set for $m_{\mathrm{N}}<90\GeV$, while for
$\ell=\Pe$ the limits extend to $m_{\mathrm{N}}<200\GeV$. Several experiments
obtained limits for low neutrino masses ($m_{\mathrm{N}}<5\GeV$), including
the LHCb Collaboration [33] at the LHC, which set limits on the mixing of a
heavy neutrino with an SM muon neutrino. The searches by L3, DELPHI, and LHCb
include the possibility of a finite heavy-neutrino lifetime, such that
$\mathrm{N}$ decays with a vertex displaced from the interaction point. In the
search reported here, however, it is assumed that $\mathrm{N}$ decays close to
the point of production, since in the mass range of this search
($m_{\mathrm{N}}>20\GeV$) the decay length is expected to be less than
$10^{-10}$m [38].
This search probes the decay of a $\PW$ boson, in which an SM neutrino
oscillates into a new state $\mathrm{N}$. In this analysis, only $\ell=\Pe$ or
$\mu$ processes are considered. In the previous CMS analyses [34, 35], only
the Drell–Yan (DY) production of $\mathrm{N}$
($\cPq\cPaq^{\prime}\to\PW^{*}\to\mathrm{N}\ell^{\pm}\to\ell^{\pm}\ell^{{}^{\prime}\pm}\cPq^{\prime}\cPaq$),
shown in Fig. 1 (left) was considered, while in this study the photon-
initiated production of $\mathrm{N}$
($\cPq\gamma\to\PW\cPq^{\prime\prime}\to\mathrm{N}\ell^{\pm}\cPq^{\prime\prime}\to\ell^{\pm}\ell^{\pm}\cPq^{\prime\prime}\cPq^{\prime}\cPaq$),
as shown in Fig. 1 (right), is also taken into account. The diagram in Fig. 1
(right) shows a possible production of $\mathrm{N}$ via $\PW\gamma$ fusion,
which we refer to by the generic term vector boson fusion (VBF). The inclusion
of the VBF channel enhances the sensitivity of this analysis for $\mathrm{N}$
masses above several hundred GeV [39], where the $t$-channel photon-initiated
processes become the dominant production mechanism for
$\PW^{*}\to\mathrm{N}\ell$ [40, 39].
Figure 1: Feynman diagram representing a resonant production of a Majorana
neutrino ($\mathrm{N}$), via the $s$-channel Drell–Yan process (left) and its
decay into a lepton and two quarks, resulting in a final state with two same-
sign leptons and two quarks from a $\PW$ boson decay. Feynman diagram for the
photon-initiated process (right).
Since $\mathrm{N}$ is a Majorana particle and can decay to a lepton of equal
or opposite charge to that of its parent $\PW$ boson, both opposite- and same-
sign (SS) lepton pairs can be produced. This search targets same-sign dilepton
(SS2$\ell$) signatures since these final states have very low SM backgrounds.
We search for events where the $\mathrm{N}$ decays to a lepton and a $\PW$
boson, and the $\PW$ boson decays hadronically, as this allows the
reconstruction of the mass of the $\mathrm{N}$ without the ambiguity
associated with the longitudinal momentum of an SM neutrino. For the DY
channel production, the final state is
$\ell^{+}\ell^{{}^{\prime}+}\cPq^{\prime}\cPaq$. The charge-conjugate decay
chain also contributes and results in an
$\ell^{-}\ell^{{}^{\prime}-}\cPaq^{\prime}\cPq$ final state. In the VBF
channel, production of an additional forward jet is produced in the event.
An observation of the
$\ell^{\pm}\ell^{{}^{\prime}\pm}\cPq^{\prime}\cPaq(\cPq^{\prime\prime})$
process would constitute direct evidence of lepton number violation. The study
of this process in different dilepton channels improves the likelihood for the
discovery of $\mathrm{N}$, and constrains the mixing matrix elements. The
dielectron ($\Pe\Pe$), dimuon ($\mu\mu$), and electron-muon ($\Pe\mu$)
channels are searched for and allow constraints to be set on
$\abs{V_{\Pe\mathrm{N}}}^{2}$, $\abs{V_{\mu\mathrm{N}}}^{2}$, and
$\abs{V_{\Pe\mathrm{N}}V^{*}_{\mu\mathrm{N}}}^{2}/(\abs{V_{\Pe\mathrm{N}}}^{2}+\abs{V_{\mu\mathrm{N}}}^{2})$,
respectively [38]. In the $\Pe\mu$ channel, the leptons from the $\PW$ boson
and the $\mathrm{N}$ decay can be either $\Pe$ and $\mu$, or $\mu$ and $\Pe$,
respectively, so the branching fraction for this channel is twice as large as
that for the $\Pe\Pe$ or $\mu\mu$ channels.
The most recent CMS search for heavy Majorana neutrinos in events with two
leptons and jets was performed for the mass range $m_{\mathrm{N}}=40$–500in
the $\Pe\Pe$, $\mu\mu$, and $\Pe\mu$ channels at $\sqrt{s}=8\TeV$ [34, 35]. A
similar search was also performed by the ATLAS Collaboration in the $\Pe\Pe$
and $\mu\mu$ channels [36]. The CMS Collaboration performed a search for heavy
Majorana neutrinos in final states with three leptons using the 2016 data set
[37], setting limits on $\abs{V_{\Pe\mathrm{N}}}^{2}$ and
$\abs{V_{\mu\mathrm{N}}}^{2}$, for the mass range $m_{\mathrm{N}}=1$–1200. In
the case of trilepton channels, events that contain both an electron and a
muon ($\Pe\Pe\mu,\mu\mu\Pe$) present an ambiguity about which of the leptons
mixes with $\mathrm{N}$, and it is thus impossible to probe
$\abs{V_{\Pe\mathrm{N}}V^{*}_{\mu\mathrm{N}}}^{2}/(\abs{V_{\Pe\mathrm{N}}}^{2}+\abs{V_{\mu\mathrm{N}}}^{2})$.
This ambiguity is not present in the current analysis with dilepton channels,
allowing limits to be set on
$\abs{V_{\Pe\mathrm{N}}V^{*}_{\mu\mathrm{N}}}^{2}/(\abs{V_{\Pe\mathrm{N}}}^{2}+\abs{V_{\mu\mathrm{N}}}^{2})$.
The CMS analysis at $\sqrt{s}=8\TeV$ showed that the efficiency for signal
events drops for masses above 400, as a consequence of the Lorentz-boosted
topology of the decay products of $\mathrm{N}$, which causes the signal jets
to overlap and be reconstructed as a single jet. The signal efficiency can be
recovered by including events containing a wide jet that is consistent with
the process $\PW\to\cPq\cPaq^{\prime}$, where the decay products of the $\PW$
boson are merged into a single jet [41]. It was also observed that the signal
efficiency dropped significantly when the mass of $\mathrm{N}$ was below the
$\PW$ boson mass ($m_{\mathrm{\PW}}$). For the $\mu\mu$ channel, the signal
acceptance was 0.65 (10.9)% for $m_{\mathrm{N}}=60\,(125)\GeV$. For
$m_{\mathrm{N}}<m_{\mathrm{\PW}}$ the final-state leptons and jets are very
soft and fail the momentum requirements applied in the 8analysis. In the
present analysis, cases where one of the signal jets fails the selection
criteria are recovered by including events with only one jet.
In this paper, a new search for $\mathrm{N}$ in the $\Pe\Pe$, $\mu\mu$, and
$\Pe\mu$ channels is presented using CMS data collected in 2016 at
$\sqrt{s}=13\TeV$. We search for events with two isolated leptons with the
same electric charge, with the presence of either a) two or more jets, b)
exactly one jet, or c) at least one wide jet. We look for an excess of events
above the expected SM background prediction by applying selection criteria to
the data to optimize the signal significance for each mass hypothesis. Heavy
Majorana neutrinos with a mass in the range of 20 to 1700are considered. There
are three potential sources of SS2$\ell$ backgrounds: SM sources in which two
prompt SS leptons are produced (a prompt lepton is defined as an electron or
muon originating from a $\PW$/$\PZ$/$\gamma^{*}$ boson or $\tau$ lepton
decay), events resulting from misidentified leptons, and opposite-sign
dilepton events (, from $\PZ\to\ell^{+}\ell^{-}$,
$\PW^{\pm}\PW^{\mp}\to\,\ell^{+}\nu\ell^{-}\overline{\nu}$) in which the sign
of one of the leptons is mismeasured. The last source is negligible for the
$\mu\mu$ and $\Pe\mu$ channels.
## 0.2 The CMS detector
The central feature of the CMS apparatus is a superconducting solenoid of 6m
internal diameter. The solenoid provides a magnetic field of 3.8T along the
direction of the counterclockwise rotating beam as viewed from above the plane
of the accelerator, taken as the $z$ axis of the detector coordinate system,
with the center of the detector defined to be at $z=0$. The azimuthal angle
$\phi$ is measured in radians in the plane perpendicular to the $z$ axis,
while the polar angle $\theta$ is measured with respect to this axis. Within
the solenoid volume are a silicon pixel and strip tracker, a lead tungstate
crystal electromagnetic calorimeter (ECAL), and a brass and scintillator
hadron calorimeter (HCAL), each composed of a barrel and two endcap sections.
The ECAL provides a coverage in pseudorapidity $\abs{\eta}<1.48$ in the barrel
region and $1.48<\abs{\eta}<3.00$ in the two endcap regions, where
pseudorapidity is defined as $\eta=-\ln[\tan(\theta/2)]$. Forward calorimetry
extends the pseudorapidity coverage provided by the barrel and endcap
detectors. Muons are detected in gas-ionization detectors, providing a
coverage of $\abs{\eta}<2.4$, and are embedded in the steel flux-return yoke
outside the solenoid. The first level of the CMS trigger system [42], composed
of custom hardware processors, uses information from the calorimeters and muon
detectors to select up to 100kHz of the most interesting events. The high-
level trigger (HLT) processor farm uses information from all CMS subdetectors
to further decrease the event rate to roughly 1kHz before data storage. A more
detailed description of the CMS detector can be found in Ref. [43].
## 0.3 Simulated samples
Samples of simulated events are used to estimate the background from SM
processes containing prompt SS leptons originating from hard-scattering
processes and to determine the heavy Majorana neutrino signal acceptance and
selection efficiency. The backgrounds from SM sources are produced using the
2.2.2 or 2.3.3 Monte Carlo (MC) generator [44] at leading order (LO) or next-
to-leading order (NLO) in perturbative quantum chromodynamics (QCD), with the
exception of $\Pg\Pg\to\PZ\PZ$ which is simulated at LO with 7.0 [45], and the
diboson production processes ($\PW\PZ$ and $\PZ\PZ$) that are generated at NLO
with the v2 [46, 47, 48, 49] generator.
The NNPDF3.0 [50] LO (NLO) parton distribution function (PDF) sets are used
for the simulated samples generated at LO (NLO). For all signal and background
samples, showering and hadronization are described using the 8.212 [51]
generator, with the CUETP8M1 [52] underlying event tune. The response of the
CMS detector is modeled using [53]. Double counting of the partons generated
with and is removed using the MLM [54] and FxFx [55] matching schemes in the
LO and NLO samples, respectively.
The $\mathrm{N}$ signals are generated using 2.6.0 at NLO precision, where the
decay of $\mathrm{N}$ is simulated with MadSpin [56], following the
implementation of Refs. [57, 58]. This includes the production of $\mathrm{N}$
via the charged-current DY and VBF processes. For the charged-current DY
production mechanism, we employ the NNPDF31_nnlo_hessian_pdfas PDF set [50],
while to include the photon PDF in the VBF ($\PW\gamma$ fusion) mechanism we
use the LUXqed17_plus_PDF4LHC15_nnlo_100 PDF set [59]. The NLO cross section,
obtained using the generator at $\sqrt{s}=13\TeV$, for the DY (VBF) process
has a value of 58.3 (0.050)pb for $m_{\mathrm{N}}=40\GeV$, dropping to 0.155
($9.65\times 10^{-4}$)pb for $m_{\mathrm{N}}=100\GeV$, and to $9.92\times
10^{-6}$ ($1.69\times 10^{-5}$)pb for $m_{\mathrm{N}}=1000\GeV$, assuming
$\abs{V_{\ell\mathrm{N}}}^{2}=0.01$. The VBF process becomes the dominant
production mode for scenarios where the mass of $\mathrm{N}$ is greater than
${\approx}700\GeV$. Only the final states with two leptons (electrons or
muons) and jets are generated.
Additional $\Pp\Pp$ collisions in the same or adjacent bunch crossings
(pileup) are taken into account by superimposing minimum bias interactions
simulated with on the hard-scattering process. The simulated events are
weighted such that the distribution of the number of additional pileup
interactions, estimated from the measured instantaneous luminosity for each
bunch crossing, matches that in data. The simulated events are processed with
the same reconstruction software as used for the data.
## 0.4 Event reconstruction and object identification
The reconstructed vertex with the largest value of summed physics-object
$\pt^{2}$ is taken to be the primary $\Pp\Pp$ interaction vertex, where is the
transverse momentum of the physics-objects. Here the physics objects are the
jets, clustered using the jet finding algorithm [60, 61] with the tracks
assigned to the vertex as inputs, and the associated missing transverse
momentum, , which is defined as the magnitude of the vector sum of the momenta
of all reconstructed particles in an event.
The global event reconstruction, based on the particle-flow algorithm [62],
aims to reconstruct and identify each individual particle in an event, with an
optimized combination of all subdetector information. In this process, the
identification of the particle type (photon, electron, muon, charged hadron,
neutral hadron) plays an important role in the determination of the particle
direction and energy. Photons are identified as ECAL energy clusters not
linked to the extrapolation of any charged-particle trajectory to the ECAL.
Electrons are identified as primary charged-particle tracks and potentially
many ECAL energy clusters corresponding to this track extrapolation to the
ECAL and to possible bremsstrahlung photons emitted along the way through the
tracker material. Muons are identified as tracks in the central tracker
consistent with either a track or several hits in the muon system, associated
with no significant associated energy deposits in the calorimeters. Charged
hadrons are identified as charged-particle tracks neither identified as
electrons, nor as muons. Finally, neutral hadrons are identified as HCAL
energy clusters not linked to any charged-hadron trajectory, or as ECAL and
HCAL energy excesses with respect to the expected charged-hadron energy
deposit.
The energy of photons is directly obtained from the ECAL measurement,
corrected for zero-suppression effects. The energy of electrons is determined
from a combination of the track momentum at the primary interaction vertex,
the corresponding ECAL cluster energy, and the energy sum of all
bremsstrahlung photons attached to the track. The energy of muons is obtained
from the corresponding track momentum. The energy of charged hadrons is
determined from a combination of the track momentum and the corresponding ECAL
and HCAL energy, corrected for zero-suppression effects and for the response
function of the calorimeters to hadronic showers. Finally, the energy of
neutral hadrons is obtained from the corresponding corrected ECAL and HCAL
energies.
### 0.4.1 Lepton selection
Electron candidates are selected in the region $\abs{\eta}<2.5$, excluding
$1.44<\abs{\eta}<1.57$. Their identification that is based on a multivariate
discriminant built from variables that characterize the shower shape and track
quality. To reject electrons originating from photon conversions in the
detector material, electrons must have no measurements missing in the
innermost layers of the tracking system and must not be matched to any
secondary vertex containing another electron [63]. To reduce the rate of the
electron sign mismeasurement, charges measured from independent techniques are
required to be the same, using the “selective method” for the charge
definition as explained in Ref. [63], which we refer to as “tight charge”.
Requiring the electrons to have tight charge reduces the signal efficiency by
1–20%, depending on $m_{\mathrm{N}}$, while the background from mismeasured
sign is reduced by a factor of 10. To ensure that electron candidates are
consistent with originating from the primary vertex, the transverse
(longitudinal) impact parameter of the leptons with respect to this vertex
must not exceed 0.1 (0.4). These electrons must also satisfy
$\abs{\mathrm{d}_{xy}}/\sigma(\mathrm{d}_{xy})<4$, where $\mathrm{d}_{xy}$ is
the transverse impact parameter relative to the primary vertex, estimated from
the track fit, and $\sigma(\mathrm{d}_{xy})$ is its uncertainty.
Muons are selected in the range $\abs{\eta}<2.4$. The muon trajectory is
required to be compatible with the primary vertex, and to have a sufficient
number of hits in the tracker and muon systems. The transverse (longitudinal)
impact parameter of the muons with respect to this vertex must not exceed 0.05
(0.40). These muons must also satisfy
$\abs{\mathrm{d}_{xy}}/\sigma(\mathrm{d}_{xy})<3$.
To distinguish between prompt leptons (a prompt lepton is defined as an
electron or muon originating in a $\PW$/$\PZ$/$\gamma^{*}$ boson or $\tau$
lepton decay) originating from decays of heavy particles, such as electroweak
(EW) bosons or heavy neutrinos, and those produced in hadron decays or hadrons
misidentified as leptons, a relative isolation variable
($I^{\ell}_{\text{rel}}$) is used. It is defined for electrons (muons) as the
pileup-corrected [63, 64] scalar $\pt$ sum of the reconstructed charged
hadrons originating from the primary vertex, the neutral hadrons, and the
photons, within a cone of $\Delta
R=\sqrt{\smash[b]{(\Delta\eta)^{2}+(\Delta\phi)^{2}}}=0.3\,(0.4)$ around the
lepton candidate’s direction at the vertex, divided by the lepton candidate’s
.
Electrons that pass all the aforementioned requirements and satisfy
$I^{\Pe}_{\text{rel}}<0.08$ are referred to as “tight electrons”. Electrons
that satisfy $I^{\Pe}_{\text{rel}}<0.4$, and pass less stringent requirements
on the multivariate discriminant and impact parameter are referred to as
“loose electrons”. Muons that pass all the aforementioned requirements and
satisfy $I^{\mu}_{\text{rel}}<0.07$ are referred to as “tight muons”. Muons
that satisfy $I^{\mu}_{\text{rel}}<0.6$, and pass a less stringent requirement
on the impact parameter and track quality requirements are referred to as
“loose muons”. Electrons within $\Delta R<0.05$ of a muon are removed, as
these particles are likely a photon radiated from the muon.
### 0.4.2 Identification of jets and missing transverse momentum
For each event, hadronic jets are clustered from the reconstructed particle-
flow objects with the infrared and collinear safe anti-jet clustering
algorithm [60], implemented in the package [65]. Two different distance
parameters, $\mathrm{R}=0.4$ and 0.8, are used with this algorithm, producing
objects referred to as AK4 and AK8 jets, respectively. The jet momentum is
determined as the vector sum of all particle momenta in the jet, and is found
from simulation to be within 5 to 10% of the true parton momentum over the
entire spectrum and detector acceptance. Additional $\Pp\Pp$ interactions
within the same or nearby bunch crossings can contribute additional tracks and
calorimetric energy depositions to the jet momentum. To mitigate this effect,
tracks identified to be originating from pileup vertices are discarded and an
offset correction is applied to correct for remaining contributions. Jet
energy corrections are derived from simulation to bring the measured response
of jets to that of particle level jets on average. In situ measurements of the
momentum balance in dijet, photon+jet, $\PZ\text{+}\text{jet}$, and multijet
events are used to estimate residual differences in jet energy scale in data
and simulation, and appropriate corrections are applied [66]. The jet energy
resolution is typically 15% at 10, 8% at 100, and 4% at 1. Additional
selection criteria are applied to remove jets potentially dominated by
anomalous contributions from various subdetector components or reconstruction
failures. The AK4 (AK8) jets must have $>20$ (200)and $\abs{\eta}<2.7$ to be
considered in the subsequent steps of the analysis. To suppress jets matched
to pileup vertices, AK4 jets must pass a selection based on the jet shape and
the number of associated tracks that point to non-primary vertices [67].
The AK8 jets are groomed using a jet pruning algorithm [68, 69]: subsequent to
the clustering of AK8 jets, their constituents are reclustered with the
Cambridge–Aachen algorithm [70, 71], where the reclustering sequence is
modified to remove soft and wide-angle particles or groups of particles. This
reclustering is controlled by a soft threshold parameter $z_{\text{cut}}$,
which is set to 0.1, and an angular separation threshold $\Delta
R>m_{\text{jet}}/p_{\text{T},\text{jet}}$. The jet pruning algorithm computes
the mass of the AK8 jet after removing the soft radiation to provide a better
mass resolution for jets, thus improving the signal sensitivity. The pruned
jet mass is defined as the invariant mass associated with the four-momentum of
the pruned jet.
In addition to the jet grooming algorithm, the “$N$-subjettiness” of jets [72]
is used to identify boosted vector bosons that decay hadronically. This
observable measures the distribution of jet constituents relative to candidate
subjet axes in order to quantify how well the jet can be divided into $N$
subjets. Subjet axes are determined by a one-pass optimization procedure that
minimizes $N$-subjettiness [72]. The separation in the phi-azimuth plane
between all of the jet constituents and their closest subjet axes are then
used to compute the $N$-subjettiness as
$\tau_{N}=1/d_{0}\Sigma_{k}p_{\mathrm{T},k}\text{min}(\Delta R_{1,k},\Delta
R_{2,k},...,\Delta R_{N,k})$ with the normalization factor
$d_{0}=\Sigma_{k}p_{\mathrm{T},k}R_{0}$ where $R_{0}$ is the clustering
parameter of the original jet, $p_{\mathrm{T},k}$ is the transverse momentum
of the $k$-constituent of the jet and $\Delta
R_{N,k}=\sqrt{\smash[b]{(\Delta\eta_{N,k})^{2}+(\Delta\phi_{N,k})^{2}}}$ is
its distance to the $N$-th subjet. In particular, the ratio between $\tau_{2}$
and $\tau_{1}$, known as $\tau_{21}$, has excellent capability for separating
jets originating from boosted vector bosons from jets originating from quarks
and gluons [72]. To select a high-purity sample of jets originating from a
hadronically decaying $\PW$ bosons, the AK8 jets are required to have
$\tau_{21}<0.6$ and a pruned jet mass between 40 and 130. We refer to these
selected jets as $\PW$-tagged jets. The efficiency of the $\tau_{21}$
selection for AK8 jets is measured in a -enriched sample in data and
simulation. To correct for observed differences between the estimated and
measured efficiencies a scale factor of $1.11\pm 0.08$ is applied to the event
for each AK8 jet that passes the $\tau_{21}$ requirement in the simulation
[67].
Identifying jets originating from a bottom quark can help suppress backgrounds
from production. To identify such jets the combined secondary vertex algorithm
[73] is used. This algorithm assigns to each jet a likelihood that it contains
a bottom hadron, using many discriminating variables, such as track impact
parameters, the properties of reconstructed decay vertices, and the presence
or absence of low-leptons. The average tagging efficiency for jets above 20is
63%, with an average misidentification probability for light-parton jets of
about 1%.
To avoid double counting due to jets matched geometrically with a lepton, any
AK8 jet that is within $\Delta R<1.0$ of a loose lepton is removed from the
event. Moreover, if an AK4 jet is reconstructed within $\Delta R<0.4$ of a
loose lepton or within $\Delta R<0.8$ of an AK8 jet, it is not used in the
analysis.
The is adjusted to account for the jet energy corrections applied to the event
[66]. The scalar sum of all activity in the event ($S_{\mathrm{T}}$) is used
in the selection of our signal region selection and is defined as the sum of
all AK4 and AK8 jets, leptons, and . The transverse mass, $m_{\text{T}}$,
which is used as a requirement to suppress backgrounds from leptonic $\PW$
boson decays, is defined as follows:
$m_{\text{T}}(\ell,\ptmiss)=\sqrt{\smash[b]{2\pt^{\ell}\ptmiss[1-\cos(\Delta\phi_{\ell,\ptvecmiss})]}},$
(1)
where $\pt^{\ell}$ is the transverse momentum of the lepton and
$\Delta\phi_{\ell,\ptvecmiss}$ is the azimuthal angle difference between the
lepton momentum and vector.
## 0.5 Event selection
Events used in this search are selected using several triggers, requiring the
presence of two charged leptons ($\Pe$ or $\mu$). All triggers require two
loosely isolated leptons, where the leading- (trailing-)lepton must have
$\pt>23\,(12)\GeV$ for the $\Pe\Pe$, $\pt>17\,(8)\GeV$ for the $\mu\mu$, and
$\pt>23\,(8)\GeV$ for the $\Pe\mu$ trigger at the HLT stage. The offline
requirements on the leading (trailing) lepton $\pt$ are governed by the
trigger thresholds, and are $\pt>25\,(15)\GeV$ for the $\Pe\Pe$,
$\pt>20\,(10)\GeV$ for the $\mu\mu$, and $\pt>25\,(10)\GeV$ for the $\Pe\mu$
channels. The efficiency for signal events to satisfy the trigger in the
$\Pe\Pe$, $\mu\mu$, and $\Pe\mu$ channels is above 0.88, 0.94, and 0.88,
respectively.
### 0.5.1 Preselection criteria
At a preselection stage, events are required to contain a pair of SS leptons.
To remove backgrounds with soft misidentified leptons, the invariant mass of
the dilepton pair is required to be above 10. Dielectron events with an
invariant mass within 20of the $\PZ$ boson mass [1] are excluded to reject
background from $\PZ$ boson decays in which one electron sign is mismeasured.
In order to suppress backgrounds from diboson production, such as $\PW\PZ$,
events with a third lepton identified using a looser set of requirements and
with $\pt>10\GeV$ are removed. Preselected events are required to have at
least one AK4 or one AK8 jet passing the full jet selection. The same
preselection is applied in all three channels ($\Pe\Pe$, $\mu\mu$, $\Pe\mu$).
### 0.5.2 Selection criteria for signals
The kinematic properties of signal events from heavy-neutrino decays depend on
its mass. To distinguish between the two $\PW$ bosons involved in the
production and decay sequence, we refer to the $\PW$ boson that produces
$\mathrm{N}$ in Fig. 1 (left) as the $\PW$ boson propagator and the $\PW$
boson that decays to a quark and anti-quark pair as the hadronically decaying
$\PW$ boson. Two search regions (SRs) are defined. In the low-mass SR
($m_{\mathrm{N}}\lesssim 80\GeV$), the $\PW$ boson propagator is on-shell and
the final-state system of dileptons and two jets should have an invariant mass
close to the $\PW$ boson mass. In the high-mass SR ($m_{\mathrm{N}}\gtrsim
80\GeV$), the $\PW$ boson propagator is off-shell but the hadronically
decaying $\PW$ boson is on-shell, so the invariant mass of the jets from the
hadronically decaying $\PW$ will be consistent with the $\PW$ boson mass.
Since the kinematic properties of the signal depend on $m_{\mathrm{N}}$, we
define four event categories to maximize the discovery potential over the full
mass range. The low- and high-mass SRs are further split based on the jet
configuration. The four signal categories used in the analysis are defined as:
* •
low- and high-mass SR1: number of AK4 jets $\geq$ 2 and number of AK8 jets =
0,
* •
low-mass SR2: number of AK4 jets = 1 and number of AK8 jets = 0,
* •
high-mass SR2: number of AK8 jets $\geq$ 1.
Taking the three flavor channels into account, the analysis has 12 separate
SRs.
In each SR, the technique of selecting jets associated with the hadronic $\PW$
boson decay is different. If there are any $\PW$-tagged AK8 jets in the event,
the AK8 jet with pruned jet mass closest to $m_{\mathrm{\PW}}$ is assumed to
be from the hadronic $\PW$ boson decay. For the high-mass SRs, if there are
two or more AK4 jets in the event and no AK8 jets, the two AK4 jets with the
invariant mass closest to $m_{\mathrm{\PW}}$ are assigned to the hadronically
decaying $\PW$ boson. In the low-mass SRs, the $\PW$ boson propagator is
reconstructed from $\mathrm{N}$ (one lepton + jet(s)) and the additional
lepton, and if there are more than two jets, the jets are selected such that
the mass is closest to $m_{\mathrm{\PW}}$. If only one jet is reconstructed in
the low-mass SR then this is assigned as being from the hadronic $\PW$ boson
decay. The jet(s) assigned to the hadronic $\PW$ boson decay are referred to
by the symbol $\PW_{\text{jet}}$ to simplify notation in the rest of the
paper.
Before optimizing the signal significance for each mass hypothesis we apply a
set of loose selections to the preselection events to select the low- and
high-mass SRs. These requirements are chosen to remove a large fraction of the
backgrounds while keeping the signal efficiency high. In the low-mass SRs, the
invariant mass of the two leptons and $\PW_{\text{jet}}$ is required to be
less than 300. To remove backgrounds from leptonic $\PW$ boson decays, events
must have less than 80. To remove backgrounds from top quark decays, events
are vetoed if they contain a -tagged AK4 jet. In the high-mass SRs, the
following selections are used. For SR1 the events are required to have
$30<m(\PW_{\text{jet}})<150\GeV$ for the invariant mass of the
$\PW_{\text{jet}}$ and $\pt>25\GeV$ for the leading AK4 jet. For SR2 the
pruned jet mass must satisfy $40<m(\PW_{\text{jet}})<130\GeV$. Since the is
correlated with the energy of the final-state objects, this requirement is not
used in high-mass SRs. Instead, we use $(\ptmiss)^{2}/S_{\mathrm{T}}$, which
has a stronger discriminating power between high-mass signals and backgrounds.
The $(\ptmiss)^{2}/S_{\mathrm{T}}$ must be less than 15. These selections are
summarized in Table 0.5.2.
Selection requirements, after applying the preselection criteria, for the low-
and high-mass signal regions. A dash indicates that the variable is not used
in the selection. Region $(\ptmiss)^{2}/S_{\mathrm{T}}$
$m(\ell^{\pm}\ell^{\pm}\PW_{\text{jet}})$ $m(\PW_{\text{jet}})$
$p_{\mathrm{T}}^{\mathrm{j}}$ () () () () () Low-mass SR1+SR2 $<$80 — $<$300 —
$>$20 High-mass SR1 — $<$15 — 30–150 $>$25 High-mass SR2 — $<$15 — 40–130
$>$200
#### Optimization of signal selection
After applying the selection criteria in Table 0.5.2, the signal significance
is optimized by combining several different variables using a modified Punzi
figure of merit [74]. The Punzi figure of merit is defined as
$\epsilon_{\mathrm{S}}/(a/2+\delta B)$ where $a$ is the number of standard
deviation, and is set equal to 2 to be consistent with the previous CMS
analysis, $\epsilon_{\mathrm{S}}$ is the signal selection efficiency, and
$\delta B$ is the uncertainty in the estimated background. The signal regions
are optimized separately for each mass hypothesis and for each of the three
flavor channels.
The variables used to optimize the signal selection, which are all optimized
simultaneously, are: the transverse momentum of the leading lepton
$\pt^{\ell_{1}}$, and of the trailing lepton $\pt^{\ell_{2}}$; the invariant
mass of the two leptons and the selected jet(s)
$m(\ell^{\pm}\ell^{\pm}\PW_{\text{jet}})$; the angular separation between the
$\PW_{\text{jet}}$ and the trailing lepton $\Delta
R(\ell_{2},\PW_{\text{jet}})$; minimum and maximum requirements on the
invariant mass of the lepton (leading or trailing) and the selected jet(s)
$m(\ell_{i}\PW_{\text{jet}})$, where $i$=1,2; and the invariant mass of the
two leptons $m(\ell^{\pm}\ell^{\pm})$. We consider the variable
$m(\ell_{i}\PW_{\text{jet}})$, as this should peak at $m_{\mathrm{N}}$ for the
signal. Since it is not known which lepton comes from the $\mathrm{N}$ decay,
the event is accepted if either $m(\ell_{i}\PW_{\text{jet}})$ satisfies the
requirements. The optimized window requirements for some SRs are enlarged to
give complete coverage of the signal parameter space at negligible loss of
sensitivity. The selection requirements for each mass hypothesis are
summarized later in Section 0.8, in Tables 0.8–0.8 for both low- and high-mass
SRs. The overall signal acceptance ranges between 0.10–0.27% and 17–33% for
$m_{\mathrm{N}}=20$–1500, respectively. Here, the lower acceptance at low
$m_{\mathrm{N}}$ is due to the selection requirements on the of the leptons
and jets in a signal with very soft jets and leptons. The overall signal
acceptance includes trigger efficiency, geometrical acceptance, and
efficiencies of all selection criteria.
## 0.6 Background estimate
The SM backgrounds leading to a final state with two SS leptons and jets are
divided into the following categories:
* •
SM processes with multiple prompt leptons: These backgrounds are mainly from
events with two vector bosons ($\PW^{\pm}\PW^{\pm}$, $\PW\PZ$, $\PZ\PZ$). We
also consider as background a $\PW$ or $\PZ$ boson decaying leptonically and
is accompanied by radiation of an initial- or final-state photon that
subsequently undergoes an asymmetric conversion. These processes produce a
final state that can have three or four leptons. If one or more of the charged
leptons fail the reconstruction or selection criteria these processes can
appear to have only two SS leptons.
* •
Misidentified leptons: These are processes that contain one or more leptons
that are either misidentified hadrons, are from heavy-flavor jets, from light
meson decays, or from a photon in a jet. These leptons are generally less
isolated than a prompt lepton from a $\PW/\PZ$ boson decay and tend to have
larger impact parameters. The main processes with a misidentified lepton in
the SRs include $\PW$+jets events and events, but multijet and DY events also
contribute.
* •
Sign mismeasurement: If the signs of leptons are mismeasured in events with
jets and two opposite-sign leptons (OS2$\ell$), these events could contaminate
a search region. When the sign of a lepton is mismeasured the lepton will on
average have a larger impact parameter in comparison to a lepton from a prompt
EW boson decay. Although the rate of mismeasuring the sign of an electron is
small, the abundance of OS2$\ell$ events from DY dilepton production means
that this background is significant. It is suppressed by tight requirements on
the impact parameter and on the charge of the electron. The muon sign
mismeasurement rate is known to be negligible, based on studies in simulation
and with cosmic ray muons [75], and is not considered in this analysis.
### 0.6.1 Background from prompt SS leptons
Background events that contain two prompt SS leptons are referred to as the
prompt-lepton background. These backgrounds are estimated using simulation. To
remove any double counting from the misidentified-lepton background estimate
based on control samples in data, the leptons have to originate in the decay
of either a $\PW/\PZ/\gamma^{*}$ boson, or a $\tau$ lepton. The largest
contribution comes from $\PW\PZ$, $\PZ\PZ$, and asymmetric photon conversions,
including those in $\PW\gamma$ and $\PZ\gamma$ events. The background from
$\PW\PZ$ and $\PW\gamma^{*}$ production, with $\PW\to\ell\nu$ and
$\PZ(\gamma^{(*)})\to\ell\ell$, can yield the same signature as $\mathrm{N}$
production: two SS isolated leptons and jets, when one of the opposite-sign
same-flavor (OSSF) leptons is not identified and QCD/pileup jets are
reconstructed in the event. This is the largest prompt contribution in both
the low- and high-mass SRs. This background is estimated from simulation, with
the simulated yield normalized to the data in a control region (CR) formed by
selecting three tight leptons with $\pt>25,15,10\GeV$ and requiring an OSSF
lepton pair with invariant mass $m(\ell^{\pm}\ell^{\mp})$ consistent with the
$\PZ$ boson mass: $\abs{m(\ell^{\pm}\ell^{\mp})-m_{\mathrm{\PZ}}}<15\GeV$. In
addition, events are required to have $\ptmiss>50\GeV$ and
$m_{\text{T}}(\ell_{\PW},\ptmiss)>20\GeV$, where the $\ell_{\PW}$ is the
lepton not used in the OSSF pair that is consistent with the $\PZ$ boson. The
ratio of the predicted to observed $\PW\PZ$ background yield in this CR is
found to be $1.051\pm 0.065$. This factor and its associated uncertainty (both
statistical and systematic) is used to normalize the corresponding simulated
sample. The systematic uncertainty on this factor is determined by varying, in
the simulation, the properties that are listed in Section 0.7.2, by $\pm 1$
standard deviation from its central value.
Production of $\PZ\PZ$ events with both $\PZ$ bosons decaying leptonically,
with two leptons not identified, results in a possible SS2$\ell$ signature.
This process is estimated from simulation, and the simulated yield is
normalized in a CR containing four leptons that form two OSSF lepton pairs
with invariant masses consistent with that of the $\PZ$ boson. The ratio of
data to simulation from the CR is found to be $0.979\pm 0.079$, and is used to
normalize the simulated $\PZ\PZ$ sample. A $\PZ$ boson -dependent EW
correction to the cross section [76, 77, 78] is not included in the simulated
samples. It would correct the cross section by at most 25%, given the range of
$\PZ$ boson probed in this analysis. Since this correction is larger than the
uncertainty on the ratio of data to simulation in the CR, we increase the
uncertainty on the normalization to 25%.
External and internal photon conversions can produce an SS2$\ell$ final state
when a photon is produced with a $\PW$ or $\PZ$ boson, and this photon
undergoes an asymmetric external or internal conversion
($\gamma^{(*)}\to\ell^{+}\ell^{-}$) in which one of the leptons has very low
and fails the lepton selection criteria. This background mostly contributes to
events in the $\Pe\Pe$ and $\Pe\mu$ channels. It is obtained from simulation
and verified in a data CR enriched in both external and internal conversions
from the $\PZ\text{+}\text{jets}$ process, with $\PZ\to\ell\ell\gamma^{(*)}$
and $\gamma^{(*)}\to\ell\ell$, where one of the leptons is outside the
detector acceptance. The CR is defined by
$\abs{m(\ell^{\pm}\ell^{\mp})-m_{\mathrm{\PZ}}}>15\GeV$ and
$\abs{m(\ell^{\pm}\ell^{\mp}\ell^{\pm})-m_{\mathrm{\PZ}}}<15\GeV$. The ratio
of data to expected background in the CR is $1.093\pm 0.075$, and this ratio
is used to normalize the MC simulation.
Other rare SM processes that can yield two SS leptons include events from EW
production of SS $\PW$ pairs, and double parton scattering, while any SM
process that yields three or more prompt leptons produces SS2$\ell$ final
states if one or more of the leptons fails to pass the selection. Processes in
the SM that can yield three or more prompt leptons include triboson processes
and production associated with a boson ($\ttbar\PW$, $\ttbar\PZ$, and
$\ttbar\PH$). Such processes generally have very small production rates (less
than 10% of total background after the preselection) and in some cases are
further suppressed by the veto on -tagged jets and requirements on . They are
estimated from simulation and assigned a conservative uncertainty of 50%,
which accounts for the uncertainties due to experimental effects, event
simulation, and theoretical calculations of the cross sections.
### 0.6.2 Background from misidentified leptons
The most important background source for low-mass signals originates from
events containing objects misidentified as prompt leptons. These originate
from hadron decays, light-quark or gluon jets, and are typically not well
isolated. Examples of these backgrounds include: multijet production, in which
one or more jets are misidentified as leptons; $\PW(\to\ell\nu)$+jets events,
in which one of the jets is misidentified as a lepton; and decays, in which
one of the top quark decays yields a prompt isolated lepton
$(\cPqt\to\PW\cPqb\to\ell\nu\cPqb)$ and the other lepton of same sign arises
from a bottom quark decay or a jet misidentified as an isolated prompt lepton.
The simulation is not reliable in estimating the misidentified-lepton
background for several reasons, including the lack of statistically large
samples (because of the small probability of a jet to be misidentified as a
lepton) and inadequate Modeling of the parton showering process. Therefore,
these backgrounds are estimated using control samples of collision data.
An independent data sample enriched in multijet events (the “measurement”
sample) is used to calculate the probability misidentifying a jet that passes
minimal lepton selection requirements (“loose leptons”) to also pass the more
stringent requirements used to define leptons after the full selection (“tight
leptons”). The misidentification probability is applied as an event-by-event
weight to the application sample. The application sample contains events in
which one lepton passes the tight selection, while the other lepton fails the
tight selection but passes the loose selection ($N_{n\overline{n}}$), as well
as events in which both leptons fail the tight selection, but pass the loose
criteria ($N_{\overline{n}\,\overline{n}}$). The total contribution to the
signal regions (, the number of events with both leptons passing the tight
selection, $N_{nn}$), is then obtained for each mass hypothesis by weighting
events of type $n\overline{n}$ and $\overline{n}\,\overline{n}$ by the
appropriate misidentification probability factors and applying the signal
selection requirements to the application sample. To account for the double
counting we correct for $\overline{n}\,\overline{n}$ events that can also be
$n\overline{n}$.
The measurement sample is selected by requiring a loose lepton and a jet,
resulting in events that are mostly dijet events, with one jet containing a
lepton. Only one lepton is allowed and requirements of $\ptmiss<80\GeV$, and
$m_{\text{T}}(\ell,\ptmiss)<25\GeV$ are applied. The loose lepton and jet are
required to be separated in azimuth by $\Delta\phi>2.5$ and the momentum of
the jet is required to be greater than the momentum of the lepton. These
requirements suppress contamination from $\PW$ and $\PZ$ boson decays.
Contamination of prompt leptons in the measurement sample from EW processes is
estimated and subtracted using simulation. The normalization of the prompt
lepton simulation is validated in a data sample enriched in $\PW$+jets events
by requiring events with a single lepton, $\ptmiss>40\GeV$, and
$60<m_{\text{T}}(\ell,\ptmiss)<100\GeV$. The minimum uncertainty that covers
the discrepancy between the data and simulation in single-lepton $\PW$+jets
events (across all $\eta$ and $\pt$ bins considered in the analysis) is 30
(13)% for electrons (muons) and is assigned as the uncertainty in the prompt
lepton normalization. The larger uncertainty for prompt electron events is to
allow for the disagreement between data and simulation in single-electron
$\PW$+jets events for high-electrons.
The method is validated using a sample of simulated , $\PW$+jets, and DY
events. The misidentification probabilities used in this validation are
obtained from simulated events comprised of jets produced via the strong
interaction, referred to as QCD multijet events. The predicted and observed
numbers of events in the $\Pe\Pe$, $\mu\mu$, and $\Pe\mu$ channels agree
within 10% for the $\PW$+jets and DY samples, and within 25% for samples. The
latter figure is reduced to 18% after rejecting events with a -tagged jet.
### 0.6.3 Background from opposite-sign leptons
To estimate backgrounds due to sign mismeasurement, the probability of
mismeasuring the lepton sign is studied. Only mismeasurement of the electron
sign is considered, and this background is estimated only in the $\Pe\Pe$
channel. The probability of mismeasuring the sign of a prompt electron is
obtained from simulated $\PZ\to\Pe^{\pm}\Pe^{\mp}$ events and is parametrized
as a function of separately for electrons in the barrel and endcap
calorimeters. The average value and statistical uncertainty for the sign
mismeasurement probabilities are found to be $(1.65\pm 0.12)\times 10^{-5}$ in
the inner ECAL barrel region ($\abs{\eta}<0.8$), $(1.07\pm 0.03)\times
10^{-4}$ in the outer ECAL barrel region ($0.8<\abs{\eta}<1.5$), and $(0.63\pm
0.01)\times 10^{-3}$ in the endcap region. The sign mismeasurement
probabilities are then validated with data separately for the barrel and
endcap regions.
To estimate the background due to sign mismeasurement in the $\Pe\Pe$ channel,
a weight $W_{p}$ is applied to data events with all the SR selections applied,
except that here the leptons are required to be oppositely signed (OS2$\ell$
events). $W_{p}$ is given by $W_{p}=p_{1}/(1-p_{1})+p_{2}/(1-p_{2})$, where
$p_{1(2)}$ is the probability for the leading (trailing) electron sign to be
mismeasured and is determined from simulated events. The of leptons with a
mismeasured sign will be misreconstructed. To correct for the misreconstructed
measurement in the OS2$\ell$ events the lepton is shifted up by 1.8%, which is
determined from simulation.
To validate the sign mismeasurement probability for the barrel (endcap)
region, a control sample of $\PZ\to\Pe^{\pm}\Pe^{\mp}$ events in the data is
selected, requiring both electrons to pass through the barrel (endcap) region
and demanding the invariant mass of the electron pair to be between 76 and
106. The difference between the observed and predicted numbers of
$\Pe^{\pm}\Pe^{\pm}$ events is used as a scale factor to account for the
modeling in the simulation. The observed number of events in the data is
determined by fitting the $\PZ$ boson mass peak. The predicted number of
events is determined by weighting the OS2$\ell$ events with the value $W_{p}$.
The scale factors and their associated statistical uncertainties in the barrel
and endcap regions are found to be $0.80\pm 0.03$ and $0.87\pm 0.03$,
respectively.
To validate the combined sign mismeasurement probability and scale factors in
the data, a control sample of $\PZ\to\Pe^{\pm}\Pe^{\mp}$ events is again
selected, as described above, but here requiring that one electron is found in
the endcap and the other, in the barrel region. The difference in the
predicted and observed numbers of $\Pe^{\pm}\Pe^{\pm}$ events in this sample
is 12%. The same procedure was performed using $\PZ\to\Pe^{\pm}\Pe^{\mp}$
events in the data but requiring no $\eta$ restrictions on the electrons and
requiring that the event has only one jet, yielding an agreement within 10%
between the predicted background and the data.
Prompt leptons and backgrounds from sign mismeasurement can contaminate the
application sample of the misidentified-lepton background, resulting in an
overprediction of this background. This contamination is removed using
simulation. The contamination from the prompt-lepton background is generally
less than 1%. However, for the backgrounds from leptons with sign
mismeasurement or leptons from photon conversions, the contamination can be as
large as 2% in the signal region and up to 30% in CR2, that is enriched in
backgrounds with mismeasured lepton sign.
### 0.6.4 Validation of background estimates
To test the validity of the background estimation methods, several signal-free
data CRs are defined. The background estimation method is applied in these
regions and the results are compared with the observed yields. These CRs are
used to validate the backgrounds separately in each of the three flavor
channels and are defined as follows:
* •
CR1: (SS2$\ell$), at least one -tagged AK4 jet,
* •
CR2: (SS2$\ell$), $\Delta R(\ell_{1},\ell_{2})>2.5$ and no -tagged AK4 jet,
* •
CR3: (SS2$\ell$), low-mass SR1 and either $\geq$ 1 -tagged jet or
$\ptmiss>100\GeV$,
* •
CR4: (SS2$\ell$), low-mass SR2 and either $\geq$ 1 -tagged jet or
$\ptmiss>100\GeV$,
* •
CR5: (SS2$\ell$), high-mass SR1 and either $\geq$ 1 -tagged jet or
$(\ptmiss)^{2}/S_{\mathrm{T}}>20\GeV$,
* •
CR6: (SS2$\ell$), high-mass SR2 and either $\geq$ 1 -tagged jet or
$(\ptmiss)^{2}/S_{\mathrm{T}}>20\GeV$.
The numbers of predicted and observed background events in each CR are shown
in Table 0.6.4. In the control regions CR1 and CR2, the backgrounds estimated
from data are dominant and validated in events both with and without -tagged
jets, while in the remaining CRs all backgrounds are validated in regions that
are close to the SRs (the misidentified-lepton background accounts for about
90% of the total background in CR1 and CR2 and about 50% across the remaining
CRs). The contribution from signal events is found to be negligible in all
control regions, with signal accounting for less than 1% of the yields in most
CRs and at most 5%, when assuming a coupling consistent with the upper limits
from previous results. In all regions the predictions are in agreement with
the observations within the statistical and systematic uncertainties described
in Section 0.7, which is dominated by the 30% uncertainty in the
misidentified-lepton background. Within each region, the observed
distributions of all relevant observables also agree with the predictions,
within the uncertainties
Observed event yields and estimated backgrounds in the control regions. The
uncertainties in the background yields are the sums in quadrature of the
statistical and systematic components. Channel Control region Estimated
background Observed $\Pe\Pe$ CR1 $\phantom{0.0}366\pm 73\phantom{0.0}$ 0378
CR2 $\phantom{0.0}690\pm 100\phantom{.0}$ 0671 CR3 $\phantom{0.0}222\pm
42\phantom{0.0}$ 0242 CR4 $\phantom{00.0}48\pm 11\phantom{0.0}$ 0038 CR5
$\phantom{0.0}334\pm 56\phantom{0.0}$ 0347 CR6 $\phantom{00}25.7\pm
4.3\phantom{00}$ 0028 $\mu\mu$ CR1 $\phantom{0.0}880\pm 230\phantom{.0}$ 0925
CR2 $\phantom{0.0}890\pm 200\phantom{.0}$ 1013 CR3 $\phantom{0.0}420\pm
100\phantom{.0}$ 0439 CR4 $\phantom{0.0}156\pm 42\phantom{0.0}$ 0174 CR5
$\phantom{0.0}560\pm 120\phantom{.0}$ 0568 CR6 $\phantom{00}35.1\pm
7.0\phantom{00}$ 0038 $\Pe\mu$ CR1 $\phantom{.0}1010\pm 240\phantom{.0}$ 1106
CR2 $\phantom{.0}1350\pm 230\phantom{.0}$ 1403 CR3 $\phantom{0.0}650\pm
140\phantom{.0}$ 0706 CR4 $\phantom{0.0}143\pm 32\phantom{0.0}$ 0150 CR5
$\phantom{0.0}920\pm 180\phantom{.0}$ 0988 CR6 $\phantom{00.0}62\pm
11\phantom{0.0}$ 0064
## 0.7 Systematic uncertainties
The estimate of backgrounds and signal efficiencies are subject to a number of
systematic uncertainties. The relative sizes of these uncertainties for each
type of background and signal, in each SR, are listed in Table 0.7. Table 0.7
shows the contributions from the uncertainty in the signal and backgrounds
(for two mass hypotheses, $m_{\mathrm{N}}=50$ and 500), expressed as a
percentage of the total uncertainty.
Summary of the relative systematic uncertainties in heavy Majorana neutrino
signal yields and in the background from prompt SS leptons, both estimated
from simulation. The relative systematic uncertainties assigned to the
misidentified-lepton and mismeasured-sign backgrounds estimated from control
regions in data and simulation are also shown. The uncertainties are given for
the low- (high-)mass selections. The range given for each systematic
uncertainty source covers the variation across the mass range. Upper limits
are presented for the uncertainty related to the PDF choice in the background
estimates, however this source of uncertainty is considered to be accounted
for via the normalization uncertainty and was not applied explicitly as an
uncertainty in the background. Source / Channel $\Pe\Pe$ signal $\Pe\Pe$
bkgd. $\mu\mu$ signal $\mu\mu$ bkgd. $\Pe\mu$ signal $\Pe\mu$ bkgd. (%) (%)
(%) (%) (%) (%) Simulation: SM cross section — .12–14 (15–27). — .13–18
(22–41). — .12–14 (16–30). Jet energy scale .02–5 (0–1).0 .002–6 (5–6).00
.02–8 (0–1).0 .003–5 (4–7).00 .01–6 (0–1).0 .001–4 (3)–.000 Jet energy
resolution .01–2 (0–0.3) .001–2 (2–6).00 .01–2 (0–0.3) 00–0.8 (1–3).00 –00.8
(0–0.3) 00–0.8 (0–3).00 Jet mass scale 0–0.3 (0–0.1) .000–1 (1–3).00 0–0.2
(0–0.1) 00–0.3 (0.7)–00 0–0.1 (0–0.1) 00–0.2 (0–5).00 Jet mass resolution
0–0.4 (0–0.3) .000–1 (0–2).00 0–0.1 (0–0.2) 00–0.1 (0–0.5)0 0–0.4 (0–0.3)
00–0.4 (0–3).00 Subjettiness .00–1 (0–8).0 00–1.0 (1–7).00 0–0.3 (0–8).0
00–0.1 (0–8).00 0–0.2 (0–8).0 00–0.4 (0–8).00 Pileup .02–3 (1)–.00 –.0002
(0–2).00 .00–1 (0–1).0 .000–1 (0–3).00 –00.7 (0.8)–0 –.0002 (2–4).00
Unclustered energy 0–0.7 (0–0.1) –.0001 (2–5).00 .00–1 (0–0.1) .000–1 (3–4).00
0–0.5 (0–0.1) –000.9 (1–2).00 Integrated luminosity –02.5 (2.5)–0 –002.5
(2.5)–00 –02.5 (2.5)–0 –002.5 (2.5)–00 –02.5 (2.5)–0 –002.5 (2.5)–00 Lepton
selection .02–4 (4)–.00 .002–4 (2–6).00 –.003 (3–4).0 –.0003 (3–5).00 –.002
(3)–.00 –.0002 (2–6).00 Trigger selection .03–4 (1)–.00 –.0003 (3–5).00 0–0.9
(0–0.4) .000–1 (0–0.8)0 –.003 (0–0.2) –.0003 (2)–.000 tagging 0–0.8 (0–1).0
–000.7 (1)–.000 0–0.5 (0–0.6) .000–1 (1–3).00 0–0.7 (0–0.7) .000–1 (1–4).00
Theory: PDF variation 0–0.7 (0–0.2) $<15$ ($<20$) 0–0.7 (0–0.1) $<15$ ($<20$)
0–0.7 (0–0.2) $<15$ ($<20$) Scale variation .01–5 (0–0.1) — .01–4 (0–0.3) —
.01–5 (0–0.2) — Estimated from data: Misidentified leptons — –.0030 (30)–.00 —
–.0030 (30)–.00 — –.0030 (30)–.00 Mismeasured sign — .29–41 (53–88). — — — —
Fractional contributions to the total background systematic uncertainties
related to the uncertainties in the prompt SS lepton, misidentified-lepton,
and mismeasured-sign backgrounds. The numbers are for the SR1 (SR2) in the
case of $m_{\mathrm{N}}$ = 50 and 500. Channel $m_{\mathrm{N}}$ Prompt-lepton
Misidentified-lepton Mismeasured-sign () (%) (%) (%) $\Pe\Pe$ 50 53 (49) 43
(46) 4.5 (4.9) 500 60 (75) 3.6 (4.6) 37 (21) $\mu\mu$ 50 38 (42) 62 (58) — 500
100 (100) 0.0 (0.0) — $\Pe\mu$ 50 52 (45) 48 (55) — 500 99 (100) 1.3 (0.0) —
### 0.7.1 Background uncertainties
The main sources of systematic uncertainties are associated with the
background estimates. The largest uncertainty is that related to the
misidentified-lepton background. The systematic uncertainty in this background
is determined by observing the change in the background estimate with respect
to variations in isolation requirement (and several other selection criteria)
for the loose leptons, modifying the requirement for the away-side jet (the
jet that is required to be back-to-back with the lepton in the measurement
region). In addition, uncertainties in the jet flavor dependence of the
misidentification probability, and in the prompt-lepton contamination in the
measurement region are taken into account By combining these sources, a
systematic uncertainty of 8.9–20% is assigned. This uncertainty depends on the
lepton flavor and the SR. The validity of the prediction of the misidentified
lepton background was checked by estimating this background using simulated
events alone. The results disagreed with those obtained from the various CRs
by up to 30%, and this value is assigned as the systematic uncertainty in this
background estimate.
The systematic uncertainties in the mismeasured electron sign background are
determined by combining weighted average of the uncertainties in barrel/endcap
scale factors from background fits, and the uncertainty on the parameterized
sign mismeasurement probabilities. To evaluate the uncertainties in the sign
mismeasurement probability scale factors, we vary the range and the number of
bins used in the fitting of the data, as well as the requirement on the
subleading lepton , and, when combining all these sources, we assign a
systematic uncertainty in the scale factors of 9%. The uncertainty in the sign
mismeasurement probability arising from the choice of parameterization
variables was estimated by considering alternative variables such as
$(\ptmiss)^{2}/S_{\mathrm{T}}$ and . A variation of up to 11% was observed.
The background estimate method was tested using only simulation, in which
OS2$\ell$ events were weighted using the sign mismeasurement probabilities
with no scale factors applied. The predicted and observed number of events in
simulation disagree by up to 7%, and this value is assigned as another source
of systematic uncertainty in estimating the sign mismeasurement background.
The three sources discussed above are combined to give a systematic
uncertainty of 16% on this background. This uncertainty covers the difference
between the predicted and observed numbers of events in both data samples
enriched in backgrounds with mismeasured electrons as discussed in Section
0.6.3.
The simulated sample used to measure the sign mismeasurement probabilities has
low statistics for events with electron above 100. When combined with the
uncertainty related to the low statistics of simulated electrons in bins with
high electron , for backgrounds from mismeasured electron sign, an overall
systematic uncertainty of 29–88% is assigned, depending on electron $\eta$ and
. The large uncertainty in this background applies only to the cases where the
SR has two high-electrons. The effect on the total systematic uncertainty in
the background is at most 5%.
### 0.7.2 Simulation uncertainties
The systematic uncertainties in the normalization of the irreducible SM
diboson backgrounds are taken from the data CR used to normalize the
backgrounds. The assigned uncertainties are 6% for $\PW\PZ$, 25% for $\PZ\PZ$
and 8% for $\PZ\gamma$ and $\PW\gamma$ backgrounds. Since other SM processes
that can yield two SS leptons, including triboson, $\ttbar\mathrm{V}$, and
$\PW^{\pm}\PW^{\pm}$, have small background yields in the SR, we assign a
conservative uncertainty of 50%, which includes the uncertainties due to
experimental effects, event simulation, and theoretical calculations of the
cross sections. The overall systematic uncertainty in the prompt-lepton
background, including the contributions discussed below, is 12–18% for the
low-mass selection and 16–43% for the high-mass selection, depending on the
lepton channel. To evaluate the uncertainty due to imperfect knowledge of the
integrated luminosity [79], jet energy/mass scale, jet energy/mass resolution
[66], tagging [73], lepton trigger and selection efficiency, as well as the
uncertainty in the total inelastic cross section used in the pileup
reweighting procedure in simulation, the input value of each parameter is
changed by $\pm 1$ standard deviation from its central value. Energy not
clustered in the detector affects the overall scale, resulting in an
uncertainty in the event yield due to the upper threshold on . The theory
uncertainties in the acceptance of the signal events are determined by varying
the renormalization and factorization scales up and down by a factor of two
relative to their nominal values, and following the PDF4LHC recommendations
[80] to estimate the uncertainty due to the choice of the PDF set. The
uncertainty related to the PDF choice in the background estimates was
evaluated, and an upper limit on the uncertainty was added to Table 0.7,
although this uncertainty was not applied explicitly in the results but
considered to be accounted for via the normalization uncertainty taken from
the normalization control regions.
## 0.8 Results and discussion
The data yields and background estimates after the application of the low- and
high-mass SR selections are shown in Table 0.8. The predicted backgrounds
contributed by events with prompt SS leptons, leptons with mismeasured sign,
and misidentified leptons are shown along with the total background estimate
and the number of events observed in data. The uncertainties shown are the
statistical and systematic components, respectively. The data yields are in
good agreement with the estimated backgrounds. Kinematic distributions also
show good agreement between data and SM expectations. Figures 2–3 show for
illustration: the invariant mass of the two leptons (of the leading lepton and
the selected jets); the invariant mass of the trailing lepton and the selected
jets; and the invariant mass of the two leptons and the selected jets for low-
(high-)mass SRs. In Fig. 2, the $m(\ell^{\pm}\ell^{\pm}\mathrm{jj})$ signal
distribution peaks somewhat below $m_{\mathrm{\PW}}$, because of the selection
requirements imposed.
Observed event yields and estimated backgrounds for the signal region
selections. The background predictions from prompt SS leptons, misidentified
leptons, leptons with mismeasured sign, and the total background are shown
together with the number of events observed in data. The uncertainties shown
are the statistical and systematic components, respectively. A dash indicates
that the background is considered negligible. SR Prompt-lepton Misidentified-
lepton Mismeasured-sign Total bkgd. $\mathrm{N}_{\text{obs}}$ $\Pe\Pe$ channel
Low-mass SR1 $206\pm 10\pm 21\phantom{0}$ $128\pm 5\pm 38\phantom{0}$ $29.8\pm
0.2\pm 12.3$ $364\pm 11\pm 45\phantom{0}$ 324 Low-mass SR2 $281\pm 12\pm
28\phantom{0}$ $143\pm 7\pm 43\phantom{0}$ $36.4\pm 0.2\pm 10.7$ $461\pm 14\pm
53\phantom{0}$ 460 High-mass SR1 $236\pm 10\pm 25\phantom{0}$ $141\pm 6\pm
42\phantom{0}$ $45.2\pm 0.3\pm 24.0$ $422\pm 12\pm 55\phantom{0}$ 382 High-
mass SR2 $8.0\pm 1.3\pm 1.6$ $2.0\pm 0.6\pm 0.6$ $0.91\pm 0.05\pm 0.80$
$10.9\pm 1.5\pm 1.9\phantom{0}$ 10 $\mu\mu$ channel Low-mass SR1 $151\pm 6\pm
16\phantom{0}$ $276\pm 7\pm 83\phantom{0}$ — $426\pm 9\pm 84\phantom{0}$ 487
Low-mass SR2 $209\pm 8\pm 19\phantom{0}$ $393\pm 9\pm 118$ — $602\pm 12\pm
120$ 663 High-mass SR1 $166\pm 6\pm 20\phantom{0}$ $244\pm 6\pm 73\phantom{0}$
— $410\pm 9\pm 76\phantom{0}$ 502 High-mass SR2 $7.1\pm 0.8\pm 1.9$ $4.4\pm
0.8\pm 1.3$ — $11.5\pm 1.1\pm 2.3\phantom{0}$ 13 $\Pe\mu$ channel Low-mass SR1
$418\pm 13\pm 37\phantom{0}$ $432\pm 10\pm 130$ — $850\pm 17\pm 135$ 907 Low-
mass SR2 $566\pm 17\pm 47\phantom{0}$ $464\pm 12\pm 139$ — $1031\pm 21\pm
147\phantom{0}$ 1042 High-mass SR1 $463\pm 14\pm 42\phantom{0}$ $409\pm 10\pm
123$ — $871\pm 17\pm 129$ 901 High-mass SR2 $16.8\pm 1.9\pm 3.6\phantom{0}$
$7.4\pm 1.3\pm 2.2$ — $24.2\pm 2.3\pm 4.2\phantom{0}$ 31
Figure 2: Observed distributions of the invariant mass of the two leptons
(upper), invariant mass of the subleading lepton and jets (middle), and the
invariant mass of the reconstructed $\PW$ propagator (lower), compared to the
expected SM background contributions, for the low-mass SR1 (left) and SR2
(right), after combining the events in the $\Pe\Pe$, $\mu\mu$, and $\Pe\mu$
channels. The hatched bands represent the sums in quadrature of the
statistical and systematic uncertainties. The solid and dashed lines show the
kinematic distributions of two possible signal hypothesis. The lower panels
show the ratio between the observed and expected events in each bin, including
the uncertainty bands that represent the statistical (cyan) and total
uncertainties (orange).
Figure 3: Observed distributions of the invariant mass of the leading lepton
and jets (upper), invariant mass of the subleading lepton and jets (middle),
and the invariant mass of the reconstructed $\PW$ propagator (lower), compared
to the expected SM background contributions, for the high-mass SR1 (left) and
SR2 (right), after combining the events in the $\Pe\Pe$, $\mu\mu$, and
$\Pe\mu$ channels. The hatched bands represent the sums in quadrature of the
statistical and systematic uncertainties. The solid and dashed lines show the
kinematic distributions of two possible signal hypothesis. The lower panels
show the ratio between the observed and expected events in each bin, including
the uncertainty bands that represent the statistical (cyan) and total
uncertainties (orange).
The expected signal depends on both $m_{\mathrm{N}}$ and the mixing matrix
elements $\abs{V_{\Pe\mathrm{N}}}^{2}$, $\abs{V_{\mu\mathrm{N}}}^{2}$, or
$\abs{V_{\Pe\mathrm{N}}V^{*}_{\mu\mathrm{N}}}^{2}/(\abs{V_{\Pe\mathrm{N}}}^{2}+\abs{V_{\mu\mathrm{N}}}^{2})$,
and the values are summarized in Table 0.8 for a few mass points. Tables
0.8–0.8 show the optimized selections applied on top of the low- and high-mass
SRs requirements for each mass hypothesis. These tables also present the
observed event counts in data and the expected background for each signal mass
hypothesis. The data are generally consistent with the predicted backgrounds
in all three flavor channels. The largest deviation observed is in the
$\mu\mu$ channel of SR1, at a signal mass of 600, and has a local significance
of 2.3 standard deviations. The corresponding point of SR2 does not show a
matching fluctuation.
Numbers of expected signal events after all the selections are applied. The
matrix element squared is equal to $1\times 10^{-4}$, $1\times 10^{-2}$, and 1
for $m_{\mathrm{N}}=50$, 200, and 1000, respectively. $m_{\mathrm{N}}$
$\Pe\Pe$ channel $\mu\mu$ channel $\Pe\mu$ channel () SR1 SR2 SR1 SR2 SR1 SR2
50 15 21 28 76 14 26 200 5.5 0.74 9.7 1.9 7.0 1.1 1000 0.43 4.0 0.80 7.5 0.57
4.5
Exclusion limits at 95% confidence level (CL) are set on the heavy Majorana
neutrino mixing matrix elements as a function of $m_{\mathrm{N}}$. The limits
are obtained using criterion [81, 82] based on the event yields in Tables
0.8–0.8. Log-normal distributions are used for both the signal and nuisance
parameters. The combined limits from SR1 and SR2, on the absolute values of
the matrix elements $\abs{V_{\Pe\mathrm{N}}}^{2}$,
$\abs{V_{\mu\mathrm{N}}}^{2}$, and
$\abs{V_{\Pe\mathrm{N}}V^{*}_{\mu\mathrm{N}}}^{2}/(\abs{V_{\Pe\mathrm{N}}}^{2}+\abs{V_{\mu\mathrm{N}}}^{2})$
are shown in Figs. 4–5, also as a function of $m_{\mathrm{N}}$. We assume the
systematic uncertainties in SR1 and SR2 to be fully correlated when
calculating these limits. The limits are calculated separately for each of the
three channels. For an $\mathrm{N}$ mass of 40the observed (expected) limits
are $\abs{V_{\Pe\mathrm{N}}}^{2}<9.5\,(8.0)\times 10^{-5}$,
$\abs{V_{\mu\mathrm{N}}}^{2}<2.3\,(1.9)\times 10^{-5}$, and
$\abs{V_{\Pe\mathrm{N}}V^{*}_{\mu\mathrm{N}}}^{2}/(\abs{V_{\Pe\mathrm{N}}}^{2}+\abs{V_{\mu\mathrm{N}}}^{2})<2.7\,(2.7)\times
10^{-5}$, and for an $\mathrm{N}$ mass of 1000the limits are
$\abs{V_{\Pe\mathrm{N}}}^{2}<0.42\,(0.32)$,
$\abs{V_{\mu\mathrm{N}}}^{2}<0.27\,(0.16)$, and
$\abs{V_{\Pe\mathrm{N}}V^{*}_{\mu\mathrm{N}}}^{2}/(\abs{V_{\Pe\mathrm{N}}}^{2}+\abs{V_{\mu\mathrm{N}}}^{2})<0.14\,(0.14)$.
The mass range below $m_{\mathrm{N}}=20\GeV$ is not considered because of the
very low selection efficiency in this region. Furthermore, since the
$\mathrm{N}$ lifetime is inversely proportional to
$m_{\mathrm{N}}^{5}\abs{V_{\ell\mathrm{N}}}^{2}$, for $m_{\mathrm{N}}<20\GeV$
it becomes significant and results in displaced decays. Thus the prompt lepton
requirement is not satisfied. The behavior of the limits around
$m_{\mathrm{N}}=80\GeV$ is caused by the fact that as the mass of the heavy
Majorana neutrino approaches the $\PW$ boson mass, the lepton produced
together with the $\mathrm{N}$ or the lepton from the $\mathrm{N}$ decay has
very low .
The present search at 13extends the previous CMS SS2$\ell$ plus jets searches
at 8 [34, 35] to both higher $\mathrm{N}$ masses as well as lower masses. In
those earlier searches, two AK4 jets were required in the low- and high-mass
SRs, while in the present analysis at $\sqrt{s}=13\TeV$, the search has been
extended in the low-mass SR to include events with exactly one AK4 jet, and in
the high-mass SR to include events with at least one AK8 jet. As seen in Figs.
4–5, the exclusion limits for the mixing matrix elements are extended both for
low and high $\mathrm{N}$ mass, and now cover $\mathrm{N}$ masses from 20 to
1600. In the range previously studied, the present limits significantly
improve over the previous results except in the region from 60–80, where they
are equivalent. The 13data were taken at higher collision rates and thus with
higher trigger thresholds and pileup rates, which impacted the sensitivity of
the search in the low-mass region. This region is covered with high efficiency
by a recent search in trilepton channels [37].
Figs. 4 shows the exclusion limits for $\abs{V_{\Pe\mathrm{N}}}^{2}$ and
$\abs{V_{\mu\mathrm{N}}}^{2}$ overlaid with the 13CMS limits from the
trilepton channel [37]. For low-mass signals the trilepton analysis is more
sensitive, since it has both fewer backgrounds from misidentified leptons and
higher signal efficiency. However for high-mass signals the signal
efficiencies are compatible, and with the inclusion of the signal region using
AK8 jets, and the larger signal cross section in the dilepton channel this
analysis has more stringent limits for masses of $\mathrm{N}$ above 100.
Selection requirements on discriminating variables determined by the
optimization for each Majorana neutrino mass point in the low-mass signal
regions. The last columns show the overall signal acceptance for the DY
channel. The quoted uncertainties include both the statistical and systematic
contributions. $m_{\mathrm{N}}$ $p^{\ell_{1}}_{\mathrm{T}}$ $\pt^{\ell_{2}}$
$m(\ell^{\pm}\ell^{\pm}\PW_{\text{jet}})$ $m(\ell_{1}\PW_{\text{jet}})$
$m(\ell_{2}\PW_{\text{jet}})$ $m(\ell^{\pm}\ell^{\pm})$ Total bkgd.
$\mathrm{N}_{\text{obs}}$ DY $A\epsilon$ () () () () () () () (%) $\Pe\Pe$
channel SR1 20 25–70 60 $<$190 $<$160 $<$160 10–60 $\phantom{}48.9\pm
9.5\phantom{0}$ 45 $0.12\pm 0.02$ 30 25–70 60 $<$190 $<$160 $<$160 10–60
$\phantom{}48.9\pm 9.5\phantom{0}$ 45 $0.13\pm 0.02$ 40 25–70 60 $<$190 $<$160
$<$160 10–60 $\phantom{}48.9\pm 9.5\phantom{0}$ 45 $0.21\pm 0.03$ 50 25–70 60
$<$190 $<$160 $<$160 10–60 $\phantom{}48.9\pm 9.5\phantom{0}$ 45 $0.24\pm
0.03$ 60 25–70 60 $<$190 $<$160 $<$160 10–60 $\phantom{}48.9\pm
9.5\phantom{0}$ 45 $0.18\pm 0.02$ 70 25–70 60 $<$190 $<$160 $<$160 10–75
$\phantom{.0}64\pm 12\phantom{.0}$ 58 $0.10\pm 0.01$ 75 25–70 60 $<$190 $<$160
$<$160 10–100 $\phantom{.0}68\pm 12\phantom{.0}$ 67 $0.13\pm 0.02$ $\Pe\Pe$
channel SR2 20 25–70 60 $<$100 $<$70 $<$70 10–60 $\phantom{}50.3\pm
8.5\phantom{0}$ 55 $0.26\pm 0.03$ 30 25–70 60 $<$100 $<$70 $<$70 10–60
$\phantom{}50.3\pm 8.5\phantom{0}$ 55 $0.30\pm 0.04$ 40 25–70 60 $<$100 $<$70
$<$70 10–60 $\phantom{}50.3\pm 8.5\phantom{0}$ 55 $0.35\pm 0.04$ 50 25–70 60
$<$100 $<$70 $<$70 10–60 $\phantom{}50.3\pm 8.5\phantom{0}$ 55 $0.32\pm 0.03$
60 25–70 60 $<$100 $<$70 $<$70 10–60 $\phantom{}50.3\pm 8.5\phantom{0}$ 55
$0.24\pm 0.03$ 70 25–70 60 $<$100 $<$70 $<$70 10–75 $\phantom{.0}65\pm
10\phantom{.0}$ 70 $0.06\pm 0.01$ 75 25–70 60 $<$100 $<$70 $<$70 10–80
$\phantom{.0}67\pm 10\phantom{.0}$ 70 $0.11\pm 0.02$ $\mu\mu$ channel SR1 20
20–80 15–50 $<$160 $<$150 $<$150 20–60 $\phantom{}15.3\pm 3.4\phantom{0}$ 18
$0.10\pm 0.02$ 30 20–80 15–50 $<$160 $<$150 $<$150 20–60 $\phantom{}15.3\pm
3.4\phantom{0}$ 18 $0.18\pm 0.03$ 40 20–80 15–50 $<$160 $<$150 $<$150 20–60
$\phantom{}15.3\pm 3.4\phantom{0}$ 18 $0.34\pm 0.05$ 50 20–80 15–50 $<$160
$<$150 $<$150 20–60 $\phantom{}15.3\pm 3.4\phantom{0}$ 18 $0.40\pm 0.04$ 60
20–80 15–50 $<$160 $<$150 $<$150 20–60 $\phantom{}15.3\pm 3.4\phantom{0}$ 18
$0.33\pm 0.04$ 70 20–80 15–50 $<$160 $<$150 $<$150 10–75 $\phantom{}20.3\pm
4.4\phantom{0}$ 21 $0.17\pm 0.02$ 75 20–80 15–50 $<$160 $<$150 $<$150 20–100
$\phantom{}18.9\pm 4.0\phantom{0}$ 19 $0.19\pm 0.03$ $\mu\mu$ channel SR2 20
20–80 15–50 $<$100 $<$70 $<$70 20–60 $\phantom{}25.9\pm 5.9\phantom{0}$ 29
$0.28\pm 0.03$ 30 20–80 15–50 $<$100 $<$70 $<$70 20–60 $\phantom{}25.9\pm
5.9\phantom{0}$ 29 $0.51\pm 0.05$ 40 20–80 15–50 $<$100 $<$70 $<$70 20–60
$\phantom{}25.9\pm 5.9\phantom{0}$ 29 $0.8\pm 0.1$ 50 20–80 15–50 $<$100 $<$70
$<$70 20–60 $\phantom{}25.9\pm 5.9\phantom{0}$ 29 $1.1\pm 0.1$ 60 20–80 15–50
$<$100 $<$70 $<$70 20–60 $\phantom{}25.9\pm 5.9\phantom{0}$ 29 $0.73\pm 0.07$
70 20–80 15–50 $<$100 $<$70 $<$70 10–75 $\phantom{}37.5\pm 7.1\phantom{0}$ 41
$0.20\pm 0.03$ 75 20–80 15–50 $<$100 $<$70 $<$70 20–80 $\phantom{}29.7\pm
6.7\phantom{0}$ 34 $0.24\pm 0.03$ $\Pe\mu$ channel SR1 20 25–60 15–40 $<$185
$<$135 $<$135 20–60 $\phantom{}34.0\pm 6.4\phantom{0}$ 34 $0.08\pm 0.02$ 30
25–60 15–40 $<$185 $<$135 $<$135 20–60 $\phantom{}34.0\pm 6.4\phantom{0}$ 34
$0.12\pm 0.02$ 40 25–60 15–40 $<$185 $<$135 $<$135 20–60 $\phantom{}34.0\pm
6.4\phantom{0}$ 34 $0.21\pm 0.02$ 50 25–60 15–40 $<$185 $<$135 $<$135 20–60
$\phantom{}34.0\pm 6.4\phantom{0}$ 34 $0.20\pm 0.03$ 60 25–60 15–40 $<$185
$<$135 $<$135 20–60 $\phantom{}34.0\pm 6.4\phantom{0}$ 34 $0.17\pm 0.02$ 70
25–60 15–40 $<$185 $<$135 $<$135 10–75 $\phantom{.0}51\pm 10\phantom{.0}$ 49
$0.09\pm 0.01$ 75 25–60 15–40 $<$185 $<$135 $<$135 20–100 $\phantom{}46.5\pm
8.7\phantom{0}$ 49 $0.17\pm 0.03$ $\Pe\mu$ channel SR2 20 25–60 15–40 $<$100
$<$65 $<$65 20–60 $\phantom{}51.7\pm 9.2\phantom{0}$ 50 $0.21\pm 0.02$ 30
25–60 15–40 $<$100 $<$65 $<$65 20–60 $\phantom{}51.7\pm 9.2\phantom{0}$ 50
$0.27\pm 0.03$ 40 25–60 15–40 $<$100 $<$65 $<$65 20–60 $\phantom{}51.7\pm
9.2\phantom{0}$ 50 $0.45\pm 0.04$ 50 25–60 15–40 $<$100 $<$65 $<$65 20–60
$\phantom{}51.7\pm 9.2\phantom{0}$ 50 $0.40\pm 0.03$ 60 25–60 15–40 $<$100
$<$65 $<$65 20–60 $\phantom{}51.7\pm 9.2\phantom{0}$ 50 $0.24\pm 0.03$ 70
25–60 15–40 $<$100 $<$65 $<$65 10–75 $\phantom{}75.8\pm 12.4\phantom{}$ 65
$0.09\pm 0.01$ 75 25–60 15–40 $<$100 $<$65 $<$65 20–80 $\phantom{}62.8\pm
10.9\phantom{}$ 57 $0.12\pm 0.03$
Selection requirements on discriminating variables determined by the
optimization for each Majorana neutrino mass point in the $\Pe\Pe$ channel
high-mass SRs. The last column shows the overall signal acceptance for the DY
and VBF channels. The quoted uncertainties include both the statistical and
systematic contributions. The dash indicates that no selection requirement is
made. $m_{\mathrm{N}}$ $p^{\ell_{1}}_{\mathrm{T}}$ $\pt^{\ell_{2}}$
$m(\ell^{\pm}\ell^{\pm}\PW_{\text{jet}})$ $m(\ell\PW_{\text{jet}})$
$(\ptmiss)^{2}/S_{\mathrm{T}}$ Total bkgd. $\mathrm{N}_{\text{obs}}$ DY
$A\epsilon$ VBF $A\epsilon$ () () () () () () (%) (%) $\Pe\Pe$ channel SR1 85
$>$25 $>$15 $>$110 45–95 $<$6 $\phantom{0}9.5\pm 2.8\phantom{0}$ 9
$\phantom{0.}0.11\pm 0.02\phantom{.}$ — 90 $>$25 $>$15 $>$110 50–100 $<$6
$\phantom{}12.5\pm 3.5\phantom{0}$ 10 $\phantom{0.}0.23\pm 0.05\phantom{.}$ —
100 $>$25 $>$15 $>$120 50–110 $<$6 $\phantom{}20.3\pm 5.0\phantom{0}$ 15
$\phantom{0.0}1.1\pm 0.1\phantom{.0}$ — 125 $>$30 $>$25 $>$120 90–140 $<$6
$\phantom{}17.7\pm 4.5\phantom{0}$ 17 $\phantom{0.0}2.6\pm 0.2\phantom{.0}$ —
150 $>$40 $>$25 $>$180 130–160 $<$6 $\phantom{}14.7\pm 3.8\phantom{0}$ 9
$\phantom{0.0}3.1\pm 0.2\phantom{.0}$ — 200 $>$55 $>$40 $>$220 160–225 $<$6
$\phantom{}12.4\pm 2.7\phantom{0}$ 10 $\phantom{0.0}4.9\pm 0.4\phantom{.0}$ —
250 $>$70 $>$60 $>$310 220–270 $<$6 $\phantom{0}6.0\pm 1.7\phantom{0}$ 4
$\phantom{0.0}5.9\pm 0.4\phantom{.0}$ — 300 $>$80 $>$60 $>$370 235–335 $<$6
$\phantom{0}8.2\pm 2.1\phantom{0}$ 6 $\phantom{0.0}7.6\pm 0.5\phantom{.0}$
$\phantom{.0}3.0\pm 0.3\phantom{.0}$ 400 $>$100 $>$65 $>$450 335–450 $<$6
$\phantom{0}2.5\pm 1.4\phantom{0}$ 4 $\phantom{0.0}6.6\pm 0.5\phantom{.0}$
$\phantom{.0}3.0\pm 0.2\phantom{.0}$ 500 $>$125 $>$65 $>$560 400–555 $<$6
$\phantom{0}1.5\pm 0.8\phantom{0}$ 5 $\phantom{0.0}5.5\pm 0.4\phantom{.0}$
$\phantom{.0}2.7\pm 0.2\phantom{.0}$ 600 $>$125 — $>$760 400–690 $<$6
$\phantom{0}0.9\pm 0.6\phantom{0}$ 1 $\phantom{0.0}3.8\pm 0.3\phantom{.0}$
$\phantom{.0}1.7\pm 0.2\phantom{.0}$ 700 $>$125 — $>$760 400–955 $<$6
$\phantom{0}1.7\pm 0.7\phantom{0}$ 1 $\phantom{0.0}4.0\pm 0.3\phantom{.0}$
$\phantom{.0}2.8\pm 0.2\phantom{.0}$ 800 $>$125 — $>$760 400–1130 $<$6
$\phantom{0}1.7\pm 0.7\phantom{0}$ 1 $\phantom{0.0}3.6\pm 0.3\phantom{.0}$
$\phantom{.0}3.0\pm 0.3\phantom{.0}$ 900 $>$125 — $>$760 400–1300 $<$6
$\phantom{0}1.7\pm 0.7\phantom{0}$ 1 $\phantom{0.0}3.2\pm 0.2\phantom{.0}$
$\phantom{.0}2.9\pm 0.2\phantom{.0}$ 1000 $>$125 — $>$760 400–1490 $<$6
$\phantom{0}1.7\pm 0.7\phantom{0}$ 1 $\phantom{0.0}2.6\pm 0.2\phantom{.0}$
$\phantom{.0}2.4\pm 0.2\phantom{.0}$ 1100 $>$125 — $>$760 400–1490 $<$6
$\phantom{0}1.7\pm 0.7\phantom{0}$ 1 $\phantom{0.0}2.2\pm 0.2\phantom{.0}$
$\phantom{.0}2.0\pm 0.2\phantom{.0}$ 1200 $>$125 — $>$760 400–1600 $<$6
$\phantom{0}1.7\pm 0.7\phantom{0}$ 1 $\phantom{0.0}2.0\pm 0.2\phantom{.0}$
$\phantom{.0}1.8\pm 0.2\phantom{.0}$ 1300 $>$125 — $>$760 400–1930 $<$6
$\phantom{0}1.7\pm 0.7\phantom{0}$ 1 $\phantom{0.0}1.8\pm 0.1\phantom{.0}$
$\phantom{.0}1.6\pm 0.2\phantom{.0}$ 1400 $>$125 — $>$760 400–1930 $<$6
$\phantom{0}1.7\pm 0.7\phantom{0}$ 1 $\phantom{0.0}1.5\pm 0.1\phantom{.0}$
$\phantom{.0}1.3\pm 0.1\phantom{.0}$ 1500 $>$125 — $>$760 400–1930 $<$6
$\phantom{0}1.7\pm 0.7\phantom{0}$ 1 $\phantom{0.0}1.3\pm 0.1\phantom{.0}$
$\phantom{.0}1.2\pm 0.2\phantom{.0}$ $\Pe\Pe$ channel SR2 85 $>$25 $>$15 — —
$<$15 $\phantom{}10.9\pm 2.9\phantom{0}$ 10 $\phantom{0.}0.001\pm
0.001\phantom{.}$ — 90 $>$25 $>$15 — 90–220 $<$15 $\phantom{0}3.4\pm
1.0\phantom{0}$ 2 $\phantom{0.}0.003\pm 0.002\phantom{.}$ — 100 $>$25 $>$15 —
100–220 $<$15 $\phantom{0}3.4\pm 1.0\phantom{0}$ 2 $\phantom{0.}0.005\pm
0.003\phantom{.}$ — 125 $>$60 $>$15 — 123–145 $<$15 $\phantom{0}0.2\pm
0.1\phantom{0}$ 0 $\phantom{0.}0.04\pm 0.01\phantom{.}$ — 150 $>$90 $>$15 —
125–185 $<$15 $\phantom{0}1.3\pm 0.5\phantom{0}$ 0 $\phantom{0.}0.19\pm
0.03\phantom{.}$ — 200 $>$100 $>$20 — 173–220 $<$15 $\phantom{0}0.8\pm
0.3\phantom{0}$ 1 $\phantom{0.}0.60\pm 0.07\phantom{.}$ — 250 $>$100 $>$25 —
220–305 $<$15 $\phantom{0}2.1\pm 1.2\phantom{0}$ 3 $\phantom{0.0}2.2\pm
0.2\phantom{.0}$ — 300 $>$100 $>$30 — 270–330 $<$15 $\phantom{0}1.3\pm
0.6\phantom{0}$ 1 $\phantom{0.0}3.5\pm 0.4\phantom{.0}$ $\phantom{0.0}0.6\pm
0.1\phantom{.0}$ 400 $>$100 $>$35 — 330–440 $<$15 $\phantom{0}3.1\pm
1.3\phantom{0}$ 3 $\phantom{0.0}9.1\pm 0.9\phantom{.0}$ $\phantom{0.0}2.9\pm
0.3\phantom{.0}$ 500 $>$120 $>$35 — 440–565 $<$15 $\phantom{0}2.8\pm
1.0\phantom{0}$ 1 $\phantom{.0}14.3\pm 1.4\phantom{.0}$ $\phantom{0.0}6.1\pm
0.6\phantom{.0}$ 600 $>$120 — — 565–675 $<$15 $\phantom{0}0.8\pm
0.3\phantom{0}$ 1 $\phantom{.0}17.4\pm 1.8\phantom{.0}$ $\phantom{.0}11.0\pm
1.0\phantom{.0}$ 700 $>$140 — — 635–775 $<$15 $\phantom{0}0.8\pm
0.3\phantom{0}$ 2 $\phantom{.0}19.4\pm 2.0\phantom{.0}$ $\phantom{.0}13.1\pm
1.3\phantom{.0}$ 800 $>$140 — — 740–1005 $<$15 $\phantom{0}0.9\pm
0.4\phantom{0}$ 0 $\phantom{.0}20.8\pm 2.1\phantom{.0}$ $\phantom{.0}14.0\pm
1.3\phantom{.0}$ 900 $>$140 — — 865–1030 $<$15 $\phantom{0}0.2\pm
0.1\phantom{0}$ 0 $\phantom{.0}19.2\pm 2.0\phantom{.0}$ $\phantom{.0}13.2\pm
1.3\phantom{.0}$ 1000 $>$140 — — 890–1185 $<$15 $\phantom{0}0.3\pm
0.1\phantom{0}$ 1 $\phantom{.0}21.5\pm 2.2\phantom{.0}$ $\phantom{.0}15.3\pm
1.5\phantom{.0}$ 1100 $>$140 — — 1035–1395 $<$15 $\phantom{0}0.1\pm
0.1\phantom{0}$ 1 $\phantom{.0}20.3\pm 2.1\phantom{.0}$ $\phantom{.0}14.7\pm
1.4\phantom{.0}$ 1200 $>$140 — — 1085–1460 $<$15 $\phantom{0}0.1\pm
0.0\phantom{0}$ 1 $\phantom{.0}20.8\pm 2.2\phantom{.0}$ $\phantom{.0}15.3\pm
1.5\phantom{.0}$ 1300 $>$140 — — 1140–1590 $<$15 $\phantom{0}0.1\pm
0.0\phantom{0}$ 1 $\phantom{.0}20.5\pm 2.2\phantom{.0}$ $\phantom{.0}15.5\pm
1.6\phantom{.0}$ 1400 $>$140 — — 1245–1700 $<$15 $\phantom{0}0.1\pm
0.0\phantom{0}$ 0 $\phantom{.0}19.6\pm 2.1\phantom{.0}$ $\phantom{.0}15.1\pm
1.6\phantom{.0}$ 1500 $>$140 — — 1300–1800 $<$15 $\phantom{0}0.04\pm
0.02\phantom{0}$ 0 $\phantom{.0}19.5\pm 2.1\phantom{.0}$ $\phantom{.0}15.2\pm
1.6\phantom{.0}$
Selection requirements on discriminating variables determined by the
optimization for each Majorana neutrino mass point in the $\mu\mu$ channel
high-mass SRs. The last column shows the overall signal acceptance for the DY
and VBF channels. The quoted uncertainties include both the statistical and
systematic contributions. The dash indicates that no selection requirement is
made. $m_{\mathrm{N}}$ $p^{\ell_{1}}_{\mathrm{T}}$ $\pt^{\ell_{2}}$
$m(\ell^{\pm}\ell^{\pm}\PW_{\text{jet}})$ $m(\ell\PW_{\text{jet}})$
$(\ptmiss)^{2}/S_{\mathrm{T}}$ Total bkgd. $\mathrm{N}_{\text{obs}}$ DY
$A\epsilon$ VBF $A\epsilon$ () () () () () () (%) (%) $\mu\mu$ channel SR1 85
$>$25 $>$10 $>$90 40–100 $<$9 $\phantom{}26.0\pm 6.3\phantom{0}$ 30
$\phantom{0.}0.50\pm 0.05\phantom{.}$ — 90 $>$25 $>$10 $>$90 45–105 $<$9
$\phantom{}34.5\pm 7.5\phantom{0}$ 35 $\phantom{0.0}1.2\pm 0.1\phantom{.0}$ —
100 $>$25 $>$15 $>$110 55–115 $<$9 $\phantom{}18.6\pm 4.2\phantom{0}$ 20
$\phantom{0.0}2.6\pm 0.2\phantom{.0}$ — 125 $>$25 $>$25 $>$140 85–140 $<$7
$\phantom{}11.7\pm 2.7\phantom{0}$ 12 $\phantom{0.0}5.1\pm 0.4\phantom{.0}$ —
150 $>$35 $>$35 $>$150 110–170 $<$7 $\phantom{0}8.9\pm 1.9\phantom{0}$ 11
$\phantom{0.0}6.6\pm 0.5\phantom{.0}$ — 200 $>$50 $>$40 $>$250 160–215 $<$7
$\phantom{0}4.6\pm 1.2\phantom{0}$ 4 $\phantom{0.0}8.1\pm 0.6\phantom{.0}$ —
250 $>$85 $>$45 $>$310 215–270 $<$7 $\phantom{0}3.0\pm 0.9\phantom{0}$ 2
$\phantom{.0}11.0\pm 0.8\phantom{.0}$ — 300 $>$100 $>$50 $>$370 225–340 $<$7
$\phantom{0}2.6\pm 1.0\phantom{0}$ 2 $\phantom{.0}13.2\pm 0.9\phantom{.0}$
$\phantom{.0}5.2\pm 0.4\phantom{.0}$ 400 $>$110 $>$60 $>$490 295–490 $<$7
$\phantom{0}0.9\pm 0.4\phantom{0}$ 3 $\phantom{.0}11.7\pm 0.8\phantom{.0}$
$\phantom{.0}5.1\pm 0.4\phantom{.0}$ 500 $>$110 $>$60 $>$610 370–550 $<$7
$\phantom{0}0.4^{\phantom{0}+\phantom{0}0.6\phantom{0}\phantom{0}}_{\phantom{0}-\phantom{0}0.4\phantom{0}\phantom{0}}$
3 $\phantom{0.0}8.6\pm 0.6\phantom{.0}$ $\phantom{.0}4.1\pm 0.3\phantom{.0}$
600 $>$110 — $>$680 370–630 $<$7
$\phantom{0}0.3^{\phantom{0}+\phantom{0}0.3\phantom{0}\phantom{0}}_{\phantom{0}-\phantom{0}0.3\phantom{0}\phantom{0}}$
3 $\phantom{0.0}7.4\pm 0.5\phantom{.0}$ $\phantom{.0}4.1\pm 0.3\phantom{.0}$
700 $>$110 — $>$800 370–885 $<$7
$\phantom{0}0.2^{\phantom{0}+\phantom{0}0.4\phantom{0}\phantom{0}}_{\phantom{0}-\phantom{0}0.2\phantom{0}\phantom{0}}$
2 $\phantom{0.0}6.7\pm 0.4\phantom{.0}$ $\phantom{.0}3.9\pm 0.3\phantom{.0}$
800 $>$110 — $>$800 370–890 $<$7
$\phantom{0}0.2^{\phantom{0}+\phantom{0}0.4\phantom{0}\phantom{0}}_{\phantom{0}-\phantom{0}0.2\phantom{0}\phantom{0}}$
2 $\phantom{0.0}6.0\pm 0.4\phantom{.0}$ $\phantom{.0}5.4\pm 0.3\phantom{.0}$
900 $>$110 — $>$800 370–1225 $<$7
$\phantom{0}0.3^{\phantom{0}+\phantom{0}0.4\phantom{0}\phantom{0}}_{\phantom{0}-\phantom{0}0.3\phantom{0}\phantom{0}}$
2 $\phantom{0.0}5.4\pm 0.4\phantom{.0}$ $\phantom{.0}5.0\pm 0.3\phantom{.0}$
1000 $>$110 — $>$800 370–1230 $<$7
$\phantom{0}0.3^{\phantom{0}+\phantom{0}0.4\phantom{0}\phantom{0}}_{\phantom{0}-\phantom{0}0.3\phantom{0}\phantom{0}}$
2 $\phantom{0.0}4.6\pm 0.3\phantom{.0}$ $\phantom{.0}4.2\pm 0.3\phantom{.0}$
1100 $>$110 — $>$800 370–1245 $<$7
$\phantom{0}0.3^{\phantom{0}+\phantom{0}0.4\phantom{0}\phantom{0}}_{\phantom{0}-\phantom{0}0.3\phantom{0}\phantom{0}}$
2 $\phantom{0.0}4.1\pm 0.3\phantom{.0}$ $\phantom{.0}3.8\pm 0.3\phantom{.0}$
1200 $>$110 — $>$800 370–1690 $<$7
$\phantom{0}0.3^{\phantom{0}+\phantom{0}0.4\phantom{0}\phantom{0}}_{\phantom{0}-\phantom{0}0.3\phantom{0}\phantom{0}}$
2 $\phantom{0.0}3.6\pm 0.2\phantom{.0}$ $\phantom{.0}3.4\pm 0.3\phantom{.0}$
1300 $>$110 — $>$800 370–1890 $<$7
$\phantom{0}0.3^{\phantom{0}+\phantom{0}0.4\phantom{0}\phantom{0}}_{\phantom{0}-\phantom{0}0.3\phantom{0}\phantom{0}}$
2 $\phantom{0.0}3.2\pm 0.2\phantom{.0}$ $\phantom{.0}3.0\pm 0.2\phantom{.0}$
1400 $>$110 — $>$800 370–1940 $<$7
$\phantom{0}0.3^{\phantom{0}+\phantom{0}0.4\phantom{0}\phantom{0}}_{\phantom{0}-\phantom{0}0.3\phantom{0}\phantom{0}}$
2 $\phantom{0.0}2.7\pm 0.2\phantom{.0}$ $\phantom{.0}2.7\pm 0.2\phantom{.0}$
1500 $>$110 — $>$800 370–2220 $<$7
$\phantom{0}0.3^{\phantom{0}+\phantom{0}0.4\phantom{0}\phantom{0}}_{\phantom{0}-\phantom{0}0.3\phantom{0}\phantom{0}}$
2 $\phantom{0.0}2.5\pm 0.2\phantom{.0}$ $\phantom{.0}2.3\pm 0.2\phantom{.0}$
$\mu\mu$ channel SR2 85 $>$25 $>$10 — — $<$15 $\phantom{}11.4\pm
3.5\phantom{0}$ 13 $\phantom{0.}0.001\pm 0.001\phantom{.}$ — 90 $>$25 $>$10 —
90–170 $<$15 $\phantom{0}4.1\pm 1.3\phantom{0}$ 4 $\phantom{0.}0.003\pm
0.003\phantom{.}$ — 100 $>$25 $>$15 — 98–145 $<$15 $\phantom{0}1.0\pm
0.3\phantom{0}$ 0 $\phantom{0.}0.006\pm 0.003\phantom{.}$ — 125 $>$60 $>$15 —
110–150 $<$15 $\phantom{0}0.8\pm 0.3\phantom{0}$ 0 $\phantom{0.}0.08\pm
0.01\phantom{.}$ — 150 $>$70 $>$15 — 145–175 $<$15 $\phantom{0}1.0\pm
0.4\phantom{0}$ 2 $\phantom{0.}0.28\pm 0.04\phantom{.}$ — 200 $>$100 $>$20 —
175–235 $<$15 $\phantom{0}1.3\pm 0.8\phantom{0}$ 0 $\phantom{0.0}1.4\pm
0.1\phantom{.0}$ — 250 $>$140 $>$25 — 226–280 $<$15 $\phantom{0}0.3\pm
0.2\phantom{0}$ 0 $\phantom{0.0}3.0\pm 0.3\phantom{.0}$ — 300 $>$140 $>$40 —
280–340 $<$15 $\phantom{0}0.4\pm 0.3\phantom{0}$ 0 $\phantom{0.0}5.4\pm
0.5\phantom{.0}$ $\phantom{0.0}0.7\pm 0.1\phantom{.0}$ 400 $>$140 $>$65 —
340–445 $<$15 $\phantom{0}0.5\pm 0.3\phantom{0}$ 2 $\phantom{.0}13.3\pm
1.3\phantom{.0}$ $\phantom{0.0}2.7\pm 0.3\phantom{.0}$ 500 $>$140 $>$65 —
445–560 $<$15 $\phantom{0}0.8\pm 0.5\phantom{0}$ 0 $\phantom{.0}22.4\pm
2.2\phantom{.0}$ $\phantom{0.0}6.8\pm 0.7\phantom{.0}$ 600 $>$140 — — 560–685
$<$15 $\phantom{0}0.7\pm 0.4\phantom{0}$ 0 $\phantom{.0}30.2\pm
2.9\phantom{.0}$ $\phantom{.0}20.4\pm 1.8\phantom{.0}$ 700 $>$140 — — 635–825
$<$15 $\phantom{0}0.8\pm 0.4\phantom{0}$ 2 $\phantom{.0}34.6\pm
3.4\phantom{.0}$ $\phantom{.0}24.7\pm 2.2\phantom{.0}$ 800 $>$140 — — 755–960
$<$15 $\phantom{0}0.4\pm 0.3\phantom{0}$ 0 $\phantom{.0}34.8\pm
3.5\phantom{.0}$ $\phantom{.0}24.9\pm 2.3\phantom{.0}$ 900 $>$140 — — 840–1055
$<$15
$\phantom{0}0.2^{\phantom{0}+\phantom{0}0.2\phantom{0}\phantom{0}}_{\phantom{0}-\phantom{0}0.2\phantom{0}\phantom{0}}$
1 $\phantom{.0}35.8\pm 3.6\phantom{.0}$ $\phantom{.0}26.9\pm 2.5\phantom{.0}$
1000 $>$140 — — 900–1205 $<$15
$\phantom{0}0.1^{\phantom{0}+\phantom{0}0.2\phantom{0}\phantom{0}}_{\phantom{0}-\phantom{0}0.1\phantom{0}\phantom{0}}$
1 $\phantom{.0}38.4\pm 3.9\phantom{.0}$ $\phantom{.0}28.9\pm 2.7\phantom{.0}$
1100 $>$140 — — 990–1250 $<$15
$\phantom{0}0.1^{\phantom{0}+\phantom{0}0.2\phantom{0}\phantom{0}}_{\phantom{0}-\phantom{0}0.1\phantom{0}\phantom{0}}$
1 $\phantom{.0}36.7\pm 3.7\phantom{.0}$ $\phantom{.0}29.2\pm 2.7\phantom{.0}$
1200 $>$140 — — 1035–1430 $<$15
$\phantom{0}0.2^{\phantom{0}+\phantom{0}0.3\phantom{0}\phantom{0}}_{\phantom{0}-\phantom{0}0.2\phantom{0}\phantom{0}}$
1 $\phantom{.0}38.5\pm 4.0\phantom{.0}$ $\phantom{.0}30.1\pm 2.8\phantom{.0}$
1300 $>$140 — — 1100–1595 $<$15 $\phantom{0}0.3\pm 0.3\phantom{0}$ 1
$\phantom{.0}38.5\pm 4.0\phantom{.0}$ $\phantom{.0}30.7\pm 3.0\phantom{.0}$
1400 $>$140 — — 1285–1700 $<$15
$\phantom{0}0.1^{\phantom{0}+\phantom{0}0.2\phantom{0}\phantom{0}}_{\phantom{0}-\phantom{0}0.1\phantom{0}\phantom{0}}$
1 $\phantom{.0}35.9\pm 3.8\phantom{.0}$ $\phantom{.0}29.4\pm 2.8\phantom{.0}$
1500 $>$140 — — 1330–1800 $<$15
$\phantom{0}0.1^{\phantom{0}+\phantom{0}0.2\phantom{0}\phantom{0}}_{\phantom{0}-\phantom{0}0.1\phantom{0}\phantom{0}}$
1 $\phantom{.0}36.4\pm 3.9\phantom{.0}$ $\phantom{.0}30.0\pm 2.9\phantom{.0}$
Selection requirements on discriminating variables determined by the
optimization for each Majorana neutrino mass point in the $\Pe\mu$ channel
high-mass SRs. The last column shows the overall signal acceptance for the DY
and VBF channels. The quoted uncertainties include both the statistical and
systematic contributions. The dash indicates that no selection requirement is
made. $m_{\mathrm{N}}$ $p^{\ell_{1}}_{\mathrm{T}}$ $\pt^{\ell_{2}}$
$m(\ell^{\pm}\ell^{\pm}\PW_{\text{jet}})$ $m(\ell\PW_{\text{jet}})$
$(\ptmiss)^{2}/S_{\mathrm{T}}$ Total bkgd. $\mathrm{N}_{\text{obs}}$ DY
$A\epsilon$ VBF $A\epsilon$ () () () () () () (%) (%) $\Pe\mu$ channel SR1 85
$>$30 $>$10 $>$120 55–95 $<$7 $\phantom{}26.1\pm 6.2\phantom{0}$ 25
$\phantom{0.}0.21\pm 0.03\phantom{.}$ — 90 $>$30 $>$10 $>$120 60–100 $<$7
$\phantom{}37.4\pm 8.4\phantom{0}$ 32 $\phantom{0.}0.59\pm 0.07\phantom{.}$ —
100 $>$25 $>$20 $>$110 60–115 $<$7 $\phantom{}23.6\pm 4.8\phantom{0}$ 21
$\phantom{0.0}1.3\pm 0.1\phantom{.0}$ — 125 $>$30 $>$30 $>$140 90–140 $<$7
$\phantom{}25.5\pm 5.9\phantom{0}$ 16 $\phantom{0.0}3.1\pm 0.2\phantom{.0}$ —
150 $>$45 $>$35 $>$150 100–170 $<$7 $\phantom{}34.1\pm 6.0\phantom{0}$ 26
$\phantom{0.0}5.1\pm 0.3\phantom{.0}$ — 200 $>$65 $>$35 $>$270 170–230 $<$7
$\phantom{}11.1\pm 2.8\phantom{0}$ 14 $\phantom{0.0}6.1\pm 0.4\phantom{.0}$ —
250 $>$75 $>$60 $>$300 200–280 $<$7 $\phantom{}11.1\pm 2.3\phantom{0}$ 9
$\phantom{0.0}8.9\pm 0.5\phantom{.0}$ — 300 $>$95 $>$60 $>$340 255–325 $<$7
$\phantom{0}5.8\pm 1.7\phantom{0}$ 8 $\phantom{0.0}9.0\pm 0.6\phantom{.0}$
$\phantom{.0}3.4\pm 0.3\phantom{.0}$ 400 $>$120 $>$60 $>$530 325–450 $<$7
$\phantom{0}2.2\pm 1.0\phantom{0}$ 7 $\phantom{0.0}7.4\pm 0.4\phantom{.0}$
$\phantom{.0}3.0\pm 0.3\phantom{.0}$ 500 $>$150 $>$60 $>$580 315–530 $<$7
$\phantom{0}1.8\pm 1.1\phantom{0}$ 6 $\phantom{0.0}6.6\pm 0.5\phantom{.0}$
$\phantom{.0}3.0\pm 0.2\phantom{.0}$ 600 $>$175 — $>$670 315–740 $<$7
$\phantom{0}1.2\pm 0.9\phantom{0}$ 4 $\phantom{0.0}5.9\pm 0.4\phantom{.0}$
$\phantom{.0}3.5\pm 0.3\phantom{.0}$ 700 $>$180 — $>$720 350–1030 $<$7
$\phantom{0}1.6\pm 1.1\phantom{0}$ 3 $\phantom{0.0}5.2\pm 0.3\phantom{.0}$
$\phantom{.0}3.8\pm 0.2\phantom{.0}$ 800 $>$180 — $>$720 400–1030 $<$7
$\phantom{0}1.6\pm 1.1\phantom{0}$ 3 $\phantom{0.0}4.5\pm 0.3\phantom{.0}$
$\phantom{.0}3.7\pm 0.2\phantom{.0}$ 900 $>$185 — $>$720 450–1040 $<$7
$\phantom{0}1.0\pm 0.7\phantom{0}$ 2 $\phantom{0.0}3.8\pm 0.2\phantom{.0}$
$\phantom{.0}3.3\pm 0.2\phantom{.0}$ 1000 $>$185 — $>$720 500–1415 $<$7
$\phantom{0}1.0\pm 0.7\phantom{0}$ 2 $\phantom{0.0}3.4\pm 0.2\phantom{.0}$
$\phantom{.0}3.0\pm 0.2\phantom{.0}$ 1100 $>$185 — $>$720 550–1640 $<$7
$\phantom{0}1.0\pm 0.7\phantom{0}$ 1 $\phantom{0.0}2.8\pm 0.2\phantom{.0}$
$\phantom{.0}2.6\pm 0.2\phantom{.0}$ 1200 $>$185 — $>$720 600–1780 $<$7
$\phantom{0}1.0\pm 0.7\phantom{0}$ 1 $\phantom{0.0}2.4\pm 0.2\phantom{.0}$
$\phantom{.0}2.3\pm 0.2\phantom{.0}$ 1300 $>$185 — $>$720 650–1880 $<$7
$\phantom{0}0.8\pm 0.7\phantom{0}$ 1 $\phantom{0.0}2.1\pm 0.1\phantom{.0}$
$\phantom{.0}1.9\pm 0.2\phantom{.0}$ 1400 $>$185 — $>$720 650–1885 $<$7
$\phantom{0}0.8\pm 0.7\phantom{0}$ 1 $\phantom{0.0}1.8\pm 0.1\phantom{.0}$
$\phantom{.0}1.7\pm 0.2\phantom{.0}$ 1500 $>$185 — $>$720 650–1885 $<$7
$\phantom{0}0.8\pm 0.7\phantom{0}$ 1 $\phantom{0.0}1.5\pm 0.1\phantom{.0}$
$\phantom{.0}1.5\pm 0.1\phantom{.0}$ 1700 $>$185 — $>$720 650–2085 $<$7
$\phantom{0}0.8\pm 0.7\phantom{0}$ 1 $\phantom{0.0}1.2\pm 0.1\phantom{.0}$
$\phantom{.0}1.3\pm 0.1\phantom{.0}$ $\Pe\mu$ channel SR2 85 $>$25 $>$10 — —
$<$15 $\phantom{}24.2\pm 6.4\phantom{0}$ 31 $\phantom{0.}0.001\pm
0.002\phantom{.}$ — 90 $>$25 $>$10 — 90–240 $<$15 $\phantom{}13.4\pm
3.7\phantom{0}$ 22 $\phantom{0.}0.003\pm 0.002\phantom{.}$ — 100 $>$30 $>$15 —
100–335 $<$15 $\phantom{}14.1\pm 4.1\phantom{0}$ 21 $\phantom{0.}0.009\pm
0.003\phantom{.}$ — 125 $>$35 $>$25 — 115–150 $<$15 $\phantom{0}0.6\pm
0.4\phantom{0}$ 2 $\phantom{0.}0.03\pm 0.01\phantom{.}$ — 150 $>$45 $>$30 —
132–180 $<$15 $\phantom{0}1.4\pm 0.5\phantom{0}$ 2 $\phantom{0.}0.14\pm
0.02\phantom{.}$ — 200 $>$70 $>$30 — 180–225 $<$15 $\phantom{0}1.5\pm
0.5\phantom{0}$ 3 $\phantom{0.}0.86\pm 0.09\phantom{.}$ — 250 $>$75 $>$55 —
225–280 $<$15 $\phantom{0}1.2\pm 0.4\phantom{0}$ 2 $\phantom{0.0}1.7\pm
0.2\phantom{.0}$ — 300 $>$95 $>$55 — 280–340 $<$15 $\phantom{0}1.2\pm
0.7\phantom{0}$ 1 $\phantom{0.0}4.4\pm 0.4\phantom{.0}$ $\phantom{0.0}0.8\pm
0.1\phantom{.0}$ 400 $>$125 $>$55 — 340–475 $<$15 $\phantom{0}2.0\pm
1.2\phantom{0}$ 1 $\phantom{.0}11.8\pm 1.1\phantom{.0}$ $\phantom{0.0}2.7\pm
0.3\phantom{.0}$ 500 $>$145 $>$60 — 460–555 $<$15 $\phantom{0}0.7\pm
0.3\phantom{0}$ 0 $\phantom{.0}16.7\pm 1.6\phantom{.0}$ $\phantom{0.0}5.2\pm
0.5\phantom{.0}$ 600 $>$160 — — 555–645 $<$15 $\phantom{0}1.4\pm
0.9\phantom{0}$ 1 $\phantom{.0}20.2\pm 1.9\phantom{.0}$ $\phantom{.0}13.2\pm
1.2\phantom{.0}$ 700 $>$170 — — 610–780 $<$15 $\phantom{0}2.0\pm
0.9\phantom{0}$ 2 $\phantom{.0}25.0\pm 2.4\phantom{.0}$ $\phantom{.0}17.6\pm
1.6\phantom{.0}$ 800 $>$170 — — 730–895 $<$15 $\phantom{0}0.8\pm
0.4\phantom{0}$ 2 $\phantom{.0}26.1\pm 2.5\phantom{.0}$ $\phantom{.0}18.3\pm
1.6\phantom{.0}$ 900 $>$180 — — 845–1015 $<$15 $\phantom{0}0.5\pm
0.2\phantom{0}$ 0 $\phantom{.0}25.6\pm 2.5\phantom{.0}$ $\phantom{.0}18.5\pm
1.7\phantom{.0}$ 1000 $>$180 — — 930–1075 $<$15 $\phantom{0}0.2\pm
0.2\phantom{0}$ 0 $\phantom{.0}23.5\pm 2.3\phantom{.0}$ $\phantom{.0}17.6\pm
1.6\phantom{.0}$ 1100 $>$180 — — 1020–1340 $<$15 $\phantom{0}0.3\pm
0.3\phantom{0}$ 0 $\phantom{.0}26.9\pm 2.7\phantom{.0}$ $\phantom{.0}19.6\pm
1.7\phantom{.0}$ 1200 $>$180 — — 1080–1340 $<$15
$\phantom{0}0.1^{\phantom{0}+\phantom{0}0.2\phantom{0}\phantom{0}}_{\phantom{0}-\phantom{0}0.1\phantom{0}\phantom{0}}$
0 $\phantom{.0}25.9\pm 2.6\phantom{.0}$ $\phantom{.0}19.9\pm 1.8\phantom{.0}$
1300 $>$180 — — 1155–1595 $<$15
$\phantom{0}0.2^{\phantom{0}+\phantom{0}0.2\phantom{0}\phantom{0}}_{\phantom{0}-\phantom{0}0.2\phantom{0}\phantom{0}}$
0 $\phantom{.0}27.1\pm 2.7\phantom{.0}$ $\phantom{.0}20.7\pm 1.9\phantom{.0}$
1400 $>$180 — — 1155–1615 $<$15
$\phantom{0}0.2^{\phantom{0}+\phantom{0}0.3\phantom{0}\phantom{0}}_{\phantom{0}-\phantom{0}0.2\phantom{0}\phantom{0}}$
0 $\phantom{.0}26.7\pm 2.7\phantom{.0}$ $\phantom{.0}20.8\pm 2.0\phantom{.0}$
1500 $>$180 — — 1345–1615 $<$15
$\phantom{0}0.0^{\phantom{0}+\phantom{0}0.1\phantom{0}\phantom{0}}_{\phantom{0}-\phantom{0}0.0\phantom{0}\phantom{0}}$
0 $\phantom{.0}21.6\pm 2.2\phantom{.0}$ $\phantom{.0}18.0\pm 1.7\phantom{.0}$
1700 $>$180 — — 1400–1800 $<$15 $\phantom{0}0.7\pm 0.6\phantom{0}$ 0
$\phantom{.0}19.8\pm 2.1\phantom{.0}$ $\phantom{.0}17.0\pm 1.7\phantom{.0}$
Figure 4: Exclusion region at 95% CL in the $\abs{V_{\Pe\mathrm{N}}}^{2}$
(upper) and $\abs{V_{\mu\mathrm{N}}}^{2}$ (lower) vs. $m_{\mathrm{N}}$ plane.
The dashed black curve is the expected upper limit, with one and two standard-
deviation bands shown in green and yellow, respectively. The solid black curve
is the observed upper limit. The brown line shows constraints from EWPD [83].
Also shown are the upper limits from other direct searches: DELPHI [30], L3
[31, 32], ATLAS [36], and the upper limits from the CMS $\sqrt{s}=8\TeV$ 2012
data [35] and the trilepton analysis [37] based on the same 2016 data set as
used in this analysis. Figure 5: Exclusion region at 95% CL in the
$\abs{V_{\Pe\mathrm{N}}V^{*}_{\mu\mathrm{N}}}^{2}/(\abs{V_{\Pe\mathrm{N}}}^{2}+\abs{V_{\mu\mathrm{N}}}^{2})$
vs. $m_{\mathrm{N}}$ plane. The dashed black curve is the expected upper
limit, with one and two standard-deviation bands shown in green and yellow,
respectively. The solid black curve is the observed upper limit. Also shown
are the upper limits from the CMS $\sqrt{s}=8\TeV$ 2012 data [35].
## 0.9 Summary
A search for heavy Majorana neutrinos, $\mathrm{N}$, in final states with
same-sign dileptons and jets has been performed in proton-proton collisions at
a center-of-mass energy of 13, using a data set corresponding to an integrated
luminosity of 35.9. No significant excess of events compared to the expected
standard model background prediction is observed. Upper limits at 95%
confidence level are set on the mixing matrix element between standard model
neutrinos and $\mathrm{N}$ ($\abs{V_{\ell\mathrm{N}}}$) in the context of a
Type-I seesaw model, as a function of $\mathrm{N}$ mass. The analysis improves
on previous 8searches by including single-jet events into the signal region,
which increases sensitivities. For an $\mathrm{N}$ mass of 40the observed
(expected) limits are $\abs{V_{\Pe\mathrm{N}}}^{2}<9.5\,(8.0)\times 10^{-5}$,
$\abs{V_{\mu\mathrm{N}}}^{2}<2.3\,(1.9)\times 10^{-5}$, and
$\abs{V_{\Pe\mathrm{N}}V^{*}_{\mu\mathrm{N}}}^{2}/(\abs{V_{\Pe\mathrm{N}}}^{2}+\abs{V_{\mu\mathrm{N}}}^{2})<2.7\,(2.7)\times
10^{-5}$, and for an $\mathrm{N}$ mass of 1000the limits are
$\abs{V_{\Pe\mathrm{N}}}^{2}<0.42\,(0.32)$,
$\abs{V_{\mu\mathrm{N}}}^{2}<0.27\,(0.16)$, and
$\abs{V_{\Pe\mathrm{N}}V^{*}_{\mu\mathrm{N}}}^{2}/(\abs{V_{\Pe\mathrm{N}}}^{2}+\abs{V_{\mu\mathrm{N}}}^{2})<0.14\,(0.14)$.
The search is sensitive to masses of $\mathrm{N}$ from 20 to 1600. The limits
on the mixing matrix elements are placed up to 1240for
$\abs{V_{\Pe\mathrm{N}}}^{2}$, 1430for the $\abs{V_{\mu\mathrm{N}}}^{2}$, and
1600for
$\abs{V_{\Pe\mathrm{N}}V^{*}_{\mu\mathrm{N}}}^{2}/(\abs{V_{\Pe\mathrm{N}}}^{2}+\abs{V_{\mu\mathrm{N}}}^{2})$.
These are the most restrictive direct limits on the $\mathrm{N}$ mixing
parameters for heavy Majorana neutrino masses greater than 430, and are the
first for masses greater than 1200.
###### Acknowledgements.
We congratulate our colleagues in the CERN accelerator departments for the
excellent performance of the LHC and thank the technical and administrative
staffs at CERN and at other CMS institutes for their contributions to the
success of the CMS effort. In addition, we gratefully acknowledge the
computing centers and personnel of the Worldwide LHC Computing Grid for
delivering so effectively the computing infrastructure essential to our
analyses. Finally, we acknowledge the enduring support for the construction
and operation of the LHC and the CMS detector provided by the following
funding agencies: the Austrian Federal Ministry of Science, Research and
Economy and the Austrian Science Fund; the Belgian Fonds de la Recherche
Scientifique, and Fonds voor Wetenschappelijk Onderzoek; the Brazilian Funding
Agencies (CNPq, CAPES, FAPERJ, FAPERGS, and FAPESP); the Bulgarian Ministry of
Education and Science; CERN; the Chinese Academy of Sciences, Ministry of
Science and Technology, and National Natural Science Foundation of China; the
Colombian Funding Agency (COLCIENCIAS); the Croatian Ministry of Science,
Education and Sport, and the Croatian Science Foundation; the Research
Promotion Foundation, Cyprus; the Secretariat for Higher Education, Science,
Technology and Innovation, Ecuador; the Ministry of Education and Research,
Estonian Research Council via IUT23-4 and IUT23-6 and European Regional
Development Fund, Estonia; the Academy of Finland, Finnish Ministry of
Education and Culture, and Helsinki Institute of Physics; the Institut
National de Physique Nucléaire et de Physique des Particules / CNRS, and
Commissariat à l’Énergie Atomique et aux Énergies Alternatives / CEA, France;
the Bundesministerium für Bildung und Forschung, Deutsche
Forschungsgemeinschaft, and Helmholtz-Gemeinschaft Deutscher
Forschungszentren, Germany; the General Secretariat for Research and
Technology, Greece; the National Research, Development and Innovation Fund,
Hungary; the Department of Atomic Energy and the Department of Science and
Technology, India; the Institute for Studies in Theoretical Physics and
Mathematics, Iran; the Science Foundation, Ireland; the Istituto Nazionale di
Fisica Nucleare, Italy; the Ministry of Science, ICT and Future Planning, and
National Research Foundation (NRF), Republic of Korea; the Lithuanian Academy
of Sciences; the Ministry of Education, and University of Malaya (Malaysia);
the Ministry of Science of Montenegro; the Mexican Funding Agencies (BUAP,
CINVESTAV, CONACYT, LNS, SEP, and UASLP-FAI); the Ministry of Business,
Innovation and Employment, New Zealand; the Pakistan Atomic Energy Commission;
the Ministry of Science and Higher Education and the National Science Center,
Poland; the Fundação para a Ciência e a Tecnologia, Portugal; JINR, Dubna; the
Ministry of Education and Science of the Russian Federation, the Federal
Agency of Atomic Energy of the Russian Federation, Russian Academy of
Sciences, the Russian Foundation for Basic Research, and the National Research
Center “Kurchatov Institute”; the Ministry of Education, Science and
Technological Development of Serbia; the Secretaría de Estado de
Investigación, Desarrollo e Innovación, Programa Consolider-Ingenio 2010, Plan
Estatal de Investigación Científica y Técnica y de Innovación 2013-2016, Plan
de Ciencia, Tecnología e Innovación 2013-2017 del Principado de Asturias, and
Fondo Europeo de Desarrollo Regional, Spain; the Ministry of Science,
Technology and Research, Sri Lanka; the Swiss Funding Agencies (ETH Board, ETH
Zurich, PSI, SNF, UniZH, Canton Zurich, and SER); the Ministry of Science and
Technology, Taipei; the Thailand Center of Excellence in Physics, the
Institute for the Promotion of Teaching Science and Technology of Thailand,
Special Task Force for Activating Research and the National Science and
Technology Development Agency of Thailand; the Scientific and Technical
Research Council of Turkey, and Turkish Atomic Energy Authority; the National
Academy of Sciences of Ukraine, and State Fund for Fundamental Researches,
Ukraine; the Science and Technology Facilities Council, UK; the US Department
of Energy, and the US National Science Foundation. Individuals have received
support from the Marie-Curie program and the European Research Council and
Horizon 2020 Grant, contract No. 675440 (European Union); the Leventis
Foundation; the A. P. Sloan Foundation; the Alexander von Humboldt Foundation;
the Belgian Federal Science Policy Office; the Fonds pour la Formation à la
Recherche dans l’Industrie et dans l’Agriculture (FRIA-Belgium); the
Agentschap voor Innovatie door Wetenschap en Technologie (IWT-Belgium); the
F.R.S.-FNRS and FWO (Belgium) under the “Excellence of Science - EOS” - be.h
project n. 30820817; the Ministry of Education, Youth and Sports (MEYS) of the
Czech Republic; the Lendület (“Momentum”) Program and the János Bolyai
Research Scholarship of the Hungarian Academy of Sciences, the New National
Excellence Program ÚNKP, the NKFIA research grants 123842, 123959, 124845,
124850 and 125105 (Hungary); the Council of Scientific and Industrial
Research, India; the HOMING PLUS program of the Foundation for Polish Science,
cofinanced from European Union, Regional Development Fund, the Mobility Plus
program of the Ministry of Science and Higher Education, the National Science
Center (Poland), contracts Harmonia 2014/14/M/ST2/00428, Opus
2014/13/B/ST2/02543, 2014/15/B/ST2/03998, and 2015/19/B/ST2/02861, Sonata-bis
2012/07/E/ST2/01406; the National Priorities Research Program by Qatar
National Research Fund; the Programa de Excelencia María de Maeztu, and the
Programa Severo Ochoa del Principado de Asturias; the Thalis and Aristeia
programs cofinanced by EU-ESF, and the Greek NSRF; the Rachadapisek Sompot
Fund for Postdoctoral Fellowship, Chulalongkorn University, and the
Chulalongkorn Academic into Its 2nd Century Project Advancement Project
(Thailand); the Welch Foundation, contract C-1845; and the Weston Havens
Foundation (USA).
## References
* [1] Particle Data Group, C. Patrignani et al., “Review of particle physics”, Chin. Phys. C 40 (2016) 100001, 10.1088/1674-1137/40/10/100001.
* [2] J. A. McCarthy, “Search for double beta decay in Ca48”, Phys. Rev. 97 (1955) 1234, 10.1103/PhysRev.97.1234.
* [3] V. R. Lazarenko and S. Y. Luk’yanov, “An attempt to detect double beta decay in Ca48”, Sov. Phys. JETP 22 (1966) 521.
* [4] E. Ma, “Pathways to naturally small neutrino masses”, Phys. Rev. Lett. 81 (1998) 1171, 10.1103/PhysRevLett.81.1171, arXiv:hep-ph/9805219.
* [5] Y. Cai, T. Han, T. Li, and R. Ruiz, “Lepton-number violation: Seesaw models and their collider tests”, Front. in Phys. 6 (2018) 40, 10.3389/fphy.2018.00040, arXiv:1711.02180.
* [6] P. Minkowski, “$\mu\to\mathrm{e}\gamma$ at a rate of one out of 1-billion muon decays?”, Phys. Lett. B 67 (1977) 421, 10.1016/0370-2693(77)90435-X.
* [7] M. Gell-Mann, P. Ramond, and R. Slansky, “Supergravity: proceedings of the supergravity workshop at Stony Brook”, in Supergravity: proceedings of the Supergravity Workshop at Stony Brook, p. 341. 1979\.
* [8] T. Yanagida, “Proceedings of the workshop on the unified theory and the baryon number in the universe”, in Proceedings of the Workshop on the Unified Theory and the Baryon Number in the Universe, p. 95. National Laboratory for High Energy Physics (KEK), 1979.
* [9] R. N. Mohapatra and G. Senjanović, “Neutrino mass and spontaneous parity violation”, Phys. Rev. Lett. 44 (1980) 912, 10.1103/PhysRevLett.44.912.
* [10] J. Schechter and J. W. F. Valle, “Neutrino masses in SU(2) $\bigotimes$ U(1) theories”, Phys. Rev. D 22 (1980) 2227, 10.1103/PhysRevD.22.2227.
* [11] R. E. Shrock, “General theory of weak processes involving neutrinos. i. leptonic pseudoscalar-meson decays, with associated tests for, and bounds on, neutrino masses and lepton mixing”, Phys. Rev. D 24 (1981) 1232, 10.1103/PhysRevD.24.1232.
* [12] J. Schechter and J. W. F. Valle, “Neutrino decay and spontaneous violation of lepton number”, Phys. Rev. D 25 (1982) 774, 10.1103/PhysRevD.25.774.
* [13] M. Magg and C. Wetterich, “Neutrino mass problem and gauge hierarchy”, Phys. Lett. B 94 (1980) 61, 10.1016/0370-2693(80)90825-4.
* [14] R. N. Mohapatra and G. Senjanović, “Neutrino masses and mixings in gauge models with spontaneous parity violation”, Phys. Rev. D 23 (1981) 165, 10.1103/PhysRevD.23.165.
* [15] R. Foot, H. Lew, X.-G. He, and G. C. Joshi, “See-saw neutrino masses induced by a triplet of leptons”, Z. Phys. C 44 (1989) 441, 10.1007/BF01415558.
* [16] R. N. Mohapatra, “Mechanism for understanding small neutrino mass in superstring theories”, Phys. Rev. Lett. 56 (1986) 561, 10.1103/PhysRevLett.56.561.
* [17] R. N. Mohapatra and J. W. F. Valle, “Neutrino mass and baryon-number nonconservation in superstring models”, Phys. Rev. D 34 (1986) 1642, 10.1103/PhysRevD.34.1642.
* [18] J. Bernabéu et al., “Lepton flavour non-conservation at high energies in a superstring inspired standard model”, Phys. Lett. B 187 (1987) 303, 10.1016/0370-2693(87)91100-2.
* [19] M. Lindner and M. Weiser, “Gauge coupling unification in left-right symmetric models”, Phys. Lett. B 383 (1996) 405, 10.1016/0370-2693(96)00775-7, arXiv:hep-ph/9605353.
* [20] T. Appelquist and R. Shrock, “Neutrino masses in theories with dynamical electroweak symmetry breaking”, Phys. Lett. B 548 (2002) 204, 10.1016/S0370-2693(02)02854-X, arXiv:hep-ph/0204141.
* [21] T. Appelquist and R. Shrock, “Dynamical symmetry breaking of extended gauge symmetries”, Phys. Rev. Lett. 90 (2003) 201801, 10.1103/PhysRevLett.90.201801, arXiv:hep-ph/0301108.
* [22] T. Asaka, S. Blanchet, and M. Shaposhnikov, “The nuMSM, dark matter and neutrino masses”, Phys. Lett. B 631 (2005) 151, 10.1016/j.physletb.2005.09.070, arXiv:hep-ph/0503065.
* [23] T. Asaka and M. Shaposhnikov, “The nuMSM, dark matter and baryon asymmetry of the universe”, Phys. Lett. B 620 (2005) 17, 10.1016/j.physletb.2005.06.020, arXiv:hep-ph/0505013.
* [24] W.-Y. Keung and G. Senjanovic, “Majorana neutrinos and the production of the right-handed charged gauge boson”, Phys. Rev. Lett. 50 (1983) 1427, 10.1103/PhysRevLett.50.1427.
* [25] D. A. Dicus, D. D. Karatas, and P. Roy, “Lepton nonconservation at supercollider energies”, Phys. Rev. D 44 (1991) 2033, 10.1103/PhysRevD.44.2033.
* [26] A. Datta, M. Guchait, and A. Pilaftsis, “Probing lepton number violation via Majorana neutrinos at hadron supercolliders”, Phys. Rev. D 50 (1994) 3195, 10.1103/PhysRevD.50.3195, arXiv:hep-ph/9311257.
* [27] F. M. L. Almeida, Y. A. Coutinho, J. A. M. Simões, and M. A. B. d. Vale, “Signature for heavy Majorana neutrinos in hadronic collisions”, Phys. Rev. D 62 (2000) 075004, 10.1103/PhysRevD.62.075004, arXiv:hep-ph/0002024.
* [28] O. Panella, M. Cannoni, C. Carimalo, and Y. N. Srivastava, “Signals of heavy Majorana neutrinos at hadron colliders”, Phys. Rev. D 65 (2002) 035005, 10.1103/PhysRevD.65.035005, arXiv:hep-ph/0107308.
* [29] F. F. Deppisch, P. S. Bhupal Dev, and A. Pilaftsis, “Neutrinos and collider physics”, New J. Phys. 17 (2015) 075019, 10.1088/1367-2630/17/7/075019, arXiv:1502.06541.
* [30] DELPHI Collaboration, “Search for neutral heavy leptons produced in Z decays”, Z. Phys. C 74 (1997) 57, 10.1007/s002880050370.
* [31] L3 Collaboration, “Search for isosinglet neutral heavy leptons in Z0 decays”, Phys. Lett. B 295 (1992) 371, 10.1016/0370-2693(92)91579-X.
* [32] L3 Collaboration, “Search for heavy isosinglet neutrino in $\mathrm{e}^{+}\mathrm{e}^{-}$ annihilation at LEP”, Phys. Lett. B 517 (2001) 67, 10.1016/S0370-2693(01)00993-5, arXiv:hep-ex/0107014.
* [33] LHCb Collaboration, “Search for Majorana neutrinos in $B^{-}\to\pi^{+}\mu^{-}\mu^{-}$ decays”, Phys. Rev. Lett. 112 (2014) 131802, 10.1103/PhysRevLett.112.131802, arXiv:1401.5361.
* [34] CMS Collaboration, “Search for heavy Majorana neutrinos in $\mu^{\pm}\mu^{\pm}+$ jets events in proton-proton collisions at $\sqrt{s}$ = 8 TeV”, Phys. Lett. B 748 (2015) 144, 10.1016/j.physletb.2015.06.070, arXiv:1501.05566.
* [35] CMS Collaboration, “Search for heavy Majorana neutrinos in ee + jets and e$\mu$ \+ jets events in proton-proton collisions at $\sqrt{s}$ = 8 TeV”, JHEP 04 (2016) 169, 10.1007/JHEP04(2016)169, arXiv:1603.02248.
* [36] ATLAS Collaboration, “Search for heavy Majorana neutrinos with the ATLAS detector in pp collisions at $\sqrt{s}$ = 8 TeV”, JHEP 07 (2015) 162, 10.1007/JHEP07(2015)162, arXiv:1506.06020.
* [37] CMS Collaboration, “Search for heavy neutral leptons in events with three charged leptons in proton-proton collisions at $\sqrt{s}=13\text{ }\text{ }\mathrm{TeV}$”, Phys. Rev. Lett. 120 (2018) 221801, 10.1103/PhysRevLett.120.221801, arXiv:1802.02965.
* [38] A. Atre, T. Han, S. Pascoli, and B. Zhang, “The search for heavy Majorana neutrinos”, JHEP 05 (2009) 030, 10.1088/1126-6708/2009/05/030, arXiv:0901.3589.
* [39] P. S. B. Dev, A. Pilaftsis, and U.-k. Yang, “New production mechanism for heavy neutrinos at the LHC”, Phys. Rev. Lett. 112 (2014) 081801, 10.1103/PhysRevLett.112.081801, arXiv:1308.2209.
* [40] D. Alva, T. Han, and R. Ruiz, “Heavy Majorana neutrinos from W$\gamma$ fusion at hadron colliders”, JHEP 02 (2015) 072, 10.1007/JHEP02(2015)072, arXiv:1411.7305.
* [41] A. Das, P. Konar, and A. Thalapillil, “Jet substructure shedding light on heavy Majorana neutrinos at the LHC”, JHEP 02 (2018) 083, 10.1007/JHEP02(2018)083, arXiv:1709.09712.
* [42] CMS Collaboration, “The CMS trigger system”, JINST 12 (2017) P01020, 10.1088/1748-0221/12/01/P01020, arXiv:1609.02366.
* [43] CMS Collaboration, “The CMS experiment at the CERN LHC”, JINST 3 (2008) S08004, 10.1088/1748-0221/3/08/S08004.
* [44] J. Alwall et al., “The automated computation of tree-level and next-to-leading order differential cross sections, and their matching to parton shower simulations”, JHEP 07 (2014) 079, 10.1007/JHEP07(2014)079, arXiv:1405.0301.
* [45] J. M. Campbell and R. K. Ellis, “MCFM for the Tevatron and the LHC”, Nucl. Phys. Proc. Suppl. 205-206 (2010) 10, 10.1016/j.nuclphysbps.2010.08.011, arXiv:1007.3492.
* [46] P. Nason, “A New method for combining NLO QCD with shower Monte Carlo algorithms”, JHEP 11 (2004) 040, 10.1088/1126-6708/2004/11/040, arXiv:hep-ph/0409146.
* [47] S. Frixione, P. Nason, and C. Oleari, “Matching NLO QCD computations with parton shower simulations: the POWHEG method”, JHEP 11 (2007) 070, 10.1088/1126-6708/2007/11/070, arXiv:0709.2092.
* [48] S. Alioli, P. Nason, C. Oleari, and E. Re, “A general framework for implementing NLO calculations in shower Monte Carlo programs: the POWHEG BOX”, JHEP 06 (2010) 043, 10.1007/JHEP06(2010)043, arXiv:1002.2581.
* [49] P. Nason and G. Zanderighi, “$W^{+}W^{-}$ , $WZ$ and $ZZ$ production in the POWHEG-BOX-V2”, Eur. Phys. J. C 74 (2014) 2702, 10.1140/epjc/s10052-013-2702-5, arXiv:1311.1365.
* [50] R. D. Ball et al., “Parton distributions for the LHC run II”, JHEP 04 (2015) 040, 10.1007/JHEP04(2015)040, arXiv:1410.8849.
* [51] T. Sj$\ddot{\mathrm{o}}$strand et al., “An introduction to PYTHIA 8.2”, Comput. Phys. Commun. 191 (2015) 159, 10.1016/j.cpc.2015.01.024, arXiv:1410.3012.
* [52] CMS Collaboration, “Event generator tunes obtained from underlying event and multiparton scattering measurements”, Eur. Phys. J. C 76 (2016) 155, 10.1140/epjc/s10052-016-3988-x, arXiv:1512.00815.
* [53] GEANT4 Collaboration, “—a simulation toolkit”, Nucl. Instrum. Meth. A 506 (2003) 250, 10.1016/S0168-9002(03)01368-8.
* [54] J. Alwall et al., “Comparative study of various algorithms for the merging of parton showers and matrix elements in hadronic collisions”, Eur. Phys. J. C 53 (2008) 473, 10.1140/epjc/s10052-007-0490-5, arXiv:0706.2569.
* [55] R. Frederix and S. Frixione, “Merging meets matching in MC@NLO”, JHEP 12 (2012) 061, 10.1007/JHEP12(2012)061, arXiv:1209.6215.
* [56] P. Artoisenet, R. Frederix, O. Mattelaer, and R. Rietkerk, “Automatic spin-entangled decays of heavy resonances in monte carlo simulations”, JHEP 03 (2013) 15, 10.1007/JHEP03(2013)015, arXiv:1212.3460.
* [57] C. Degrande, O. Mattelaer, R. Ruiz, and J. Turner, “Fully automated precision predictions for heavy neutrino production mechanisms at hadron colliders”, Phys. Rev. D 94 (2016) 053002, 10.1103/PhysRevD.94.053002, arXiv:1602.06957.
* [58] A. Das, P. Konar, and S. Majhi, “Production of heavy neutrino in next-to-leading order QCD at the LHC and beyond”, JHEP 06 (2016) 019, 10.1007/JHEP06(2016)019, arXiv:1604.00608.
* [59] A. Manohar, P. Nason, G. P. Salam, and G. Zanderighi, “How bright is the proton? a precise determination of the photon parton distribution function”, Phys. Rev. Lett. 117 (2016) 242002, 10.1103/PhysRevLett.117.242002, arXiv:1607.04266.
* [60] M. Cacciari, G. P. Salam, and G. Soyez, “The anti-$k_{\mathrm{T}}$ jet clustering algorithm”, JHEP 04 (2008) 063, 10.1088/1126-6708/2008/04/063, arXiv:0802.1189.
* [61] M. Cacciari, G. P. Salam, and G. Soyez, “FastJet user manual”, Eur. Phys. J. C 72 (2012) 1896, 10.1140/epjc/s10052-012-1896-2, arXiv:1111.6097.
* [62] CMS Collaboration, “Particle-flow reconstruction and global event description with the CMS detector”, JINST 12 (2017) P10003, 10.1088/1748-0221/12/10/P10003, arXiv:1706.04965.
* [63] CMS Collaboration, “Performance of electron reconstruction and selection with the CMS detector in proton-proton collisions at $\sqrt{s}$ = 8 TeV”, JINST 10 (2015) P06005, 10.1088/1748-0221/10/06/P06005, arXiv:1502.02701.
* [64] CMS Collaboration, “Performance of CMS muon reconstruction in pp collision events at $\sqrt{s}$ = 7 TeV”, JINST 7 (2012) P10002, 10.1088/1748-0221/7/10/P10002, arXiv:1206.4071.
* [65] E. Akhmedov et al., “Improving electro-weak fits with TeV-scale sterile neutrinos”, JHEP 05 (2013) 081, 10.1007/JHEP05(2013)081, arXiv:1302.1872.
* [66] CMS Collaboration, “Jet energy scale and resolution in the CMS experiment in pp collisions at 8 TeV”, JINST 12 (2017) P02014, 10.1088/1748-0221/12/02/P02014, arXiv:1607.03663.
* [67] CMS Collaboration, “Jet algorithms performance in 13 TeV data”, CMS Physics Analysis Summary CMS-PAS-JME-16-003, CERN, 2017.
* [68] S. D. Ellis, C. K. Vermilion, and J. R. Walsh, “Techniques for improved heavy particle searches with jet substructure”, Phys. Rev. D 80 (2009) 051501, 10.1103/PhysRevD.80.051501, arXiv:0903.5081.
* [69] S. D. Ellis, C. K. Vermilion, and J. R. Walsh, “Recombination algorithms and jet substructure: Pruning as a tool for heavy particle searches”, Phys. Rev. D 81 (2010) 094023, 10.1103/PhysRevD.81.094023, arXiv:0912.0033.
* [70] Y. L. Dokshitzer, G. D. Leder, S. Moretti, and B. R. Webber, “Better jet clustering algorithms”, JHEP 08 (1997) 001, 10.1088/1126-6708/1997/08/001, arXiv:hep-ph/9707323.
* [71] M. Wobisch and T. Wengler, “Hadronization corrections to jet cross-sections in deep inelastic scattering”, in Monte Carlo generators for HERA physics. Proceedings, Workshop, Hamburg, Germany, 1998-1999, p. 270. 1998\. arXiv:hep-ph/9907280.
* [72] J. Thaler and K. Van Tilburg, “Maximizing boosted top identification by minimizing $N$-subjettiness”, JHEP 02 (2012) 093, 10.1007/JHEP02(2012)093, arXiv:1108.2701.
* [73] CMS Collaboration, “Identification of heavy-flavour jets with the CMS detector in pp collisions at 13 TeV”, JINST 13 (2018) P05011, 10.1088/1748-0221/13/05/P05011, arXiv:1712.07158.
* [74] G. Punzi, “Sensitivity of searches for new signals and its optimization”, in Statistical Problems in Particle Physics, Astrophysics, and Cosmology, p. 79. 2003\. arXiv:physics/0308063.
* [75] CMS Collaboration, “Performance of CMS muon reconstruction in cosmic-ray events”, JINST 5 T03022.
* [76] A. Bierweiler, T. Kasprzik, and J. H. K$\ddot{\mathrm{u}}$hn, “Vector-boson pair production at the LHC to $\mathcal{O}$($\alpha^{3}$) accuracy”, JHEP 12 (2013) 071, 10.1007/JHEP12(2013)071, arXiv:1305.5402.
* [77] S. Gieseke, T. Kasprzik, and J. H. K$\ddot{\mathrm{u}}$hn, “Vector-boson pair production and electroweak corrections in HERWIG++”, Eur. Phys. J. C 74 (2014) 2988, 10.1140/epjc/s10052-014-2988-y, arXiv:1401.3964.
* [78] J. Baglio, L. D. Ninh, and M. M. Weber, “Massive gauge boson pair production at the LHC: A next-to-leading order story”, Phys. Rev. D 88 (2013) 113005, 10.1103/PhysRevD.88.113005, arXiv:1307.4331.
* [79] CMS Collaboration, “CMS luminosity measurements for the 2016 data taking period”, CMS Physics Analysis Summary CMS-PAS-LUM-17-001, CERN, 2017.
* [80] J. Butterworth et al., “PDF4LHC recommendations for LHC run II”, J. Phys. G 43 (2016) 023001, 10.1088/0954-3899/43/2/023001, arXiv:1510.03865.
* [81] T. Junk, “Confidence level computation for combining searches with small statistics”, Nucl. Instrum. Meth. A 434 (1999) 435, 10.1016/S0168-9002(99)00498-2, arXiv:hep-ex/9902006.
* [82] A. L. Read, “Presentation of search results: the CLs technique”, J. Phys. G 28 2693\.
* [83] J. de Blas, “Electroweak limits on physics beyond the Standard Model”, in European Physical Journal Web of Conferences, volume 60 of European Physical Journal Web of Conferences, p. 19008. nov, 2013. arXiv:1307.6173. 10.1051/epjconf/20136019008.
## .10 The CMS Collaboration
Yerevan Physics Institute, Yerevan, Armenia
A.M. Sirunyan, A. Tumasyan Institut für Hochenergiephysik, Wien, Austria
W. Adam, F. Ambrogi, E. Asilar, T. Bergauer, J. Brandstetter, M. Dragicevic,
J. Erö, A. Escalante Del Valle, M. Flechl, R. Frühwirth1, V.M. Ghete, J.
Hrubec, M. Jeitler1, N. Krammer, I. Krätschmer, D. Liko, T. Madlener, I.
Mikulec, N. Rad, H. Rohringer, J. Schieck1, R. Schöfbeck, M. Spanring, D.
Spitzbart, A. Taurok, W. Waltenberger, J. Wittmann, C.-E. Wulz1, M. Zarucki
Institute for Nuclear Problems, Minsk, Belarus
V. Chekhovsky, V. Mossolov, J. Suarez Gonzalez Universiteit Antwerpen,
Antwerpen, Belgium
E.A. De Wolf, D. Di Croce, X. Janssen, J. Lauwers, M. Pieters, M. Van De
Klundert, H. Van Haevermaet, P. Van Mechelen, N. Van Remortel Vrije
Universiteit Brussel, Brussel, Belgium
S. Abu Zeid, F. Blekman, J. D’Hondt, I. De Bruyn, J. De Clercq, K. Deroover,
G. Flouris, D. Lontkovskyi, S. Lowette, I. Marchesini, S. Moortgat, L.
Moreels, Q. Python, K. Skovpen, S. Tavernier, W. Van Doninck, P. Van Mulders,
I. Van Parijs Université Libre de Bruxelles, Bruxelles, Belgium
D. Beghin, B. Bilin, H. Brun, B. Clerbaux, G. De Lentdecker, H. Delannoy, B.
Dorney, G. Fasanella, L. Favart, R. Goldouzian, A. Grebenyuk, A.K. Kalsi, T.
Lenzi, J. Luetic, N. Postiau, E. Starling, L. Thomas, C. Vander Velde, P.
Vanlaer, D. Vannerom, Q. Wang Ghent University, Ghent, Belgium
T. Cornelis, D. Dobur, A. Fagot, M. Gul, I. Khvastunov2, D. Poyraz, C. Roskas,
D. Trocino, M. Tytgat, W. Verbeke, B. Vermassen, M. Vit, N. Zaganidis
Université Catholique de Louvain, Louvain-la-Neuve, Belgium
H. Bakhshiansohi, O. Bondu, S. Brochet, G. Bruno, C. Caputo, P. David, C.
Delaere, M. Delcourt, B. Francois, A. Giammanco, G. Krintiras, V. Lemaitre, A.
Magitteri, A. Mertens, M. Musich, K. Piotrzkowski, A. Saggio, M. Vidal Marono,
S. Wertz, J. Zobec Centro Brasileiro de Pesquisas Fisicas, Rio de Janeiro,
Brazil
F.L. Alves, G.A. Alves, L. Brito, M. Correa Martins Junior, G. Correia Silva,
C. Hensel, A. Moraes, M.E. Pol, P. Rebello Teles Universidade do Estado do Rio
de Janeiro, Rio de Janeiro, Brazil
E. Belchior Batista Das Chagas, W. Carvalho, J. Chinellato3, E. Coelho, E.M.
Da Costa, G.G. Da Silveira4, D. De Jesus Damiao, C. De Oliveira Martins, S.
Fonseca De Souza, H. Malbouisson, D. Matos Figueiredo, M. Melo De Almeida, C.
Mora Herrera, L. Mundim, H. Nogima, W.L. Prado Da Silva, L.J. Sanchez Rosas,
A. Santoro, A. Sznajder, M. Thiel, E.J. Tonelli Manganote3, F. Torres Da Silva
De Araujo, A. Vilela Pereira Universidade Estadual Paulista a, Universidade
Federal do ABC b, São Paulo, Brazil
S. Ahujaa, C.A. Bernardesa, L. Calligarisa, T.R. Fernandez Perez Tomeia, E.M.
Gregoresb, P.G. Mercadanteb, S.F. Novaesa, SandraS. Padulaa, D. Romero Abadb
Institute for Nuclear Research and Nuclear Energy, Bulgarian Academy of
Sciences, Sofia, Bulgaria
A. Aleksandrov, R. Hadjiiska, P. Iaydjiev, A. Marinov, M. Misheva, M. Rodozov,
M. Shopova, G. Sultanov University of Sofia, Sofia, Bulgaria
A. Dimitrov, L. Litov, B. Pavlov, P. Petkov Beihang University, Beijing, China
W. Fang5, X. Gao5, L. Yuan Institute of High Energy Physics, Beijing, China
M. Ahmad, J.G. Bian, G.M. Chen, H.S. Chen, M. Chen, Y. Chen, C.H. Jiang, D.
Leggat, H. Liao, Z. Liu, F. Romeo, S.M. Shaheen6, A. Spiezia, J. Tao, C. Wang,
Z. Wang, E. Yazgan, H. Zhang, J. Zhao State Key Laboratory of Nuclear Physics
and Technology, Peking University, Beijing, China
Y. Ban, G. Chen, A. Levin, J. Li, L. Li, Q. Li, Y. Mao, S.J. Qian, D. Wang, Z.
Xu Tsinghua University, Beijing, China
Y. Wang Universidad de Los Andes, Bogota, Colombia
C. Avila, A. Cabrera, C.A. Carrillo Montoya, L.F. Chaparro Sierra, C. Florez,
C.F. González Hernández, M.A. Segura Delgado University of Split, Faculty of
Electrical Engineering, Mechanical Engineering and Naval Architecture, Split,
Croatia
B. Courbon, N. Godinovic, D. Lelas, I. Puljak, T. Sculac University of Split,
Faculty of Science, Split, Croatia
Z. Antunovic, M. Kovac Institute Rudjer Boskovic, Zagreb, Croatia
V. Brigljevic, D. Ferencek, K. Kadija, B. Mesic, A. Starodumov7, T. Susa
University of Cyprus, Nicosia, Cyprus
M.W. Ather, A. Attikis, M. Kolosova, G. Mavromanolakis, J. Mousa, C. Nicolaou,
F. Ptochos, P.A. Razis, H. Rykaczewski Charles University, Prague, Czech
Republic
M. Finger8, M. Finger Jr.8 Escuela Politecnica Nacional, Quito, Ecuador
E. Ayala Universidad San Francisco de Quito, Quito, Ecuador
E. Carrera Jarrin Academy of Scientific Research and Technology of the Arab
Republic of Egypt, Egyptian Network of High Energy Physics, Cairo, Egypt
Y. Assran9,10, S. Elgammal10, S. Khalil11 National Institute of Chemical
Physics and Biophysics, Tallinn, Estonia
S. Bhowmik, A. Carvalho Antunes De Oliveira, R.K. Dewanjee, K. Ehataht, M.
Kadastik, M. Raidal, C. Veelken Department of Physics, University of Helsinki,
Helsinki, Finland
P. Eerola, H. Kirschenmann, J. Pekkanen, M. Voutilainen Helsinki Institute of
Physics, Helsinki, Finland
J. Havukainen, J.K. Heikkilä, T. Järvinen, V. Karimäki, R. Kinnunen, T.
Lampén, K. Lassila-Perini, S. Laurila, S. Lehti, T. Lindén, P. Luukka, T.
Mäenpää, H. Siikonen, E. Tuominen, J. Tuominiemi Lappeenranta University of
Technology, Lappeenranta, Finland
T. Tuuva IRFU, CEA, Université Paris-Saclay, Gif-sur-Yvette, France
M. Besancon, F. Couderc, M. Dejardin, D. Denegri, J.L. Faure, F. Ferri, S.
Ganjour, A. Givernaud, P. Gras, G. Hamel de Monchenault, P. Jarry, C. Leloup,
E. Locci, J. Malcles, G. Negro, J. Rander, A. Rosowsky, M.Ö. Sahin, M. Titov
Laboratoire Leprince-Ringuet, Ecole polytechnique, CNRS/IN2P3, Université
Paris-Saclay, Palaiseau, France
A. Abdulsalam12, C. Amendola, I. Antropov, F. Beaudette, P. Busson, C.
Charlot, R. Granier de Cassagnac, I. Kucher, S. Lisniak, A. Lobanov, J. Martin
Blanco, M. Nguyen, C. Ochando, G. Ortona, P. Paganini, P. Pigard, R. Salerno,
J.B. Sauvan, Y. Sirois, A.G. Stahl Leiton, A. Zabi, A. Zghiche Université de
Strasbourg, CNRS, IPHC UMR 7178, Strasbourg, France
J.-L. Agram13, J. Andrea, D. Bloch, J.-M. Brom, E.C. Chabert, V. Cherepanov,
C. Collard, E. Conte13, J.-C. Fontaine13, D. Gelé, U. Goerlach, M. Jansová,
A.-C. Le Bihan, N. Tonon, P. Van Hove Centre de Calcul de l’Institut National
de Physique Nucleaire et de Physique des Particules, CNRS/IN2P3, Villeurbanne,
France
S. Gadrat Université de Lyon, Université Claude Bernard Lyon 1, CNRS-IN2P3,
Institut de Physique Nucléaire de Lyon, Villeurbanne, France
S. Beauceron, C. Bernet, G. Boudoul, N. Chanon, R. Chierici, D. Contardo, P.
Depasse, H. El Mamouni, J. Fay, L. Finco, S. Gascon, M. Gouzevitch, G.
Grenier, B. Ille, F. Lagarde, I.B. Laktineh, H. Lattaud, M. Lethuillier, L.
Mirabito, A.L. Pequegnot, S. Perries, A. Popov14, V. Sordini, M. Vander
Donckt, S. Viret, S. Zhang Georgian Technical University, Tbilisi, Georgia
|
# Molecular machines for quantum error correction
Thiago Guerreiro<EMAIL_ADDRESS>Department of Physics, Pontifícia
Universidade Católica, Rio de Janeiro, Brazil
###### Abstract
Inspired by biological molecular machines we explore the idea of an active
quantum robot whose purpose is delaying decoherence. A conceptual model
capable of partially protecting arbitrary logical qubit states against single
physical qubit errors is presented. Implementation of an instance of that
model - the entanglement qubot - is proposed using laser-dressed Rydberg
atoms. Dynamics of the system is studied using stochastic wavefunction
methods.
## I Introduction
The living cell can be seen as a Brownian computer [1]. At its core, machines
of molecular dimensions store, correct and process information in the presence
of noise, with the goal of keeping the state of the living creature away from
thermodynamical equilibrium. The machinery of life [2] is responsible for gene
expression, matter transport across the cell and energy harvesting, among a
vast number of other tasks [3]. An example of such molecular devices is RNA
polymerase (RNAP): an enzyme with $\sim 40.000$ atoms, roughly $\SI{10}{nm}$
of linear size, capable of synthesising a strand of RNA from a DNA template in
the presence of Brownian noise, at error rates as low as $10^{-7}$ [4].
Molecular devices such as RNAP have inspired nanotechnology [5, 6] and various
artificial molecular machines were built, such as molecular ratchets [7],
pumps [8], motors [9], and gene editing tools [10].
Detailed unified understanding of biological molecular machines according to
the tradition of theoretical physics is yet to be achieved [11], but there is
little doubt that experimental [12] and computational methods [13] in physics
play a key role in that endeavour. It is also expected that the coming age of
quantum information processing will illuminate biological systems through
simulation of quantum chemistry [14] and quantum enhanced learning [15, 16].
Conversely [17], one could ask whether biological molecular machines will
inspire new ideas for engineering autonomous molecular-sized quantum
information processing devices with the goal of keeping quantum states away
from thermodynamical equilibrium. It is the purpose of this work to explore
this idea.
Figure 1: (a) Schematics for a conceptual qubot model capable of partially
protecting an arbitrary logical qubit state against decoherence. (b) Example
of a possible potential landscape describing the interaction between the
nucleus atoms; for this plot the radial dependence of (2) is considered
$J_{\alpha}(R)=(d^{2}/R^{3})j_{\alpha}$, with $j_{y}=-3j_{x},j_{z}=6j_{x}$.
A quantum molecular machine would be a device composed of at most a few
thousand atoms capable of autonomously storing, protecting and/or processing
quantum states in the presence of external decoherence and thermalization. We
refer to these bio-inspired devices as quantum robots, or qubots [18].
Devising qubots is a problem in coherent quantum chemistry [19, 20] much like
engineering artificial molecular machines is a problem in synthetic chemistry
[21]. Hence, the ultracold atom [22] and molecular toolbox [23, 24] is
expected to play a key role in the conception of these active quantum devices.
As we will see, qubots exploit open system dynamics to achieve their purpose
and thus have a close connection to the idea of engineered environments
constructed to produce desired quantum states [25, 26, 27, 28, 29, 30, 31,
32]. Their nature, however, is much closer to that of artificial molecular
ratchets and pumps that respond to the environment and consume resources to
maintain nonequilibrium states [33].
In what follows, we explore various aspects around the idea of qubots. We
begin by introducing a conceptual model for a quantum robot capable of
partially protecting a logical qubit state against single physical qubit
errors. It is interesting that the model can handle almost all combinations of
phase and bit-flip errors since, as pointed out by Kitaev, it is generally
easy to get rid of one kind of errors, but not both [34]. The construction is
somewhat inspired by the surface code [35], only here syndrome detection and
correction are part of the system’s dynamics rather than a consequence of
measurement followed by external conditional action. Next, a specific physical
implementation of instances of the model based on laser-dressed Rydberg atoms
is discussed. More specifically, we exhibit potential landscapes implementing
an entanglement qubot, a device that stabilizes a Bell state against single
qubit errors. The stabilized Bell state is only one possible state of the
logical qubit, but in this case we can view the qubot as preserving a
maximally entangled state. An ensemble of entanglement qubots could therefore
preserve vast amounts of entanglement, a useful resource. Simulation of the
entanglement qubot dynamics is performed with the help of stochastic
wavefunction methods, and we evaluate the effects of coupling the motional
degrees of freedom of the robot to an external heat bath. We conclude with a
discussion on potential future developments regarding active quantum matter.
## II Conceptual model
We would like to introduce the conceptual model of a quantum robot capable of
protecting an arbitrary logical qubit state against errors. Our quantum robot
consists of two parts, called the nucleus and the correctors. See Figure 1(a)
for a schematic representation. A pair of particles denoted $a$ and $b$
constitute the nucleus. Quantum information is stored in the particles’
internal spin degrees of freedom taken to be two spin 1/2 systems with Hilbert
space $\mathbb{C}^{2}\otimes\mathbb{C}^{2}$ and basis states denoted
$\\{|0\rangle|0\rangle,|0\rangle|1\rangle,|1\rangle|0\rangle,|1\rangle|1\rangle\\}$.
Particle $a$ is held fixed at the origin by an optical tweezer while $b$ is
subject to the potential
$\displaystyle V(R)=V_{t}(R)+V_{I}(R)\ ,$ (1)
where $R$ is the relative distance between $a$ and $b$, $V_{t}(R)$ is a trap
potential for particle $b$ and
$\displaystyle V_{I}(R)=J_{z}Z_{a}Z_{b}+J_{x}X_{a}X_{b}+J_{y}Y_{a}Y_{b}\ ,$
(2)
is the interaction energy between the particles, where
$X_{\lambda},Y_{\lambda},Z_{\lambda}$ are the Pauli operators for particle
$\lambda$ ($=a,b$) and the coefficients $J_{\alpha}=J_{\alpha}(R)$ form a
spatial-dependent spin-spin interaction pattern. We assume for simplicity that
particle $b$ can only move along the direction $\hat{R}$.
As an example of trap potential one may consider an optical tweezer,
$\displaystyle V_{t}(R)$ $\displaystyle=$ $\displaystyle
V_{0}\left(R-\delta\right)^{2}\ .$ (3)
where $V_{0}$ and $\delta$ are constants. Tunneling outside the confining
potential is considered negligible. Note also that dipole-dipole interactions
among atoms and polar molecules is of the form (2), and typically for
molecules [36, 37] and spin impurities in diamond [38],
$\displaystyle J_{\alpha}=(d^{2}/R^{3})j_{\alpha}\ ,$ (4)
where $d$ is the dipole moment [39] and $j_{\alpha}$ a proportionality
constant with $\alpha=x,y,z$. Through the remaining of this section we will
consider this radial dependence as an illustration of the qubot functioning.
Note however that effective spin interactions of the so-called $XYZ$ form with
more general radial dependencies can be engineered within a number of
different systems, including trapped ions [40, 41], atoms in dressed Rydberg
states [42, 43] and microwave-excited polar molecules in optical lattices [45,
46]. In the next section an implementation using laser dressed Rydberg atoms
will be discussed.
Bell states of the particles’ spins are eigenstates of $V_{I}$ with
eigenvalues given by
$\displaystyle V_{I}|\psi^{-}\rangle$ $\displaystyle=$
$\displaystyle\left(-J_{x}-J_{y}-J_{z}\right)|\psi^{-}\rangle\ ,$ (5)
$\displaystyle V_{I}|\phi^{-}\rangle$ $\displaystyle=$
$\displaystyle\left(-J_{x}+J_{y}+J_{z}\right)|\phi^{-}\rangle\ ,$ (6)
$\displaystyle V_{I}|\psi^{+}\rangle$ $\displaystyle=$
$\displaystyle\left(J_{x}+J_{y}-J_{z}\right)|\psi^{+}\rangle\ ,$ (7)
$\displaystyle V_{I}|\phi^{+}\rangle$ $\displaystyle=$
$\displaystyle\left(J_{x}-J_{y}+J_{z}\right)|\phi^{+}\rangle\ .$ (8)
This implies that the total potential $V(R)$ exhibits collective spin-
dependent landscapes.
As an example consider the trap potential (3) and the spin pattern (4). If
local equilibrium positions $R_{0}(|\psi\rangle)$ exist, they satisfy the
condition
$\displaystyle
R_{0}^{4}(R_{0}-\delta)=\dfrac{3d^{2}\langle\psi|W|\psi\rangle}{2V_{0}}\ ,$
(9)
where
$\langle\psi|W|\psi\rangle=\langle\psi|\left(j_{z}Z_{a}Z_{b}+j_{x}X_{a}X_{b}+j_{y}Y_{a}Y_{b}\right)|\psi\rangle$
are possible expectation values with respect to each of the four Bell states.
Figure 1(b) shows the total potential landscape seen by particle $b$ for each
of the spin Bell states, displaying the spin-dependent potentials. Note that
the state $|\psi^{+}\rangle$ does not exhibit a minimum; this is not a problem
provided the protected logical qubit states do not involve $|\psi^{+}\rangle$.
In between equilibrium points of the potential landscapes in Figure 1(b) there
are corrective sites, where devices we call correctors are present. Correctors
are represented in Figure 1(a) as loops. The function of the corrective
devices is executing a unitary operation on the spin subspace once the
particle approaches their site. There are two correctors, denoted $L1$ and
$L2$. For illustration of the device functioning, in the remaining of this
section we treat the correctors $L1$ and $L2$ as qubits. Note however that
there are a number of ways of implementing such devices and alternatives to
the qubit model will be discussed in the following implementation section.
Consider the $L1$ device has basis states
$\\{|\mu_{0}^{1}\rangle,|\mu_{1}^{1}\rangle\\}$. Whenever the particle enters
one of the $L1$ loops, the unitary operation $Z_{b}X_{L1}$ is executed, where
$X_{L1}=|\mu_{0}^{1}\rangle\langle\mu_{1}^{1}|+|\mu_{1}^{1}\rangle\langle\mu_{0}^{1}|$.
It is important that $L1$ is insensitive to whether particle $b$ entered the
innermost or outermost loop, since obtaining that information would collapse
the spin state of the system as it is correlated to motion. The $L2$ system,
or middle corrector, has basis states
$\\{|\mu_{0}^{2}\rangle,|\mu_{1}^{2}\rangle\\}$ and whenever particle $b$
enters $L2$, the unitary $X_{b}X_{L2}$ is executed, where $X_{L2}$ is once
again the bit-flip operator on the corresponding basis states of $L2$.
We have the following operations:
$\displaystyle L1:\ Z_{b}X_{L1}\ ,\ \ L2:\ X_{b}X_{L2}\ .$ (10)
Note these unitaries act on the spins conditional on the particle’s position.
Hence, when tracing out the position degree of freedom, action of the
corrective sites manifests as dissipative maps on the spin subspace.
Logical basis states of the nucleus are defined as
$\displaystyle|\bar{0}\rangle$ $\displaystyle=$
$\displaystyle|\psi^{-}\rangle$ (11) $\displaystyle|\bar{1}\rangle$
$\displaystyle=$ $\displaystyle|\phi^{-}\rangle$ (12)
and an arbitrary logical qubit state is
$\displaystyle|\Psi\rangle=\alpha|\bar{0}\rangle+\beta|\bar{1}\rangle$ (13)
Note that a superposition of the $|\bar{0}\rangle,|\bar{1}\rangle$ states
implies particle $b$ is in a superposition of singlet and triplet spin states,
implying a superposition of different spatial equilibrium points.
To understand how the qubot delays decoherence and partially protects the
logical qubit, one must follow carefully what happens to the particles when a
physical error occurs in one of the spins. Single physical qubit errors are
assumed to be much more likely than multi-qubit errors [35] and the
depolarizing channel is considered as decoherence model. A summary of possible
errors and how they act on logical basis states is shown in Table 1.
Error | $\ \ |\psi^{-}\rangle$ | $\ \ |\phi^{-}\rangle$ | Corrected state
---|---|---|---
$X_{a}$ | $-|\phi^{-}\rangle$ | $-|\psi^{-}\rangle$ | $-\alpha|\bar{0}\rangle-\beta|\bar{1}\rangle$
$X_{b}$ | $+|\phi^{-}\rangle$ | $+|\psi^{-}\rangle$ | $\alpha|\bar{0}\rangle+\beta|\bar{1}\rangle$
$Z_{a}$ | $+|\psi^{+}\rangle$ | $+|\phi^{+}\rangle$ | $\alpha|\bar{0}\rangle-\beta|\bar{1}\rangle$
$Z_{b}$ | $-|\psi^{+}\rangle$ | $+|\phi^{+}\rangle$ | $-\alpha|\bar{0}\rangle-\beta|\bar{1}\rangle$
$Z_{a}X_{a}$ | $-|\phi^{+}\rangle$ | $-|\psi^{+}\rangle$ | $\alpha|\bar{0}\rangle-\beta|\bar{1}\rangle$
$Z_{b}X_{b}$ | $+|\phi^{+}\rangle$ | $-|\psi^{+}\rangle$ | $-\alpha|\bar{0}\rangle-\beta|\bar{1}\rangle$
Table 1: Effect of physical errors on logical basis states and the final
corrected state after action of the qubot. Figure 2: Example of a qubot
cycle: (a) particle $b$ rests in its equilibrium position, while the spin
states form a singlet $|\bar{0}\rangle=|\psi^{-}\rangle$; (b) an error occurs,
changing the potential landscape seen by $b$; (c) the particle is forced into
loop $L2$, which restores the original spin state and (d) the particle goes
back to the original equilibrium position.
As an illustration, consider the example of a bit-flip in the first spin
described by the $X_{a}$ operator. Initially, an arbitrary logical qubit state
$|\Psi\rangle=\alpha|\psi^{-}\rangle+\beta|\phi^{-}\rangle$ is in a
superposition of equilibrium positions $R_{0}(|\psi^{-}\rangle)$ and
$R_{0}(|\phi^{-}\rangle)$ given by solutions of (9). The $X_{a}$ error changes
the spin state of the particles according to
$\displaystyle\alpha|\psi^{-}\rangle+\beta|\phi^{-}\rangle\rightarrow-\alpha|\phi^{-}\rangle-\beta|\psi^{-}\rangle$
(14)
and hence the particles’ interaction potential is changed accordingly. After
the error, the possible positions of particle $b$ are no longer equilibrium
points of the potential landscapes. For the case in which $b$ was initially at
$R_{0}(|\psi^{-}\rangle)$, the particles repel, forcing $b$ into $L2$.
Similarly, for $R_{0}(|\phi^{-}\rangle)$, occurrence of the error causes an
attractive force which pulls $b$ into $L2$. Once $b$ reaches the loop, the
operator $X_{b}X_{L2}$ is applied, restoring the logical qubit to the original
state and driving the system back to the initial superposition of equilibrium
points. Naturally this process introduces kinetic energy in the form of
phonons, which must be removed if particle $b$ is to settle back in the
original state. This implies the need for a dissipative force acting on $b$
which could be provided by state-independent cooling of the atom motion. For
now, we will assume that such cooling is present, and this phonon issue will
be discussed further in the implementation section.
Similar processes occur for $X_{b}$ and $Z_{b}$ errors: a combination of spin-
motion dynamics and subsequent application of the loop operators corrects
errors and restores the system to the initial arbitrary logical state. The
qubot is also able to correct a concatenation of phase and bit-flip errors,
given by $Y_{b}$. Note that this requires a passage through two correctors.
The present qubot model is not able to correct all errors. As can be seen in
Table 1, logical basis states transform under $Z_{a}$ with opposite parity,
thus inducing a phase error in the logical qubit. This imparts on the $Y_{a}$
error since $iY_{a}=Z_{a}X_{a}$. This imperfection can be traced back to the
fact that the qubot uses two physical qubits to encode a logical state. The
quantum Hamming bound [47] implies that for single qubit errors, a minimum of
five qubits are required to achieve complete fault tolerance for one logical
qubit. Despite this partial fault tolerance the qubot can delay decoherence of
arbitrary logical qubit states, and for some specific states it is even able
to preserve it regardless of the error, as for example the singlet
$|\psi\rangle=|\psi^{-}\rangle$. More general models implementing perfect
quantum error correcting codes [48] can nevertheless be devised at the expense
of more particles or higher spin states. Note that to protect arbitrary
logical qubit states, the qubot potential landscapes must distinguish between
all the four elements of the Bell basis, as in Figure 1(b). If the landscape
for two or more Bell states is indistinguishable, certain errors will cause no
effect upon the atom preventing the action of the correctors. Note also that
the order of the potential minima for each Bell state defines the choice of
position and action for the corrective sites, as well as the choice of logical
basis states.
It is instructive to consider the qubot operation under a depolarizing channel
acting on particle $b$ alone. Denote environment states as $|e_{j}\rangle$.
Decoherence causes the joint particle-environment-corrector state to evolve
according to,
$\displaystyle|\Psi\rangle|e_{0}\rangle|\mu_{0}^{1}\mu_{0}^{2}\rangle\rightarrow\sqrt{1-p}\left(\alpha|\psi^{-}\rangle+\beta|\phi^{-}\rangle\right)|e_{0}\rangle|\mu_{0}^{1}\mu_{0}^{2}\rangle$
$\displaystyle+\sqrt{\dfrac{p}{3}}\left(\alpha|\phi^{-}\rangle+\beta|\psi^{-}\rangle\right)|e_{1}\rangle|\mu_{0}^{1}\mu_{0}^{2}\rangle$
$\displaystyle+\sqrt{\dfrac{p}{3}}\left(-\alpha|\psi^{+}\rangle+\beta|\phi^{+}\rangle\right)|e_{2}\rangle|\mu_{0}^{1}\mu_{0}^{2}\rangle$
$\displaystyle+\sqrt{\dfrac{p}{3}}\left(\alpha|\phi^{+}\rangle-\beta|\psi^{+}\rangle\right)|e_{3}\rangle|\mu_{0}^{1}\mu_{0}^{2}\rangle$
(15)
where $p$ denotes the error probability. Equation (15) describes the
depolarizing dynamics suffered by the logical qubit, with the first term
proportional to $\sqrt{1-p}$ corresponding to no decoherence and the
subsequent terms proportional to $\sqrt{p/3}$ corresponding to errors on the
logical qubit. Note that at this stage, the corrective devices remain
unaffected while the system undergoes errors and the environment learns when
an error has occurred. Tracing out the environment, the above evolution
induces a dissipative map on the spin system increasing its entropy and
causing decoherence of the original state.
With the occurrence of errors the potential landscapes acting on $b$ undergo a
change forcing the action of the correctors upon the spin state of the
nucleus. Purity of the logical qubit is restored at the expense of an increase
in entropy for the correctors; after a correction event, (15) evolves to
$\displaystyle|\Psi\rangle\left(\sqrt{1-p}|e_{0}\rangle|\mu_{0}^{1}\mu_{0}^{2}\rangle+\sqrt{\dfrac{p}{3}}|e_{1}\rangle|\mu_{0}^{1}\mu_{1}^{2}\rangle\right.$
$\displaystyle\left.-\sqrt{\dfrac{p}{3}}|e_{2}\rangle|\mu_{1}^{1}\mu_{0}^{2}\rangle-\sqrt{\dfrac{p}{3}}|e_{3}\rangle|\mu_{1}^{1}\mu_{1}^{2}\rangle\right)$
(16)
where we can see that the original logical qubit state is restored and the
environment gets correlated to the correctors’ state. After a single error
correction, the correctors’ states must be reset to the pure initial state
$|\mu_{0}^{1}\mu_{0}^{2}\rangle$. This is a non-unitary operation which
requires energy expenditure, similar to erasing a quantum state [49, 50] and
can be implemented as a non-equilibrium stochastic process. This corresponds
to a consumption of resources by the qubot analogous to the consumption of
resources by biological molecular machines and living organisms. Irrespective
of the physical implementation of the corrective sites, such consumption of
resources is a mandatory part of the qubot operation in accordance to the laws
of thermodynamics.
## III Implementation
Potential engineering. Spin-spin interactions of the form (2) suitable for
implementing quantum robots could be engineered in a number of different
atomic and molecular systems. In this section a physical implementation using
laser-dressed Rydberg atoms [42, 43, 44] is discussed. As will be shown,
instances of the qubot described in the previous section can be realized for
realistic experimental parameters, provided one chooses the correct logical
basis elements and position of corrective sites. We will focus on a qubot that
stabilizes an effective entangled spin state against a depolarizing
environment similar to the one outlined in [18]. We shall refer to this device
as an entanglement qubot.
Figure 3: Level schematics for the entanglement qubot. Figure 4: (a) Spin
pattern, corresponding to the coefficients of Eq. (20) for the parameters
$n=60$, $\Delta_{-}=-\Delta_{+}=2\pi\times\SI{50}{MHz}$ and
$\Omega_{-}=\Omega_{+}/3=\SI{2\pi\times 3}{MHz}$. (b) Collective spin-
dependent potential landscapes. Each trace corresponds to a Bell state of the
qubot nucleus. Corrective devices $L1$ and $L2$ are positioned outside the
potential minima, for example as the dashed vertical lines indicate.
A pair of 87Rb atoms labelled $a$ and $b$ constitute the qubot nucleus.
Effective spin states are provided by hyperfine levels of $b$, specifically
$\displaystyle|0\rangle$ $\displaystyle=$
$\displaystyle|5^{2}S_{1/2},F=1,m_{F}=1\rangle\ ,$ (17)
$\displaystyle|1\rangle$ $\displaystyle=$
$\displaystyle|5^{2}S_{1/2},F=2,m_{F}=2\rangle\ ,$ (18)
with energy difference $\omega_{01}$. The atom-atom interaction potential is
induced by dressing the $|0\rangle,|1\rangle$ states with two strongly
interacting Rydberg Zeeman sublevels in the $n^{2}P_{1/2}$ manifold via Rabi
oscillations with detunings $\Delta_{\pm}$ and frequencies $\Omega_{\pm}$
using $\sigma^{\pm}$ polarized light. The interaction between Rydberg states
arises from a van der Waals potential of the form $C_{6}R^{-6}$, and a fixed
orientation of the two particles is considered, with the atoms polarized
perpendicular to the plane. Large detunings guarantee that only a small
fraction of the Rydberg states is admixed to the $|0\rangle,|1\rangle$ levels
while maintaining a long lifetime. Following [42], the Rydberg states are
$\displaystyle|r_{\pm}\rangle=|n^{2}P_{1/2},m_{j}=\pm
1/2\rangle|m_{I}=3/2\rangle\ ,$ (19)
with an energy difference $\Delta E_{r}$. Detunings are chosen such that the
energy conservation condition $\Delta E_{r}=(\Delta_{+}-\Delta_{-})$ is
satisfied. A level diagram is shown in Figure 3. The atoms are trapped in one
dimensional potentials, insensitive to their internal states. State-
independent trapping of Rydberg dressed atoms can be achieved in so-called
magic [51, 52] and magnetic traps [53]. While atom $a$ is fixed at the origin,
$b$ is able to move under the influence of a force resulting from the
combination of an external tweezer and the atom-atom interaction potential.
As in quantum chemistry [19, 20], the time scale associated to electronic
dynamics is much shorter than the time scale of nuclei motion. An effective
spin dependent Born-Oppenheimer potential can therefore be derived at fixed
atomic separations $R$. In the limit of large detunings
$\Omega_{\pm}\ll\Delta_{\pm}$ and for
$\Delta_{+}/\Delta_{-}<0,\Delta_{+}+\Delta_{-}<0$, adiabatic elimination [66]
can be used in the rotating frame to obtain an effective interaction acting on
the subspace generated by the $|0\rangle,|1\rangle$ states to fourth order in
$\Omega_{\pm}/\Delta_{\pm}$,
$\displaystyle
V_{I}(R)=J_{z}Z_{a}Z_{b}+J_{x}X_{a}X_{b}+J_{y}Y_{a}Y_{b}+J_{\parallel}\left(Z_{a}+Z_{b}\right)\
,$ (20)
where $J_{\alpha}(R)$ ($\alpha=x,y,z$) are radial steplike coefficients
depending on the Rabi frequencies $\Omega_{\pm}$, detunings $\Delta_{\pm}$ and
van der Waals $C_{6}$ coefficients for the $n^{2}P_{1/2}$ manifold.
$J_{\parallel}$ is an effective magnetic field, which we assume can be
cancelled by an additional weak non-homogeneous field on the order of 2G. See
Appendices A and B for explicit definitions, formulas and details on the
potential and effective magnetic field, respectively.
A plot of the $J_{\alpha}$ spin pattern for $n=60$, detunings
$\Delta_{-}=-\Delta_{+}=2\pi\times\SI{50}{MHz}$ and Rabi frequencies
$\Omega_{-}=\Omega_{+}/3=\SI{2\pi\times 3}{MHz}$ can be seen in Figure 4(a).
Note these are in the same parameter region as used for realizing the quantum
spin ice Hamiltonian on a kagome lattice in [42, 54]. The parameters defining
a qubot potential are not unique, allowing some freedom in the construction;
for an example of a different set of numbers and the resulting spin pattern
see the Appendix C.
From the spin pattern coefficients together with Eqs.(5)-(8) and a trap
potential $V_{t}(R)$ we can derive the collective spin-dependent potentials
acting on particle $b$. Consider a trap potential provided by two neighboring
optical tweezers,
$\displaystyle
V_{t}(R)=V_{0}\left[\left(R-\delta_{1}\right)^{2}+\left(R-\delta_{2}\right)^{2}\right]$
(21)
where $V_{0}=\SI{15}{kHz/\mu m^{2}}$, $\delta_{1}=\SI{1.6}{\mu m}$ and
$\delta_{2}=\SI{2.0}{\mu m}$. The resulting spin-dependent potential
landscapes $V(R)$ can be seen in Figure 4(b), where each trace corresponds to
a different Bell state of the two atoms. Note equilibrium positions are
separated by approximately $\SI{0.3}{\mu m}$. Trap frequencies are
approximately $\omega_{t}/2\pi\approx\SI{1}{kHz}$. Possible positions for the
corrective sites $L1$ and $L2$, corresponding to the transformations (10), are
represented by dashed vertical lines. Note the potential landscapes for the
Bell states $|\psi^{-}\rangle$ and $|\psi^{+}\rangle$ overlap. This implies
that one cannot choose either $|\psi^{-}\rangle$ or $|\psi^{+}\rangle$ as
protected states, as in this case, phase errors could not be corrected. The
protected logical state is chosen to be $|\phi^{+}\rangle$.
Corretive sites. Correctors $L1$ and $L2$ were previously considered to be
qubits acting as an entropy sink for maintaining the purity of the protected
logical qubit state carried by the nucleus. The interaction between
superconducting quantum electronics and atomic [55], molecular [56] and
mesoscopic particles [57] has been extensively studied in the context of
hybrid quantum systems and the coupling between NV centers and superconductors
has been observed [58]. A number of different implementations involving
superconducting qubit systems is therefore expected.
Beyond qubits, one may consider additional atoms as candidates for
implementing corrective devices. Controlled atomic collisions [59] would
provide the mechanism for position-dependent unitary operations. One could
envision a lattice with arrays of data particles interpolated with corrective
particles, analogous to the surface code [35]; occurrence of errors would
alter the interaction between data particles, enabling or inhibiting motion
and tunneling - and consequently interactions - with neighboring corrective
sites. It would be as a surface code in motion, where errors induce controlled
motion leading to correction feedbacks. It is important to stress that in the
course of the qubot action, entropy of the corrective atoms would increase and
a dissipative map for restarting the correctors in their original state would
have to be continuously enforced, for example through an amplitude damping
channel [18].
Corrective devices could also be implemented using Rabi oscillations between
the $|0\rangle,|1\rangle$ levels. By carefully tuning the Rabi frequency of
the transition and the profile of the spin-dependent potentials in Figure 4(b)
it is in principle possible to engineer the transit time of atom $b$ through
$L1$ and $L2$ such that $Z_{b}$ and $X_{b}$ operations are applied, analogous
to the transit time stimulated decay in ammonia masers [60] and Ramsey
interferometry in atomic fountain clocks [61]. In this implementation -
probably the most practical from an experimental point-of-view - the
electromagnetic field assumes the role of entropy sink since conditional $X$
and $Z$ operations on the atom would introduce uncertainties in the intensity
and phase of the field, respectively. A schematics of this implementation is
shown in Figure 5.
Figure 5: Corrective sites as Rabi oscillations.
Operation, cooling and lifetime. Operation of the qubot proceeds as described
in the previous section: occurrence of an error induces a change in the
potential landscape seen by atom $b$ thus forcing it into one of the
corrective sites $L1$ or $L2$. Note that errors can occur due to external
environmental influence or intrinsically due to thermal and quantum
fluctuations of the atomic motion. Consider atom $b$ in a thermal state. For
temperatures on the order of 10nK, reachable for atomic ensembles [62], the
occupation number of atomic motion is $\bar{n}\approx 0.1$ pointing out that
the atom is effectively in the trap ground state. Zero point motion of the
atom is approximately
$R_{\mathrm{zpm}}\simeq\sqrt{\hbar/2m\omega_{t}}\approx\SI{0.23}{\mu m}$,
indicating that at 10nK quantum fluctuations can cause the atom to reach the
corrective sites even when no environmental error took place, inducing change
in the qubot state. Hence, intrinsic fluctuation errors are expected to
constitute a portion of total errors. In the next section, a model of the
qubot operation taking into account intrinsic and external errors will be
discussed.
Errors can be effectively corrected provided the qubot nucleus undergoes
constant cooling of its motional degrees of freedom to dissipate the kinetic
energy gained by mechanical forces due to potential changes. Such cooling
mechanism needs to preserve the quantum information stored in the nucleus, so
it must be insensitive to the quantum state stored in the spins. State-
insensitive cooling of neutral atoms can be achieved via superfluid immersion
[63], cavity cooling [64] or sympathetic cooling through spin-independent
Rydberg interactions with neighboring atoms [65].
What is the order of magnitude of the expected lifetime for the protected
entangled state? The $60P_{1/2}$ Rydberg state has a lifetime on the order of
$\tau_{r}\approx\SI{133}{\mu s}$ [70]. This implies a bare lifetime for the
effective spin state of
$\tau_{s}\approx(2\Delta_{-}/\Omega_{-})^{2}\tau_{r}\approx\SI{9}{ms}$ [42],
corresponding to a spin decoherence rate $\Gamma\approx\SI{111}{Hz}$. A decay
process to the ground state $|0\rangle$ is defined by the following
transformations,
$\displaystyle|0\rangle|e_{0}\rangle$ $\displaystyle\rightarrow$
$\displaystyle|0\rangle|e_{0}\rangle$ (22)
$\displaystyle|1\rangle|e_{0}\rangle$ $\displaystyle\rightarrow$
$\displaystyle\sqrt{1-\tau_{s}^{-1}dt}|1\rangle|e_{0}\rangle+\sqrt{\tau_{s}^{-1}dt}|0\rangle|e_{1}\rangle$
(23)
where the first ket corresponds to the spin of the particle while the second
ket represents the environment state. Action of this quantum channel upon the
elements of the Bell basis can be written in terms of strings of Pauli errors
[71]. It is thus expected that the qubot is able to extend the lifetime of
Rydberg dressed entangled states.
## IV Dynamics Simulation
Exploration of the qubot requires simulation of its error-correction dynamics.
Any such simulation must take into account the effects of quantum fluctuations
of atomic motion, as these fluctuations are in themselves a source of errors
that can disturb the protected Bell state. A first principles description of
the spin and motion degrees of freedom is intricate as the spin state is
subject to transformations conditional on the motion state, which in itself is
conditioned on the spin through the spin-dependent potential. As Wheeler would
say [72]: spin tells matter how to move, matter tells spin how to turn.
To capture the essential features of the qubot we propose an open quantum
system model in which the motion and spin degrees of freedom follow a set of
discrete-time coupled stochastic Schrodinger equations. Each realization of
the evolution is described in terms of sequences of quantum state pairs,
denoted $|\psi\rangle$ for the spin and $|\phi\rangle$ for the motion degree
of freedom. Averaging over many realizations of the stochastic process results
in the mean behavior of the system.
The spin and motion degrees of freedom act as environments for each other.
This idea can be used to motivate the model as follows. For simplicity,
discretize (1D) space into a set of points $R_{k}$. The position state reads
$\displaystyle|\phi\rangle=\sum_{k}\phi(R_{k})|R_{k}\rangle$ (24)
where $|\phi(R_{k})|^{2}$ gives the probability of finding the particle at
position $R_{k}$. The initial state evolves in a small time increment $\delta
t$ according to
$\displaystyle|\psi\rangle|\phi\rangle\xrightarrow{\delta
t}\sum_{i}\phi(R_{i})(T(R_{i})|\psi\rangle)(W(|\psi\rangle)|R_{i}\rangle)$
$\displaystyle=|\Psi(t+\delta t)\rangle$ (25)
where $T(R_{i})$ is the identity operator unless $R_{i}=R_{L1}$ or
$R_{i}=R_{L2}$, for which
$\displaystyle T(R_{L1})=Z_{b}$ (26) $\displaystyle T(R_{L2})=X_{b}$ (27)
The operator $W(|\psi\rangle)$ contains information on the spin-dependent
potential and is responsible for the evolution of the motion state. Expanding
$|\Psi(t+dt)\rangle$,
$\displaystyle|\Psi(t+\delta t)\rangle$ $\displaystyle=$
$\displaystyle\sum_{i\neq
L1,L2}\phi(R_{i})|\psi\rangle(W(|\psi\rangle)|R_{i}\rangle)$ (28)
$\displaystyle+$
$\displaystyle\phi(R_{L1})(Z_{b}|\psi\rangle)(W(|\psi\rangle)|R_{L1}\rangle)$
$\displaystyle+$
$\displaystyle\phi(R_{L2})(X_{b}|\psi\rangle)(W(|\psi\rangle)|R_{L2}\rangle)$
Assuming the spin state is continuously monitored in the Bell basis, the above
state continuously collapses to a random separable state allowing the phase
information and correlations of the global state to be ignored. Note that
under this monitoring assumption one can describe the dynamics of the system
within a simpler scenario and yet verify the error correction capability of
the proposed qubot. Moreover, monitoring of the joint spin state in the Bell
basis can be achieved by continuous measurement of the force acting on
particle $a$, since the interaction between the particles is given by their
joint spin state. The motion state then acts as an environment for the spin,
inducing corrective jump operators,
$\displaystyle L_{1}=\sqrt{\gamma_{L1}}Z_{b}$ (29) $\displaystyle
L_{2}=\sqrt{\gamma_{L2}}X_{b}$ (30)
where we define correction rates as
$\displaystyle\gamma_{L1}dt$ $\displaystyle=$
$\displaystyle|\phi(R_{L1})|^{2}$ (31) $\displaystyle\gamma_{L2}dt$
$\displaystyle=$ $\displaystyle|\phi(R_{L2})|^{2}$ (32)
Note that the probability of a given corrective jump occuring is also the
probability of finding the particle in the corresponding corrective site. In
addition to corrective jumps the spin state is also under the effect of a
depolarizing channel due to an external decoherence environment, defined in
terms of the collapse operators
$\displaystyle L_{3}=\sqrt{\dfrac{\Gamma}{3}}X_{b}\ ,\
L_{4}=\sqrt{\dfrac{\Gamma}{3}}Y_{b}\ ,L_{5}=\sqrt{\dfrac{\Gamma}{3}}Z_{b}\ ,\
$ (33)
where $\Gamma$ is the decoherence rate.
Conversely spin acts as an environment to the motion state. If no spin
corrective jump occurs the motion state is left almost unperturbed, according
to (28), and evolves through the unitary predicted by the spin state
$|\psi\rangle$ plus the effects of a damping collapse operator provided by an
additional spin-insensitive cooling environment with damping rate $\kappa$
acting as a drain of kinetic energy, as discussed previously. On the other
hand, if a corrective jump $L_{1}$ or $L_{2}$ happens the motion state
collapses to $|R_{L1}\rangle$ or $|R_{L2}\rangle$, respectively. The collapsed
state subsequently evolves according to the unitary predicted by the spin
state $|\psi\rangle$ plus the additional damping collapse operator. When spin
jumps happen, the motion Hamiltonian must be updated accordingly for the next
time iteration.
This evolution can be implemented via a coupled Monte-Carlo method. First,
define the Motion Monte-Carlo procedure (MMC) for a damped harmonic oscillator
as following:
* (1)
Define motion state $|\phi\rangle$ and Hamiltonian $H$;
* (2)
Compute $\delta v=\kappa\delta t\langle\phi|a^{\dagger}a|\phi\rangle$;
* (3)
Choose uniformly distributed random number $q\in[0,1]$;
* (4)
If $q<\delta v$, update $|\phi\rangle\leftarrow a|\phi\rangle/\sqrt{\delta
v/\delta t}$;
* (5)
If $q>\delta v$, update $|\phi\rangle\leftarrow e^{-i\hat{H}\delta
t}|\phi\rangle/\sqrt{1-\delta v}$, where $\hat{H}=H-\frac{i}{2}a^{\dagger}a$;
We denote by MMC$(|\phi\rangle,H,\delta t)$ the output of the above procedure
for input state $|\phi\rangle$, Hamiltonian $H$, over a time step $\delta t$.
This output consists of the updated motion state after one time step.
The following algorithm, dubbed Spin-Motion Monte Carlo (SMMC), summarizes one
time iteration of the qubot dynamics:
* (1)
Define (update) motion and spin states $|\phi\rangle$ and $|\psi\rangle$ and
motion Hamiltonian $H=H(|\psi\rangle)$;
* (2)
Define correction rates $\gamma_{L1}\delta t=|\langle
R_{L1}|\phi\rangle|^{2},\gamma_{L2}\delta t=|\langle R_{L2}|\phi\rangle|^{2}$,
where $\delta t$ is the discrete time increment;
* (3)
Compute $\delta p_{k}=\delta t\langle\psi|L_{k}^{\dagger}L_{k}|\psi\rangle$
and $\delta p=\sum_{k}\delta p_{k}$;
* (4)
Choose uniformly distributed random number $r\in[0,1]$;
* (5)
If $r<\delta p$, update $|\psi\rangle\leftarrow
L_{k}|\psi\rangle/\sqrt{dp_{k}/\delta t}$ with probability $\delta
p_{k}/\delta p$;
(5.1) If jumps $L_{k}$ with $k=1$ or $2$ occurred, update
$|\phi\rangle\leftarrow|R_{L_{k}}\rangle$ and run
MMC$(|R_{L_{k}}\rangle,H,\delta t)$. After MMC update the motion state and the
motion Hamiltonian to $H=H(L_{k}|\psi\rangle)$;
(5.2) If jumps $L_{k}$ with $k=3,4$ or $5$ occurred, run
MMC$(|\phi\rangle,H,\delta t)$. After MMC update the motion state and the
motion Hamiltonian to $H=H(L_{k}|\psi\rangle)$;
* (6)
If $r>\delta p$, update $|\psi\rangle\leftarrow e^{-iH_{s}\delta
t}|\psi\rangle/\sqrt{1-\delta p}$, where
$H_{s}=-i\sum_{k}L_{k}^{\dagger}L_{k}$;
(6.1) Run MMC$(|\phi\rangle,H,\delta t)$. After MMC update the motion state
and the motion Hamiltonian to $H=H(|\psi\rangle)$;
* (7)
Go to (1) for next iteration.
A time series of quantum states $\\{|\psi(t)\rangle,|\phi(t)\rangle\\}$ is
called a quantum trajectory of the system, and can be obtained by iterating
SMMC. Mean behavior of the qubot can be obtained by averaging quantities of
interest over many quantum trajectories. For example, we can define the
overlap between the qubot spin state and the protected Bell state as
$F=\mathbb{E}\left[|\langle\psi(t)|\phi^{+}\rangle|^{2}\right]$, where
$\mathbb{E}\left[...\right]$ denotes the ensemble average over all quantum
trajectories. The quantity $F$ then measures how close the qubot spin state is
on average to the protected state and hence quantifies how well the qubot
functions.
To simplify the dynamics simulation, spin-dependent potentials are taken to be
harmonic traps of equal resonance frequency. This removes any issues due to
anharmonicity in the potentials and allows for the definition of fixed phonon
creation and annihilation operators. The potentials shown in Figure 4(b) are
approximated as
$\displaystyle
V(|\psi\rangle,R)=\dfrac{m\omega_{t}^{2}}{2}\left[R-R_{0}(|\psi\rangle)\right]^{2}$
(34)
where $\omega_{t}/2\pi=\SI{1}{kHz}$ and the trap position
$R_{0}(|\psi\rangle)$ is given by
Figure 6: Coupled spin-motion Monte-Carlo simulation of the qubot, $10^{3}$
quantum trajectories. Top: average fidelity to the $|\phi^{+}\rangle$ Bell
state as a function of time for the qubot plus a depolarizing channel (thick
green line) compared to the action of a depolarizing channel alone (thin
purple line). Middle: average position of the atom with corresponding quantum
uncertainty (light blue shade). Bottom: average correction rates $\gamma_{L1}$
(light yellow line) and $\gamma_{L2}$ (thick green line). The parameters used
in the plot are: decoherence rate $\Gamma=\SI{100}{Hz}$, trap frequency
$\omega_{t}=\SI{1}{kHz}$, damping rate
$\kappa=\SI{0.1}{ms}\times\omega_{t}^{2}$, initial wavepacket uncertainty
$\Delta R=\SI{0.22}{\mu m}$, $R_{L2}=-R_{L1}=\SI{0.63}{\mu m}$.
$\displaystyle
R_{0}(|\psi\rangle)=\left\\{\begin{array}[]{ll}R_{01},&\mathrm{if}\
|\phi^{+}\rangle\\\ R_{10},&\mathrm{if}\ |\phi^{-}\rangle\\\
R_{00},&\mathrm{if}\ |\psi^{\pm}\rangle\\\ \end{array}\right.$ (38)
The positions $R_{\alpha\beta}$ are dependent on the details of the
experimental implementation. Inspired by Figure 4(b) we consider
$R_{01}=\SI{1.90}{\mu m}$, $R_{10}=\SI{2.20}{\mu m}$ and $R_{00}=\SI{1.64}{\mu
m}$. Since the Hamiltonian always appears inside a commutator, constant terms
can be neglected without affecting the dynamics. Defining the origin of our
reference frame at the minimum of the potential $V(|\phi^{+}\rangle)$ and
neglecting constant shifts, the Hamiltonian reads
$\displaystyle H(|\psi\rangle)=\omega_{t}a^{\dagger}a-m\omega_{t}^{2}\Delta
R_{0}(|\psi\rangle)R_{\mathrm{zpm}}\left(a^{\dagger}+a\right)$ (39)
with $a^{\dagger},a$ the creation and annihilation operators for the
$|\phi^{+}\rangle$ potential, given by,
$\displaystyle
a=\sqrt{\dfrac{m\omega_{t}}{2}}\left(R+\dfrac{i}{m\omega_{t}}P\right)$ (40)
$\displaystyle
a^{\dagger}=\sqrt{\dfrac{m\omega_{t}}{2}}\left(R-\dfrac{i}{m\omega_{t}}P\right)$
(41)
with $R,P$ the atom position and momentum operators of particle $b$,
respectively, $R_{\mathrm{zpm}}$ the corresponding zero-point motion and
$\Delta R_{0}(|\psi\rangle)=R_{0}(|\psi\rangle)-R_{0}(|\phi^{+}\rangle)$. The
effect of a change in the spin state can be interpreted as the appearance of
an additional force acting on particle $b$.
Figure 6 shows the result of iterating SMMC averaged over $10^{3}$ quantum
trajectories, implemented using QuTiP [73], for the initial Bell-position
state $|\phi^{+}\rangle|\chi\rangle$, where $|\chi\rangle$ is a Gaussian
wavepacket in position with uncertainty $\Delta R$. See the Figure caption for
details on the parameters used in the simulation. The top graph shows the mean
overlap $F=\mathbb{E}\left[|\langle\psi(t)|\phi^{+}\rangle|^{2}\right]$ as a
function of time for the qubot (thick green line) compared to the depolarizing
channel alone (thin purple line). We can see that initially the qubot overlap
drops faster than the free spins, but it stabilizes at about $70\%$, while
free decohering spins decrease significantly below. The middle plot shows the
atom position and its quantum uncertainty as a function of time: action of the
qubot stabilizes the location of the atom. Note that motion of the atom
towards one corrective site is expected to increase correction rates of that
site and decrease correction rates of the other. This behavior can be seen in
the bottom graph, where rates are shown as a function of time. As expected,
$\gamma_{L1}$ (light yellow line) displays significant anti-correlation with
$\gamma_{L2}$ (thick green line).
The effect of finite temperature can be evaluated by adapting SMMC to include
motion collapse operators $\sqrt{\kappa(\bar{n}+1)}a$ and
$\sqrt{\kappa\bar{n}}a^{\dagger}$ representing contact with a thermal bath of
phonons at temperature $T$ with coupling $\kappa$ and thermal occupation
number $\bar{n}$, where $\bar{n}=1/(e^{\hbar\omega_{t}/k_{B}T}-1)$. When in
contact with a thermal bath, the particle initially in the ground state
evolves to a thermal state with mean number of phonons $\bar{n}$, increasing
the position spread and consequently the intrinsic qubot error rate. The spin
overlap is thus expected to decrease with temperature.
Figure 7: Effect of contact with a thermal bath at temperature $T$ upon the
steady state time-averaged overlap $\langle F\rangle_{s}$. Time average is
considered starting at 10ms, when the overlap has already achieved its steady
value. Error bars correspond to one standard deviation. Each point is
evaluated from $10^{2}$ quantum trajectories. Coupling to the heat bath is
$\kappa=\SI{0.1}{ms}\times\omega_{t}^{2}$ and all remaining parameters are the
same as in Figure 6.
The time-averaged steady state overlap $\langle F\rangle_{s}$ as a function of
temperature is plotted in Figure 7. Each point is the result of time-averaging
$10^{2}$ quantum trajectories with error bars corresponding to one standard
deviation. As expected, the effect of contact with a heat bath is to decrease
the overlap.
Figure 8: Influence of corrector positioning. Top: steady state overlap.
Bottom: mean correction rates. Averages are considered from $\SI{10}{ms}$
onward, when the device is well settled in the steady state. Error bars
correspond to one standard deviation. Each point is evaluated from $10^{2}$
quantum trajectories. All remaining parameters are the same as in Figure 6.
Quantum fluctuations of the atomic motion can induce internal errors if the
atom interacts with the correctors when no external (decoherence) error has
taken place. To quantify that effect, the steady state overlap $\langle
F\rangle_{s}$ and correction rates $\langle\gamma\rangle_{s}$ are numerically
calculated for different values of the $L1$ position $|R_{L1}|$, shown in
Figure 8; $R_{L1}=-R_{L2}$ is assumed. Note that if the correctors are too
close to the equilibrium position of $|\phi^{+}\rangle($
$|R_{L1}|<\SI{0.40}{\mu m}$), the steady state overlap $\langle F\rangle_{s}$
falls below 50%, while the mean rate for ‘correction’ events are on the order
of 1kHz, due to the atom fluctuating towards $L1$ or $L2$ even in the absence
of an error. As $|R_{L1}|$ is increased, the steady state overlap increases,
reaching a maximum value $\langle F\rangle_{s}\approx 0.7$ for
$|R_{L1}|\approx 0.63$, and then decreases again as the correctors are placed
further apart from the atom. The mean correction rates can be seen to decrease
as the position $|R_{L1}|$ is further increased, which is intuitive since
larger distances imply longer correction times. The optimal operation point
$|R_{L1}|\approx 0.63$ is such that the mean correction rates
$\langle\gamma\rangle$ are of the same order of the decoherence rate
$\Gamma=\SI{100}{Hz}$. See Appendix D for more details.
## V Discussion
Throughout this work we discussed quantum robots, devices as the one
conceptualized in [18], capable of harnessing interactions between its
constituent parts and the surrounding environment to achieve targeted tasks
such as state protection against decoherence. We have introduced for the first
time a model of a qubot capable of partially protecting an arbitrary logical
qubit state against general single physical qubit errors. The first physical
implementation of an instance of such device, capable of protecting a Bell
state against the detrimental action of a depolarizing environment has been
described, as well as Monte-Carlo simulations of the qubot dynamics and the
inclusion of effects due to contact of the device with a thermal bath.
From where we stand, several directions for future exploration can be sighted.
For instance, a more thorough investigation of the capabilities of the
proposed entanglement qubot remains to be done: by tuning the relevant
parameters such as the Rydberg level detunings $\Delta_{\pm}$ and trap
potential $V_{t}(R)$ can we engineer a qubot capable of protecting entangled
states other than the $|\phi^{+}\rangle$ state? What about implementing a
system analogous to the conceptual model, capable of protecting an arbitrary
logical qubit? Could we extend the device to handle multiple qubits? Would the
protection work against general physical errors? We have focused on the
implementation using Rydberd-dressed atoms, but that is certainly not the only
possibility. What other opportunities are offered by considering different
physical setups for qubots? Polar molecules provide a promising platform [36,
45, 46, 19] with the possibility of coupling to superconducting quantum
electronics [56].
Synthetic molecular machines are one of the frontiers of nanotechnology [6, 8,
9, 21]. Enabled by the idea of a quantum robot we can envision extensions of
the molecular machinery toolbox where the quantum states of the nanomachines
play a fundamental role in their dynamics. These devices would combine
resources from the environment, stochasticity and non-equilibrium to execute
coupled quantum motion and processing of quantum information entering the
realm of quantum nanomechanics. For example, in the entanglement qubot one
could set the correction sites to perform the operation $L1=L2=X_{b}$, and
initiate the spins in the state $|\psi^{+}\rangle$. This would cause a
periodic spin-driven motion of the atom. It would be interesting to
investigate the possibility of building quantum time crystals [38, 74, 75]
using this scheme.
Quantum robots with no moving parts are also a hitherto unexplored direction.
In such devices an error in one degree of freedom would unleash a chain of
reactions in other internal non-mechanical parts of the system, which would
act back on the affected degree of freedom and steer it to a desired state.
This touches upon the theoretical issue of quantum feedback [76, 77], in a
situation where the feedback itself is carried by quantum mechanical
information, rather than the usual classical information scheme in which a
measurement result is used to counter-act on the system.
Finally, a very intriguing thought is the combination of a large number of
quantum robots interacting with each other. Large numbers of interacting
classical active agents display fascinating emergent behavior [78, 79].
Ensembles of active quantum agents on the other hand remain unexplored. Qubots
offer a concrete path towards experimentally uncovering the physics of quantum
active matter.
## Acknowledgements
I thank Bruno Melo, Lucianno Defaveri, Bruno Suassuna, Igor Brandão and George
Svetlichny for useful discussions and feedback on the manuscript. This work
was financed in part by the Coordenacão de Aperfeiçoamento de Pessoal de Nível
Superior - Brasil (CAPES) - Finance Code 001, Conselho Nacional de
Desenvolvimento Científico e Tecnológico (CNPq) and the FAPERJ Scholarship No.
E-26/202.830/2019.
## VI Appendix A: effective potentials
As described in the main text, admixing strongly interacting Rydberg states
from the $n^{2}P_{1/2}$ manifold to the low-lying $5^{2}S_{1/2}$ Zeeman
sublevels induces spatial dependent spin-spin interactions of the form (20).
For completeness we reproduce the main results of [42] outlining the toolbox
for engineering a wide range of effective spin interactions.
The interaction coefficients $J_{\alpha}$ are calculated by adiabatic
elimination of the Rydberg levels $|r_{\pm}\rangle$ up to fourth order in
$\Delta/\Omega$, and are given by
$\displaystyle
J_{z}(R)=\dfrac{1}{4}\left(\tilde{V}_{--}(R)-2\tilde{V}_{+-}(R)+\tilde{V}_{++}(R)\right)\
,$ (42) $\displaystyle
J_{x}(R)=2\left(\tilde{W}_{+-}(R)+\tilde{W}_{++}(R)\right)\ ,$ (43)
$\displaystyle J_{y}(R)=2\left(\tilde{W}_{+-}(R)-\tilde{W}_{++}(R)\right)\ ,$
(44) $\displaystyle
J_{\parallel}(R)=\dfrac{1}{4}\left(\tilde{V}_{--}(R)-\tilde{V}_{++}(R)\right)\
,$ (45)
where the functions $\tilde{W}_{\alpha\beta},\tilde{V}_{\alpha\beta}$ are
effective radial dependent steplike potentials,
$\displaystyle\tilde{V}_{\alpha\alpha}(R)=\dfrac{\Omega^{2}_{\bar{\sigma}}}{2\Delta_{\bar{\sigma}}}-\dfrac{\Omega^{4}_{\bar{\sigma}}}{4\Delta_{\bar{\sigma}}^{3}}+\dfrac{\Omega_{\bar{\alpha}}^{4}}{4\Delta_{\bar{\alpha}}^{2}}\dfrac{V_{++}-2\Delta_{\alpha}}{W_{++}^{2}-(V_{++}-2\Delta_{+})(V_{++}-2\Delta_{-})}$
(46)
$\displaystyle\tilde{V}_{+-}(R)=\dfrac{\Omega^{2}_{-}}{4\Delta_{-}}+\dfrac{\Omega^{2}_{+}}{4\Delta_{+}}-\dfrac{\Omega_{+}^{2}\Omega_{-}^{2}}{16\Delta_{+}^{2}\Delta_{-}}-\dfrac{\Omega_{+}^{2}\Omega_{-}^{2}}{16\Delta_{-}^{2}\Delta_{+}}-\dfrac{\Omega^{4}_{-}}{16\Delta_{-}^{3}}-\dfrac{\Omega^{4}_{+}}{16\Delta_{+}^{3}}+\dfrac{\Delta_{\pm}^{2}\Omega_{+}^{2}\Omega_{-}^{2}}{16\Delta_{+}^{2}\Delta_{-}^{2}}\dfrac{(\Delta_{\pm}-V_{+-})}{(\Delta_{\pm}-V_{+-})^{2}-W_{+-}^{2}}$
(47) $\displaystyle\tilde{W}_{+-}(R)$ $\displaystyle=$
$\displaystyle\dfrac{\Omega_{+}^{2}\Omega_{-}^{2}}{16\Delta_{+}^{2}\Delta_{-}^{2}}\dfrac{\Delta_{\pm}^{2}W_{+-}}{(\Delta_{\pm}-V_{+-})^{2}-W_{+-}^{2}}$
(48) $\displaystyle\tilde{W}_{++}(R)$ $\displaystyle=$
$\displaystyle\dfrac{\Omega_{+}^{2}\Omega_{-}^{2}}{4\Delta_{+}\Delta_{-}}\dfrac{W_{++}}{W_{++}^{2}-(V_{++}-2\Delta_{+})(V_{++}-2\Delta_{-})}$
(49)
written in terms of the $n^{2}P_{1/2}$ van der Waals potentials
$V_{\alpha\beta},W_{\alpha\beta}$. Note the single particle light-shifts have
been included in the above expressions. Moreoever,
$\tilde{V}_{+-}=\tilde{V}_{-+}$, and we have defined
$\Delta_{\pm}=\Delta_{+}+\Delta_{-}$ and $\bar{\alpha}=-\alpha$. In the
parameter region $\Delta_{+-}<0$, $\Delta_{+}/\Delta_{-}<0$ resonant Rydberg
excitations are avoided for all values of $R$. For atomic orientation
$\theta=\pi/2$ (polar), $\phi=0$ (azimuthal) the van der Waals potentials are
$\displaystyle V_{\alpha\beta}=\dfrac{c_{\alpha\beta}}{R^{6}}\ ,\
W_{+-}=\dfrac{w}{R^{6}}=-\dfrac{1}{3}W_{++}\ .$ (50)
where the so-called $C_{6}$ coefficients $c_{\alpha\beta}$ and $w$ are
obtained from second order perturbation theory, and are given by
$\displaystyle c_{++}$ $\displaystyle=$
$\displaystyle\dfrac{2}{81}\left(5C_{6}^{(a)}+14C_{6}^{(b)}+8C_{6}^{(c)}\dfrac{}{}\right)$
(51) $\displaystyle c_{+-}$ $\displaystyle=$
$\displaystyle\dfrac{2}{81}\left(C_{6}^{(a)}+10C_{6}^{(b)}+16C_{6}^{(c)}\dfrac{}{}\right)$
(52) $\displaystyle w$ $\displaystyle=$
$\displaystyle\dfrac{2}{81}\left(C_{6}^{(a)}+C_{6}^{(b)}-2C_{6}^{(c)}\dfrac{}{}\right)$
(53)
The indivitual channel coefficients $C_{6}^{(\nu)}$, $\nu=a,b,c$ are not
dependent of magnetic quantum numbers and characterize the interaction
strengh. There is one channel for each non-vanishing matrix element of the
dipole-dipole interaction potential [42],
$\displaystyle a$ $\displaystyle:$ $\displaystyle\ P_{1/2}+P_{1/2}\rightarrow
S_{1/2}+S_{1/2}$ (54) $\displaystyle b$ $\displaystyle:$ $\displaystyle\
P_{1/2}+P_{1/2}\rightarrow D_{3/2}+D_{3/2}$ (55) $\displaystyle c$
$\displaystyle:$ $\displaystyle\ P_{1/2}+P_{1/2}\rightarrow D_{3/2}+S_{1/2}$
(56)
and each $C_{6}^{(\nu)}$ is calculated from the radial part of the dipole-
dipole matrix element [68],
$\displaystyle
C^{(\nu)}_{6}=\sum_{n_{\alpha}n_{\beta}}\dfrac{e^{4}}{\delta_{\alpha\beta}}\left(R_{nl}^{n_{\alpha}l_{\alpha}}R_{nl}^{n_{\beta}l_{\beta}}\right)^{2}$
(57)
where
$\displaystyle R_{nl}^{n_{i}l_{i}}=\int
drr^{2}\psi_{n,l,j}(r)^{*}r\psi_{n_{i},l_{i},j_{i}}(r)\ ,$ (58)
and $\delta_{\alpha\beta}$is the energy defect between levels $n_{\alpha}$ and
$n_{\beta}$.
Figure 9: $C_{6}^{(\nu)}$ coefficients as a function of principal quantum
number for the $n^{2}P_{1/2}$ manifold.
To numerically obtain the coefficients (57), and consequently the step-like
potentials (47) and (49), we use the ARC python library for alkali Rydberg
atoms [69]. Numerical calculation results are shown in Figure 9 as a function
of the principal quantum number for the $n^{2}P_{1/2}$ manifold. For $n=60$,
as used in the main text, we find
$\displaystyle-C_{6}^{(a)}\approx 2\pi\times\SI{2.7E5}{MHz\cdot\mu m^{6}}$
(59) $\displaystyle C_{6}^{(b)}\approx 2\pi\times\SI{1.1E3}{MHz\cdot\mu
m^{6}}$ (60) $\displaystyle C_{6}^{(c)}\approx
2\pi\times\SI{4.9E4}{MHz\cdot\mu m^{6}}$ (61)
## VII Appendix B: Magnetic field $J_{\parallel}$
Besides the $J_{\alpha}(R)$ coefficients, the Rydberg dressing generates an
effective magnetic field term $J_{\parallel}(Z_{a}+Z_{b})$ in the interaction
energy. Under the influence of this term, Bell states of the $ab$ pair are no
longer eigenstates of the interaction. To obtain the spin dependent potential
landscapes given by the eigenvalues in Eqs.(5)-(8), we need to cancel
$J_{\parallel}$ by applying an external spatial dependent static field. How
large such a field needs to be? A plot of $J_{\parallel}$ can be seen in
Figure 10.
Figure 10: $J_{\parallel}$ profile.
Note that $\langle J_{\parallel}\rangle\approx\SI{1401}{kHz}$. Considering the
Landé factor $|g_{F}|\approx\SI{0.70}{MHz/G}$ for the $5^{2}S_{1/2}$ states
[67] this effective magnetic field can be cancelled by an additional weak non-
homogeneous field of order of magnitude $|B_{c}|\approx\SI{2}{G}$.
## VIII Appendix C: Alternative spin pattern
Alternative spin dependent potentials, defined by parameters different from
the ones employed in the main text are shown in Figure 11. Here, we consider
detunings $\Delta_{+}=-2\pi\times\SI{70}{MHz}$,
$\Delta_{-}=2\pi\times\SI{30}{MHz}$, Rabi frequencies
$\Omega_{+}=\Omega_{-}=-2\pi\times\SI{7}{MHz}$ and the trap potential
$\displaystyle V_{t}(R)=V_{0}\left(R-\delta\right)^{2}$ (62)
where $V_{0}=\SI{15}{kHz/\mu m^{2}}$ and $\delta=\SI{2.30}{\mu m}$.
Figure 11: (a) Alternative spin pattern profile. (b) Resulting spin dependent
potentials. (c) Resulting spin dependent potentials after adding the spin-
independent harmonic potential.
Note the resulting landscapes in Figure 11(c) suggest $|\phi^{-}\rangle$ as
protected state, while corrective loops $L1$ and $L2$ should be reversed with
respect to the choice discussed in the main text. The effective magnetic field
has a mean value $\langle J_{\parallel}\rangle\approx\SI{1803}{kHz}$, which
requires a slightly higher compensating magnetic field, but still on the order
of a few Gauss. The spatial profile $J_{\parallel}(R)$ is shown in Figure 12.
Figure 12: Alternative effective magnetic field.
## IX Appendix D: Optimal operation
To evaluate the effect of positioning of the correctors $L1$ and $L2$, we ran
SMMC, as described in the main text, for different values of the positions
$R_{L1}=-R_{L2}$.
Figure 13: Overlap $F$ for different values of the corrector position
$|R_{L1}|=|R_{L2}|$ obtained from simulating $10^{2}$ quantum trajectories.
For each trace position values are $(0.47,0.52,0.58,0.63,0.69,0.74,0.80)\
\SI{}{\mu m}$ from top to bottom, respectively. Grey dashed lines represent
the overlap of free decohering spins, for comparison. All remaining parameters
are the same as in Figure 6 in the main text.
Figure 13 shows traces of the overlap $F$ as a function of time. Each trace
corresponds to a different corrector position (see caption), and the overlap
of free spins under the action of the depolarizing channel is shown as the
grey dashed line for comparison. The points in Figure 8 (see main text) are
obtained by time-averaging the overlap above $\SI{10}{ms}$ for each of the
traces in Figure 13.
We can see that if the corrector’s positions are too close to the atom’s
equilibrium position, the overlap quickly decays due to internal errors,
occurring when a quantum fluctuation in the atomic position places it near the
corrective site. This fast drop in overlap can be mitigated by positioning the
correctors further apart from the $|\phi^{+}\rangle$ equilibrium point. There
is, however, a trade-off: the maximum steady state overlap $\approx 70\%$ is
reached for a position $|R_{L1}|\approx\SI{0.63}{\mu m}$, while placing the
correctors further than that reduce the correction rates below the decoherence
rate and consequently the steady state overlap.
Decoherence causes the overlap to decrease exponentially according to
$e^{-\Gamma t}=e^{-t/\tau_{D}}$, where
$\tau_{D}=\Gamma^{-1}=(\SI{100}{Hz})^{-1}=\SI{10}{ms}$ is the characteristic
decay time of the system. Decoherence effectivelly freezes when the system
reaches its steady state, which happens after a stabilization time $t_{s}$
elapses. From Figure 6 in the main text, we see that $t_{s}\approx\SI{4}{ms}$,
yielding an expected overlap of $F\approx e^{-t_{s}/\tau_{D}}\approx 0.67$, in
accordance to the simulation results.
## References
* [1] C. H. Bennet, The thermodynamics of computation - a review, International Journal of Theoretical Physics volume 21, 905 (1982)
* [2] D. S. Goodsell, The machinery of life, Copernicus
* [3] B. Alberts et. al., Essential Cell Biology, Garland Science; 3rd edition (March 27, 2009)
* [4] R. Milo and R. Phillips, Cell Biology by the Numbers, Garland Science
* [5] R. P. Feynman, There’s plenty of room at the bottom, (1959)
* [6] L. Zhang et. al., Molecular machines with bio-inspired mechanisms, PNAS 115, 9397 (2018)
* [7] V. Serreli et. al., A molecular information ratchet, Nature 445, 523 (2007)
* [8] C. Cheng et. al., An artificial molecular pump, Nature Nanotechnology 10, 547 (2015)
* [9] S. Kassem et. al., Artificial molecular motors, Chemical Society Reviews. 46, 2592 (2017)
* [10] J. Doudna et. and S. H. Sternberg, A Crack in Creation: Gene Editing and the Unthinkable Power to Control Evolution, Houghton Mifflin Harcourt
* [11] W. Bialek, Biophysics: searching for principles, Princeton univdersity press
* [12] A. Chakraborty, C. A. M. Steven and M. Block, Observing Single RNA Polymerase Molecules Down to Base-Pair Resolution, Optical Tweezers 391 (2016)
* [13] P. C. Bressloff, Stochastic Processes in Cell Biology, Springer
* [14] Google AI Quantum, Hartree-Fock on a superconducting qubit quantum computer, Science 369 (6507), 1084 (2020)
* [15] C. Outeiral et. al., The prospects of quantum computing in computational molecular biology, WIREs Computational Molecular Science (2020)
* [16] P. S. Emani et. al., Quantum computing at the frontiers of biological sciences, Nature Methods (2021)
* [17] H. Frauenfelder, Ask not what physics can do for biology - ask what biology can do for physics , Phys. Biol. 11, 053004 (2014)
* [18] T. Guerreiro, Quantum molecular robots, Quantum Science and Technology (2020)
* [19] L. D. Carr et. al., Cold and ultracold molecules: science, technology and applications, New Journal of Physics, 11 (2009)
* [20] R. V. Krems, Molecules in Electromagnetic Fields: From Ultracold Physics to Controlled Chemistry, John Wiley & Sons (2018)
* [21] B. Lau et. al., An introduction to ratchets in chemistry and biology, Material Horizons, 3 (2017)
* [22] D. Budker, D. Kimball, D. DeMille, Atomic physics: An exploration through problems and solutions, Oxford University Press
* [23] S. Ospelkaus et. al., Quantum-State Controlled Chemical Reactions of Ultracold Potassium-Rubidium Molecules, Science 327, 853 (2010)
* [24] L. R. Liu et. al., Ultracold Molecular Assembly, arXiv:1701.03121
* [25] M. B. Plenio, S. F. Huelga, A. Beige, and P. L. Knight, Cavity loss induced generation of entangled atoms, Phys. Rev. A 59, 2468 (1999)
* [26] M. B. Plenio and S. F. Huelga, Entangled Light from White Noise, Phys. Rev. Lett. 88, 197901
* [27] S. Diehl et. al., Quantum States and Phases in Driven Open Quantum Systems with Cold Atoms, Nature Physics 4, 878 (2008)
* [28] F. Verstraete et. al., Quantum computation, quantum state engineering, and quantum phase transitions driven by dissipation, Nature Physics 5, 633 (2009)
* [29] G. Vacanti and A. Beige, Cooling atoms into entangled states, New J. Phys. 11 083008 (2009)
* [30] F. Reiter, M. J Kastoryano and A. S. Sorensen, Driving two atoms in an optical cavity into an entangled steady state using engineered decay, New J. Phys. 14 (2012)
* [31] J. B. Brask, N. Brunner, G. Haack and M. Huber, Autonomous quantum thermal machine for generating steady-state entanglement, New Journal of Physics 17, 113029 (2015)
* [32] F. Reiter, A. S. Sorensen, P. Zoller and C. A. Muschik , Dissipative quantum error correction and application to quantum sensing with trapped ions, Nature Communications 8, 1822 (2017)
* [33] C. Cheng et. al., Design and Synthesis of Nonequilibrium Systems, ACS Nano 9, 8672 (2015)
* [34] A. Y. Kitaev, Unpaired Majorana fermions in quantum wires, Phys.-Usp. 44 131 (2001)
* [35] A. G. Fowler, M. Mariantoni, J. M. Martinis and A. N. Cleland, Surface codes: Towards practical large-scale quantum computation, Phys. Rev. A 86, 032324 (2012)
* [36] Q. Wei et. al., Entanglement of polar molecules in pendular states, J. Chem. Phys. 134, 124107 (2011)
* [37] J. Pietraszewicz et. al., Spin dynamics of two bosons in an optical lattice site: A role of anharmonicity and anisotropy of the trapping potential, Phys. Rev. A 88, 013608 (2013)
* [38] S. Choi et. al., Observation of discrete time-crystalline order in a disordered dipolar many-body system, Nature 543, 221- 225 (2017)
* [39] S. Weinberg, Lectures on quantum mechanics, Cambridge university press (2012)
* [40] D. Porras and J.I. Cirac, Effective quantum spin systems with ion traps, Phys. Rev. Lett. 92, 207901 (2004)
* [41] K. Kim et. al., Quantum simulation of frustrated Ising spins with trapped ions, Nature volume 465, pages590–593(2010)
* [42] A. W. Glaetzle et. al., Designing Frustrated Quantum Magnets with Laser-Dressed Rydberg Atoms, Phys. Rev. Lett. 114, 173002 (2015)
* [43] R. M. W. van Bijnen and T. Pohl, Quantum magnetism and topological ordering via enhanced Rydberg-dressing near Forster-resonances, Phys. Rev. Lett. 114, 243002 (2015)
* [44] Y.-Y. Jau et. al., Entangling atomic spins with a Rydberg-dressed spin-flip blockade, Nature Physics 12, 71 (2016)
* [45] A. Micheli, G.K. Brennen and P. Zoller, A toolbox for lattice spin models with polar molecules, Nature Physics, 2, 341-347 (2006)
* [46] G.K. Brennen, A. Micheli and P. Zoller, Designing spin-1 lattice models using polar molecules, New J. Phys. 9 138 (2007)
* [47] D. Gottesman, A Class of Quantum Error-Correcting Codes Saturating the Quantum Hamming Bound, Phys.Rev. A54 1862 (1996)
* [48] R. Laflamme et. al., Perfect Quantum Error Correcting Code, Phys. Rev. Lett. 77, 198 (1996)
* [49] C. H. Bennett, Notes on Landauer’s principle, Reversible Computation and Maxwell’s Demon, Studies in History and Philosophy of Modern Physics vol. 34 pp. 501-510 (2003)
* [50] J. Bub, Maxwell’s demon and the thermodynamics of computation, Studies in History and Philosophy of Modern Physics 32, 569-579 (2001)
* [51] Jun Ye, H. J. Kimble, H. Katori, Quantum State Engineering and Precision Metrology using State-Insensitive Light Traps, Science 320, 1734-1738 (2008)
* [52] S. Zhang et. al., Magic-wavelength optical traps for Rydberg atoms, Phys. Rev. A 84, 043408 (2011)
* [53] A. G. Boetes et. al., Trapping of Rydberg atoms in tight magnetic microtraps, Phys. Rev. A 97, 013430 (2018)
* [54] A, Glaetzle et. al., Quantum Spin Ice and dimer models with Rydberg atoms, Phys. Rev. X 4, 041037 (2014)
* [55] D. Kielpinski et. al., Quantum Interface between an Electrical Circuit and a Single Atom, Phys. Rev. Lett. 108, 130504 (2012)
* [56] A. Andre et. al., A coherent all-electrical interface between polar molecules and mesoscopic superconducting resonators, Nature Physics 2, 636 (2006)
* [57] L. Martinetz et. al., Quantum electromechanics with levitated nanoparticles, npj Quantum Information volume 6, 101 (2020)
* [58] Y. Kubo et. al., Hybrid Quantum Circuit with a Superconducting Qubit Coupled to a Spin Ensemble, Phys. Rev. Lett. 107, 220501 (2011)
* [59] D. Jaksch et. al., Entanglement of atoms via cold controlled collisions, Phys.Rev.Lett. 82 1975 (1999)
* [60] R. P. Feynman, Feynman lectures on physics, Vol. III, Chapter 9, Addison-Wesley (2006)
* [61] R. Wynands and S. Weyers, Atomic fountain clocks, Metrologia, 42, 3 (2005)
* [62] D. M. Weld et. al., Spin Gradient Thermometry for Ultracold Atoms in Optical Lattices, Phys. Rev. Lett. 103, 245301 (2009)
* [63] A. J. Daley, P. O. Fedichev, and P. Zoller, Single-atom cooling by superfluid immersion: A nondestructive method for qubits, Phys. Rev. A 69, 022306 (2004)
* [64] A. Griessner, D. Jaksch and P. Zoller, Cavity-assisted nondestructive laser cooling of atomic qubits, J. Phys. B: At. Mol. Opt. Phys. 37 1419 (2004)
* [65] R. Belyansky et. al., Nondestructive cooling of an atomic quantum register via state-insensitive Rydberg interactions, Phys. Rev. Lett. 123, 213603 (2019)
* [66] V. Paulisch et. al., Beyond adiabatic elimination: A hierarchy of approximations for multi-photon processes, EPJ-Plus 129:12 (2014)
* [67] S. Bize et. al., High-accuracy measurement of the 87Rb ground-state hyperfine splitting in an atomic fountain, Europhys. Lett. 45, 558 (1999)
* [68] T. G. Walker and M. Saffman, Consequences of Zeeman Degeneracy for van der Waals Blockade between Rydberg Atoms, Phys. Rev. A 77, 032723 (2008)
* [69] M. Sibalic et. al., ARC: An open source library for calculating properties of alkali Rydber gatoms, Computer Physics Communications, 220, 319-331 (2017)
* [70] I. Beterov et. al., Quasiclassical calculations of blackbody-radiation-induced depopulation rates and effective lifetimes of Rydberg $nS$, $nP$ and $nD$ alkali-metal atoms with $n\leq 80$ , Phys. Rev. A 79, 052504 (2009)
* [71] J. Preskill, Quantum information lecture notes, Chapter 7, http://theory.caltech.edu/~preskill/ph229/notes/chap7.pdf
* [72] C. Misner, K. Thorne and W. Zurek, John Wheeler, relativity, and quantum information , Physics Today 62, 4, 40 (2009)
* [73] J. R. Johansson, P. D. Nation, and F. Nori, QuTiP 2: A Python framework for the dynamics of open quantum systems, Comp. Phys. Comm. 184, 1234 (2013)
* [74] F. Wilczek, Quantum Time Crystals, Phys. Rev. Lett. 109, 160401 (2012)
* [75] J. Zhang et. al., Observation of a Discrete Time Crystal, Nature volume 543, pages217–220(2017)
* [76] G. J. Milburn and H. M. Wiseman, Quantum Measurement and Control, Cambridge University Press
* [77] C. Ahn et. al., Continuous quantum error correction via quantum feedback control, Phys. Rev. A 65, 042301 (2002)
* [78] T. Vicsek et. al., Novel Type of Phase Transition in a System of Self-Driven Particles, Phys. Rev. Lett. 75, 1226 (1995)
* [79] D. Geyer et. al., Sounds and hydrodynamics of polar active fluids, Nature Materials volume 17, 789–793 (2018)
|
# What drives the European carbon market? Macroeconomic factors and
forecasts††thanks: The authors gratefully acknowledge the participants at the
4th IWEEE in Bolzano for their useful feedback. Luca Rossini acknowledges
financial support from the Italian Ministry MIUR under the PRIN-PNRR project
Mapping and Pricing of Methane Emissions from the European Electricity Sector
(MAP-of-MeLEES) (grant P2022H483A). This research used the Computational
resources provided by the Core Facility INDACO, which is a project of High-
Performance Computing at the University of Milan.
Andrea Bastianin University of Milan, Italy and Fondazione Eni Enrico Mattei
(FEEM<EMAIL_ADDRESS>Elisabetta Mirto University of Milan, Italy.
<EMAIL_ADDRESS>Yan Qin London Stock Exchange Group, Norway.
<EMAIL_ADDRESS>Luca Rossini University of Milan, Italy and Fondazione Eni
Enrico Mattei (FEEM<EMAIL_ADDRESS>
Abstract: Putting a price on carbon – with taxes or developing carbon markets
– is a widely used policy measure to achieve the target of net-zero emissions
by 2050. This paper tackles the issue of producing point, direction-of-change,
and density forecasts for the monthly real price of carbon within the EU
Emissions Trading Scheme (EU ETS). We aim to uncover supply- and demand-side
forces that can contribute to improving the prediction accuracy of models at
short- and medium-term horizons. We show that a simple Bayesian Vector
Autoregressive (BVAR) model, augmented with either one or two factors
capturing a set of predictors affecting the price of carbon, provides
substantial accuracy gains over a wide set of benchmark forecasts, including
survey expectations and forecasts made available by data providers. We extend
the study to verified emissions and demonstrate that, in this case, adding
stochastic volatility can further improve the forecasting performance of a
single-factor BVAR model. We rely on emissions and price forecasts to build
market monitoring tools that track demand and price pressure in the EU ETS
market. Our results are relevant for policymakers and market practitioners
interested in monitoring the carbon market dynamics.
Key Words: Bayesian inference; Carbon prices; Climate Changes; EU ETS;
Forecasting.
JEL Codes: C11; C32; C53; Q02; Q50.
## 1 Introduction
Climate change is one of the greatest challenges currently addressed by
governments, central banks, and other national and supranational regulators.
The size and complexity of the call to fight climate change are evident, even
when – setting aside its environmental consequences – we focus solely on its
direct socio-economic impacts (Carleton and Hsiang,, 2016). The literature has
highlighted a multitude of effects and transmission channels that connect the
environmental influences of climate change to various economic aggregates
(Dell et al.,, 2014; Ciccarelli and Marotta,, 2024). Moreover, when designing
mitigation and adaptation policies, a trade-off emerges between the stringency
of measures and the undesired outcomes that these might induce (Fullerton and
Muehlegger,, 2019; Hsiang et al.,, 2019).
Among the policy measures implemented to achieve the target of net-zero
emissions by 2050 – as laid down in the 2015 Paris Agreement – carbon taxes
and carbon pricing have lately received a lot of attention from academics and
policymakers interested in quantifying the desired and unintended
macroeconomic effects of putting a price on carbon (see e.g. European Central
Bank, 2021a, ; Parry et al.,, 2021; Pan et al.,, 2023). Given this literature
and the need to incorporate climate and carbon market modules into
macroeconomic models used by central banks, having access to reliable short-
and medium-term forecasts of the price of carbon is becoming increasingly
important (European Central Bank, 2021b, ; NGFS,, 2022).
This paper focuses on producing point, direction-of-change, and density
forecasts of the real price of carbon at a monthly sampling frequency in the
world’s largest carbon market: the EU Emission Trading Scheme (EU ETS). Our
research pursues two intertwined targets. First, we aim to uncover supply- and
demand-side forces that can contribute to improving the prediction accuracy of
models at short- and medium-term horizons. The second main objective of the
paper is to highlight which methodological choices have the potential to
improve the forecast accuracy of the estimated models.
Since the inception of the EU ETS in 2005, the research on carbon pricing has
addressed a variety of issues (see Chevallier,, 2012; Pan et al.,, 2023, for a
survey). In particular, Bjørnland et al., (2023) and Känzig, (2023) use
macroeconomic and financial data, along with microdata to evaluate the impact
of carbon price shocks. Recently Känzig and Konradt, (2023) and Moessner,
(2022) focus on European countries and the inflationary effects in a set of
OECD countries, respectively.
Our paper directly contributes to the literature dealing with forecasting the
price of carbon and identifying its determinants such as weather, energy
prices, macroeconomic and financial conditions (see e.g. Chevallier, 2011a, ;
Chevallier, 2011b, ; Koop and Tole,, 2013; Lei et al.,, 2022; Mansanet-
Bataller et al.,, 2007; Tan et al.,, 2022). Whether the results that emerge
from this strand of the literature on carbon price forecasting – mostly based
on daily and weekly data – can be directly applied to building monthly or
quarterly econometric models through appropriate time aggregation of variables
has to be empirically determined.
Differently from the literature, we focus on the real monthly price of carbon,
which enables us to provide results that are relevant to the dialogue about
the macroeconomic consequences of carbon pricing. Indeed, our analysis can be
closely related to Chevallier, 2011b , which evaluate the role of energy
prices and business cycle movements in driving the price of carbon, and to
Bjørnland et al., (2023), which use a Bayesian structural vector
autoregressive (VAR) model with endogenous variables (i.e. real price of
carbon, verified emissions, and industrial production) to identify demand and
supply shocks driving the emissions and the real price of carbon. Moreover,
while most of the papers dealing with carbon price forecasting rely on
frequentist methods, we use Bayesian techniques to readily incorporate
Stochastic Volatility (SV) dynamics into the models and to generate density
forecasts.111One notable exception is Koop and Tole, (2013) that also rely on
Bayesian methods but consider daily data. Reliance on SV models is another
novelty of our paper; in fact, several previous analyses have focused either
on conditionally homoskedastic models or have modeled conditional volatility
dynamics using Generalized AutoRegressive Conditional Heteroskedasticity
models.
After having assessed a suite of benchmark univariate time series models, we
consider small-scale Vector Autoregressive (VAR) models with endogenous
variables capturing forces affecting the real price of carbon. First, we focus
on the model put forth by Bjørnland et al., (2023) as a starting point. Next,
we extend their VAR model to include factors capturing the influence of
multiple predictors that affect the EU ETS real price.
While we initially compare models based on their ability to deliver accurate
point forecasts, we subsequently extend these results in several directions.
The first extension evaluates different specifications for their ability to
yield accurate sign forecasts and prediction densities. Directional accuracy
is mostly relevant at short-term forecast horizons, while accurate prediction
densities are useful for assessing the uncertainty surrounding point forecasts
and for quantifying the probability of extreme price movements both in the
short- and medium-run. Next, we examine the role of time-varying volatility in
improving forecast accuracy (Clark and Ravazzolo,, 2015; Chan,, 2023) and
assess the presence of forecast instabilities (Rossi,, 2021).
We show that a simple VAR model, augmented with either one or two factors
capturing key predictors of the price of carbon, provides substantial accuracy
gains over a wide set of benchmark forecasts, including survey expectations
and forecasts made available by data providers. We extend the forecasting
study to verified emissions and demonstrate that, in this case, SV can further
improve the performance of a single-factor VAR model. Lastly, we show how
model-based forecasts can be used to build market monitoring tools that track
demand and price pressure in the EU ETS market.
The rest of the paper is organized as follows. Section 2 offers an overview of
the EU ETS; data and details of the forecasting exercise are illustrated in
Section 3. Results for real price and verified emissions are presented in
Sections 4 and 5, respectively. Section 6 concludes.
## 2 The EU Emissions Trading System
The EU ETS is a cap and trade system that started in 2005 intending to reduce
carbon emissions. In this system, the maximum quantity of emissions (the cap)
is set through unit permits (European Unit Allowances, EUA), which allow the
owner of the permit to produce 1 ton of CO2 or an equivalent quantity of other
greenhouse gases. The European Commission sets a yearly cap on the total
greenhouse gas emissions that can be produced by actors participating in the
system. Since the aim is to decrease emissions over time, every year the cap
is lower than the year before, and consequently, the maximum allocation of EUA
is reduced.
The EU ETS system is a financial market where actors can acquire EUA on the
primary market through an auctioning system, and trade derivatives on the
secondary market. A certain amount of permits is originally granted for free
each year according to the needs of specific sector emissions, although the
remaining amount of available allowances is allocated on the primary market
through uniform price auctions with single rounds and sealed bids, conducted
daily by the European Energy Exchange (EEX). Since EUA have been classified as
financial instruments, the associated derivatives - such as spot, futures,
options, and forward contracts - can be traded on secondary markets, both on
exchange and over the counter. While auctions take place on the EEX, trading
takes place also on the Intercontinental Exchange (ICE).
EUA are handed out to the market through a system of benchmark-based
allocation or auctions. If emissions at the end of the year result to be lower
than the cap set for each installation participating in the market, permits
can be traded among actors for an economic value to be determined on the
secondary market. In case emissions exceed the threshold, sanctions are
applied to economic agents participating in the market.
Actors participating in the EU ETS market entail industries belonging to high
emissions sectors: electricity and heat generation, energy-intensive industry
sectors (including oil refineries, steel works, and production of iron,
aluminum, metals, paper, etc.), aviation within the European Economic Area
and, starting from 2024, maritime transport. Participation in the EU ETS is
mandatory for companies in the covered sectors, however, for some of the
sectors only production plants bigger than a threshold size are included.
Historically, the EU ETS evolution went through four phases. To meet the
objectives set by the European Commission in terms of emissions reduction,
each phase aims to reduce the number of EUA granted to each participating
sector. The cap can be lowered by setting a decreasing number of allowances to
be allocated each year or by establishing a yearly linear reduction factor,
e.g. a linear reduction of 1.74% and 2.2% of the baseline 2008-2012 emissions
have been set respectively from 2013 and 2021 onwards, with no end date,
resulting in a year-on-year reduction by up to 43 million allowances
(International Carbon Action Partnership,, 2023).
The pilot phase (2005-2007) aimed to verify rules, regulations, emission
detection systems, as well as the regulatory framework. In this phase, the
allocation system was grandfathering: all EUA were allocated freely to
industries, up to the cap set for each regulated sector. The second phase,
which lasted from 2008 to 2012, was characterized by the introduction of the
allocation of permits employing an auctioning system. In this phase, roughly 2
up to 5% of the total permits were allocated through auctions. This share
increased to reach 54$\%$ in the third phase, which lasted from 2013 to 2020,
and it includes more sectors and gases. Lastly, the fourth phase (2021-2030)
has the aim of reducing net emissions by at least 55% by 2030 compared to
1990, as set in the European Climate Law, by further lowering the cap and
targeting the carbon leakage phenomenon.
In July 2021, the European Commission adopted a series of legislative
proposals regarding EU ETS aimed at increasing the pace of emissions cuts.
These include, among others, covering more sectors and gases, gradually
lowering the number of emission allowances each year, and reinforcing the
Market Stability Reserve (MSR), which aims at reducing the surplus of
allowances in the market. An excessive allowances surplus would lead to lower
carbon prices, rendering, therefore, the ETS system less effective, by
decreasing incentives of the economic actors participating in the market to
lower emissions. The MSR is automatically applied when the total number of
allowances on the market exceeds a certain threshold. In Phase IV, the free
EUA allocation system was granted a ten-year extension, and specific measures
have been taken for sectors exposed to a higher risk of carbon leaking.
Figure 1: Allocated and verified emissions for all stationary installations in
EU-27 countries (excluding the aviation sector). Notes: authors’ elaboration
of data from https://www.eea.europa.eu.
The way allowance allocation has evolved throughout the years is described in
Figure 1. In Phase II, auctions were introduced, but between 95% and 97% of
allowances were still distributed for free. This changed completely during
Phase III, where auctions became the predominant allocation method for most
sectors, covering between 40% and 55% of the total allocated permits. Phase
III also needed to deal with the aftermath of an excessive surplus of unused
EUA in Phase II. The number of allocated EUA decreased in 2014, increasing the
price of allowances and a reduction in the number of verified emissions and
unused EUA (Bjørnland et al.,, 2023). In Phase IV, auctions remain the main
allocation method on the primary market.
## 3 Data and methods
Following Bjørnland et al., (2023), we focus on forecasting the end-of-month
price of the one-month ahead futures contract traded on ICE that represents
the most closely watched series by practitioners. Moreover, we deflate the
nominal futures price by using the Euro area harmonized index of consumer
prices. The left panel of Figure 2 shows the real price series jointly with
shaded grey areas representing the recessions in the Euro Area.
As for the predictors, we follow the approach of Boivin and Ng, (2006) and
Baumeister et al., (2022). Therefore, instead of collecting a large number of
series, we carefully select 21 predictors that capture demand and supply-side
forces driving the price of carbon. More precisely, we concentrate on
variables within the following categories:
* •
Economic activity (8 series): we collect data on aggregate industrial
production (IP) for the EU-19 area, as well as indices for sectors covered by
the EU ETS (i.e. electricity, gas, steam, and air conditioning supply, basic
metals, manufacture of paper and its products, coke and refined petroleum
products, chemical products, non-metallic mineral products) from Eurostat.
Moreover, we consider the Euro Stoxx 50 stock price index, sourced from
Refinitiv Eikon.
* •
Energy prices (7 series): we consider the prices of Brent crude oil, TTF
natural gas (front-month and front-year), ARA API-2 coal (front-year), German
power price (front-year), clean dark, and clean spark spreads (front month).
These variables are sourced from Refinitiv Eikon.
* •
Technical indicators (3 series): we select some of the variables that
practitioners use to track the functioning of the EU ETS market (Marcu et
al.,, 2023). The auction coverage ratio, defined as the total number of bids
in an auction divided by the number of available EUA, is a proxy for the
actual auction demand relative to supply on the primary market. As a rule of
thumb, a value greater (lower) than two indicates a high (low) auction demand
relative to supply. The auction clearing price, and a volatility proxy based
on the monthly auction price range. These variables are sourced from the EUA
Primary Market Auction Report maintained by EEX.
* •
Weather conditions (2 series): temperature and precipitation anomalies for
EU-19 countries are constructed as differences from long-term moving averages
using data sourced from the Weather for Energy Tracker maintained by the
International Energy Agency (IEA).
Figure 2: Real EU ETS price (left), actual and interpolated verified emissions
for EU-19 countries (right) from March 2005 to September 2023.
Notes: Interpolated emissions are obtained using IP indices of six sectors
covered by the EU ETS. Shaded areas represent recessions in the Euro Area, as
determined by the CEPR-EABCN (https://eabcn.org).
Moreover, we also collect information on verified emissions for all stationary
installations in six sectors covered by the EU ETS222We do not include the
aviation sector because it is covered by the EU ETS since 2012, while we
interpolate data on verified emissions using IP indices starting from 2006. in
the EU-19 countries. These data, available on an annual basis, refer to the
actual amount of greenhouse gas emissions produced by a company or entity, as
reported in its emissions report and verified by an accredited verifier by 31
March of the following year.333Based on verified emissions, companies must
surrender a corresponding number of emission allowances by the end of April of
that year. If a company’s verified emissions exceed the number of allowances
it holds, it may need to purchase additional allowances from the market to
cover the excess emissions. Following Bjørnland et al., (2023) and Känzig,
(2023), we temporally disaggregate annual data with the Chow and Lin, (1971)
method. Annual and interpolated monthly verified emissions are shown in the
right panel of Figure 2.444More precisely, we consider the IP indices and
emissions for all sectors covered by the EU ETS (except aviation) and
construct an emission-weighted IP index to interpolate annual verified
emissions. The MATLAB library provided by Quilis, (2013) is used to implement
the Chow-Lin method. Since verified emissions for 2023 were not available, we
used the 2022 value to interpolate data for 2023.
The forecasting exercise is based on data spanning from June 2012 to September
2023, comprising a total of 136 monthly observations. The start date of the
sample is dictated by the availability of data on auctions, which are
particularly relevant for explaining the allocation mechanism of emission
allowances during Phases III and IV of the EU ETS (see Figure 1). Forecasts
are generated using an expanding window approach: each time a new forecast is
produced, the estimation sample is updated by adding a new observation. The
first estimation sample ends in December 2017, and the last forecast is issued
in September 2022. We consider forecast horizons of one month up to one year
ahead. The forecast evaluation sample is the same for all forecast horizons
and consists of 58 observations spanning from December 2018 to September 2023.
We denote the level of the real price of carbon in month $t$ as $R_{t}$ and
the log price as $r_{t}=\log R_{t}$. Models are estimated using the first
difference of the logarithm of the real price of carbon, $\Delta r_{t}$, and
forecasts are constructed iteratively from the estimated models and converted
into levels as follows:
$\displaystyle\hat{R}_{t+h|t}$
$\displaystyle=\exp\left(r_{t}+\sum_{\ell=1}^{h}\Delta\hat{r}_{t+\ell|t}\right),$
where $\Delta\hat{r}_{t+\ell|t}$ is the $\ell$-step ahead forecasted value.
The evaluation of point forecasts relies on the relative Root Mean Squared
Forecast Error (RMSFE) that represents the ratio of the RMSFE of a model to
the RMSFE of the benchmark, such as the Random Walk (RW). Therefore, a
relative RMSFE lower than unity is taken as evidence that a certain model is
more accurate than the benchmark.
Sign, or direction-of-change forecasts, are defined as:
$\text{sign}\left(\hat{R}_{t+h|t}-R_{t}\right)$, where $\mathrm{sign}(x)$
equals -1 if $x<0$, 0 if $x=0$, and 1 if $x>0$. We use the Success Ratio (SR),
defined as the proportion of correctly predicted signs, to gauge the
directional accuracy. A SR greater than 0.5 indicates a gain in accuracy
relative to the RW model that implies a no-change forecast.
As a further measure of forecasting, we rely on the quantile-based continuous
ranked probability score (qCRPS) of Gneiting and Ranjan, (2011), which is a
density forecasting measure denoted by
$\widehat{QS}_{t}=\frac{1}{J-1}\sum_{j=1}^{J-1}\widehat{QS}_{t}^{\alpha_{j}}=\frac{1}{J-1}\sum_{j=1}^{J-1}2\left[\mathbb{I}\left(R_{t+h}\leq\hat{q}^{\alpha_{j}}_{t+h|t}\right)-\alpha_{j}\right]\times\left(\hat{q}^{\alpha_{j}}_{t+h|t}-R_{t+h}\right),$
(1)
where $\mathbb{I}(\cdot)$ denotes the indicator function and
$\hat{q}^{\alpha_{j}}_{t+h|t}$ is the $h$-step ahead quantile forecast for
$R_{t+h}$ at level $\alpha_{j}=j/J$ with $J=20$, which corresponds to
$\alpha_{j}=0.05,0.10,\ldots,0.95$. We construct weighted versions of the
qCRPS, where the weights are selected to emphasize specific regions, such as
the center or one of the tails of the distribution:
$\displaystyle\widehat{wQS}_{t}$
$\displaystyle=\frac{1}{J-1}\sum_{j=1}^{J-1}\nu(\alpha_{j})\widehat{QS}_{t}^{\alpha_{j}},\quad\text{with
}\nu(\alpha_{j})=\begin{cases}\alpha_{j}(1-\alpha_{j})&\text{(center)},\\\
\alpha_{j}^{2}&\text{(right tail)},\\\ (1-\alpha_{j})^{2}&\text{(left
tail)}.\end{cases}$ (2)
In a pairwise comparison, the model with the lowest score is ranked as the
most accurate.
## 4 Results
### 4.1 Univariate time series models
Most previous papers rely on daily or weekly nominal data, therefore in the
case of monthly real carbon prices, it is not clear which benchmark to use. In
the literature concerning the forecasting of commodity prices, particularly
for crude oil, the simple Random Walk (RW) model, or no-change forecast, is
commonly used as a benchmark. In this initial phase of the analysis, we focus
on point and sign forecasts, while we will consider density forecasts at a
later stage. Results are summarized in Table 1, which shows the relative RMFSE
and the SR of different models for selected forecast horizons of
$h=1,2,3,6,9,12$ months ahead.555In the Supplement, we display results for all
forecast horizons from 1 to 12 months ahead and for additional univariate time
series models.
In the short run, including a drift in the RW model (RWD) provides modest
gains in point forecast accuracy, while for forecasts ranging from one-quarter
up to one year ahead, accuracy is comparable to that of the RW model. Coupled
with an SR below 0.5 at the 12-month horizon, this implies that the RWD model
does not qualify as a more accurate benchmark for point and sign forecasting.
If Autoregressive (AR) and/or Moving Average (MA) components are included,
models perform poorly at horizons 1 and 2, but they lead to RMSFE reductions
at horizons from one quarter up to one year that can be as large as 16.98%.
See columns 3 and 4 of Table 1. Looking at the success ratio (panel b), ARIMA
models have some directional accuracy, especially at longer horizons.
Table 1: Relative RMSFE and Success Ratio of univariate time series models (a)
Relative RMSFE
---
| | First-difference, $\Delta r_{t}$
h | RWD | ARIMA(1,1,1) | ARIMA(0,1,1) | BAR(1) | BAR(3) | BAR(12) | BAR(aic)
1 | 0.993 | 1.069 | 1.071 | 1.043 | 1.050 | 1.046 | 1.047
2 | 0.999 | 1.010 | 1.002 | 1.003 | 1.003 | 1.013 | 1.004
3 | 1.0000 | 0.998 | 0.993 | 0.990 | 0.988 | 0.998 | 0.992
6 | 1.0000 | 0.905 | 0.901 | 0.903 | 0.903 | 0.923 | 0.911
9 | 1.0000 | 0.856 | 0.854 | 0.855 | 0.858 | 0.864 | 0.858
12 | 1.0000 | 0.830 | 0.831 | 0.830 | 0.834 | 0.848 | 0.834
(b) Success Ratio
| | First-difference, $\Delta r_{t}$
h | RWD | ARIMA(1,1,1) | ARIMA(0,1,1) | BAR(1) | BAR(3) | BAR(12) | BAR(aic)
1 | 0.603 | 0.535 | 0.535 | 0.517 | 0.535 | 0.517 | 0.517
2 | 0.621 | 0.517 | 0.517 | 0.517 | 0.517 | 0.621 | 0.517
3 | 0.638 | 0.535 | 0.569 | 0.603 | 0.621 | 0.638 | 0.603
6 | 0.776 | 0.759 | 0.759 | 0.776 | 0.776 | 0.776 | 0.776
9 | 0.569 | 0.828 | 0.828 | 0.828 | 0.828 | 0.828 | 0.828
12 | 0.431 | 0.897 | 0.897 | 0.897 | 0.897 | 0.897 | 0.897
Notes: Panel (a) shows the ratio of RMSFE of model $m$ over the RW model.
Values below one suggest superior forecast performance of model $m$ to the RW
(in bold). ∗ denotes that the null hypothesis of the Diebold-Mariano test is
rejected at the 90% (95%) confidence level. Panel (b) reports the success
ratios; entries in bold suggest that the model can accurately predict the
direction of change over 50% of the time. ∗ indicates that the p-value of the
Pesaran and Timmermann, (2009) test of the null of no directional accuracy is
below 0.1, hence providing evidence of statistical accuracy at the 10%
significance level. Models yielding the lowest RMSFE or highest SR are
underlined.
We conclude with Autoregressive, AR($p$), models specified for the log first
difference estimated with Bayesian natural conjugate prior and denoted as BAR
(columns 5-8). We consider fixed lag orders $p$ = 1, 3, and 12, as well as lag
order selection based on the Akaike Information Criterion (AIC). In this case,
we set the maximum lag order to 12 and select $p$ each time a new forecast is
issued. Indeed, BAR models for $\Delta r_{t}$ display directional accuracy and
lead to RMSFE reductions that reach 17% at horizon 12 in the case of a simple
BAR(1). Moreover, it is worth pointing out that fixed and small lag orders are
preferable to either setting $p=12$ or selecting $p$ recursively with the AIC.
To sum up, given that at horizons 1 and 2, not even BAR models for $\Delta
r_{t}$ outperform the RW, we cannot definitively discard it as a benchmark.
In Table 1, we have used the Diebold and Mariano, (1995) test, as modified by
Coroneo and Iacone, (2020), to verify the statistical significance of RMSFE
reductions, and the Pesaran and Timmermann, (2009) test to assess the
statistical significance of SR. Possibly due to the small size of the
evaluation sample or because of instabilities in forecasting performance, we
are never able to provide evidence of statistically significant improvements
over the RW. We address the presence of forecast instabilities in Sections 4.4
and 4.5.
### 4.2 VAR models of the EU ETS carbon market
To link the real price of carbon to its determinants, a natural starting point
is a small-scale VAR(p) model of the EU ETS market:
$\mathbf{y}_{t}=\mathbf{a}+\sum_{j=1}^{p}\mathbf{A}_{j}\mathbf{y}_{t-j}+\mathbf{u}_{t},$
(3)
where $\mathbf{a}$ is a $n\times 1$ vector of intercepts, $\mathbf{A}_{j}$ are
$n\times n$ matrices of coefficients for $j=1,\ldots,p$, and $\mathbf{u}_{t}$
is an $n\times 1$ vector of zero-mean innovations Normally distributed with
covariance matrix $\boldsymbol{\Sigma}$. Following Bjørnland et al., (2023),
we consider a baseline VAR specification in which $\mathbf{y}_{t}$ is a
$3\times 1$ vector including $\Delta r_{t}$ (or alternatively, $r_{t}$), the
first difference of the logarithm of interpolated verified emissions,
$emis_{t}$, and first difference of the logarithm of aggregate industrial
production for EU-19 countries, $\Delta ip_{t}$.
The baseline model does not include many predictors tracked by practitioners,
such as energy prices, technical indicators related to the auctioning of EU
ETS allowances, and weather anomalies that might affect the demand for
electricity and gas. Differently from Bjørnland et al., (2023), we consider
also the Euro Stoxx stock price index as an additional proxy of real economic
activity, as well as a wider set of IP indices. In particular, we add IP
indices for the main sectors covered by the EU ETS to better capture demand-
side pressures affecting the real price of carbon.
Given the relatively small size of our sample of data, including the set of 21
predictors described in Section 3 in the model would not be feasible.
Similarly, we want to avoid sparse model representations implied by the
selection of only a handful of the potentially relevant predictors. Thus, we
consider a Factor-augmented VAR model (FAVAR), where we estimate a modified
version of the baseline VAR specification, in which we replace the aggregate
IP index for EU-19 with up to three factors extracted from the set of 21
predictors.
In detail, we pool the information from the 21 predictors based on principal
component analysis.666Before the analysis, all variables have been transformed
to induce stationarity and then standardized. IP indices, the Euro Stoxx price
index, energy prices, and the auction price are transformed into monthly
growth rates (first differences of logarithms). Clean dark and clean spark
spreads are first-differenced, while the monthly auction price range is log-
transformed. We do not apply any transformation to the auction cover ratio,
temperature, and precipitation anomalies. A preliminary screening for outliers
was also carried out. However, given that only a few extreme observations
(i.e. observations exceeding 20 times the interquartile range from the median)
were detected during the COVID pandemic, we decided to keep them in the
sample. We decided to focus on the first three factors since they account for
48% of the variance of the 21-time series. The percentages of total variance
explained by the first, second, and third factors are 22%, 17.8%, and 8.6%,
respectively.
Figure 3: $R^{2}$ between factors and individual predictors obtained from
regressing a factor on an individual predictor using data from June 2012 to
December 2017.
Figure 3 presents the $R^{2}$ from regressing a factor on an individual
predictor using data from June 2012 to December 2017. Broadly speaking, the
first factor loads mostly on energy prices. The second-factor loads on a
combination of a few IP indices and the cover ratio, while the third loads
heavily on the two spreads, precipitation anomalies, and the remaining IP
indices.
As shown in the coming sections, the forecasting exercise will show that a
BVAR(1) model augmented with one factor predicts both the real price of carbon
and verified emissions adequately and better than alternative models.
Therefore, it is useful to characterize the variables underlying the dynamics
of the first factor. Figure 4 indicates that the first factor tracks the
European business cycle as well as the evolution of energy prices and is
highly correlated with both IP growth and returns of TTF natural gas. Indeed,
during the COVID-19 recession, the first factor is negative and well below its
mean, while it becomes positive in the recovery phase and for most of the
first semester of 2022 when natural gas prices rise due to the Russian
invasion of Ukraine.
Figure 4: Correlation between Factor 1 and industrial production (top) and
growth rate of the price of TTF natural gas (bottom) from December 2017 to
September 2023. Notes: we depict centered 3-month moving average of Factor 1
and predictors (scaled to have same variance as Factor 1).
Figure 5 displays the evolution of the first factor over time, along with bars
representing the contribution of different predictors grouped by class (i.e.
economic activity, energy prices, technical indicators, and weather
anomalies). Interestingly, the drivers underlying the movements of the first
factor change over time, but are always mostly related to economic activity
proxies and energy prices. Measures of real economic activity account for the
largest share of the downward movement of the factor in 2020 when European
economies were frozen during lockdowns. The upward movement of the factor in
2022 and its successive decline in 2023 are attributed to the pressures
related to the war in Ukraine and the subsequent easing of conditions in
energy markets.
Figure 5: Contribution of predictors (grouped by class) to Factor 1,
represented as bars, along with the evolution of Factor 1 (green line) from
December 2017 to September 2023. Notes: we depict the centered 3-month moving
average of both Factor 1 and the predictors’ contributions to it.
### 4.3 Can VAR models forecast the real price of carbon?
In this section, we focus on Bayesian techniques (BVAR or BFAVAR) with
Minnesota prior to forecast the real price of carbon. Overall, both for
baseline BVAR models and factor-augmented specifications, setting the lag
order to 1 or 3 yields the most accurate point and direction-of-change
forecasts. Indeed, inspecting the leftmost column of Table 2, the baseline
BVAR model with $p=12$ – as set by Bjørnland et al., (2023) – is regularly
outperformed by specifications with $p=1$ and $p=3$.777In the Supplement, we
have estimated the same models with the real price of carbon log-levels, and
point forecasts were always less accurate than growth rates estimation. As for
the lag length of BVAR models, we consider pre-specified values of $p=1,3,12$
along with a data-driven approach that sets $p$ by minimizing the AIC at each
forecast origin, selecting among BVAR($p$) models up to order 12.
Table 2 shows that at horizons up to 2 months, BVAR and BFAVAR models are as
accurate as no-change point forecasts. However, as the forecast horizon grows,
BVAR and BFAVAR models lead to sizable accuracy gains both for point and sign
forecasts. For instance, at a one-year horizon, the BVAR(1) and factor-
augmented BVAR(1) model result in a reduction in RMSFE in the range 17.8-18.2%
and can accurately predict the direction of price movements 89.7% of the time.
Table 2: Relative RMSFE and Success Ratio for Bayesian Vector Autoregressive
and Factor models. (a) Relative RMSFE
---
| | 1 Factor | 2 Factors | 3 Factors
h | BVAR(1) | BVAR(3) | BVAR(12) | BVAR(1) | BVAR(3) | BVAR(1) | BVAR(3) | BVAR(1) | BVAR(3)
1 | 1.057 | 1.075 | 1.074 | 1.062 | 1.073 | 1.039 | 1.077 | 1.062 | 1.114
2 | 1.006 | 1.019 | 1.028 | 1.010 | 1.028 | 1.024 | 1.102 | 1.066 | 1.154
3 | 0.995 | 0.996 | 1.002 | 1.002 | 1.019 | 0.999 | 1.052 | 1.015 | 1.062
4 | 0.953 | 0.952 | 0.970 | 0.947 | 0.953 | 0.963 | 0.960 | 0.976 | 0.975
5 | 0.965 | 0.964 | 0.984 | 0.957 | 0.960 | 0.970 | 0.971 | 0.985 | 0.983
6 | 0.915 | 0.916 | 0.937 | 0.906 | 0.916 | 0.921 | 0.919 | 0.935 | 0.927
7 | 0.897 | 0.899 | 0.917 | 0.886 | 0.899 | 0.902 | 0.894 | 0.916 | 0.899
8 | 0.844 | 0.852 | 0.858 | 0.831 | 0.847 | 0.850 | 0.856 | 0.867 | 0.863
9 | 0.857 | 0.865 | 0.873 | 0.851 | 0.872 | 0.856 | 0.901 | 0.863 | 0.894
10 | 0.825 | 0.836 | 0.838 | 0.824 | 0.850 | 0.830 | 0.887 | 0.833 | 0.881
11 | 0.822 | 0.837 | 0.843 | 0.824 | 0.849 | 0.828 | 0.886 | 0.831 | 0.883
12 | 0.818 | 0.834 | 0.848 | 0.822 | 0.849 | 0.828 | 0.892 | 0.834 | 0.891
(b) Success Ratio
| | 1 Factor | 2 Factors | 3 Factors
h | BVAR(1) | BVAR(3) | BVAR(12) | BVAR(1) | BVAR(3) | BVAR(1) | BVAR(3) | BVAR(1) | BVAR(3)
1 | 0.552 | 0.569 | 0.517 | 0.552 | 0.500 | 0.569 | 0.569 | 0.552 | 0.569
2 | 0.535 | 0.552 | 0.569 | 0.535 | 0.569 | 0.552 | 0.569 | 0.552 | 0.535
3 | 0.603 | 0.621 | 0.621 | 0.621 | 0.603 | 0.621 | 0.655∗ | 0.603 | 0.603
4 | 0.707 | 0.672 | 0.672 | 0.690 | 0.655 | 0.707 | 0.690 | 0.690 | 0.672
5 | 0.672 | 0.672 | 0.655 | 0.672 | 0.638 | 0.672 | 0.655 | 0.672 | 0.672
6 | 0.776 | 0.759 | 0.759 | 0.776 | 0.741 | 0.776 | 0.724 | 0.776 | 0.759
7 | 0.793 | 0.776 | 0.793 | 0.793 | 0.759 | 0.793 | 0.759 | 0.793 | 0.776
8 | 0.845 | 0.828 | 0.845 | 0.845 | 0.810 | 0.845 | 0.810 | 0.845 | 0.828
9 | 0.828 | 0.810 | 0.828 | 0.828 | 0.793 | 0.828 | 0.793 | 0.828 | 0.810
10 | 0.862 | 0.828 | 0.862 | 0.862 | 0.828 | 0.862 | 0.810 | 0.862 | 0.828
11 | 0.914 | 0.879 | 0.914 | 0.914 | 0.879 | 0.914 | 0.862 | 0.914 | 0.874
12 | 0.897 | 0.879 | 0.897 | 0.897 | 0.862 | 0.897 | 0.845 | 0.897 | 0.862
Notes: see notes to Table 1
At intermediate forecast horizons (i.e., 4 up to 10 months ahead), the BVAR(1)
model augmented with a single factor regularly yields the most accurate point
forecasts, and its performance tends to improve as the horizon grows.
Increasing the number of factors above one generally leads to less accurate
point forecasts. In particular, adding the third factor does not seem to
provide any advantage to models with one or two factors. The model with two
factors, on the other hand, seems to be slightly better than the single-factor
specification at shorter forecast horizons.
### 4.4 Are there instabilities in forecasting performance?
There is widespread evidence that the relative forecasting performance of
models changes over time due to parameter instability, shocks with time-
varying volatilities, and changes in the variance of the predictors (Giacomini
and Rossi,, 2010; Rossi,, 2021). In such cases, averaging the difference in
forecasting performance over the full evaluation sample, as we did in the
previous sections, results in a loss of information that might lead to
standard tests of predictive ability to conclude that two competing models are
equally accurate.
In Figure 6, we investigate the existence of instabilities in forecasting
performance. We do so by implementing the fluctuation test statistic,
$\mathcal{F}_{t}^{OOS}$, of Giacomini and Rossi, (2010) considering a centered
moving average over a $19$-month window.888The test has a nonstandard
distribution, and the critical values provided by Giacomini and Rossi, (2010)
depend on the ratio between the size of the window used to compute the moving
average and the number of observations in the evaluation sample. In the
Supplement, we show that the results are robust to changes in the size of the
window. The test is based on the (standardized) difference between the MSFE of
a benchmark model and of the factor-augmented BVAR(1) model with one or two
factors. Positive values of the test statistic indicate that the BFAVAR model
has a lower MSFE than the benchmark. All tests are one-sided: the null
hypothesis is the factor VAR model has the same MSFE as the benchmark, while
the alternative is that the former is more accurate than the latter. The
dashed line indicates the 5% critical value, $CV_{0.05}$, and the null
hypothesis is rejected when $\max\mathcal{F}_{t}^{OOS}>CV_{0.05}$.
In our comparison, we focus on one month, one quarter, and one year ahead
forecast horizons. To raise the bar of forecast evaluation, we do not solely
focus on the RW model. Instead, we compare the performance of factor models
against the BAR(1) and the BVAR(1) models, which, as shown in previous
sections, appear to be more competitive benchmarks than the RW model.
Figure 6: Fluctuation test statistic for a BFAVAR with 1 factor (left) and 2
factors (right) against different benchmarks for forecast horizons $h=1,3,12$
months from September 2019 to December 2022. Notes: the fluctuation test
statistic, $\mathcal{F}_{t}^{OOS}$, of Giacomini and Rossi, (2010), is
calculated with a 19-month centered rolling window. Positive values indicate
that the BFAVAR model is better than the benchmark. All tests are one-sided,
with the null hypothesis being that the BFAVAR(1) model has the same MSFE as
the benchmark; the alternative is that the BFAVAR(1) model forecasts better
than the benchmark. The dashed line indicates the one-sided 5% critical value,
$CV_{0.05}$. The null hypothesis is rejected when
$\max\mathcal{F}_{t}^{OOS}>CV_{0.05}$.
Several interesting results emerge from Figure 6. First, regardless of the
benchmark model, the relative forecasting performance of BVAR specifications –
augmented with either one or two factors – changes over time and tends to
deteriorate at the end of 2022. Indeed, especially at shorter forecast
horizons, the test statistic often becomes negative in this period. Second,
the model with one factor (left panel) is usually preferable to the model with
two factors (right panel), in that the latter rarely beats the benchmarks.
Third, at a forecast horizon of one month, the single-factor model performs
better than the RW model until the beginning of 2021; while it never beats the
BAR(1). However, at longer forecast horizons and especially for $h=12$, the
single-factor model performs better than the benchmarks during most of the
evaluation sample. Moreover, for $h=12$, the fluctuation test leads to a
rejection of the null hypothesis of indistinguishable forecasting performance;
in all cases, the test statistic lies above the 5% critical value in late 2020
and in part of 2021.
Thus the single-factor BVAR(1) model, condensing information of a broad set of
predictors related to the EU ETS, is a promising specification for forecasting
the magnitude as well as the direction-of-change of the real price of carbon
at longer horizons. At horizons shorter than a quarter, VAR models do not
offer a real advantage over simple univariate specifications, especially in
point forecasting.
### 4.5 Density forecasts and the role of stochastic volatility
To gauge the uncertainty associated with point forecasts and following
evidence that modeling SV improves density forecasts of macroeconomic
aggregates (Clark and Ravazzolo,, 2015; Chan,, 2023), we evaluate density
forecasting for future values of the real price of carbon.
We also decided to include in BVAR models a Choleski multivariate SV process
(Chan,, 2023, for details):
$\displaystyle\mathbf{u}_{t}$ $\displaystyle\sim
N(\mathbf{0},\boldsymbol{\Sigma}_{t}),\quad\boldsymbol{\Sigma}_{t}^{-1}=\mathbf{B}_{0}^{\prime}\mathbf{D}_{t}^{-1}\mathbf{B}_{0},$
(4) $\displaystyle h_{i,t}$
$\displaystyle=\mu_{i}+\phi_{i}\left(h_{i,t-1}-\mu_{i}\right)+\varepsilon_{i,t},\quad\varepsilon_{i,t}\sim
N(0,\sigma^{2}_{\varepsilon}),$ (5)
where $\mathbf{B}_{0}$ is a lower triangular matrix of size $n\times n$ with
ones along the main diagonal and
$\mathbf{D}_{t}=\text{diag}\left(e^{h_{1,t}},\cdots,e^{h_{n,t}}\right)$, and
Equation (5) specifies an independent autoregressive volatility process for
each variable in the model for $t=2,\ldots,T$ with initial condition
$h_{i,1}\sim N(\mu_{i},\sigma^{2}_{i}/(1-\phi_{i}^{2}))$.
To evaluate the density performance, we rely on qCRPS (Gneiting and Ranjan,,
2011), where the model with the lowest qCRPS is ranked as the most
accurate.999Note that, contrary to the previous tables, these numbers are not
ratios to a benchmark but are expressed in the same scale as real prices. In
Panel (a) of Table 3, we focus on the center of the distribution, while the
accuracy in forecasting the right and left tails of the distribution, which
are of interest to assess the probability of extreme price movements, is
evaluated in Panel (b) and (c), respectively.
Table 3: Quantile-weighted Continuous Ranked Probability Score (qCRPS). (a)
Quantile-weighted CRPS: center
---
| | 1 Factor | 2 Factors | | 1 Factor | 2 Factors
h | BVAR(1) | BVAR(1) | BVAR(1) | BVAR-SV(1) | BVAR-SV(1) | BVAR-SV(1)
1 | 0.729 | 0.721 | 0.692 | 0.712 | 0.714 | 0.704
2 | 0.978 | 0.966 | 0.968 | 0.977 | 0.974 | 0.983
3 | 1.231 | 1.226 | 1.200 | 1.257 | 1.252 | 1.253
4 | 1.322 | 1.310 | 1.318 | 1.349 | 1.341 | 1.355
5 | 1.625 | 1.611 | 1.623 | 1.648 | 1.639 | 1.659
6 | 1.670 | 1.649 | 1.663 | 1.701 | 1.689 | 1.712
7 | 1.790 | 1.761 | 1.782 | 1.804 | 1.793 | 1.819
8 | 1.860 | 1.831 | 1.861 | 1.886 | 1.877 | 1.905
9 | 2.121 | 2.102 | 2.109 | 2.144 | 2.139 | 2.161
10 | 2.180 | 2.169 | 2.185 | 2.245 | 2.248 | 2.277
11 | 2.325 | 2.325 | 2.336 | 2.407 | 2.404 | 2.432
12 | 2.454 | 2.472 | 2.480 | 2.587 | 2.585 | 2.612
(b) Quantile-weighted CRPS: right tail
| | 1 Factor | 2 Factors | | 1 Factor | 2 Factors
h | BVAR(1) | BVAR(1) | BVAR(1) | BVAR-SV(1) | BVAR-SV(1) | BVAR-SV(1)
1 | 1.286 | 1.217 | 1.124 | 1.237 | 1.220 | 1.198
2 | 1.733 | 1.676 | 1.663 | 1.681 | 1.664 | 1.678
3 | 2.207 | 2.131 | 2.088 | 2.166 | 2.139 | 2.170
4 | 2.392 | 2.315 | 2.327 | 2.358 | 2.332 | 2.389
5 | 2.915 | 2.835 | 2.847 | 2.845 | 2.815 | 2.880
6 | 3.111 | 3.042 | 3.064 | 3.036 | 3.015 | 3.084
7 | 3.324 | 3.219 | 3.268 | 3.235 | 3.208 | 3.299
8 | 3.515 | 3.394 | 3.452 | 3.462 | 3.437 | 3.544
9 | 3.984 | 3.890 | 3.917 | 3.914 | 3.900 | 4.008
10 | 4.304 | 4.230 | 4.264 | 4.302 | 4.293 | 4.417
11 | 4.630 | 4.576 | 4.596 | 4.638 | 4.634 | 4.757
12 | 5.015 | 4.9996 | 5.008 | 5.090 | 5.098 | 5.222
(b) Quantile-weighted CRPS: left tail
| | 1 Factor | 2 Factors | | 1 Factor | 2 Factors
h | BVAR(1) | BVAR(1) | BVAR(1) | BVAR-SV(1) | BVAR-SV(1) | BVAR-SV(1)
1 | 1.270 | 1.282 | 1.235 | 1.278 | 1.288 | 1.268
2 | 1.599 | 1.605 | 1.588 | 1.677 | 1.671 | 1.680
3 | 1.897 | 1.914 | 1.834 | 2.046 | 2.026 | 2.001
4 | 1.943 | 1.958 | 1.953 | 2.116 | 2.091 | 2.097
5 | 2.404 | 2.412 | 2.410 | 2.606 | 2.572 | 2.598
6 | 2.303 | 2.292 | 2.306 | 2.506 | 2.464 | 2.491
7 | 2.471 | 2.467 | 2.471 | 2.642 | 2.609 | 2.629
8 | 2.534 | 2.506 | 2.562 | 2.689 | 2.656 | 2.686
9 | 2.901 | 2.888 | 2.863 | 3.082 | 3.052 | 3.048
10 | 2.806 | 2.803 | 2.811 | 3.021 | 2.998 | 3.001
11 | 2.946 | 2.960 | 2.963 | 3.179 | 3.155 | 3.165
12 | 3.007 | 3.029 | 3.053 | 3.289 | 3.266 | 3.282
Notes: The best forecasts, associated with the lowest scores, are underlined.
Figure 7: Fluctuation test statistic for the right tail for BFAVAR with 1
factor (left) and 2 factors (right). September 2019 to December 2022.
With very few exceptions, homoskedastic factor-augmented BVAR models – with
either one or two factors – are more accurate at forecasting the center and
both tails than the alternative specifications. In the case of the real price
of carbon, incorporating SV does not seem to offer any sizable advantage on
any horizon. Interestingly, the homoskedastic BVAR(1) model augmented with two
factors is more accurate than any other specifications at horizons up to a
quarter-ahead. This result stands in contrast to what is observed for point
and sign forecasts, for which factor-augmented models do not yield accuracy
gains at shorter horizons. Moreover, it emphasizes that, for short-term
density forecasts, the second factor, which captures predictors beyond energy
prices and economic activity that are the main drivers of the first factor
mostly, might be relevant.101010See the Supplement for results concerning the
inclusion of SV in BVAR models.
We complement these results with evidence of instabilities in the (right) tail
forecasting. When evaluating the single-factor model, against the BAR(1) and
BVAR(1) models, results in the first column of Figure 7 allow rejecting the
null of equal accuracy in right tail forecasting at all forecast horizons.
Therefore, the relative forecasting performance is not stable, but in several
months, the single-factor model offers accuracy gains. As for the
specification with two factors, we can see that the null is rejected at all
horizons when the benchmark is the BAR(1) model and at horizons 1 and 12 when
the benchmark is the BVAR(1) model. Results concerning the center of the
distribution largely mimic what is observed for point forecasts: the single-
factor model does better than the benchmarks at horizons of one quarter and
one year.111111In the Supplement, we consider the center and the left tail as
well as robustness checks involving changes in the implementation of the
fluctuation test. We have considered if dropping verified emissions from BVAR
and BFAVAR models could alter our main conclusions, but the relative ranking
of models is not affected.
## 5 Expert forecasts, verified emissions and market monitoring
### 5.1 Expert forecasts of the nominal price of carbon
As a further step, we provide a qualitative comparison of the single-factor
BVAR(1) point forecasts against those issued by the Carbon Team at the London
Stock Exchange Group (LSEG; formerly Refinitiv), and its survey forecasts. In
both cases, LSEG provides forecasts expressed in current euros; thus, we need
to transform our real price forecasts into nominal terms by producing
inflation forecasts from an Unobserved Component SV model and using them to
get nominal price forecasts.121212Given the short time span of our evaluation
sample, using different inflation forecasts, such as RW forecasts, does not
alter the results.
LSEG Carbon Team’s forecasts. The Carbon Team at LSEG produces forecasts of
nominal EU ETS price with an irregular cadence for the period 2014-2023
ranging from 3 to 6 times per year, where the forecast horizon can be either
the current year or several years in the future. One challenge of working with
these data is that they are “fixed event forecasts”, while our models produce
“fixed horizon forecasts”. The characteristic of “fixed event forecasts” is
that the forecast horizon changes as the forecast origin moves forward. At
each forecast origin, the LSEG team produces forecasts for the current year,
$f^{FE}_{t+k|t}$, and for the next year, $f^{FE}_{t+k+12|t}$ where
$k=1,...,12$ represents the number of months until the end of the year (e.g.
$k=12$ in January and $k=1$ in December). To approximate one year ahead fixed
horizon forecasts, $f_{t+12|t}^{FH}$, using LSEG’s fixed event forecasts, we
follow Dovern et al., (2012):
Figure 8: Nominal EU ETS price and forecasts. Notes: Nominal EU ETS price (red
line), LSEG (blue dots), RW (dashed black line), BFAVAR(1) one-year-ahead
forecasts (dotted blue line). The bars represent the difference between the
RMSFE of the LSEG forecast and RW (black) or BFAVAR(1) (blue). Positive values
indicate that LSEG is less accurate than the alternative forecast.
$f_{t+12|t}^{FH}=\frac{k}{12}f^{FE}_{t+k|t}+\frac{12-k}{12}f^{FE}_{t+k+12|t},$
(6)
where weights are proportional to the degree of overlap of the two fixed event
forecasts.131313A fixed horizon forecast issued in January, $k=12$, would
therefore be equal to $f_{t+12|t}^{FH}=f^{FE}_{t+12|t}$. We obtain a set of 39
one-year-ahead forecasts irregularly spaced over the period January 2015 - May
2023; only 19 of these forecasts overlap with those in our evaluation period
spanning from December 2018 to September 2023.
Figure 8 shows that LSEG and RW forecasts are remarkably similar; indeed, the
correlation of the respective forecast errors is 0.95 over the period January
2015 - May 2023. For the period when the single-factor BVAR(1) model forecasts
overlap with those from LSEG, the model-based RMSFE is smaller than LSEG’s
RMSFE in 15 cases out of 19. All in all, while the short sample of data only
allows for a qualitative comparison, these results show that the model-based
forecasts are strikingly different from those issued by LSEG, which, on the
other hand, are similar to those from a RW model.
Figure 9: Survey and model-based one-year-ahead density forecasts of the
nominal EU ETS price from 2021 to 2023. Notes: each year the survey provides a
different set of price ranges across which respondents can choose. The title
indicates the date of the forecast, while the data come from the survey of the
previous year. The 2020 survey – leftmost plot – provided only three
categories “$<25$”, “About 25” and “$>25$” Euro per tCO2e.
Survey forecasts. Carbon Market Surveys run by LSEG each year (from 2020 to
2022) between February and April (in 2020 it was closed in March) capture the
market sentiment of respondents from a multitude of countries who are mostly
stakeholders with tangible and financial interests in carbon markets (e.g.,
traders, firms covered by an ETS). Participants who have answered the section
of the survey concerning EU ETS price expectations are 60 in 2020, 119 in
2021, and 88 in 2022. Since also survey forecasts are of the fixed event type,
we approximate fixed horizon one-year-ahead prediction densities derived from
surveys and compare them with prediction densities from the BFAVAR model and
the realized price at the target date (see Figure 9).
Only in 2023 survey and model-based forecasts are aligned and centered around
the realized price. In 2021 and 2022, both the survey and the single-factor
BVAR model tend to under-forecast and hence largely miss the run-up in the
European carbon price over this period. Nevertheless, we notice that the
realized prices lie in the right tail of the one-year-ahead prediction density
of the BFAVAR model. The fact that forecasting the price for March 2021 a year
in advance was a hard task appears evident, noticing that the price lies
outside the 95% BFAVAR prediction interval only in this case. The 2021
forecast is based on survey data collected in March 2020, at the beginning of
the COVID-19 pandemic, when overall macroeconomic uncertainty was at a record
level, especially in European countries. Similarly, also BFAVAR’s one-year-
ahead forecasts for 2021 and 2022 are based on macroeconomic data for the
COVID-19 recession period, without making any adjustments to them. These
factors should be kept in mind when evaluating the prediction accuracy for
2021 and 2022 in absolute terms.
### 5.2 Verified emissions forecasts
Forecasts of verified emissions are as crucial as price forecasts, and they
have received essentially no attention in the academic literature. Entities
participating in the EU ETS need reliable emission forecasts to plan emission
reduction strategies and ensure compliance with regulatory requirements, while
environmental agencies rely on these forecasts to assess the impact of
emissions reduction initiatives. BVAR models considered in this paper might
represent a valuable tool to produce verified emission forecasts.
Table 4 shows that, as far as point forecasts are concerned, the baseline
BVAR(1)-SV model represents the most accurate model at all horizons above 1
month. Interestingly, since data on verified emissions are published once a
year in April, the one-year-ahead forecast horizon appears particularly
relevant. Indeed, the baseline BVAR(1)-SV model and the factor-augmented
BFAVAR(1)-SV models yield RMSFE reductions over 7%. Moving to density
forecasting, Table 5 highlights that the BVAR(1)-SV and the single-factor
BFAVAR(1)-SV models yield the most accurate forecasts of the center of the
distribution of verified emissions.
Table 4: Relative RMSFE for verified emissions. | | | 1 Factor | 2 Factors | | | 1 Factor | 2 Factors
---|---|---|---|---|---|---|---|---
h | BAR(1) | BVAR(1) | BVAR(1) | BVAR(1) | BAR(1)-SV | BVAR(1)-SV | BVAR(1)-SV | BVAR(1)-SV
1 | 1.102 | 1.241 | 1.182 | 1.237 | 1.005 | 1.019 | 1.029 | 1.018
2 | 1.055 | 1.230 | 1.166 | 1.204 | 0.999 | 0.997 | 1.002 | 0.998
3 | 1.046 | 1.252 | 1.174 | 1.218 | 0.998 | 0.986 | 0.989 | 0.988
4 | 1.042 | 1.268 | 1.175 | 1.226 | 0.997 | 0.975 | 0.979 | 0.978
5 | 1.033 | 1.247 | 1.154 | 1.204 | 0.996 | 0.963 | 0.965 | 0.965
6 | 1.029 | 1.238 | 1.141 | 1.193 | 0.996 | 0.956 | 0.958 | 0.957
7 | 1.023 | 1.209 | 1.117 | 1.166 | 0.995 | 0.948 | 0.950 | 0.950
8 | 1.021 | 1.195 | 1.104 | 1.153 | 0.996 | 0.944 | 0.946 | 0.946
9 | 1.018 | 1.173 | 1.086 | 1.133 | 0.996 | 0.939 | 0.942 | 0.941
10 | 1.014 | 1.149 | 1.071 | 1.112 | 0.996 | 0.935 | 0.938 | 0.938
11 | 1.014 | 1.143 | 1.065 | 1.106 | 0.996 | 0.930 | 0.934 | 0.933
12 | 1.012 | 1.129 | 1.053 | 1.093 | 0.996 | 0.926 | 0.929 | 0.929
Notes: see notes to Table 1 Table 5: Quantile-weighted Continuous Ranked
Probability Score (qCRPS) for verified emissions. (a) Quantile-weighted CRPS:
center
---
| | | 1 Factor | 2 Factors | | | 1 Factor | 2 Factors
h | BAR(1) | BVAR(1) | BVAR(1) | BVAR(1) | BAR(1)-SV | BVAR-SV(1) | BVAR-SV(1) | BVAR-SV(1)
1 | 154.876 | 154.834 | 154.813 | 154.799 | 154.837 | 154.819 | 154.812 | 154.824
2 | 154.875 | 154.813 | 154.787 | 154.774 | 154.791 | 154.775 | 154.759 | 154.781
3 | 154.890 | 154.810 | 154.786 | 154.771 | 154.757 | 154.743 | 154.725 | 154.750
4 | 154.929 | 154.836 | 154.814 | 154.798 | 154.746 | 154.734 | 154.716 | 154.742
5 | 154.975 | 154.868 | 154.853 | 154.839 | 154.743 | 154.730 | 154.712 | 154.741
6 | 155.076 | 154.958 | 154.946 | 154.931 | 154.795 | 154.782 | 154.765 | 154.795
7 | 155.180 | 155.055 | 155.046 | 155.035 | 154.851 | 154.839 | 154.825 | 154.856
8 | 155.332 | 155.200 | 155.194 | 155.191 | 154.953 | 154.9380 | 154.928 | 154.966
9 | 155.424 | 155.284 | 155.282 | 155.301 | 154.997 | 154.983 | 154.974 | 155.020
10 | 155.583 | 155.442 | 155.441 | 155.451 | 155.109 | 155.093 | 155.091 | 155.134
11 | 155.698 | 155.551 | 155.550 | 155.568 | 155.175 | 155.156 | 155.160 | 155.207
12 | 155.792 | 155.632 | 155.633 | 155.658 | 155.226 | 155.201 | 155.208 | 155.257
Notes: see notes to Table 3
### 5.3 Market monitoring
Following Baumeister et al., (2022), we construct indices of demand pressure,
upward, and downward price pressure for the EU ETS market. For approximating
demand pressure, we take the difference between one-year and one-month-ahead
verified emission forecasts from the BFAVAR(1)-SV model. A negative value of
the proxy signals expectations of loosening market conditions over the next
year.
Using the predictive densities delivered by Bayesian estimation of the single-
factor BVAR(1) model, upward and downward price pressure indices are defined
as follows:
$\displaystyle PP^{+}_{t}$
$\displaystyle=\frac{1}{12}\sum_{h=1}^{12}\mathbb{I}\left[\hat{R}_{t+h|t}>\max\left(R_{t},R_{t-1},\ldots,R_{t-11}\right)\right],$
(7) $\displaystyle PP^{-}_{t}$
$\displaystyle=\frac{1}{12}\sum_{h=1}^{12}\mathbb{I}\left[\hat{R}_{t+h|t}<\min\left(R_{t},R_{t-1},\ldots,R_{t-11}\right)\right].$
(8)
These proxies estimate the probability that over the next 12 months, the real
price of carbon is above (below) the maximum (minimum) value observed in the
previous year.
The upper panel of Figure 10 displays the demand pressure proxy as a set of
bars along with a line representing verified emissions. Note that, although
verified emissions are available at an annual sampling frequency, BVAR models
are estimated with monthly data; therefore, the demand pressure index allows
monitoring expectations about the EU ETS market in real time each month. The
index is always negative to track the long-term decline of verified emissions,
interspersed with an increase in 2021 in the aftermath of the COVID-19
recession.
Figure 10: Demand and price pressure indices for the EU ETS market from March
2018 to August 2023. Notes: we plot the backward 3-month moving average of
price and demand pressure indices. Indices are aligned with the forecast
origin.
As shown by Bjørnland et al., (2023), the dynamics of verified emissions are
elicited by three main forces: a supply shock and two distinct demand shocks.
On the supply side, we have EU ETS regulation: the cap and trade scheme has
progressively tightened the limit of total greenhouse gas emissions of covered
entities. This is the first element factored in the negative demand pressure
index. The two demand shocks are related to economic activity and transition
demand. The first essentially depends on the business cycle and, therefore,
can be associated with positive or negative changes in the demand pressure
index. For instance, during the COVID pandemic, expectations of a future quick
recovery associated with an increase in industrial activity are captured by a
rebound of the demand pressure index. Much like the supply-side shock, the
transition-demand shock – capturing, among other things, the increased usage
of renewables – contributes to the steady decline in verified emissions and
therefore keeps the pressure index in negative territory.
The two price pressure indices, along with the evolution of the real EU ETS
prices, are shown in the middle and bottom panels of Figure 10. Given that
between 2018 and 2023, the real price has steadily increased – peaking at
almost 120 Euros in early 2023 – the upward pressure index in the middle panel
always signals expectations of soaring prices. Two exceptions are recorded in
2020 and in early 2023 when the downward price pressure index rises
temporarily, capturing expectations of price decreases.
## 6 Conclusions
The fact that the EU ETS regulation is strengthening over time calls for even
more analyses focusing on the macroeconomic effects of carbon price shocks.
Indeed, the ECB already embeds technical assumptions on carbon pricing in its
projections (European Central Bank, 2021b, ). Technical assumptions boil down
to setting the trajectory for key variables entering the ECB’s macroeconomic
models over the projection horizon and are derived in a variety of ways,
including univariate and multivariate econometric models and using the price
of futures contracts (European Central Bank,, 2016).
In this paper, we have identified carbon price drivers and methodological
choices that can directly inform projections and scenario analyses used to
gauge the macroeconomic effect of carbon price shocks. Our results show that
EU ETS prices and verified emissions can be forecasted with relatively simple
BVAR models estimated with monthly data.
There are at least two aspects of our analysis that deserve further
investigation. First, the use of time series sampled at different frequencies.
Mixed Frequency and Mixed-data sampling (MIDAS) models could be used to
further improve monthly forecasts relying on data sampled at a higher
frequency, such as weather and financial variables. Moreover, given that
several key predictors of real carbon prices – such as the level of verified
emissions and some macroeconomic aggregates – are available only at the lower
sampling frequency, it is also interesting to rely on reverse-MIDAS approaches
to exploit these data (Foroni et al.,, 2023). Lastly, our results could be
extended in the direction of real-time data to assess the impact of data
revisions on monthly forecasts of the price of carbon.
## Supplementary Materials
An online Supplement provides further forecasting results and different
fluctuation tests for various models and rolling windows.
## References
* Baumeister et al., (2022) Baumeister, C., Korobilis, D., and Lee, T. K. (2022). Energy markets and global economic conditions. Review of Economics and Statistics, 104(4):828–844.
* Bjørnland et al., (2023) Bjørnland, H., Cross, J. L., and Kapfhammer, F. (2023). The drivers of emission reductions in the European carbon market. CAMP Working Paper Series 8/2023, BI Norwegian Business School.
* Boivin and Ng, (2006) Boivin, J. and Ng, S. (2006). Are more data always better for factor analysis? Journal of Econometrics, 132(1):169–194.
* Carleton and Hsiang, (2016) Carleton, T. A. and Hsiang, S. M. (2016). Social and economic impacts of climate. Science, 353(6304):aad9837.
* Chan, (2023) Chan, J. C. (2023). Comparing stochastic volatility specifications for large Bayesian VARs. Journal of Econometrics, 235(2):1419–1446.
* (6) Chevallier, J. (2011a). Macroeconomics, finance, commodities: interactions with carbon markets in a data-rich model. Economic Modelling, 28(1-2):557–567.
* (7) Chevallier, J. (2011b). A model of carbon price interactions with macroeconomic and energy dynamics. Energy Economics, 33(6):1295–1312.
* Chevallier, (2012) Chevallier, J. (2012). Econometric analysis of carbon markets. Springer Netherlands.
* Chow and Lin, (1971) Chow, G. C. and Lin, A.-L. (1971). Best linear unbiased interpolation, distribution, and extrapolation of time series by related series. The Review of Economics and Statistics, 53(4):372–375.
* Ciccarelli and Marotta, (2024) Ciccarelli, M. and Marotta, F. (2024). Demand or supply? An empirical exploration of the effects of climate change on the macroeconomy. Energy Economics, 129:107163.
* Clark and Ravazzolo, (2015) Clark, T. E. and Ravazzolo, F. (2015). Macroeconomic forecasting performance under alternative specifications of time-varying volatility. Journal of Applied Econometrics, 30(4):551–575.
* Coroneo and Iacone, (2020) Coroneo, L. and Iacone, F. (2020). Comparing predictive accuracy in small samples using fixed-smoothing asymptotics. Journal of Applied Econometrics, 35(4):391–409.
* Dell et al., (2014) Dell, M., Jones, B. F., and Olken, B. A. (2014). What do we learn from the weather? The new climate-economy literature. Journal of Economic literature, 52(3):740–798.
* Diebold and Mariano, (1995) Diebold, F. X. and Mariano, R. S. (1995). Comparing predictive accuracy. Journal of Business & Economic Statistics, 13(1):253.
* Dovern et al., (2012) Dovern, J., Fritsche, U., and Slacalek, J. (2012). Disagreement among forecasters in g7 countries. Review of Economics and Statistics, 94(4):1081–1096.
* European Central Bank, (2016) European Central Bank (2016). A guide to the Eurosystem/ECB staff macroeconomic projection exercises. Available at: https://data.europa.eu/doi/10.2866/35667.
* (17) European Central Bank (2021a). Climate change and monetary policy in the Euro Area. Occasional Paper Series 271, European Central Bank.
* (18) European Central Bank (2021b). ECB presents action plan to include climate change considerations in its monetary policy strategy. Technical report, Press Release. Available at: https://www.ecb.europa.eu.
* Foroni et al., (2023) Foroni, C., Ravazzolo, F., and Rossini, L. (2023). Are low frequency macroeconomic variables important for high frequency electricity prices? Economic Modelling, 120:106160.
* Fullerton and Muehlegger, (2019) Fullerton, D. and Muehlegger, E. (2019). Who bears the economic burdens of environmental regulations? Review of Environmental Economics and Policy, 13(1):62–82.
* Giacomini and Rossi, (2010) Giacomini, R. and Rossi, B. (2010). Forecast comparisons in unstable environments. Journal of Applied Econometrics, 25(4):595–620.
* Gneiting and Ranjan, (2011) Gneiting, T. and Ranjan, R. (2011). Comparing density forecasts using threshold- and quantile-weighted scoring rules. Journal of Business & Economic Statistics, 29(3):411–422.
* Hsiang et al., (2019) Hsiang, S., Oliva, P., and Walker, R. (2019). The distribution of environmental damages. Review of Environmental Economics and Policy, 13(1):83–103.
* International Carbon Action Partnership, (2023) International Carbon Action Partnership (2023). Emissions trading worldwide. Status Report 2023.
* Känzig, (2023) Känzig, D. R. (2023). The unequal economic consequences of carbon pricing. Working Paper 31221, National Bureau of Economic Research.
* Känzig and Konradt, (2023) Känzig, D. R. and Konradt, M. (2023). Climate policy and the economy: evidence from Europe’s carbon pricing initiatives. Working Paper 31260, National Bureau of Economic Research.
* Koop and Tole, (2013) Koop, G. and Tole, L. (2013). Forecasting the European carbon market. Journal of the Royal Statistical Society Series A: Statistics in Society, 176(3):723–741.
* Lei et al., (2022) Lei, H., Xue, M., and Liu, H. (2022). Probability distribution forecasting of carbon allowance prices: a hybrid model considering multiple influencing factors. Energy Economics, 113:106189.
* Mansanet-Bataller et al., (2007) Mansanet-Bataller, M., Pardo, A., and Valor, E. (2007). CO2 prices, energy and weather. The Energy Journal, 28(3).
* Marcu et al., (2023) Marcu, A., Hernández, J. F. L., Romeo, G., Alberola, E., Faure, A., Obienu, C., Qin, B., O’Neill, M., Caneill, J. Y., and Schleicher, S. (2023). 2023 State of the EU ETS Report. Available at: https://ercst.org/2023-state-of-the-eu-ets-report/.
* Moessner, (2022) Moessner, R. (2022). Effects of carbon pricing on inflation. Working Papers 9563, CESifo.
* NGFS, (2022) NGFS (2022). NGFS scenarios for central banks and supervisors. NGFS publications, Central Banks and Supervisors Network for Greening the Financial System (NGFS). Available at: https://www.ngfs.net/en.
* Pan et al., (2023) Pan, J., Cross, J. L., and Zhang, B. (2023). A review of carbon emissions reduction mechanisms. Working paper.
* Parry et al., (2021) Parry, I., Black, S., and Roaf, J. (2021). Proposal for an international carbon price floor among large emitters. Staff Climate Notes 2021/001, International Monetary Fund.
* Pesaran and Timmermann, (2009) Pesaran, M. H. and Timmermann, A. (2009). Testing dependence among serially correlated multicategory variables. Journal of the American Statistical Association, 104(485):325–337.
* Quilis, (2013) Quilis, E. M. (2013). Temporal disaggregation library. MATLAB Central File Exchange.
* Rossi, (2021) Rossi, B. (2021). Forecasting in the presence of instabilities: How we know whether models predict well and how to improve them. Journal of Economic Literature, 59(4):1135–1190.
* Tan et al., (2022) Tan, X., Sirichand, K., Vivian, A., and Wang, X. (2022). Forecasting European carbon returns using dimension reduction techniques: commodity versus financial fundamentals. International Journal of Forecasting, 38(3):944–969.
|
# On the Kohayakawa–Kreuter conjecture
Eden Kuperwasser School of Mathematical Sciences, Tel Aviv University, Tel
Aviv 6997801, Israel<EMAIL_ADDRESS>,
Wojciech Samotij and Yuval Wigderson
###### Abstract.
Let us say that a graph $G$ is Ramsey for a tuple $(H_{1},\dots,H_{r})$ of
graphs if every $r$-coloring of the edges of $G$ contains a monochromatic copy
of $H_{i}$ in color $i$, for some $i\in\llbracket{r}\rrbracket$. A famous
conjecture of Kohayakawa and Kreuter, extending seminal work of Rödl and
Ruciński, predicts the threshold at which the binomial random graph $G_{n,p}$
becomes Ramsey for $(H_{1},\dots,H_{r})$ asymptotically almost surely. In this
paper, we resolve the Kohayakawa–Kreuter conjecture for almost all tuples of
graphs. Moreover, we reduce its validity to the truth of a certain
deterministic statement, which is a clear necessary condition for the
conjecture to hold. All of our results actually hold in greater generality,
when one replaces the graphs $H_{1},\dots,H_{r}$ by finite families
$\mathcal{H}_{1},\dots,\mathcal{H}_{r}$. Additionally, we pose a natural
(deterministic) graph-partitioning conjecture, which we believe to be of
independent interest, and whose resolution would imply the Kohayakawa–Kreuter
conjecture.
EK, WS, and YW are supported by ERC Consolidator Grant 101044123
(RandomHypGra), by Israel Science Foundation Grant 2110/22, and by NSF–BSF
Grant 2019679. YW is additionally supported by ERC Consolidator Grant 863438
(LocalGlobal).
## 1\. Introduction
### 1.1. Symmetric Ramsey properties of random graphs
Given graphs $G$ and $H_{1},\dotsc,H_{r}$, one says that $G$ is _Ramsey for
the tuple $(H_{1},\dotsc,H_{r})$_ if, for every $r$-coloring of the edges of
$G$, there is a monochromatic copy of $H_{i}$ in some color
$i\in\llbracket{r}\rrbracket$. In the symmetric case $H_{1}=\dotsb=H_{r}=H$,
we simply say that $G$ is _Ramsey for $H$ in $r$ colors_. Ramsey’s theorem
[24] implies that the complete graph $K_{n}$ is Ramsey for
$(H_{1},\dotsc,H_{r})$ whenever $n$ is sufficiently large. The fundamental
question of graph Ramsey theory is to determine, for a given tuple
$(H_{1},\dotsc,H_{r})$, which graphs $G$ are Ramsey for it. For more on this
question, as well as the many fascinating sub-questions it contains, we refer
the reader to the survey [3].
In this paper, we are interested in Ramsey properties of random graphs, a
topic that was initiated in the late 1980s by Frankl–Rödl [6] and
Łuczak–Ruciński–Voigt [31]. The main question in this area is, for a given
tuple $(H_{1},\dotsc,H_{r})$, which functions $p=p(n)$ satisfy that $G_{n,p}$
is Ramsey for $(H_{1},\dots,H_{r})$ a.a.s.111As usual, $G_{n,p}$ denotes the
binomial random graph with edge probability $p$ and we say that an event
happens _asymptotically almost surely (a.a.s.)_ if its probability tends to
$1$ as $n\to\infty$. In the case $H_{1}=\dotsb=H_{r}$, this question was
resolved in the remarkable work of Rödl and Ruciński [25, 26, 27]. In order to
state their result, we need the following terminology and notation. For a
graph $J$, we denote by $v_{J}$ and $e_{J}$ the number of vertices and edges,
respectively, of $J$. The _maximal $2$-density_ of a non-empty graph $H$ with
$v_{H}\geqslant 3$ is then defined222We also define $m_{2}(K_{2})\coloneqq
1/2$ and $m_{2}(H)\coloneqq 0$ if $H$ has no edges. to be
$m_{2}(H)\coloneqq\max\left\\{\frac{e_{J}-1}{v_{J}-2}:J\subseteq
H,v_{J}\geqslant 3\right\\}.$
With this notation, we can state the random Ramsey theorem of Rödl and
Ruciński [27].
###### Theorem 1.1 (Rödl–Ruciński [27]).
For every graph $H$ which is not a forest333Rödl and Ruciński also determined
the Ramsey threshold when $H$ is a forest, but for simplicity we do not state
this more general result. and every integer $r\geqslant 2$, there exist
constants $c,C>0$ such that
$\lim_{n\to\infty}\operatorname{Pr}(G_{n,p}\text{ is Ramsey for $H$ in $r$
colors})=\begin{cases}1&\text{if }p\geqslant Cn^{-1/m_{2}(H)},\\\ 0&\text{if
}p\leqslant cn^{-1/m_{2}(H)}.\end{cases}$
As with many such threshold results for random graph properties, Theorem 1.1
really consists of two statements: the _$1$ -statement_, which says that
$G_{n,p}$ satisfies the desired property a.a.s. once $p$ is above some
threshold, and the _$0$ -statement_, which says that $G_{n,p}$ a.a.s. fails to
satisfy the desired property if $p$ is below some threshold.
In recent years, there has been a great deal of work on transferring
combinatorial theorems, such as Ramsey’s theorem or Turán’s theorem [30], to
sparse random settings. As a consequence, several new proofs of the
$1$-statement of Theorem 1.1 have been found. Two such proofs were first given
by Conlon–Gowers [4] and, independently, by Friedgut–Rödl–Schacht [8] (see
also Schacht [29]) with the use of their transference principles. More
recently, Nenadov and Steger [22] found a very short proof of the 1-statement
of Theorem 1.1 that uses the hypergraph container method of Saxton–Thomason
[28] and Balogh–Morris–Samotij [1].
However, these techniques are not suitable for proving the respective
0-statements such as that in Theorem 1.1. Furthermore, whereas the 0-statement
of the aforementioned sparse random analogue of Turán’s theorem is very easy
to establish, proving the 0-statement of Theorem 1.1 requires a significant
amount of work. To understand this, suppose that $G$ is some graph that is
Ramsey for $H$ in $r$ colors. As is well-known (see e.g. [14, Theorem 3.4]),
the probability that $G_{n,p}$ contains $G$ as a subgraph is bounded away from
zero if (and only if) $p=\Omega(n^{-1/m(G)})$, where $m(G)$ is the _maximal
density_ of $G$, defined by
$m(G)\coloneqq\max\left\\{\frac{e_{J}}{v_{J}}:J\subseteq G,v_{J}\geqslant
1\right\\}.$
In particular, if $m(G)\leqslant m_{2}(H)$, then the $0$-statement of Theorem
1.1 cannot hold. Therefore, a prerequisite for any proof of the $0$-statement
is the following result, which Rödl–Ruciński [25] termed the _deterministic
lemma_ : If $G$ is Ramsey for $H$ in $r$ colors, then $m(G)>m_{2}(H)$. We
stress that this result is by no means trivial; in particular, it turns out to
be false if we remove the assumption that $H$ is not a forest [27, 7], or if
we move from graphs to hypergraphs [9].
To complement the deterministic lemma, Rödl–Ruciński also proved what they
termed a _probabilistic lemma_. Loosely speaking, this is a result that says
that the $0$-statement of Theorem 1.1 is actually _equivalent_ to the
deterministic lemma. In other words, an obvious necessary condition for the
validity of the $0$-statement—the non-existence of a graph $G$ that is Ramsey
for $H$ and satisfies $m(G)\leqslant m_{2}(H)$—is also a sufficient condition.
### 1.2. Asymmetric Ramsey properties of random graphs
Given our good understanding of Ramsey properties of random graphs in the
symmetric case, provided by Theorem 1.1, it is natural to ask what happens if
we remove the assumption that $H_{1}=\dotsb=H_{r}$. This question was first
raised by Kohayakawa and Kreuter [15], who proposed a natural conjecture for
the threshold controlling when $G_{n,p}$ is Ramsey for an arbitrary tuple
$(H_{1},\dotsc,H_{r})$. To state their conjecture, we need the notion of the
_mixed $2$-density_: For graphs $H_{1},H_{2}$ with $m_{2}(H_{1})\geqslant
m_{2}(H_{2})$, their mixed $2$-density is defined as
$m_{2}(H_{1},H_{2})\coloneqq\max\left\\{\frac{e_{J}}{v_{J}-2+1/m_{2}(H_{2})}:J\subseteq
H_{1},v_{J}\geqslant 2\right\\}.$
With this terminology, we may state the conjecture of Kohayakawa and Kreuter
[15].
###### Conjecture 1.2 (Kohayakawa–Kreuter [15]).
Let $H_{1},\dots,H_{r}$ be graphs satisfying
$m_{2}(H_{1})\geqslant\dotsb\geqslant m_{2}(H_{r})$ and $m_{2}(H_{2})>1$.
There exist constants $c,C>0$ such that
$\lim_{n\to\infty}\operatorname{Pr}(G_{n,p}\text{ is Ramsey for
}(H_{1},\dotsc,H_{r}))=\begin{cases}1&\text{if }p\geqslant
Cn^{-1/m_{2}(H_{1},H_{2})},\\\ 0&\text{if }p\leqslant
cn^{-1/m_{2}(H_{1},H_{2})}.\end{cases}$
The assumption $m_{2}(H_{2})>1$ is equivalent to requiring that $H_{1}$ and
$H_{2}$ are not forests; it was added by Kohayakawa, Schacht, and Spöhel [16]
to rule out sporadic counterexamples, in analogy with the assumption that $H$
is not a forest in Theorem 1.1.
The role of the mixed $2$-density $m_{2}(H_{1},H_{2})$ in the context of
Conjecture 1.2 can seem a little mysterious at first, but there is a natural
(heuristic) explanation. Since one can color all edges that do not lie in a
copy of $H_{1}$ with color $1$, the only important edges are those that do lie
in copies of $H_{1}$. The mixed $2$-density is defined in such a way that
$p=\Theta(n^{-1/m_{2}(H_{1},H_{2})})$ is the threshold at which the number of
copies of (the densest subgraph of) each of $H_{2},\dotsc,H_{r}$ is at least
of the same order of magnitude as the number of edges in the union of all
copies of (the densest subgraph of) $H_{1}$ in $G_{n,p}$. Since at least one
edge in each copy of $H_{1}$ must receive a color from $\\{2,\dotsc,r\\}$,
this is the point where avoiding monochromatic copies of $H_{2},\dotsc,H_{r}$
becomes difficult.
Conjecture 1.2 has received a great deal of attention over the years, and has
been proved in a number of special cases. Following a sequence of partial
results [15, 19, 16, 9, 11], the $1$-statement of Conjecture 1.2 was proved by
Mousset, Nenadov, and Samotij [20] with the use of the container method as
well as a randomized “typing” procedure. We henceforth focus on the
$0$-statement, where progress has been more limited.
Note that, in order to prove the $0$-statement, one can make several
simplifying assumptions. First, one can assume that $r$, the number of colors,
is equal to $2$. Indeed, if one can a.a.s. $2$-color the edges of $G_{n,p}$
and avoid monochromatic copies of $H_{1},H_{2}$ in colors $1,2$, respectively,
then certainly $G_{n,p}$ is not Ramsey for $(H_{1},\dots,H_{r})$. Furthermore,
if $H_{2}^{\prime}\subseteq H_{2}$ is a subgraph satisfying
$m_{2}(H_{2}^{\prime})=m_{2}(H_{2})$, then the $0$-statement for the pair
$(H_{1},H_{2}^{\prime})$ implies the $0$-statement for $(H_{1},H_{2})$, as any
coloring with no monochromatic copy of $H_{2}^{\prime}$ in particular has no
monochromatic copy of $H_{2}$. Thus, we may assume that $H_{2}$ is _strictly
$2$-balanced_, meaning that $m_{2}(H_{2}^{\prime})<m_{2}(H_{2})$ for any
$H_{2}^{\prime}\subsetneq H_{2}$. For exactly the same reason, we may assume
that $H_{1}$ is _strictly $m_{2}(\cdot,H_{2})$-balanced_, meaning that
$m_{2}(H_{1}^{\prime},H_{2})<m_{2}(H_{1},H_{2})$ for any
$H_{1}^{\prime}\subsetneq H_{1}$. Let us say that the pair $(H_{1},H_{2})$ is
_strictly balanced_ if $H_{2}$ is strictly $2$-balanced and $H_{1}$ is
strictly $m_{2}(\cdot,H_{2})$-balanced. Additionally, let us say that
$(H_{1}^{\prime},H_{2}^{\prime})$ is a _strictly balanced pair of subgraphs_
of $(H_{1},H_{2})$ if $(H_{1}^{\prime},H_{2}^{\prime})$ is strictly balanced
and satisfies $m_{2}(H_{2}^{\prime})=m_{2}(H_{2})$ and
$m_{2}(H_{1}^{\prime},H_{2}^{\prime})=m_{2}(H_{1},H_{2})$. All previous works
on the $0$-statement of Conjecture 1.2 have made these simplifying
assumptions, working in the case $r=2$ and with a strictly balanced pair
$(H_{1},H_{2})$.
The original paper of Kohayakawa and Kreuter [15] proved the $0$-statement of
Conjecture 1.2 when $H_{1}$ and $H_{2}$ are cycles. This was extended to the
case when both $H_{1}$ and $H_{2}$ are cliques in [19], and to the case when
$H_{1}$ is a clique and $H_{2}$ is a cycle in [18]. To date, the most general
result is due to Hyde [13], who proved the $0$-statement of Conjecture 1.2 for
almost all pairs of regular graphs $(H_{1},H_{2})$; in fact, this follows from
Hyde’s main result [13, Theorem 1.9], which establishes a certain
deterministic condition whose validity implies the $0$-statement of Conjecture
1.2. Finally, the first two authors [17] recently proved the $0$-statement of
Conjecture 1.2 in the case where $m_{2}(H_{1})=m_{2}(H_{2})$. Because of this,
we henceforth focus on the case that $m_{2}(H_{1})>m_{2}(H_{2})$.
### 1.3. New results
As in the symmetric setting, a necessary prerequisite for proving the
$0$-statement of Conjecture 1.2 is proving the following _deterministic lemma_
: If $G$ is Ramsey for $(H_{1},H_{2})$, then $m(G)>m_{2}(H_{1},H_{2})$. The
main result in this paper is a corresponding probabilistic lemma, which states
that this obvious necessary condition is also sufficient.
###### Theorem 1.3.
The $0$-statement of Conjecture 1.2 holds if and only if, for every strictly
balanced pair $(H_{1},H_{2})$, every graph $G$ that is Ramsey for
$(H_{1},H_{2})$ satisfies $m(G)>m_{2}(H_{1},H_{2})$.
More precisely, we prove that if $(H_{1},H_{2})$ is any pair of graphs and
$(H_{1}^{\prime},H_{2}^{\prime})$ is a strictly balanced pair of subgraphs of
$(H_{1},H_{2})$, then the $0$-statement of Conjecture 1.2 holds for
$(H_{1},H_{2})$ if every graph $G$ which is Ramsey for
$(H_{1}^{\prime},H_{2}^{\prime})$ satisfies
$m(G)>m_{2}(H_{1}^{\prime},H_{2}^{\prime})=m_{2}(H_{1},H_{2})$.
While we believe that the probabilistic lemma, Theorem 1.3, is our main
contribution, we are able to prove the deterministic lemma in a wide range of
cases. This implies that the $0$-statement of Conjecture 1.2 is true for
almost all pairs of graphs. The most general statement we can prove is
slightly tricky to state because of the necessity of passing to a strictly
balanced pair of subgraphs; however, here is a representative example of our
results, which avoids this technicality and still implies Conjecture 1.2 for
almost all pairs of graphs. We state the more general result in Theorem 1.7
below.
###### Theorem 1.4.
Conjecture 1.2 holds for all sequences $H_{1},\dots,H_{r}$ of graphs
satisfying $m_{2}(H_{1})\geqslant\dotsb\geqslant m_{2}(H_{r})$ and
$m_{2}(H_{2})>\frac{11}{5}$.
As discussed above, Theorem 1.4 follows easily from Theorem 1.3 and a
deterministic lemma for strictly balanced pairs $(H_{1},H_{2})$ satisfying
$m_{2}(H_{1})\geqslant m_{2}(H_{2})>\frac{11}{5}$. The deterministic lemma in
this setting is actually very straightforward and follows from standard
coloring techniques.
Using a number of other coloring techniques, we can prove the deterministic
lemma (and thus Conjecture 1.2) in several additional cases, which we discuss
below. However, let us first propose a conjecture, which we believe to be of
independent interest, and whose resolution would immediately imply Conjecture
1.2 in all cases.
###### Conjecture 1.5.
For any graph $G$, there exists a forest $F\subseteq G$ such that
$m_{2}(G\setminus F)\leqslant m(G).$
Here, $G\setminus F$ denotes the graph obtained from $G$ by deleting the edges
of $F$ (but not deleting any vertices). To give some intuition for Conjecture
1.5, we note that $m(G)\leqslant m_{2}(G)\leqslant m(G)+1$ for any graph $G$,
and that $m_{2}(F)=1$ for any forest $F$ which is not a matching. Thus, it is
natural to expect that by deleting the edges of a forest, we could decrease
$m_{2}(G)$ by roughly $1$. Conjecture 1.5 says that this is roughly the case,
in that the deletion of an appropriately-chosen forest can decrease $m_{2}(G)$
to lie below $m(G)$.
Moreover, we note that Conjecture 1.5 easily implies the deterministic lemma
in all cases444Recall that the case of $m_{2}(H_{1})=m_{2}(H_{2})$ was settled
in [17], so we may freely make this assumption. with
$m_{2}(H_{1})>m_{2}(H_{2})$, and thus implies Conjecture 1.2. Indeed, it is
straightforward to verify in this case that $m_{2}(H_{1})>m_{2}(H_{1},H_{2})$
(see Lemma 3.4 below). Now, suppose that $G$ is some graph with $m(G)\leqslant
m_{2}(H_{1},H_{2})<m_{2}(H_{1})$. If Conjecture 1.5 is true, we may partition
the edges of $G$ into a forest $F$ and a graph $K$ with $m_{2}(K)\leqslant
m(G)<m_{2}(H_{1})$. This latter condition implies, in particular, that $K$
contains no copy of $H_{1}$. Additionally, by the assumption $m_{2}(H_{2})>1$
in Conjecture 1.2, we know that $H_{2}$ contains a cycle and thus $F$ contains
no copy of $H_{2}$. In other words, coloring the edges of $K$ with color $1$
and the edges of $F$ with color $2$ witnesses that $G$ is not Ramsey for
$(H_{1},\dots,H_{r})$.
Because of this, it would be of great interest to prove Conjecture 1.5.
Somewhat surprisingly, we know how to prove Conjecture 1.5 under the extra
assumption that $m(G)$ is an integer. This extra condition seems fairly
artificial, but we do not know how to remove it—our technique uses tools from
matroid theory that seem to break down once $m(G)$ is no longer an integer. We
present this proof in Appendix B, in the hope that it may serve as a first
step to the full resolution of Conjecture 1.5, and thus Conjecture 1.2.
Although we are not able to resolve Conjecture 1.5, we do have a number of
other techniques for proving the deterministic lemma, and thus Conjecture 1.2,
under certain assumptions. First, we are able to resolve the case when the
number of colors is at least three and $m_{2}(H_{2})=m_{2}(H_{3})$.
###### Theorem 1.6.
Let $H_{1},\dots,H_{r}$ be a sequence of graphs with $r\geqslant 3$ and
suppose that $m_{2}(H_{1})\geqslant
m_{2}(H_{2})=m_{2}(H_{3})\geqslant\dotsb\geqslant m_{2}(H_{r})$ and
$m_{2}(H_{2})>1$. Then Conjecture 1.2 holds for $H_{1},\dots,H_{r}$.
We can also prove Conjecture 1.2 in a number of additional cases, expressed in
terms of the properties of (a strictly balanced pair of subgraphs of) the pair
$(H_{1},H_{2})$ of two densest graphs.
Recall that the _degeneracy_ of $H$ is the maximum over all $J\subseteq H$ of
the minimum degree of $J$.
###### Theorem 1.7.
Suppose that $(H_{1},H_{2})$ is strictly balanced. Suppose additionally that
one of the following conditions holds:
1. (a)
$\chi(H_{2})\geqslant 3$, or
2. (b)
$H_{2}$ is not the union of two forests, or
3. (c)
$\chi(H_{1})>m_{2}(H_{1},H_{2})+1$, or
4. (d)
$H_{1}$ has degeneracy at least $\lfloor 2m_{2}(H_{1},H_{2})\rfloor$, or
5. (e)
$H_{1}=K_{s,t}$ for some $s,t\geqslant 2$, or
6. (f)
$m_{2}(H_{1})>\lceil m_{2}(H_{1},H_{2})\rceil$.
In any of these cases, Conjecture 1.2 holds for $(H_{1},H_{2})$.
###### Remark.
The only graphs $H_{2}$ which do not satisfy (a) or (b) are sparse bipartite
graphs, such as even cycles. On the other hand, (c) applies whenever $H_{1}$
is a clique555Note that $m_{2}(H_{1},H_{2})\leqslant m_{2}(H_{1})$, hence (c)
holds if $\chi(H_{1})>m_{2}(H_{1})+1$, and cliques satisfy
$m_{2}(K_{k})=\frac{k+1}{2}$. or, more generally, a graph obtained from a
clique by deleting few edges. Moreover, (d) applies to reasonably dense
graphs, as well as all $d$-regular bipartite graphs with $d\geqslant 2$, and
(e) handles all cases when $H_{1}$ is a biclique666In fact, our proof of (e)
applies to a larger class of graphs, which we call _$(s,t)$ -graphs_; see
Section 5 for details.. Thus, very roughly speaking, the strictly balanced
cases that remain open in Conjecture 1.2 are those in which $H_{2}$ is
bipartite and very sparse and $H_{1}$ is not “too dense”.
Case (f) is somewhat stranger and it is not obvious that there exist graphs to
which it applies. However, one can check that, for example, it applies if
$H_{1}=K_{3,3,3,3}$ and $H_{2}=C_{8}$, and that none of the other cases of
Theorem 1.7 (or any of the earlier results on Conjecture 1.2) apply in this
case. However, the main reason we include (f) is that it is implied by our
partial progress on Conjecture 1.5; since we believe that this conjecture is
the correct approach to settling Conjecture 1.2 in its entirety, we wanted to
highlight (f).
We remark that, unfortunately, the conditions in Theorem 1.7 do not exhaust
all cases. While it is quite likely that simple additional arguments could
resolve further cases, Conjecture 1.5 remains the only (conjectural) approach
we have found to resolve Conjecture 1.2 in all cases. Moreover, our proof of
the probabilistic lemma implies that, in order to prove Conjecture 1.2 for a
pair $(H_{1},H_{2})$, it is enough to prove the deterministic lemma for graphs
$G$ of order not exceeding an explicit constant $K=K(H_{1},H_{2})$. In
particular, the validity of Conjecture 1.2 for any specific pair of graphs
reduces to a finite computation.
### 1.4. Ramsey properties of graph families
All of the results discussed in the previous subsection hold in greater
generality, when we replace $H_{1},\dots,H_{r}$ with $r$ finite families of
graphs. In addition to being interesting in its own right, such a
generalization also has important consequences in the original setting of
Conjecture 1.2; indeed, our proof of the three-color result, Theorem 1.6,
relies on our ability to work with graph families. Before we state our more
general results, we need the following definitions.
###### Definition 1.8.
Let $\mathcal{H}_{1},\dots,\mathcal{H}_{r}$ be finite families of graphs. We
say that a graph $G$ is _Ramsey_ for $(\mathcal{H}_{1},\dots,\mathcal{H}_{r})$
if every $r$-coloring of $E(G)$ contains a monochromatic copy of some
$H_{i}\in\mathcal{H}_{i}$ in some color $i\in\llbracket{r}\rrbracket$.
We now define the appropriate generalizations of the notions of maximum
$2$-density and mixed $2$-density to families of graphs. First, given a finite
family of graphs $\mathcal{H}$, we let
$m_{2}(\mathcal{H})\coloneqq\min_{H\in\mathcal{H}}m_{2}(H).$
Second, given a graph $H$ and a (finite) family $\mathcal{L}$ of graphs, we
let
$m_{2}(H,\mathcal{L})\coloneqq\max\left\\{\frac{e_{J}}{v_{J}-2+1/m_{2}(\mathcal{L})}:J\subseteq
H,v_{J}\geqslant 2\right\\}.$
Third, given two finite families of graphs $\mathcal{H}$ and $\mathcal{L}$
with $m_{2}(\mathcal{H})\geqslant m_{2}(\mathcal{L})$, we define
$m_{2}(\mathcal{H},\mathcal{L})\coloneqq\min_{H\in\mathcal{H}}m_{2}(H,\mathcal{L}).$
Finally, continuing the terminology above, let us say that the pair
$(\mathcal{H},\mathcal{L})$ is _strictly balanced_ if every graph in
$\mathcal{L}$ is strictly $2$-balanced and every graph in $\mathcal{H}$ is
strictly $m_{2}(\cdot,\mathcal{L})$-balanced.
The following conjecture is a natural generalization of Conjecture 1.2 to
families of graphs.
###### Conjecture 1.9 (Kohayakawa–Kreuter conjecture for families).
Let $\mathcal{H}_{1},\dots,\mathcal{H}_{r}$ be finite families of graphs with
$m_{2}(\mathcal{H}_{1})\geqslant\dotsb\geqslant m_{2}(\mathcal{H}_{r})$ and
suppose that $m_{2}(\mathcal{H}_{2})>1$. There exist constants $c,C>0$ such
that
$\lim_{n\to\infty}\operatorname{Pr}(G_{n,p}\text{ is Ramsey for
}(\mathcal{H}_{1},\dots,\mathcal{H}_{r}))=\begin{cases}1&\text{if }p\geqslant
Cn^{-1/m_{2}(\mathcal{H}_{1},\mathcal{H}_{2})},\\\ 0&\text{if }p\leqslant
cn^{-1/m_{2}(\mathcal{H}_{1},\mathcal{H}_{2})}.\end{cases}$
Note that, for any $H_{1}\in\mathcal{H}_{1},\dots,H_{r}\in\mathcal{H}_{r}$,
the property of being Ramsey for $(H_{1},\dots,H_{r})$ implies the property of
being Ramsey for $(\mathcal{H}_{1},\dots,\mathcal{H}_{r})$. Therefore, the
$1$-statement of Conjecture 1.9 follows from the $1$-statement of Conjecture
1.2, which we know to be true by the result of Mousset, Nenadov, and Samotij
[20].
The $0$-statement of Conjecture 1.9 remains open; the only progress to date is
due to the first two authors [17], who proved Conjecture 1.9 whenever
$m_{2}(\mathcal{H}_{1})=m_{2}(\mathcal{H}_{2})$. We make further progress on
this conjecture: as in the case of single graphs, we prove a probabilistic
lemma that reduces the $0$-statement to a deterministic lemma, which is
clearly a necessary condition.
###### Theorem 1.10 (Probabilistic lemma for families).
The $0$-statement of Conjecture 1.9 holds if and only if, for every strictly
balanced pair $(\mathcal{H}_{1},\mathcal{H}_{2})$ of finite families of
graphs, every graph $G$ that is Ramsey for $(\mathcal{H}_{1},\mathcal{H}_{2})$
satisfies $m(G)>m_{2}(\mathcal{H}_{1},\mathcal{H}_{2})$.
As in Theorems 1.4 and 1.7, we can prove the deterministic lemma for families
in a wide variety of cases, namely when every graph $H_{1}\in\mathcal{H}_{1}$
or every graph $H_{2}\in\mathcal{H}_{2}$ satisfies one of the conditions in
Theorem 1.7. In particular, we resolve Conjecture 1.9 in many cases. However,
we believe that the right way to resolve Conjecture 1.9 in its entirety is the
same as the right way to resolve the original Kohayakawa–Kreuter conjecture,
Conjecture 1.2. Namely, if Conjecture 1.5 is true, then Conjecture 1.9 is true
for all families of graphs.
### 1.5. Organization
Most of the rest of this paper is dedicated to proving Theorem 1.10, and thus
also Theorem 1.3. Our technique is inspired by recent work of the first two
authors [17], who proved Conjecture 1.9 in the case
$m_{2}(\mathcal{H}_{1})=m_{2}(\mathcal{H}_{2})$. Therefore, we assume
henceforth that $m_{2}(\mathcal{H}_{1})>m_{2}(\mathcal{H}_{2})$. We will now
change notation and denote $\mathcal{H}_{1}=\mathcal{H}$ and
$\mathcal{H}_{2}=\mathcal{L}$. The names stand for _heavy_ and _light_ ,
respectively, and are meant to remind the reader that
$m_{2}(\mathcal{L})<m_{2}(\mathcal{H})$. We also assume henceforth that
$(\mathcal{H},\mathcal{L})$ is a strictly balanced pair of families.
The rest of this paper is organized as follows. In Section 2, we present a
high-level overview of our proof of Theorem 1.10. Section 3 contains a number
of preliminaries for the proof, including the definitions and basic properties
of _cores_ —a fundamental notion in our approach—as well as several simple
numerical lemmas. The proof of Theorem 1.10 is carried out in detail in
Section 4. In Section 5, we prove the deterministic lemma under various
assumptions, which yields Theorems 1.7 and 1.4 as well as their
generalizations to families. We conclude with two appendices: Appendix A
proves Theorem 1.6 by explaining what in our proof needs to be adapted to deal
with the three-color setting; and Appendix B presents our partial progress on
Conjecture 1.5.
#### Additional note
As this paper was being written, we learned that very similar results were
obtained independently by Bowtell, Hancock, and Hyde [2], who also resolve
Conjecture 1.2 in the vast majority of cases. As with this paper, they first
prove a probabilistic lemma, showing that resolving the Kohayakawa–Kreuter
conjecture is equivalent to proving a deterministic coloring result. By using
a wider array of coloring techniques, they are able to prove more cases of
Conjecture 1.2 than we can. Additionally, they consider a natural
generalization of the Kohayakawa–Kreuter to uniform hypergraphs (a topic that
we chose not to pursue here) and establish its $0$-statement for almost all
pairs of hypergraphs; see also [9] for more on such hypergraph questions. In
contrast, their work does not cover families of graphs, a generalization that
falls out naturally from our approach.
#### Acknowledgments
We would like to thank Anita Liebenau and Letícia Mattos for fruitful
discussions on Ramsey properties of random graphs. We are also indebted to
Candida Bowtell, Robert Hancock, and Joseph Hyde for sharing an early draft of
their paper [2] with us, and for their many invaluable comments.
## 2\. Proof outline
We now sketch, at a very high level, the proof of the probabilistic lemma. Let
us fix a strictly balanced pair of families $(\mathcal{H},\mathcal{L})$. We
wish to upper-bound the probability that $G_{n,p}$ is Ramsey for
$(\mathcal{H},\mathcal{L})$, where $p\leqslant
cn^{-1/m_{2}(\mathcal{H},\mathcal{L})}$ for an appropriately chosen constant
$c=c(\mathcal{H},\mathcal{L})>0$. Our approach is modeled on the recent proof
of the $0$-statement of Theorem 1.1 due to the first two authors [17];
however, there are substantial additional difficulties that arise in the
asymmetric setting.
One can immediately make several simplifying assumptions. First, if $G_{n,p}$
is Ramsey for $(\mathcal{H},\mathcal{L})$, then there exists some $G\subseteq
G_{n,p}$ that is _minimally_ Ramsey for $(\mathcal{H},\mathcal{L})$, in the
sense that any proper subgraph $G^{\prime}\subsetneq G$ is not Ramsey for
$(\mathcal{H},\mathcal{L})$. It is not hard to show (see Lemma 3.2 below) that
every minimally Ramsey graph has a number of interesting properties. In
particular, if $G$ is minimally Ramsey, then every edge of $G$ lies in at
least one copy of some $H\in\mathcal{H}$, and at least one copy of some
$L\in\mathcal{L}$. Our arguments will exploit a well-known strengthening of
this property, which we call _supporting a core_ ; see Definition 3.1 for the
precise definition.
We would ideally like to union-bound over all possible minimally Ramsey graphs
$G$ in order to show that a.a.s. none of them appears in $G_{n,p}$.
Unfortunately, there are potentially too many minimally Ramsey graphs for this
to be possible. To overcome this, we construct a smaller family $\mathcal{S}$
of subgraphs of $K_{n}$ such that every Ramsey graph $G$ contains some element
of $\mathcal{S}$ as a subgraph. Since $\mathcal{S}$ is much smaller than the
family of minimally Ramsey graphs, we can effectively union-bound over
$\mathcal{S}$. This basic idea also underlies the container method [28, 1] and
the recent work of Harel, Mousset, and Samotij on the upper tail problem for
subgraph counts [12]. The details here, however, are slightly subtle; there
are actually three different types of graphs in $\mathcal{S}$ and a different
union-bound argument is needed to handle each type.
We construct our family $\mathcal{S}$ with the use of an exploration process
on minimally Ramsey graphs, each of which supports a core. This exploration
process starts with a fixed edge of $K_{n}$ and gradually adds to it copies of
graphs in $\mathcal{H}\cup\mathcal{L}$. As long as the subgraph
$G^{\prime}\subseteq G$ of explored edges is not yet all of $G$, we add to
$G^{\prime}$ a copy of some graph in $\mathcal{H}\cup\mathcal{L}$ that
intersects $G^{\prime}$ but is not fully contained in it. By choosing this
copy in a principled manner (more on this momentarily), we can ensure that
$\mathcal{S}$ satisfies certain conditions which enable this union-bound
argument.
Since our goal is to show that the final graph $G^{\prime}$ is rather dense
(and thus unlikely to appear in $G_{n,p}$), we always prefer to add copies of
graphs in $\mathcal{H}$, as these boost the density of $G^{\prime}$. If there
are no available copies of $H\in\mathcal{H}$, we explore along some
$L\in\mathcal{L}$. As $L$ may be very sparse, this can hurt us; however, the
“core” property guarantees that each copy of $L$ comes with at least one copy
of some $H\in\mathcal{H}$ per new edge. An elementary (but fairly involved)
computation shows that the losses and the gains pencil out, which is the key
fact showing that $\mathcal{S}$ has the desired properties.
## 3\. Preliminaries
### 3.1. Ramsey graphs and cores
Given a graph $G$, denote by
$\mathcal{F}_{\mathcal{H}}[G],\mathcal{F}_{\mathcal{L}}[G]$ the set of all
copies of members of $\mathcal{H},\mathcal{L}$, respectively, in $G$. We think
of $\mathcal{F}_{\mathcal{H}}[G],\mathcal{F}_{\mathcal{L}}[G]$ as hypergraphs
on the ground set $E(G)$; in particular, we think of an element of
$\mathcal{F}_{\mathcal{H}}[G],\mathcal{F}_{\mathcal{L}}[G]$ as a collection of
edges of $G$ that form a copy of some $H\in\mathcal{H},L\in\mathcal{L}$,
respectively. To highlight the (important) difference between the members of
$\mathcal{H}\cup\mathcal{L}$ and their copies (i.e. the elements of
$\mathcal{F}_{\mathcal{H}}[G]\cup\mathcal{F}_{\mathcal{L}}[G]$), we will
denote the former by $H$ and $L$ and the latter by $\widehat{H}$ and
$\widehat{L}$.
Given a graph $G$ and
$\mathcal{F}_{\mathcal{H}}\subseteq\mathcal{F}_{\mathcal{H}}[G],\mathcal{F}_{\mathcal{L}}\subseteq\mathcal{F}_{\mathcal{L}}[G]$,
we say that the tuple
$(G,\mathcal{F}_{\mathcal{H}},\mathcal{F}_{\mathcal{L}})$ is _Ramsey_ if, for
every two-coloring of $E(G)$, there is an element of
$\mathcal{F}_{\mathcal{H}}$ that is monochromatic red or an element of
$\mathcal{F}_{\mathcal{L}}$ that is monochromatic blue. In particular, we see
that $G$ is Ramsey for $(\mathcal{H},\mathcal{L})$ if and only if
$(G,\mathcal{F}_{\mathcal{H}}[G],\mathcal{F}_{\mathcal{L}}[G])$ is Ramsey.
Having said that, allowing tuples
$(G,\mathcal{F}_{\mathcal{H}},\mathcal{F}_{\mathcal{L}})$ where
$\mathcal{F}_{\mathcal{H}}$ and $\mathcal{F}_{\mathcal{L}}$ are proper subsets
of $\mathcal{F}_{\mathcal{H}}[G]$ and $\mathcal{F}_{\mathcal{L}}[G]$,
respectively, enables us to deduce further useful properties. These are
encapsulated in the following definition.
###### Definition 3.1.
An _$(\mathcal{H},\mathcal{L})$ -core_ (or _core_ for short) is a tuple
$(G,\mathcal{F}_{\mathcal{H}},\mathcal{F}_{\mathcal{L}})$, where $G$ is a
graph and
$\mathcal{F}_{\mathcal{H}}\subseteq\mathcal{F}_{\mathcal{H}}[G],\mathcal{F}_{\mathcal{L}}\subseteq\mathcal{F}_{\mathcal{L}}[G]$,
with the following properties:
* •
The hypergraph $\mathcal{F}_{\mathcal{H}}\cup\mathcal{F}_{\mathcal{L}}$ is
connected and spans $E(G)$.
* •
For every $\widehat{H}\in\mathcal{F}_{\mathcal{H}}$ and every edge
$e\in\widehat{H}$, there exists an $\widehat{L}\in\mathcal{F}_{\mathcal{L}}$
such that $\widehat{H}\cap\widehat{L}=\\{e\\}$.
* •
For every $\widehat{L}\in\mathcal{F}_{\mathcal{L}}$ and every edge
$e\in\widehat{L}$, there exists an $\widehat{H}\in\mathcal{F}_{\mathcal{H}}$
such that $\widehat{H}\cap\widehat{L}=\\{e\\}$.
We say that $G$ _supports a core_ if there exist
$\mathcal{F}_{\mathcal{H}}\subseteq\mathcal{F}_{\mathcal{H}}[G],\mathcal{F}_{\mathcal{L}}\subseteq\mathcal{F}_{\mathcal{L}}[G]$
such that $(G,\mathcal{F}_{\mathcal{H}},\mathcal{F}_{\mathcal{L}})$ is a core.
The reason we care about cores is that minimal Ramsey graphs support cores, as
shown in the following lemma. Essentially the same lemma appears in the work
of Rödl and Ruciński [25], where it is given as an exercise. The same idea was
already used in several earlier works, including [15, Claim 6] and [18, Lemma
4.1].
###### Lemma 3.2.
Suppose that a graph $G$ is Ramsey for $(\mathcal{H},\mathcal{L})$, but none
of its proper subgraphs are Ramsey for $(\mathcal{H},\mathcal{L})$. Then $G$
supports an $(\mathcal{H},\mathcal{L})$-core.
###### Proof.
As $G$ is Ramsey for $(\mathcal{H},\mathcal{L})$, we know that
$(G,\mathcal{F}_{\mathcal{H}}[G],\mathcal{F}_{\mathcal{L}}[G])$ is a Ramsey
tuple. Let
$\mathcal{F}_{\mathcal{H}}\subseteq\mathcal{F}_{\mathcal{H}}[G],\mathcal{F}_{\mathcal{L}}\subseteq\mathcal{F}_{\mathcal{L}}[G]$
be inclusion-minimal subfamilies such that
$(G,\mathcal{F}_{\mathcal{H}},\mathcal{F}_{\mathcal{L}})$ is still a Ramsey
tuple. In other words, this tuple is Ramsey, but for any
$\mathcal{F}_{\mathcal{H}}^{\prime}\subseteq\mathcal{F}_{\mathcal{H}},\mathcal{F}_{\mathcal{L}}^{\prime}\subseteq\mathcal{F}_{\mathcal{L}}$
such that at least one inclusion is strict, the tuple
$(G,\mathcal{F}_{\mathcal{H}}^{\prime},\mathcal{F}_{\mathcal{L}}^{\prime})$ is
not Ramsey. We will show that
$(G,\mathcal{F}_{\mathcal{H}},\mathcal{F}_{\mathcal{L}})$ is a core.
If some $e\in E(G)$ is not contained in any edge of
$\mathcal{F}_{\mathcal{H}}\cup\mathcal{F}_{\mathcal{L}}$, then $(G\setminus
e,\mathcal{F}_{\mathcal{H}},\mathcal{F}_{\mathcal{L}})$ is still Ramsey, and
thus $G\setminus e$ is Ramsey for $(\mathcal{H},\mathcal{L})$, contradicting
the minimality of $G$. Furthermore, if
$\mathcal{F}_{\mathcal{H}}\cup\mathcal{F}_{\mathcal{L}}$ is not connected,
then at least one of its connected components induces a Ramsey tuple, which
contradicts the minimality of
$(\mathcal{F}_{\mathcal{H}},\mathcal{F}_{\mathcal{L}})$. Thus, the first
condition in the definition of a core is satisfied. We now turn to the next
two parts of the definition.
To see that the second condition in the definition of a core is satisfied, fix
some $\widehat{H}\in\mathcal{F}_{\mathcal{H}}$ and some $e\in\widehat{H}$. By
minimality, we can find a two-coloring of $E(G)$ such that no element of
$\mathcal{F}_{\mathcal{L}}$ is blue and no element of
$\mathcal{F}_{\mathcal{H}}\setminus\\{\widehat{H}\\}$ is red. Note that all
edges of $\widehat{H}$ are colored red, as otherwise our coloring would
witness $(G,\mathcal{F}_{\mathcal{H}},\mathcal{F}_{\mathcal{L}})$ being not
Ramsey. Flip the color of $e$ from red to blue. Since $\widehat{H}$ is now no
longer monochromatic red, we must have created a monochromatic blue element
$\widehat{L}$ of $\mathcal{F}_{\mathcal{L}}$. As all edges of
$\widehat{H}\setminus e$ are still red, we see that
$\widehat{H}\cap\widehat{L}=\\{e\\}$, as required. Interchanging the roles of
$\mathcal{F}_{\mathcal{H}},\mathcal{F}_{\mathcal{L}}$, and the colors yields
the third condition in the definition of a core. ∎
### 3.2. Numerical lemmas
In this section, we collect a few useful numerical lemmas, all of which are
simple combinatorial facts about vertex- and edge-counts in graphs. We begin
with the following well-known result, which we will use throughout.
###### Lemma 3.3 (The mediant inequality).
Let $a,c\geqslant 0$ and $b,d>0$ be real numbers with $a/b\leqslant c/d$. Then
$\frac{a}{b}\leqslant\frac{a+c}{b+d}\leqslant\frac{c}{d}.$
Moreover, if one inequality is strict, then so is the other (which happens if
and only if $a/b<c/d$).
###### Proof.
Both inequalities are easily seen to be equivalent to the inequality
$ad\leqslant bc$, which is itself the same as $a/b\leqslant c/d$. ∎
###### Lemma 3.4.
Let $(\mathcal{H},\mathcal{L})$ be a strictly balanced pair. If
$m_{2}(\mathcal{L})<m_{2}(\mathcal{H})$, then
$m_{2}(\mathcal{L})<m_{2}(\mathcal{H},\mathcal{L})<m_{2}(\mathcal{H})$.
###### Proof.
To see the second inequality, let $H\in\mathcal{H}$ be a graph with
$m_{2}(H)=m_{2}(\mathcal{H})$ and observe that the strict
$m_{2}(\cdot,\mathcal{L})$-balancedness of $H$ implies that
$m_{2}(H,\mathcal{L})=\frac{e_{H}}{v_{H}-2+1/m_{2}(\mathcal{L})}=\frac{(e_{H}-1)+1}{(v_{H}-2)+1/m_{2}(\mathcal{L})}\leqslant\frac{m_{2}(H)\cdot(v_{H}-2)+1}{(v_{H}-2)+1/m_{2}(\mathcal{L})}.$
Since $m_{2}(H)=m_{2}(\mathcal{H})>m_{2}(\mathcal{L})$, Lemma 3.3 implies that
$m_{2}(\mathcal{H},\mathcal{L})\leqslant
m_{2}(H,\mathcal{L})<m_{2}(\mathcal{H})$.
For the first inequality, let $H\in\mathcal{H}$ be a graph for which
$m_{2}(H,\mathcal{L})=m_{2}(\mathcal{H},\mathcal{L})$ and let $J\subseteq H$
be its subgraph with $\frac{e_{J}-1}{v_{J}-2}=m_{2}(H)$. By the strict
$m_{2}(\cdot,\mathcal{L})$-balancedness of $H$, we have
$m_{2}(H,\mathcal{L})\geqslant
m_{2}(J,\mathcal{L})=\frac{(e_{J}-1)+1}{(v_{J}-2)+1/m_{2}(\mathcal{L})}=\frac{m_{2}(H)\cdot(v_{J}-2)+1}{(v_{J}-2)+1/m_{2}(\mathcal{L})}.$
Since $m_{2}(H)>m_{2}(\mathcal{L})$, Lemma 3.3 implies that
$m_{2}(\mathcal{H},\mathcal{L})=m_{2}(H,\mathcal{L})\geqslant
m_{2}(J,\mathcal{L})>m_{2}(\mathcal{L})$. ∎
###### Lemma 3.5.
Let $H\in\mathcal{H}$ be strictly $m_{2}(\cdot,\mathcal{L})$-balanced. Then
for any $F\subsetneq H$ with $v_{F}\geqslant 2$, we have
$e_{H}-e_{F}>m_{2}(H,\mathcal{L})\cdot(v_{H}-v_{F})\geqslant
m_{2}(\mathcal{H},\mathcal{L})\cdot(v_{H}-v_{F}).$
###### Proof.
The second inequality follows from the definition of
$m_{2}(\mathcal{H},\mathcal{L})$. Since $e_{F}<e_{H}$, we may assume that
$v_{F}<v_{H}$, as otherwise the claimed inequality holds vacuously. Since $H$
is strictly $m_{2}(\cdot,\mathcal{L})$-balanced, we have
$m_{2}(H,\mathcal{L})=\frac{e_{H}}{v_{H}-2+1/m_{2}(\mathcal{L})}=\frac{(e_{H}-e_{F})+e_{F}}{(v_{H}-v_{F})+(v_{F}-2+1/m_{2}(\mathcal{L}))}$
whereas
$\frac{e_{F}}{v_{F}-2+1/m_{2}(\mathcal{L})}<m_{2}(H,\mathcal{L}).$
Since $v_{H}>v_{F}$, we may use Lemma 3.3 to conclude that
$(e_{H}-e_{F})/(v_{H}-v_{F})>m_{2}(H,\mathcal{L})$. ∎
###### Lemma 3.6.
Let $L\in\mathcal{L}$ be strictly 2-balanced. Then for any $J\subsetneq L$
with $e_{L}\geqslant 1$, we have
$e_{L}-e_{J}\geqslant m_{2}(L)\cdot(v_{L}-v_{J})\geqslant
m_{2}(\mathcal{L})\cdot(v_{L}-v_{J}).$
Moreover, the first inequality is strict unless $J=K_{2}$.
###### Proof.
The second inequality is immediate since $m_{2}(\mathcal{L})\leqslant
m_{2}(L)$. Since $e_{J}<e_{L}$, we may assume that $v_{J}<v_{L}$, as otherwise
the claimed (strict) inequality holds vacuously. We clearly have equality if
$J=K_{2}$ and strict inequality if $v_{J}=2$ and $e_{J}=0$, so we may assume
henceforth that $v_{J}>2$. Since $L$ is strictly $2$-balanced,
$m_{2}(L)=\frac{e_{L}-1}{v_{L}-2}=\frac{(e_{L}-e_{J})+(e_{J}-1)}{(v_{L}-v_{J})+(v_{J}-2)}$
whereas $(e_{J}-1)/(v_{J}-2)<m_{2}(L)$. Since $v_{J}>2$, we may apply Lemma
3.3 to conclude the desired result, with a strict inequality. ∎
###### Lemma 3.7.
Suppose that $(\mathcal{H},\mathcal{L})$ is a strictly balanced pair. Defining
$\alpha\coloneqq m_{2}(\mathcal{H},\mathcal{L})$ and
$X\coloneqq\min_{H\in\mathcal{H}}\\{(e_{H}-1)-\alpha\cdot(v_{H}-2)\\}$, we
have that
$X+(v_{K}-2)(\alpha-1)\geqslant
e_{K}\cdot\left(\frac{\alpha}{m_{2}(L)}-1\right)$
for every $L\in\mathcal{L}$ and every non-empty $K\subseteq L$. Moreover, the
inequality is strict unless $K=K_{2}$.
###### Proof.
Without loss of generality, we may assume that $m_{2}(L)<\alpha$ and that
$v_{K}>2$, as otherwise the statement holds vacuously (recall from Lemma 3.4
that $\alpha=m_{2}(\mathcal{H},\mathcal{L})>m_{2}(\mathcal{L})>1$). Fix some
$L\in\mathcal{L}$ and a nonempty $K\subseteq L$. Recall that each
$H\in\mathcal{H}$ is strictly $m_{2}(\cdot,\mathcal{L})$-balanced and
satisfies $m_{2}(H,\mathcal{L})\geqslant
m_{2}(\mathcal{H},\mathcal{L})=\alpha$. This implies that
$\frac{e_{H}}{v_{H}-2+1/m_{2}(\mathcal{L})}\geqslant\alpha$
or, equivalently,
$e_{H}\geqslant\alpha\cdot(v_{H}-2)+\frac{\alpha}{m_{2}(\mathcal{L})}.$
Consequently,
$X=\min_{H\in\mathcal{H}}\\{(e_{H}-1)-\alpha\cdot(v_{H}-2)\\}\geqslant\frac{\alpha}{m_{2}(\mathcal{L})}-1\geqslant\frac{\alpha}{m_{2}(L)}-1,$
where the final inequality uses that $m_{2}(L)\geqslant m_{2}(\mathcal{L})$.
Since $L$ is strictly $2$-balanced and we assumed that $m_{2}(L)<\alpha$, we
have
$(e_{K}-1)\cdot\left(\frac{\alpha}{m_{2}(L)}-1\right)\leqslant
m_{2}(L)\cdot(v_{K}-2)\cdot\left(\frac{\alpha}{m_{2}(L)}-1\right)=(v_{K}-2)(\alpha-
m_{2}(L)).$
Rearranging the above inequality, we obtain
$\displaystyle
e_{K}\cdot\left(\frac{\alpha}{m_{2}(L)}-1\right)-(v_{K}-2)(\alpha-1)$
$\displaystyle\leqslant(1-m_{2}(L))(v_{K}-2)+\left(\frac{\alpha}{m_{2}(L)}-1\right)$
$\displaystyle<\frac{\alpha}{m_{2}(L)}-1\leqslant X,$
where the penultimate inequality uses the assumption that $v_{K}>2$. ∎
## 4\. Proof of the probabilistic lemma
In this section, we prove Theorem 1.10. We in fact prove the following more
precise statement.
###### Lemma 4.1 (Theorem 1.10, rephrased).
Let $(\mathcal{H},\mathcal{L})$ be a strictly balanced pair of finite families
of graphs satisfying $m_{2}(\mathcal{H})>m_{2}(\mathcal{L})$. There exists a
constant $c>0$ such that the following holds. If $p\leqslant
cn^{-1/m_{2}(\mathcal{H},\mathcal{L})}$, then a.a.s. every $G\subseteq
G_{n,p}$ which supports a core satisfies $m(G)\leqslant
m_{2}(\mathcal{H},\mathcal{L})$.
Note that this immediately implies the difficult direction in Theorem 1.10.
Indeed, suppose that the $0$-statement of 1.9 fails for some tuple
$(\mathcal{H}_{1},\dotsc,\mathcal{H}_{r})$, i.e., the random graph $G_{n,p}$
is Ramsey for $(\mathcal{H}_{1},\dotsc,\mathcal{H}_{r})$ with probability
bounded away from zero when
$p=cn^{-1/m_{2}(\mathcal{H}_{1},\mathcal{H}_{2})}$, for an arbitrarily small
constant $c>0$. In particular, with probability bounded away from zero,
$G_{n,p}$ contains a graph that is also Ramsey for any pair
$(\mathcal{H},\mathcal{L})$ of families of subgraphs of
$(\mathcal{H}_{1},\mathcal{H}_{2})$. For an appropriately chosen pair
$(\mathcal{H},\mathcal{L})$, Lemma 3.2 implies that some subgraph $G\subseteq
G_{n,p}$ supports an $(\mathcal{H},\mathcal{L})$-core. By the assumed
assertion of Lemma 4.1, a.a.s. any such $G\subseteq G_{n,p}$ satisfies
$m(G)\leqslant m_{2}(\mathcal{H},\mathcal{L})$. However, by the deterministic
lemma (i.e. the assumption of Theorem 1.10), we know that no such $G$ can be
Ramsey for $(\mathcal{H},\mathcal{L})$, a contradiction.
Our proof of Lemma 4.1 follows closely the proof of the probabilistic lemma in
recent work of the first two authors [17]. Fix a strictly balanced pair
$(\mathcal{H},\mathcal{L})$ of families satisfying
$m_{2}(\mathcal{H})>m_{2}(\mathcal{L})$, and let $\alpha\coloneqq
m_{2}(\mathcal{H},\mathcal{L})$. Let $\mathcal{G}_{\mathrm{bad}}$ denote the
set of graphs $G\subseteq K_{n}$ which support a core and satisfy
$m(G)>m_{2}(\mathcal{H},\mathcal{L})$. The key lemma, which implies Lemma 4.1,
is as follows.
###### Lemma 4.2.
There exist constants $\Lambda,K>0$ and a collection $\mathcal{S}$ of
subgraphs of $K_{n}$ satisfying the following properties:
1. (a)
Every element of $\mathcal{G}_{\mathrm{bad}}$ contains some $S\in\mathcal{S}$
as a subgraph.
2. (b)
Every $S\in\mathcal{S}$ satisfies at least one of the following three
conditions:
1. (i)
$v_{S}\geqslant\log n$ and $e_{S}\geqslant\alpha\cdot(v_{S}-2)$;
2. (ii)
$v_{S}<\log n$ and $e_{S}\geqslant\alpha\cdot v_{S}+1$;
3. (iii)
$v_{S}\leqslant K$ and $m(S)>\alpha$.
3. (c)
For every $k\in\llbracket{n}\rrbracket$, there are at most $(\Lambda n)^{k}$
graphs $S\in\mathcal{S}$ with $v_{S}=k$.
Before we prove Lemma 4.2, let us see why it implies Lemma 4.1.
###### Proof of Lemma 4.1.
Recall that $p\leqslant cn^{-1/\alpha}$, for a small constant
$c=c(\mathcal{H},\mathcal{L})$ to be chosen later. We wish to prove that
a.a.s. $G_{n,p}$ contains no element of $\mathcal{G}_{\mathrm{bad}}$. By Lemma
4.2(a), it suffices to prove that a.a.s. $G_{n,p}$ contains no element of
$\mathcal{S}$. By (b), the elements of $\mathcal{S}$ are of three types, each
of which we deal with separately. First, recall that for any fixed graph $S$
with $m(S)>\alpha$, we have that $\operatorname{Pr}(S\subseteq G_{n,p})=o(1)$
(see e.g. [14, Theorem 3.4]). As there are only a constant number of graphs on
at most $K$ vertices, we may apply the union bound and conclude that a.a.s. no
graph $S$ satisfying $v_{S}\leqslant K$ and $m(S)>\alpha$ appears in
$G_{n,p}$. This deals with the elements of $\mathcal{S}$ corresponding to case
(b)(iii).
Let $\mathcal{S}^{\prime}\subseteq\mathcal{S}$ be the set of $S\in\mathcal{S}$
which lie in cases (b)(i) or (b)(ii). We have that
$\displaystyle\operatorname{Pr}(S\subseteq G_{n,p}\text{ for some
}S\in\mathcal{S}^{\prime})$
$\displaystyle\leqslant\sum_{S\in\mathcal{S}^{\prime}}p^{e_{S}}$
$\displaystyle\leqslant\sum_{k=1}^{\lceil\log n\rceil-1}(\Lambda
n)^{k}p^{\alpha k+1}+\sum_{k=\lceil\log n\rceil}^{\infty}(\Lambda
n)^{k}p^{\alpha(k-2)}$ $\displaystyle\leqslant p\sum_{k=1}^{\infty}(\Lambda
c^{\alpha})^{k}+c^{-2\alpha}n^{2}\sum_{k=\lceil\log n\rceil}^{\infty}(\Lambda
c^{\alpha})^{k}$
We now choose $c$ so that $\Lambda c^{\alpha}=e^{-3}$. Then the first sum
above can be bounded by $p$, which tends to $0$ as $n\to\infty$. The second
term can be bounded by $2c^{-2\alpha}n^{-1}$, which also tends to $0$ as
$n\to\infty$. All in all, we find that a.a.s. $G_{n,p}$ does not contain any
graph in $\mathcal{S}$, as claimed. ∎
### 4.1. The exploration process and the proof of Lemma 4.2
In this section, we prove Lemma 4.2. We will construct the family
$\mathcal{S}$ by considering an exploration process on the set $\mathcal{G}$
of graphs $G\subseteq K_{n}$ which support a core. For each such
$G\in\mathcal{G}$, let us arbitrarily choose collections
$\mathcal{F}_{\mathcal{H}}\subseteq\mathcal{F}_{\mathcal{H}}[G]$ and
$\mathcal{F}_{\mathcal{L}}\subseteq\mathcal{F}_{\mathcal{L}}[G]$ such that
$(G,\mathcal{F}_{\mathcal{H}},\mathcal{F}_{\mathcal{L}})$ is a core. From now
on, by copies of graphs from $\mathcal{H},\mathcal{L}$ in $G$, we mean only
those copies that belong to the families
$\mathcal{F}_{\mathcal{H}},\mathcal{F}_{\mathcal{L}}$, respectively. This
subtlety will be extremely important in parts of the analysis.
We first fix arbitrary orderings on the graphs in $\mathcal{H}$ and
$\mathcal{L}$. Additionally, we fix a labeling of the vertices of $K_{n}$,
which induces an ordering of all subgraphs according to the lexicographic
order. Together with the ordering on $\mathcal{H},\mathcal{L}$, we obtain a
lexicographic ordering on all copies in $K_{n}$ of graphs in
$\mathcal{H},\mathcal{L}$. Now, given a $G\in\mathcal{G}$, we build a sequence
$G_{0}\subsetneq G_{1}\subsetneq\dotsb\subseteq G$ as follows. We start with
$G_{0}$ being the graph comprising only the smallest edge of $G$. As long as
$G_{i}\neq G$, do the following: Since $G\neq G_{i}$ and
$(G,\mathcal{F}_{\mathcal{H}},\mathcal{F}_{\mathcal{L}})$ is a core, there
must be some copy of a graph from $\mathcal{H}\cup\mathcal{L}$ which belongs
to $\mathcal{F}_{\mathcal{H}}\cup\mathcal{F}_{\mathcal{L}}$ that intersects
$G_{i}$ but is not fully contained in $G_{i}$. Call such an _overlapping_ copy
_regular_ if it intersects $G_{i}$ in exactly one edge, called its _root_ ;
otherwise, call the copy _degenerate_. We form $G_{i+1}$ from $G_{i}$ as
follows:
1. (1)
Suppose first that there is an overlapping copy of some graph in
$\mathcal{H}$. We form $G_{i+1}$ by adding to $G_{i}$ the smallest (according
to the lexicographic order) such copy. We call $G_{i}\to G_{i+1}$ a
_degenerate $\mathcal{H}$-step_.
2. (2)
Otherwise, there must be an overlapping copy $\widehat{L}$ of some
$L\in\mathcal{L}$. Note that, for every edge $e\in\widehat{L}\setminus G_{i}$,
there must be a copy of some $H\in\mathcal{H}$ that meets $\widehat{L}$ only
at $e$, as $(G,\mathcal{F}_{\mathcal{H}},\mathcal{F}_{\mathcal{L}})$ is a
core. Note further that this copy of $H$ does not intersect $G_{i}$, as
otherwise we would perform a degenerate $\mathcal{H}$-step. We pick the
smallest such copy for every $e\in\widehat{L}\setminus G_{i}$, and call it
$\widehat{H_{e}}$ (note that the graphs $H_{e}\in\mathcal{H}$ such that
$H_{e}\cong\widehat{H_{e}}$ may be different for different choices of $e$). We
say that $\widehat{L}$ is _pristine_ if it is regular and the graphs
$\\{\widehat{H_{e}}\\}_{e\in\widehat{L}\setminus G_{i}}$ are all vertex-
disjoint (apart from the intersections that they are forced to have in
$V(\widehat{L})$).
1. (2.1)
If there is a pristine copy of some graph in $\mathcal{L}$, we pick the
smallest one in the following sense: First, among all edges of $G_{i}$ that
are roots of a pristine copy of some graph in $\mathcal{L}$, we choose the one
that arrived to $G_{i}$ earliest. Second, among all pristine copies that are
rooted at this edge, we pick the smallest (according to the lexicographic
order). We then form $G_{i+1}$ by adding to $G_{i}$ this smallest copy
$\widehat{L}$ as well as all $\widehat{H_{e}}$ where $e\in\widehat{L}\setminus
G_{i}$. We call $G_{i}\to G_{i+1}$ a _pristine step_.
2. (2.2)
If there are no pristine copies of any graph in $\mathcal{L}$, we pick the
smallest (according to the lexicographic order) overlapping copy $\widehat{L}$
of a graph in $\mathcal{L}$ and we still form $G_{i+1}$ by adding to $G_{i}$
the union of $\widehat{L}$ and all its $\widehat{H_{e}}$ with
$e\in\widehat{L}\setminus G_{i}$. We call $G_{i}\to G_{i+1}$ a _degenerate
$\mathcal{L}$-step_.
We define the _balance_ of $G_{i}$ to be
$b(G_{i})\coloneqq e_{G_{i}}-\alpha\cdot v_{G_{i}},$
where we recall that $\alpha=m_{2}(\mathcal{H},\mathcal{L})$. The key result
we will need in order to prove (b) is the following lemma. We remark that a
similar result was proved by Hyde [13, Claims 6.2 and 6.3]; it plays an
integral role in his approach to the Kohayakawa–Kreuter conjecture.
###### Lemma 4.3.
For every $i$, we have that $b(G_{i+1})\geqslant b(G_{i})$. Moreover, there
exists some $\delta=\delta(\mathcal{H},\mathcal{L})>0$ such that
$b(G_{i+1})\geqslant b(G_{i})+\delta$ if $G_{i+1}$ was obtained from $G_{i}$
by a degenerate step.
As the proof of Lemma 4.3 is somewhat technical, we defer it to Section 4.2.
For the moment, we assume the result and continue the discussion of how we
construct the family $\mathcal{S}$. We now let $\Gamma\coloneqq\lceil
2\alpha/\delta\rceil$, where $\delta$ is the constant from Lemma 4.3. For
$G\in\mathcal{G}$, let
$\tau(G)\coloneqq\min\\{i:v_{G_{i}}\geqslant\log n\text{ or }G_{i}=G\text{ or
}G_{i-1}\to G_{i}\text{ is the $\Gamma$th degenerate step}\\}$
and let
$\mathcal{S}\coloneqq\\{G_{\tau(G)}:G\in\mathcal{G}_{\mathrm{bad}}\\}.$ (1)
Having defined the family $\mathcal{S}$, we are ready to prove Lemma 4.2.
Since the definition of $\mathcal{S}$ clearly guarantees property (a), it
remains to establish properties (b) and (c). We begin by showing that, if $K$
is sufficiently large (depending only on $\mathcal{H}$ and $\mathcal{L}$),
then (b) holds.
###### Proof of Lemma 4.2(b).
Let $\delta$ be the constant from Lemma 4.3, let $M\coloneqq\max\\{e_{L}\cdot
v_{H}:H\in\mathcal{H},L\in\mathcal{L}\\}$, and let $K\coloneqq 2M^{2}\Gamma$;
note that each of these parameters depends only on $\mathcal{H}$ and
$\mathcal{L}$.
Every $S\in\mathcal{S}$ is of the form $G_{\tau(G)}$ for some
$G\in\mathcal{G}_{\mathrm{bad}}$. We split into cases depending on which of
the three conditions defining $\tau(G)$ caused us to stop the exploration.
Suppose first that we stopped the exploration because $v_{S}\geqslant\log n$.
By Lemma 4.3, we have that
$e_{S}-\alpha\cdot v_{S}=b(S)=b(G_{\tau(G)})\geqslant b(G_{0})=1-2\alpha,$
and therefore $e_{S}\geqslant\alpha\cdot(v_{S}-2)$. This yields case (b)(i).
Next, suppose we stopped the exploration because step $G_{\tau(G)-1}\to
G_{\tau(G)}$ was the $\Gamma$th degenerate step. As we are not in the previous
case, we may assume that $v_{S}<\log n$. By Lemma 4.3 and our choice of
$\Gamma$, we have that
$e_{S}-\alpha\cdot v_{S}=b(S)=b(G_{\tau(G)})\geqslant
b(G_{0})+\Gamma\delta\geqslant 1-2\alpha+2\alpha=1.$
Rearranging, we see that $e_{S}\geqslant\alpha\cdot v_{S}+1$, yielding case
(b)(ii).
The remaining case is when we stop because $S=G\in\mathcal{G}_{\mathrm{bad}}$.
Since the definition of $\mathcal{G}_{\mathrm{bad}}$ implies that
$m(G)>\alpha$, in order to establish (b)(iii), we only need to show that
$v_{G}\leqslant K$. For this proof, we need to keep track of another parameter
during the exploration process, which we term the _pristine boundary_. Recall
that at every pristine step, we add to $G_{i}$ a copy $\widehat{L}$ of some
$L\in\mathcal{L}$ that intersects $G_{i}$ in a single edge (the root), and
then add copies $\widehat{H_{e}}$ of graphs $H_{e}\in\mathcal{H}$, one for
every edge of $\widehat{L}$ apart from the root. Let us say that the
_boundary_ of this step is the set of all newly added vertices that are not in
$\widehat{L}$, that is, the set $V(G_{i+1})\setminus(V(G_{i})\cup
V(\widehat{L}))=(\bigcup_{e\in\widehat{L}\setminus
G_{i}}V(\widehat{H_{e}}))\setminus V(\widehat{L})$. Note that the size of the
boundary is equal to
$Y_{i}\coloneqq\sum_{e\in\widehat{L}\setminus G_{i}}(v_{H_{e}}-2);$
indeed, by the definition of pristine steps, the copies $\widehat{H_{e}}$ are
vertex-disjoint outside of $V(\widehat{L})$.
We claim that $Y_{i}\geqslant 3$. To see this, note first that $L$ has at
least three edges, as it is not a forest. Similarly, each $H_{e}$ has at least
three vertices. Putting these together, we see that there are at least two
terms in the sum, and every term in the sum is at least one. Thus,
$Y_{i}\geqslant 3$ unless $e_{L}=3$ and $v_{H_{e}}=3$ for all $e$. But in this
case, $L=K_{3}=H_{e}\in\mathcal{H}$ for all $e$, which means that
$\widehat{L}$ should have been added to $G_{i}$ as a degenerate
$\mathcal{H}$-step.
We now inductively define the pristine boundary $\partial G_{i}$ of $G_{i}$ as
follows. We set $\partial G_{0}\coloneqq\varnothing$. If $G_{i}\to G_{i+1}$ is
a pristine step, then we delete from $\partial G_{i}$ the two endpoints of the
root and add to $\partial G_{i}$ the boundary of this pristine step. Note that
$\lvert\partial G_{i+1}\rvert\geqslant\lvert\partial
G_{i}\rvert+Y_{i}-2\geqslant\lvert\partial G_{i}\rvert+1$. On the other hand,
if $G_{i}\to G_{i+1}$ is a degenerate step, then we only remove vertices from
$\partial G_{i}$, without adding any new vertices. Namely, we remove from
$\partial G_{i}$ all the vertices which are included in the newly added
graphs. In other words, if we performed a degenerate $\mathcal{H}$-step by
adding a copy $\widehat{H}$ of some graph in $\mathcal{H}$, we set $\partial
G_{i+1}\coloneqq\partial G_{i}\setminus V(\widehat{H})$. Similarly, if we
performed a degenerate $\mathcal{L}$-step by adding a copy $\widehat{L}$ of
some graph in $\mathcal{L}$ along with the graphs $\widehat{H_{e}}$ for all
$e\in\widehat{L}\setminus G_{i}$, we set $\partial G_{i+1}\coloneqq\partial
G_{i}\setminus(V(\widehat{L})\cup\bigcup_{e}V(\widehat{H_{e}}))$. Note that in
either case $\lvert\partial G_{i+1}\rvert\geqslant\lvert\partial
G_{i}\rvert-M$, as the union of all graphs added in each degenerate step can
have at most $M$ vertices.
We now argue that $\partial G_{\tau(G)}=\varnothing$. Indeed, suppose we had
some vertex $v\in\partial G_{\tau(G)}$. By definition, $v$ was added during a
pristine step, as a vertex of a copy $\widehat{H_{e}}$ of some graph
$H_{e}\in\mathcal{H}$, and was never touched again. Observe that $v$ is
incident to some edge $uv$ of $\widehat{H_{e}}$ that was not touched by any
later step of the exploration. However, as
$(G,\mathcal{F}_{\mathcal{H}},\mathcal{F}_{\mathcal{L}})$ is a core and
$\widehat{H_{e}}\in\mathcal{F}_{\mathcal{H}}$, there must be some
$\widehat{L_{uv}}\in\mathcal{F}_{\mathcal{L}}$ that intersects
$\widehat{H_{e}}$ only at $uv$. Moreover, as $\widehat{L_{uv}}$ has minimum
degree at least two (by the strict $2$-balancedness assumption), there is some
edge $vw\in\widehat{L_{uv}}\setminus uv$ that is incident to $v$. Since we
assumed that $G_{\tau(G)}=G$, the edge $vw$ must have been added at some
point, a contradiction to the assumption that $v$ was never touched again.
Finally, since $\lvert\partial G_{i}\rvert$ increases by at least one during
every pristine step and decreases by at most $M$ during each of the at most
$\Gamma$ degenerate steps, in order to achieve $\partial
G_{\tau(G)}=\varnothing$, there can be at most $M\Gamma$ pristine steps. In
particular, the total number of exploration steps is at most $M\Gamma+\Gamma$.
As each exploration step adds at most $M$ vertices to $G_{i}$, we conclude
that $v_{G}\leqslant M(M\Gamma+\Gamma)+2\leqslant K$. This completes the proof
of (b)(iii). ∎
###### Proof of Lemma 4.2(c).
Suppose $S$ has $k$ vertices and let $G\in\mathcal{G}_{\mathrm{bad}}$ be such
that $S=G_{\tau(G)}$. We consider the exploration process on $G$. Note that in
every step we add an overlapping copy of a graph from a finite family
$\mathcal{F}$ that comprises all graphs in $\mathcal{H}$ (for the cases where
we made a degenerate $\mathcal{H}$-step) and graphs in $\mathcal{L}$ that have
graphs from $\mathcal{H}$ glued on subsets of their edges, with all
intersection patterns (for the pristine and degenerate $\mathcal{L}$-steps).
Let $\mathcal{F}^{\times}$ denote the graphs in $\mathcal{F}$ that correspond
to a pristine step.
Now, every degenerate step can be described by specifying the graph
$F\in\mathcal{F}$ whose copy $\widehat{F}$ we are adding, the subgraph
$F^{\prime}\subseteq F$ and the embedding $\varphi\colon V(F^{\prime})\to
V(G_{i})$ that describe the intersection $\widehat{F}\cap G_{i}$, and the
sequence of $v_{F}-v_{F^{\prime}}$ vertices of $K_{n}$ that complete $\varphi$
to an embedding of $F$ into $K_{n}$. Every pristine step is uniquely described
by the root edge in $G_{i}$, the graph $F\in\mathcal{F}^{\times}$, the edge of
$F$ corresponding to the root, and the (ordered sequence of) $v_{F}-2$
vertices of $K_{n}$ that complete the root edge to a copy of $F$ in $K_{n}$.
There are at most $n^{k}$ ways to choose the sequence of vertices that were
added through this exploration process, in the order that they are introduced
to $G$. Each pristine step adds at least one new vertex, so there are at most
$k$ pristine steps. Furthermore, there are always at most $\Gamma$ degenerate
steps, meaning that $\tau(G)\leqslant k+\Gamma$. In particular, there are at
most $(k+\Gamma)\cdot 2^{k+\Gamma}$ ways to choose $\tau(G)$ and to specify
which steps were pristine.
For every degenerate step, there are at most
$\sum_{F\in\mathcal{F}}\sum_{\ell=2}^{v_{F}}\binom{v_{F}}{\ell}k^{\ell}\leqslant\lvert\mathcal{F}\rvert\cdot(k+1)^{M_{v}}$
ways of choosing $F\in\mathcal{F}$ and describing the intersection of its copy
$\widehat{F}$ with $G_{i}$ (the set $V(F^{\prime})\subseteq V(F)$ and the
embedding $\varphi$ above), where $M_{v}\coloneqq\max\\{v_{F}\colon
F\in\mathcal{F}\\}$. As for the pristine steps, note that, in the course of
our exploration, the sequence of the arrival times of the roots to
$G_{\tau(G)}$ must be non-decreasing. This is because as soon as an edge
appears in some $G_{i}$, every pristine step that includes it as a root at any
later step is already available, and we always choose the one rooted at the
edge that arrived to $G$ the earliest. Therefore, there are at most
$\binom{e_{S}+k}{k}$ possible sequences of root edges, since this is the
number of non-decreasing sequences of length $k$ in
$\llbracket{e_{S}}\rrbracket$. To supplement this bound, remember that every
step increases the number of edges in $G_{i}$ by at most
$M_{e}\coloneqq\max\\{e_{F}:F\in\mathcal{F}\\}$, which means that
$e_{S}\leqslant 1+\tau(G)\cdot M_{e}\leqslant 1+(k+\Gamma)\cdot M_{e}.$
To summarize, the number of $S\in\mathcal{S}$ with $k$ vertices is at most
$n^{k}\cdot(k+\Gamma)\cdot
2^{k+\Gamma}\cdot\left(\lvert\mathcal{F}\rvert\cdot(k+1)^{M_{v}}\right)^{\Gamma}\cdot\binom{(k+\Gamma)\cdot
M_{e}+k+1}{k}\cdot\left(\lvert\mathcal{F}\rvert\cdot M_{e}\right)^{k}.$
Every term in this product, apart from the first, is bounded by an exponential
function of $k$, since $\Gamma,\lvert\mathcal{F}\rvert,M_{v}$, and $M_{e}$ are
all constants. Therefore, if we choose
$\Lambda=\Lambda(\mathcal{H},\mathcal{L})$ sufficiently large, we find that
the number of $S\in\mathcal{S}$ with $v_{S}=k$ is at most $(\Lambda n)^{k}$,
as claimed. ∎
### 4.2. Proof of Lemma 4.3
In this section, we prove Lemma 4.3. The proof is divided into a number of
claims. Recall Lemma 3.5, which asserts that
$e_{H}-e_{F}>m_{2}(\mathcal{H},\mathcal{L})\cdot(v_{H}-v_{F})=\alpha\cdot(v_{H}-v_{F})$
for all $H\in\mathcal{H}$ and all $F\subsetneq H$. This implies that we can
choose some $\delta_{1}=\delta_{1}(\mathcal{H},\mathcal{L})>0$ so that
$e_{H}-e_{F}\geqslant\alpha\cdot(v_{H}-v_{F})+\delta_{1}$ (2)
for all $H\in\mathcal{H}$ and all $F\subsetneq H$; we henceforth fix such a
$\delta_{1}>0$.
Our first claim deals with the (easy) case that $G_{i}\to G_{i+1}$ is a
degenerate $\mathcal{H}$-step.
###### Claim 4.4.
If $G_{i}\to G_{i+1}$ is a degenerate $\mathcal{H}$-step, then
$b(G_{i+1})\geqslant b(G_{i})+\delta_{1}$.
###### Proof.
Suppose we add to $G_{i}$ a copy of some $H\in\mathcal{H}$ that intersects
$G_{i}$ on a subgraph $F\subseteq H$. This means that
$e_{G_{i+1}}=e_{G_{i}}+(e_{H}-e_{F})\qquad\text{ and }\qquad
v_{G_{i+1}}=v_{G_{i}}+(v_{H}-v_{F})$
and thus
$b(G_{i+1})-b(G_{i})=(e_{H}-e_{F})-\alpha\cdot(v_{H}-v_{F})\geqslant\delta_{1},$
where the inequality follows from (2), as $F$ must be a proper subgraph of
$H$. ∎
Now, suppose that $G_{i}\to G_{i+1}$ is an $\mathcal{L}$-step, either
degenerate or pristine, which means that we add a copy $\widehat{L}$ of some
$L\in\mathcal{L}$ and then add, for every edge $e\in\widehat{L}\setminus
G_{i}$, a copy $\widehat{H_{e}}$ of some $H_{e}\in\mathcal{H}$. Let
$G_{i}^{\prime}\coloneqq G_{i}\cup\widehat{L}$ and let $\widehat{J}\coloneqq
G_{i}\cap\widehat{L}$, so that $\widehat{J}\cong J$ for some $J\subsetneq L$
with at least one edge. Note that
$b(G_{i}^{\prime})-b(G_{i})=(e_{L}-e_{J})-\alpha\cdot(v_{L}-v_{J}),$ (3)
as we add $e_{L}-e_{J}$ edges and $v_{L}-v_{J}$ vertices to $G_{i}$ when
forming $G_{i}^{\prime}$.
In order to analyze $b(G_{i+1})-b(G_{i}^{\prime})$, we now define an auxiliary
graph $\mathcal{I}$ as follows. Its vertices are the edges of
$\widehat{L}\setminus\widehat{J}$. Recall that, for every such edge $e$, the
graph $\widehat{H_{e}}\cong H_{e}$ intersects $G_{i}^{\prime}$ only in the
edge $e$. A pair $e,f$ of edges of $\widehat{L}\setminus\widehat{J}$ will be
adjacent in $\mathcal{I}$ if and only if their corresponding graphs
$\widehat{H_{e}}$ and $\widehat{H_{f}}$ share at least one edge (equivalently,
the graphs $\widehat{H_{e}}\setminus e$ and $\widehat{H_{f}}\setminus f$ share
an edge).
Denote the connected components of $\mathcal{I}$ by $K_{1},\dotsc,K_{m}$ and
note that each of them corresponds to a subgraph of
$\widehat{L}\setminus\widehat{J}$. For each $j\in\llbracket{m}\rrbracket$, let
$U_{j}\coloneqq\bigcup_{e\in K_{j}}(\widehat{H_{e}}\setminus e).$
Note that the graphs $G_{i}^{\prime}$ and $U_{1},\dotsc,U_{m}$ are pairwise
edge-disjoint and that each $U_{j}$ shares at least $v_{K_{j}}$ vertices (the
endpoints of all the edges of $K_{j}$) with $G_{i}^{\prime}$. It follows that
$b(G_{i+1})-b(G_{i}^{\prime})\geqslant\sum_{j=1}^{m}(e_{U_{j}}-\alpha\cdot(v_{U_{j}}-v_{K_{j}}))=\sum_{j=1}^{m}(b(U_{j})+\alpha\cdot
v_{K_{j}}).$ (4)
Finally, as in the statement of Lemma 3.7, define
$X\coloneqq\min\\{(e_{H}-1)-\alpha\cdot(v_{H}-2):H\in\mathcal{H}\\}.$
The following claim lies at the heart of the matter.
###### Claim 4.5.
For every $j\in\llbracket{m}\rrbracket$, we have
$b(U_{j})\geqslant
X-2\alpha-(v_{K_{j}}-2)+\min\\{\delta_{1},1\\}\cdot\mathbf{1}_{v_{K_{j}}>2}.$
###### Proof.
Since $K_{j}$ is connected in $\mathcal{I}$, we may order its edges as
$e_{1},\dotsc,e_{\ell}$ so that, for each $r\in\llbracket{\ell-1}\rrbracket$,
the edge $e_{r+1}$ is $\mathcal{I}$-adjacent to $\\{e_{1},\dotsc,e_{r}\\}$.
Letting $F\subseteq H_{e_{r+1}}$ be the subgraph corresponding to this
intersection, we define, for each $r\in\\{0,\dotsc,\ell\\}$,
$U_{j}^{r}\coloneqq\bigcup_{s=1}^{r}(\widehat{H_{e_{s}}}\setminus e_{s}),$
so that $\varnothing=U_{j}^{0}\subseteq\dotsb\subseteq U_{j}^{\ell}=U_{j}$.
Observe that
$b(U_{j}^{1})=e_{U_{j}^{1}}-\alpha\cdot
v_{U_{j}^{1}}=(e_{H_{e_{1}}}-1)-\alpha\cdot v_{H_{e_{1}}}\geqslant X-2\alpha,$
where the inequality follows from the definition of $X$.
Suppose now that $r\geqslant 1$ and let $\widehat{F}$ be the intersection of
$\widehat{H_{e_{r+1}}}\setminus e_{r+1}$ with $U_{j}^{r}$; note that this
intersection is non-empty as $e_{r+1}$ is $\mathcal{I}$-adjacent to
$\\{e_{1},\dots,e_{r}\\}$. We have
$b(U_{j}^{r+1})-b(U_{j}^{r})=(e_{H_{e_{r+1}}}-1-e_{F})-\alpha\cdot(v_{H_{e_{r+1}}}-v_{F}).$
Let $t_{r+1}$ be the number of endpoints of $e_{r+1}$ that are not in
$U_{j}^{r}$. Suppose first that $t_{r+1}=0$, that is, both endpoints of
$e_{r+1}$ are already in $U_{j}^{r}$. In this case, both endpoints of
$e_{r+1}$ also belong to $\widehat{F}$ and thus $\widehat{F}\cup e_{r+1}$ is
isomorphic to a subgraph $F^{+}\subseteq H_{e_{r+1}}$ with $e_{F}+1$ edges and
$v_{F}$ vertices, which means that
$b(U_{j}^{r+1})-b(U_{j}^{r})=(e_{H_{e_{r+1}}}-e_{F^{+}})-\alpha\cdot(v_{H_{e_{r+1}}}-v_{F^{+}})\geqslant
0,$
by Lemma 3.5. In case $t_{r+1}>0$, $F$ is a proper subgraph of $H_{e_{r+1}}$
and thus we have
$b(U_{j}^{r+1})-b(U_{j}^{r})\geqslant\delta_{1}-1\geqslant\delta_{1}-t_{r+1},$
see (2). We may thus conclude that
$b(U_{j})=b(U_{j}^{1})+\sum_{r=1}^{\ell-1}(b(U_{j}^{r+1})-b(U_{j}^{r}))\geqslant
X-2\alpha-\sum_{r=1}^{\ell-1}t_{r+1}+\delta_{1}\cdot\mathbf{1}_{t_{2}+\dotsb+t_{\ell}>0}.$
The desired inequality follows as $t_{2}+\dotsb+t_{\ell}=\lvert
V(K_{j})\setminus V(U_{j}^{1})\rvert\leqslant v_{K_{j}}-2$ and, further,
$v_{K_{j}}>2$ implies that the sum $t_{2}+\dotsb+t_{r}$ is either positive or
at most $v_{K_{j}}-3$. ∎
We are now ready to show that the balance only increases when we perform an
$\mathcal{L}$-step.
###### Claim 4.6.
If $G_{i}\to G_{i+1}$ is an $\mathcal{L}$-step, then $b(G_{i+1})\geqslant
b(G_{i})$. Moreover, if this $\mathcal{L}$-step is degenerate, then
$b(G_{i+1})\geqslant b(G_{i})+\delta_{2}$ for some $\delta_{2}>0$ that depends
only on $\mathcal{H}$ and $\mathcal{L}$.
###### Proof.
By (3), (4), and Claim 4.6, we have
$\displaystyle b$
$\displaystyle(G_{i+1})-b(G_{i})=b(G_{i}^{\prime})-b(G_{i})+b(G_{i+1})-b(G_{i}^{\prime})$
$\displaystyle\geqslant(e_{L}-e_{J})-\alpha\cdot(v_{L}-v_{J})+\sum_{j=1}^{m}(b(U_{j})+\alpha\cdot
v_{K_{j}})$
$\displaystyle\geqslant(e_{L}-e_{J})-\alpha\cdot(v_{L}-v_{J})+\sum_{j=1}^{m}\left(X+(v_{K_{j}}-2)(\alpha-1)\right)+\min\\{\delta_{1},1\\}\cdot\mathbf{1}_{\mathcal{I}\neq\varnothing},$
since $\mathcal{I}$ is nonempty only if one of its components has more than
two vertices. We now apply Lemma 3.7 to each component $K_{j}$ to conclude
that
$\sum_{j=1}^{m}\left(X+(v_{K_{j}}-2)(\alpha-1)\right)\geqslant\sum_{j=1}^{m}e_{K_{j}}\cdot\left(\frac{\alpha}{m_{2}(L)}-1\right)=(e_{L}-e_{J})\left(\frac{\alpha}{m_{2}(L)}-1\right).$
Therefore,
$b(G_{i+1})-b(G_{i})\geqslant(e_{L}-e_{J})\cdot\frac{\alpha}{m_{2}(L)}-\alpha\cdot(v_{L}-v_{J})+\min\\{\delta_{1},1\\}\cdot\mathbf{1}_{\mathcal{I}\neq\varnothing}\geqslant\min\\{\delta_{1},1\\}\cdot\mathbf{1}_{\mathcal{I}\neq\varnothing},$
where the last inequality follows from Lemma 3.6. This implies the desired
result if the $\mathcal{L}$-step is pristine. If the $\mathcal{L}$-step is not
pristine but $\mathcal{I}$ has no edges, it means that some vertex was
repeated between different $\widehat{H_{e}}$. In that case, the first
inequality in (4) is strict (we assumed there that the graphs $U_{j}$ share no
vertices outside of $V(K_{j})$). All in all, we obtain the desired boost in
the degenerate case. ∎
Combining Claims 4.4 and 4.6, we obtain Lemma 4.3. This completes the proof of
the probabilistic lemma.
## 5\. Proof of the deterministic lemma
Given the probabilistic lemma and the work of the first two authors on the
symmetric case [17], in order to prove Conjecture 1.9, which generalizes the
Kohayakawa–Kreuter conjecture, we only need to show the following. For every
strictly balanced pair $(\mathcal{H},\mathcal{L})$ of finite families of
graphs with $m_{2}(\mathcal{H})>m_{2}(\mathcal{L})>1$, we can two-color the
edges of every graph $G$ satisfying $m(G)\leqslant
m_{2}(\mathcal{H},\mathcal{L})$ so that there are neither red monochromatic
copies of any $H\in\mathcal{H}$ nor blue monochromatic copies of any
$L\in\mathcal{L}$. As discussed in the introduction, we do not know how to do
this in all cases. However, the following proposition lists a number of extra
assumptions under which we are able to find such a coloring. We recall the
notion of the _$1$ -density_ (or _fractional arboricity_) of a graph $L$,
defined by
$m_{1}(L)\coloneqq\max\left\\{\frac{e_{J}}{v_{J}-1}:J\subseteq
L,v_{J}\geqslant 2\right\\}.$
We also make the following definition.
###### Definition 5.1.
Given positive integers $s\leqslant t$, we say that a graph is an _$(s,t)$
-graph_ if its minimum degree is at least $s$, and every edge contains a
vertex of degree at least $t$. We say that a graph is _$(s,t)$ -avoiding_ if
none of its subgraphs is an $(s,t)$-graph.
###### Proposition 5.2.
Let $(\mathcal{H},\mathcal{L})$ be a strictly balanced pair of finite families
of graphs satisfying $m_{2}(\mathcal{H})>m_{2}(\mathcal{L})$ and suppose that
at least one of the following conditions holds:
1. (a)
$\chi(L)\geqslant 3$ for all $L\in\mathcal{L}$;
2. (b)
$\chi(H)>m_{2}(\mathcal{H},\mathcal{L})+1$ for every $H\in\mathcal{H}$;
3. (c)
$m_{1}(L)>2$ for all $L\in\mathcal{L}$;
4. (d)
every $H\in\mathcal{H}$ contains an $(s,t)$-graph as a subgraph, for some
integers $s\leqslant t$ satisfying
$\frac{1}{s+1}+\frac{1}{t+1}<\frac{1}{m_{2}(\mathcal{H},\mathcal{L})};$
5. (e)
$\lceil m_{2}(\mathcal{H},\mathcal{L})\rceil<m_{2}(\mathcal{H})$;
Then any graph $G$ with $m(G)\leqslant m_{2}(\mathcal{H},\mathcal{L})$ is not
Ramsey for $(\mathcal{H},\mathcal{L})$.
Cases (a)–(c) all follow fairly easily from known coloring techniques; we
supply the details in the remainder of this section. Case (d) is proved by a
short inductive argument, see below. Case (e) follows from our partial
progress on Conjecture 1.5, namely, that we are able to prove it when $m(G)$
is an integer; we present the proof of this result in Appendix B. We end this
section with short derivations of Theorems 1.7 and 1.4 from the proposition.
###### Proof of Theorem 1.4.
Assume that $m_{2}(L)>\frac{11}{5}$. By passing to a subgraph with the same
$2$-density, we may assume that $L$ is strictly $2$-balanced. Thanks to cases
(a) and (c) of Proposition 5.2, we are done unless $m_{1}(L)\leqslant 2$ and
$L$ is bipartite. The bounds on $m_{1}(L)$ and $m_{2}(L)$ imply that
$2v_{L}-2\geqslant e_{L}>\frac{11}{5}(v_{L}-2)+1$, which yields $v_{L}<7$.
However, as $L$ is bipartite on at most six vertices, we have
$m_{2}(L)\leqslant m_{2}(K_{3,3})=2$, a contradiction. ∎
###### Proof of Theorem 1.7.
Cases (a), (b), (c), and (f) follow immediately777Proposition 5.2(c) implies
Theorem 1.7(b) thanks to Nash-Williams’s theorem (Theorem 5.7 below). from
Proposition 5.2. For Theorem 1.7(d), note that a graph with minimum degree $d$
is a $(d,d)$-graph. Thus, if $H_{1}$ has degeneracy at least $d$, then it
contains some $(d,d)$-graph as a subgraph. Similarly, Theorem 1.7(e) follows,
since if $s\leqslant t$, then $K_{s,t}$ is an $(s,t)$-graph satisfying
$1/m_{2}(K_{s,t})=(s+t-2)/(st-1)\geqslant 1/(s+1)+1/(t+1)$. ∎
### 5.1. Auxiliary results
We start with a helpful observation relating $m(G)$ and the degeneracy of $G$.
We say that a graph is _$d$ -degenerate_ if its degeneracy is at most $d$.
###### Lemma 5.3.
Every graph $G$ is $\lfloor 2m(G)\rfloor$-degenerate.
###### Proof.
For every $G^{\prime}\subseteq G$, we have
$\delta(G^{\prime})\leqslant\left\lfloor\frac{2e_{G^{\prime}}}{v_{G^{\prime}}}\right\rfloor\leqslant\lfloor
2m(G)\rfloor,$
where $\delta(G^{\prime})$ is the minimum degree of $G^{\prime}$. ∎
Our second lemma allows us to compare between the various densities.
###### Lemma 5.4.
For every graph $H$, we have $m_{2}(H)\leqslant m_{1}(H)+\frac{1}{2}\leqslant
m(H)+1$.
###### Proof.
Notice that both $\frac{e-1}{v-2}\leqslant\frac{e}{v-1}+\frac{1}{2}$ and
$\frac{e}{v-1}\leqslant\frac{e}{v}+\frac{1}{2}$ are equivalent to
$e\leqslant\binom{v}{2}$, so both inequalities hold whenever $v,e$ are the
numbers of vertices and edges, respectively, of any graph. In particular, if
$v,e$ correspond to the subgraph of $H$ that achieves $m_{2}(H)$, we find that
$m_{2}(H)=\frac{e-1}{v-2}\leqslant\frac{e}{v-1}+\frac{1}{2}\leqslant
m_{1}(H)+\frac{1}{2}$. The second inequality follows in the same way, now
passing to the subgraph that achieves $m_{1}(H)$. ∎
Our next lemma gives a lower bound on the average degree of an $(s,t)$-graph.
We remark that this inequality is tight for $K_{s,t}$ and that it can be
restated as $e_{H}/v_{H}\geqslant m(K_{s,t})$.
###### Lemma 5.5.
If $H$ is an $(s,t)$-graph, then
$\frac{1}{s}+\frac{1}{t}\geqslant\frac{v_{H}}{e_{H}}.$
###### Proof.
The assumption that $H$ is an $(s,t)$-graph implies that, for every $uv\in
E(H)$, we have $1/\deg(u)+1/\deg(v)\leqslant 1/s+1/t$. This means that
$e_{H}\cdot\left(\frac{1}{s}+\frac{1}{t}\right)\geqslant\sum_{uv\in
H}\left(\frac{1}{\deg(u)}+\frac{1}{\deg(v)}\right)=v_{H}.\qed$
The next lemma supplies a decomposition of a graph of bounded degeneracy.
###### Lemma 5.6.
If a graph $G$ is $(dk-1)$-degenerate, for some positive integers $d,k$, then
there is a partition $V(G)=V_{1}\cup\dotsb\cup V_{k}$ such that the graphs
$G[V_{1}],\dotsc,G[V_{k}]$ are all $(d-1)$-degenerate.
###### Proof.
We may construct the desired partition in the following way. Initialize
$V_{1}=\dotsb=V_{k}=\varnothing$ and let $v_{1},\dotsc,v_{n}$ be an ordering
of the vertices of $G$ such that every $v_{i}$ has at most $dk-1$ neighbors
preceding it. We distribute the vertices one-by-one, each time putting $v_{i}$
in a set $V_{j}$ where, at the time, $v_{i}$ has the smallest number of
neighbors. By the pigeonhole principle, this number is at most
$\lfloor\frac{dk-1}{k}\rfloor=d-1$. ∎
Finally, we quote Nash-Williams’s theorem on partitions of graphs into
forests.
###### Theorem 5.7 (Nash-Williams [21]).
A graph $G$ can be partitioned into $t$ forests if and only if $\lceil
m_{1}(G)\rceil\leqslant t$.
### 5.2. Proof of Proposition 5.2
We are now ready to prove Proposition 5.2. Denote $\alpha\coloneqq
m_{2}(\mathcal{H},\mathcal{L})$ and let $G$ be an arbitrary graph satisfying
$m(G)\leqslant\alpha$. We will argue that (the edge set of) $G$ can be
partitioned into an $\mathcal{H}$-free graph and an $\mathcal{L}$-free graph.
We split into cases, depending on which condition is satisfied by the pair
$(\mathcal{H},\mathcal{L})$.
#### Cases (a) and (b).
Let $k\coloneqq\lfloor\alpha\rfloor+1$, so that $m(G)\leqslant\alpha<k$, and
note that Lemma 5.3 implies that $G$ is $(2k-1)$-degenerate. Consequently,
Lemma 5.6 yields two partitions of the edges of $G$: a partition into a
$1$-degenerate graph and a $k$-colorable graph; and a partition into a
$(k-1)$-degenerate graph and a bipartite graph. The existence of the first
partition proves (b), as every $1$-degenerate graph is $\mathcal{L}$-free
whereas the assumption on $\mathcal{H}$ implies that $\chi(H)>k$ for every
$H\in\mathcal{H}$. We now argue that the existence of the second partition
proves (a). To this end, note that the assumption there implies that every
bipartite graph is $\mathcal{L}$-free, so it is enough to show that
$\delta(H)\geqslant k$ for every $H\in\mathcal{H}$ and thus every
$(k-1)$-degenerate graph is $\mathcal{H}$-free. To see that this is the case,
consider an arbitrary $H\in\mathcal{H}$ and let $v\in V(H)$ be its vertex with
smallest degree. As $H$ is strictly $m_{2}(\cdot,\mathcal{L})$-balanced, Lemma
3.5 gives $\delta(H)=e_{H}-e_{H\setminus v}>\alpha$, unless $v_{H}=3$, in
which case $H=K_{3}$ and we still have $\delta(H)\geqslant 2=m_{2}(H)\geqslant
m_{2}(\mathcal{H})>\alpha$. Since $\delta(H)$ is an integer, we actually have
$\delta(H)\geqslant\lfloor\alpha\rfloor+1=k$, as needed.
#### Case (c).
It is enough to show that $G$ can be partitioned into an $\mathcal{H}$-free
graph and a union of two forests; indeed, if $m_{1}(L)>2$ for all
$L\in\mathcal{L}$, then no union of two forests can contain a member of
$\mathcal{L}$ as a subgraph, by (the easy direction of) Theorem 5.7. Let
$m_{1}(\mathcal{H})\coloneqq\min\\{m_{1}(H):H\in\mathcal{H}\\}$. By Lemma 5.4
and the assumption $m(G)\leqslant
m_{2}(\mathcal{H},\mathcal{L})<m_{2}(\mathcal{H})$, we find that
$m_{1}(G)\leqslant m(G)+\frac{1}{2}\leqslant
m_{2}(\mathcal{H})+\frac{1}{2}\leqslant m_{1}(\mathcal{H})+1.$
As a result, if we let $t\coloneqq\lceil m_{1}(\mathcal{H})\rceil$, we find
that $\lceil m_{1}(G)\rceil\leqslant t+1$ and therefore Theorem 5.7 supplies a
partition $G$ into $t+1$ forests $G_{1},\dots,G_{t+1}$. Taking
$G^{\prime}\coloneqq G_{1}\cup\dots\cup G_{t-1}$, we arrive at a partition
$G=G^{\prime}\cup(G_{t}\cup G_{t+1})$. By (the easy direction of) Theorem 5.7,
we know that $m_{1}(G^{\prime})\leqslant t-1<m_{1}(\mathcal{H})$, so
$G^{\prime}$ is $\mathcal{H}$-free. As $G_{t}$ and $G_{t+1}$ are forests, we
get the desired decomposition.
#### Case (d).
It is enough to show that $G$ can be decomposed into a forest and an
$(s,t)$-avoiding graph. Assume that this is not the case and let $G$ be a
smallest counterexample with $m(G)\leqslant\alpha$. It is enough to show that
$G$ is an $(s+1,t+1)$-graph, as then Lemma 5.5 gives
$\frac{1}{s+1}+\frac{1}{t+1}\geqslant\frac{v_{G}}{e_{G}}\geqslant\frac{1}{m(G)}\geqslant\frac{1}{\alpha},$
a contradiction.
Suppose first that $G$ has a vertex $v$ of degree at most $s$. By minimality
of $G$, we can decompose the edges of $G\setminus v$ into an $(s,t)$-avoiding
graph $K$ and a forest $F$. Adding an arbitrary edge incident with $v$ to $F$
and the remaining edges to $K$ maintains $F$ being a forest and $K$ being
$(s,t)$-avoiding, as any $(s,t)$-subgraph of $K$ would have to use $v$, which
has degree at most $s-1$ in $K$. This contradicts our assumption on
indecomposability of $G$.
Second, suppose that $G$ contains an edge $uv$ with $\deg(u),\deg(v)\leqslant
t$. By minimality of $G$, we can decompose $G^{\prime}\coloneqq G\setminus uv$
into a forest $F$ and an $(s,t)$-avoiding graph $K$. Adding $uv$ to $F$ must
close a cycle, meaning that both $u$ and $v$ are incident to at least one
$F$-edge of $G^{\prime}$ and thus the $K$-degrees of $u$ and $v$ in
$G^{\prime}$ are at most $t-2$. This means, however, that we can add $uv$ to
$K$ while still keeping the degrees of both its endpoints strictly below $t$.
Again, we find that $K$ contains no $(s,t)$-subgraph, a contradiction.
#### Case (e).
Let $k\coloneqq\lceil m_{2}(\mathcal{H},\mathcal{L})\rceil$. Since we assume
that $m_{2}(\mathcal{H})>k$, it is enough to decompose $G$ into a forest and a
graph $K$ with $m_{2}(K)\leqslant k$. The following theorem, which implies
Conjecture 1.5 in the case that $m(G)$ is an integer, supplies such a
decomposition.
###### Theorem 5.8.
Let $k$ be an integer, and let $G$ be a graph with $m(G)\leqslant k$. Then
there exists a forest $F\subseteq G$ such that $m_{2}(G\setminus F)\leqslant
k$.
The proof of Theorem 5.8 is substantially more involved, as it relies on
techniques from matroid theory. We are hopeful that similar techniques may be
used to prove Conjecture 1.5 in its entirety. We defer the proof of Theorem
5.8 to Appendix B.
## References
* [1] J. Balogh, R. Morris, and W. Samotij, Independent sets in hypergraphs, _J. Amer. Math. Soc._ 28 (2015), 669–709.
* [2] C. Bowtell, R. Hancock, and J. Hyde, Proof of the Kohayakawa–Kreuter conjecture for the majority of cases, 2023. Preprint available at arXiv.
* [3] D. Conlon, J. Fox, and B. Sudakov, Recent developments in graph Ramsey theory, in _Surveys in combinatorics 2015_ , _London Math. Soc. Lecture Note Ser._ , vol. 424, Cambridge Univ. Press, Cambridge, 2015, 49–118.
* [4] D. Conlon and W. T. Gowers, Combinatorial theorems in sparse random sets, _Ann. of Math. (2)_ 184 (2016), 367–454.
* [5] J. Edmonds, Minimum partition of a matroid into independent subsets, _J. Res. Nat. Bur. Standards Sect. B_ 69B (1965), 67–72.
* [6] P. Frankl and V. Rödl, Large triangle-free subgraphs in graphs without $K_{4}$, _Graphs Combin._ 2 (1986), 135–144.
* [7] E. Friedgut and M. Krivelevich, Sharp thresholds for certain Ramsey properties of random graphs, _Random Structures Algorithms_ 17 (2000), 1–19.
* [8] E. Friedgut, V. Rödl, and M. Schacht, Ramsey properties of random discrete structures, _Random Structures Algorithms_ 37 (2010), 407–436.
* [9] L. Gugelmann, R. Nenadov, Y. Person, N. Škorić, A. Steger, and H. Thomas, Symmetric and asymmetric Ramsey properties in random hypergraphs, _Forum Math. Sigma_ 5 (2017), Paper No. e28, 47.
* [10] S. L. Hakimi, On the degrees of the vertices of a directed graph, _J. Franklin Inst._ 279 (1965), 290–308.
* [11] R. Hancock, K. Staden, and A. Treglown, Independent sets in hypergraphs and Ramsey properties of graphs and the integers, _SIAM J. Discrete Math._ 33 (2019), 153–188.
* [12] M. Harel, F. Mousset, and W. Samotij, Upper tails via high moments and entropic stability, _Duke Math. J._ 171 (2022), 2089–2192.
* [13] J. Hyde, Towards the 0-statement of the Kohayakawa-Kreuter conjecture, _Combin. Probab. Comput._ 32 (2023), 225–268.
* [14] S. Janson, T. Łuczak, and A. Ruciński, _Random graphs_ , Wiley-Interscience Series in Discrete Mathematics and Optimization, Wiley-Interscience, New York, 2000.
* [15] Y. Kohayakawa and B. Kreuter, Threshold functions for asymmetric Ramsey properties involving cycles, _Random Structures Algorithms_ 11 (1997), 245–276.
* [16] Y. Kohayakawa, M. Schacht, and R. Spöhel, Upper bounds on probability thresholds for asymmetric Ramsey properties, _Random Structures Algorithms_ 44 (2014), 1–28.
* [17] E. Kuperwasser and W. Samotij, The list-Ramsey threshold for families of graphs, 2023. Preprint available at arXiv:2305.19964.
* [18] A. Liebenau, L. Mattos, W. Mendonça, and J. Skokan, Asymmetric Ramsey properties of random graphs involving cliques and cycles, _Random Structures Algorithms_ 62 (2023), 1035–1055.
* [19] M. Marciniszyn, J. Skokan, R. Spöhel, and A. Steger, Asymmetric Ramsey properties of random graphs involving cliques, _Random Structures Algorithms_ 34 (2009), 419–453.
* [20] F. Mousset, R. Nenadov, and W. Samotij, Towards the Kohayakawa-Kreuter conjecture on asymmetric Ramsey properties, _Combin. Probab. Comput._ 29 (2020), 943–955.
* [21] C. S. J. A. Nash-Williams, Decomposition of finite graphs into forests, _J. London Math. Soc._ 39 (1964), 12.
* [22] R. Nenadov and A. Steger, A short proof of the random Ramsey theorem, _Combin. Probab. Comput._ 25 (2016), 130–144.
* [23] J. Oxley, _Matroid theory_ , _Oxford Graduate Texts in Mathematics_ , vol. 21, second ed., Oxford University Press, Oxford, 2011.
* [24] F. P. Ramsey, On a Problem of Formal Logic, _Proc. London Math. Soc. (2)_ 30 (1929), 264–286.
* [25] V. Rödl and A. Ruciński, Lower bounds on probability thresholds for Ramsey properties, in _Combinatorics, Paul Erdős is eighty, Vol. 1_ , Bolyai Soc. Math. Stud., János Bolyai Math. Soc., Budapest, 1993, 317–346.
* [26] V. Rödl and A. Ruciński, Random graphs with monochromatic triangles in every edge coloring, _Random Structures Algorithms_ 5 (1994), 253–270.
* [27] V. Rödl and A. Ruciński, Threshold functions for Ramsey properties, _J. Amer. Math. Soc._ 8 (1995), 917–942.
* [28] D. Saxton and A. Thomason, Hypergraph containers, _Invent. Math._ 201 (2015), 925–992.
* [29] M. Schacht, Extremal results for random discrete structures, _Ann. of Math. (2)_ 184 (2016), 333–365.
* [30] P. Turán, Eine Extremalaufgabe aus der Graphentheorie, _Mat. Fiz. Lapok_ 48 (1941), 436–452.
* [31] T. Łuczak, A. Ruciński, and B. Voigt, Ramsey properties of random graphs, _J. Combin. Theory Ser. B_ 56 (1992), 55–68.
## Appendix A The three-color setting
In this section, we explain what about the proof needs to change to handle the
case $r\geqslant 3$, and prove Theorem 1.6. As many of these results are
essentially identical to the results discussed previously, we omit or shorten
several of the proofs. We begin by defining a natural three-color analogue of
cores.
###### Definition A.1.
Let $\mathcal{H}_{1},\mathcal{H}_{2},\mathcal{H}_{3}$ be three finite families
of graphs. A tuple $(G,\mathcal{F}_{1},\mathcal{F}_{2},\mathcal{F}_{3})$ is an
_$(\mathcal{H}_{1},\mathcal{H}_{2},\mathcal{H}_{3})$ -core_ if $G$ is a graph
and $\mathcal{F}_{i}\subseteq\mathcal{F}_{\mathcal{H}_{i}}[G]$ for all
$i\in\llbracket{3}\rrbracket$ are families satisfying the following
properties:
* •
The hypergraph $\mathcal{F}_{1}\cup\mathcal{F}_{2}\cup\mathcal{F}_{3}$ is
connected and spans $E(G)$.
* •
For every $i\in\llbracket{3}\rrbracket$, every
$\widehat{H_{i}}\in\mathcal{F}_{i}$, every edge $e\in\widehat{H_{i}}$, and
every $j\in\llbracket{3}\rrbracket\setminus\\{i\\}$, there is some
$\widehat{H_{j}}\in\mathcal{F}_{j}$ with
$\widehat{H_{i}}\cap\widehat{H_{j}}=\\{e\\}$.
We say that $G$ _supports a core_ if there exists a core
$(G,\mathcal{F}_{1},\mathcal{F}_{2},\mathcal{F}_{3})$.
The following simple lemma is a straightforward generalization of Lemma 3.2,
so we omit the proof.
###### Lemma A.2.
Let $G$ be a graph that is minimally Ramsey for
$(\mathcal{H}_{1},\mathcal{H}_{2},\mathcal{H}_{3})$, in the sense that any
proper subgraph $G^{\prime}\subsetneq G$ is not Ramsey for
$(\mathcal{H}_{1},\mathcal{H}_{2},\mathcal{H}_{3})$. Then $G$ supports a core.
It would be very convenient if every
$(\mathcal{H}_{1},\mathcal{H}_{2},\mathcal{H}_{3})$-core were also an
$(\mathcal{H}_{1},\mathcal{H}_{2})$-core. At first glance this seems true,
since the intersection property in Definition A.1 easily implies the
intersection property in Definition 3.1. Unfortunately, it may be the case
that the hypergraph $\mathcal{F}_{1}\cup\mathcal{F}_{2}\cup\mathcal{F}_{3}$ is
connected, but that the hypergraph $\mathcal{F}_{1}\cup\mathcal{F}_{2}$ is
disconnected. Nonetheless, this is the only obstruction, and the following
result is true.
###### Lemma A.3.
Let $(G,\mathcal{F}_{1},\mathcal{F}_{2},\mathcal{F}_{3})$ be
$(\mathcal{H}_{1},\mathcal{H}_{2},\mathcal{H}_{3})$-core for some families of
graphs $\mathcal{H}_{1},\mathcal{H}_{2},\mathcal{H}_{3}$. Then
$(G,\mathcal{F}_{1},\mathcal{F}_{2}\cup\mathcal{F}_{3})$ is an
$(\mathcal{H}_{1},\mathcal{H}_{2}\cup\mathcal{H}_{3})$-core.
###### Proof.
First note that the hypergraph
$\mathcal{F}_{1}\cup(\mathcal{F}_{2}\cup\mathcal{F}_{3})$ is simply the same
as the hypergraph $\mathcal{F}_{1}\cup\mathcal{F}_{2}\cup\mathcal{F}_{3}$, so
it is connected and spans $E(G)$ by assumption. For every
$\widehat{H_{1}}\in\mathcal{F}_{1}$ and every edge $e\in\widehat{H_{1}}$, we
may apply Definition A.1 with $j=2$ to see that there exists some
$\widehat{H_{2}}\in\mathcal{F}_{2}\subseteq\mathcal{F}_{2}\cup\mathcal{F}_{3}$
such that $\widehat{H_{1}}\cap\widehat{H_{2}}=\\{e\\}$. Similarly, applying
Definition A.1 with $j=1$, we see that for every
$\widehat{H_{23}}\in\mathcal{F}_{2}\cup\mathcal{F}_{3}$ and every edge
$e\in\widehat{H_{23}}$, there is some $\widehat{H_{1}}\in\mathcal{F}_{1}$ such
that $\widehat{H_{1}}\cap\widehat{H_{23}}=\\{e\\}$. Thus,
$(G,\mathcal{F}_{1},\mathcal{F}_{2}\cup\mathcal{F}_{3})$ is an
$(\mathcal{H}_{1},\mathcal{H}_{2}\cup\mathcal{H}_{3})$-core. ∎
The key (trivial) observation is that if
$m_{2}(\mathcal{H}_{2})=m_{2}(\mathcal{H}_{3})$, then
$m_{2}(\mathcal{H}_{2}\cup\mathcal{H}_{3})$ is also equal to both these
numbers, as
$m_{2}(\mathcal{H}_{2}\cup\mathcal{H}_{3})=\min\\{m_{2}(\mathcal{H}_{2}),m_{2}(\mathcal{H}_{3})\\}$.
Now, suppose we are given families
$\mathcal{H}_{1},\mathcal{H}_{2},\mathcal{H}_{3}$ with
$m_{2}(\mathcal{H}_{1})>m_{2}(\mathcal{H}_{2})=m_{2}(\mathcal{H}_{3})$. By
passing to families of subgraphs, we may assume that
$\mathcal{H}_{2},\mathcal{H}_{3}$ are strictly 2-balanced and that
$\mathcal{H}_{1}$ is strictly $m_{2}(\cdot,\mathcal{H}_{2})$-balanced. We now
define $\mathcal{H}=\mathcal{H}_{1}$ and
$\mathcal{L}=\mathcal{H}_{2}\cup\mathcal{H}_{3}$. By Lemma 4.1, we know that
there exists some $c>0$ such that if $p\leqslant
cn^{-1/m_{2}(\mathcal{H},\mathcal{L})}$, then a.a.s. $G_{n,p}$ contains no
subgraph $G$ which supports an $(\mathcal{H},\mathcal{L})$-core and satisfies
$m(G)>m_{2}(\mathcal{H},\mathcal{L})$.
On the other hand, if $G_{n,p}$ is Ramsey for
$(\mathcal{H}_{1},\mathcal{H}_{2},\mathcal{H}_{3})$, then it must contain some
minimally Ramsey subgraph $G$. By Lemmas A.3 and A.2, $G$ supports an
$(\mathcal{H},\mathcal{L})$-core. Moreover, by the above, we must have
$m(G)\leqslant
m_{2}(\mathcal{H},\mathcal{L})=m_{2}(\mathcal{H}_{1},\mathcal{H}_{2})$, for
otherwise $G\nsubseteq G_{n,p}$ a.a.s. Given this, the following deterministic
lemma concludes the proof.
###### Lemma A.4.
Let $\mathcal{H}_{1},\mathcal{H}_{2},\mathcal{H}_{3}$ satisfy
$m(\mathcal{H}_{1})\geqslant m(\mathcal{H}_{2})\geqslant
m(\mathcal{H}_{3})>1$. If $G$ is Ramsey for
$(\mathcal{H}_{1},\mathcal{H}_{2},\mathcal{H}_{3})$, then
$m(G)>m_{2}(\mathcal{H}_{1},\mathcal{H}_{2})$.
###### Proof.
We will actually prove that $m(G)>m_{2}(\mathcal{H}_{1})$, which implies the
desired result since $m_{2}(\mathcal{H}_{1})\geqslant
m_{2}(\mathcal{H}_{1},\mathcal{H}_{2})$. Suppose for contradiction that
$m(G)\leqslant m_{2}(\mathcal{H}_{1})$. By Theorem 5.7 (cf. the proof of
Proposition 5.2(c)), we know that $G$ is the union of an
$\mathcal{H}_{1}$-free graph and two forests. As
$m_{2}(\mathcal{H}_{2})\geqslant m_{2}(\mathcal{H}_{3})>1$, every graph in
$\mathcal{H}_{2}\cup\mathcal{H}_{3}$ contains a cycle, and hence each of these
forests is $\mathcal{H}_{2}\cup\mathcal{H}_{3}$-free. Using one color for the
$\mathcal{H}_{1}$-free graph and one color for each of the two forests, we see
that $G$ is not Ramsey for
$(\mathcal{H}_{1},\mathcal{H}_{2},\mathcal{H}_{3})$. ∎
## Appendix B Proof of Conjecture 1.5 in the integer case
In this section, we present the proof of Theorem 5.8, which implies Conjecture
1.5 in the case that $m(G)$ is an integer. We will use some well-known results
from matroid theory; all definitions and proofs can be found in any standard
reference on matroid theory, such as Oxley’s book [23].
The main result we will need is the following matroid partitioning theorem,
originally due to Edmonds [5]. We remark that this theorem easily implies
Nash-Williams’s theorem (Theorem 5.7), which was used in the proof of
Proposition 5.2(c).
###### Theorem B.1.
Let $M_{1},M_{2}$ be matroids on the same ground set $E$, with rank functions
$r_{1},r_{2}$, respectively. Then $E$ can be partitioned as $E=I_{1}\cup
I_{2}$, with $I_{i}$ independent in $M_{i}$ for $i=1,2$, if and only if
$r_{1}(X)+r_{2}(X)\geqslant\lvert X\rvert$
for every $X\subseteq E$.
A slightly weaker statement appears as [23, Theorem 11.3.12], where the result
is only stated when $M_{1}=M_{2}$. However, it is clear and well-known that
the same proof proves Theorem B.1, using the formula for the rank of a matroid
union, as given in [23, Theorem 11.3.1].
In our application, we will set $E=E(G)$ and let $M_{1}$ be the graphic
matroid of $G$, whose independent sets are precisely the acyclic subgraphs of
$G$. We may view any subset of $E(G)$ as a subgraph $J$ of $G$; we then use
$e_{J}$ rather than $\lvert J\rvert$ to denote the size of this subset of
$E(G)$. Additionally, we use $v_{J}$ to denote the number of vertices incident
to any edge of $J$, and $\omega_{J}$ to denote the number of connected
components of $J$. It is well-known (e.g. [23, equation 1.3.8]) that the rank
function of $M_{1}$ is given by $r_{1}(J)=v_{J}-\omega_{J}$ for all
$J\subseteq E(G)$.
The second matroid we use will be one whose independent sets are precisely
those subgraphs $K\subseteq G$ with $m_{2}(K)\leqslant k$. The fact that this
is a matroid is the content of the next lemma.
###### Lemma B.2.
Let $G$ be a graph and let $k$ be a positive integer. Then the family of
subgraphs $K\subseteq G$ with $m_{2}(K)\leqslant k$ is the collection of
independent sets of a matroid.
###### Proof.
Define a function $f\colon 2^{E(G)}\to\mathbb{Z}$ by $f(J)=k(v_{J}-2)+1$, for
every $J\subseteq E(G)$. Note that this function is integer-valued since
$k\in\mathbb{Z}$. Additionally, it is clear that $f$ is increasing, in the
sense that $f(J)\leqslant f(J^{\prime})$ whenever $J\subseteq J^{\prime}$.
Finally, we claim that $f$ is submodular. This is easiest to see by recalling
that the function $g(J)=v_{J}$ is submodular (see e.g. [23, Proposition
11.1.6]); as $f$ is obtained from $g$ by multiplying by a positive constant
and adding a constant, we find that $f$ is submodular as well.
Now, by [23, Corollary 11.1.2], we find that there exists a matroid $M(f)$ on
$E(G)$ whose independent sets are precisely those $K\subseteq E(G)$ with the
property that $e_{J}\leqslant f({J})$ for all non-empty $J\subseteq K$. Note
that, for a graph $J$ with at least three vertices, the inequality
$e_{J}\leqslant f(J)$ is equivalent to $d_{2}(J)\leqslant k$, where
$d_{2}(J)=(e_{J}-1)/(v_{J}-2)$. If $J$ is non-empty and has only two vertices,
then it must have one edge and $e_{J}\leqslant f(J)$ always holds. Thus, we
see that $K$ is independent in $M(f)$ if and only if
$\max\\{(e_{J}-1)/(v_{J}-2):J\subseteq K,v_{J}\geqslant 3\\}\leqslant k$. This
condition is precisely the condition that $m_{2}(K)\leqslant k$. ∎
In order to apply Theorem B.1 to the matroids $M_{1},M_{2}$, we need a way of
lower-bounding the rank function of $M_{2}$. This is achieved by the following
lemma.
###### Lemma B.3.
Let $k$ be a positive integer. If $J$ is a graph with $m(J)\leqslant k$, then
there is a subgraph $J^{\prime}\subseteq J$ with $m_{2}(J^{\prime})\leqslant
k$ and $e_{J}\leqslant e_{J^{\prime}}+v_{J}-1.$
###### Proof.
A well-known theorem of Hakimi [10], which is itself a simple consequence of
Theorem B.1, implies that since $m(J)\leqslant k$, we can partition $J$ into
graphs $J_{1},\dots,J_{k}$, with $m(J_{i})\leqslant 1$ for all $i$ (i.e. every
component of every $J_{i}$ has at most one cycle). We may assume without loss
of generality that $J_{k}$ is non-empty. Let $e$ be an edge of $J_{k}$ and
define $J^{\prime}=J_{1}\cup\dotsb\cup J_{k-1}\cup\\{e\\}$. We claim that
$m_{2}(J^{\prime})\leqslant k$ and $e_{J}\leqslant e_{J^{\prime}}+v_{J}-1$.
The second claim is fairly easy to see, as
$e_{J^{\prime}}=1+\sum_{i=1}^{k-1}e_{J_{i}}=1+(e_{J}-e_{J_{k}})\geqslant
1+e_{J}-v_{J_{k}}\geqslant e_{J}-v_{J}+1,$
where the second equality uses the fact that $J_{1},\dots,J_{k}$ partition
$J$, and the two inequalities follow from $e_{J_{k}}\leqslant
v_{J_{k}}\leqslant v_{J}$, since $m(J_{k})\leqslant 1$ and $J_{k}\subseteq J$.
So it remains to prove that $m_{2}(J^{\prime})\leqslant k$, i.e. that
$d_{2}(L)\leqslant k$ for all $L\subseteq J^{\prime}$. If $v_{L}\leqslant
2k-1$, then
$d_{2}(L)\leqslant\frac{\binom{v_{L}}{2}-1}{v_{L}-2}=\frac{1}{2}\cdot\frac{v_{L}^{2}-v_{L}-2}{v_{L}-2}=\frac{1}{2}(v_{L}+1)\leqslant
k,$
as claimed. So we may assume that $v_{L}\geqslant 2k$. As $m(J_{i})\leqslant
1$ for all $i$, we see that $e_{L}\leqslant(k-1)v_{L}+1$. Therefore,
$d_{2}(L)=\frac{e_{L}-1}{v_{L}-2}\leqslant\frac{(k-1)v_{L}}{v_{L}-2}\leqslant\frac{kv_{L}-2k}{v_{L}-2}=k.\qed$
With all of these preliminaries, we are ready to prove Theorem 5.8.
###### Proof of Theorem 5.8.
Let $G$ be a graph with $m(G)\leqslant k$ and let $E=E(G)$. Let $M_{1}$ be the
graphic matroid on the ground set $E$ and let $M_{2}$ be the matroid given by
Lemma B.2, whose independent sets are those $K\subseteq G$ with
$m_{2}(K)\leqslant k$. We wish to prove that $E$ can be partitioned into an
independent set from $M_{1}$ and an independent set from $M_{2}$; by Theorem
B.1, it suffices to prove that $r_{1}(J)+r_{2}(J)\geqslant e_{J}$ for all
$J\subseteq G$.
So fix some $J\subseteq G$, and let its connected components be
$J_{1},\dots,J_{t}$. We then have that $r_{1}(J)=v_{J}-\omega_{J}=v_{J}-t$. As
$m(G)\leqslant k$, we certainly have that $m(J_{i})\leqslant k$ for all $i$,
and hence Lemma B.3 implies that there exist $J^{\prime}_{i}\subseteq J_{i}$
with $m_{2}(J^{\prime}_{i})\leqslant k$ and $e_{J_{i}}\leqslant
e_{J^{\prime}_{i}}+v_{J_{i}}-1$. Let $J^{\prime}=J_{1}^{\prime}\cup\dotsb\cup
J_{t}^{\prime}$. If $J^{\prime}$ is a matching, then
$m_{2}(J^{\prime})\leqslant 1\leqslant k$. If not, then its maximal
$2$-density is attained on some connected component, hence
$m_{2}(J^{\prime})=\max_{i}m_{2}(J_{i}^{\prime})\leqslant k$. Therefore,
$J^{\prime}$ is independent in $M_{2}$, which implies that
$r_{2}(J)\geqslant
r_{2}(J^{\prime})=e_{J^{\prime}}=\sum_{i=1}^{t}e_{J_{i}^{\prime}}\geqslant\sum_{i=1}^{t}(e_{J_{i}}-(v_{J_{i}}-1))=e_{J}-(v_{J}-t).$
Recalling that $r_{1}(J)=v_{J}-t$, we conclude that
$r_{1}(J)+r_{2}(J)\geqslant e_{J}$, as claimed. ∎
|
1
# Dependence-Aware, Unbounded Sound Predictive Race Detection
This extended arXiv version of an OOPSLA 2019 paper adds Appendices A–C
Kaan Genç Ohio State UniversityUSA<EMAIL_ADDRESS>, Jake Roemer Ohio State
UniversityUSA<EMAIL_ADDRESS>, Yufan Xu Ohio State UniversityUSA
<EMAIL_ADDRESS>and Michael D. Bond Ohio State UniversityUSA
<EMAIL_ADDRESS>
(2019)
# Dependence-Aware, Unbounded Sound Predictive Race Detection
This extended arXiv version of an OOPSLA 2019 paper adds Appendices A–C
Kaan Genç Ohio State UniversityUSA<EMAIL_ADDRESS>, Jake Roemer Ohio State
UniversityUSA<EMAIL_ADDRESS>, Yufan Xu Ohio State UniversityUSA
<EMAIL_ADDRESS>and Michael D. Bond Ohio State UniversityUSA
<EMAIL_ADDRESS>
(2019)
###### Abstract.
Data races are a real problem for parallel software, yet hard to detect. Sound
predictive analysis observes a program execution and detects data races that
exist in some _other, unobserved_ execution. However, existing predictive
analyses miss races because they do not scale to full program executions or do
not precisely incorporate data and control dependence.
This paper introduces two novel, sound predictive approaches that incorporate
data and control dependence and handle full program executions. An evaluation
using real, large Java programs shows that these approaches detect more data
races than the closest related approaches, thus advancing the state of the art
in sound predictive race detection.
data race detection, dynamic predictive analysis
††copyright: rightsretained††doi: 10.1145/3360605††journalyear: 2019††journal:
PACMPL††journalvolume: 3††journalnumber: OOPSLA††article:
179††publicationmonth: 10††ccs: Software and its engineering Dynamic
analysis††ccs: Software and its engineering Software testing and debugging
## 1\. Introduction
With the rise in parallel software, _data races_ represent a growing hazard.
Programs with data races written in shared-memory languages including Java and
C++ have weak or undefined semantics, as a result of assuming data race
freedom for performance reasons (Manson et al., 2005; Boehm and Adve, 2008;
Adve and Boehm, 2010). Data races are culprits in real software failures,
resulting in substantial financial losses and even harm to humans (Boehm,
2011; Kasikci et al., 2015; Lu et al., 2008; Kasikci et al., 2012;
Narayanasamy et al., 2007; Cao et al., 2016; Flanagan and Freund, 2010a; Sen,
2008; Burnim et al., 2011; Zhivich and Cunningham, 2009; U.S.–Canada Power
System Outage Task Force, 2004; Leveson and Turner, 1993; PCWorld, 2012).
Writing scalable, data-race-free code is challenging, as is detecting data
races, which occur nondeterministically depending on shared-memory
interleavings and program inputs and environments. The most common approach
for dealing with data races is to detect them during in-house testing using
dynamic _happens-before (HB)_ analysis (Flanagan and Freund, 2009; Pozniansky
and Schuster, 2007; Elmas et al., 2007; Serebryany and Iskhodzhanov, 2009;
Serebryany et al., 2012; Intel Corporation, 2016), which detects conflicting
accesses (two memory accesses, at least one of which is a write, to the same
variable by different threads) unordered by the HB partial order (Lamport,
1978). However, HB analysis misses data races when accesses _could_ race in
some _other_ execution but are ordered by critical sections on the same lock
in the observed execution.
A promising alternative to HB analysis is _sound predictive analysis_ , which
detects additional predictable data races from an observed execution (Huang et
al., 2014; Said et al., 2011; Huang and Rajagopalan, 2016; Chen et al., 2008;
Şerbănuţă et al., 2013; Liu et al., 2016; Smaragdakis et al., 2012; Kini et
al., 2017; Roemer et al., 2018; Pavlogiannis, 2019); an analysis is _sound_ if
it detects no false races (Section 2). Some predictive analyses rely on
generating and solving SMT constraints, so in practice they cannot scale to
full program executions and instead analyze _bounded windows_ of execution,
missing races between accesses that do not execute close together (Huang et
al., 2014; Said et al., 2011; Huang and Rajagopalan, 2016; Chen et al., 2008;
Şerbănuţă et al., 2013; Liu et al., 2016) (Section 8). In contrast,
_unbounded_ predictive analyses avoid this limitation by detecting races based
on computing a partial order weaker than HB, using analyses with linear
running time in the length of the trace (Kini et al., 2017; Roemer et al.,
2018). However, these partial-order-based analyses miss predictable races
because they do not incorporate precise notions of _data and control
dependence_. More precisely, existing predictive partial orders do _not_
encode the precise conditions for reordering memory accesses to expose a race:
The reordering can change the last writer of a memory read (data dependence)
if the read in turn cannot affect whether the racing accesses execute (control
dependence). Encoding data and control dependence precisely in a partial order
is fundamentally challenging (Section 2).
##### Contributions.
This paper designs and evaluates new predictive analyses, making the following
contributions:
* •
A partial order called _strong-dependently-precedes (SDP)_ that improves over
the highest-coverage sound partial order from prior work (Kini et al., 2017)
by incorporating data dependence more precisely (Section 4).
* •
A proof that SDP is sound, i.e., detects no false races (Section 4.2).
* •
A partial order called _weak-dependently-precedes (WDP)_ that improves over
the previous highest-coverage partial order from prior work (Roemer et al.,
2018) by incorporating data and control dependence precisely (Section 4).
* •
A proof that WDP is complete (sometimes called _maximal_ (Huang et al., 2014;
Şerbănuţă et al., 2013)), detecting all races knowable from an observed
execution (Section 4.2).
* •
Dynamic analyses that compute SDP and WDP and detect SDP- and WDP-races
(Section 5).
* •
An algorithm for filtering out WDP-races that are false races, by extending
prior work’s _vindication_ algorithm (Roemer et al., 2018), yielding an
overall sound approach (Section 6).
* •
An implementation and evaluation of SDP and WDP analyses and WDP-race
vindication on benchmarked versions of real, large Java programs (Section 7).
The evaluation shows that the analyses find predictable races missed by the
closest related approaches (Kini et al., 2017; Roemer et al., 2018; Huang et
al., 2014).
## 2\. Background and Motivation
Recent partial-order-based predictive analyses can scale to full program
executions, enabling detection of predictable races that are millions of
executed operations apart (Kini et al., 2017; Roemer et al., 2018). However,
these partial orders are fundamentally limited and miss predictable races, as
this section explains. First, we introduce formalisms used throughout the
paper.
### 2.1. Execution Model
An _execution trace_ $\mathit{tr}$ is a sequence of events, ordered by the
total order $<_{\textsc{$\mathit{tr}$}}$, that represents a multithreaded
execution without loss of generality, corresponding to a linearization of a
sequentially consistent (SC) execution.111Although programs with data races
may violate SC (Manson et al., 2005; Boehm and Adve, 2008; Adve and Boehm,
2010; Dolan et al., 2018), dynamic race detection analyses (including ours)
add synchronization instrumentation before accesses, generally ensuring SC. We
assume every event in $\mathit{tr}$ is a unique object (e.g., has a unique
identifier), making it possible to identify the same event $e$ across other,
predicted traces. Each event has two attributes: (1) an identifier for the
thread that executed the operation; and (2) an operation, which is one of
wr(x), rd(x), acq(m), rel(m), or br, where x is a program memory location and
m is a program lock. (Later we consider how to extend analyses to handle lock-
free accesses and Java volatile / C++ atomic accesses.) An execution trace
must be _well formed_ : a thread may only acquire an unheld lock and may only
release a lock it has acquired.
Each br (branch) event $b$ represents an executed conditional operation—such
as a conditional jump, polymorphic call, or array element access—that may be
dependent on some prior read event(s) by the same thread. We assume a helper
function $\mathit{brDepsOn}(b,r)$ exists that returns true if the value read
by read event $r$ may affect $b$’s outcome. An implementation could use static
dependence analysis to identify reads on which a branch is data dependent. For
simplicity, the paper’s examples assume $\mathit{brDepsOn}(b,r)$ always
returns true, i.e., every branch is assumed dependent on preceding reads by
the same thread. Our implementation and evaluation make the same assumption,
as explained later. This assumption limits the capability of predictive
analysis to predict different executions; in other words, it limits the number
of knowable data races from a single execution.
Three example traces are shown in Figures 1(b), 1(c), and 1(e), in which top-
to-bottom order represents trace order, and column placement denotes an
event’s executing thread. We discuss these examples in detail later.
Two read or write events to the same variable are _conflicting_ , notated
$e\asymp e^{\prime}$, if the events are executed by different threads and at
least one is a write.
_Program-order (PO)_ is a partial order that orders events in the same thread:
$e\prec_{\textsc{\tiny{PO}}}e^{\prime}$ if
$e<_{\textsc{$\mathit{tr}$}}e^{\prime}$ and the events are executed by the
same thread.
The function $\mathit{CS(}e)$ returns the set of events in the critical
section started or ended by acquire or release event $e$, including the
bounding acquire and release events. $\mathit{R}(a)$ returns the release event
ending the critical section started by acquire event $a$, and $\mathit{A}(r)$
returns the acquire event starting the critical section ended by release event
$r$. The function $\mathit{lockset}(e)$ returns the set of locks held at a
read or write event $e$ by its executing thread.
### 2.2. Predictable Traces and Predictable Races
⬇ int z = 0, y = 0; Object m = new Object(); new Thread(() -> { synchronized
(m) { int t = z; y = 1; } }).start(); new Thread(() -> { synchronized (m) { z
= 1; x = 1; } }).start(); new Thread(() -> { synchronized (m) { int t = x; if
(t == 0) return; } int t = y; } }).start();
((a)) Java code that could lead to the executions in (b) and (c).
Thread 1 Thread 2 Thread 3 acq(m) rd(z) wr(y) rel(m) acq(m) wr(z) wr(x) rel(m)
acq(m) rd(x) br rel(m) rd(y)
((b)) Execution with a predictable race
Thread 1 Thread 2 Thread 3 acq(m) wr(z) wr(x) rel(m) acq(m) rd(x) br rel(m)
acq(m) rd(z) wr(y) rd(y)
((c)) Predictable trace of (b)
⬇ int z = 0, y = 0; Object m = new Object(); new Thread(() -> { synchronized
(m) { int t = z; if (t == 0) y = 1; } }).start(); new Thread(() -> {
synchronized (m) { z = 1; x = 1; } }).start(); new Thread(() -> { synchronized
(m) { int t = x; if (t == 0) return; } int t = y; } }).start();
((d)) Java code that could lead to the execution in (e).
Thread 1 Thread 2 Thread 3 acq(m) rd(z) br wr(y) rel(m) acq(m) wr(z) wr(x)
rel(m) acq(m) rd(x) br rel(m) rd(y)
((e)) Execution with no predictable race
Figure 1. Two code examples, with potential executions they could lead to. The
execution in (b) has a predictable race, as demonstrated by the predictable
trace in (c). The execution in (e) has no predictable race.
By observing one execution of a program, it is possible to predict data races
in both the observed execution and some _other_ executions of the program. The
information present in the observed execution implies the existence of other,
different executions, called _predictable traces_. To define what traces can
be predicted from an observed trace, we first define several relevant
concepts.
###### Definition 2.1 (Last writer).
Given a trace $\mathit{tr}$, let $\mathit{lastwr}_{\mathit{tr}}(r)$ for a read
event $r$ be the last write event before $r$ in $\mathit{tr}$ that accesses
the same variable as $r$, or $\varnothing$ if no such event exists.
To ensure that a predictable trace is feasible, each read in a predictable
trace must have the same last writer as in the observed trace—with one
exception: a read can have a different last writer if the read cannot take the
execution down a different control-flow path than the observed execution. An
example of such a read is Thread 1’s rd(z) event in Figure 1(c). Next, we
introduce a concept that helps in identifying reads whose last writer must be
preserved in a predictable trace.
###### Definition 2.2 (Causal events).
Given a trace $\mathit{tr}$, set of events $\mathit{S}$, and event $e$, let
$\mathit{causal(\mathit{tr},\mathit{S},e)}$ be a function that returns true if
at least one of the following properties holds, and false otherwise.
* •
$e$ is a read, and there exists a branch event $b$ such that
$b\in\mathit{S}\land e\prec_{\textsc{\tiny{PO}}}b\land\mathit{brDepsOn}(b,e)$.
* •
$e$ is a write, and there exists a read event $e^{\prime}$ such that
$e^{\prime}\in\mathit{S}\land e=\mathit{lastwr}_{\mathit{tr}}(e^{\prime})$.
* •
$e$ is a read, and there exists a write event $e^{\prime}$ such that
$e^{\prime}\in\mathit{S}\land e\prec_{\textsc{\tiny{PO}}}e^{\prime}$ ($e$ and
$e^{\prime}$ may access different variables).
Intuitively, $\mathit{causal(\mathit{tr},\mathit{S},e)}$ tells us whether an
event $e$ could have affected some event $e^{\prime}$ in $\mathit{S}$
directly. For example, $\mathit{causal(\mathit{tr},\mathit{S},e)}$ if read $e$
may affect a branch event in $\mathit{S}$; if $e$ writes a variable later read
by an event in $\mathit{S}$; or if $e$ reads a value that may affect a later
write by the same thread in $\mathit{S}$ (even if the read and write are to
different variables, to account for intra-thread data flow).
We can now define a predictable trace of an observed trace, which is a trace
that is definitely a feasible execution of the program, given the existence of
the observed execution.
###### Definition 2.3 (Predictable trace).
An execution trace $\mathit{tr^{\prime}}$ is a _predictable trace_ of trace
$\mathit{tr}$ if $\mathit{tr^{\prime}}$ contains only events in $\mathit{tr}$
(i.e., $\forall e:e\in\mathit{tr^{\prime}}\implies e\in\mathit{tr}$) and all
of the following rules hold:
_Program order (PO) rule:_ For any events $e_{1}$ and $e_{2}$, if
$e_{1}\prec_{\textsc{\tiny{PO}}}e_{2}$, then
$e_{1}<_{\textsc{$\mathit{tr^{\prime}}$}}e_{2}\lor
e_{2}\notin\mathit{tr^{\prime}}$.
_Last writer (LW) rule:_ For every read event $e$ such that
$\mathit{causal(\mathit{tr},\mathit{tr^{\prime}},e)}$,
$\mathit{lastwr}_{\mathit{tr^{\prime}}}(e)=\mathit{lastwr}_{\mathit{tr}}(e)$.
(In this context, $\mathit{tr^{\prime}}$ means the set of events in the trace
$\mathit{tr^{\prime}}$.)
_Lock semantics (LS) rule:_ For acquire events $e_{1}$ and $e_{2}$ on the same
lock, if $e_{1}<_{\textsc{$\mathit{tr^{\prime}}$}}e_{2}$ then
$e_{1}<_{\textsc{$\mathit{tr^{\prime}}$}}\mathit{R}(e_{1})<_{\textsc{$\mathit{tr^{\prime}}$}}e_{2}$.
The PO and LW rules ensure key properties from $\mathit{tr}$ also hold in
$\mathit{tr^{\prime}}$, while the LS rule ensures that $\mathit{tr^{\prime}}$
is well formed. The intuition behind the LW rule is that any read that may
(directly or indirectly) affect the control flow of the program must have the
same last writer in predictable trace $\mathit{tr^{\prime}}$ as in observed
trace $\mathit{tr}$.
Note that throughout the paper, partial ordering notation such as $e\prec
e^{\prime}$ refers to the order of $e$ and $e^{\prime}$ in the _observed_
trace $\mathit{tr}$ (not a predictable trace $\mathit{tr^{\prime}}$).
Predictable traces do not in general contain every event in the observed trace
they are based on. For the purposes of race detection, a predictable trace
will _conclude with_ a pair of conflicting events, which are preceded by
events necessary according to the definition of predictable trace. For
example, consider Figure 1(c), which is a predictable trace of Figure 1(b)
that excludes Thread 1’s event after wr(y). The PO rule is satisfied, and the
LS rule is satisfied after reordering Thread 2 and 3’s critical sections
before Thread 1’s. The LW rule is satisfied because rd(z) is not a causal
event in $\mathit{tr^{\prime}}$.
###### Definition 2.4 (Predictable race).
An execution $\mathit{tr}$ has a predictable race if a predictable trace
$\mathit{tr^{\prime}}$ of $\mathit{tr}$ has two _conflicting_ , _consecutive_
events: $e_{1}\asymp e_{2}\land
e_{1}<_{\textsc{$\mathit{tr^{\prime}}$}}e_{2}\land(\nexists
e:e_{1}<_{\textsc{$\mathit{tr^{\prime}}$}}e<_{\textsc{$\mathit{tr^{\prime}}$}}e_{2})$.
Figure 1(b) has a predictable race, as demonstrated by the predictable trace
in Figure 1(c). In contrast, Figure 1(e) has no predictable race. The
difference between Figures 1(b) and 1(e) is the br event in Thread 1. Since no
br exists in Thread 1 in Figure 1(b), rd(z) is not a causal event, which in
turn allows the critical sections in Threads 2 and 3 to be reordered above the
critical section in Thread 1, allowing wr(y) and rd(y) to be consecutive in
the predictable trace. In contrast, a br event executes before wr(y) in Figure
1(e). No predictable trace of this example can exclude br without excluding
wr(y); otherwise the PO rule would be violated. rd(z) is a causal event in any
predictable trace where wr(y) is included, which makes it impossible to
reorder the critical sections. As a result, no predictable trace exists in
which wr(y) and rd(y) are consecutive. Figures 1(a) and 1(d) show source code
that could lead to the executions in Figures 1(b) and 1(e), respectively. The
code in Figure 1(d) has no race; in fact, any deviation of critical section
ordering from Figure 1(e)’s causes wr(y) or rd(y) _not_ to execute.
### 2.3. Existing Predictive Partial Orders
Here we overview three relations introduced in prior work, called _happens-
before (HB)_ , _weak-causally-precedes (WCP)_ , and _doesn’t-commute (DC)_ ,
that can be computed in time linearly proportional to the length of the
execution trace (Kini et al., 2017; Roemer et al., 2018). Intuitively, each
relation orders events that may not be legal to reorder in a predictable
trace, so that two unordered conflicting events represent a true or potential
data race (depending on whether the relation is sound). An execution trace has
an _HB-race_ , _WCP-race_ , or _DC-race_ if it contains two conflicting events
that are unordered by HB, WCP, or DC, respectively.
##### Definitions of relations.
Table 1 gives definitions of HB, WCP, and DC by presenting their properties
comparatively. The first two rows of the table say how the relations order
critical sections on the same lock. HB orders all critical sections on the
same lock, and it orders the first critical section’s rel(m) to the second
critical section’s acq(m). WCP and DC order only _conflicting_ critical
sections (critical sections on the same lock containing conflicting events),
and they order from the first critical section’s rel(m) to the second critical
section’s conflicting access event. That is, if $r_{1}$ and $r_{2}$ are
release events on the same lock such that
$r_{1}<_{\textsc{$\mathit{tr}$}}r_{2}$, and $e_{1}$ and $e_{2}$ are
conflicting events ($e_{1}\asymp e_{2}$) such that
$e_{1}\in\mathit{CS(}r_{1})\land e_{2}\in\mathit{CS(}r_{2})$, then
$r_{1}\prec_{\textsc{\tiny{WCP}}}e_{2}$ and
$r_{1}\prec_{\textsc{\tiny{DC}}}e_{2}$. The intuition behind these properties
of WCP and DC is that non-conflicting critical sections can generally be
reordered in a predictable trace; and even in the case of conflicting critical
sections, the second critical section can be “reordered” so that it executes
only up to its conflicting access and the first critical section does not
execute at all in the predictable trace.
Property | $\prec_{\textsc{\tiny{HB}}}$ | $\prec_{\textsc{\tiny{WCP}}}$ | $\prec_{\textsc{\tiny{DC}}}$
---|---|---|---
Same-lock critical section ordering | All | Confl. | Confl.
Orders rel to… | acq | wr/rd | wr/rd
Includes $\prec_{\textsc{\tiny{PO}}}$? | Yes | No | Yes
Left-and-right composes with $\prec_{\textsc{\tiny{HB}}}$? | Yes | Yes | No
$\textsf{acq(m)}\prec\textsf{rel(m)}$ implies $\textsf{rel(m)}\prec\textsf{rel(m)}$? | Yes | Yes | Yes
Transitive? | Yes | Yes | Yes
Table 1. Definitions of three strict partial orders over events in an
execution trace. Each order is the minimum relation satisfying the listed
properties.
The next two table rows show whether the relations include PO or compose with
HB. HB and DC include (i.e., are supersets of) PO: if
$e_{1}\prec_{\textsc{\tiny{PO}}}e_{2}$, then
$e_{1}\prec_{\textsc{\tiny{HB}}}e_{2}$ and
$e_{1}\prec_{\textsc{\tiny{DC}}}e_{2}$. In contrast, WCP does not include PO
but instead _composes with_ the stronger HB: if
$e_{1}\prec_{\textsc{\tiny{WCP}}}e_{2}\prec_{\textsc{\tiny{HB}}}e_{3}$ or
$e_{1}\prec_{\textsc{\tiny{HB}}}e_{2}\prec_{\textsc{\tiny{WCP}}}e_{3}$, then
$e_{1}\prec_{\textsc{\tiny{WCP}}}e_{2}$. (By virtue of being transitive, HB
composes with itself.) The intuition behind including or composing with PO (a
subset of HB) is that PO-ordered events cannot be reordered in a predictable
trace. The intuition behind WCP composing with HB, in essence, is to avoid
predicting traces that violate the LS rule of predictable traces. As a result,
WCP is sound while DC is unsound, as we will see.
The last two rows show properties shared by all relations. First, if two
critical sections on the same lock
$a_{1}\prec_{\textsc{\tiny{PO}}}r_{1}<_{\textsc{$\mathit{tr}$}}a_{2}\prec_{\textsc{\tiny{PO}}}r_{2}$
are ordered at all (meaning simply $a_{1}\prec_{\ast}r_{2}$ because all
relations minimally compose with PO), then their release events are ordered
($r_{1}\prec_{\ast}r_{2}$). Second, all of the relations are transitive. As a
result of being transitive, antisymmetric, and irreflexive, all of the
relations are strict partial orders.
##### Example
As an example of WCP and DC ordering, consider the execution in Figure 2(b).
Both relations order Thread 1’s rel(m) to Thread 2’s wr(x) because the
critical sections on m contain conflicting accesses to x. By WCP’s composition
with HB (and thus PO) and DC’s inclusion of PO, both WCP and DC transitively
order rd(x) to wr(x) and wr(y) to rd(y)
($\textsf{wr(y)}\prec_{\textsc{\tiny{WCP}}}\textsf{rd(y)}$ and
$\textsf{wr(y)}\prec_{\textsc{\tiny{DC}}}\textsf{rd(y)}$), so the execution
has no WCP- or DC-races.
##### Soundness and completeness.
A relation or analysis is _sound_ if it detects a race only for an execution
trace with a predictable race or deadlock.222A trace has a _predictable
deadlock_ if there exists a valid reordering with a deadlock. We define
soundness to include predictable deadlocks because prior work’s WCP relation
(Kini et al., 2017) and our SDP relation are sound in this way. A relation or
analysis is _complete_ if it detects a race for every execution trace with a
predictable race.
WCP (and HB) are sound: a WCP-race (HB-race) indicates a predictable race or
deadlock. DC is unsound: an execution with a DC-race may have no predictable
race or deadlock. However, prior work shows that DC-races are generally true
predictable races in practice, and an efficient _vindication_ algorithm can
verify DC-races as predictable races by computing additional constraints and
building a predictable trace exposing the race (Roemer et al., 2018). Later in
the paper, we provide more details about vindication, when introducing a new
relation that (like DC) is unsound and makes use of a vindication algorithm.
### 2.4. Limitations of Existing Predictive Partial Orders
WCP and DC analyses are the state of the art in detecting as many predictable
races as possible using online, unbounded analysis (Kini et al., 2017; Roemer
et al., 2018). However, WCP and DC are _incomplete_ , failing to detect some
predictable races. WCP and DC are overly strict because they order all
conflicting accesses, conservatively ruling out some predictable traces that
still preserve the last writer of each causal read. This strictness arises
from imprecise handling of data and control dependence:
##### Data dependence:
WCP and DC order all conflicting accesses, which is imprecise because the
order of a write–write or read–write conflict does not necessarily need to be
preserved to satisfy the last-writer (LW) rule of predictable traces.
Thread 1 Thread 2 acq(m) rd(x) br wr(y) rel(m) acq(m) wr(x) rel(m) rd(y)
((a)) WDP-race but no WCP-, DC-, or SDP-race. No predictable race exists.
Thread 1 Thread 2 acq(m) wr(y) rd(x) br rel(m) acq(m) wr(x) rel(m) rd(y)
((b)) WDP-race but no WCP-, DC-, or SDP-race. A predictable race exists, as
(c) shows.
Thread 1 Thread 2 acq(m) wr(x) rel(m) acq(m) wr(y) rd(y)
((c)) Predictable trace of (b) showing that (b) has a predictable race
DCSDPDCSDP
Figure 2. Executions showing WCP and DC’s overly strict handling of read–write
dependencies. Edges represent ordering, labeled using the weakest applicable
relation(s) (and omitting ordering established by HB alone), implying ordering
by strictly stronger relations (see Figure 3(a) for comparison of relations).
Thread 1 Thread 2 acq(m) wr(x) wr(y) rel(m) acq(m) wr(x) rel(m) rd(y) acq(m)
rd(x) rel(m) br
((a)) Execution with SDP- and WDP-race but no WCP- or DC-race
DC SDP DC WDP SDP
Thread 1 Thread 2 acq(m) wr(x) rel(m) acq(m) wr(x) wr(y) rd(y)
((b)) Valid reordering of (a) showing that (a) has a predictable race
Thread 1 Thread 2 acq(m) wr(x) sync(o) wr(y) rel(m) acq(m) wr(x) rel(m)
sync(o) rd(x) br rd(y)
((c)) Execution with WDP-race but no SDP-race or predictable race
Figure 3. Executions showing WCP and DC’s overly strict handling of
write–write dependencies. sync(o) is an abbreviation for the sequence acq(o);
rd(oVar); br; wr(oVar); rel(o).
Consider the executions in Figures 2(a) and 2(b), in which WCP and DC order
rd(x) to wr(x). WCP and DC’s read–write ordering assumes that no predictable
trace exists where a pair of conflicting accesses are reordered. This
rationale works for Figure 2(a), in which no predictable race exists. However,
conflicting accesses may be reordered as long as the LW rule of predictable
traces is satisfied. Figure 2(c) is a predictable trace of Figure 2(b) that
reorders the critical sections and exposes a race on y.
Similarly for write–write conflicts, consider Figure 2(a), in which WCP and DC
order the two wr(x) events, leading to no WCP- or DC-race on accesses to y.
However, the wr(x) events can be reordered, as the predictable trace in Figure
2(b) shows, exposing a race. (The reader can ignore Figure 2(c) until Section
4.)
It is difficult to model read–write and write–write dependencies more
precisely using a partial order. In the case of a read–write dependency, the
accesses can be reordered _as long as the read cannot impact a branch’s
outcome in the predictable trace_ (i.e., the read is not a causal event in the
predictable trace, or is not part of the predictable trace). For a write–write
dependency, the accesses can be reordered _as long as they do not change a
causal read’s last writer in the predictable trace._ Incorporating either kind
of constraint into a partial order is challenging but also desirable because
partial orders can be computed efficiently.
##### Control dependence:
Thread 1 Thread 2 wr(y) acq(m) wr(x) rel(m) acq(m) rd(x) rel(m) rd(y)
((a)) WDP-race but no WCP-, DC-, or SDP-race
Thread 1 Thread 2 acq(m) rd(x) rel(m) wr(y) rd(y)
((b)) Valid reordering of (a) showing that (a) has a predictable race
Thread 1 Thread 2 wr(y) acq(m) wr(x) rel(m) acq(m) rd(x) rd(y) br rel(m)
((c)) WDP-race but no WCP-, DC-, or SDP-race
Thread 1 Thread 2 acq(m) rd(x) wr(y) rd(y)
((d)) Valid reordering of (c) showing that (c) has a predictable race
DC
SDP
DC
SDP
WDP
Figure 4. Executions showing WCP and DC’s overly strict handling of control
dependencies.
WCP and DC order true (write–read) dependencies even when the read may not
affect a branch outcome that affects whether a race happens.
Figure 4(a) shows an execution with a predictable race, as the predictable
trace in Figure 4(b) demonstrates. Note that in Figure 4(b), rd(x) has a
different last writer than in Figure 4(a), but the lack of a following br
event means that rd(y) is still guaranteed to happen (i.e., rd(x) is not a
causal event). A variant of this example is in Figure 4(c), which has a branch
event dependent on a read outcome, but the branch can be absent from a
predictable trace demonstrating a predictable race (Figure 4(d)).
Thread 1 | Thread 2 | Thread 3
---|---|---
this.method = …; | |
acq(this); | |
this.time += time | |
rel(this); | |
| acq(this); |
| this.time += time; |
| rel(this); |
| br |
| | acq(this);
| | this.time += time;
| | rel(this);
| | br
| | …= this.method;
DCDCWDPWDP
Figure 5. A predictable race in the Java program pmd that was detected by WDP,
but not WCP, DC, or SDP. Note the transitive edges formed by DC, which WDP
avoids as the branches are outside the critical sections. The code has been
simplified and abbreviated.
WCP and DC miss the predictable races in Figures 4(a) and 4(c) by
conservatively assuming that any event after a rd(x) may be control dependent
on the read value. Similarly, WCP and DC miss the predictable race in Figure
5, which our implementation found in the Java program pmd (Section 7).
Essentially, WCP and DC conservatively assume that a dependent branch
immediately follows each read. This limitation is unsurprising considering the
challenge of modeling control dependencies using a partial order. In
particular, it is difficult for a partial order to model the fact that _a read
must have the same last writer only if the read may affect a branch in the
predictable trace_.
This work develops partial orders that are weaker than WCP and DC and thus
predict more races. At the same time, these new partial orders retain key
properties of the existing relations: WCP’s soundness and DC’s amenability to
a vindication algorithm that ensures soundness, respectively.
## 3\. Overview
The previous section introduced prior work’s weak-causally-precedes (WCP)
(Kini et al., 2017) and doesn’t-commute (DC) (Roemer et al., 2018), and
explained their limitations that lead to missing predictable races. The next
two sections introduce new relations and analyses that overcome these
limitations. Section 4 introduces the _strong-dependently-precedes (SDP)_ and
_weak-dependently-precedes (WDP)_ relations, which are weaker than WCP and DC,
respectively. Section 5 presents online dynamic analyses for computing SDP and
WDP and detecting SDP- and WDP-races.
## 4\. New Dependence-Aware Predictive Relations
This section introduces new partial orders called _strong-dependently-precedes
(SDP)_ and _weak-dependently-precedes (WDP)_ that overcome the limitations of
prior work’s predictive relations (Roemer et al., 2018; Kini et al., 2017)
(Section 2.4) by incorporating more precise notions of data and control
dependence.
### 4.1. The SDP and WDP Partial Orders
SDP is weaker than WCP 333It may seem confusing that SDP is _weaker_ than WCP.
SDP is so named because it is stronger than WDP, while WCP is so named because
it is weaker than prior work’s _causally-precedes (CP)_ (Smaragdakis et al.,
2012). by not ordering write–write conflicts, based on the insight that writes
can be unordered unless they can affect the outcome of a read, but write–read
and read–write ordering already handles that ordering soundly. WDP only orders
the last writer of a read to a branch that depends on that read, which is the
only reordering constraint that does not lead to missing predictable races.
| Prior work | This paper
---|---|---
Property | $\prec_{\textsc{\tiny{HB}}}$ | $\prec_{\textsc{\tiny{WCP}}}$ | $\prec_{\textsc{\tiny{DC}}}$ | $\prec_{\textsc{\tiny{SDP}}}$ | $\prec_{\textsc{\tiny{WDP}}}$
Same-lock critical section ordering | All | Confl. | Confl. | Confl. | Last wr–rd only
Orders rel to… | acq | wr/rd | wr/rd | wr/rd or to next rd* | Next br
Includes $\prec_{\textsc{\tiny{PO}}}$? | Yes | No | Yes | No | Yes
Left-and-right composes with $\prec_{\textsc{\tiny{HB}}}$? | Yes | Yes | No | Yes | No
$\textsf{acq(m)}\prec\textsf{rel(m)}$ implies $\textsf{rel(m)}\prec\textsf{rel(m)}$? | Yes | Yes | Yes | Yes | Yes
Transitive? | Yes | Yes | Yes | Yes | Yes
Table 2. Definitions of five strict partial orders over events in an execution
trace. Each order is the minimum relation satisfying the listed properties.
This table adds columns $\prec_{\textsc{\tiny{SDP}}}$ and
$\prec_{\textsc{\tiny{WDP}}}$ to Table 1 (page 1).
* As the text explains, SDP adds release–access ordering for write–read and read–write conflicts, and adds ordering from the release to the next read for write–write conflicts.
Table 2 defines SDP and WDP. The table shows that SDP is like WCP and WDP is
like DC except in how they order critical sections (first two table rows).
##### The SDP relation.
SDP only orders conflicting critical sections when one critical section
contains a read. Like WCP, SDP orders the first critical section’s rel(m) to
the second critical section’s access. The intuition behind this property is
that write–write conflicts generally do not impose any limitations on what
traces can be predicted. Figure 2(a) shows an example in which two conflicting
writes can be safely reordered in a predictable trace.
However, ignoring write–write conflicts altogether would be unsound, as Figure
2(c) shows: the execution has no predictable race. To ensure soundness, SDP
handles write–write conflicts by ordering the first critical section to the
second thread’s next _read_ to the same variable.
More formally, SDP handles conflicting critical sections as follows. If
$r_{1}$ and $r_{2}$ are release events on the same lock,
$r_{1}<_{\textsc{$\mathit{tr}$}}r_{2}$, $e_{1}$ and $e_{2}$ are write events
and $e_{3}$ is a read event, $e_{1}\asymp e_{2}$, $e_{1}\asymp e_{3}$,
$e_{2}\prec_{\textsc{\tiny{PO}}}e_{3}$, $e_{1}\in\mathit{CS(}r_{1})$, and
$e_{2}\in\mathit{CS(}r_{2})$, then $r_{1}\prec_{\textsc{\tiny{SDP}}}e_{3}$.
SDP addresses a limitation of WCP via more precise handling of data
dependencies. SDP certainly does not address all imprecise data dependencies
(e.g., read–write dependencies), and it does not address control dependence.
SDP is the weakest known sound partial order.
##### The WDP relation
A separate but worthwhile goal is to develop a partial order that is weaker
than DC but produces few false positives so that it is practical to vindicate
potential races. WDP achieves this goal and is in fact complete, detecting all
predictable races. WDP orders the last writer of each read to the earliest
branch that depends on that read (and orders no other conflicting critical
sections). The intuition behind this behavior is that the only constraint that
is universally true for all predictable traces is that the last writer of a
read must not occur after the read if there is a branch that depends on the
read.
More formally, if $r_{1}$ and $r_{2}$ are releases on the same lock,
$e_{1}\in\mathit{CS(}r_{1})$, $e_{2}\in\mathit{CS(}r_{2})$,
$e_{1}=\mathit{lastwr}_{tr}(e_{2})$, $e_{2}\prec_{\textsc{\tiny{PO}}}b$, and
$\mathit{brDepsOn}(b,e_{2})$, then $r_{1}\prec_{\textsc{\tiny{WDP}}}b$.
Unlike DC, WDP integrates control dependence by ordering the write’s critical
section to the first branch dependent on the read. WDP does not model _local_
data dependencies, where a read affects the value written by a write in the
same thread. As a result, WDP may find false races, but Section 6 describes a
method for ruling out such false races. These properties make WDP complete (as
we show). WDP is the strongest known complete partial order.
##### SDP- and WDP-races.
Unlike WCP and DC, SDP and WDP do not inherently order all conflicting
accesses that hold a common lock. Thus the following definition of SDP- and
WDP-races explicitly excludes conflicting accesses holding a common lock.
A trace has a _SDP-race_ (or _WDP-race_) if it has two conflicting events
unordered by SDP (WDP) that hold no lock in common. That is, $\mathit{tr}$ has
an SDP-race (WDP-race) on events $e$ and $e^{\prime}$ if
$e<_{\textsc{$\mathit{tr}$}}e^{\prime}$, $e\asymp e^{\prime}$,
$e\not\prec_{\textsc{\tiny{SDP}}}e^{\prime}$
($e\not\prec_{\textsc{\tiny{WDP}}}e^{\prime}$), and
$\mathit{lockset}(e)\cap\mathit{lockset}(e^{\prime})=\emptyset$.
##### Examples.
To illustrate SDP and WDP, we refer back to examples from pages 3–5.
The executions in Figures 2(a) and 2(b) have no SDP-races: SDP orders the
read–write conflicts. In contrast, these executions have WDP-races: there is
no cross-thread WDP ordering because the executions have no lock-protected
write–read conflicts.
Figure 2(a) has SDP- and WDP-races. SDP and WDP do not order the write–write
conflict on x. Nor does WDP order events on the write–read conflict on x,
since Thread 1’s wr(x) is not the last writer of rd(x).
Figure 2(c), which has a WDP-race but no SDP-race or predictable race, shows
the need for SDP’s release–read ordering for write–write conflicts.
The executions in Figures 4(a) and 4(c) have no SDP-race since SDP does not
take branches into account. On the other hand, both executions have WDP-races:
Figure 4(a) has no branch dependent on the read, and Figure 4(c) has a branch,
but it occurs after rd(y).
WDP analysis discovers the predictable race in Figure 5. In this case, the
fact that there is no branch within the critical section allows WDP to avoid
creating an unnecessary transitive edge that otherwise would hide the race.
$\prec_{\textsc{\tiny{HB}}}$$(\prec_{\textsc{\tiny{WCP}}}\cup\prec_{\textsc{\tiny{PO}}})$$\prec_{\textsc{\tiny{DC}}}$$(\prec_{\textsc{\tiny{SDP}}}\cup\prec_{\textsc{\tiny{PO}}})$$\prec_{\textsc{\tiny{WDP}}}$
((a)) Weaker above stronger relations
HB-racesWCP-racesDC-racesSDP-racesWDP-races ((b)) Supersets above subsets
Table 3. Lattices showing the relationships among the relations and
corresponding kinds of races. Only WCP and SDP do not include PO and thus do
not in general order events within the same thread, but this property is
irrelevant for comparing relation strength because same-thread accesses cannot
race in any case, so the relation lattice uses
$(\prec_{\textsc{\tiny{WCP}}}\cup\prec_{\textsc{\tiny{PO}}})$ and
$(\prec_{\textsc{\tiny{SDP}}}\cup\prec_{\textsc{\tiny{PO}}})$ to make the
relations more directly comparable. (WCP and SDP analyses in fact detect races
by comparing access events using
$(\prec_{\textsc{\tiny{WCP}}}\cup\prec_{\textsc{\tiny{PO}}})$ and
$(\prec_{\textsc{\tiny{SDP}}}\cup\prec_{\textsc{\tiny{PO}}})$, respectively.)
| Sound? | Complete?
---|---|---
HB | Yes (Lamport, 1978) | No (Fig. 2(b), 2(a), 4(a), 4(c))
WCP | Yes (Kini et al., 2017) | No (Fig. 2(b), 2(a), 4(a), 4(c))
DC | No (Roemer et al., 2018) | No (Fig. 2(b), 2(a), 4(a), 4(c))
SDP | Yes (Section 4.2) | No (Fig. 2(b), 4(a), 4(c))
WDP | No (Fig. 2(a) and 2(c)) | Yes (Section 4.2)
Table 4. Soundness and completeness of each relation.
### 4.2. Soundness and Completeness
Figure 4 and Table 4 illustrate the relationships among the different
relations and corresponding race types. SDP never misses a race that WCP
finds, and WDP never misses a race that DC or SDP finds. SDP is sound but
incomplete (never reports a false race but may miss predictable races), while
WDP is unsound but complete (may report false races but never misses a
predictable race).
Here we prove that SDP is sound and WDP is complete. The proofs are manual and
have not been verified by a theorem prover.
###### Theorem 0 (SDP soundness).
If an execution trace has a SDP-race, then it has a predictable race or a
predictable deadlock.
###### Proof.
We define $\prec_{\textsc{\tiny{SDP}}(i)}$ to be a variant of
$\prec_{\textsc{\tiny{WCP}}}$ and $\prec_{\textsc{\tiny{SDP}}}$ that orders
critical sections like SDP for the first $i$ conflicting writes in
$\mathit{tr}$, and orders critical sections like WCP otherwise (Table 2).
Formally, for conflicting events $e$ and $e^{\prime}$ in critical sections on
the same lock, $e\prec_{\textsc{\tiny{SDP}}(i)}e^{\prime}$ if either:
* •
$e$ or $e^{\prime}$ is a read; or
* •
there are $i$ many conflicting pairs of write events (i.e., two conflicting
write events without an intervening conflicting write event) before the write
pair $(e,e^{\prime})$ (a write pair $(w,w^{\prime})$ is before a write pair
$(e,e^{\prime})$ if $w^{\prime}<_{tr}e^{\prime}$).
Note that $\prec_{\textsc{\tiny{SDP}}(0)}\equiv\prec_{\textsc{\tiny{WCP}}}$
and $\prec_{\textsc{\tiny{SDP}}(\infty)}\equiv\prec_{\textsc{\tiny{SDP}}}$.
The rest of the proof proceeds by induction to show that SDP($i$) is sound for
all $i$, i.e., if an execution trace has an SDP($i$)-race, then it has a
predictable race or deadlock.
##### Base case:
Since $\prec_{\textsc{\tiny{SDP}}(0)}\equiv\prec_{\textsc{\tiny{WCP}}}$ and
WCP is sound (Kini et al., 2017), SDP($0$) is sound.
##### Inductive step:
Let $\sigma$ be an execution trace whose first SDP($i$)-race is between events
$e_{1}$ and $e_{2}$, where _first_ means that $e_{2}$ is as early as possible
in $\sigma$, and among SDP($i$)-races whose second event is $e_{2}$, $e_{1}$
is as late as possible in $\sigma$. Proceeding with proof by contradiction,
suppose $\sigma$ has no predictable race or deadlock.
Now let $\mathit{tr}$ be a trace equivalent to $\sigma$ that moves all events
between $e_{1}$ and $e_{2}$ that are not HB ordered with both events, to
outside of $e_{1}$ and $e_{2}$, and additionally removes all events after
$e_{2}$. Specifically:
* •
if $e_{1}<_{\sigma}e\land
e_{1}\not\prec_{\textsc{\tiny{HB}}}e\prec_{\textsc{\tiny{HB}}}e_{2}$, move $e$
before $e_{1}$ in $\mathit{tr}$;
* •
if
$e_{1}<_{\sigma}e<_{\sigma}e_{2}\land{e\not\prec_{\textsc{\tiny{HB}}}e_{2}}$,
omit $e$ from $\mathit{tr}$;
* •
if $e_{2}<_{\sigma}e$, omit $e$ from $\mathit{tr}$.
Thus the last event in $\mathit{tr}$ is $e_{2}$. Like $\sigma$,
$\mathit{tr}$’s first SDP($i$)-race is between events $e_{1}$ and $e_{2}$, and
$\mathit{tr}$ has no predictable race or deadlock. That is,
$e_{1}<_{\textsc{$\mathit{tr}$}}e_{2}$, $e_{1}\asymp e_{2}$, and
$e_{1}\not\prec_{\textsc{\tiny{SDP}}}e_{2}$.
Because $\mathit{tr}$ has no predictable race or deadlock,
$e_{1}\prec_{\textsc{\tiny{SDP}}(i-1)}e_{2}$ by the induction hypothesis.
Because of the disparity between $e_{1}\prec_{\textsc{\tiny{SDP}}(i-1)}e_{2}$
and $e_{1}\not\prec_{\textsc{\tiny{SDP}}(i)}e_{2}$, it must be that
$l\prec_{\textsc{\tiny{SDP}}(i-1)}w^{\prime}$ and
$l\not\prec_{\textsc{\tiny{SDP}}(i)}w^{\prime}$, where $w^{\prime}$ is the
$i$th conflicting write in $\mathit{tr}$, $w$ is the latest write before
$w^{\prime}$ such that $w\asymp w^{\prime}$, and $l$ is the outermost release
event of a critical section containing $w$ that releases the same lock as any
critical section containing $w^{\prime}$.
If there is a read event $r$ that reads the same variable as $w$ and
$w^{\prime}$ such that
$w<_{\textsc{$\mathit{tr}$}}r<_{\textsc{$\mathit{tr}$}}e_{2}$, then either
* •
$w^{\prime}\asymp r\land r<_{\textsc{$\mathit{tr}$}}w^{\prime}$, in which case
$e_{1}\prec_{\textsc{\tiny{HB}}}r\prec_{\textsc{\tiny{SDP}}(i)}w^{\prime}\prec_{\textsc{\tiny{HB}}}e_{2}$;
* •
$w^{\prime}\asymp r\land w^{\prime}<_{\textsc{$\mathit{tr}$}}r$, in which case
$e_{1}\prec_{\textsc{\tiny{HB}}}w^{\prime}\prec_{\textsc{\tiny{SDP}}(i)}r\prec_{\textsc{\tiny{HB}}}e_{2}$;
* •
$w\asymp r\land w^{\prime}\prec_{\textsc{\tiny{PO}}}r$, in which case
$e_{1}\prec_{\textsc{\tiny{HB}}}l\prec_{\textsc{\tiny{SDP}}(i)}r\prec_{\textsc{\tiny{HB}}}e_{2}$;
or
* •
$w\asymp r\land r\prec_{\textsc{\tiny{PO}}}w^{\prime}$, in which case
$e_{1}\prec_{\textsc{\tiny{HB}}}w\prec_{\textsc{\tiny{SDP}}(i)}r\prec_{\textsc{\tiny{HB}}}e_{2}$.
In each of these cases, by SDP($i$)’s composition with HB,
$e_{1}\prec_{\textsc{\tiny{SDP}}(i)}e_{2}$, a contradiction.
Therefore there is no read event $r$ that reads the same variable as $w$ and
$w^{\prime}$ such that $w<_{\textsc{$\mathit{tr}$}}r$. Any read event that
reads the same variable as $w$ and $w^{\prime}$ must occur _before_ $w$.
Now consider the trace $\mathit{tr^{\prime}}$ that is equivalent to
$\mathit{tr}$ except:
* •
$w^{\prime}$ is replaced by a wr(x) event, where x is a brand-new variable not
used in $\mathit{tr}$.
* •
For every read $r$ in $\mathit{tr}$ that reads the same variable as
$w^{\prime}$, an event $r^{\prime}$ is appended immediately after $r$ such
that $r^{\prime}$ is a rd(x) event and $r\prec_{\textsc{\tiny{PO}}}r^{\prime}$
in $\mathit{tr^{\prime}}$.
Note that the SDP($i-1$) ordering for $\mathit{tr^{\prime}}$ is the same as
the SDP($i$) ordering for $\mathit{tr}$: the rd(x)–wr(x) conflicts introduce
the same ordering in $\mathit{tr^{\prime}}$ as the original read–write
conflicts between $w^{\prime}$ and its prior reads in $\mathit{tr}$, and
$\mathit{tr^{\prime}}$ does not contain the write–write conflict on $w$ and
$w^{\prime}$ found in $\mathit{tr}$. Thus in $\mathit{tr^{\prime}}$,
$e_{1}\not\prec_{\textsc{\tiny{SDP}}(i-1)}e_{2}$. By the induction hypothesis,
$\mathit{tr^{\prime}}$ has a predictable race or deadlock. Let
$\mathit{tr^{\prime\prime}}$ be a predictable trace of $\mathit{tr^{\prime}}$
that exposes a race or deadlock. However, if we modify
$\mathit{tr^{\prime\prime}}$ by removing the rd(x) events and replacing the
wr(x) event with $w^{\prime}$, the resulting trace is a predictable trace of
$\mathit{tr}$ that exposes a race or deadlock. Thus $\mathit{tr}$ has a
predictable race or deadlock, which is a contradiction.
Thus for all $i$, SDP($i$) is sound. Since
$\prec_{\textsc{\tiny{SDP}}(\infty)}\equiv\prec_{\textsc{\tiny{SDP}}}$,
therefore SDP is sound. ∎
###### Theorem 4.1 (WDP completeness).
If an execution trace has a predictable race, then it has a WDP-race.
To prove the theorem, we use the following helper lemma:
###### Lemma 4.2 (WDP-ordered events cannot be reordered).
Given an execution trace $\mathit{tr}$, for any events $e_{1}$ and $e_{2}$ in
$\mathit{tr}$ such that $e_{1}\prec_{\textsc{\tiny{WDP}}}e_{2}$, let
$\mathit{tr^{\prime}}$ be a reordering of $\mathit{tr}$ where $e_{1}$ and
$e_{2}$ have been reordered: either
$e_{2}<_{\textsc{$\mathit{tr^{\prime}}$}}e_{1}$ or
$e_{2}\in\mathit{tr^{\prime}}\land e_{1}\notin\mathit{tr^{\prime}}$. Then,
$\mathit{tr^{\prime}}$ must not be a valid predictable trace of $\mathit{tr}$.
The overall proof strategy is analogous to a corresponding proof for DC
(Roemer et al., 2018), so we have relegated the proof of Lemma 4.2 to Appendix
A.
###### Proof of Theorem 4.1.
Let us prove this theorem by contradiction. Let $\mathit{tr}$ be a trace with
a predictable race on conflicting events $e_{1}$ and $e_{2}$ such that
$e_{1}<_{\textsc{$\mathit{tr}$}}e_{2}$, but no WDP-race. Let
$\mathit{tr^{\prime}}$ be a predictable trace of $\mathit{tr}$ in which
$e_{1}$ and $e_{2}$ are consecutive:
$e_{1}<_{\textsc{$\mathit{tr^{\prime}}$}}e_{2}$ and $\nexists
e:e_{1}<_{\textsc{$\mathit{tr^{\prime}}$}}e<_{\textsc{$\mathit{tr^{\prime}}$}}e_{2}$.
Applying the definition of a WDP-race (Section 4), either
$e_{1}\prec_{\textsc{\tiny{WDP}}}e_{2}$ or
$\mathit{lockset}(e_{1})\cap\mathit{lockset}(e_{2})\neq\emptyset$. If
$\mathit{lockset}(e_{1})\cap\mathit{lockset}(e_{2})\neq\emptyset$, then
$\mathit{tr^{\prime}}$ violates the LS rule of predictable traces.
Thus $e_{1}\prec_{\textsc{\tiny{WDP}}}e_{2}$. By the definition of a
predictable race, $e_{1}$ and $e_{2}$ must be read or write events, and must
be on different threads. As a result, the WDP ordering between $e_{1}$ and
$e_{2}$ cannot be established by WDP conflicting critical section ordering or
“$\textsf{acq(m)}\prec_{\textsc{\tiny{WDP}}}\textsf{rel(m)}\implies\textsf{rel(m)}\prec_{\textsc{\tiny{WDP}}}\textsf{rel(m)}$,”
which require $e_{1}$ to be a release, and not by PO since the events are on
different threads. Therefore, $e_{1}\prec_{\textsc{\tiny{WDP}}}e_{2}$ by WDP
transitivity, so there must exist an event $e$ such that
$e_{1}\prec_{\textsc{\tiny{WDP}}}e\prec_{\textsc{\tiny{WDP}}}e_{2}$.
Since $e_{1}$ and $e_{2}$ are consecutive in $\mathit{tr^{\prime}}$, either
$e<_{\textsc{$\mathit{tr^{\prime}}$}}e_{1}$,
$e_{2}<_{\textsc{$\mathit{tr^{\prime}}$}}e$, or
$e_{2}\in\mathit{tr^{\prime}}\land e\notin\mathit{tr^{\prime}}$. By Lemma 4.2,
any of these possibilities implies $\mathit{tr^{\prime}}$ is an invalid
predictable trace of $\mathit{tr}$, a contradiction. ∎
### 4.3. Using Precise Dependence Information
Up to this point, we have assumed that a branch event depends on every
preceding read event in the same thread, meaning that the condition
$\mathit{brDepsOn}(b,e_{2})$ in WDP’s handling of write–read critical sections
holds for every read $e_{2}$ and branch $b$. This assumption is needed unless
static control dependence information is available from conservative static
program analysis (e.g., (Huang and Huang, 2017; Ferrante et al., 1987)). We
tried out one kind of static analysis to compute static control dependencies
but found it provided no benefit, so our experiments do not use it (Section
7). Here we show some examples of how WDP uses static control dependence
information if it is available.
Figure 6 shows two executions that differ only in whether precise static
control dependency information is available. Figure 6(a) has no control
dependency information available, so each branch is conservatively dependent
on all prior reads, or the information is available but the branch outcome
_may depend_ on the prior read. Figure 6(b) has control dependency information
that says that the branch outcome does _not_ depend on the prior read. As a
result, Figure 6(b) has weaker WDP ordering than Figure 6(a), leading to a
detected WDP-race in Figure 6(b) only.
Thread 1 Thread 2 wr(y) acq(m) wr(x) rel(m) acq(m) rd(x) br rel(m) rd(y)
((a)) No WDP-race exists if the branch event is dependent on the prior read.
Thread 1 Thread 2 wr(y) acq(m) wr(x) rel(m) acq(m) rd(x) br rel(m) rd(y)
((b)) A WDP-race exists if the branch event does not depend on the prior read.
Thread 1 Thread 2 acq(m) rd(x) rel(m) br wr(y) rd(y)
((c)) A valid reordering of (b) demonstrating that (b) has a race.
WDP
Figure 6. Example executions that differ only in the static control
dependencies between branches and reads. Dotted edges indicate reads that a
branch depends on, i.e., $\mathit{brDepsOn}(b,e)$. If precise control
dependence information rules out read–branch dependencies, WDP can find
additional races, such as the race on y in (b).
## 5\. SDP and WDP Analyses
Algorithm 1 SDP analysis at each event type, with differences from WCP
analysis
1:procedure acquire($t,l$)
2:
$\use@mathgroup\M@U\symAMSb{H}_{t}\leftarrow\use@mathgroup\M@U\symAMSb{H}_{t}\sqcup\use@mathgroup\M@U\symAMSb{H}_{l}$
3:
$\use@mathgroup\M@U\symAMSb{C}_{t}\leftarrow\use@mathgroup\M@U\symAMSb{C}_{t}\sqcup\use@mathgroup\M@U\symAMSb{C}_{l}$
4: foreach $t^{\prime}\neq t$ do
$Acq_{l}(t^{\prime}).Enque(\use@mathgroup\M@U\symAMSb{C}_{t}[t:=\use@mathgroup\M@U\symAMSb{H}_{t}(t)])$
5:procedure release($t,l,R,W$)
6: while
$Acq_{l}(t).Front()\sqsubseteq\use@mathgroup\M@U\symAMSb{C}_{t}[t:=\use@mathgroup\M@U\symAMSb{H}_{t}(t)]$
do
7: $Acq_{l}(t).Deque()$
8:
$\use@mathgroup\M@U\symAMSb{C}_{t}\leftarrow\use@mathgroup\M@U\symAMSb{C}_{t}\sqcup
Rel_{l}(t).Deque()$
9: foreach $x\in R$ do
$\use@mathgroup\M@U\symAMSb{L}^{r}_{l,x}\leftarrow\use@mathgroup\M@U\symAMSb{L}^{r}_{l,x}\sqcup\use@mathgroup\M@U\symAMSb{H}_{t}$
10: foreach $x\in W$ do
$\use@mathgroup\M@U\symAMSb{L}^{w}_{l,x}\leftarrow\use@mathgroup\M@U\symAMSb{L}^{w}_{l,x}\sqcup\use@mathgroup\M@U\symAMSb{H}_{t}$
11:
$\use@mathgroup\M@U\symAMSb{H}_{l}\leftarrow\use@mathgroup\M@U\symAMSb{H}_{t}$
12:
$\use@mathgroup\M@U\symAMSb{C}_{l}\leftarrow\use@mathgroup\M@U\symAMSb{C}_{t}$
13: foreach $t^{\prime}\neq t$ do
$Rel_{l}(t^{\prime}).Enque(\use@mathgroup\M@U\symAMSb{H}_{t})$
14:
$\use@mathgroup\M@U\symAMSb{H}_{t}(t)\leftarrow\use@mathgroup\M@U\symAMSb{H}_{t}(t)+1$
15:procedure read($t,x,L$)
16: $+$
$\use@mathgroup\M@U\symAMSb{C}_{t}\leftarrow\use@mathgroup\M@U\symAMSb{C}_{t}\sqcup\use@mathgroup\M@U\symAMSb{B}_{t,x}$$\triangleright$
Apply prior write–write conflict
17:
$\use@mathgroup\M@U\symAMSb{C}_{t}\leftarrow\use@mathgroup\M@U\symAMSb{C}_{t}\sqcup_{l\in
L}\use@mathgroup\M@U\symAMSb{L}^{w}_{l,x}$
18: check
$\use@mathgroup\M@U\symAMSb{W}_{x}\sqsubseteq\use@mathgroup\M@U\symAMSb{C}_{t}[t:=\use@mathgroup\M@U\symAMSb{H}_{t}(t)]$$\triangleright$
Write–read race?
19:
$\use@mathgroup\M@U\symAMSb{R}_{x}(t)\leftarrow\use@mathgroup\M@U\symAMSb{H}_{t}(t)$
20:procedure write($t,x,L$)
21:
$\use@mathgroup\M@U\symAMSb{C}_{t}\leftarrow\use@mathgroup\M@U\symAMSb{C}_{t}\sqcup_{l\in
L}\use@mathgroup\M@U\symAMSb{L}^{r}_{l,x}$
22:$-$
$\use@mathgroup\M@U\symAMSb{C}_{t}\leftarrow\use@mathgroup\M@U\symAMSb{C}_{t}\sqcup_{l\in
L}\use@mathgroup\M@U\symAMSb{L}^{w}_{l,x}$
23: check
$\use@mathgroup\M@U\symAMSb{W}_{x}\sqsubseteq\use@mathgroup\M@U\symAMSb{C}_{t}[t:=\use@mathgroup\M@U\symAMSb{H}_{t}(t)]\sqcup_{l\in
L}\use@mathgroup\M@U\symAMSb{L}^{w}_{l,x}$$\triangleright$ Write–write race?
24: $+$ $\use@mathgroup\M@U\symAMSb{B}_{t,x}\leftarrow\sqcup_{l\in
L}\use@mathgroup\M@U\symAMSb{L}^{w}_{l,x}$$\triangleright$ Records write–write
conflict for future read
25: check
$\use@mathgroup\M@U\symAMSb{R}_{x}\sqsubseteq\use@mathgroup\M@U\symAMSb{C}_{t}[t:=\use@mathgroup\M@U\symAMSb{H}_{t}(t)]$$\triangleright$
Read–write race?
26:
$\use@mathgroup\M@U\symAMSb{W}_{x}(t)\leftarrow\use@mathgroup\M@U\symAMSb{H}_{t}(t)$
27:procedure branch($t,L$)
28: skip $\triangleright$ No analysis at branch events
Algorithm 2 WDP analysis at each event type, with differences from DC analysis
1:procedure acquire($t,l$)
2: foreach $t^{\prime}\neq t$ do
$Acq_{l,t^{\prime}}(t).Enque(\use@mathgroup\M@U\symAMSb{C}_{t})$
3:procedure release($t,l,R,W$)
4: foreach $t^{\prime}\neq t$ do
5: while
$Acq_{l,t}(t^{\prime}).Front()\sqsubseteq\use@mathgroup\M@U\symAMSb{C}_{t}$ do
6: $Acq_{l,t}(t^{\prime}).Deque()$
7:
$\use@mathgroup\M@U\symAMSb{C}_{t}\leftarrow\use@mathgroup\M@U\symAMSb{C}_{t}\sqcup
Rel_{l,t}(t^{\prime}).Deque()$
8: foreach $x\in W$ do
$\use@mathgroup\M@U\symAMSb{L}_{l,x}\leftarrow\use@mathgroup\M@U\symAMSb{C}_{t}$
$\triangleright$ Record release time for writes in critical section
9:$-$ foreach $x\in R$ do
$\use@mathgroup\M@U\symAMSb{L}^{r}_{l,x}\leftarrow\use@mathgroup\M@U\symAMSb{C}_{t}$
10: foreach $t^{\prime}\neq t$ do
$Rel_{l,t^{\prime}}(t).Enque(\use@mathgroup\M@U\symAMSb{C}_{t})$
11:
$\use@mathgroup\M@U\symAMSb{C}_{t}(t)\leftarrow\use@mathgroup\M@U\symAMSb{C}_{t}(t)+1$
12:procedure read($t,x,e,L$)
13:$-$
$\use@mathgroup\M@U\symAMSb{C}_{t}\leftarrow\use@mathgroup\M@U\symAMSb{C}_{t}\bigsqcup_{l\in(L\cap
L^{w}_{x,t^{\prime}})}\use@mathgroup\M@U\symAMSb{L}_{l,x}$
14:$-$ check
$\use@mathgroup\M@U\symAMSb{W}_{x}\sqsubseteq\use@mathgroup\M@U\symAMSb{C}_{t}$
15: $+$ foreach thread $t^{\prime}\neq t$ check
$\use@mathgroup\M@U\symAMSb{W}_{x}(t^{\prime})\leq\use@mathgroup\M@U\symAMSb{C}_{t}(t^{\prime})\lor
L^{w}_{x,t^{\prime}}\cap L\neq\emptyset$ $\triangleright$ Check write–read
race
16: $+$ let $t^{\prime}\leftarrow T_{x}$ $\triangleright$ Get last writer
thread of $x$
17: if $+$ $t^{\prime}\notin\\{\varnothing,t\\}\land L\cap
L^{w}_{x,t^{\prime}}\neq\emptyset$ then $\triangleright$ Write–read conflict
18: $+$ $\use@mathgroup\M@U\symAMSb{B}_{t,x}\leftarrow\bigsqcup_{l\in(L\cap
L^{w}_{x,t^{\prime}})}\use@mathgroup\M@U\symAMSb{L}_{l,x}$ $\triangleright$
Record time of writer thread’s release for later use
19: if $+$
$\use@mathgroup\M@U\symAMSb{B}_{t,x}\nsqsubseteq\use@mathgroup\M@U\symAMSb{C}_{t}$
then $D_{t}\leftarrow D_{t}\cup\\{\langle x,e\rangle\\}$ $\triangleright$
Record read
20:
$\use@mathgroup\M@U\symAMSb{R}_{x}(t)\leftarrow\use@mathgroup\M@U\symAMSb{C}_{t}(t)$
21: $+$ $L^{r}_{x,t}\leftarrow L$
22:procedure write($t,x,L$)
23:$-$
$\use@mathgroup\M@U\symAMSb{C}_{t}\leftarrow\use@mathgroup\M@U\symAMSb{C}_{t}\sqcup\bigsqcup_{l\in(L\cap
L^{w}_{x,t^{\prime}})}\use@mathgroup\M@U\symAMSb{L}_{l,x}\sqcup\bigsqcup_{l\in(L\cap
L^{r}_{x,t^{\prime}})}\use@mathgroup\M@U\symAMSb{L}^{r}_{l,x}$
24:$-$ check
$\use@mathgroup\M@U\symAMSb{W}_{x}\sqsubseteq\use@mathgroup\M@U\symAMSb{C}_{t}$
25: $+$ foreach thread $t^{\prime}\neq t$ check
$\use@mathgroup\M@U\symAMSb{W}_{x}(t^{\prime})\leq\use@mathgroup\M@U\symAMSb{C}_{t}(t^{\prime})\lor
L^{w}_{x,t^{\prime}}\cap L\neq\emptyset$ $\triangleright$ Check write–write
race
26:$-$ check
$\use@mathgroup\M@U\symAMSb{R}_{x}\sqsubseteq\use@mathgroup\M@U\symAMSb{C}_{t}$
27: $+$ foreach thread $t^{\prime}\neq t$ check
$\use@mathgroup\M@U\symAMSb{R}_{x}(t^{\prime})\leq\use@mathgroup\M@U\symAMSb{C}_{t}(t^{\prime})\lor
L^{r}_{x,t^{\prime}}\cap L\neq\emptyset$ $\triangleright$ Check read–write
race
28:
$\use@mathgroup\M@U\symAMSb{W}_{x}(t)\leftarrow\use@mathgroup\M@U\symAMSb{C}_{t}(t)$
29: $+$ $L^{w}_{x,t}\leftarrow L$
30: $+$ $T_{x}\leftarrow t$ $\triangleright$ Set last writer thread of $x$
31:procedure branch($t,e,L$)
32:$-$ skip
33: $+$ $\textbf{foreach}\langle x,r\rangle\in
D_{t}:\mathit{brDepsOn}(e,r)\textbf{ do
}\use@mathgroup\M@U\symAMSb{C}_{t}\leftarrow\use@mathgroup\M@U\symAMSb{C}_{t}\sqcup\use@mathgroup\M@U\symAMSb{B}_{t,x}$
$\triangleright$ Add release–branch ordering
34: $+$ $D_{t}\leftarrow D_{t}\setminus\\{\langle x,r\rangle\in
D_{t}:\mathit{brDepsOn}(e,r)\\}$ $\triangleright$ Remove dependencies that
were applied
_SDP analysis_ and _WDP analysis_ are new online dynamic program analyses that
compute SDP and WDP and detect SDP- and WDP-races, respectively. Algorithms 1
and 2 show SDP and WDP analyses, respectively, for each kind of event. This
section’s notation and terminology follow the WCP and DC papers’ to some
extent (Roemer et al., 2018; Kini et al., 2017).
Both algorithms show the differences relative to prior analyses (SDP versus
WCP and WDP versus DC) by labeling lines with “$+$” to show logic added by our
analyses and “$-$” with grayed-out text to show lines removed by our analyses.
Algorithm 1 shows that SDP analysis requires few changes to WCP analysis.
These changes are for tracking write–write conflicts to add ordering when a
future read is detected on the second write’s thread. In addition, SDP
analysis avoids reporting write–write races for writes in critical sections on
the same lock by using $\use@mathgroup\M@U\symAMSb{L}^{w}_{l,x}$ at line 23.
Algorithm 2 shows that WDP analysis makes several significant changes to DC
analysis. These changes are primarily to deal with branches, by recording
information about write–read dependencies at read events (lines 16–19) and
using the recorded information at branch events (lines 33–34). Unlike DC
analysis, WDP analysis does not establish ordering at any conflicting
accesses, and it never needs to track ordering from a read to another access
(since it detects only write–read conflicts). In addition, WDP analysis
ensures it does not report races on accesses in critical sections on the same
lock, by maintaining and using the locksets $L^{w}_{x,t^{\prime}}$ and
$L^{r}_{x,t^{\prime}}$ when checking for races.
##### Analysis details
In both SDP and WDP analyses, the procedural parameters $t$ and $l$ are the
current thread and lock; $L$ is the set of locks held by the thread performing
the current event; $R$ and $W$ are the sets of variables that were read and
written in the ending critical section on $l$; and $e$ represents the current
read or branch event (for detecting branch dependencies).
The analysis uses _vector clocks_ (Mattern, 1988) to represent logical SDP or
WDP time. A vector clock $C:\mathit{Tid}\mapsto\mathcal{N}$ maps each thread
to a nonnegative integer. Operations on vector clocks are pointwise comparison
($C_{1}\sqsubseteq C_{2}\iff\forall t.C_{1}(t)\leq C_{2}(t)$) and pointwise
join ($C_{1}\sqcup C_{2}\equiv\lambda t.\mathit{\max(C_{1}(t),C_{2}(t))}$):
Both analyses maintain the following state:
* •
$\use@mathgroup\M@U\symAMSb{C}_{t}$ is a vector clock that represents the
current SDP or WDP time for thread $t$.
* •
$\use@mathgroup\M@U\symAMSb{R}_{x}$ and $\use@mathgroup\M@U\symAMSb{W}_{x}$
are vector clocks that represent the SDP or WDP time of the last reads and
writes to $x$.
* •
$\mathit{Acq}_{l}(t)$ and $\mathit{Rel}_{l}(t)$ (SDP) and
$\mathit{Acq}_{l,t^{\prime}}(t)$ and $\mathit{Rel}_{l,t^{\prime}}(t)$ (WDP)
are queues of vector clocks that help compute the
“$\textsf{acq(m)}\prec\textsf{rel(m)}$ implies
$\textsf{rel(m)}\prec\textsf{rel(m)}$” property (Table 2).
In addition, SDP analysis maintains the following state:
* •
$\use@mathgroup\M@U\symAMSb{H}_{t}$ is a vector clock that represents the
current HB time for thread $t$.
$\use@mathgroup\M@U\symAMSb{C}_{t}[t:=\use@mathgroup\M@U\symAMSb{H}_{t}(t)]$,
which evaluates to a vector clock with every element equal to
$\use@mathgroup\M@U\symAMSb{C}_{t}$ except that element $t$ is equal to
$\use@mathgroup\M@U\symAMSb{H}_{t}(t)$, represents
$\prec_{\textsc{\tiny{SDP}}}\cup\prec_{\textsc{\tiny{PO}}}$.
* •
$\use@mathgroup\M@U\symAMSb{B}_{t,x}$ is a vector clock that represents the
SDP time of a release event $e$ of a critical section on lock $m$ containing a
write event $w$ to $x$ such that a later write event $w^{\prime}$ to $x$ by
$t$ conflicts with the write and
$m\in\mathit{lockset}(w)\cap\mathit{lockset}(w^{\prime})$.
* •
$\use@mathgroup\M@U\symAMSb{L}^{r}_{l,x}$ and
$\use@mathgroup\M@U\symAMSb{L}^{w}_{l,x}$ are vector clocks that represent the
HB times of critical sections on $l$ that read and wrote $x$, respectively.
WDP analysis maintains the following additional state:
* •
$\use@mathgroup\M@U\symAMSb{L}_{l,x}$ is a vector clock that represents the
WDP time of critical sections on $l$ that wrote $x$.
* •
$\use@mathgroup\M@U\symAMSb{B}_{t,x}$ is a vector clock that represents the
WDP time of the last write event $e$ to $x$ such that a later read event
$e^{\prime}$ to $x$ by $t$ conflicts with $e$ and
$\mathit{lockset}(e)\cap\mathit{lockset}(e^{\prime})\neq\emptyset$.
* •
$T_{x}$ is the last thread to write to $x$, or $\varnothing$ if no thread has
yet written $x$.
* •
$D_{t}$ is a set of pairs $\langle x,e\rangle$ such that event $e$ is a read
to $x$ by a thread that has not (yet) executed a branch $b$ such that
$\mathit{brDepsOn}(b,e)$.
* •
$L^{r}_{x,t}$ and $L^{w}_{x,t}$ are sets of locks that were held by thread $t$
when it last read and wrote variable $x$, respectively.
Initially, every vector clock maps every thread to 0, except $\forall
t.\use@mathgroup\M@U\symAMSb{H}_{t}(t)=1$ for SDP analysis, and $\forall
t.\use@mathgroup\M@U\symAMSb{C}_{t}(t)=1$ for WDP analysis. Every queue and
set is initially empty.
The analyses provide SDP and WDP’s handling of conflicting critical sections
by detecting some kinds of conflicts on accesses holding a common lock. SDP
analysis orders the earlier release of a common lock to the current event for
write–read and read–write conflicts using
$\use@mathgroup\M@U\symAMSb{L}^{r}_{l,x}$ and
$\use@mathgroup\M@U\symAMSb{L}^{w}_{l,x}$ (lines 17 and 21 in Algorithm 1).
For write–write conflicts, SDP analysis stores the time of the earlier release
of a common lock in $\use@mathgroup\M@U\symAMSb{B}_{t,x}$ (line 24) to order
the earlier write to a later read of $x$ by the current thread (line 16).
WDP analysis orders the release on the writer’s executing thread to a later
branch dependent on the read. The analysis does so by recording the time of
the last writer’s release in $\use@mathgroup\M@U\symAMSb{L}_{l,x}$ (line 8 in
Algorithm 2). Later, when a conflicting read occurs on thread $t$ holding $l$,
the analysis uses $\use@mathgroup\M@U\symAMSb{L}_{l,x}$ to get the time for
the last conflicting writer $T_{x}$’s release, and stores this time in
$\use@mathgroup\M@U\symAMSb{B}_{t,x}$ (line 18). When $t$ executes a branch
dependent on the prior conflicting read, WDP adds ordering from the release to
the current branch (line 33). The analysis detects the dependent branch using
$D_{t}$, which contains a set of $\langle x,e\rangle$ pairs for which a branch
dependent on read event $e$ has not yet executed (line 30). The exact
representation of $e$ and behavior of $\mathit{brDepsOn}(b,e)$ are
implementation dependent.
The analyses compute the “$\textsf{acq(m)}\prec\textsf{rel(m)}$ implies
$\textsf{rel(m)}\prec\textsf{rel(m)}$” property (Table 2) in the same way as
WCP and DC analyses, respectively (Kini et al., 2017; Roemer et al., 2018).
Briefly, $\mathit{Acq}_{l}(t)$ and $\mathit{Rel}_{l}(t)$ contain times of
acq(l) and rel(l) operations (by any thread other than $t$) such that the
acq(l) operation is not yet SDP ordered to a following rel(l) by thread $t$.
$\mathit{Acq}_{l,t^{\prime}}(t)$ and $\mathit{Rel}_{l,t^{\prime}}(t)$ contain
times of acq(l) and rel(l) operations by thread $t$ such that the acq(l)
operation is not yet WDP ordered to a following rel(l) by thread $t^{\prime}$.
SDP analysis provides composition with HB using
$\use@mathgroup\M@U\symAMSb{C}_{t}$, $\use@mathgroup\M@U\symAMSb{H}_{t}$, and
$\use@mathgroup\M@U\symAMSb{C}_{t}[t:=\use@mathgroup\M@U\symAMSb{H}_{t}(t)]$.444In
Algorithm 1, $\use@mathgroup\M@U\symAMSb{C}_{t}$ and
$\use@mathgroup\M@U\symAMSb{C}_{t}[t:=\use@mathgroup\M@U\symAMSb{H}_{t}(t)]$
are analogous to the WCP paper’s $\use@mathgroup\M@U\symAMSb{P}_{t}$ and
$\use@mathgroup\M@U\symAMSb{C}_{t}$, respectively (Kini et al., 2017). WDP
analysis includes PO with the increment of
$\use@mathgroup\M@U\symAMSb{C}_{t}(t)$ at line 11.
The analyses check the conditions for a SDP- or WDP-race by using
$\use@mathgroup\M@U\symAMSb{R}_{x}$ and $\use@mathgroup\M@U\symAMSb{W}_{x}$.
Since the analyses do not order all pairs of conflicting accesses, unordered
conflicting accesses are not sufficient to report a race. SDP analysis uses
the vector clock $\use@mathgroup\M@U\symAMSb{L}^{w}_{l,x}$ and WDP analysis
uses the locksets $L^{r}_{x,t}$ and $L^{w}_{x,t}$ to check if the current and
prior conflicting accesses’ held locks overlap (lines 18, 23, and 25 in
Algorithm 1; lines 15, 25, and 27 in Algorithm 2).
##### Atomic accesses and operations.
We extend SDP and WDP analyses to handle accesses that have ordering or
atomicity semantics: _atomic accesses_ that introduce ordering such as Java
volatile and C++ atomic accesses, and _atomic read-modify-write operations_
such as atomic test-and-set.
The following pseudocode shows how we extend WDP analysis (Algorithm 2) to
handle atomic reads and writes and atomic operations. (The extensions to SDP
analysis are similar but also add conflicting read–write and write–write–read
ordering.)
1:procedure atomicRead($t,x,e$)
2: let $t^{\prime}\leftarrow T_{x}$ $\triangleright$ Get last writer thread of
$x$
3: if
$t^{\prime}\notin\\{\varnothing,t\\}\land\use@mathgroup\M@U\symAMSb{W}_{x}\nsqsubseteq\use@mathgroup\M@U\symAMSb{C}_{t}$
then $\triangleright$ Write–read conflict
4:
$\use@mathgroup\M@U\symAMSb{B}_{t,x}\leftarrow\use@mathgroup\M@U\symAMSb{W}_{x}$
$\triangleright$ Record the write
5: if
$\use@mathgroup\M@U\symAMSb{B}_{t,x}\nsqsubseteq\use@mathgroup\M@U\symAMSb{C}_{t}$
then $D_{t}\leftarrow D_{t}\cup\\{\langle x,e\rangle\\}$
6:procedure atomicWrite($t,x$)
7:
$\use@mathgroup\M@U\symAMSb{W}_{x}\leftarrow\use@mathgroup\M@U\symAMSb{C}_{t}$
8: $T_{x}\leftarrow t$ $\triangleright$ Set last writer thread of $x$
9:procedure atomicReadModifyWrite($t,x,e$)
10: atomicRead($t$, $x$, $e$)
11: atomicWrite($t$, $x$)
In essence, the analysis handles atomic accesses like regular accesses
contained in single-access critical sections on a unique lock to the accessed
variable. The analysis treats an atomic operation as an atomic read followed
by an atomic write.
##### Handling races.
The behavior of programs with data races is unreliable (Adve and Boehm, 2010;
Dolan et al., 2018), but our analyses’ instrumentation performs
synchronization operations before accesses, which generally ensures sequential
consistency (SC) for all program executions. A different problem is that if an
analysis continues detecting races after the first race, then additional
detected races are not necessarily real races because they may depend on an
earlier race (i.e., if the earlier race were ordered, the later race would not
exist). Our implementation (Section 7.1) addresses this issue by treating
racing accesses as if they were contained in single-access critical sections
on the same lock. Specifically, SDP analysis orders one racing event to the
other for write–read and read–write races, and WDP analysis orders write–read
races to a branch that depends on the read if the write is the last writer of
the read. For example, after detecting a race in line 15, WDP analysis
performs the following:
$\textbf{if}\>t^{\prime}=T_{x}\>\textbf{then}\;\use@mathgroup\M@U\symAMSb{B}_{t,x}\leftarrow\use@mathgroup\M@U\symAMSb{W}_{x}(t^{\prime});\>D_{t}\leftarrow
D_{t}\cup\\{\langle x,e\rangle\\}$.
##### Time and space complexity.
SDP analysis and WDP analysis’s run times are each linear in the number of
events. Like WCP and DC analyses (Kini et al., 2017; Roemer et al., 2018), for
$N$ events, $L$ locks, and $T$ threads, time complexity for an entire
execution trace is $O(N\times(L\times T+T^{2}))$. Briefly, SDP or WDP analysis
at a read or write takes $O(L\times T+T^{2})$ time even considering the
additional computation it performs compared with WCP or DC analysis; and WDP
analysis’s run time at branch events can be amortized over read events.
## 6\. Verifying WDP-Races
WDP analysis is unsound, so a WDP-race may not indicate a predictable race
(i.e., there may be no race exposed in any predictable trace). To avoid
reporting false races, our approach post-processes each detected WDP-race with
an algorithm called VindicateWDPRace. Here we overview VindicateWDPRace;
Appendix B presents VindicateWDPRace in detail with an algorithm and examples.
To support performing VindicateWDPRace on WDP-races, WDP analysis builds a
_constraint graph_ in which execution events are nodes, and initially edges
represent WDP ordering. VindicateWDPRace discovers and adds additional
constraints to the graph that enforce lock semantics (LS) and last-writer (LW)
rules. VindicateWDPRace uses the constraint graph to attempt to construct a
predictable trace that exposes the WDP-race as a predictable race.
VindicateWDPRace extends prior work’s VindicateDCRace algorithm for checking
DC-races (Roemer et al., 2018). VindicateWDPRace differs from VindicateDCRace
primarily in the following way. VindicateWDPRace computes and adds LW
constraints to the constraint graph for all reads that must be causal events
in the predictable trace. Importantly, VindicateWDPRace computes causal reads
and adds LW constraints at each iteration of adding constraints and at each
attempt at building a predictable trace.
Algorithm 3 shows VindicateWDPRace at a high level; Appendix B presents
VindicateWDPRace in detail. VindicateWDPRace takes the initial constraint
graph ($G$) and a WDP-race ($e_{1},e_{2}$) as input (line 1). It first calls
AddConstraints (line 2), which adds necessary constraints to $G$ and returns
an updated constraint graph. AddConstraints first adds _consecutive-event_
constraints (i.e., edges) to $G$ to enforce that any predictable trace must
execute $e_{1}$ and $e_{2}$ consecutively (line 9). AddConstraints then
computes the set of causal events for any predictable trace constrained by
$G$, which it uses to add LW constraints to $G$, ensuring that every causal
read in a predictable trace can have the same last writer as in the original
trace (line 11). Next, AddConstraints adds LS constraints to $G$, by
identifying critical sections on the same lock that are partially ordered and
thus must be fully ordered to obey LS rules (line 12). Since added LW and LS
constraints may lead to new LS and LW constraints being detected,
respectively, AddConstraints iterates until it finds no new constraints to add
(lines 10–13).
Algorithm 3 Check if WDP-race is a true predictable race (high-level version
of algorithm)
1:procedure VindicateWDPRace($G,e_{1},e_{2}$) $\triangleright$ Inputs:
constraint graph and WDP-race events
2: $G\leftarrow\textsc{AddConstraints}(G,e_{1},e_{2})$
3: if $G$ has a cycle reaching $e_{1}$ or $e_{2}$ then return No predictable
race
4: else
5:
$\mathit{tr^{\prime}}\leftarrow\textsc{ConstructReorderedTrace}(G,e_{1},e_{2})$
$\triangleright$ Non-empty iff predictable trace constructed
6: if $\mathit{tr^{\prime}}\neq\langle\,\rangle\,$ then return Predictable
race witnessed by $\mathit{tr^{\prime}}$ $\triangleright$ Check for non-empty
trace
7: else return Don’t know
8:procedure AddConstraints($G,e_{1},e_{2}$)
9: Add consecutive-event constraints to $G$
10: do
11: Compute causal reads and add last-writer (LW) constraints to $G$
12: Add lock-semantics (LS) constraints to $G$
13: while $G$ has changed
14: return $G$
The constraints added by AddConstraints are necessary but insufficient
constraints on any trace exposing a predictable race on $e_{1}$ and $e_{2}$.
Thus if $G$ has a cycle that must be part of any predictable trace, then the
original trace has no predictable race on $e_{1}$ and $e_{2}$ (line 3).
Otherwise, AddConstraints calls ConstructReorderedTrace (line 5), which
attempts to construct a legal predictable trace $\mathit{tr^{\prime}}$.
ConstructReorderedTrace is a greedy algorithm that starts from $e_{1}$ and
$e_{2}$ and works backward, adding events in reverse order that satisfy $G$’s
constraints and also conform to LS and LW rules ($G$’s constraints are
necessary but insufficient). If ConstructReorderedTrace returns a (non-empty)
trace $\mathit{tr^{\prime}}$, it is a legal predictable trace exposing a race
on $e_{1}$ and $e_{2}$ (line 6). Otherwise, ConstructReorderedTrace returns an
empty trace, which means that it could not find a predictable race, although
one may exist (line 7).
## 7\. Evaluation
This section evaluates the predictive race detection effectiveness and run-
time performance of this paper’s approaches.
### 7.1. Implementation
We implemented SDP and WDP analyses and VindicateWDPRace by extending the
publicly available _Vindicator_ implementation, which includes HB, WCP, and DC
analyses and VindicateDCRace (Roemer et al.,
2018).555https://github.com/PLaSSticity/Vindicator Vindicator is built on top
of _RoadRunner_ , a dynamic analysis framework for concurrent Java programs
(Flanagan and Freund,
2010b).666https://github.com/stephenfreund/RoadRunner/releases/tag/v0.5 We
extended RoadRunner to instrument branches to enable WDP analysis at program
branches. RoadRunner operates on the Java bytecode of analyzed programs, so
analysis properties such as SDP soundness and WDP completeness hold with
respect to the execution of the bytecode, even if the JVM compiler optimizes
away control or data dependencies.
Our implementation of SDP and WDP analyses and VindicateWDPRace is publicly
available.777https://github.com/PLaSSticity/SDP-WDP-implementation
We evaluated _Joana_ to perform static analysis for detecting whether a branch
depends on prior reads or not (Giffhorn and Hammer, 2008), following the
system dependency graphs used in MCR-S (Huang and Huang, 2017). We found no
practical advantages to using Joana. In most programs, for the vast majority
of the write–read branch dependencies executed, the next branch after the read
is dependent on the read according to Joana. In pmd and sunflow, static
analysis reported many write–read dependencies where the following branch did
not depend on the read, but this did not lead to any additional WDP-races
being detected. It is unclear whether these results are mainly due to
properties of the evaluated programs (i.e., if almost all branches do depend
on prior reads) or imprecision of Joana’s static analysis. Our implementation
and evaluation do not use static analysis, and instead assume that branches
always depend on prior reads.
##### SDP and WDP analyses.
We implemented a single analysis tool within RoadRunner that can perform HB,
WCP, DC, SDP, and WDP analyses on a single observed execution. The
implementation of HB, WCP, and DC analyses are taken from the Vindicator
implementation, and implementation of SDP and WDP analyses follows Algorithms
1 and 2. For thread fork and join (including implicitly forked threads (Roemer
et al., 2018)) and static class initializer edges (Lindholm and Yellin, 1999),
each analysis adds appropriate ordering between the two synchronizing events.
The analyses treat calls to m.wait() as a release of m followed by an acquire
of m. The analyses instrument volatile variable accesses as _atomic accesses_
as described in Section 5. The analyses can in theory handle lock-free data
structures, such as data structures in java.util.concurrent, by handling
atomic operations as in Section 5. However, RoadRunner instruments only
application code, not Java library code, and it does not intercept underlying
atomic operations (e.g., by instrumenting calls to atomic sun.misc.Unsafe
methods). The analyses may thus miss some synchronization in the evaluated
programs.
The analyses can determine that some observed events are “redundant” and
cannot affect the analysis results. For a read or write event, if the same
thread has performed a read or write, respectively, to the same variable
without an intervening synchronization operation, then the access is
redundant. For a branch event, if the same thread has not performed a read
event since the last branch event, then the branch is redundant (since our
implementation assumes that a branch is dependent on all prior reads). The
implementation “fast path” detects and filters redundant events, and does not
perform analysis for them.
The implementation is naturally parallel because application threads running
in parallel perform analysis. The implementation uses fine-grained
synchronization on metadata to ensure atomicity of the analysis for an event.
For WDP analysis, to obtain an approximation of $<_{\textsc{$\mathit{tr}$}}$
(needed by vindication; see line 44 of Algorithm 4 in Appendix B), the
implementation assigns each event node in the constraint graph a Lamport
timestamp (Lamport, 1978) that respects HB order:
$e\prec_{\textsc{\tiny{HB}}}e^{\prime}\implies\mathit{ts}(e)<\mathit{ts}(e^{\prime})$.
##### Handling races.
To keep finding real races after the first detected race, whenever an analysis
detects a race, it updates vector clocks (and WDP’s constraint graph) so that
the execution so far is race free. SDP and WDP analyses treat racing accesses
as though minimal critical sections on the same lock protected them, as
described in Section 5. HB, WCP, and DC analyses handle detected races by
adding ordering between all accesses.
If an analysis detects multiple races involving the current access, it reports
only one of the races but adds ordering to eliminate all of the races.
##### Vindication.
WDP analysis constructs a constraint graph representing the observed
execution’s WDP ordering. When the execution completes, the implementation
calls VindicateWDPRace on a configurable subset of the WDP-races, e.g., each
WDP-race that is not also a SDP-race.
### 7.2. Methodology
The experiments execute large, real Java programs harnessed as the DaCapo
benchmarks (Blackburn et al., 2006), version 9.12-bach. We use a version of
the DaCapo programs that the RoadRunner authors have modified to work with
RoadRunner;888https://github.com/stephenfreund/RoadRunner/releases/tag/v0.5
the resulting workloads are approximately equal to DaCapo’s default workload.
The experiments exclude DaCapo programs eclipse, tradebeans, and tradesoap,
which the RoadRunner authors have not modified to run with RoadRunner; jython,
which failed to run with RoadRunner in our environment; and the single-
threaded program fop.
The experiments execute on a quiet Intel Xeon E5-4620 with four 8-core
processors with hyperthreading disabled and 256 GB of main memory, running
Linux 3.10.0. We execute RoadRunner with the HotSpot 1.8.0 JVM and set the
maximum heap size to 245 GB.
We run various combinations of the analyses to collect race results and
statistics and measure performance. To account for run-to-run variation, each
reported result is the mean of five trials.
Each WDP-race in an execution is a _dynamic_ WDP-race (similarly for SDP-,
DC-, WCP-, and HB-races). Among dynamic WDP-races, some may be detected at the
same static accesses. If two dynamic WDP-races have the same two static source
location regardless of order, then they are the same _static_ WDP-race
(similarly for SDP-, DC-, WCP-, and HB-races).
### 7.3. Dynamic Characteristics
Table 5 shows properties of the analyzed programs. The _#Thr_ column reports
total threads created by an execution and, in parentheses, threads active at
termination. The rest of the columns count events from WDP analysis; other
analyses are similar but exclude branch events. _Total events_ are all
executed events instrumented by the analysis.
| | | Total | Analyzed events
---|---|---|---|---
| #Thr | events | All ( | acq/rel | wr | rd | br )
avrora | 7 | (7) | 2,400 M | 260 M ( | 1.2% | 17.7% | 42.2% | 38.4% )
batik | 7 | (7) | 490 M | 17 M ( | 0.6% | 26.3% | 38.4% | 34.0% )
h2 | 34 | (33) | 9,368 M | 768 M ( | 0.5% | 17.1% | 43.1% | 39.1% )
luindex | 3 | (3) | 910 M | 72 M ( | 0.6% | 20.1% | 42.4% | 36.9% )
lusearch | 34 | (34) | 2,746 M | 301 M ( | 0.9% | 19.5% | 43.6% | 35.6% )
pmd | 33 | (33) | 403 M | 41 M ( | $<\,$0.1% | 28.5% | 37.3% | 34.2% )
sunflow | 65 | (33) | 14,452 M | 887 M ( | $<\,$0.1% | 44.7% | 41.4% | 13.8% )
tomcat | 106 | (67) | 113 M | 29 M ( | 2.8% | 18.7% | 42.1% | 36.1% )
xalan | 33 | (33) | 1,306 M | 436 M ( | 2.1% | 12.0% | 48.8% | 37.1% )
Table 5. Dynamic characteristics of the analyzed programs. Event counts (shown
in millions) and percentages are collected from WDP analysis; other analyses
do not analyze branch events.
_Analyzed events_ are the events _not_ filtered by the fast path that detects
redundant events. The rest of the columns show the breakdown of analyzed
events by event type. The percentages do not add up to 100% because they do
not include other events (e.g., fork, join, wait, volatile access, and static
class initializer events), which are always less than 1% of analyzed events.
Unsurprisingly, most analyzed events are memory accesses or branches.
### 7.4. Race Detection Effectiveness
Table 6 reports detected races for two different experiments that each run a
combination of analyses on the same executions. Table 6(a)’s results are from
an experiment that runs HB, WCP, and SDP analyses together on the same
executions, to compare these analyses’ race detection capabilities directly.
Likewise, a separate experiment runs DC and WDP analyses together on the same
executions to make them directly comparable, resulting in Table 6(b)’s
results.
For each race count, the first value is static races, followed by dynamic
races in parentheses. For example, on average over the five trials, the
analysis detects about 406,000 WDP-races for avrora, which each correspond to
one of 5 different unordered pairs of static program locations.
Program | HB-races | WCP-races | SDP-races | |
---|---|---|---|---|---
avrora | 5 | (205 K) | 5 | (206 K) | 5 | (206 K) | |
batik | 0 | (0) | 0 | (0) | 0 | (0) | |
h2 | 9 | (52 K) | 9 | (52 K) | 9 | (53 K) | |
luindex | 1 | (1) | 1 | (1) | 1 | (1) | |
lusearch | 0 | (0) | 0 | (0) | 0 | (0) | |
pmd | 6 | (351) | 6 | (354) | 8 | (562) | |
sunflow | 2 | (19) | 2 | (25) | 2 | (25) | |
tomcat | 85 | (34 K) | 86 | (34 K) | 91 | (38 K) | |
xalan | 6 | (203) | 21 | (520 K) | 52 | (2.2 M) | |
((a)) HB, WCP, and SDP analyses on the same executions.
Program | | | | DC-races | WDP-races
---|---|---|---|---|---
avrora | | | | 5 | (203 K) | 5 | (406 K)
batik | | | | 0 | (0) | 0 | (0)
h2 | | | | 11 | (54 K) | 12 | (63 K)
luindex | | | | 1 | (1) | 1 | (1)
lusearch | | | | 0 | (0) | 1 | (30)
pmd | | | | 9 | (2 K) | 10 | (3 K)
sunflow | | | | 2 | (49) | 2 | (100)
tomcat | | | | 94 | (36 K) | 284 | (125 K)
xalan | | | | 17 | (649 K) | 170 | (15 M)
((b)) DC and WDP analyses on the same executions.
Table 6. Static and dynamic (in parentheses) race counts from two different
experiments.
Table 6(a) shows that SDP analysis finds significantly more races than not
only HB analysis but also WCP analysis—the state of the art in unbounded sound
predictive race detection (Flanagan and Freund, 2017; Kini et al., 2017).
These additional races are due to SDP incorporating data dependence more
precisely than WCP by not ordering write–write conflicting critical sections,
essentially permitting predictable traces that swap writes without changing a
causal read’s last writer.
Likewise, Table 6(b) shows that WDP analysis finds more races than DC
analysis, the state of the art in high-coverage unbounded predictive race
detection (Roemer et al., 2018). These additional races result from WDP being
more precise with respect to both data and control dependence than DC, and in
fact being complete.
The counts of HB-, WCP-, and DC-races we report here are significantly
different from those reported by the Vindicator paper (Roemer et al., 2018).
(While the counts are not directly comparable, both papers show similar trends
between relations.) The most significant cause of this effect is that
RoadRunner stops tracking a field after the field has 100 races, a behavior
that Vindicator used but that we disabled for these results to avoid
artificially underreporting race counts. Furthermore, our analyses do not use
a Vindicator optimization that merges events, reducing the number of races
reported when there are multiple races between synchronization-free regions.
We disabled this optimization because WDP analysis must track variable access
information for each event, negating the advantages of this optimization.
Another difference is that the Vindicator experiments spawned fewer threads
for some benchmarks by setting the number of available cores to 8.
Table 8 reports results from an experiment that runs SDP and WDP analyses
together and then performs VindicateWDPRace on _WDP-only races_ , which are
WDP-races that are not also SDP-races. The _SDP-races_ and _WDP-races_ columns
report static and dynamic races, as in Table 6. The _WDP-only_ column is
_static_ WDP-only races, which are static WDP-races that have no dynamic
instances that are SDP-races. The last column, _WDP-only $\rightarrow$
Verified_, reports how many static WDP-only races are detected and how many
are successfully vindicated as true races by VindicateWDPRace. In this
experiment, the implementation tries to vindicate up to 10 dynamic instances
of each static WDP-only race. The implementation first attempts to vindicate
the five earliest dynamic instances of a static WDP-only race, then five
random dynamic instances, stopping as soon as it verifies any dynamic instance
of the static race.
Program | SDP-races | WDP-races | WDP-only$\;\;\rightarrow\;\;$ | Verified
---|---|---|---|---
avrora | 5 | (202 K) | 5 | (407 K) | 0$\;\;{\color[rgb]{1,1,1}\definecolor[named]{pgfstrokecolor}{rgb}{1,1,1}\pgfsys@color@gray@stroke{1}\pgfsys@color@gray@fill{1}\rightarrow}\;\;$ |
batik | 0 | (0) | 0 | (0) | 0$\;\;{\color[rgb]{1,1,1}\definecolor[named]{pgfstrokecolor}{rgb}{1,1,1}\pgfsys@color@gray@stroke{1}\pgfsys@color@gray@fill{1}\rightarrow}\;\;$ |
h2 | 12 | (53 K) | 13 | (63 K) | 1$\;\;\rightarrow\;\;$ | 0
luindex | 1 | (1) | 1 | (1) | 0$\;\;{\color[rgb]{1,1,1}\definecolor[named]{pgfstrokecolor}{rgb}{1,1,1}\pgfsys@color@gray@stroke{1}\pgfsys@color@gray@fill{1}\rightarrow}\;\;$ |
lusearch | 0 | (0) | 1 | (30) | 1$\;\;\rightarrow\;\;$ | 0
pmd | 9 | (456) | 10 | (3 K) | 1$\;\;\rightarrow\;\;$ | 1
sunflow | 2 | (32) | 2 | (100) | 0$\;\;{\color[rgb]{1,1,1}\definecolor[named]{pgfstrokecolor}{rgb}{1,1,1}\pgfsys@color@gray@stroke{1}\pgfsys@color@gray@fill{1}\rightarrow}\;\;$ |
tomcat | 98 | (37 K) | 334 | (128 K) | 236$\;\;\rightarrow\;\;$ | 60
xalan | 31 | (1.7 M) | 170 | (15 M) | 139$\;\;\rightarrow\;\;$ | 137
Table 7. Static and dynamic (in parentheses) race counts from an experiment
running SDP and WDP analyses together and vindicating dynamic instances of
static WDP-only races. The _WDP-only $\rightarrow$ Verified_ column reports
static WDP-only races, followed by how many static WDP-only races were
verified as predictable races by VindicateWDPRace.
| Mean$\;\;\pm\;$ | Stdev | Max
---|---|---|---
pmd | 24,200$\;\;\pm\;$ | 14,100 | 40,294
tomcat | 4,830,000$\;\;\pm\;$ | 5,540,000 | 28,734,020
xalan | 52,100$\;\;\pm\;$ | 90,500 | 751,701
Table 8. Characteristics of the distribution of event distances of WDP-only
races that are verified predictable races. The table rounds the mean and
standard deviation to three significant digits.
Around half of the static WDP-only races are verified predictable races: out
of 378 static WDP-only races on average, 198 are verified predictable races.
As Section 7.5 shows, it can take a few minutes for VindicateWDPRace to check
a WDP-race. Given the difficulty and importance of detecting unknown, hard-to-
expose data races in mature software—and the amount of time developers
currently spend on testing and debugging—the time for VindicateWDPRace is
reasonable.
We confirmed that in all of the experiments, SDP analysis detected every race
detected by WCP analysis, and WDP analysis detected every race detected by DC
or SDP analysis.
##### Race characteristics.
SMT-solver-based predictive race detectors can be as precise as SDP and WDP
analyses, but cannot scale to unbounded program executions (Huang et al.,
2014; Said et al., 2011; Huang and Rajagopalan, 2016; Chen et al., 2008;
Şerbănuţă et al., 2013; Liu et al., 2016) (Section 8). These approaches
typically analyze bounded windows of an execution trace, missing races
involving “far apart” events. We can estimate whether SMT-based approaches
would miss a predictable race by computing the race’s _event distance_ , which
is the number of events in the execution trace between the race’s two events.
Since our implementation does not compute a total order of events, it
approximates event distance using Lamport timestamps: event distance is the
number of events $e$ such that
$\mathit{ts}(e_{1})<\mathit{ts}(e)<\mathit{ts}(e_{2})$.
Table 8 reports the distribution of event distances between accesses in each
successfully vindicated WDP-only race (i.e., the last column of Table 8). The
average distance and standard deviation are across all trials. The last column
reports the greatest distance found among all trials.
### 7.5. Performance
Table 9 reports the run-time performance of various combinations of analyses.
_Base_ is native execution time without any instrumentation. Other columns
(excluding _Failed_ and _Verified_) are slowdowns relative to _Base_.
The _Instr. only_ columns are RoadRunner configurations that instrument events
(excluding or including branches) but perform no analysis in the
instrumentation.
The _Analyses w/o constraint graph_ show configurations that do not construct
a constraint graph. Only configurations including WDP analysis instrument
branch events. The _WCP_ and _SDP_ columns show the slowdowns from running the
WCP and SDP analyses independently; the performance difference between them is
modest, suggesting that there is no significant performance penalty from using
SDP analysis over WCP analysis. (SDP’s performance improvement over WCP for h2
is not statistically significant, according to confidence intervals in Table
14 in Appendix C.)
| | Instr. only | Analyses w/o constraint graph | SDP+WDP+graph
---|---|---|---|---
| Base | w/o br | w/ br | WCP | SDP | SDP+DC | SDP+WDP | Slowdown | Failed | Verified
avrora | 6.0 s | 2.7$\;\\!\times$ | 3.4$\;\\!\times$ | 21$\;\\!\times$ | 23$\;\\!\times$ | 32$\;\\!\times$ | 38$\;\\!\times$ | 50$\;\\!\times$ | - | -
batik | 4.2 s | 3.2$\;\\!\times$ | 4.4$\;\\!\times$ | 14$\;\\!\times$ | 14$\;\\!\times$ | 16$\;\\!\times$ | 22$\;\\!\times$ | 25$\;\\!\times$ | - | -
h2 | 9.0 s | 6.7$\;\\!\times$ | 9.4$\;\\!\times$ | 125$\;\\!\times$ | 135$\;\\!\times$ | 213$\;\\!\times$ | 208$\;\\!\times$ | 241$\;\\!\times$ | 386 s | -
luindex | 1.6 s | 5.0$\;\\!\times$ | 9.5$\;\\!\times$ | 66$\;\\!\times$ | 69$\;\\!\times$ | 82$\;\\!\times$ | 100$\;\\!\times$ | 120$\;\\!\times$ | - | -
lusearch | 4.1 s | 3.8$\;\\!\times$ | 4.5$\;\\!\times$ | 17$\;\\!\times$ | 17$\;\\!\times$ | 21$\;\\!\times$ | 23$\;\\!\times$ | 32$\;\\!\times$ | ¡ 0.1 s | -
pmd | 3.0 s | 6.4$\;\\!\times$ | 8.7$\;\\!\times$ | 21$\;\\!\times$ | 22$\;\\!\times$ | 23$\;\\!\times$ | 27$\;\\!\times$ | 29$\;\\!\times$ | - | 0.9 s
sunflow | 2.8 s | 8.6$\;\\!\times$ | 12$\;\\!\times$ | 104$\;\\!\times$ | 106$\;\\!\times$ | 116$\;\\!\times$ | 124$\;\\!\times$ | 189$\;\\!\times$ | - | -
tomcat | 1.9 s | 5.1$\;\\!\times$ | 5.4$\;\\!\times$ | 23$\;\\!\times$ | 22$\;\\!\times$ | 37$\;\\!\times$ | 40$\;\\!\times$ | 43$\;\\!\times$ | 1.8 s | 49 s
xalan | 5.5 s | 2.4$\;\\!\times$ | 3.0$\;\\!\times$ | 34$\;\\!\times$ | 35$\;\\!\times$ | 51$\;\\!\times$ | 66$\;\\!\times$ | 113$\;\\!\times$ | 29 s | 0.2 s
Table 9. Slowdowns of program instrumentation and various analyses over
uninstrumented execution, and the average time taken to vindicate WDP-only
races.
_SDP+DC_ represents performing SDP and DC analyses together. We run DC
analysis with SDP analysis to minimize DC-races that need vindication.
(Vindicator combined DC analysis with WCP analysis for this purpose (Roemer et
al., 2018), but SDP analysis is more powerful.)
_SDP+WDP+graph_ represents the canonical use case for WDP analysis. This
configuration performs SDP and WDP analyses and constructs the constraint
graph to enable vindication. It uses SDP to reduce how many WDP-races need
vindication. For comparison purposes, _SDP+WDP_ forgoes constructing the
constraint graph, showing the cost of constructing the graph, which we have
not optimized. _SDP+WDP_ is slower than _SDP+DC_ because WDP analysis is
generally more complex than DC analysis.
Finally, _Failed_ and _Verified_ are the average times taken for each dynamic
race that VindicateWDPRace fails to verify or successfully verifies,
respectively. Vindication times vary significantly across programs;
vindication is particularly slow for tomcat because most of its racing
accesses are separated by millions of events (Table 8). Vindication is slow
for h2, even though its races are not far apart, because VindicateWDPRace
discovers new critical section constraints that require analyzing over 500
million events.
### 7.6. Summary and Discussion
Our SDP- and WDP-based approaches are slower than other predictive approaches,
but they find more races, some of which are millions of events apart. SMT-
based approaches would not be able to find these far-apart races because they
cannot scale past analyzing bounded windows of executions (Huang et al., 2014;
Said et al., 2011; Huang and Rajagopalan, 2016; Chen et al., 2008; Şerbănuţă
et al., 2013; Liu et al., 2016) (Section 8). Notably, _RVPredict_ , which
(like WDP) incorporates precise control and data dependence, uses an analysis
window of 10,000 events (Huang et al., 2014), meaning it would miss many of
the predictable races detected and verified by our approach.
Our evaluation demonstrates the power of SDP and WDP analyses to find more
races than prior approaches _in a single execution_. A potential limitation of
the evaluation is that it does not compare our analyses with approaches that
perform HB analysis on multiple executions (e.g., using one of the many
schedule exploration approaches; Section 8). Our work’s aim is to push the
limit on what can be found in a single execution, which is essentially
complementary to approaches that explore multiple schedules. No other known
sound technique could have predicted all of these races from the observed
executions.
## 8\. Related Work
This section describes and compares with prior work, starting with the most
closely related work.
##### Unbounded predictive analysis.
Prior work introduces unbounded predictive analyses, weak-causally-precedes
(WCP) and doesn’t-commute (DC) analyses (Kini et al., 2017; Roemer et al.,
2018), which Sections 2 and 7 covered and evaluated in detail. SDP and WDP
analyses predict more races in real programs than WCP and DC analyses,
respectively (Section 7).
The WCP relation is weaker (i.e., detects more races) than Smaragdakis et
al.’s earlier _causally-precedes (CP)_ relation (Smaragdakis et al., 2012).
Smaragdakis et al.’s implementation detects races within bounded windows of
500 events because of the difficulty of developing an efficient unbounded
analysis for CP (Smaragdakis et al., 2012; Roemer and Bond, 2019).
Recent work introduces the _afterward-confirm (AC)_ relation and an approach
called _DigHR_ for computing AC (Luo et al., 2018). AC is the same as _CP_
except that it removes write–write conflicting critical section ordering.
Despite this similarity with our work, the contributions differ significantly,
and the DigHR work has major correctness issues. Foremost, the DigHR paper
claims incorrectly that AC is sound. AC is, to the best of our understanding,
unsound: its removal of write–write ordering leads to detecting false races,
including for the execution in Figure 2(c) (with the br event omitted; DigHR’s
event model does not include branches). The DigHR paper provides a soundness
proof, which we believe is incorrect as a result of informality leading to not
covering cases such as Figure 2(c). In contrast with DigHR, our work
introduces a sound relaxation of WCP (SDP). Additionally, our work introduces
a complete relation (WDP), handles control dependencies (br events), and
presents linear-time analyses for SDP and WDP (DigHR is superlinear, like
existing CP analyses (Smaragdakis et al., 2012; Roemer and Bond, 2019)).
Concurrently with our work, Pavlogiannis introduces a predictive race
detection approach called _M2_ that is related to vindication (Pavlogiannis,
2019). Pavlogiannis uses lockset analysis as an imprecise filter for potential
races checked by M2, while our work introduces linear-time WDP analysis as a
less-imprecise filter for potential races checked by VindicateWDPRace.
Although Pavlogiannis reports performance that is sometimes competitive with
the performance of HB, WCP, and DC analyses, Pavlogiannis’s implementations of
these analyses perform extra passes over execution traces in addition to the
efficient single-pass vector-clock-based analyses from prior work (Kini et
al., 2017; Roemer et al., 2018). It is unclear to us how M2 and
VindicateWDPRace would compare in terms of detection capability (aside from
the fact that only VindicateWDPRace takes branches and control dependence into
account). In addition to these differences, our work incorporates branches and
control dependence sensitivity, while Pavlogiannis’s work does not and thus
would miss races such as Figures 1(b) and 5; and our work introduces a sound
partial order and linear-time analysis (SDP and SDP analysis).
Our concurrent work introduces the _SmartTrack_ algorithm, which optimizes the
performance of WCP and DC analyses (Roemer et al., 2019). SmartTrack’s
optimizations apply to analyses that compute predictive relations that order
all pairs of conflicting accesses—a property that SDP and WDP do not conform
to. In any case, optimizing WDP analysis would have limited benefit because
the analysis still must construct a constraint graph in order to perform
vindication (a necessity considering that so many WDP-races fail vindication
in practice). The SmartTrack paper also introduces a new relation _weak-
doesn’t-commute (WDC)_ that is a weak variant of DC (Roemer et al., 2019).
Unlike SDP and WDP, WDC does _not_ help find more races than DC, but rather
serves to improve analysis performance.
##### Bounded predictive approaches.
Other approaches predict data races by generating and solving satisfiability
modulo theories (SMT) constraints (Huang et al., 2014; Said et al., 2011;
Huang and Rajagopalan, 2016; Chen et al., 2008; Şerbănuţă et al., 2013; Liu et
al., 2016). These SMT-based approaches cannot analyze long execution traces in
reasonable time, because constraints are quadratic or cubic in trace length,
and constraint-solving time often grows superlinearly with constraints. These
approaches thus break traces into bounded “windows” of traces (e.g.,
500–10,000 events each), missing predictable races involving accesses that
occur further apart in the trace.
One advantage of SMT-based approaches is that they can be both sound and
complete (within a bounded window) by encoding precise constraints. Notably,
_RVPredict_ includes branches in its execution model (Huang et al., 2014).
(RVPredict also incorporates _values_ into its event model, so a read in a
predictable trace can have a different last writer as long as it writes the
same value (Huang et al., 2014).) RVPredict is thus complete by Section 2’s
definition, except that it is practically limited to bounded windows of
execution. WDP analysis, on the other hand, is complete without the windowing
limitation, but VindicateWDPRace is not guaranteed to vindicate a WDP-race
even when a predictable race exists.
##### Schedule exploration.
In contrast to predictive analysis, _schedule exploration_ approaches execute
the program multiple times to explore more program behaviors (Huang, 2015;
Huang and Huang, 2017; Musuvathi and Qadeer, 2007; Burckhardt et al., 2010;
Eslamimehr and Palsberg, 2014; Sen, 2008; Cai and Cao, 2015; Henzinger et al.,
2004). These approaches may be systematic (often called _model checking_) or
be based on randomness or heuristics. Schedule exploration is complementary to
predictive analysis, which aims to glean as much as possible from a given
execution. _Maximal causality reduction (MCR)_ combines schedule exploration
with predictive analysis (Huang, 2015; Huang and Huang, 2017). _MCR-S_
incorporates static control flow information to reduce the complexity of MCR’s
generated SMT constraints (Huang and Huang, 2017).
##### Other analyses.
Widely used data race detectors typically use dynamic _happens-before (HB)_
analysis (Lamport, 1978; Flanagan and Freund, 2009; Pozniansky and Schuster,
2007; Elmas et al., 2007; Serebryany and Iskhodzhanov, 2009; Serebryany et
al., 2012; Intel Corporation, 2016). HB analysis cannot predict races
involving reordered critical sections on the same lock. The detected races
thus depend heavily on the scheduling of the analyzed program. Other analyses
find a subset of HB-races by detecting simultaneously executing conflicting
accesses or regions (Veeraraghavan et al., 2011; Biswas et al., 2015;
Effinger-Dean et al., 2012; Erickson et al., 2010; Sen, 2008; Biswas et al.,
2017).
_Lockset_ analysis checks a locking discipline, ensuring that all pairs of
conflicting accesses hold some common lock (Savage et al., 1997; Nishiyama,
2004; Choi et al., 2002; von Praun and Gross, 2001). Lockset analysis is
predictive but unsound, reporting false races for synchronization patterns
other than the locking discipline. _Hybrid_ lockset–HB analyses generally
incur disadvantages of one or both kinds of analysis (Dinning and Schonberg,
1991; O’Callahan and Choi, 2003; Yu et al., 2005; Pozniansky and Schuster,
2007).
_Sampling-based_ analysis trades coverage for performance (opposite of
predictive analysis) in order to detect data races in production (Bond et al.,
2010; Erickson et al., 2010; Marino et al., 2009; Kasikci et al., 2013; Biswas
et al., 2017; Sheng et al., 2011; Zhang et al., 2017). Custom hardware support
can detect data races with low performance cost but has not been implemented
(Devietti et al., 2012; Zhou et al., 2007; Wood et al., 2014; Segulja and
Abdelrahman, 2015; Peng et al., 2017; Lucia et al., 2010; Marino et al., 2010;
Singh et al., 2011).
Dynamic analysis can estimate the likely harm of a data race (Boehm, 2011;
Kasikci et al., 2015; Narayanasamy et al., 2007; Cao et al., 2016; Flanagan
and Freund, 2010a; Burnim et al., 2011), which is orthogonal to detection. All
data races are erroneous under language memory models that ascribe them
undefined semantics (Adve and Boehm, 2010; Boehm and Adve, 2008; Manson et
al., 2005; Boehm and Demsky, 2014; Boehm, 2012; Boehm and Adve, 2012; Ševčík
and Aspinall, 2008). Java’s memory model defines weak semantics for data races
(Manson et al., 2005), but inadvertently prohibits common JVM optimizations
(Ševčík and Aspinall, 2008; Boehm and Demsky, 2014).
_Static_ program analysis can detect all races across all feasible executions
of a program (von Praun and Gross, 2003; Naik et al., 2006; Naik and Aiken,
2007; Pratikakis et al., 2006; Engler and Ashcraft, 2003; Voung et al., 2007),
but it reports thousands of false races for real programs (Biswas et al.,
2017; Lee et al., 2012).
##### Avoiding or tolerating data races.
New languages and type systems can ensure data race freedom, but require
significant programmer effort (Bocchino et al., 2009; Boyapati et al., 2002;
Rinard and Lam, 1998; Matsakis and Klock, 2014; Abadi et al., 2006; Flanagan
and Freund, 2007). Compilers and hardware can provide well-defined semantics
for data races, but incur high run-time costs or hardware complexity (Segulja
and Abdelrahman, 2015; Lucia et al., 2010; Marino et al., 2010; Singh et al.,
2011; Marino et al., 2011; Singh et al., 2012; Sura et al., 2005; Sengupta et
al., 2015; Ahn et al., 2009; Ouyang et al., 2013).
## 9\. Conclusion
SDP and WDP analyses improve over existing predictive analyses by
incorporating precise notions of data and control dependence, finding more
races both in theory and in practice while retaining linear (in trace length)
run time and thus unbounded analysis. SDP analysis maintains WCP analysis’s
soundness while increasing race coverage. WDP analysis finds all data races
that can be predicted from an observed execution; not all WDP-races are
predictable races, but VindicateWDPRace can efficiently filter false races.
Experiments show that our new approaches find many predictable races in real
programs that prior approaches are unable to find. These properties and
results suggest that our contributions advance the state of the art in
predictive race detection analysis.
## Acknowledgments
We thank Rob LaTour for early help with this project. Thanks to Steve Freund
for making RoadRunner publicly available and providing help with using and
modifying it. Thanks to the anonymous reviewers for their thorough and
insightful feedback.
## References
* (1)
* Abadi et al. (2006) Martín Abadi, Cormac Flanagan, and Stephen N. Freund. 2006\. Types for Safe Locking: Static Race Detection for Java. _TOPLAS_ 28, 2 (2006), 207–255.
* Adve and Boehm (2010) Sarita V. Adve and Hans-J. Boehm. 2010. Memory Models: A Case for Rethinking Parallel Languages and Hardware. _CACM_ 53 (2010), 90–101. Issue 8.
* Ahn et al. (2009) W. Ahn, S. Qi, M. Nicolaides, J. Torrellas, J.-W. Lee, X. Fang, S. Midkiff, and D. Wong. 2009. BulkCompiler: High-Performance Sequential Consistency through Cooperative Compiler and Hardware Support. In _MICRO_. 133–144.
* Bessey et al. (2010) Al Bessey, Ken Block, Ben Chelf, Andy Chou, Bryan Fulton, Seth Hallem, Charles Henri-Gros, Asya Kamsky, Scott McPeak, and Dawson Engler. 2010\. A Few Billion Lines of Code Later: Using Static Analysis to Find Bugs in the Real World. _CACM_ 53, 2 (2010), 66–75.
* Biswas et al. (2017) Swarnendu Biswas, Man Cao, Minjia Zhang, Michael D. Bond, and Benjamin P. Wood. 2017. Lightweight Data Race Detection for Production Runs. In _CC_. 11–21.
* Biswas et al. (2015) Swarnendu Biswas, Minjia Zhang, Michael D. Bond, and Brandon Lucia. 2015. Valor: Efficient, Software-Only Region Conflict Exceptions. In _OOPSLA_. 241–259.
* Blackburn et al. (2006) S. M. Blackburn, R. Garner, C. Hoffman, A. M. Khan, K. S. McKinley, R. Bentzur, A. Diwan, D. Feinberg, D. Frampton, S. Z. Guyer, M. Hirzel, A. Hosking, M. Jump, H. Lee, J. E. B. Moss, A. Phansalkar, D. Stefanović, T. VanDrunen, D. von Dincklage, and B. Wiedermann. 2006\. The DaCapo Benchmarks: Java Benchmarking Development and Analysis. In _OOPSLA_. 169–190.
* Bocchino et al. (2009) Robert L. Bocchino, Jr., Vikram S. Adve, Sarita V. Adve, and Marc Snir. 2009. Parallel Programming Must Be Deterministic by Default. In _HotPar_. 4–9.
* Boehm (2011) Hans-J. Boehm. 2011\. How to miscompile programs with “benign” data races. In _HotPar_. 6.
* Boehm (2012) Hans-J. Boehm. 2012\. Position paper: Nondeterminism is Unavoidable, but Data Races are Pure Evil. In _RACES_. 9–14.
* Boehm and Adve (2008) Hans-J. Boehm and Sarita V. Adve. 2008. Foundations of the C++ Concurrency Memory Model. In _PLDI_. 68–78.
* Boehm and Adve (2012) Hans-J. Boehm and Sarita V. Adve. 2012. You Don’t Know Jack about Shared Variables or Memory Models. _CACM_ 55, 2 (2012), 48–54.
* Boehm and Demsky (2014) Hans-J. Boehm and Brian Demsky. 2014. Outlawing Ghosts: Avoiding Out-of-Thin-Air Results. In _MSPC_. Article 7, 6 pages.
* Bond et al. (2010) Michael D. Bond, Katherine E. Coons, and Kathryn S. McKinley. 2010. Pacer: Proportional Detection of Data Races. In _PLDI_. 255–268.
* Boyapati et al. (2002) Chandrasekhar Boyapati, Robert Lee, and Martin Rinard. 2002\. Ownership Types for Safe Programming: Preventing Data Races and Deadlocks. In _OOPSLA_. 211–230.
* Burckhardt et al. (2010) Sebastian Burckhardt, Pravesh Kothari, Madanlal Musuvathi, and Santosh Nagarakatte. 2010\. A Randomized Scheduler with Probabilistic Guarantees of Finding Bugs. In _ASPLOS_. 167–178.
* Burnim et al. (2011) Jacob Burnim, Koushik Sen, and Christos Stergiou. 2011\. Testing Concurrent Programs on Relaxed Memory Models. In _ISSTA_. 122–132.
* Cai and Cao (2015) Yan Cai and Lingwei Cao. 2015. Effective and Precise Dynamic Detection of Hidden Races for Java Programs. In _ESEC/FSE_. 450–461.
* Cao et al. (2016) Man Cao, Jake Roemer, Aritra Sengupta, and Michael D. Bond. 2016\. Prescient Memory: Exposing Weak Memory Model Behavior by Looking into the Future. In _ISMM_. 99–110.
* Chen et al. (2008) Feng Chen, Traian Florin Şerbănuţă, and Grigore Roşu. 2008\. jPredictor: A Predictive Runtime Analysis Tool for Java. In _ICSE_. 221–230.
* Choi et al. (2002) Jong-Deok Choi, Keunwoo Lee, Alexey Loginov, Robert O’Callahan, Vivek Sarkar, and Manu Sridharan. 2002\. Efficient and Precise Datarace Detection for Multithreaded Object-Oriented Programs. In _PLDI_. 258–269.
* Devietti et al. (2012) Joseph Devietti, Benjamin P. Wood, Karin Strauss, Luis Ceze, Dan Grossman, and Shaz Qadeer. 2012. RADISH: Always-On Sound and Complete Race Detection in Software and Hardware. In _ISCA_. 201–212.
* Dinning and Schonberg (1991) Anne Dinning and Edith Schonberg. 1991. Detecting Access Anomalies in Programs with Critical Sections. In _PADD_. 85–96.
* Dolan et al. (2018) Stephen Dolan, KC Sivaramakrishnan, and Anil Madhavapeddy. 2018\. Bounding Data Races in Space and Time. In _PLDI_. 242–255.
* Effinger-Dean et al. (2012) Laura Effinger-Dean, Brandon Lucia, Luis Ceze, Dan Grossman, and Hans-J. Boehm. 2012\. IFRit: Interference-Free Regions for Dynamic Data-Race Detection. In _OOPSLA_. 467–484.
* Elmas et al. (2007) Tayfun Elmas, Shaz Qadeer, and Serdar Tasiran. 2007\. Goldilocks: A Race and Transaction-Aware Java Runtime. In _PLDI_. 245–255.
* Engler and Ashcraft (2003) Dawson Engler and Ken Ashcraft. 2003. RacerX: Effective, Static Detection of Race Conditions and Deadlocks. In _SOSP_. 237–252.
* Erickson et al. (2010) John Erickson, Madanlal Musuvathi, Sebastian Burckhardt, and Kirk Olynyk. 2010. Effective Data-Race Detection for the Kernel. In _OSDI_. 1–16.
* Eslamimehr and Palsberg (2014) Mahdi Eslamimehr and Jens Palsberg. 2014. Race Directed Scheduling of Concurrent Programs. In _PPoPP_. 301–314.
* Ferrante et al. (1987) Jeanne Ferrante, Karl J. Ottenstein, and Joe D. Warren. 1987\. The Program Dependence Graph and Its Use in Optimization. _TOPLAS_ 9, 3 (1987), 319–349.
* Flanagan and Freund (2007) Cormac Flanagan and Stephen N. Freund. 2007. Type Inference Against Races. _SCP_ 64, 1 (2007), 140–165.
* Flanagan and Freund (2009) Cormac Flanagan and Stephen N. Freund. 2009. FastTrack: Efficient and Precise Dynamic Race Detection. In _PLDI_. 121–133.
* Flanagan and Freund (2010a) Cormac Flanagan and Stephen N. Freund. 2010a. Adversarial Memory for Detecting Destructive Races. In _PLDI_. 244–254.
* Flanagan and Freund (2010b) Cormac Flanagan and Stephen N. Freund. 2010b. The RoadRunner Dynamic Analysis Framework for Concurrent Programs. In _PASTE_. 1–8.
* Flanagan and Freund (2017) Cormac Flanagan and Stephen N. Freund. 2017. _The FastTrack2 Race Detector_. Technical Report. Williams College.
* Giffhorn and Hammer (2008) Dennis Giffhorn and Christian Hammer. 2008. Precise Analysis of Java Programs Using JOANA. In _SCAM_. 267–268.
* Godefroid and Nagappan (2008) P. Godefroid and N. Nagappan. 2008. Concurrency at Microsoft – An Exploratory Survey. In _EC 2_.
* Henzinger et al. (2004) Thomas A. Henzinger, Ranjit Jhala, and Rupak Majumdar. 2004\. Race Checking by Context Inference. In _PLDI_. 1–13.
* Huang (2015) Jeff Huang. 2015\. Stateless Model Checking Concurrent Programs with Maximal Causality Reduction. In _PLDI_. 165–174.
* Huang et al. (2014) Jeff Huang, Patrick O’Neil Meredith, and Grigore Roşu. 2014\. Maximal Sound Predictive Race Detection with Control Flow Abstraction. In _PLDI_. 337–348.
* Huang and Rajagopalan (2016) Jeff Huang and Arun K. Rajagopalan. 2016. Precise and Maximal Race Detection from Incomplete Traces. In _OOPSLA_. 462–476.
* Huang and Huang (2017) Shiyou Huang and Jeff Huang. 2017. Speeding Up Maximal Causality Reduction with Static Dependency Analysis. In _ECOOP_. 16:1–16:22.
* Intel Corporation (2016) Intel Corporation. 2016\. Intel Inspector. https://software.intel.com/en-us/intel-inspector-xe.
* Kasikci et al. (2012) Baris Kasikci, Cristian Zamfir, and George Candea. 2012\. Data Races vs. Data Race Bugs: Telling the Difference with Portend. In _ASPLOS_. 185–198.
* Kasikci et al. (2013) Baris Kasikci, Cristian Zamfir, and George Candea. 2013\. RaceMob: Crowdsourced Data Race Detection. In _SOSP_. 406–422.
* Kasikci et al. (2015) Baris Kasikci, Cristian Zamfir, and George Candea. 2015\. Automated Classification of Data Races Under Both Strong and Weak Memory Models. _TOPLAS_ 37, 3, Article 8 (May 2015), 44 pages.
* Kini et al. (2017) Dileep Kini, Umang Mathur, and Mahesh Viswanathan. 2017\. Dynamic Race Prediction in Linear Time. In _PLDI_. 157–170.
* Lamport (1978) Leslie Lamport. 1978\. Time, Clocks, and the Ordering of Events in a Distributed System. _CACM_ 21, 7 (1978), 558–565.
* Lee et al. (2012) Dongyoon Lee, Peter M. Chen, Jason Flinn, and Satish Narayanasamy. 2012. Chimera: Hybrid Program Analysis for Determinism. In _PLDI_. 463–474.
* Leveson and Turner (1993) N. G. Leveson and C. S. Turner. 1993. An Investigation of the Therac-25 Accidents. _IEEE Computer_ 26, 7 (1993), 18–41.
* Lindholm and Yellin (1999) Tim Lindholm and Frank Yellin. 1999. _The Java Virtual Machine Specification_ (2nd ed.). Prentice Hall PTR.
* Liu et al. (2016) Peng Liu, Omer Tripp, and Xiangyu Zhang. 2016. IPA: Improving Predictive Analysis with Pointer Analysis. In _ISSTA_. 59–69.
* Lu et al. (2008) Shan Lu, Soyeon Park, Eunsoo Seo, and Yuanyuan Zhou. 2008\. Learning from Mistakes: A Comprehensive Study on Real World Concurrency Bug Characteristics. In _ASPLOS_. 329–339.
* Lucia et al. (2010) Brandon Lucia, Luis Ceze, Karin Strauss, Shaz Qadeer, and Hans-J. Boehm. 2010. Conflict Exceptions: Simplifying Concurrent Language Semantics with Precise Hardware Exceptions for Data-Races. In _ISCA_. 210–221.
* Luo et al. (2018) Peng Luo, Deqing Zou, Hai Jin, Yajuan Du, Long Zheng, and Jinan Shen. 2018. DigHR: precise dynamic detection of hidden races with weak causal relation analysis. _J. Supercomputing_ (2018).
* Manson et al. (2005) Jeremy Manson, William Pugh, and Sarita V. Adve. 2005\. The Java Memory Model. In _POPL_. 378–391.
* Marino et al. (2009) Daniel Marino, Madanlal Musuvathi, and Satish Narayanasamy. 2009\. LiteRace: Effective Sampling for Lightweight Data-Race Detection. In _PLDI_. 134–143.
* Marino et al. (2010) Daniel Marino, Abhayendra Singh, Todd Millstein, Madanlal Musuvathi, and Satish Narayanasamy. 2010\. DRFx: A Simple and Efficient Memory Model for Concurrent Programming Languages. In _PLDI_. 351–362.
* Marino et al. (2011) Daniel Marino, Abhayendra Singh, Todd Millstein, Madanlal Musuvathi, and Satish Narayanasamy. 2011\. A Case for an SC-Preserving Compiler. In _PLDI_. 199–210.
* Matsakis and Klock (2014) Nicholas D. Matsakis and Felix S. Klock, II. 2014. The Rust Language. In _HILT_. 103–104.
* Mattern (1988) Friedemann Mattern. 1988\. Virtual Time and Global States of Distributed Systems. In _Workshop on Parallel and Distributed Algorithms_. 215–226.
* Musuvathi and Qadeer (2007) Madanlal Musuvathi and Shaz Qadeer. 2007. Iterative Context Bounding for Systematic Testing of Multithreaded Programs. In _PLDI_. 446–455.
* Naik and Aiken (2007) Mayur Naik and Alex Aiken. 2007. Conditional Must Not Aliasing for Static Race Detection. In _POPL_. 327–338.
* Naik et al. (2006) Mayur Naik, Alex Aiken, and John Whaley. 2006. Effective Static Race Detection for Java. In _PLDI_. 308–319.
* Narayanasamy et al. (2007) Satish Narayanasamy, Zhenghao Wang, Jordan Tigani, Andrew Edwards, and Brad Calder. 2007\. Automatically Classifying Benign and Harmful Data Races Using Replay Analysis. In _PLDI_. 22–31.
* Nishiyama (2004) Hiroyasu Nishiyama. 2004\. Detecting Data Races using Dynamic Escape Analysis based on Read Barrier. In _VMRT_. 127–138.
* O’Callahan and Choi (2003) Robert O’Callahan and Jong-Deok Choi. 2003. Hybrid Dynamic Data Race Detection. In _PPoPP_. 167–178.
* Ouyang et al. (2013) Jessica Ouyang, Peter M. Chen, Jason Flinn, and Satish Narayanasamy. 2013. …and region serializability for all. In _HotPar_.
* Pavlogiannis (2019) Andreas Pavlogiannis. 2019\. Fast, Sound and Effectively Complete Dynamic Race Detection. arXiv:1901.08857 http://arxiv.org/abs/1901.08857
* PCWorld (2012) PCWorld. 2012. Nasdaq’s Facebook Glitch Came From Race Conditions. http://www.pcworld.com/article/255911/nasdaqs_facebook_glitch_came_from_race_conditions.html.
* Peng et al. (2017) Yuanfeng Peng, Benjamin P. Wood, and Joseph Devietti. 2017\. PARSNIP: Performant Architecture for Race Safety with No Impact on Precision. In _MICRO_. 490–502.
* Pozniansky and Schuster (2007) Eli Pozniansky and Assaf Schuster. 2007. MultiRace: Efficient On-the-Fly Data Race Detection in Multithreaded C++ Programs. _CCPE_ 19, 3 (2007), 327–340.
* Pratikakis et al. (2006) Polyvios Pratikakis, Jeffrey S. Foster, and Michael Hicks. 2006\. LOCKSMITH: Context-Sensitive Correlation Analysis for Race Detection. In _PLDI_. 320–331.
* Rinard and Lam (1998) Martin C. Rinard and Monica S. Lam. 1998. The Design, Implementation, and Evaluation of Jade. _TOPLAS_ 20 (1998), 483–545. Issue 3.
* Roemer and Bond (2019) Jake Roemer and Michael D. Bond. 2019. Online Set-Based Dynamic Analysis for Sound Predictive Race Detection. _CoRR_ abs/1907.08337 (2019). arXiv:1907.08337 http://arxiv.org/abs/1907.08337
* Roemer et al. (2019) Jake Roemer, Kaan Genç, and Michael D. Bond. 2019\. Practical Predictive Race Detection. _CoRR_ abs/1905.00494 (2019). arXiv:1905.00494 http://arxiv.org/abs/1905.00494
* Roemer et al. (2018) Jake Roemer, Kaan Genç, and Michael D. Bond. 2018\. High-Coverage, Unbounded Sound Predictive Race Detection. In _PLDI_. 374–389.
* Said et al. (2011) Mahmoud Said, Chao Wang, Zijiang Yang, and Karem Sakallah. 2011\. Generating Data Race Witnesses by an SMT-based Analysis. In _NFM_. 313–327.
* Savage et al. (1997) Stefan Savage, Michael Burrows, Greg Nelson, Patrick Sobalvarro, and Thomas Anderson. 1997\. Eraser: A Dynamic Data Race Detector for Multi-Threaded Programs. In _SOSP_. 27–37.
* Segulja and Abdelrahman (2015) Cedomir Segulja and Tarek S. Abdelrahman. 2015. Clean: A Race Detector with Cleaner Semantics. In _ISCA_. 401–413.
* Sen (2008) Koushik Sen. 2008\. Race Directed Random Testing of Concurrent Programs. In _PLDI_. 11–21.
* Sengupta et al. (2015) Aritra Sengupta, Swarnendu Biswas, Minjia Zhang, Michael D. Bond, and Milind Kulkarni. 2015. Hybrid Static–Dynamic Analysis for Statically Bounded Region Serializability. In _ASPLOS_. 561–575.
* Şerbănuţă et al. (2013) Traian Florin Şerbănuţă, Feng Chen, and Grigore Roşu. 2013\. Maximal Causal Models for Sequentially Consistent Systems. In _RV_. 136–150.
* Serebryany and Iskhodzhanov (2009) Konstantin Serebryany and Timur Iskhodzhanov. 2009. ThreadSanitizer – data race detection in practice. In _WBIA_. 62–71.
* Serebryany et al. (2012) Konstantin Serebryany, Alexander Potapenko, Timur Iskhodzhanov, and Dmitriy Vyukov. 2012. Dynamic Race Detection with LLVM Compiler. In _RV_. 110–114.
* Ševčík and Aspinall (2008) Jaroslav Ševčík and David Aspinall. 2008. On Validity of Program Transformations in the Java Memory Model. In _ECOOP_. 27–51.
* Sheng et al. (2011) Tianwei Sheng, Neil Vachharajani, Stephane Eranian, Robert Hundt, Wenguang Chen, and Weimin Zheng. 2011. RACEZ: A Lightweight and Non-Invasive Race Detection Tool for Production Applications. In _ICSE_. 401–410.
* Singh et al. (2011) Abhayendra Singh, Daniel Marino, Satish Narayanasamy, Todd Millstein, and Madan Musuvathi. 2011\. Efficient Processor Support for DRFx, a Memory Model with Exceptions. In _ASPLOS_. 53–66.
* Singh et al. (2012) Abhayendra Singh, Satish Narayanasamy, Daniel Marino, Todd Millstein, and Madanlal Musuvathi. 2012\. End-to-End Sequential Consistency. In _ISCA_. 524–535.
* Smaragdakis et al. (2012) Yannis Smaragdakis, Jacob Evans, Caitlin Sadowski, Jaeheon Yi, and Cormac Flanagan. 2012. Sound Predictive Race Detection in Polynomial Time. In _POPL_. 387–400.
* Sura et al. (2005) Zehra Sura, Xing Fang, Chi-Leung Wong, Samuel P. Midkiff, Jaejin Lee, and David Padua. 2005\. Compiler Techniques for High Performance Sequentially Consistent Java Programs. In _PPoPP_. 2–13.
* U.S.–Canada Power System Outage Task Force (2004) U.S.–Canada Power System Outage Task Force. 2004. _Final Report on the August 14th Blackout in the United States and Canada_. Technical Report. Department of Energy.
* Veeraraghavan et al. (2011) Kaushik Veeraraghavan, Peter M. Chen, Jason Flinn, and Satish Narayanasamy. 2011. Detecting and Surviving Data Races using Complementary Schedules. In _SOSP_. 369–384.
* von Praun and Gross (2001) Christoph von Praun and Thomas R. Gross. 2001. Object Race Detection. In _OOPSLA_. 70–82.
* von Praun and Gross (2003) Christoph von Praun and Thomas R. Gross. 2003. Static Conflict Analysis for Multi-Threaded Object-Oriented Programs. In _PLDI_. 115–128.
* Voung et al. (2007) Jan Wen Voung, Ranjit Jhala, and Sorin Lerner. 2007\. RELAY: Static Race Detection on Millions of Lines of Code. In _ESEC/FSE_. 205–214.
* Wood et al. (2014) Benjamin P. Wood, Luis Ceze, and Dan Grossman. 2014\. Low-Level Detection of Language-Level Data Races with LARD. In _ASPLOS_. 671–686.
* Yu et al. (2005) Yuan Yu, Tom Rodeheffer, and Wei Chen. 2005. RaceTrack: Efficient Detection of Data Race Conditions via Adaptive Tracking. In _SOSP_. 221–234.
* Zhang et al. (2017) Tong Zhang, Changhee Jung, and Dongyoon Lee. 2017\. ProRace: Practical Data Race Detection for Production Use. In _ASPLOS_. 149–162.
* Zhivich and Cunningham (2009) M. Zhivich and R. K. Cunningham. 2009. The Real Cost of Software Errors. _IEEE Security & Privacy_ 7 (03 2009), 87–90.
* Zhou et al. (2007) Pin Zhou, Radu Teodorescu, and Yuanyuan Zhou. 2007\. HARD: Hardware-Assisted Lockset-based Race Detection. In _HPCA_. 121–132.
## Appendix A Proof of WDP Completeness Helper Lemma
###### Proof of Lemma 4.2.
We proceed by induction on the _WDP-distance_ of two events,
$\mathit{d(e,e^{\prime})}$, defined as follows (the WDP properties refer to
Table 2):
$\displaystyle\mathit{d(e,e^{\prime})}=\min\begin{cases}0&\textnormal{ if
}e\prec_{\textsc{\tiny{WDP}}}e^{\prime}\textnormal{ by WDP conflicting
critical section ordering}\\\
1+\mathit{d(\mathit{A}(e),e^{\prime})}&\textnormal{ if
}e\prec_{\textsc{\tiny{WDP}}}e^{\prime}\textnormal{ by
``$\textsf{acq(m)}\prec_{\textsc{\tiny{WDP}}}\textsf{rel(m)}\implies\textsf{rel(m)}\prec_{\textsc{\tiny{WDP}}}\textsf{rel(m)}$''}\\\
0&\textnormal{ if }e\prec_{\textsc{\tiny{WDP}}}e^{\prime}\textnormal{ by
PO}\\\ \lx@intercol
1+\min_{e^{\prime\prime}}(\mathit{d(e,e^{\prime\prime})}+\mathit{d(e^{\prime\prime},e^{\prime})})\hfil\lx@intercol\\\
&\textnormal{ if }\exists
e^{\prime\prime}:e\prec_{\textsc{\tiny{WDP}}}e^{\prime\prime}\land
e^{\prime\prime}\prec_{\textsc{\tiny{WDP}}}e^{\prime}\textnormal{ by WDP
transitivity}\\\ \infty&\textnormal{ otherwise}\end{cases}$
##### Base case
Let $e_{1}$ and $e_{2}$ be events in a trace $\mathit{tr}$ such that
$\mathit{d(e_{1},e_{2})}=0$ and $e_{1}\prec_{\textsc{\tiny{WDP}}}e_{2}$. Since
$\mathit{d(e_{1},e_{2})}=0$, $e_{1}$ and $e_{2}$ are ordered directly, using
PO or WDP’s conflicting critical section ordering:
If $e_{1}\prec_{\textsc{\tiny{WDP}}}e_{2}$ by PO, then
$e_{1}\prec_{\textsc{\tiny{PO}}}e_{2}$. Let $\mathit{tr^{\prime}}$ be a
predictable trace of $\mathit{tr}$ in which
$e_{2}<_{\textsc{$\mathit{tr^{\prime}}$}}e_{1}$ or
$e_{2}\in\mathit{tr^{\prime}}\land e_{1}\notin\mathit{tr^{\prime}}$, either of
which violates the PO rule of a predictable traces.
If $e_{1}\prec_{\textsc{\tiny{WDP}}}e_{2}$ by WDP’s conflicting critical
section ordering, then $e_{1}$ is a release event, $e_{2}$ is a branch event,
there exists a release event $r_{2}$ over the same lock as $e_{1}$, and there
exists write event $e$ and read event $e^{\prime}$ such that
$e\in\mathit{CS(}e_{1})$, $e^{\prime}\in\mathit{CS(}r_{2})$, $e\asymp
e^{\prime}$, $\mathit{lastwr}_{\mathit{tr}}(e^{\prime})=e$, and
$\mathit{brDepsOn}(e_{2},e^{\prime})$. Let $\mathit{tr^{\prime}}$ be a
predictable trace of $\mathit{tr}$ where
$e_{2}<_{\textsc{$\mathit{tr^{\prime}}$}}e_{1}$ or
$e_{2}\in\mathit{tr^{\prime}}\land e_{1}\notin\mathit{tr^{\prime}}$. Then in
either case, $r_{2}<_{\textsc{$\mathit{tr^{\prime}}$}}a$ or
$r_{2}\in\mathit{tr^{\prime}}\land a\notin\mathit{tr^{\prime}}$, where $a$ is
the matching acquire of $e_{1}$; otherwise $\mathit{tr^{\prime}}$ would be an
invalid predictable trace due to the LS rule of predictable traces. As a
result, $e^{\prime}<_{\textsc{$\mathit{tr^{\prime}}$}}e$ or
$e^{\prime}\in\mathit{tr^{\prime}}\land e\notin\mathit{tr^{\prime}}$, which
means
$\mathit{lastwr}_{\mathit{tr^{\prime}}}(e^{\prime})\neq\mathit{lastwr}_{\mathit{tr}}(e^{\prime})$.
|
# Imperfect Narrow Escape problem
T. Guérin Laboratoire Ondes et Matière d’Aquitaine, CNRS, UMR 5798,
Université de Bordeaux, F-33400 Talence, France M. Dolgushev O. Bénichou
Sorbonne Université, CNRS, Laboratoire de Physique Théorique de la Matière
Condensée, LPTMC, F-75005 Paris, France R. Voituriez Sorbonne Université,
CNRS, Laboratoire de Physique Théorique de la Matière Condensée, LPTMC,
F-75005 Paris, France Sorbonne Université, CNRS, Laboratoire Jean Perrin,
LJP, F-75005 Paris, France
###### Abstract
We consider the kinetics of the imperfect narrow escape problem, i.e. the time
it takes for a particle diffusing in a confined medium of generic shape to
reach and to be adsorbed by a small, imperfectly reactive patch embedded in
the boundary of the domain, in two or three dimensions. Imperfect reactivity
is modeled by an intrinsic surface reactivity $\kappa$ of the patch, giving
rise to Robin boundary conditions. We present a formalism to calculate the
exact asymptotics of the mean reaction time in the limit of large volume of
the confining domain. We obtain exact explicit results in the two limits of
large and small reactivities of the reactive patch, and a semi-analytical
expression in the general case. Our approach reveals an anomalous scaling of
the mean reaction time as the inverse square root of the reactivity in the
large reactivity limit, valid for an initial position near the extremity of
the reactive patch. We compare our exact results with those obtained within
the “constant flux approximation”; we show that this approximation turns out
to give exactly the next-to-leading order term of the small reactivity limit,
and provides a good approximation of the reaction time far from the reactive
patch for all reactivities, but not in the vicinity of the boundary of the
reactive patch due to the above mentioned anomalous scaling. These results
thus provide a general framework to quantify the mean reaction times for the
imperfect narrow escape problem.
## I Introduction
How much time does it take for a random walker to reach a target point? The
answer to this question has received a lot of attention in the last decade in
the physics literature [1, 2, 3, 4, 5, 6, 7, 8, 9]. First passage problems
appear in various areas of biological and soft matter physics and are in
particular relevant to the problem of reaction kinetics, since two reactants
have to meet before being able to react [10, 11]. When the reaction is
“perfect”, i.e. when it occurs instantaneously upon each encounter, its
kinetics is controlled by the first passage statistics of one reactant
molecule, seen as a random walker, to the second reactant, seen as a “target”.
However, many reactions do not occur at first contact between the random
walker and the targets, leading to imperfect reactivity. Imperfect reactivity
can have diverse origins at the microscopic scale, such as orientational
constraints on the reactive particles [11], the fact that the surface of the
reactive particles is not entirely covered by reactive patches (such as in the
chemoreception problem [12]), the need to overcome an energetic [13] (or
entropic [14]) activation barrier before reaction, the presence of a gate that
can be randomly closed or opened when the reactant meets the target [15, 16],
etc (see Ref. [17] and references therein for a recent review on imperfect
reactivity).
Imperfect reactivity was early investigated for molecules diffusing in
infinite space [18, 19, 11, 20] (with an imposed concentration at infinity).
The search problem for a single random walker moving in a confining volume for
an imperfect target, initially considered in Ref. [21] for centered spheres,
has also attracted recent attention and several asymptotic results for
imperfect search kinetics have been derived [22, 23, 24, 25, 26, 27, 28, 29,
30]. Recently, explicit asymptotics of the reaction time statistics have been
obtained for general Markovian random walks [31]. Besides the case of reactive
targets located in the bulk of a confining domain, the narrow escape problem
(NEP) consists in calculating the escape time of a random walker out of a
confining domain, through a small window at the boundary of the domain (see
Figure 1(a)). While the NEP is now well characterized for perfect reactions,
for spherical domains [8, 32, 33, 34, 35] and large domains of arbitrary
shapes [36], fewer results are available for imperfect reactions (i.e. for a
partially adsorbing patch). The imperfect narrow escape problem has been
investigated for particular geometries in cylindrical [37, 38] or spherical
domains [39][40] in which case the analysis depends on the eigenfunctions of
the particular confining volume that is considered and relies on the so-called
uniform flux approximation introduced in Ref. [20].
Figure 1: (a) Illustration of the imperfect narrow escape problem. A
partially reactive patch (thick red line) is embedded in the boundary of a
confining domain. A random walker, starting from the initial position (red
sphere) diffuses in the domain and is eventually adsorbed on the patch. (b)
Zoom on the portion of space delimited by the dashed blue lines around the
reactive patch.
The aim of the present paper is to apply the formalism introduced in Ref. [31]
to cover the case of the imperfect narrow escape problem in a domain of
generic shape. Our formalism is asymptotically exact in the limit of large
confining volume – it does not involve the constant flux approximation – and
provides explicit results in both regimes of small and large reactivity. Of
particular interest for imperfect reaction problems is the mean reaction time
when the initial position is located on the reactive patch; this time is
exactly zero for perfect reactions and scales as $1/\kappa$ for targets in the
interior of the volume. We identify a region for which the reaction time
behaves anomalously with the reactivity $\kappa$. This region, which does not
exist for targets in the bulk of the confining domain, is located at the
boundary of the imperfectly reactive patch. While one would naively expect
this time to be inversely proportional to $\kappa$, we find instead that when
the initial position is at the boundary of the reactive domain, the mean
reaction time $\langle T\rangle_{e}$ is actually anomalously high
($\propto\kappa^{-1/2}$) and follows the exact asymptotics
$\displaystyle\langle
T\rangle_{e}\underset{\kappa\to\infty}{\sim}\frac{V}{(d-1)(2\pi\ \kappa
Da)^{1/2}a^{d-2}}.$ (1)
Here, $d=2$ or $d=3$ is the spatial dimension, $D$ is the diffusion
coefficient, $a$ is the radius of the reactive patch, and $V$ the volume of
the confining domain. Here, we assume that the confining volume is taken large
enough, and the patch small enough, so that the confining boundaries at the
vicinity of the target can be considered as a flat wall in which the reactive
patch is embedded, the latter being considered as a line segment of length
$2a$ in $d=2$ or as a flat disk of radius $a$ in $d=3$. We show below how to
obtain this anomalous scaling relation by solving a Wiener-Hopf integral
equation. We will also show how this “anomalous” behavior (1) of $\langle
T\rangle$ with $\kappa$ can be related to the divergence of fields in
Laplacian problems near surfaces presenting asperities, as occurs in
electrostatics near conducting edges [41] or in the coffee ring effect [42].
More generally, we show that the mean escape time for an arbitrary initial
position far from the target, and for any finite reactivity $\kappa$ satisfies
the following exact asymptotics:
$\displaystyle\langle T\rangle/V\underset{r\gg
a}{\sim}\begin{cases}\frac{1}{\pi D}\ln(r/a)+C_{\infty}&(d=2)\\\
-\frac{1}{2\pi Dr}+C_{\infty}&(d=3)\end{cases},$ (2)
where $C_{\infty}$ is independent of the initial distance $r$ from the target.
For finite values of the reactivity $\kappa$, we show that $C_{\infty}$ can be
obtained through a semi analytical procedure. In the limit of large
reactivity, we show that $C_{\infty}$ can be determined explicitly and is
given by :
$\displaystyle C_{\infty}\underset{\kappa a\gg D}{\sim}\begin{cases}\frac{\ln
2}{\pi D}+\frac{1}{\pi^{2}\kappa a}\left(\ln\frac{8\kappa
a}{D}+\gamma_{e}+1\right)&(d=2)\\\ \frac{1}{4Da}+\frac{1}{4\pi\kappa
a^{2}}\left(\ln\frac{2\kappa a}{D}+\gamma_{e}+1\right)&(d=3)\end{cases}$ (3)
where $\gamma_{e}$ is Euler’s constant. This expression is understood as the
first two terms in the expansion of $C_{\infty}$ in powers of $1/\kappa$.
Interestingly, this result shows that the term $C_{\infty}$ is not analytic in
powers of $\kappa$, which originates from the anomalous scaling (1). Finally,
we also give exact results in the small reactivity limit, which will be found
to be exactly the same (at first order) as the results obtained within the
self-consistent, “constant flux approximation” that has been invoked to study
the imperfect narrow escape in the literature [20, 39, 40, 37]. It is found
that, far from the reactive patch this approximation is very accurate (for any
reactivity), while it fails for initial positions close to the reactive patch,
and in particular does not predict the behavior (1) near the boundary of the
patch.
The outline of the paper is as follows. First, we recall the formalism of Ref.
[31] in the particular case of the imperfect narrow escape problem for
diffusing particles to obtain equations for the mean escape time in the large
volume limit (Section II.1). We show how the formalism can be presented under
the form of an integral equation that is suitable for studying the large and
small reactivity limits in Section II.2. The large reactivity limit is
investigated in Section III where Eqs. (1) and (3) are derived. In this
Section, we also give a simple scaling argument that relates the anomalous
behavior of $\langle T\rangle$ near the extremity of the patch to the
divergence of electric fields near the edges of conducting objects. The small
reactivity limit is examined in Section IV. Last, we study briefly how the
constant flux approximation can be implemented within our formalism in Section
V. An exact, but formal solution for any reactivity parameter (that requires
numerical tools, however) is presented in Appendix A.
## II Formalism for the imperfect narrow escape problem in the large volume
limit
### II.1 General formalism
We consider the stochastic motion of an overdamped particle moving with
diffusion coefficient $D$ in a confining volume $\Omega$. The boundary of the
volume is $\partial\Omega$ and contains a small window $S_{r}$ which is
partially reactive, the rest of the confining boundary is assumed to be smooth
and perfectly reflecting, see Fig. 1(a). We assume that $S_{r}$ is formed by
the region of the surface at geodesic distance less than $a$ from the center,
and $a$ is called the radius of the patch $S_{r}$. The Fokker-Planck equation
for the probability density $p(\mathbf{r},t)$ to observe the particle at
position $\mathbf{r}$ and time $t$ is
$\displaystyle\partial_{t}p=D\ \nabla^{2}p$
$\displaystyle(\mathbf{r}\in\Omega),$ (4) $\displaystyle\mathbf{n}\cdot\nabla
p=0$ $\displaystyle(\mathbf{r}\in\partial\Omega\backslash S_{r}),$ (5)
$\displaystyle D\ \mathbf{n}\cdot\nabla p+\kappa\ p=0$
$\displaystyle(\mathbf{r}\in S_{r}).$ (6)
where $\mathbf{n}$ is the unit vector normal to the surface, pointing to the
exterior of the volume. For a partially reactive surface, the reactivity
parameter $\kappa$ is defined in such a way that the probability that the
particle is absorbed by an infinitesimal surface element $dS$ located around
$\mathbf{r}_{s}$ during $dt$ is $\kappa\ p(\mathbf{r}_{s},t)dSdt$. We assume
that the space dimension is $d=2$ (2D) or $d=3$ (3D). It is very well known
that an equation for the mean first passage time can be obtained by
identifying the adjoint transport operator [1], which in our case leads to the
following equation for the mean reaction time $\langle T\rangle(\mathbf{r})$
to the target, where $\mathbf{r}$ now represents the initial position of the
particle:
$\displaystyle D\nabla^{2}\langle T\rangle=-1$
$\displaystyle(\mathbf{r}\in\Omega),$ (7)
$\displaystyle\mathbf{n}\cdot\nabla\langle T\rangle=0$
$\displaystyle(\mathbf{r}\in\partial\Omega\backslash S_{r}),$ (8)
$\displaystyle D\ \mathbf{n}\cdot\nabla\langle T\rangle+\kappa\langle
T\rangle=0$ $\displaystyle(\mathbf{r}\in S_{r}).$ (9)
Integrating Eq. (7) over the whole volume, and using the divergence formula
and the boundary conditions leads to the exact integral relation:
$\displaystyle\kappa\int_{S_{r}}dS(\mathbf{r})\langle T\rangle=V,$ (10)
where $V=|\Omega|$ is the volume of the domain. Now, we consider the large
volume limit, which is obtained when the confining volume extends without
changing its shape, keeping constant the size of the target and the initial
distance to the target. We define the rescaled mean escape time $\Phi$ by
$\displaystyle\Phi(\mathbf{r})=\lim_{V\to\infty}\langle
T(\mathbf{r})\rangle/V.$ (11)
In the large volume limit, the boundary at the vicinity of the reactive target
becomes increasingly similar to a flat surface in which the reactive patch is
a flat disk of radius $a$ in 3D (or a flat segment of length $2a$ in 2D).
Here, we denote the distance to the reflecting surface containing reactive
patch as $z$, see Fig. 1(b). With this in mind, inserting the ansatz (11) into
the above equations yields a closed system of equations in the large volume
limit:
$\displaystyle\nabla^{2}\Phi=0\ \ (\text{if }z>0),$ (12)
$\displaystyle\kappa\int_{S_{r}}dS(\mathbf{r})\Phi=1,$ (13) $\displaystyle
D\partial_{z}\Phi=\begin{cases}0&(\text{if }\ z=0,|\mathbf{r}|>a),\\\ \kappa\
\Phi&(\text{if }\ z=0,|\mathbf{r}|<a).\end{cases}$ (14)
Importantly, we see that in the large volume limit, Eq. (11), the obtained
equations are independent of the shape of the confining volume, which is
present only though the scale factor $V$ in the definition of $\Phi$. We have
directly controlled this aspect by performing numerical stochastic simulations
of trajectories in the confined domain. The results of such simulations are
shown on Fig. 2 and confirm that our formalism correctly predicts the mean
reaction time in the large volume limit, independently on the shape of the
confining domain.
Equations (12), (13), (14) generalize the formalism of Ref. [36] to the case
of imperfect reactions. The fact that the mean first reaction time scales with
the volume is actually more general than the specific diffusive walk that we
have considered here [31]. To solve the above equations, we may be tempted to
use spheroidal coordinates, which can be used to solve the problem for either
infinite or vanishing reactivity. For finite reactivity however, the resulting
equations in such coordinates involve Robin conditions with non-uniform
coefficients, so that the mean reaction time can be obtained only in terms of
the solution of an infinite linear system. This procedure is described in
Appendix A, and it indeed leads to a generic numerical solution that will be
useful to test our analytical insights in all the paper. However, it is not
suitable for analytical calculations. For this reason, we adopt a different
approach, consisting in deriving an integral equation satisfied by $\Phi$ on
the reactive patch.
Figure 2: (a) Geometry of the confining domains (called $A$ and $B$) that are
considered for stochastic simulations. In 2D, these domains are defined in
polar coordinates by $r(\theta)=Rf(\theta)$ with $f=1.6(1+0.5\cos^{2}\theta)$
for domain $A$ and $f=1.6(1+0.1\sin\theta+0.3\sin 3\theta)$ for domain $B$.
Domains in 3D are obtained by considering revolution of 2D curves around the
vertical dashed line. The reactive patch is indicated by a thick red line, and
the initial position is taken at a distance $r$ from the center of the patch
along the black dashed line. In the figure, we have used $R=6a$. (b),(c)
Results of Brownian dynamics simulations for the mean reaction time in 2D/3D
(parameters are indicated in the legend) compared to general theoretical
expressions as obtained in Appendix A. In all simulations, we used a time step
$\Delta t=10^{-4}a^{2}/D$. Boundary conditions are implemented as follows: if,
at the end of a time step, the random walker falls outside the domain, then if
it is “behind” a reflecting wall it is reflected with respect to this wall,
and if it falls “behind” the reactive patch, it is absorbed with probability
$P_{a}=\kappa\sqrt{\pi dt/D}$ (in which case the trajectory ends) and it is
reflected with probability $1-P_{a}$, see Ref. [43].
### II.2 Obtaining an integral equation for the mean reaction time
Let us first characterize the large distance behavior of the rescaled mean
first reaction time $\Phi$. The condition (13), combined with the boundary
condition at $z=0$, implies that
$\displaystyle\int_{S_{0}}dS\ \mathbf{n}\cdot\nabla\Phi=1/D,$ (15)
for any surface $S_{0}$ whose intercept with the plane $z=0$ encloses the
reactive patch. Taking such surface $S_{0}$ to be a half-disk of radius $R$
(in 2D) or a half-sphere (in 3D), we see that
$\displaystyle\partial_{r}\Phi\underset{r\to\infty}{\sim}\begin{cases}1/[\pi
Dr]&(d=2),\\\ 1/[2\pi Dr^{2}]&(d=3),\end{cases}$ (16)
where $r$ is the distance to the center of the reactive patch. Hence, the
behavior of $\Phi$ for large $r$ takes the form
$\displaystyle\Phi(\mathbf{r})\underset{|\mathbf{r}|\to\infty}{\sim}\begin{cases}\frac{1}{\pi
D}\ln|\mathbf{r}|+C_{\infty}+o(1)&(d=2),\\\ -\frac{1}{2\pi
Dr}+C_{\infty}+o(1)&(d=3),\end{cases}$ (17)
where $C_{\infty}$ does not depend on $\mathbf{r}$. The quantity $C_{\infty}$
thus characterizes the behavior of the mean reaction time far from the target,
it could be used in matched asymptotics expansions if one aims to identify the
first passage times distributions, as in ref. [22] (which deals with interior
targets).
Let us now introduce the Green’s function $G_{N}$ for the Laplace problem with
Neumann boundary conditions at $z=0$ (including on the reactive region). Such
Green’s function satisfy
$\displaystyle\nabla^{2}G_{N}(\mathbf{r}|\mathbf{r}_{0})=-\delta(\mathbf{r}-\mathbf{r}_{0}),$
(18) $\displaystyle\mathbf{n}\cdot\nabla
G_{N}(\mathbf{r}|\mathbf{r}_{0})=0\hskip 56.9055pt(z=0).$ (19)
The expression of $G_{N}$ is easily found by using the image method [44]:
$\displaystyle
G_{N}(\mathbf{r}|\mathbf{r}_{0})=\begin{cases}-\frac{1}{2\pi}[\ln|\mathbf{r}-\mathbf{r}_{0}|+\ln|\mathbf{r}-\mathbf{r}_{0}^{*}|]&(d=2),\\\
\frac{1}{4\pi}\left[\frac{1}{|\mathbf{r}-\mathbf{r}_{0}|}+\frac{1}{|\mathbf{r}-\mathbf{r}_{0}^{*}|}\right]&(d=3),\end{cases}$
(20)
where $\mathbf{r}_{0}^{*}$ represents the symmetric image of $\mathbf{r}_{0}$
with respect to the plane $z=0$. We note that the large $r$ behavior of
$G_{N}$ is
$\displaystyle
G_{N}(\mathbf{r}|\mathbf{r}_{0})\underset{|\mathbf{r}|\to\infty}{=}\begin{cases}-\frac{1}{\pi}\ln|\mathbf{r}|+o(1/r)&(d=2),\\\
\frac{1}{2\pi r}+o(1/r^{2})&(d=3),\end{cases}$ (21)
We now use manipulations that are standard in Green’s function problems [44]
to put the problem for $\Phi$ on the form of an integral equation. Using the
Eq. (18) and $\nabla^{2}\Phi=0$ we see that the following equality holds:
$\displaystyle\Phi(\mathbf{r}_{0})=\int_{z\geq
0}d\mathbf{r}[G_{N}(\mathbf{r}|\mathbf{r}_{0})\nabla_{\mathbf{r}}^{2}\Phi-\Phi(\mathbf{r})\nabla_{\mathbf{r}}^{2}G_{N}(\mathbf{r}|\mathbf{r}_{0})].$
(22)
Using the divergence formula, we obtain
$\displaystyle\Phi(\mathbf{r}_{0})=$
$\displaystyle\int_{S}dS(\mathbf{r})\mathbf{n}\cdot[G_{N}(\mathbf{r}|\mathbf{r}_{0})\nabla_{\mathbf{r}}\Phi(\mathbf{r})-\Phi(\mathbf{r})\nabla_{\mathbf{r}}G_{N}(\mathbf{r}|\mathbf{r}_{0})],$
(23)
where $S$ is any closed surface in the half-space $z\geq 0$. Taking this
surface to be a half-circle (or half-sphere in 3D) of radius $R$ joined with a
segment of size $2R$ on the axis $z=0$, we see that in the limit $R\to\infty$
$\displaystyle\Phi(\mathbf{r}_{0})=C_{\infty}-\frac{\kappa}{D}\int_{S_{r}}dS(\mathbf{r}_{s})\Phi(\mathbf{r}_{s})G_{N}(\mathbf{r}_{s}|\mathbf{r}_{0}),$
(24)
where we have used Eqs. (17),(21),(19) and (14) to simplify the integrals over
the surfaces in Eq. (23). The above equation means that $\Phi(\mathbf{r})$ can
be constructed for any position as soon as one knows its value on the reactive
patch. Taking $\mathbf{r}_{0}$ to be on the reactive patch yields an integral
equation for $\Phi_{s}$, defined to be the value of $\Phi$ on the reactive
patch. Since the above equation involves an unknown constant $C_{\infty}$ it
must be accompanied by a supplementary condition, which is provided by the
relation (13). Let us finally write explicitly the integral equations for
$\Phi$ for the 2D and the 3D cases. For $d=2$, we obtain:
$\displaystyle\Phi_{s}(x_{0})=C_{\infty}+\frac{\kappa}{D\pi}\int_{-a}^{a}dx\
\Phi_{s}(x)\ \ln|x-x_{0}|,$ (25) $\displaystyle\int_{-a}^{a}dx\
\Phi_{s}(x)=1/\kappa,$ (26)
In the case $d=3$, we note that $\Phi_{s}(\mathbf{r})$ depends only on the
radial distance to the disk center, $\Phi_{s}(\mathbf{r})=\Phi_{s}(r)$. The
kernel of the integral equation can be simplified after a few algebraic
manipulations detailed in Appendix B, leading to
$\displaystyle\Phi_{s}(r_{0})=C_{\infty}-\frac{\kappa}{D}\int_{0}^{a}dr\frac{2\
r\ \Phi_{s}(r)}{\pi(r+r_{0})}K\left(\frac{2\sqrt{rr_{0}}}{r+r_{0}}\right),$
(27) $\displaystyle 2\pi\int_{0}^{a}dr\ r\ \Phi_{s}(r)=1/\kappa,$ (28)
where $K(k)$ is the complete elliptic integral of the first kind, defined as
$K(k)=\int_{0}^{1}dt[(1-t^{2})(1-k^{2}t^{2})]^{-1/2}$ with $k$ the Elliptic
modulus (to be distinguished from the parameter $m=k^{2}$). These integral
equations admit no known analytical solution in general. In the next sections,
we focus on their asymptotic study. From now on, without loss of generality,
we set the units of length and time so that $a=1$ and $D/a^{2}=1$. The
remaining parameter $\kappa$ then represents $\kappa a/D$ in full units.
## III The limit of large reactivity
### III.1 A scaling argument for the anomalous behavior of the mean reaction
time for large reactivity.
Here we present a brief scaling argument that leads to the anomalous scaling
(1). In the case of perfect reactions $\kappa=\infty$, it is clear from Eqs.
(12),(14),(15) that $\Phi$ can be seen as the electrostatic potential
generated by a charged conducting disk that is embedded in an insulating
surface, with the prescription that $\Phi=0$ on this disk. In fact, using the
image methods, it is easy to show that $\Phi$ is also the electrostatic
potential in infinite space generated by an infinitely thin disk. It is well
known [41] that, for this electrostatic problem, the “electric field”
$-\nabla\Phi$ diverges near the edge as $1/\rho^{1/2}$, where $\rho$ is the
distance from the disk extremity. Therefore, at a distance $\rho\ll a$ from
the edges, $\Phi\propto\rho^{1/2}$. In the case that $\kappa$ is large but
finite, since the natural length scale associated to finite reactivity is
$\ell^{*}=D/\kappa$ [17], we may therefore assume that the mean reaction time
is comparable to the mean first passage time when the starting position is
located at a distance $\ell^{*}$ from the reactive patch. In this condition,
with $\rho=\ell^{*}$ we obtain the anomalous scaling $\Phi\propto
1/\sqrt{\kappa}$, as announced in Eq. (1). In what follows, we show how to
rigorously derive this scaling law, with the prefactor.
### III.2 2D case
#### III.2.1 Leading order
Consider now the limit $\kappa\to\infty$ in the case $d=2$. Since this
situation corresponds to a first passage problem, we know that the asymptotic
value of $C_{\infty}$ does not depend on $\kappa$. The fact that $\Phi_{s}$
vanishes in the large $\kappa$ limit leads us to postulate in line with Eqs.
(25)-(26) that
$\displaystyle\Phi_{s}(x_{0})\underset{\kappa\to\infty}{\sim}\frac{1}{\kappa}\Phi_{1}(x_{0}),\hskip
28.45274ptC_{\infty}\underset{\kappa\to\infty}{\sim}C_{1},$ (29)
where $C_{1}$ and $\Phi_{1}$ do not depend on $\kappa$. Inserting these ansatz
into the integral equation (25) and the normalisation condition (26), we
obtain
$\displaystyle\int_{-1}^{1}dx\ \Phi_{1}(x)\ln|x-x_{0}|=-\pi C_{1},$ (30)
$\displaystyle\int_{-1}^{1}dx\ \Phi_{1}(x)=1.$ (31)
The integral equation (30) for $\Phi_{1}(x)$ is known as Carleman’s equation
and its analytical solution is known explicitly [45]. Using also the
normalisation condition, we obtain the final expression for $C_{1}$ and
$\Phi_{1}$:
$\displaystyle C_{1}=\frac{\ln 2}{\pi},\
\Phi_{1}(x)=\frac{1}{\pi\sqrt{1-x^{2}}}.$ (32)
#### III.2.2 Boundary layer near the extremities of the reactive patch
We now note that $\Phi_{1}(x)$ is formally infinite at $x=\pm 1$, i.e. near
the boundary of the reactive patch. This means that our expansion (29) is not
valid near these points, suggesting a behavior similar to those obtained for
boundary layer problems. Since the reaction length [17] is $1/\kappa$ in our
units, we expect processes happening at such scales. Therefore, we assume the
behavior
$\displaystyle\Phi_{s}(x)=\kappa^{\alpha}\psi((1-|x|)\kappa),\hskip
28.45274pt(1-|x|)\ll 1,$ (33)
with $\psi$ a scaling function. The exponent $\alpha$ will be set such that
the behavior of $\Phi$ in the boundary layer matches with that far from the
boundary layer. Namely, the compatibility of the above ansatz with Eq. (32)
imposes the choice
$\displaystyle\alpha=-1/2,\hskip
28.45274pt\psi(X)\underset{X\to\infty}{\sim}\frac{1}{\pi\sqrt{2X}}.$ (34)
Hence, the structure of the solution in the limit $\kappa\to\infty$ is
$\displaystyle\Phi_{s}(x)=\begin{cases}\kappa^{-1}\Phi_{1}(x)+...&(1-|x|)\gg
1/\kappa\\\ \kappa^{-1/2}\psi((1-|x|)\kappa)+...&(1-|x|)\ll 1\end{cases}$ (35)
and the condition (34) ensures that these two expressions give the same result
in their common validity regime $\kappa^{-1}\ll 1-|x|\ll 1$. A key point here
is that the mean return time, starting from the boundary of the reactive
region scales as $1/\kappa^{1/2}$ and is thus infinitely larger than the mean
return time starting from the center of the target, which scales as
$1/\kappa$. This suggests to set $C_{\infty}=C_{1}+C_{1/2}/\sqrt{\kappa}+...$
in the limit of large reactivity (even though the constant $C_{1/2}$ will turn
out to vanish).
Let us now find the set of equations satisfied by $\psi$. First, we consider
the normalisation condition (26), which we write under the form
$\displaystyle\int_{-1}^{1}dx\left[\Phi_{s}(x)-\frac{\Phi_{1}(x)}{\kappa}\right]=0.$
(36)
Here we remark that the integrand in the above integral is maximal near $x=\pm
1$. Let us define an intermediate length scale $\ell$ such that
$\displaystyle 1/\kappa\ll\ell\ll 1.$ (37)
We will keep this notation in the whole paper. We write Eq. (36) by separating
the integral into two regions: when $(1-|x|)>\ell$ then we approximate
$\Phi_{s}(x)$ by the first line of Eq. (35) and if $(1-|x|)<\ell$ we use the
expressions on the second line of Eq. (35) for $\Phi_{s}$ and we approximate
by its behavior near $x=\pm 1$, which reads $\Phi_{1}\simeq
1/[\pi\sqrt{2(1-|x|)}]$. This leads to
$\displaystyle 0$
$\displaystyle=\int_{-1}^{-1+\ell}dx\left[\frac{\psi((1-|x|)\kappa)}{\sqrt{\kappa}}-\frac{1}{\kappa\pi\sqrt{2(1-|x|)}}\right]$
$\displaystyle+\int_{1-\ell}^{1}dx\left[\frac{\psi((1-|x|)\kappa)}{\sqrt{\kappa}}-\
\frac{1}{\kappa\pi\sqrt{2(1-|x|)}}\right].$
Setting $X=(1-|x|)\kappa$, and using $\ell\kappa\gg 1$, we obtain
$\displaystyle\int_{0}^{\infty}dX\left[\psi(X)-\frac{1}{\pi\sqrt{2X}}\right]=0.$
(38)
In order to find the equation satisfied by $\psi$ it is useful to write the
difference between the general equation (25) and the equation satisfied by
$\Phi_{1}$, to find
$\displaystyle\Phi_{s}(x_{0})=$ $\displaystyle\ C_{\infty}-C_{1}$
$\displaystyle+\frac{\kappa}{\pi}\int_{-1}^{1}dx\left[\Phi_{s}(x)-\frac{\Phi_{1}(x)}{\kappa}\right]\ln|x-x_{0}|.$
(39)
This leads to
$\displaystyle\Phi_{s}(x_{0})=\ C_{\infty}-C_{1}$
$\displaystyle+\frac{\kappa}{\pi}\int_{-1}^{-1+\ell}dx\left[\frac{\psi((1-|x|)\kappa)}{\sqrt{\kappa}}-\frac{\Phi_{1}(x)}{\kappa}\right]\ln|x-x_{0}|$
$\displaystyle+\frac{\kappa}{\pi}\int_{1-\ell}^{1}dx\left[\frac{\psi((1-|x|)\kappa)}{\sqrt{\kappa}}-\frac{\Phi_{1}(x)}{\kappa}\right]\ln|x-x_{0}|.$
(40)
For $x_{0}=1-X_{0}/\kappa$, if we set $X=(1-|x|)\kappa$, this yields
$\displaystyle\psi(X_{0})=C_{1/2}+\int_{0}^{\ell\kappa}\frac{dX}{\pi}\left[\psi(X)-\frac{1}{\pi\sqrt{2X}}\right]\ln\frac{|X-X_{0}|}{\kappa}$
$\displaystyle+\frac{1}{\pi}\int_{0}^{\ell\kappa}dX\left[\psi(X)-\frac{1}{\pi\sqrt{2X}}\right]\ln\left|2-\frac{X+X_{0}}{\kappa}\right|.$
(41)
Now, in the limit $\kappa\to\infty$, noting that $\ell\kappa\to\infty$ [by
definition of $\ell$, see Eq. (37)] and using the previously found condition
(38) we obtain
$\displaystyle\psi(X_{0})$ $\displaystyle=C_{1/2}$
$\displaystyle+\int_{0}^{\infty}dX\left[\psi(X)-\frac{1}{\pi\sqrt{2X}}\right]\frac{\ln|X-X_{0}|}{\pi}.$
(42)
This is a Wiener-Hopf integral equation for the unknown function $\psi(X)$. We
solve it in the next section.
#### III.2.3 Solution of the Wiener-Hopf equation (42)
We solve the Wiener-Hopf integral equation with Carleman’s method, as
described in Ref. [45]. We note that a similar equation has appeared in
viscous flow theory [46, 47, 48] but the differences between our equations and
the equation studied in these references justify the fact to solve it here in
detail. First, let us introduce the following notations. We denote $f_{+}(X)$
all functions (depending on the real variable $X$) that vanish for all $X<0$,
and $f_{-}(X)$ all functions that vanish for $X>0$. For any function $f(X)$
one can write $f(X)=f_{+}(X)+f_{-}(X)$, with $f_{+}(X)=f(X)\theta(X)$ and
$f_{-}(X)=f(X)\theta(-X)$, where $\theta$ is the Heaviside step function. We
introduce the complex Fourier transform and its inverse:
$\displaystyle\hat{f}(z)=\int_{-\infty}^{\infty}dXf(X)e^{-izX},$ (43)
$\displaystyle f(X)=\frac{1}{2\pi}\int_{-\infty}^{\infty}du\hat{f}(u)e^{+iuX}$
(44)
where $z$ represents a complex number and $u$ a real number. We denote
$\hat{f}_{+}(z)$ the Fourier transform of the function $f_{+}(X)$, and
$\hat{f}_{-}(z)$ the Fourier transform of $f_{-}(z)$. Typically, Fourier
transforms of the form $\hat{f}_{+}(z)$ are defined in the lower complex half-
plane $\text{Im}(z)\leq 0$, Fourier transforms of the form $\hat{f}_{-}(z)$
are defined in the upper complex half-plane $\text{Im}(z)\geq 0$ (as long as
$f_{\pm}(x)$ does not diverge exponentially at $x\to\pm\infty$). Now, we can
define $\psi_{+}(X)\equiv\psi(X)\theta(X)$, and we introduce
$\displaystyle K(X)=\frac{1}{\pi}\ln|X|,\hskip
28.45274pt\chi_{+}(X)=\frac{1}{\pi\sqrt{2X}}\theta(X).$ (45)
The integral equation (42) can be generalized for negative $X_{0}$ by writing
$\displaystyle\psi_{+}(X_{0})=$
$\displaystyle\int_{0}^{\infty}dX\left[\psi_{+}(X)-\chi_{+}(X)\right]K(X-X_{0})+y_{-}(X_{0}),$
(46)
where the only remarkable property of $y_{-}(X_{0})$ is that it vanishes for
positive $X_{0}$. Note that we have assumed that $C_{1/2}=0$, this will be
justified at the end of the calculation by the fact that the obtained solution
satisfies the normalization condition (38) for this value of $C_{1/2}$. Taking
the Fourier transform of the above equation, we obtain
$\displaystyle\hat{\psi}_{+}(u)=[\hat{\psi}_{+}(u)-\hat{\chi}_{+}(u)]\hat{K}(u)+\hat{y}_{-}(u).$
(47)
Calculating the Fourier transforms leads to
$\displaystyle\hat{\psi}_{+}(u)\left(1+\frac{1}{|u|}\right)=-\frac{1-i\
\text{sign}(u)}{2\sqrt{\pi}|u|^{3/2}}+\hat{y}_{-}(u),$ (48)
where $\text{sign}(u)=\theta(u)-\theta(-u)$ is the sign function. This
equation can be considerably simplified by introducing an auxilliary function
$S_{-}(X)$ defined by
$\displaystyle S_{-}(X)=\theta(-X)\frac{\sqrt{2\left|X\right|}}{\pi},\
\hat{S}_{-}(u)=\frac{1-i\ \text{sign}(u)}{2\sqrt{\pi}\left|u\right|^{3/2}},$
(49)
so that the Wiener-Hopf equation can be written as
$\displaystyle\hat{\psi}_{+}(u)\left(1+\frac{1}{|u|}\right)=-\hat{S}_{-}(u)+\hat{y}_{-}(u)=\hat{f}_{-}(u),$
(50)
where $\hat{f}_{-}(u)$ is the Fourier transform of another unknown function
$f_{-}(X)$, whose only remarkable property is to vanish for positive $X$. This
kind of equations is known as a homogeneous Wiener-Hopf equation, the method
to solve it consists in obtaining a factorization of the form
$\hat{\psi}_{+}(u)\hat{g}_{+}(u)=\hat{g}_{-}(u)$. To this end, we write the
Wiener-Hopf equation under the form
$\displaystyle\hat{\psi}_{+}(u)e^{\hat{W}(u)}=\hat{f}_{-}(u),$ (51)
with
$\displaystyle\hat{W}(u)=\ln[1+1/|u|],$ (52)
and we seek a factorization $\hat{W}(u)=\hat{W}_{+}(u)+\hat{W}_{-}(u)$. This
can be done by calculating its inverse Fourier transform:
$\displaystyle W(X)=\frac{\cos(X)[2\
\text{Si}(\left|X\right|)-\pi]+\pi}{2\pi\left|X\right|}-\frac{\text{Ci}(\left|X\right|)\sin(X)}{\pi
X}$ (53)
where Ci and Si are the integral cosine and integral sine functions
$\displaystyle\text{Ci}(X)=-\int_{X}^{\infty}dt\frac{\cos(t)}{t},\hskip
8.5359pt\text{Si}(X)=\int_{0}^{X}dt\frac{\sin(t)}{t}.$ (54)
A factorization may thus be obtained by setting $W(x)=W_{+}(x)+W_{-}(x)$, i.e
$W_{+}(x)=W(x)\theta(x)$ and $W_{-}(x)=W(x)\theta(-x)$. Now, we write the
equation (51) as
$\displaystyle\hat{\psi}_{+}(u)e^{\hat{W}_{+}(u)}=\hat{f}_{-}(u)e^{-\hat{W}_{-}(u)}.$
(55)
We are now in the favorable case: the terms on left-hand-side are analytic
functions in the upper complex plane, those on the right are analytic in the
lower complex plane except for one pole at $z=0$, and these terms are equal on
the real axis. According to the theorem of analytic continuation, combined
with the Cauchy theorem, we conclude that both terms are equal to a constant
plus a $1/z$ term on the whole complex plane [45]. We thus have
$\displaystyle\hat{\psi}_{+}(u)e^{\hat{W}_{+}(u)}=a_{0}+\frac{a_{1}}{u},$ (56)
where the constants $a_{0},a_{1}$ will be found by requiring that $\psi(X)$ is
a solution to our problem. Since $\hat{W}_{+}(z)$ is defined on the lower
complex plane, we may consider the above equations on the lower imaginary axis
$u=-is$, in which case the above equality can be written in terms of Laplace
transforms, with the usual notation
$\tilde{f}_{+}(s)=\int_{0}^{\infty}dte^{-st}f_{+}(t)=\hat{f}_{+}(-is)$:
$\displaystyle\tilde{\psi}_{+}(s)=\left(a_{0}+\frac{i\
a_{1}}{s}\right)e^{-\tilde{W}^{+}(s)}.$ (57)
The Laplace transform $\tilde{W}^{+}(s)$ can be identified by calculating its
derivative, i.e. the Laplace transform of $-xW(x)$, and then by integrating
over $s$; this leads to
$\displaystyle\tilde{W}^{+}(s)=\frac{1}{4}\ln\frac{1+s^{2}}{s^{2}}+m(s),$ (58)
with
$\displaystyle m(s)=-\int_{0}^{s}\frac{dw\ \ln w}{\pi(1+w^{2})}.$ (59)
We know that the behavior of $\psi$ for large arguments is given by the
matching condition Eq. (34), which translates to the small-$s$ behavior:
$\displaystyle\tilde{\psi}_{+}(s)\underset{s\to 0}{\sim}\frac{1}{\sqrt{2\pi
s}}.$ (60)
Inserting Eq. (58) into (57) and taking the small-$s$ limit, we see that the
above behavior is obtained for $ia_{1}=1/\sqrt{2\pi}$. Next, the value of
$a_{0}$ is found by requiring that $\psi(0)$ is finite, so that
$\tilde{\psi}(s)$ vanishes in the limit $s\to\infty$; this leads to $a_{0}=0$.
Hence, the final expression for the function $\psi$ is given in the Laplace
domain by
$\displaystyle\tilde{\psi}_{+}(s)=\frac{1}{\sqrt{2\pi
s}(1+s^{2})^{1/4}}e^{-m(s)}.$ (61)
Finally, we must check that the normalization condition (38) holds, this is
readily done by noting that $\tilde{\psi}(s)-1/\sqrt{2\pi
s}=\mathcal{O}(\sqrt{s}\ln s)$ vanishes for small $s$. This justifies our
hypothesis $C_{1/2}=0$. Unfortunately, the Laplace inversion cannot be
performed so that we know only $\psi(X)$ in closed form, however we can easily
derive the asymptotic behavior of $\psi(X)$ for small and large arguments. The
study of the asymptotic behavior of $\tilde{\psi}(s)$ for large $s$ leads
readily to the initial value of $\psi(X)$:
$\psi(0)=\lim_{s\to\infty}s\tilde{\psi}(s)=\frac{1}{\sqrt{2\pi}},$ (62)
this justifies the previously announced result (1). The large $X$ behavior can
be computed by noting that the Laplace transform of $X\psi(X)$ is
$d\tilde{\psi}/ds$ and by expanding this one for small $s$, with the result:
$\displaystyle\psi(X)\underset{X\to\infty}{\simeq}\frac{1}{\sqrt{2X}\pi}+\frac{\ln(4X)+\gamma_{e}-1}{\pi^{2}(2X)^{3/2}}+...$
(63)
with $\gamma_{e}$ the Euler-Mascheroni constant. A formula for $\psi(X)$ can
be obtained by considering the inverse Laplace transform of $\tilde{\psi}(s)$
with the Mellin’s inverse formula, by using a contour that follows the
negative real axis (above and below), such Laplace inversion is obtained in
Appendix 4 of Ref. [48]:
$\displaystyle\psi(X)=\frac{1}{\pi}\int_{0}^{\infty}dp\
\frac{e^{-pX+m(p)}}{\sqrt{2\pi p}(1+p^{2})^{3/4}}.$ (64)
In summary, here we have obtained an analytic expression in Laplace space for
the scaling function $\psi(X)$ which characterizes the behavior of the mean
first passage time near the extremities of the reactive patch in two
dimensions. The validity of our approach is checked on Figure 3 by comparing
with exact numerical results obtained from the general form of the solution.
Figure 3: Behavior of the mean first reaction time near the extremities of the
reactive patch. Symbols: exact general solution obtained numerically in
Appendix A in 2D (upper symbols) and 3D (lower symbols). We also represent the
values of $\psi$ and $\psi^{3d}$ obtained from Eqs. (64) and (78).
#### III.2.4 Next-to-leading order expansion
Up to now, the constant $C_{\infty}$, which characterizes the behavior of the
mean reaction time when the initial position is far from the target, has been
obtained at leading order only in the large reactivity limit, with the same
result as in the case of a perfectly reactive patch. Here we show how to
obtain the first non-trivial correction for $C_{\infty}$ for large reactivity,
with the result that $C_{\infty}(\kappa)$ is not analytic in $\kappa$. First,
we note that when $1/\kappa\ll 1-|x|\ll 1$, Eqs. (63) and (35) indicate that
$\displaystyle\Phi_{s}(x)\simeq\frac{1}{\kappa\pi\sqrt{2(1-|x|)}}+\frac{{\color[rgb]{0,0,0}\ln[4(1-|x|)\kappa]}+\gamma_{e}-1}{\kappa^{2}\pi^{2}[2(1-|x|)]^{3/2}}+...$
(65)
This suggests that, outside the boundary layer, the next-to-leading order
behavior of $\Phi_{s}$ reads:
$\displaystyle\Phi_{s}(x)=\frac{\Phi_{1}}{\kappa}+\frac{\Phi_{2}^{*}\ln\kappa+\Phi_{2}}{\kappa^{2}}+...,$
(66)
because this expression can be matched with (65) by imposing that
$\displaystyle\Phi_{2}^{*}(x)\simeq\frac{1}{\pi^{2}[2(1-|x|)]^{3/2}},$
$\displaystyle(x\to\pm 1),$ (67)
$\displaystyle\Phi_{2}(x)\simeq\frac{\ln[4(1-|x|)]+\gamma_{e}-1}{\pi^{2}[2(1-|x|)]^{3/2}},$
$\displaystyle(x\to\pm 1).$ (68)
These expansions also lead us to assume that, for large reactivity the
constant $C_{\infty}$ behaves as
$\displaystyle C_{\infty}=C_{1}+\frac{C_{2}+C_{2}^{*}\ln\kappa}{\kappa}+...$
$\displaystyle(\kappa\to\infty)$ (69)
The equation for $\Phi_{2}$ and $\Phi_{2}^{*}$ can be identified as follows.
We consider again the intermediate length scale $\ell$ satisfying (37), and we
start from the integral equation (39), which we write as
$\displaystyle\int_{-1+\ell}^{1-\ell}dx\left[\frac{\Phi_{2}^{*}(x)\ln\kappa+\Phi_{2}(x)}{\kappa}\right]\ln|x-x_{0}|=$
$\displaystyle-\pi\frac{C_{2}+C_{2}^{*}\ln\kappa}{\kappa}+\frac{\pi\phi_{1}(x_{0})}{\kappa}+B(x_{0})+B(-x_{0}),$
(70)
where $B$ contains all the terms which appear due to the fact that the
integral over $\Phi_{2}$ in the above equation is evaluated over a truncated
interval $]-1+\ell;1+\ell[$ instead of $]-1;1[$, so that
$\displaystyle B=$
$\displaystyle\int_{0}^{\ell\kappa}\frac{dX}{\sqrt{\kappa}}\left(\frac{1}{\pi\sqrt{2X}}-\psi(X)\right)\ln\left(1-x_{0}-\frac{X}{\kappa}\right).$
(71)
To proceed further, we consider (70) as an integral equation for
$\Phi_{2}+\ln\kappa\Phi_{2}^{*}$ over the truncated interval
$]-1+\ell;1+\ell[$. Its solution is analytically known and we identify the
constants $C_{2}$ and $C_{2}^{*}$ by requiring that the normalisation
condition is satisfied at this order of $\kappa$. This procedure requires to
evaluate $B$ in the limit $\ell\to 0$ without assuming that $\ell\ll
1-|x_{0}|$, and it turns out to be relatively tedious. The calculation is
described in Appendix C.1, and leads to the explicit results
$\displaystyle C_{2}^{*}=\frac{1}{\pi^{2}},$ $\displaystyle
C_{2}=\frac{1+\gamma_{e}+\ln 8}{\pi^{2}}.$ (72)
These values of $C_{2}$ and $C_{2}^{*}$ are in excellent agreement with the
exact solution for $\Phi_{s}$ obtained numerically, as shown in Fig. 4(a).
Figure 4: Comparison of the values of $C_{\infty}$ obtained numerically
(symbols, see Appendix A for details), with the analytical predictions in Eqs.
(72) and (81) (black lines), for the two-dimensional (a) and three-dimensional
(b) domains.
### III.3 3D case
We now adapt the approach to the 3D case. It turns out that the solution
admits the same scaling behaviors than in 2D:
$\displaystyle\Phi_{s}(r)=\begin{cases}\frac{\Phi_{1}(r)}{\kappa}+\frac{\Phi_{2}^{*}(r)\ln\kappa+\Phi_{2}(r)}{\kappa^{2}}&(1-r)\gg
1/\kappa,\\\ \frac{1}{\sqrt{\kappa}}\psi^{3d}((1-r)\kappa)&(1-r)\ll
1,\end{cases}$ (73)
where the first line is the expansion of $\Phi_{s}(r)$ in powers of $\kappa$
at fixed $r$, and the second line the expansion of $\Phi$ in powers of
$\kappa$ at fixed $X=(1-r)\sqrt{\kappa}$. At leading order, the integral
equation for $\Phi_{1}$ reads
$\displaystyle 0=C_{1}-\int_{0}^{1}dr\
\frac{2r}{\pi(r+r_{0})}K\left(\frac{2\sqrt{rr_{0}}}{r+r_{0}}\right)\Phi_{1}(r),$
(74) $\displaystyle 2\pi\int_{0}^{1}dr\ r\ \Phi_{1}(r)=1.$ (75)
The solution of the above integral equation (where $r\ \Phi_{1}(r)$ is
considered to be the unknown function) is known [45] and this leads to the
solution
$\displaystyle\Phi_{1}(r)=\frac{1}{2\pi\sqrt{1-r^{2}}},\hskip
28.45274ptC_{1}=\frac{1}{4}.$ (76)
We thus note that
$\displaystyle\Phi_{1}(r\to 1)\sim\frac{1}{2\pi\sqrt{2(1-r)}}.$ (77)
In the boundary layer near $r=1$, we set $r=1-X/\kappa$,
$r_{0}=1-X_{0}/\kappa$, and $\Phi_{s}(r)=1/\sqrt{\kappa}\psi(X)$. With these
scalings we can expand the integral equation (27) with the result that
$\psi^{3d}$ satisfies exactly the same equation than in 2D, the only
difference is that it has to match with $\psi^{3d}(X)\sim 1/(2\pi\sqrt{2X})$
for large $X$ [due to Eq. (77)] and there is thus a factor of 2 that arises
when we compare to the 2D case:
$\displaystyle\psi^{3d}(X)=\frac{1}{2}\psi(X).$ (78)
This relation is checked on Fig. 3. Let us now identify the next order terms
in 3D. Inserting the ansatz (73) into the integral equation (27) and expanding
at second order, we obtain
$\displaystyle\int_{0}^{1-\ell}\ \frac{dr\
r}{r+r_{0}}K\left(\frac{2\sqrt{rr_{0}}}{r+r_{0}}\right)[\Phi_{2}(r)+\Phi_{2}^{*}(r)\ln\kappa]=$
$\displaystyle-\frac{\pi\Phi_{1}}{2}+\frac{\pi}{2}[C_{2}^{*}\ln\kappa+C_{2}]+B$
(79)
where the term $B$ compensates the fact that the above integrals are evaluated
over the truncated interval $[0;1-\ell[$, so that:
$\displaystyle B(r_{0},\ell)=-\sqrt{\kappa}\int_{1-\ell}^{1}\ \frac{dr\
r}{r+r_{0}}K\left(\frac{2\sqrt{rr_{0}}}{r+r_{0}}\right)$
$\displaystyle\times\left[\psi^{3d}((1-r)\kappa)-\frac{1}{2\pi\sqrt{2(1-r)\kappa}}\right].$
(80)
As in the 2D situation, we consider (79) as an integral equation for which the
solution is analytically known; and we then chose $C_{2}$ and $C_{2}^{*}$ so
that the normalisation condition for $\Phi_{s}$ holds at all orders of
$\kappa$. The final result is
$\displaystyle C_{2}^{*}=\frac{1}{4\pi},\
C_{2}=\frac{\gamma_{e}+1+\ln(2)}{4\pi},$ (81)
and it agrees perfectly with numerical solutions, as shown in Fig. 4(b).
## IV The limit of small reactivity
Let us now consider the limit $\kappa\to 0$. At leading order, the mean
reaction time is homogeneous. We seek a solution under the form
$\displaystyle\Phi_{s}(x)=\frac{1}{\kappa}\sum_{n\geq 0}f_{n}(x)\kappa^{n},\
C_{\infty}=\frac{1}{\kappa}\sum_{n\geq 0}c_{n}\kappa^{n}.$ (82)
At leading order, we obtain
$\displaystyle f_{0}=c_{0}=1/|S_{r}|.$ (83)
where $S_{r}$ is the length (in 2D) or the area (in 3D) of the reactive patch.
Furthermore, next-orders can be found iteratively by using
$\displaystyle
f_{n}(\mathbf{r})=c_{n}-\frac{1}{D}\int_{S_{r}}dS(\mathbf{r}^{\prime})f_{n-1}(\mathbf{r^{\prime}})G_{N}(\mathbf{r}|\mathbf{r}^{\prime})$
(84)
with the condition for $n\geq 1$:
$\displaystyle\int_{S_{r}}dSf_{n}=0.$ (85)
For $d=2$, the explicit computations can be done for the first orders, and we
find
$\displaystyle c_{0}=1/2;\ c_{1}=\frac{3-\ln 4}{2\pi},\
c_{2}=\frac{2}{9}-\frac{7}{3\pi^{2}}.$ (86)
In the 3D situation, the leading order is simply
$\displaystyle f_{0}=c_{0}=1/\pi;$ (87)
and the recurrence relation is
$\displaystyle
f_{n}(x)=c_{n}-\frac{2}{\pi}\int_{0}^{1}dyf_{n-1}(y)\frac{y}{x+y}K\left(\frac{2\sqrt{xy}}{x+y}\right).$
(88)
Unfortunately, it seems very difficult to calculate these integrals, and even
at first order the coefficient $c_{1}$ can be calculated only numerically:
$\displaystyle
c_{1}=\frac{4}{\pi^{2}}\int_{0}^{1}dx\int_{0}^{1}dy^{\prime}\frac{xy}{x+y}K\left(\frac{2\sqrt{xy}}{x+y}\right)\simeq
0.27.$ (89)
## V Comparison with the constant flux approximation
The constant flux approximation (CFA) [20] has been used in many recent
studies [39, 40, 37] on imperfect reactivity in confinement, and here we
consider how this approximation compares to the exact results in our
formalism. First, we need to adapt this approximation to our situation of
large volume limit. In the CFA, one replaces the Robin condition (14) by
inhomogeneous Neumann conditions:
$\displaystyle D\partial_{z}\Phi=\begin{cases}0&(z=0,r>a)\\\
-Q&(z=0,r<a)\end{cases},$ (90)
where the flux $Q$ is assumed to be constant on the reactive patch and will be
determined self-consistently with a closure relation. A natural choice of
closure relation is to impose that Robin condition is satisfied on average,
hence
$\displaystyle Q=\kappa\int_{S_{r}}dS\ \Phi_{s},$ (91)
but we also have the normalization condition (13), so that
$\displaystyle Q=1.$ (92)
Now, inserting (90) into (23) leads directly to a solution for $\Phi$ within
CFA:
$\displaystyle\Phi(\mathbf{r}_{0})=C_{\infty}+\frac{Q}{D}\int_{S_{r}}dS(\mathbf{r}_{s})G_{N}(\mathbf{r}_{s}|\mathbf{r}_{0}).$
(93)
Integrating over $S$ and using (92) and (91), we obtain the CFA value of
$C_{\infty}$:
$\displaystyle C_{\infty}^{\text{cfa}}=\frac{1}{\kappa
S_{r}}\left(1-\frac{\kappa}{D}\int_{S_{r}}dS(\mathbf{r})\int_{S_{r}}dS(\mathbf{r}_{0})G_{N}(\mathbf{r}|\mathbf{r}_{0})\right)$
(94)
Comparing with the results of Sec. IV, we see that in the CFA, $C_{\infty}$ is
exactly the same as the next-to-leading order expansion of $C_{\infty}$ in the
limit of low reactivity, i.e.,
$\displaystyle C_{\infty}^{\text{cfa}}=\frac{c_{0}}{\kappa}+c_{1}$ (95)
It may be therefore surprising that CFA works for $C_{\infty}$ even for rather
large values of the reactivity (Fig. 5), but this comes from the fact that the
value of $c_{1}$ turns out to be extremely close to the exact value of
$C_{\infty}(\kappa=\infty)$ (the difference is of the order of a few
percents). This might be the reason why the CFA approach can be implemented to
yield accurate results in other contexts. However, the value of the mean first
passage time near the extremity of the reactive patch is not well captured by
this approximation, since it is obvious in Eq. (93) that it does not scale as
$1/\sqrt{\kappa}$ for large $\kappa$, contrary to what we have found.
Figure 5: Values of $C_{\infty}$ in 2D (a) and 3D (b), as found from the exact
numerical solution compared to the exact large and small reactivity
asymptotics. Note that the constant flux approximation (CFA) is exactly
equivalent to the first order expansion in the limit of low reactivity. Here
we use the units so that $D=1$, $a=1$.
## VI Conclusion
In this paper we have considered the imperfect narrow escape problem for
diffusive particles in confinement. We have established a general formalism
which provides the mean reaction time in the large volume limit for any value
of the reactivity parameter. We have obtained explicit results in $d=2$ and
$d=3$ in the respective limits of low and large reactivity parameter. Our most
surprising result is the scaling of the mean reaction time when the initial
position is at the extremity of the imperfect patch; this mean return time
scales as $\kappa^{-1/2}$ and is thus much larger than the naively expected
scaling $1/\kappa$. Interestingly, we have shown that this anomalous scaling
is closely related to the divergence of the electric field near corners and
edges of conducting objects [41], which is also responsible for the existence
of coffee rings [42] or the crispiness of the extremities of cooked potatoes
[49]. We have explicitly identified the prefactor of this scaling law by
solving a Wiener-Hopf equation. We have also identified a non-analytic
behavior for the capacitances of the imperfect patches as a function of the
reactivity. We note that we have restricted ourselves to the case of circular
patches, but we believe that for the more general patches with a smooth
boundary the asymptotic scaling laws should remain unchanged. Finally, we have
made a link between the results obtained within the Constant Flux
Approximation (CFA) and the low reactivity limit. It turns out that the CFA
gives very accurate predictions of the mean reaction time when the initial
position is far from the target, but fails to predict the correct behavior of
the mean return times; when the initial position is on the reactive patch. In
the future, one could adapt our formalism to multiple targets, for example to
generalize the classical calculation [12] of the absorption time by a sphere
covered by reactive patches to imperfect patches. Our results provide a
general framework to quantify the mean reaction times for the imperfect narrow
escape problem.
###### Acknowledgements.
Computer time for this study was provided by the computing facilities MCIA
(Mesocentre de Calcul Intensif Aquitain) of the Université de Bordeaux and of
the Université de Pau et des Pays de l’Adour. T. G. acknowledges the support
of the grant ComplexEncounters, ANR-21-CE30-0020.
## Appendix A Exact general form of the solution $\Phi(\mathbf{r})$
### A.1 Imperfect narrow escape problem in 2D
Here we describe a way to obtain the general solution of the problem formed by
Eqs. (12)-(14). It consists in writing the equations in a set of orthogonal
coordinates and using the standard method of separation of variables. We first
describe this approach in the 2D case, for which we use the elliptic
coordinates $\mu,\nu$ defined as
$\displaystyle x=a\ \mathrm{ch}(\mu)\cos(\nu),\ y=a\
\mathrm{sh}(\mu)\sin(\nu).$ (96)
We calculate the scale factors $h_{i}=|\partial_{i}\mathbf{r}|$ with
$i\in\\{\mu,\nu\\}$:
$\displaystyle h_{\mu}=h_{\nu}=a\sqrt{\mathrm{ch}^{2}(\mu)-\cos^{2}\nu}.$ (97)
The Laplace equation satisfied by $\Phi$ and the reflecting boundary
conditions outside the reactive patch are written in this coordinate system as
$\displaystyle\partial_{\nu}^{2}\Phi+\partial_{\mu}^{2}\Phi=0,\ $
$\displaystyle\partial_{\nu}\Phi|_{\nu=0}=\partial_{\nu}\Phi|_{\nu=\pi/2}=0.$
(98)
The general solution for these equations can be written by using the method of
separation of variables, which leads to
$\displaystyle\Phi=B\mu+\sum_{n=0}^{\infty}\phi_{n}\ e^{-2n\mu}\cos(2n\nu).$
(99)
Furthermore, the normalization condition (13) can also be written $D\int
dS\partial_{n}\Phi=1$ for any surface surrounding the target. Far from the
target, this means that $\partial_{r}\Phi=1/(\pi rD)$. Noting that
$\mu\simeq\ln(2r/a)$ for large $r$, we thus find
$\displaystyle B=1/(\pi D).$ (100)
We also note that the quantity $C_{\infty}$ is given, in this mode
decomposition, by
$\displaystyle C_{\infty}=\frac{\ln 2}{\pi D}+\Phi_{0}.$ (101)
Finally, the Robin condition at the target surface reads
$\displaystyle
D\partial_{n}\Phi+\kappa\Phi=\left(-\frac{D}{h_{\mu}}\partial_{\mu}\Phi+\kappa\Phi\right)_{\mu=0}=0,$
(102)
so that
$\displaystyle D\partial_{\mu}\Phi|_{\mu=0}=\kappa\ a\ \sin\nu\
\Phi|_{\mu=0}.$ (103)
Using this condition and the form of the general solution (99), we find that
the coefficients $\phi_{n}$ are solution of the infinite linear system
$\displaystyle\pi m\phi_{m}+\frac{\kappa
a}{D}\sum_{n=0}^{\infty}A_{mn}\phi_{n}=\delta_{m,0},$ (104)
which is satisfied for all positive integers $m$, with
$\displaystyle A_{nm}=\int_{0}^{\pi}d\nu\sin\nu\cos(2m\nu)\cos(2n\nu)$
$\displaystyle=\frac{2[1-4(m^{2}+n^{2})]}{16(m^{4}+n^{4})+1-8(m^{2}+n^{2})-32m^{2}n^{2}}.$
(105)
In practice, this linear system (104) can be solved numerically by taking into
account only a finite number of modes $N$, and checking that the obtained
quantities do not depend on $N$ for large $N$. Note also that $C_{\infty}$ can
be directly calculated by using Eq. (101).
### A.2 3D case
This approach can be adapted to the 3D case, for which we use orthogonal
coordinates defined as
$\displaystyle x=a\sqrt{(1+\alpha^{2})(1-\beta^{2})}\cos\varphi,$ (106)
$\displaystyle y=a\sqrt{(1+\alpha^{2})(1-\beta^{2})}\sin\varphi,$ (107)
$\displaystyle z=a\alpha\beta,$ (108)
where $\varphi$ is the azimuthal angle. Note that $\alpha>0$ and
$\beta\in[0;1]$ are related to the standard oblate spheroidal coordinates
($\mu,\nu,\varphi)$ by $\alpha=\sinh(\mu)$ and $\beta=\sin(\nu)$. Inversion
formulas read
$\displaystyle\alpha=\sqrt{\frac{\left(\frac{r}{a}\right)^{2}-1+\sqrt{1+\left(\frac{r}{a}\right)^{4}+2\left(\frac{r}{a}\right)^{2}\cos(2\theta)}}{2}},$
(109)
$\displaystyle\beta=\sqrt{\frac{1-\left(\frac{r}{a}\right)^{2}+\sqrt{1+\left(\frac{r}{a}\right)^{4}+2\left(\frac{r}{a}\right)^{2}\cos(2\theta)}}{2}},$
(110)
with $(r,\theta,\varphi)$ the usual spherical coordinates. It is useful to
calculate the scale factors $h_{i}=|\partial{\mathbf{r}}/\partial i|$, with
$i=\\{\alpha,\beta,\varphi\\}$,
$\displaystyle
h_{\alpha}=a\left(\frac{\alpha^{2}+\beta^{2}}{1+\alpha^{2}}\right)^{1/2},\
h_{\beta}=a\left(\frac{\alpha^{2}+\beta^{2}}{1-\beta^{2}}\right)^{1/2},$
$\displaystyle h_{\varphi}=a\left[(1+\alpha^{2})(1-\beta^{2})\right]^{1/2}.$
(111)
For axisymmetric functions, the Laplacian reads in this orthogonal coordinates
$\displaystyle\nabla^{2}\Phi=\frac{1}{h_{\alpha}h_{\beta}h_{\varphi}}\left(\frac{\partial}{\partial\alpha}\frac{h_{\beta}h_{\varphi}}{h_{\alpha}}\frac{\partial\Phi}{\partial\alpha}+\frac{\partial}{\partial\beta}\frac{h_{\alpha}h_{\varphi}}{h_{\beta}}\frac{\partial\Phi}{\partial\beta},\right),$
(112)
so that $\Phi$ satisfies the equation
$\displaystyle\frac{\partial}{\partial\alpha}(1+\alpha^{2})\frac{\partial\Phi}{\partial\alpha}+\frac{\partial}{\partial\beta}(1-\beta^{2})\frac{\partial\Phi}{\partial\beta}=0.$
(113)
We impose Neumann conditions for $\beta=0$ and $\beta=1$, at which
$\partial_{\beta}\Phi=0$. With these conditions, the general solution can be
found by the method of separation of variables, which leads to
$\displaystyle\Phi(\alpha,\beta)=\Phi_{\infty}+\sum_{q=0}^{\infty}a_{q}\
g_{q}(\alpha)\ P_{2q}(\beta),$ (114)
where $P_{2q}$ are even Legendre polynomials (satisfying both Neumann
conditions at $\beta=0$ and $\beta=1$), and
$\displaystyle
g_{q}(\alpha)=\frac{1}{i}Q_{2q}(i\alpha)-\frac{\pi}{2}P_{2q}(i\alpha),$ (115)
where $i^{2}=-1$ and $Q_{2q}$ are Legendre functions of the second kind. Let
us give here additional details on the function $g_{q}$. To see that $g_{q}$
is real it is useful to write $Q_{2q}$ as [50]
$\displaystyle Q_{2q}(x)=\frac{P_{2q}(x)}{2}\ln\frac{1+x}{1-x}-W_{2q-1}(x),$
(116)
where $W$ is the polynomial
$\displaystyle
W_{2q-1}(x)=\sum_{m=1}^{q}\frac{4q-(1+4(m-1))}{(2m-1)(2q-m+1)}P_{2q-(2m-1)}(x).$
(117)
For purely imaginary arguments $x=i\alpha$, we have
$\displaystyle Q_{2q}(i\alpha)=iP_{2q}(i\alpha)\arctan(x)-W_{2q-1}(i\alpha),$
(118)
and we thus see that
$\displaystyle
g_{q}(\alpha)=P_{2q}(i\alpha)\arctan(\alpha)-\frac{W_{2q-1}(i\alpha)}{i}-\frac{\pi}{2}P_{2q}(i\alpha).$
(119)
Using the parity of $P$ and $W$, it becomes clear that $g_{q}$ is real.
Furthermore it can be checked that it decreases to zero at infinity (and
$g_{0}\sim 1/\alpha$ for large $\alpha$).
Now, the equation satisfied by the coefficients $a_{q}$ is identified by using
the Robin condition. In these coordinates, the partially absorbing disk
corresponds to $\alpha=0$, and the Robin conditions can be deduced from
$\partial_{n}\Phi=-\left(h_{\alpha}^{-1}\partial_{\alpha}\Phi\right)_{\alpha=0}$
so that the boundary conditions read
$\displaystyle(D\partial_{\alpha}\Phi-a\beta\kappa\Phi)_{\alpha=0}=0.$ (120)
Inserting the general solution (114) into the above boundary condition,
multiplying by $P_{2k}(\beta)$ and integrating, we obtain the linear system:
$\displaystyle\sum_{q=0}^{\infty}a_{q}g_{q}^{\prime}(0)\int_{0}^{1}d\beta
P_{2q}(\beta)P_{2k}(\beta)=\kappa\Phi_{\infty}\int_{0}^{1}d\beta\beta
P_{2k}(\beta)$
$\displaystyle+\kappa\sum_{q=0}^{\infty}a_{q}g_{q}(0)\int_{0}^{1}d\beta\beta
P_{2q}(\beta)P_{2k}(\beta),$ (121)
for all positive integers $k$. Finally, we calculate the surface element at
$\alpha=0$, $dS_{\alpha}=h_{\varphi}h_{\beta}d\varphi d\beta$ so that the
normalization condition reads
$\displaystyle 2\pi\kappa a^{2}\int_{0}^{1}\Phi(0,\beta)\beta d\beta=1,$ (122)
which leads to the equation
$\displaystyle\kappa a^{2}\pi\Phi_{\infty}+2\pi\kappa
a^{2}\sum_{q=0}^{\infty}a_{q}g_{q}(0)\int_{0}^{1}d\beta\beta P_{2q}(\beta)=1.$
(123)
A numerical solution for $\Phi_{\infty}$ can thus be found by solving the
linear system (121) for the coefficients $a_{q}$ (completed by the above
normalisation condition). Note also that $C_{\infty}=\Phi_{\infty}$.
## Appendix B Identification of the integral equation (27) in 3D
Here we briefly show how to obtain the integral equation (27). Using Eq. (20)
for $d=3$, we see that Eq. (24) writes
$\displaystyle\Phi(r_{0})=C_{\infty}-\frac{\kappa}{D}\int_{0}^{a}drK(r,r_{0})\Phi(r),$
(124)
with
$\displaystyle K(r,r_{0})$
$\displaystyle=\frac{1}{2\pi}\int_{0}^{2\pi}\frac{r\
d\theta}{\sqrt{r^{2}+r_{0}^{2}-2rr_{0}\cos\theta}}.$ (125)
The quantity $K(r,r_{0})$ can be simplified as follows:
$\displaystyle K(r,r_{0})$ $\displaystyle=\frac{1}{\pi}\int_{0}^{\pi}\frac{r\
d\theta}{\sqrt{r^{2}+r_{0}^{2}-2rr_{0}\cos\theta}}$
$\displaystyle=\frac{2r}{\pi}\int_{0}^{\pi/2}\frac{du}{\sqrt{r^{2}+r_{0}^{2}-2rr_{0}\cos(2u)}}$
$\displaystyle=\frac{2r}{\pi}\int_{0}^{\pi/2}\frac{du}{\sqrt{r^{2}+r_{0}^{2}-2rr_{0}[2\cos^{2}u-1]}}$
$\displaystyle=\frac{2r}{\pi(r+r_{0})}\int_{0}^{\pi/2}\frac{du}{\sqrt{1-\frac{4rr_{0}}{(r+r_{0})^{2}}\cos^{2}u}}$
$\displaystyle=\frac{2r}{\pi(r+r_{0})}K\left(\frac{2\sqrt{rr_{0}}}{r+r_{0}}\right),$
(126)
where $u=\theta/2$ and in the last line we have recognized the definition of
the elliptic function $K$. Inserting the above result into Eq. (124) finally
leads to Eq. (27).
## Appendix C Calculation details for the identification of $C_{2}$ and
$C_{2}^{*}$
### C.1 2D situation
Here we describe the details of calculations leading to the identification of
the constants $C_{2}$ and $C_{2}^{*}$ in the 2D situation. Let us first
evaluate the term $B$ in Eq. (71):
$\displaystyle
B\simeq-\frac{1}{\sqrt{\kappa}}\int_{0}^{\ell\kappa}dX\left(\psi(X)-\frac{1}{\pi\sqrt{2X}}\right)\ln(1-x_{0})$
$\displaystyle-\int_{0}^{\ell\kappa}\frac{dX}{\sqrt{\kappa}}\left(\psi(X)-\frac{1}{\pi\sqrt{2X}}\right)\ln\frac{1-x_{0}-X/\kappa}{1-x_{0}}.$
(127)
Here we have only assumed that one can use the leading order of the scaling
form for the solution $\Phi$ for arguments lower than $\ell$. Now, to evaluate
the first line we use the normalisation condition (38), and to evaluate the
terms on the second line, we change variable $X=u\kappa\ell$:
$\displaystyle
B\simeq\frac{1}{\sqrt{\kappa}}\int_{\ell\kappa}^{\infty}dX\left(\psi(X)-\frac{1}{\pi\sqrt{2X}}\right)\ln(1-x_{0})$
$\displaystyle-\sqrt{\kappa}\ell\int_{0}^{1}du\left(\psi(\kappa
u\ell)-\frac{1}{\pi\sqrt{2\kappa
u\ell}}\right)\ln\frac{1-x_{0}-u\ell}{1-x_{0}}.$ (128)
Using Eq. (63), we can write $B$ under the form
$\displaystyle B$
$\displaystyle\simeq\frac{f_{2}^{*}(x_{0})\ln\kappa+f_{2}(x_{0})}{\kappa},$
(129)
with
$\displaystyle f_{2}^{*}(x_{0})$
$\displaystyle=\frac{\ln(1-x_{0})}{\sqrt{\ell}\sqrt{2}\pi^{2}}-\int_{0}^{1}\frac{du\
\ell}{\pi^{2}(2u\ell)^{\frac{3}{2}}}\ln\frac{1-x_{0}-u\ell}{1-x_{0}},$ (130)
$\displaystyle f_{2}(x_{0})$
$\displaystyle=\frac{\ln(4\ell)+\gamma_{e}+1}{\sqrt{2\ell}\pi^{2}}\ln(1-x_{0})$
$\displaystyle-\ell\int_{0}^{1}du\frac{\ln(4u\ell)+\gamma_{e}-1}{\pi^{2}(2u\ell)^{3/2}}\ln\frac{1-x_{0}-u\ell}{1-x_{0}}.$
(131)
Let us precise a few properties of $f_{2}^{*}$ (similar properties hold for
$f_{2}$). In the limit $\ell\to 0$ at fixed $x_{0}$, we see that
$\displaystyle f_{2}^{*}(x_{0})$ $\displaystyle\underset{\ell\to
0}{\sim}\frac{\ln(1-x_{0})}{\sqrt{\ell}\sqrt{2}\pi^{2}}$ (132)
Near the extremity of the patch, we set $x_{0}=1-v\ell$ to determine the
behavior of $f_{2}^{*}$. In the limit $\ell\to 0$ at fixed $v=(1-x_{0})/\ell$,
we obtain
$\displaystyle f_{2}^{*}(x_{0}=1-v\ell)$ $\displaystyle\underset{\ell\to
0}{\sim}\frac{\ln(v\ell)}{\sqrt{\ell}\sqrt{2}\pi^{2}}-\int_{0}^{1}\frac{du\
\ell}{\pi^{2}(2u\ell)^{\frac{3}{2}}}\ln\frac{v-u}{v}$ (133)
Collecting the terms $O(\ln\kappa/\kappa)$ in the integral equation (70) leads
to
$\displaystyle\int_{-1+\ell}^{1-\ell}dx\
\Phi_{2}^{*}(x)\ln|x-x_{0}|=F_{2}^{*}(x_{0}),$ (134) $\displaystyle
F_{2}^{*}(x_{0})=-\pi C_{2}^{*}+f_{2}^{*}(x_{0})+f_{2}^{*}(-x_{0}).$ (135)
We consider this equation as an integral equation over the interval
$[-1+\ell;1-\ell]$, for which the solution is analytically known [45]:
$\displaystyle\Phi_{2}^{*}(x)=\frac{1}{\pi^{2}\sqrt{b^{2}-x^{2}}}\Bigg{[}\int_{-b}^{b}dt\frac{\sqrt{b^{2}-t^{2}}\partial_{t}F_{2}^{*}(t)}{t-x}$
$\displaystyle+\frac{1}{\ln[b/2]}\int_{-b}^{b}dt\frac{F_{2}^{*}(t)}{\sqrt{b^{2}-t^{2}}}\Bigg{]},$
(136)
where we have set $b=1-\ell$. As a consequence, the integral of $\Phi_{2}^{*}$
over the truncated interval $]-b;b[$ reads
$\displaystyle I_{2}^{*}(\ell)$ $\displaystyle=\int_{-b}^{b}dx\
\Phi_{2}^{*}(x)=\frac{1}{\pi\ln[b/2]}\int_{-b}^{b}\frac{dx\
F_{2}^{*}(x)}{\sqrt{b^{2}-x^{2}}}.$ (137)
When $\ell\to 0$, the behavior of $I_{2}^{*}(\ell)$ is obtained by inserting
the small $\ell$ limit of $f_{2}^{*}(x_{0})$ at fixed $x_{0}$ given by Eq.
(132) into Eq. (135) and inserting the result into the above equation, leading
to
$\displaystyle I_{2}^{*}(\ell)\underset{\ell\to 0}{\simeq}\frac{(-1)}{\pi\ln
2}\int_{-1}^{1}\frac{dt\
\ln\left(1-t^{2}\right)}{\sqrt{2}\pi^{2}\sqrt{\ell}\sqrt{1-t^{2}}}=\frac{\sqrt{2}}{\pi^{2}\sqrt{\ell}},$
(138)
this result is consistent with the fact that the behavior of $\Phi_{2}^{*}$
near $x=\pm 1$ is given by (67). Now, the fact that the normalisation
condition (26) holds at all powers of $\kappa$ implies that
$\displaystyle\lim_{\ell\to
0}\left[I_{2}^{*}(\ell)-\frac{\sqrt{2}}{\pi^{2}\sqrt{\ell}}\right]=0.$ (139)
We thus evaluate
$\displaystyle\Delta
I_{2}^{*}(\ell)=I_{2}^{*}(\ell)-\frac{\sqrt{2}}{\pi^{2}\sqrt{\ell}}$
$\displaystyle\simeq\frac{(-1)}{\pi\ln
2}\int_{-1}^{1}dt\left[\frac{F_{2}^{*}(t)\theta(b-|t|)}{\sqrt{b^{2}-t^{2}}}-\frac{\
\ln\left(1-t^{2}\right)}{\pi^{2}\sqrt{2\ell(1-t^{2})}}\right].$ (140)
The contributions of $f_{2}^{*}(t)$ in this integral are negligible except for
$x_{0}$ at the vicinity of $1$. Thus, we set $t=1-v\ell$ to evaluate the above
integral, and the integral can be evaluated by integrating $v$ over
$[0,\infty]$ (except for the term $C_{2}^{*}$ which is evaluated without this
change of variable). Using (133) yields
$\displaystyle\Delta I_{2}^{*}(\ell)\simeq\frac{(-1)}{\pi\ln
2}\Bigg{[}\int_{-1}^{1}dt\frac{-\pi C_{2}^{*}}{\sqrt{1-t^{2}}}$
$\displaystyle+\frac{2}{\pi^{2}\sqrt{2}}\int_{0}^{\infty}dv\ln(2v\ell)\left(\frac{\theta(v-1)}{\sqrt{2(v-1)}}-\frac{1}{\sqrt{2v}}\right)$
$\displaystyle-\int_{1}^{\infty}dv\int_{0}^{1}d{u}\frac{2}{\pi^{2}(2{u})^{3/2}\sqrt{2(v-1)}}\ln\frac{v-{u}}{v}.\Bigg{]}$
(141)
All the integrals appearing in the above equations can be evaluated. Imposing
$\Delta I_{2}^{*}(\ell)=0$ then leads to
$\displaystyle C_{2}^{*}=1/\pi^{2}.$ (142)
To identify $C_{2}$ we proceed in the same way. The integral equation is
$\displaystyle\int_{-1+\ell}^{1-\ell}dx\ \Phi_{2}(x)\ln|x-x_{0}|=F_{2}(x),$
(143) $\displaystyle F_{2}(x)=-\pi
C_{2}+\pi\phi_{1}+f_{2}(x_{0})+f_{2}(-x_{0}),$ (144)
so that
$\displaystyle I_{2}(\ell)=\int_{-b}^{b}dx\
\Phi_{2}(x)=\frac{1}{\pi\ln(b/2)}\int_{-b}^{b}\frac{dt\
F_{2}(t)}{\sqrt{b^{2}-t^{2}}}.$ (145)
As before we need to evaluate the behavior of $F_{2}(x,\ell)$ for small $\ell$
$\displaystyle F_{2}(x)\underset{\ell\to
0}{\sim}\frac{\ln(4\ell)+\gamma+1}{\sqrt{2\ell}\pi^{2}}\ln(1-x_{0}^{2})=\frac{F_{2}^{0}(x)}{\sqrt{\ell}}.$
(146)
At leading order for small $\ell$ we obtain
$\displaystyle I_{2}(\ell)\underset{\ell\to
0}{\sim}\frac{1}{\pi\ln(1/2)}\int_{-1}^{1}dt\frac{F_{2}^{0}(t)}{\sqrt{\ell}\sqrt{1-t^{2}}}=\frac{I_{2}^{0}}{\sqrt{\ell}}.$
(147)
The next-to-leading order is
$\displaystyle I_{2}-\frac{I_{2}^{0}}{\sqrt{\ell}}\simeq\frac{(-1)}{\pi\ln
2}\int_{-1}^{1}dt\left[\frac{F_{2}(t)\theta(b-|t|)}{\sqrt{b^{2}-t^{2}}}-\frac{F_{2}^{0}(t)}{\sqrt{\ell(1-t^{2})}}\right]$
(148)
Again, we evaluate it by setting $t=1-v\ell$ and taking the small $\ell$ limit
of the obtained integrand at fixed $v$, leading to
$\displaystyle I_{2}(\ell)-I_{2}^{0}(\ell)\simeq\frac{(-1)}{\pi\ln
2}\Bigg{\\{}\int_{-b}^{b}dx\frac{-\pi C_{2}+\Phi_{1}(x)}{\sqrt{b^{2}-x^{2}}}$
$\displaystyle+2\int_{0}^{\infty}dv\frac{\ln(4\ell)+\gamma+1}{\sqrt{2}\pi^{2}}\ln(2v\ell)\left[\frac{\theta(v-1)}{\sqrt{2(v-1)}}-\frac{1}{\sqrt{2v}}\right]$
$\displaystyle-2\int_{1}^{\infty}dv\int_{0}^{\infty}du\frac{\ln(4{u}\ell)+\gamma_{e}-1}{\pi^{2}(2{u})^{3/2}\sqrt{2(v-1)}}\ln\frac{v-{u}}{v}\Bigg{\\}}$
(149)
To evaluate the term involving $\Phi_{1}$ we introduce a variable
$\varepsilon$ so that $\ell\ll\varepsilon\ll 1$ and we calculate
$\displaystyle\int_{-b}^{b}dx$
$\displaystyle\frac{\Phi_{1}(x)}{\sqrt{b^{2}-x^{2}}}=\int_{-b}^{b}\frac{dx}{\pi\sqrt{1-x^{2}}\sqrt{b^{2}-x^{2}}}$
$\displaystyle\simeq\int_{-1+\varepsilon}^{1-\varepsilon}\frac{dx}{\pi(1-x^{2})}+\int_{1}^{\varepsilon/\ell}\frac{2\ell\
dv}{2\pi\sqrt{v(v-1}}$ $\displaystyle\simeq\frac{\ln(8/\ell)}{\pi},$ (150)
where the last equality follows from the evaluation of the integrals with
$\ell\ll\varepsilon\ll 1$. Finally, evaluating all terms in Eq. (149) and
requiring that the $I_{2}(\ell)-I_{2}^{0}(\ell)\to 0$ for small $\ell$, we
obtain
$\displaystyle C_{2}=\frac{1+\gamma_{e}+\ln 8}{\pi^{2}},$ (151)
as announced in the main text.
### C.2 Second-order calculation in the limit of large reactivity in 3D
We evaluate the term $B$ in Eq. (80) by writing
$\displaystyle
B(r_{0},\ell)\simeq\sqrt{\kappa}\frac{K\left(\frac{2\sqrt{r_{0}}}{1+r_{0}}\right)}{(1+r_{0})}\int_{\ell\kappa}^{\infty}dX\left[\psi(X)-\frac{1}{2\pi\sqrt{2X}}\right]$
$\displaystyle-\ell\int_{0}^{1}\frac{du}{\sqrt{\kappa}(1+r_{0})}\left[\psi(u\ell\kappa)-\frac{1}{2\pi\sqrt{2u\ell\kappa}}\right]$
$\displaystyle\times\left[K\left(\frac{2\sqrt{(1-u\ell)r_{0}}}{1-u\ell+r_{0}}\right)-K\left(\frac{2\sqrt{r_{0}}}{1+r_{0}}\right)\right].$
(152)
Hence
$\displaystyle
B(r_{0},\ell)=\frac{K\left(\frac{2\sqrt{r_{0}}}{1+r_{0}}\right)}{(1+r_{0})}\frac{1+\gamma_{e}+\ln(4\ell\kappa)}{2\sqrt{2\ell}\pi^{2}}$
$\displaystyle-\ell\int_{0}^{1}du\frac{[-1+\gamma_{e}+\ln(4u\ell\kappa)]}{2\pi^{2}(2u\ell)^{3/2}(1+r_{0})}$
$\displaystyle\times\left[K\left(\frac{2\sqrt{(1-u\ell)r_{0}}}{1-u\ell+r_{0}}\right)-K\left(\frac{2\sqrt{r_{0}}}{1+r_{0}}\right)\right].$
(153)
Note that here, for conciseness we will treat $\ln\kappa$ as being of order
$1$ in powers of $\kappa$, the result will be exactly the same as in the case
where one separates the $\ln\kappa$ terms and the $O(1)$ terms.
In the small $\ell$ limit at fixed $r_{0}$ we obtain
$\displaystyle B(r_{0},\ell)\underset{\ell\to
0}{\sim}\frac{K\left(\frac{2\sqrt{r_{0}}}{1+r_{0}}\right)}{1+r_{0}}\frac{1+\gamma_{e}+\ln(4\ell\kappa)}{2\sqrt{2\ell}\pi^{2}}=\frac{B^{0}(r_{0})}{\sqrt{\ell}},$
(154)
whereas if we set $r_{0}=1-\ell v$, in the limit $\ell\to 0$ at fixed $v$ we
obtain
$\displaystyle B(1-v\ell,\ell)\underset{\ell\to
0}{\sim}\frac{1}{8}\frac{1+\gamma_{e}+\ln(4\ell\kappa)}{\sqrt{2\ell}\pi^{2}}\ln\frac{8^{2}}{(v\ell)^{2}}$
$\displaystyle-\int_{0}^{1}d{u}\
\frac{[-1+\gamma_{e}+\ln(4\ell\kappa{u})]}{4\pi^{2}(2{u})^{3/2}}\ln\frac{v}{v-{u}},$
(155)
where we have used $K(1-y)\simeq\frac{1}{2}\ln(8/y)$ for small $y$. Let us
write the integral equation (79) under the form
$\displaystyle\int_{0}^{1-\ell}\ \frac{dr\ r\
\tilde{\Phi}_{2}(r)}{r+r_{0}}K\left(\frac{2\sqrt{rr_{0}}}{r+r_{0}}\right)=\frac{\pi[\tilde{C}_{2}-\Phi_{1}]}{2}+B$
(156)
with $\tilde{\Phi}_{2}=\Phi_{2}(r)+\Phi_{2}^{*}(r)\ln\kappa$,
$\tilde{C}_{2}=C_{2}^{*}\ln\kappa+C_{2}$. Let us define
$\displaystyle I_{2}(r,\ell)=\int_{0}^{1-\ell}dr\ r\ \tilde{\Phi}_{2}(r).$
(157)
Using the analytically known solution [45] of the integral equation (156), we
obtain
$\displaystyle
I_{2}(\ell)=\frac{4}{\pi^{2}}\int_{0}^{1-\ell}ds\frac{s\left(\frac{\pi}{2}(C_{2}-\Phi_{1})+B(s,\ell)\right)}{\sqrt{(1-\ell)^{2}-s^{2}}}.$
(158)
When $\ell\to 0$ we obtain at leading order
$\displaystyle I_{2}(r,\ell)\underset{\ell\to
0}{\sim}\frac{4}{\pi^{2}\sqrt{\ell}}\int_{0}^{1}ds\frac{s}{\sqrt{1-s^{2}}}B^{0}(s,\ell)$
(159)
and this integral diverges for $\ell\to 0$, as it should due to the known
behavior for $\tilde{\Phi}_{2}(r)$ when $r$ approaches $1$. At next-to-leading
order, we evaluate the terms involving $B$ by setting $s=1-v\ell$ and take the
small $\ell$ limit at fixed $v$, so that we can use Eq. (155):
$\displaystyle I_{2}(\ell)-\frac{4}{\pi^{2}\sqrt{\ell}}\int_{0}^{1}\frac{ds\
sB^{0}}{\sqrt{1-s^{2}}}\simeq\frac{4}{\pi^{2}}\Bigg{\\{}\int_{0}^{1}\frac{ds\
s}{\sqrt{1-s^{2}}}\frac{\pi}{2}C_{2}$
$\displaystyle-\int_{1}^{\infty}dv\int_{0}^{1}d{u}\
\frac{[-1+\gamma_{e}+\ln(4\ell\kappa{u})]}{4\pi^{2}(2u)^{3/2}\sqrt{2(v-1)}}\ln\frac{v}{v-{u}}$
$\displaystyle+\int_{0}^{\infty}dv\frac{1+\gamma_{e}+\ln(4\ell\kappa)}{4\sqrt{2}\pi^{2}}\ln\frac{8}{v\ell}\left(\frac{\theta(v-1)}{\sqrt{2(v-1)}}-\frac{1}{\sqrt{2v}}\right)$
$\displaystyle+\int_{0}^{1-\ell}ds\frac{s\pi\Phi_{1}(s)}{2\sqrt{(1-\ell)^{2}-s^{2}}}\Bigg{\\}}.$
(160)
To evaluate the term containing $\Phi_{1}$, defined in Eq. (76) we can use
again a trick where we use a variable $\varepsilon$ with
$\ell\ll\varepsilon\ll 1$:
$\displaystyle\int_{0}^{1-\ell}\frac{dr\ r\
\Phi_{1}(r)}{\sqrt{(1-\ell)^{2}-r^{2}}}=\int_{0}^{1-\varepsilon}dr\frac{r\Phi_{1}(r)}{\sqrt{1-r^{2}}}$
$\displaystyle+\int_{1}^{\varepsilon/\ell}\frac{dv\
\ell}{4\pi\sqrt{v(v-1}}=\frac{\ln(2/\ell)}{4\pi}.$ (161)
Finally, all the integrals in (160) can be evaluated, requiring that it
vanishes for small $\ell$ leads to
$\displaystyle\tilde{C}_{2}=\frac{\gamma_{e}+1+\ln(2\kappa)}{4\pi},$ (162)
which is exactly Eq. (81).
## References
* Redner [2001] S. Redner, _A guide to First- Passage Processes_ (Cambridge University Press, Cambridge, England, 2001).
* Condamin et al. [2007] S. Condamin, O. Bénichou, V. Tejedor, R. Voituriez, and J. Klafter, Nature 450, 77 (2007).
* Pal and Reuveni [2017] A. Pal and S. Reuveni, Phys. Rev. Lett. 118, 030603 (2017).
* Grebenkov [2016] D. S. Grebenkov, Phys. Rev. Lett. 117, 260201 (2016).
* Bénichou et al. [2010] O. Bénichou, D. Grebenkov, P. Levitz, C. Loverdo, and R. Voituriez, Phys. Rev. Lett. 105, 150606 (2010).
* Vaccario et al. [2015] G. Vaccario, C. Antoine, and J. Talbot, Phys. Rev. Lett. 115, 240601 (2015).
* Metzler et al. [2014] R. Metzler, S. Redner, and G. Oshanin, _First-passage phenomena and their applications_ (World Scientific, 2014).
* Schuss et al. [2007] Z. Schuss, A. Singer, and D. Holcman, Proc. Natl. Acad. Sci. U. S. A. 104, 16098 (2007).
* Newby and Allard [2016] J. Newby and J. Allard, Phys. Rev. Lett. 116, 128101 (2016).
* Rice [1985] S. Rice, _Diffusion-Limited Reactions_ (Elsevier, 1985).
* Berg and von Hippel [1985] O. G. Berg and P. H. von Hippel, Annu. Rev. Biophys. Biophys. Chem. 14, 131 (1985).
* Berg and Purcell [1977] H. C. Berg and E. M. Purcell, Biophys. J. 20, 193 (1977).
* Shoup and Szabo [1982] D. Shoup and A. Szabo, Biophys. J. 40, 33 (1982).
* Zhou and Zwanzig [1991] H.-X. Zhou and R. Zwanzig, J. Chem. Phys. 94, 6147 (1991).
* Reingruber and Holcman [2009] J. Reingruber and D. Holcman, Phys. Rev. Lett. 103, 148102 (2009).
* Bénichou et al. [2000] O. Bénichou, M. Moreau, and G. Oshanin, Phys. Rev. E 61, 3388 (2000).
* Grebenkov [2019] D. S. Grebenkov, Chemical Kinetics: Beyond the Textbook pp. 191–219 (2019).
* Collins and Kimball [1949] F. C. Collins and G. E. Kimball, J. Colloid Sci. 4, 425 (1949).
* Traytak and Price [2007] S. D. Traytak and W. S. Price, J. Chem. Phys. 127, 184508 (2007).
* Shoup et al. [1981] D. Shoup, G. Lipari, and A. Szabo, Biophys. J. 36, 697 (1981).
* Szabo et al. [1980] A. Szabo, K. Schulten, and Z. Schulten, J. Chem. Phys. 72, 4350 (1980).
* Isaacson et al. [2016] S. A. Isaacson, A. J. Mauro, and J. Newby, Phys. Rev. E 94, 042414 (2016).
* Isaacson and Newby [2013] S. A. Isaacson and J. Newby, Phys. Rev. E 88, 012820 (2013).
* Lindsay et al. [2017] A. E. Lindsay, A. J. Bernoff, and M. J. Ward, Multiscale Model. & Simul. 15, 74 (2017).
* Grebenkov [2022] D. S. Grebenkov, J. Stat. Mech. 2022, 083205 (2022).
* Grebenkov and Skvortsov [2022] D. S. Grebenkov and A. T. Skvortsov, Phys. Rev. E 105, 054107 (2022).
* Chaigneau and Grebenkov [2022] A. Chaigneau and D. S. Grebenkov, Phys. Rev. E 105, 054146 (2022).
* Grebenkov et al. [2018a] D. S. Grebenkov, R. Metzler, and G. Oshanin, Commun. Chem. 1, 96 (2018a).
* Bressloff [2022] P. C. Bressloff, Phys. Rev. E 105, 034141 (2022).
* Lindsay et al. [2015] A. Lindsay, T. Kolokolnikov, and J. Tzou, Phys. Rev. E 91, 032111 (2015).
* Guérin et al. [2021] T. Guérin, M. Dolgushev, O. Bénichou, and R. Voituriez, Commun. Chem. 4, 1 (2021).
* Singer et al. [2006a] A. Singer, Z. Schuss, D. Holcman, and R. Eisenberg, J. Stat. Phys. 122, 437 (2006a).
* Singer et al. [2006b] A. Singer, Z. Schuss, and D. Holcman, J. Stat. Phys. 122, 465 (2006b).
* Singer et al. [2006c] A. Singer, Z. Schuss, and D. Holcman, J. Stati. Phys. 122, 491 (2006c).
* Mangeat and Rieger [2021] M. Mangeat and H. Rieger, Phys. Rev. E 104, 044124 (2021).
* Bénichou and Voituriez [2008] O. Bénichou and R. Voituriez, Phys. Rev. Lett. 100, 168105 (2008).
* Grebenkov et al. [2018b] D. Grebenkov, R. Metzler, and G. Oshanin, Phys. Chem. Chem. Phys. (2018b).
* Grebenkov et al. [2017] D. S. Grebenkov, R. Metzler, and G. Oshanin, New J. Phys. 19, 103025 (2017).
* Grebenkov and Oshanin [2017] D. S. Grebenkov and G. Oshanin, Phys. Chem. Chem. Phys. 19, 2723 (2017).
* Grebenkov et al. [2019] D. S. Grebenkov, R. Metzler, and G. Oshanin, New J. Phys. 21, 122001 (2019).
* Jackson [1977] J. D. Jackson, _Classical electrodynamics_ , vol. 31999 (Wiley New York, 1977).
* Deegan et al. [1997] R. D. Deegan, O. Bakajin, T. F. Dupont, G. Huber, S. R. Nagel, and T. A. Witten, Nature 389, 827 (1997).
* Singer et al. [2008] A. Singer, Z. Schuss, A. Osipov, and D. Holcman, SIAM J. Appl. Math. 68, 844 (2008).
* Barton [1989] G. Barton, _Elements of Green’s functions and propagation_ (Clarendon Press, Oxford, 1989).
* Polyanin and Manzhirov [2008] A. D. Polyanin and A. V. Manzhirov, _Handbook of integral equations_ (CRC press, 2008).
* Brown [1977] S. N. Brown, J. Eng. Math. 11, 219 (1977).
* Boersma [1978] J. Boersma, J. Eng. Math. 12, 237 (1978).
* Kida and Take [1994] T. Kida and T. Take, JSME Int. J. Ser. B 37, 538 (1994).
* Bocquet [2007] L. Bocquet, Am. J. Phys. 75, 148 (2007).
* Abramowitz and Stegun [1964] M. Abramowitz and I. A. Stegun, _Handbook of Mathematical Functions With formulas, Graphs, and Mathematica Tables_ (1964).
|
# A kinematic calibration of the O-rich Mira variable period–age relation from
Gaia
Hanyuan Zhang and Jason L. Sanders
Department of Physics and Astronomy, University College London, London WC1E
6BT, UK
<EMAIL_ADDRESS><EMAIL_ADDRESS>
(Accepted XXX. Received YYY; in original form ZZZ)
###### Abstract
Empirical and theoretical studies have demonstrated that the periods of Mira
variable stars are related to their ages. This, together with their brightness
in the infrared, makes them powerful probes of the formation and evolution of
highly-extincted or distant parts of the Local Group. Here we utilise the Gaia
DR3 catalogue of long-period variable candidates to calibrate the period–age
relation of the Mira variables. Dynamical models are fitted to the O-rich Mira
variable population across the extended solar neighbourhood and then the
resulting solar neighbourhood period–kinematic relations are compared to
external calibrations of the age–kinematic relations to derive a Mira variable
period–age relation of $\tau\approx(6.9\pm
0.3)\,\mathrm{Gyr}(1+\tanh((330\,\mathrm{d}-P)/(400\pm 90)\mathrm{d})$. Our
results compare well with previous calibrations using smaller datasets as well
as the period–age properties of Local Group cluster members. This calibration
opens the possibility of accurately characterising the star formation and the
impact of different evolutionary processes throughout the Local Group.
###### keywords:
stars: variables: general – stars: AGB – Galaxy: disc – Galaxy: kinematics and
dynamics – Galaxy: evolution
††pubyear: 2023††pagerange: A kinematic calibration of the O-rich Mira
variable period–age relation from Gaia–References
## 1 Introduction
In the study of the formation and evolution of the Milky Way, one crucial
ingredient is accurate stellar ages (Freeman & Bland-Hawthorn, 2002; Bland-
Hawthorn & Gerhard, 2016). With this information, we can begin disentangling
the series of events that have led to the observed Milky Way today, as well as
directly measure the dynamical restructuring of the Galaxy. However, despite
their clear advantages in analysing the Galaxy, stellar ages are awkward
quantities due to their indirect measurement only via stellar models. Many
stellar age indicators exist (Soderblom, 2010) which often provide different
levels of accuracy for different stellar types and different stellar
populations. With the availability of Gaia astrometry (Gaia Collaboration et
al., 2016, 2018) and complementary large-scale spectroscopic surveys (e.g. De
Silva et al., 2015; Majewski et al., 2017), two methods applicable to large
collections of stars are comparisons to isochrone models (e.g. Xiang et al.,
2017; Sanders & Das, 2018; Xiang & Rix, 2022, which operates most successfully
for subgiant stars that have recently turned off the main sequence), and
indirect mass measurements of giant stars through spectroscopic measurements
of the products of dredge-up episodes calibrated via asteroseismology (e.g.
Masseron & Gilmore, 2015; Martig et al., 2016).
Mira variables are high-amplitude thermally-pulsing asymptotic giant branch
(AGB) stars. Their study in the Large Magellanic Cloud (e.g. Glass & Evans,
1981; Wood et al., 1999; Groenewegen, 2004) demonstrated that they follow a
tight period–luminosity relation (believed to be associated with fundamental
mode pulsation) making them interesting tracers both for local Galactic and
cosmological studies (Catchpole et al., 2016; Grady et al., 2019, 2020; Huang
et al., 2020). The chemistry of Mira variables is either oxygen or carbon-
dominated depending on the C/O ratio (Höfner & Olofsson, 2018), but O-rich
Mira variables are significantly more common in the Milky Way and are found to
follow tighter period–luminosity relations due potentially to less
circumstellar dust (Ita & Matsunaga, 2011). It has long been empirically known
that groups of Mira variables binned by period show correlations between
period and scaleheight/velocity dispersion (Merrill, 1923; Feast, 1963), which
is typically interpreted as a correlation between the period and age of a Mira
variable where the older stars have longer periods. This opens the possibility
of using Mira variables as age indicators within the Galaxy and beyond (e.g.
Grady et al., 2020). A limited number of Mira variables in clusters also
validate the period–age connection although confident assignment of membership
has only been possible recently with Gaia data (Grady et al., 2019; Marigo et
al., 2022). Although the period–age relation has been approximately calibrated
empirically (Feast & Whitelock, 2000b), relatively few theoretical models
reproducing the behaviour exist (Wyatt & Cahn, 1983; Eggen, 1998; Trabucchi &
Mowlavi, 2022) and the lack of detailed reproduction of the period–luminosity
relations of fundamental mode pulsation from theoretical models suggests the
period–age relations still have some associated uncertainty and there is a
need for accurate data-driven calibrations.
Encounters in the stellar discs of galaxies cause stellar populations to
slowly kinematically heat giving rise to age–velocity dispersion relations
(Wielen, 1977) such as those suggested for Mira variable stars (Feast, 1963).
There are multiple suggested perturbers that give rise to disc heating
including molecular clouds, spiral arms or merger events (Spitzer &
Schwarzschild, 1951, 1953; Barbanis & Woltjer, 1967; Velazquez & White, 1999;
Hänninen & Flynn, 2002; Aumer et al., 2016) that likely have differing
relative contributions across the Galactic disc (Mackereth et al., 2019). In
the solar neighbourhood, the stellar velocity dispersion is approximately a
power law in age with exponent $\sim 0.3$ for the radial dispersion and $\sim
0.5$ for the vertical dispersion (Holmberg et al., 2009; Aumer & Binney, 2009;
Sharma et al., 2021). A common picture (Binney & Tremaine, 2008) for this
behaviour is that the spiral arms are efficient in-plane heating sources
giving rise to the increase in radial velocity dispersion and molecular clouds
are efficient in converting this radial energy into vertical energy (Aumer et
al., 2016). There is the further complication that the stellar populations
could have been born hotter in the past, which could play a part in the
observed correlations (Bird et al., 2021). Now with Gaia data, the
age–velocity dispersion relations can be inspected across the Galactic disc
(Sanders & Das, 2018; Mackereth et al., 2019; Sharma et al., 2021; Gaia
Collaboration et al., 2021). For our purposes, the fact that correlations
between age and kinematics exist is sufficient and we need not necessarily
understand the underlying cause. In this way, kinematics can be used as an age
proxy for groups of stars. Note that for this procedure to operate well, we
are perhaps implicitly assuming that the kinematic–age relations are monotonic
as evidenced in the solar neighbourhood (e.g. Holmberg et al., 2009).
With the publication of large catalogues of variable stars from Gaia with
associated proper motions (Eyer et al., 2022), there is now the possibility of
thorough characterizations of the dynamical properties of different families
of Mira variable stars (Alvarez et al., 1997). Kinematic characterization then
opens up the possibility of mutual age calibration of different age tracers.
By assuming kinematics are solely a function of age, we can anchor different
age indicators to each other by requiring they all reproduce the same
age–kinematic relations (e.g Angus et al., 2015; Angus et al., 2020). In this
way, we can characterise the Mira variable period–age relation. This
simplifying assumption can be complicated by metallicity dependence,
particularly if different tracers are biased toward different metallicity
populations. The Mira variable stage occurs in stars of all metallicities
although C-rich Mira variables are only formed through dredge-up in young,
metal-poor stars (Boyer et al., 2013). This strategy of mutual age calibration
via age–kinematic relations has been utilised successfully in the study of
gyrochronology (Angus et al., 2015) and chromospheric activity in late-type
stars (Wilson & Woolley, 1970; West et al., 2015), and promises a route to the
mutual calibration of all stellar age indicators.
In this work, we utilise the astrometry of the latest Gaia DR3 long-period
variable candidate catalogue to characterise the kinematic behaviour of O-rich
Mira variables separated by period and combine this information with
literature age–velocity dispersion relations in the solar neighbourhood to
characterise the period–age relation for O-rich Mira variable stars. In
Section 2 we describe the dataset we use focusing on the cuts required to
isolate both O-rich AGB stars and those high-amplitude long-period variables
that are likely Mira variables. In Section 3 we describe our modelling
procedure and tests on mock data, before showing the results applied to data
in Section 4 and the resulting period–age relation in Section 5. We critically
discuss our approach and compare to other Mira variable period–age relations
in Section 6 before summarising our conclusions in Section 7.
## 2 The Gaia DR3 O-rich Mira variable sample
Figure 1: Colour–magnitude diagrams computed using a $3\sigma$-adjusted
parallax, $\varpi-3\sigma_{\varpi}$. We define the region occupied by AGB
stars as
$G-5\log_{10}(100/(\varpi-3\sigma_{\varpi}))<2.5(G_{\mathrm{BP}}-G_{\mathrm{RP}})-5$:
any star outside this is likely a YSO. The right panel shows those only those
stars with $\texttt{best\\_class\\_score}>0.8$ which effectively removes any
likely YSO contaminants. Figure 2: Properties of our O-rich Mira sample: the
top left panel shows the distribution of the Wesenheit index difference from
Lebzelter et al. (2018) used to separate O-rich and C-rich Mira. The lower
left panel shows the distribution of this quantity vs. period. The right two
panels show the period and distance error for the O-rich Mira sample.
Figure 3: The contour plot of the C-rich (black) and O-rich (red) Mira
variable population selected by their spectrum on period–amplitude plane and
period–colour plane respectively. Candidates below the blue line were removed
from the sample.
We begin by describing how we form our O-rich Mira variable sample. It is
important to note that our analysis relies on characterising the velocity
distributions at each Galactic location. In this way, considerations on the
completeness of our sample are unimportant provided we do not perform any
specific selections on the velocities of the stars. Our primary objective with
the selection is to form a low-contamination subset.
We use the long period variable (LPV) candidate catalogue from Gaia DR3
(Lebzelter et al., 2022). This catalogue has been constructed in a two-stage
process – likely variable stars are identified by comparison to literature
variable sources and reference non-variable Gaia sources, and then classified
based on literature classifications and features including the Lomb-Scargle
period, time summary statistics, colours and parallax (Holl et al., 2018;
Rimoldini et al., 2019, 2022). Stars classified as LPVs with $G$ $5$th$-95$th
percentile greater than $0.1\,\mathrm{mag}$ and
$G_{\mathrm{BP}}-G_{\mathrm{RP}}>0.5$ (along with other less important cuts
for our purposes) were further considered by the specific object study (SOS).
Candidate LPVs from the SOS have published generalised Lomb–Scargle periods
(and Fourier amplitudes) in Gaia DR3 if the period is greater than
$35\,\mathrm{day}$ and shorter than the $34$ month time series duration, the
$G$-band signal-to-noise $>15$ and there is no correlation between the image
determination parameters and the time series. Infrared photometric
measurements were acquired from the 2MASS catalogue (Skrutskie et al., 2006)
using the precomputed cross-match provided on the Gaia archive. There are
$1\,657\,987$ variable star observations in the Gaia DR3 LPV candidate SOS
catalogue after the cross-match with 2MASS. We first remove stars without
measured periods or without $J$ and $K_{s}$ photometric measurements which are
needed for later selection pipelines. These requirements reduce the size of
the sample to $387\,419$ objects.
To isolate a sample of likely Mira variables, we employ cuts in period and
magnitude. We retain stars with $80<\mathrm{Period}/\,\mathrm{day}<1000$
(Matsunaga et al., 2009) and in amplitude we employ a similar cut to Grady et
al. (2019), which removes stars with $\texttt{amplitude}<0.5\,\mathrm{mag}$
(compared to Grady et al. 2019 cut at $0.43\,\mathrm{mag}$). Here amplitude is
the $G$ semi-amplitude computed from a Fourier fit. Note that around the
problematic period of $190$ day, the Fourier fit can significantly
overestimate the amplitude of the LPVs leading to lower-amplitude semi-regular
variable contaminants in a Mira variable selection. We remove stars with
$170<\mathrm{Period(days)}<200$ and $\texttt{amplitude}>1.3$, and
$350<\mathrm{Period(days)}<400$ and $\texttt{amplitude}>1.6$ to mitigate
against this.
As highlighted by Mowlavi et al. (2018), young stellar objects (YSOs) can be a
contaminant in the LPV processing as they have similar colours, amplitudes and
periods to LPVs. In the classification pipeline from Holl et al. (2018) and
Rimoldini et al. (2019), the probability of the object being of the reported
class, best_class_score, seems an effective indicator of YSOs. In Fig. 1, we
show the colour–absolute magnitude diagram for our sample computed using a
parallax adjusted by $3$ times the parallax uncertainty. This gives the
brightest possible magnitude for each star within the parallax uncertainties
so any star consistent with being near the main sequence using this measure is
likely a YSO. Many of these objects also have
$\texttt{best\\_class\\_score}<0.8$ so we choose to only consider stars with
$\texttt{best\\_class\\_score}>0.8$. From this series of cuts, we end up with
$75\,874$ Mira variable star candidates.
### 2.1 O-rich/C-rich classification
LPVs can be either oxygen-rich or carbon-rich depending on the metallicity and
the strength of the dredge-ups which is controlled by the initial mass (Höfner
& Olofsson, 2018). The O-rich stars follow a tighter period-luminosity
relation (due to increased circumstellar dust in the C-rich stars, Ita &
Matsunaga, 2011) and are significantly more common in the Milky Way (with
C-rich stars contributing more in the outer disc, Blanco et al., 1984;
Ishihara et al., 2011). As shown by Lebzelter et al. (2022), the Gaia DR3
BP/RP (XP) spectra can be used to effectively separate O-rich and C-rich AGB
stars due to the differing set of band heads and features in their spectra
arising primarily from the TiO and CN absorption features. Sanders & Matsunaga
(submitted) have provided an unsupervised classification approach for these
spectra that effectively separates O-rich and C-rich LPV stars and performs
better than the Gaia DR3 classifications for highly-extincted sources. We
adopt their classifications where Gaia DR3 XP spectra are available. Lebzelter
et al. (2018) showed that, within the LMC, O-rich and C-rich Mira variables
can be separated in the plane of $W_{\mathrm{BPRP}}-W_{JK_{s}}$ vs. $K_{s}$.
Here the two Wesenheit indices are
$W_{\mathrm{BPRP}}=G_{\mathrm{RP}}-1.3(G_{\mathrm{BP}}-G_{\mathrm{RP}})$ and
$W_{JK_{s}}=K_{s}-0.686(J-K_{s})$. Although the boundary employed by Lebzelter
et al. (2018) is slightly curved, we can employ a very similar cut to select
O-rich Mira as $W_{\mathrm{BPRP}}-W_{JK_{s}}<1$. The left two panels of Fig. 2
show that this Wesenheit index difference against period for the selected Mira
sample, whilst the right panels are the period and distance percentage error
of the O-rich Mira after further selections. The performance and purpose of
these two cuts are very alike, but we employed both cuts here to maximally
remove C-rich Mira contamination.
Aided by the XP spectrum classifications, we have found that O-rich and C-rich
sources are separated in the period–amplitude plane and period–colour plane as
shown in Fig. 3. Hence, we make a further two cuts to remove those C-rich Mira
variables when an XP classification is not available:
$\texttt{amplitude}>1.2\log_{10}(\mathrm{Period}/\mathrm{days})-2.22$;
$G_{\mathrm{BP}}-G_{\mathrm{RP}}>7\log_{10}(\mathrm{Period}/\mathrm{days})-13.20$.
The resulting number of O-rich Mira variable candidates was $46\,107$.
### 2.2 Assigning distances
The distance modulus, $m$, of O-rich Mira stars are estimated from the
period–luminosity relation
$M_{KJK}=\left\\{\begin{array}[]{rcl}-7.53-4.05(\log_{10}P-2.3),&&\log_{10}P<2.6,\\\
-8.75-6.99(\log_{10}P-2.6),&&\log_{10}P\geq 2.6,\end{array}\right.$ (1)
where $P$ is the period in days and $M_{KJK}$ the absolute Wesenheit
magnitude, and the corresponding apparent Wesenheit magnitude $W_{KJK}$ is
$W_{KJK}=K_{s}-0.473(J-K_{s}).$ (2)
The extinction coefficient is taken from Wang & Chen (2019). This extinction
coefficient does not include the reddening caused by the circumstellar dust if
its properties are different from the interstellar dust. Instead, because the
period-luminosity relation is calibrated with respect to the O-rich Mira
variables in the LMC, the reddening from circumstellar dust has already been
considered in equation (1). The only caveat left is the potential difference
in properties of the circumstellar dust between O-rich Mira variables in the
LMC and the Milky Way possibly arising due to the difference in metallicity.
We consider this a minor effect in our analysis, particularly at shorter
periods where significant circumstellar dust is uncommon (Ita & Matsunaga,
2011).
The intrinsic scatter $\sigma$ of the period-luminosity relation is
$\sigma=\left\\{\begin{array}[]{rcl}\sigma_{23}+m_{\sigma_{1}}(\log_{10}P-2.3),&&\log_{10}P<2.6,\\\
\sigma_{23}+0.3m_{\sigma_{1}}+m_{\sigma_{2}}(\log_{10}P-2.6),&&\log_{10}P\geq
2.6,\end{array}\right.$ (3)
where $\ln\sigma_{23}=-1.47$, $m_{\sigma_{1}}=0.20$ and $m_{\sigma_{2}}=0.89$.
These relationships are taken from fits of the single-epoch 2MASS data for
Mira variables in the LMC (Sanders, in prep.). The scatter is a combination of
the single-epoch scatter and the intrinsic scatter due to variance in the
population. Whitelock et al. (2008) has argued from a comparison of LMC Mira
variables with local Mira variables with Hipparcos and VLBI parallaxes that
the Mira variable period-luminosity relation is metallicity-independent,
validating our use of the LMC relations for the Milky Way disc Mira variables.
Sanders (in prep.) has shown that the $W_{KJK}$ relations for the Milky Way
are quite similar to the LMC relations. To compute the uncertainties in
distance modulus, $\sigma_{m}$, we combine in quadrature the intrinsic scatter
of the period–luminosity relation from equation (3) with the uncertainty
propagated from the photometric and period measurement uncertainties. The
typical period uncertainties give rise to a median scatter of
$0.06\,\mathrm{mag}$ but the scatter arising from the single-epoch
measurements is $\gtrsim 0.22\,\mathrm{mag}$. Note that the period
uncertainties are only meaningful if the correct periodogram peak has been
identified. In the case of aliases, the reported period can be formally
inconsistent with the true period. Lebzelter et al. (2022) show the impact of
aliasing is low. Additionally, in our modelling, we allow for the possibility
of a star to be an ‘outlier’ which will capture any incorrectly assigned
periods.
### 2.3 Gaia astrometric data quality
LPV stars are one of the most challenging regimes for the Gaia astrometric
pipeline for a number of reasons. First, these sources are very red and Gaia’s
image parameter determination is not well characterised for sources redder
than $\nu_{\mathrm{eff}}=1.24\,\mu\mathrm{m}^{-1}$ (Rowell et al., 2021).
Secondly, LPVs are variable whilst the current Gaia astrometric pipelines
utilise a fixed colour in the modelling that could lead to systematics
(Pourbaix et al., 2003). Finally and possibly most importantly, LPVs can have
radii of $1\,\mathrm{AU}$ or larger, and in the optical the photocentres
wobble on the order of $\lesssim 10\,\mathrm{per\,cent}$ the radius of the
star Chiavassa et al. (2011); Chiavassa et al. (2018). This additional
photocentre wobble can lead to biases in the recovered astrometry (e.g.
Andriantsaralaza et al., 2022) but as the motion is somewhat random and
importantly not aligned in any special directions with respect to the
parallactic and proper motion directions, particularly when averaging over
many stars, the predominant effect is that the reported astrometric
uncertainties are underestimates of the true uncertainties.
Sanders (in prep.) has looked at the expected performance of Gaia on a set of
modelled Mira variable stars and found that the parallax uncertainties must be
inflated for higher parallax objects. This analysis agreed approximately with
a full characterisation of the period–luminosity relation and Gaia parallaxes
for the Mira variable stars for which Sanders (in prep.) measured an inflation
factor of $1+\exp[-(m-8.5)/0.8]$ for the parallax uncertainties. Here, we
assume that the proper motion uncertainties must be inflated by the same
factor (as validated by Sanders, in prep.). We do not consider the parallaxes
in this work.
In addition to the inflation of the astrometric uncertainties on purely
physical grounds, any mischaracterisation of Gaia’s performance gives rise to
misestimated astrometric uncertainties. Steps are taken to mitigate against
this in the Gaia pipeline (Lindegren et al., 2012) but several studies have
shown that problems likely still exist (e.g. El-Badry et al., 2021; Maíz
Apellániz, 2022). Again, this is particularly a concern for the redder sources
due to the image parameter determination. Sanders (in prep.) has modelled the
period–luminosity relation using the Gaia parallaxes including a flexible
model for the factor by which Gaia’s parallax errors must be inflated. The
model is two quadratics in $G$ and $\nu_{\mathrm{eff}}$ for the $5-$ or
$6-$parameter astrometric solutions respectively. We adopt their models for
the $W_{KJK}$ period–luminosity fits which typically require the parallax
uncertainties to be inflated by a factor $\sim 1.5$. Although the inflation
factor is appropriate for parallax errors, the astrometric modelling is a
linear regression so underestimates in the output parameters reflect
misestimates of the individual epoch astrometric (along-scan) measurements. It
is therefore appropriate to assume all the astrometric uncertainties must be
scaled in a similar way to the parallax uncertainties.
### 2.4 Final spatial cuts
We adopt a final series of spatial cuts to focus on Galactic disc members. We
remove stars with $270^{\circ}<\ell<290^{\circ}$, $-42^{\circ}<b<-22^{\circ}$,
and $40<\mathrm{heliocentric\ distance}\,(\mathrm{kpc})<60$ to remove
potential LMC candidates. As we only consider Mira variables from the Galactic
disc, we removed possible bar-bulge contribution by cutting stars with
$R<5\,\mathrm{kpc}$, where $R$ is the galactocentric radius. For the interest
of kinematic modelling, we only looked at stars with $\mathrm{heliocentric\
distance}<8\,\mathrm{kpc}$ and $R<10\,\mathrm{kpc}$. Stars with
$\sigma_{m}>0.6$ are removed to avoid stars with extremely large spatial
uncertainties. With all of the cuts described in this section, there remain
$8\,290$ O-rich Mira variable star candidates in the sample.
## 3 Kinematic modelling using dynamical models
Due primarily to the specifics of the scanning law, Gaia’s detection of
variable stars is a strong function of on-sky location and magnitude. This
makes fitting density, or full dynamical, models to any Gaia variable dataset
difficult without a careful characterisation of the selection function. Here
we employ a simpler approach by only considering the velocity,
$\boldsymbol{v}$, distribution of our sample at each observed Galactic
location, $\boldsymbol{x}$ i.e. $p(\boldsymbol{v}|\boldsymbol{x})$. Except in
the most extreme cases, a Mira variable star will not fail to be in the
catalogue as a result of its proper motion such that we can safely model the
conditional distribution of the proper motions given position. We opt to work
with full dynamical models $f(\boldsymbol{J})$ expressed as functions of the
actions $\boldsymbol{J}$ due to their ability to capture the detailed shapes
of the velocity distributions and their necessary linking of the radial and
azimuthal velocity profiles.
In Fig. 4, we plot the latitudinal velocity dispersion profile for several
period bins of the selected O-rich Mira as shown. A clear trend in
period–dispersion relation is seen implying that the O-rich Mira variables
follow a period–age relationship. In our modelling procedure, we will model
populations of stars in period bins. Note that the periods are uncertain (as
described in the previous section), but typically the uncertainty in the
period is small ($\sim 10\,\mathrm{day}$, except in the case of aliases) and
mixing between bins is a small effect. Working with binned data significantly
simplifies our procedure and allows us to fully explore the kinematics with
period rather than imposing some functional form. We discuss this latter
possibility later.
Figure 4: The transverse latitudinal velocity, $v_{b}$, dispersion profiles of
O-rich Mira separated into different period bins. Stars in this figure are
only from $|b|<5^{\circ}$, so $v_{b}$ is approximately equal to the Galactic
vertical velocity, or $v_{z}$, dispersion.
For a given population of stars with similar periods, we wish to fit the
probability distribution function $p(\boldsymbol{\mu}|\ell,b,m)$ where
$\boldsymbol{\mu}$ is the proper motion vector, $(\ell,b)$ the Galactic
coordinates and $m$ the distance modulus (as described in the previous
section). We begin by writing
$p(\boldsymbol{\mu}|\ell,b,m)=\frac{p(\ell,b,m,\boldsymbol{\mu})}{p(\ell,b,m)}=\frac{\int\mathrm{d}v_{||}p(\ell,b,m,\boldsymbol{\mu},v_{||})}{\int\mathrm{d}^{2}\boldsymbol{\mu}\,\mathrm{d}v_{||}\,p(\ell,b,m,\boldsymbol{\mu},v_{||})}.$
(4)
The proper motions and distance moduli are measured quantities with some
associated uncertainties characterised by the proper motion covariance matrix
$\boldsymbol{\Sigma}_{\mu}$ and the uncertainty in distance modulus
$\sigma_{m}$. We, therefore, marginalize over the uncertainties by writing
$\begin{split}p(\ell,&b,m,\boldsymbol{\mu},v_{||})\\\
&=\int\mathrm{d}^{2}\boldsymbol{\mu}^{\prime}\mathrm{d}m^{\prime}\mathcal{N}(\boldsymbol{\mu}|\boldsymbol{\mu}^{\prime},\boldsymbol{\Sigma}_{\mu})\mathcal{N}(m|m^{\prime},\sigma^{2}_{m})p(\ell,b,m^{\prime},\boldsymbol{\mu}^{\prime},v_{||}),\end{split}$
(5)
where $\mathcal{N}(x|\mu,\sigma^{2})$ are Gaussians with mean $\mu$ and
variance $\sigma^{2}$. We then relate the distribution in observable
coordinates to the dynamical distribution function in actions as
$p(\ell,b,m^{\prime},\boldsymbol{\mu}^{\prime},v_{||})=\left|\frac{\partial(\boldsymbol{J},\boldsymbol{\theta})}{\partial(\ell,b,m,\boldsymbol{\mu},v_{||})}\right|f(\boldsymbol{J})\propto
s^{5}\cos b\,f(\boldsymbol{J}),$ (6)
where $\boldsymbol{J}=(J_{r},J_{\phi},J_{z})$ is the set of actions
corresponding to the observed 6d coordinate (with corresponding angle
coordinates $\boldsymbol{\theta}$) and $s$ is the distance corresponding to
distance modulus $m$. Note the Jacobian between
$(\boldsymbol{x},\boldsymbol{v})$ and $(\boldsymbol{J},\boldsymbol{\theta})$
is unity due to the canonical nature of the action-angle coordinates.
We choose $f(\boldsymbol{J})$ as a quasi-isothermal distribution function,
which is suitable for warm discs (Binney, 2010). We follow the implementation
in Agama (Vasiliev, 2019) which has a functional form given by
$\displaystyle
f(\boldsymbol{J})=\frac{\tilde{\Sigma}\,\Omega}{2\pi^{2}\,\kappa^{2}}\times\frac{\kappa}{\tilde{\sigma}_{r}^{2}}\exp\left(-\frac{\kappa\,J_{r}}{\tilde{\sigma}_{r}^{2}}\right)\times\frac{\nu}{\tilde{\sigma}_{z}^{2}}\exp\left(-\frac{\nu\,J_{z}}{\tilde{\sigma}_{z}^{2}}\right)\times
B(J_{\phi}),$ $\displaystyle B(J_{\phi})=\left\\{\begin{array}[]{ll}1&\mbox{if
}J_{\phi}\geq 0,\\\
\exp\left(\frac{2\Omega\,J_{\phi}}{\tilde{\sigma}_{r}^{2}}\right)&\mbox{if
}J_{\phi}<0,\end{array}\right.,$ (9)
$\displaystyle\tilde{\Sigma}(R_{\mathrm{c}})\equiv\Sigma_{0}\exp(-R_{\mathrm{c}}/R_{\mathrm{disc}}),$
$\displaystyle\tilde{\sigma}_{r}^{2}(R_{\mathrm{c}})\equiv\sigma_{r,0}^{2}\exp(-2(R_{\mathrm{c}}-R_{0})/R_{\sigma,r}),$
$\displaystyle\tilde{\sigma}_{z}^{2}(R_{c})\equiv\sigma_{z,0}^{2}\exp(-2(R_{c}-R_{0})/R_{\sigma,z}),$
(10)
where $R_{\mathrm{c}}$ is the radius corresponding to a circular orbit of
angular momentum $J_{\phi}\equiv L_{z}$ and $(\kappa,\Omega,\nu)$ are the
epicyclic frequencies at this angular momentum. This distribution function
describes an approximately exponential disc in radius which is
broadened/warmed vertically and radially by two exponential terms. There are
five key free parameters for the model: (i) the scalelength of the disc,
$R_{\mathrm{disc}}$, (ii) the radial ($\sigma_{r,0}$) and vertical
($\sigma_{z,0}$) normalizations of the velocity dispersions at the Sun
($R=R_{0}$), and (iii) their corresponding scalelengths ($R_{\sigma,r}$ and
$R_{\sigma,z}$). The actions are evaluated using the ‘Stäckel fudge’ algorithm
described by Binney (2012), summarized and critically assessed against
alternatives in Sanders & Binney (2016) and implemented in Agama (Vasiliev,
2019). We adopt a fixed axisymmetric gravitational potential for the Galaxy
from McMillan (2017). Fixing the potential could lead to sub-optimal model
fits (as we will discuss later) but it significantly simplifies the
computation and incorporates external constraints from the analysis of other
datasets.
### 3.1 Computational considerations
The computational difficulty in evaluating equation (4) is computing the
integrals efficiently. Here we use Monte Carlo integration. For the numerator,
we generate a set of $N$ samples for each star from the proper motion and
distance modulus error ellipses. The unknown $v_{||}$ is sampled from a
probability distribution $G(v_{||}|\ell,b,m,\boldsymbol{\mu})$ which is
proportional to a quasi-isothermal distribution function with fixed parameters
$f^{\prime}(\boldsymbol{J})$ at a given $(\ell,b,m,\boldsymbol{\mu})$,
$G(v_{||}|\ell,b,m,\boldsymbol{\mu})=\frac{p(\ell,b,m,\boldsymbol{\mu},v_{||})}{\int\mathrm{d}v_{||}\,p(\ell,b,m,\boldsymbol{\mu},v_{||})}=A_{v_{||}}f^{\prime}(\boldsymbol{J}).$
(11)
Samples are generated from this distribution using the inverse cumulative
distribution. The value of $f^{\prime}(\boldsymbol{J}_{i})$ for each sample is
stored to reweight the Monte Carlo sum. For the denominator, we sample
$\boldsymbol{v}=(v_{x},v_{y},v_{z})$ directly at a given observed position
$(\ell,b,m)$ in a similar way to the numerator as
$G(\boldsymbol{v}|\ell,b,m)=\frac{p(\ell,b,m,\boldsymbol{v})}{\int\mathrm{d^{3}}\boldsymbol{v}\,p(\ell,b,m,\boldsymbol{v})}=A_{\boldsymbol{v}}f^{\prime}(\boldsymbol{J}),$
(12)
from which samples are generated using Markov Chain Monte Carlo (MCMC,
Foreman-Mackey et al., 2013), and once again $f^{\prime}(\boldsymbol{J}_{i})$
are stored. $A_{v_{||}}$ and $A_{\boldsymbol{v}}$ defined in equation (11) and
(12) are constant factors which can be computed for each individual star. Only
the ratio of these two normalisation factors is important:
$A\equiv\frac{A_{\boldsymbol{v}}}{A_{v_{||}}}=\frac{\int\mathrm{d}v_{||}\,p(\ell,b,m,\boldsymbol{\mu},v_{||})}{\int\mathrm{d^{3}}\boldsymbol{v}\,p(\ell,b,m,\boldsymbol{v})}=\frac{\int\mathrm{d}v_{||}\,f^{\prime}(\boldsymbol{J})}{\int\mathrm{d^{3}}\boldsymbol{v}\,f^{\prime}(\boldsymbol{J})}.$
(13)
$A$ is evaluated using Monte Carlo integration: $v_{||}$ and $\boldsymbol{v}$
are sampled from a Gaussian distribution centred on zero in the radial and
vertical velocities, and on the rotation curve in the azimuthal velocity. As
$f^{\prime}(\boldsymbol{J})$ is fixed, $A$ can be precomputed once for each
individual star to a desired accuracy. The $f^{\prime}(\boldsymbol{J})$ we use
throughout this paper has fixed parameters:
$R_{\mathrm{disc}}=2.5\,\mathrm{kpc}$, $\sigma_{r,0}=50\,\mathrm{km/s}$,
$\sigma_{z,0}=50\,\mathrm{km/s}$, $R_{\sigma,r}=5.0\,\mathrm{kpc}$ and
$R_{\sigma,z}=5.0\,\mathrm{kpc}$. These parameters are chosen such that the
distributions of the integration samples are typically broader than the
modelled distributions to minimise bias in the Monte Carlo integration.
Sampling from the distribution $G$, instead of a Gaussian distribution
increases the computational efficiency by reducing the noise in the Monte
Carlo integration for a fixed number of sampling. Now for each star, the
integrals (up to a normalization constant) are given by
$p(\ell,b,m,\boldsymbol{\mu})\approx\frac{1}{NA_{v_{||}}}\color[rgb]{0,0,0}\sum^{\begin{subarray}{c}\mathrm{errors\,in\,}m,\boldsymbol{\mu}\\\
v_{||}\mathrm{\,from\,}G(v_{||}|\dots)\end{subarray}}_{i}s_{i}^{5}\cos
b\frac{f(\boldsymbol{J}_{i})}{f^{\prime}(\boldsymbol{J}_{i})},$ (14)
and
$p(\ell,b,m)\approx\frac{1}{NA_{\boldsymbol{v}}}\color[rgb]{0,0,0}\sum^{\begin{subarray}{c}\mathrm{errors\,in\,}m\\\
\boldsymbol{v}\mathrm{\,from\,}G(\boldsymbol{v}|\dots)\end{subarray}}_{i}s_{i}^{3}\cos
b\frac{f(\boldsymbol{J}_{i})}{f^{\prime}(\boldsymbol{J}_{i})}.$ (15)
Note in the second expression we only have $3$ powers of $s$ as the integral
has been rewritten in terms of the 3d space velocity $\boldsymbol{v}$ (as
opposed to the observable space of proper motion and radial velocity). As we
are using a fixed potential, we precompute $\boldsymbol{J}_{i}$,
$R_{\mathrm{c},i}$ and the epicyclic frequencies for all samples using the
routines from Vasiliev (2019) and Bovy (2015).
### 3.2 Outlier component
Another complexity is to introduce an outlier distribution to overcome the
contamination of samples by stars which are members of the halo, are possibly
not Mira variable stars or have poorly-measured periods. To do this, we assume
that the velocity distribution of the contamination stars is described by a 3D
spherically symmetric Gaussian distribution that is centred on Galactocentric
$\boldsymbol{v}=\boldsymbol{0}$ with standard deviation in each dimension
$\sigma_{v}$. Similar to the previous approach, we calculate the
$p_{\mathrm{outlier}}(\boldsymbol{\mu}|\ell,b,m)$ using equation (4) and (5),
but replacing $p(\ell,b,m^{\prime},\boldsymbol{\mu}^{\prime},v_{||})$ with
$p_{\mathrm{outlier}}(\ell,b,m^{\prime},\boldsymbol{\mu}^{\prime},v_{||})$
which is chosen to be
$p_{\mathrm{outlier}}(\ell,b,m^{\prime},\boldsymbol{\mu}^{\prime},v_{||})=s^{5}\cos
b\,\mathcal{N}(\boldsymbol{v}|\boldsymbol{0},\sigma_{{v}}^{2}\boldsymbol{I})\mathcal{U}(x,y,z),$
(16)
where $\mathcal{U}(x,y,z)$ is the uniform distribution in Galactocentric
Cartesian spatial coordinates $(x,y,z)$. For each star,
$p_{\mathrm{outlier}}(\boldsymbol{\mu}|\ell,b,m)$ is evaluated numerically by
$p_{\mathrm{outlier}}(\boldsymbol{\mu}|\ell,b,m)=\frac{\sum^{\mathrm{errors\,in\,}m,\boldsymbol{\mu}}_{i}s_{i}^{5}\mathcal{N}(s_{i}\boldsymbol{\mu}_{i}+\boldsymbol{v}_{t,\odot,i}|\boldsymbol{0},\sigma^{2}_{v}\boldsymbol{I})}{\sum^{\mathrm{errors\,in\,}m}_{i}s_{i}^{3}},$
(17)
where $v_{t,\odot}$ is the solar velocity in the Galactocentric frame
projected in the plane perpendicular to the line-of-sight between the star and
the Sun. To include this distribution in the log-likelihood, we rewrite the
probability for each individual star as
$p_{\mathrm{tot},j}=(1-\epsilon)p_{j}(\boldsymbol{\mu}|\ell,b,m)+\epsilon
p_{\mathrm{outlier},j}(\boldsymbol{\mu}|\ell,b,m).$ (18)
Note with this definition, the outlier fraction at each spatial location,
$\epsilon$, is approximately constant. We choose the Gaussian because the
contamination could come from a variety of sources, and the Gaussian
distribution is a general, easily-computed way to characterise those sources.
### 3.3 Likelihood
We have now fully specified our model. The full log-likelihood for each
population of stars is
$\ln L=\sum^{\mathrm{stars}}_{j}\ln
p_{\mathrm{tot}}(\boldsymbol{\mu}_{j}|\ell_{j},b_{j},m_{j}),$ (19)
For each population of stars, we optimize the likelihood with respect to the
five parameters of the quasi-isothermal ($R_{\mathrm{disc}}$, $\sigma_{r,0}$,
$\sigma_{z,0}$, $R_{\sigma,r}$, $R_{\sigma,z}$) and the two parameters of the
outlier distribution $(\epsilon,\sigma_{v})$. The log-likelihood is explored
using MCMC performed with emcee (Foreman-Mackey et al., 2013). We adopt priors
on the radial scale lengths as
$R_{\mathrm{disc}}\sim\mathcal{N}(3.8\,\mathrm{kpc},(2\,\mathrm{kpc})^{2})$
and $R_{\sigma,r/z}\sim\mathcal{N}(4.5\,\mathrm{kpc},(3\,\mathrm{kpc})^{2})$
and a prior for velocity dispersion of the outlier component is a normal
distribution
$\sigma_{\boldsymbol{v}}\sim\mathcal{N}(200\,\mathrm{km\,s}^{-1},(150\,\mathrm{km\,s}^{-1})^{2})$.
The priors for the other three parameters are uniform:
$\sigma_{r/z,0}\sim\mathcal{U}(0,120\,\mathrm{km\,s}^{-1})$ and
$\epsilon\sim\mathcal{U}(0,1)$.
A final step in our procedure is converting the modelled distribution function
parameters to the physical measures of the velocity dispersion in the solar
neighbourhood. It is these quantities we compare with previous
characterisations of the age–velocity dispersion relation. For each set of
($R_{\mathrm{disc}}$, $\sigma_{r,0}$, $\sigma_{z,0}$, $R_{\sigma,r}$,
$R_{\sigma,z}$), we generate mock stars using the Agama DF sampling routines
and fit an exponential profile
$\sigma_{i}=\widetilde{\sigma}_{i,0}\exp[{(R_{0}-R)/\widetilde{R}_{\sigma,i}}]$
to the radial and vertical velocity dispersions binned in radius. The
normalization $\widetilde{\sigma}_{i,0}$ and scalelength
$\widetilde{R}_{\sigma,i}$ give the physical velocity dispersion and its
radial gradient in the solar neighbourhood.
### 3.4 Mock samples and validation
Figure 5: Results of fits on mock data: the lower left corner plot is the
posterior of the fitting parameters from the test on mock data including an
outlier distribution. The red lines are the parameters that generated the mock
sample, and the black dashed lines are the $16$th, $50$th and $86$th
percentiles of the posterior, respectively. The upper right corner plot gives
the posteriors of the physical velocity dispersion parameters corresponding to
the sets of fitted parameters. The physical velocity dispersion parameters are
propagated from the fitted parameters using the routine described in Section
4.
Given a fitted $f(\boldsymbol{J})$ model, we wish to draw mock samples to
compare with the data and validate our fitting procedure. We use the Agama DF
sampling routine to generate a large number of mock stars. For each generated
mock star, we find the nearest observed star in our dataset in $(R,z)$, place
the mock star at the azimuth $\phi$ of the real star and transform the mock
polar velocities to $\boldsymbol{\mu}$. This procedure exploits the
axisymmetry of the models. We further scatter the proper motions and distance
moduli of the mock stars by the corresponding uncertainties of the real stars.
The previous Mira selection criteria in the heliocentric distance and $R$ are
also applied to the mock sample. Note this procedure produces a mock dataset
with each real star corresponding to multiple mock stars in proportion to the
local stellar density at the location of the real star. This reduces the shot
noise in our mock samples but means the mock sample has a different spatial
density to the data. To reproduce the spatial distribution of the dataset, we
record the index of the closest matched real star for each mock star and then
count the number of times that this real star is the closest match to any mock
star. A weight is calculated for each mock star as the reciprocal of this
number count. The weight will be used when we compare our fitted model to the
dataset. When directly comparing to a fitted dataset, we further remove mock
stars which do not reside within $100\,\mathrm{pc}$ of any real star (this
requirement is not imposed on the mock test set described below but makes
little practical difference). Our procedure does not fully generate the data
as we have not accounted for uncertainty in the data $(R,z)$. However, it is
sufficient for validation purposes.
We can use the generated mock observations to test the validity of our method.
We generate a mock sample of $614$ stars from $f(\boldsymbol{J})$ with known
parameters chosen arbitrarily as $R_{\mathrm{disc}}=3.8\,\mathrm{kpc}$,
$\sigma_{r,0}=45.0\,\mathrm{km\,s}^{-1}$,
$\sigma_{z,0}=35.0\,\mathrm{km\,s}^{-1}$, $R_{\sigma,r}=4.5\,\mathrm{kpc}$ and
$R_{\sigma,z}=4.4\,\mathrm{kpc}$. We then replace velocities of $10\%$ of the
generated data with $\boldsymbol{v}$ sampled from a spherically symmetric
Gaussian
$\mathcal{N}(\boldsymbol{v}|\boldsymbol{0},(100\,\mathrm{km/s})^{2}\boldsymbol{I})$
which is the assumed velocity distribution of outlier stars. Stars sampled
from the outlier distribution have a chance to be unbound from the potential,
so after removing those unbound stars, the actual proportion of outlier stars
can be smaller than $10\%$, i.e. $\epsilon<10\%$. Without those high-velocity
stars in the mock sample, the velocity dispersion of the generated outlier
stars is reduced so a fitted $\sigma_{{v}}<100\,\mathrm{km/s}$ is expected but
the recovered parameters of the $f(\boldsymbol{J})$ model should be unbiased.
The posteriors from the MCMC are shown in the low left of Fig. 5. The
parameters $\sigma_{z,0}$ and $R_{\sigma_{z,0}}$ both deviate slightly from
the default parameters but only around the 1$\sigma$ level. In the upper right
corner of Fig. 5, we convert each set of fitted parameters into the physical
velocity dispersion profile parameters, $\widetilde{\sigma}_{i,0}$ and
$\widetilde{R}_{\sigma,i}$. Although there are small differences in the
distribution function parameters, the resulting physical velocity dispersions
and scalelengths at the solar position are well recovered. We also produced
the posterior of the same sample using the log-likelihood without the outlier
distribution. The medians of the parameters are
$R_{\mathrm{disc}}=3.56\,\mathrm{kpc}$,
$\sigma_{r,0}=65.74\,\mathrm{km\,s}^{-1}$,
$\sigma_{z,0}=43.16\,\mathrm{km\,s}^{-1}$, $R_{\sigma,r}=3.14\,\mathrm{kpc}$
and $R_{\sigma,z}=2.48\,\mathrm{kpc}$. As expected, $\sigma_{r,0}$ and
$\sigma_{z,0}$ are overestimated. This demonstrates that adding the outlier
distribution is necessary when the contamination of the sample is significant.
## 4 Velocity dispersion of O-rich Mira variable stars in different age bins
Figure 6: Velocity histograms for O-rich Mira variables separated by period (as given in days above each column). The top panels show $v_{\ell}$ and bottom $v_{b}$. The points are data and black lines the models. Table 1: Distribution function parameter estimates for the Mira variable model fits. The left column gives the considered period bin and the other columns show the median and uncertainties estimated from the $16$th and $84$th percentiles. Period range (day) | Mean period (day) | Number of stars | $R_{\mathrm{disc}}\,\mathrm{(kpc)}$ | $\sigma_{r,0}\,\mathrm{(km/s)}$ | $\sigma_{z,0}\,\mathrm{(km/s)}$ | $R_{\sigma,r}\,\mathrm{(kpc)}$ | $R_{\sigma,z}\,\mathrm{(kpc)}$ | $\epsilon$ | $\sigma_{\boldsymbol{v}}\,\mathrm{(km/s)}$
---|---|---|---|---|---|---|---|---|---
$80-150$ | 126.4 | 230 | $3.55^{+2.40}_{-1.19}$ | $48.05^{+6.17}_{-4.83}$ | $30.45^{+2.14}_{-2.01}$ | $8.32^{+3.29}_{-1.64}$ | $9.65^{+2.28}_{-2.41}$ | $0.06^{+0.02}_{-0.02}$ | $151.49^{+30.72}_{-22.91}$
$150-200$ | 179.3 | 430 | $3.56^{+1.27}_{-0.74}$ | $40.59^{+3.06}_{-3.26}$ | $41.07^{+3.62}_{-3.46}$ | $3.54^{+0.25}_{-0.19}$ | $6.47^{+1.14}_{-1.10}$ | $0.02^{+0.02}_{-0.02}$ | $105.58^{+61.84}_{-40.22}$
$200-225$ | 212.8 | 442 | $5.11^{+1.00}_{-1.35}$ | $53.14^{+5.51}_{-5.39}$ | $51.72^{+3.39}_{-2.89}$ | $5.05^{+1.19}_{-0.60}$ | $9.52^{+2.06}_{-1.94}$ | $0.01^{+0.04}_{-0.01}$ | $46.46^{+111.19}_{-21.72}$
$225-250$ | 237.7 | 494 | $3.79^{+1.43}_{-0.82}$ | $37.66^{+2.66}_{-2.18}$ | $55.99^{+3.74}_{-4.19}$ | $3.97^{+0.32}_{-0.31}$ | $10.36^{+2.89}_{-2.19}$ | $0.01^{+0.01}_{-0.01}$ | $78.35^{+71.38}_{-36.07}$
$250-275$ | 263.3 | 708 | $3.53^{+1.95}_{-0.91}$ | $52.38^{+3.59}_{-3.72}$ | $42.22^{+1.83}_{-2.10}$ | $12.16^{+3.37}_{-2.72}$ | $7.41^{+1.30}_{-0.92}$ | $0.03^{+0.02}_{-0.01}$ | $79.23^{+28.35}_{-12.22}$
$275-300$ | 287.2 | 909 | $4.47^{+1.55}_{-0.98}$ | $51.79^{+2.49}_{-2.11}$ | $39.57^{+2.00}_{-1.75}$ | $13.57^{+3.08}_{-2.17}$ | $7.24^{+1.27}_{-0.87}$ | $0.02^{+0.01}_{-0.01}$ | $104.36^{+22.33}_{-15.63}$
$300-325$ | 313.0 | 907 | $2.72^{+0.71}_{-0.58}$ | $46.46^{+2.04}_{-2.18}$ | $34.53^{+1.90}_{-1.72}$ | $11.68^{+2.13}_{-1.49}$ | $8.12^{+1.50}_{-1.15}$ | $0.00^{+0.01}_{-0.00}$ | $109.87^{+45.84}_{-36.79}$
$325-350$ | 337.7 | 970 | $2.49^{+0.67}_{-0.47}$ | $43.15^{+1.68}_{-1.75}$ | $32.94^{+1.34}_{-1.18}$ | $12.10^{+2.50}_{-1.54}$ | $9.08^{+1.55}_{-1.17}$ | $0.01^{+0.01}_{-0.00}$ | $114.63^{+37.64}_{-21.27}$
$350-375$ | 362.3 | 861 | $5.29^{+1.78}_{-1.38}$ | $42.44^{+1.82}_{-1.78}$ | $28.84^{+1.26}_{-1.38}$ | $13.52^{+4.08}_{-2.42}$ | $11.35^{+2.55}_{-2.39}$ | $0.01^{+0.01}_{-0.01}$ | $79.10^{+79.99}_{-45.18}$
$375-400$ | 387.3 | 784 | $4.69^{+2.07}_{-1.41}$ | $42.33^{+2.26}_{-1.62}$ | $23.89^{+1.54}_{-1.49}$ | $12.01^{+3.08}_{-1.96}$ | $7.89^{+2.12}_{-1.27}$ | $0.01^{+0.01}_{-0.01}$ | $79.19^{+48.55}_{-32.91}$
$400-450$ | 422.5 | 1015 | $2.87^{+1.40}_{-0.69}$ | $41.45^{+1.81}_{-1.61}$ | $25.77^{+0.86}_{-1.04}$ | $14.43^{+3.11}_{-2.01}$ | $13.69^{+2.48}_{-1.58}$ | $0.00^{+0.01}_{-0.00}$ | $88.41^{+66.37}_{-43.88}$
$450-500$ | 470.9 | 396 | $3.18^{+2.57}_{-1.28}$ | $37.42^{+1.90}_{-2.27}$ | $19.56^{+1.59}_{-1.71}$ | $13.22^{+2.70}_{-2.64}$ | $15.03^{+4.98}_{-4.03}$ | $0.04^{+0.02}_{-0.02}$ | $82.44^{+21.10}_{-18.52}$
$500-600$ | 527.5 | 144 | $4.68^{+2.68}_{-2.10}$ | $34.27^{+3.22}_{-2.85}$ | $16.85^{+1.92}_{-2.41}$ | $11.45^{+3.82}_{-3.06}$ | $11.34^{+5.52}_{-3.99}$ | $0.02^{+0.03}_{-0.01}$ | $121.84^{+69.88}_{-44.77}$
Table 2: Solar neighbourhood velocity dispersions and local spatial gradients of the velocity dispersions for the Mira variable fits. The age estimations are also provided, where $\tau_{r}$ is the age estimation from the radial velocity dispersion while $\tau_{z}$ is that from the vertical velocity dispersion. Period (days) | $\widetilde{\sigma}_{r,0}\,\mathrm{(km/s)}$ | $\widetilde{\sigma}_{z,0}\,\mathrm{(km/s)}$ | $\widetilde{R}_{\sigma,r}\,\mathrm{(kpc)}$ | $\widetilde{R}_{\sigma,z}\,\mathrm{(kpc)}$ | $\tau_{r}\,\mathrm{(Gyr)}$ | $\tau_{z}\,\mathrm{(Gyr)}$
---|---|---|---|---|---|---
$80-150$ | $49.83^{4.39}_{3.78}$ | $24.59^{1.47}_{1.38}$ | $10.54^{3.94}_{2.13}$ | $9.04^{1.64}_{1.64}$ | $8.57^{+1.13}_{-0.98}$ | $6.34^{+0.36}_{-0.34}$
$150-200$ | $67.20^{4.70}_{4.20}$ | $34.44^{1.86}_{1.43}$ | $6.88^{0.62}_{0.61}$ | $7.46^{0.85}_{0.92}$ | $10.82^{+0.56}_{-0.52}$ | $8.07^{+0.82}_{-0.75}$
$200-225$ | $62.96^{2.99}_{3.11}$ | $38.24^{1.55}_{1.56}$ | $7.92^{1.47}_{0.91}$ | $9.37^{1.48}_{1.27}$ | $10.41^{+0.45}_{-0.45}$ | $9.34^{+0.64}_{-0.66}$
$225-250$ | $52.83^{3.00}_{3.20}$ | $40.24^{1.67}_{1.78}$ | $6.16^{0.53}_{0.53}$ | $10.10^{1.66}_{1.42}$ | $9.25^{+0.73}_{-0.89}$ | $9.72^{+0.61}_{-0.67}$
$250-275$ | $51.90^{2.48}_{2.29}$ | $32.92^{1.14}_{1.16}$ | $15.31^{4.51}_{3.61}$ | $7.78^{0.94}_{0.77}$ | $9.09^{+0.72}_{-0.78}$ | $7.58^{+0.60}_{-0.55}$
$275-300$ | $50.54^{2.07}_{1.91}$ | $31.08^{1.21}_{1.08}$ | $16.61^{4.20}_{2.97}$ | $7.54^{1.05}_{0.76}$ | $8.75^{+0.78}_{-0.70}$ | $7.25^{+0.52}_{-0.44}$
$300-325$ | $46.67^{1.73}_{1.72}$ | $27.57^{1.11}_{1.01}$ | $14.35^{2.78}_{2.08}$ | $8.05^{1.15}_{0.87}$ | $7.80^{+0.55}_{-0.57}$ | $6.77^{+0.35}_{-0.35}$
$325-350$ | $43.27^{1.50}_{1.54}$ | $26.27^{0.86}_{0.88}$ | $14.41^{3.15}_{2.00}$ | $8.66^{1.10}_{0.91}$ | $7.01^{+0.59}_{-0.54}$ | $6.61^{+0.34}_{-0.34}$
$350-375$ | $41.74^{1.67}_{1.69}$ | $23.11^{0.89}_{0.93}$ | $15.35^{4.79}_{2.96}$ | $10.24^{1.74}_{1.75}$ | $6.66^{+0.52}_{-0.49}$ | $6.20^{+0.33}_{-0.30}$
$375-400$ | $41.98^{1.84}_{1.47}$ | $20.02^{1.00}_{0.94}$ | $13.73^{3.78}_{2.23}$ | $7.82^{1.65}_{1.12}$ | $6.67^{+0.51}_{-0.49}$ | $5.66^{+0.34}_{-0.32}$
$400-450$ | $41.08^{1.58}_{1.50}$ | $20.92^{0.71}_{0.76}$ | $16.45^{3.56}_{2.46}$ | $11.61^{1.65}_{1.17}$ | $6.43^{+0.47}_{-0.43}$ | $5.86^{+0.31}_{-0.31}$
$450-500$ | $37.21^{1.87}_{2.03}$ | $16.46^{1.21}_{1.16}$ | $14.60^{3.38}_{2.91}$ | $12.58^{3.00}_{2.60}$ | $5.52^{+0.57}_{-0.57}$ | $4.60^{+0.61}_{-0.96}$
$500-600$ | $34.24^{2.99}_{2.78}$ | $14.65^{1.34}_{1.85}$ | $12.44^{4.03}_{3.23}$ | $10.46^{3.60}_{3.11}$ | $4.50^{+0.86}_{-1.12}$ | $3.62^{+0.76}_{-1.05}$
Figure 7: Velocity dispersion profiles as a function of Galactocentric radius
for O-rich Mira variables separated by period (as given in days above each
column). The top panels show longitudinal, $\ell$, and bottom latitudinal,
$b$. The points are data and black lines the models. Figure 8: Velocity
histograms for O-rich Mira variables with periods in the range $275-300$ day
separated into bins of Galactocentric radius (as given above each column). The
top panels show the longitudinal velocity $v_{\ell}$ and the bottom the
latitudinal velocity, $v_{b}$. The red points are data and the black lines are
the models. Figure 9: Vertical density distribution profile for O-rich Mira
variables separated by period bins (as given in days above each panel). Each
panel shows the dataset (points) compared to the unweighted distribution of
mock samples (black lines). All histograms are normalised, and subplots do not
share the same y-axis. The discrepancy between the distributions is a
reflection of the completeness of the dataset.
To investigate the kinematic properties of the sample defined in Section 2, we
put the O-rich Mira variables into period bins and treat stars in each bin as
a sub-population drawn from the same DF. We choose the period bins to be wider
than the typical uncertainties in the period measurements, and hence we
neglect the period uncertainties that scatter stars from bin to bin (the
impact of the period uncertainties on the distance uncertainties _have_ been
considered). The median of the period uncertainties is $11.6$ days and $7.1$
days for those stars with periods less than $300$ days. We have also tried to
bin stars with a wider period bin ($50$ days instead of $25$), which gives
very similar results to the presented binning strategy. The adopted priors on
the radial scale lengths are
$R_{\mathrm{disc}}\sim\mathcal{N}(4\,\mathrm{kpc},(3\,\mathrm{kpc})^{2})$ and
$R_{\sigma,r/z}\sim\mathcal{N}(10\,\mathrm{kpc},(6\,\mathrm{kpc})^{2})$, the
prior for velocity dispersion of the outlier component is a normal
distribution
$\sigma_{{v}}\sim\mathcal{N}(100\,\mathrm{km\,s}^{-1},(80\,\mathrm{km\,s}^{-1})^{2})$
and the other priors are uniform as defined in the previous section. The
posterior distributions for the fits of each period bin are given in the
supplementary material and are summarised by the medians and percentiles in
Table 1. The contamination fraction $\epsilon$ is generally small and
$\sigma_{{v}}$ generally large for all period bins. Table 2 reports the
physical radial and vertical velocity dispersion normalization and scalelength
in the solar neighbourhood, $\widetilde{\sigma}_{i,0}$ and
$\widetilde{R}_{\sigma,i}$ respectively.
To verify the results of the MCMC fitting, we generate mock samples for the
best-fit parameters according to the procedure from Section 3.4, and we make
use of the weights for the mock sample to compare the kinematics of the fitted
model with the dataset under the same spatial distribution. In Fig. 6, we have
plotted the $v_{\ell}$ and $v_{b}$ distributions of these mock samples
compared to that of the observations, where $v_{\ell/b}=s\cdot\mu_{\ell/b}$.
We have chosen to omit the lowest period bin ($80-150$ days) from this plot
and in later plots and analysis because the contamination rate, $\epsilon$ is
the highest among other period bins (see Table 1) and it is likely it does not
follow the broad trend of increasing dispersion with decreasing period due to
contamination from short-period-red stars as we will discuss in Section 6.3.
For the displayed period bins, the mock samples generally agree with the
observations. For some period bins, the shape of the observed $v_{b}$ is
sharper than the mock sample implying that our modelling has some caveats.
Three reasons could lead to this: first, the assumed outlier distribution did
not characterise the contamination accurately and underestimated the outlier
star contribution consequently. Secondly, the period binning strategy needs to
be improved. Bins at long periods cover Mira variables of a broader range of
ages than the bins at short periods. Hence, if the younger stars in the period
bin have much smaller velocity dispersion than the average of the bin, the
sharper peak in observation would appear while the general shape of the
overall distribution is still correct. Thirdly, the assumed functional form
for the velocity dispersion parameters
$\sigma_{i}=\sigma_{i,0}\exp{(R_{0}-R)/R_{\sigma,i}}$ may be inappropriate. We
illustrate this final possibility by plotting the radial profile of the
longitudinal and latitudinal velocity dispersions $\sigma_{\ell}$ and
$\sigma_{b}$ in Fig. 7. For one or two period bins, the large $R$ radial
behaviour of $\sigma_{b}$ is not completely in agreement with the
observations. The $\sigma_{\ell}$ distribution is relatively more poorly
fitted than the $v_{b}$ distribution. Again, this could be due to the adopted
form of the distribution function. However, apart from these very minor
discrepancies, our modelling is in agreement with the observations. This is
reinforced by the comparison of the $v_{\ell}$ and $v_{b}$ distribution for
$275<\mathrm{Period/day}<300$ in Fig. 8. The model is in good agreement with
the observations. We will discuss further limitations of our approach in
Section 6.
As noted previously, the spatial distribution of stars has not been considered
in the modelling as it is subject to completeness effects arising from Gaia’s
scanning law and the effects of extinction. As a result, the spatial
distribution of the (unweighted) mock samples and the observations are in
disagreement when the completeness of the dataset is not considered. Our
weighting of the mock samples reproduces the spatial distribution of the data
enabling comparison of the kinematic fits as shown in Fig. 6, for example.
When the weights are not considered, the mock sample distribution can be
considered as the approximate underlying completeness-corrected distribution
of the data (only up to a point as according to our procedure, where there is
no data there will also be no mock stars). The weights are thus giving the
proportion of stars at each $\boldsymbol{x}$ that have been observed. This is
demonstrated in Fig. 9 by comparing the unweighted Galactic height
distribution of the mock sample to the dataset. Note that our procedure only
gives access to the relative completeness so the histograms have been chosen
to be normalized. The distributions of the data points are generally broader
than the unweighted mock distributions, which we interpret as incompleteness
in the dataset towards the Galactic midplane, possibly arising from
extinction. This interpretation of the unweighted mock samples assumes the
distribution functions well describe the Milky Way sub-populations. We discuss
the shortcomings of the approach later, but the good agreement in Fig. 9 also
demonstrates that even without considering incompleteness, the distribution
functions do a good job of describing the data.
## 5 Period–age relationship
Figure 10: The calibrated age–period relationship of the O-rich Mira
variables. The orange and violet points are the velocity dispersion from the
kinematic modelling. The orange, purple, and black lines are the fitted
period–age relations using radial, vertical velocity dispersions, and two
together respectively, with fitted parameters given in Table 3.
With the dynamical distribution functions in each Mira variable period bin
well characterised, we now turn to what this implies for the corresponding age
of each period bin. To do this we must adopt an age–velocity dispersion
relation (AVR). We choose the AVR measured by Yu & Liu (2018) from LAMOST data
of $\sim 3500$ sub-giant/red giant stars. Yu & Liu (2018) characterised the
velocity dispersions of their sample split into age bins using the entirety of
their dataset and also for two sets split by Galactic height:
$|z|<0.27\,\mathrm{kpc}$ and $|z|>0.27\,\mathrm{kpc}$. The ages of stars in Yu
& Liu (2018) were estimated by comparing the stellar parameters
($[\mathrm{Fe/H}]$, $\mathrm{T_{eff}}$, $\log g$) measured by LAMOST to a grid
of isochrone models. Age estimates were found by marginalizing the likelihood
over initial mass and absolute magnitude. The AVRs were produced by further
binning stars in their sample by age. This procedure accounts for
uncertainties arising from the velocities but _not_ the ages. We discuss the
impact of this later.
We estimate the corresponding AVR of our sample by averaging the two
$|z|$-separated AVRs in Yu & Liu (2018) weighted by the number of stars in our
sample that are above and below $|z|=0.27\,\mathrm{kpc}$ in each bin.
Consequently, the final AVR was slightly different for each bin. At low ages,
the corresponding AVR is not monotonic due in part to uncertainties and the
low numbers of stars in some low-age bins. Thus, we remove points in the AVR
if the age is less than that of the previous age bin so that we could
interpolate a monotonic AVR to find an age at each radial and vertical
dispersion, $\widetilde{\sigma}_{r,0}$ and $\widetilde{\sigma}_{z,0}$. The
uncertainty is again propagated using Monte Carlo samples. The final
calibrated age–period relationship is shown in Fig. 10.
Yu & Liu (2018) discussed that the uncertainties in the estimated ages of
stars would broaden the measured AVR. Liu et al. (2015) argued that the age
estimation method used in Yu & Liu (2018) could have uncertainties at the
$30\,\mathrm{per\,cent}$ level which propagate from the uncertainties of the
LAMOST stellar parameters. Here, we will discuss how much this effect would
affect the period–age relationship. We generate $500\,000$ stars with
uniformly distributed ages and assign each star a radial and vertical velocity
from a Gaussian distribution centred at $0$ with standard deviations of
$\sigma_{r}$ and $\sigma_{z}$ calculated from the AVR. Then, the ages of the
stars are scattered by $(10,20,30)\%$ uncertainties. We then bin the stars
with the scattered age and calculate the measured radial and vertical velocity
dispersion. The ratio of the measured to actual velocity dispersion for the
AVR is given in Fig. 11, where the left and right panels are made for the AVR
of $|z|<0.27\,\mathrm{kpc}$ and $|z|>0.27\,\mathrm{kpc}$ respectively. We
divide this ratio by the corresponding velocity dispersions in the AVR as a
correction. In Fig. 12 we show the period–age relations calibrated using AVRs
with different levels of age uncertainty. We see that with $30\%$ uncertainty
in AVR the maximum correction could be up to $20\%$ in age as calibrated from
$\sigma_{R,0}$ and $34\%$ from $\sigma_{z,0}$.
Figure 11: The ratio of the age–velocity dispersion relation broadened by
different age uncertainties ($10,20$ and $30\,\mathrm{per\,cent}$ denoted by
dotted, dashed and solid) relative to the ‘true’ age–velocity dispersion
relation without age uncertainties. The left panel shows results for the
$|z|<0.27\,\mathrm{kpc}$ AVR from Yu & Liu (2018) and the right panel their
age–velocity dispersion relation for $|z|>0.27\,\mathrm{kpc}$. Yellow lines
correspond to $\sigma_{r}$ and blue $\sigma_{z}$. Figure 12: The calibrated
period–age relationship using age–velocity dispersion relations broadened by
different age uncertainties (as labelled in the legend). The relation
calibrated by $\widetilde{\sigma}_{r,0}$ is shown in the left panel while
$\widetilde{\sigma}_{z,0}$ is on the right. The error bars are not shown in
this figure. The black dashed lines in both panels are the fitted period–age
relations shown by the orange and pink lines in Fig. 10 respectively.
We have also considered other recent AVR calibrations available in the
literature. For example, Sharma et al. (2021) have provided a fit of the
radial and vertical dispersions in a separable form in terms of the age,
angular momentum, metallicity and Galactic height. Their relations produce
significantly smaller dispersions at fixed age such that the derived
period–age relation will assign significantly larger ages at fixed period
which in the extreme can be $\gg 14\,\mathrm{Gyr}$. We are therefore inclined
to use the Yu & Liu (2018) relations and the applicability of the Sharma et
al. (2021) relations merits further investigation.
## 6 Discussion
We now turn to the interpretation and understanding of our results, in
particular concentrating on the comparison with previous period–age estimates
for Mira variable stars and possible future model improvements.
### 6.1 Comparison with Mira variable cluster members and previous results
Figure 13: Comparison of the derived period–age relations with other
literature results. The orange squares and pink triangles show our Mira
variable period–age measurements from Table 2. The small grey points are from
the models of Wyatt & Cahn (1983), the green short-dashed line from the model
of Eggen (1998) and the orange long-dashed line from the model of Trabucchi &
Mowlavi (2022, along with the associated scatter shown by the shaded region).
The light blue squares are Mira variable globular cluster members from Clement
et al. (2001), the brown diamonds C-rich Mira variable open cluster members
from Marigo et al. (2022) and the light-blue triangles LMC cluster members.
The solid blue line is a fit from Grady et al. (2019) to a broader sample of
LMC cluster members. The grey points are period–age estimates for disc
populations from Feast et al. (2006), Feast (2009) and Feast & Whitelock
(2014). The black line is the joint fit of our results and the globular
cluster members from Table 3 and the thinner orange, pink and grey lines show
the other three fits from that same table.
In Fig. 10 we display a series of period–age indicators of Mira variable
stars. The age–kinematic method for period–age calibration has been utilised
by Feast et al. (2006), Feast (2009) and Feast & Whitelock (2014). Feast &
Whitelock (2000b) demonstrated that Mira variables in the solar neighbourhood
exhibited clear correlations between period and kinematics. These have been
translated approximately into period–age measurements using results from the
solar neighbourhood in the cited works. However, it should be said that all of
the quoted results are only approximate due to the absence of robust
age–kinematics calibrations.
Mira variables in clusters give a more direct measurement of the period–age
relation than the indirect method using the age–kinematic calibrations.
Unfortunately, there are comparatively few cluster Mira variables. Those in
globular clusters have been studied by Sloan et al. (2010) whilst those with
good evidence of Milky Way open cluster membership from Gaia have been studied
by Marigo et al. (2022). There are also many candidates for LMC cluster
membership as studied by Grady et al. (2019). However, membership of an LMC
cluster is difficult to discern purely from projected coordinates (as used by
Grady et al., 2019) and proper motion data. We compile Mira variable globular
cluster members using the globular cluster variable star compilation from
Clement et al. (2001). We consider all stars flagged as ‘M’ or ‘M?’, and not
flagged as a likely field star (‘f’ or ‘f?’). Furthermore, if available, we
ensure the Gaia DR3 proper motion is within $3\sigma$ of the measured cluster
mean proper motion from Vasiliev & Baumgardt (2021). Here $\sigma$ is a
quadrature sum of the measurement uncertainty and the central velocity
dispersion. We complement with ages primarily from VandenBerg et al. (2013)
and Dotter et al. (2010), and from Beaulieu et al. (2001) for NGC 6553,
Geisler et al. (2007) for Terzan 7, Ortolani et al. (1999) for Terzan 1,
Marín-Franch et al. (2009) and Forbes & Bridges (2010) for NGC 6441 and Santos
& Piatti (2004) for NGC 6356, NGC 6388, NGC 6642 and NGC 6760. Terzan 5 has
evidence of multiple star formation events (Ferraro et al., 2016) so we assign
stars with periods $<400$ day an age of $12\,\mathrm{Gyr}$ and longer-period
stars an age of $4.5\,\mathrm{Gyr}$. There is a carbon-rich Mira variable in
the old globular cluster Lyngå 7 that has been suggested as a product of
binary evolution (Feast et al., 2013). However, its Gaia DR3 proper motion is
not consistent with being a cluster member. Its radial velocity is perfectly
consistent so one possibility is that the Gaia measurement is spurious. This
seems quite likely as there are two nearby Gaia DR3 sources with only two-
parameter astrometric solutions suggesting contamination in the Lyngå 7 C-rich
Mira variable measurement.
For Mira variable open cluster members, we use the compilation from Marigo et
al. (2022) adopting their measured periods and the cluster ages from Cantat-
Gaudin et al. (2020). Marigo et al. (2022) identify some cluster members on
the fundamental period–luminosity relation followed by Mira variable stars but
with too low an amplitude for traditional Mira variable classification. We
consider all stars that Marigo et al. (2022) identify as fundamental pulsators
and with $G$ band amplitudes greater than $0.865\,\mathrm{mag}$ (Grady et al.,
2019) estimated from the photometric uncertainties. There are two such stars
with are both C-rich.
Finally, we consider possible LMC and SMC cluster members from the Gaia DR3
LPV candidate catalogue. We combine the list of cluster ages from Baumgardt et
al. (2013) and Bonatto & Bica (2010). To limit contaminants, we conservatively
find all Gaia DR3 LPV candidates within one cluster radius as determined by
Bica et al. (2008) (adopting the median cluster radius of
$0.45\,\mathrm{arcmin}$ when a radius is not available). We further limit to
those with proper motions within $3\sigma$ of
$(\mu_{\alpha}*,\mu_{\delta})=(1.910,0.229)\mathrm{mas\,yr}^{-1}$
(Kallivayalil et al., 2013) where $\sigma$ is the quadrature sum of the
uncertainties and $100\,\mathrm{km\,s}^{-1}$ at the distance of the LMC, and
those with distances between $30$ and $70\,\mathrm{kpc}$ as determined from
equation (1). We isolate Mira variables by restricting to stars with $G$
amplitudes $>0.865\,\mathrm{mag}$ as determined by the $G$ photometric
uncertainties and the Fourier light curve fits. This results in $4$ high-
confidence LMC cluster members.
The described combination of cluster measurements is shown in Fig. 13. We see
in general the good agreement between the results derived from the
age–kinematic relation and the cluster members. There are some globular
cluster members with longer periods but higher ages (most notably the $312$
day period Mira in NGC 5927 which has an age of $12.25\,\mathrm{Gyr}$ from
Dotter et al. 2010 and $10.75\,\mathrm{Gyr}$ from VandenBerg et al. 2013).
This may reflect metallicity dependence in the period–age relation or these
could be the results of binary evolution in these clusters producing slightly
more massive AGB stars than expected at fixed age.
There are several theoretical period–age relations from the literature. The
earliest of these are the results from Wyatt & Cahn (1983) who found ages for
local Mira variable stars via main-sequence mass estimates derived from models
of Mira variables as fundamental pulsators which were fitted to optical and
infrared photometry and periods. Eggen (1998) similarly provided a
theoretically-motivated period–age relation by supposing fundamental Mira-like
pulsations occur once a star of a given mass (age) reaches some critical
radius. Most recently, Trabucchi & Mowlavi (2022) have used theoretical models
to produce period–age calibrations for O-rich and C-rich Mira variable stars.
They highlighted one expectation of the models is a large spread of age at
fixed period. Furthermore, their period–age relations agreed very well with
the cluster member measurements mostly compiled by Grady et al. (2019).
However, as we have hinted at above, there is perhaps good reason to believe
that the LMC cluster members are quite a contaminated set and that LMC field
stars coincident on the sky with the clusters are likely to be incorrectly
identified as cluster members. The field stars will typically be older than
the cluster members, having already left their parent clusters, and so these
contaminants will act to decrease the typical age at fixed period. It could be
that there is an additional variable controlling the period–age relation that
produces the discrepancy between the LMC clusters and the local age–kinematic
relations. The spread in models from Trabucchi & Mowlavi (2022) is almost
consistent with the measurements made here. However, the globular clusters
suggest any metallicity dependence would go the other way. Furthermore, binary
evolution produces higher periods at fixed age so would not explain the
discrepancy.
A further supporting piece of evidence for the age–period relation we have
derived here is the properties of the LMC population as a whole and the
Galactic bulge sample. In both sets, there are stars with $\sim
500-600\,\mathrm{day}$ periods. From our calibrations, these stars are $\sim
3-4\,\mathrm{Gyr}$ old. The LMC has a tail towards longer-periods consistent
with even more recent star formation. The Galactic bulge is primarily
considered as an old population (Zoccali et al., 2003) although there has been
significant evidence that there are intermediate-age populations as young as
$3\,\mathrm{Gyr}$ (Bensby et al., 2013, 2017; Nataf, 2016). Our calibration is
entirely consistent with these results. A lower age–period relation would mean
a significant population of stars in the Galactic bulge with $\lesssim
1\,\mathrm{Gyr}$ old populations although again we should stress the expected
spread in ages at each period could still produce some consistency in the
results.
### 6.2 A parametric period–age relation
Our fitting has provided the approximate ages of O-rich Mira variable
populations in a series of period bins. It is more convenient to work with an
analytic relation that approximately fits the results. The flexible form
$\tau=\tau_{0}\frac{1}{2}\left(1+\tanh\Big{[}\frac{330-P(\mathrm{days})}{P_{s}}\Big{]}\right),$
(20)
provides an approximate fit to the data. We take the data reported in Table 2
and fit equation (20) allowing for an additional fractional scatter in the
ages of $f_{\tau}$ such that the age errors are
$\sqrt{\sigma_{\tau}^{2}+f_{\tau}^{2}\tau^{2}}$. $(\tau_{0},P_{s},f_{\tau})$
are given logarithmic flat priors and we sample using emcee (Foreman-Mackey et
al., 2013). We fit for $\sigma_{r}$ and $\sigma_{z}$ both separately and
jointly and report the results in Table 3. Although the dispersion parameters
are derived from the same model fit, the corner plots in the supplementary
material demonstrate the parameter constraints are uncorrelated for nearly all
period bins validating treating the results in this way. We also perform a
joint fit of the dispersion results together with the globular cluster member
compilation described in the previous section, again reporting the results in
Table 3. All four sets of results are quite consistent with $\sigma_{z}$-only
fits producing the lowest age at fixed period and the combination with the
globular clusters producing the highest. As expected, the scatter is largest
for the combined fit with the globular clusters but nevertheless, the scatter
is only around $10\,\mathrm{per\,cent}$ in age.
Table 3: Functional form for the period–age relation fitted to our results. We adopt the form $\tau=(\tau_{0}/2)\left(1+\tanh((330-P(\mathrm{day}))/P_{s}\right)$ with a fractional age uncertainty of $f_{\tau}$. Subset | $\tau_{0}$ | $P_{s}$ | $\ln f_{\tau}$
---|---|---|---
$\sigma_{r}$ | $14.9\pm 0.7$ | $389\pm 77$ | $-6.70\pm 0.01$
$\sigma_{z}$ | $13.0\pm 0.5$ | $404\pm 111$ | $-5.71\pm 0.03$
Both | $13.7\pm 0.6$ | $401\pm 88$ | $-2.63\pm 0.04$
With GC | $14.7\pm 0.7$ | $308\pm 54$ | $-2.17\pm 0.04$
### 6.3 Model limitations and future improvements
Before concluding, we will discuss some of the limitations of our modelling
and possible improvements that could be adopted in future analyses.
_Binning in period_ : We have opted to bin our data in period and analyse each
period bin independently. This is a valid approach as the period uncertainties
are typically quite small: the median period uncertainty is $11.6$ day and
$7.1$ day for $\mathrm{period}<300\,\mathrm{day}$. Hence, our strategy is
valid for most of the period bins considered. A further generalization is to
express the models in terms of the period as a continuous subpopulation label.
We then have to introduce hyper-parametrizations for the parameters in
$f(\boldsymbol{J})$ to express $f(\boldsymbol{J}|P)$. The integrals would
involve an additional integral over the label $P$ and we would have a
weighting of the populations $f(P)$ (which if we are considering periods as
proxies for age is akin to a star formation rate and could be an exponential
in age, for example). The advantage of this approach is a more principled
accounting of the period uncertainties as well as providing a route to
consider the spread in age (kinematics) at each period that might arise from
helium flashes, hot-bottom burning or the presence of short-period red stars.
The downside of such an approach is that we would have to fit a parametrized
form for the parameters as a function of $P$ making the models significantly
more complicated and potentially producing biased by our choice of functional
form.
_Velocity dispersion profile_ : We have here adopted a simple pure exponential
decay for the velocity dispersion of each period bin. This form gives a good
fit of the models to the data, particularly as we have chosen a rather limited
Galactocentric radial range. It has been suggested that the velocity
dispersion in the outer disc flattens or even increases with radius (Sanders &
Das, 2018; Mackereth et al., 2019). For example, Sharma et al. (2021) argues
that the pure exponential decay of the velocity dispersions is not well
motivated by the data, which shows signs of a rising dispersion beyond the
solar radius. To incorporate this possibility, one possible change is to
modify $\tilde{\sigma}_{i}(R_{c})$ as
$\tilde{\sigma}_{i}(R_{c})\equiv\sigma_{i,0}(\exp[-(R_{c}-R_{0})/R_{\sigma,i}]+\alpha_{i}(R_{i}/R_{0})^{2})/(1+\alpha_{i}),$
(21)
with the additional fitting parameters $\alpha_{i}$ to match the
flattening/upturning dispersion profiles in the outer disc as suggested by
Sharma et al. (2021). This may be a necessary enhancement when modelling the
data beyond the extended solar neighbourhood. For example, if one were to
consider investigating possible metallicity dependence of the period–age
relation. However, such an enhancement does not seem necessary for our data.
_Limitations of equilibrium axisymmetric distribution function approach_ : It
is reassuring to note that the age estimates from the radial and vertical
dispersions separately give very similar results for the period–age relation
of the O-rich Mira variables. However, the relation derived from the radial
dispersion is consistently higher than that derived from the vertical
dispersion. We have seen how our dynamical models capture well both the
longitudinal and latitudinal velocity distributions of the sample but
typically the latitudinal distributions are better modelled suggesting our
results are more reliable for the period–age relation derived from
$\sigma_{z}$. This occasional mismatch of the longitudinal dispersion in Fig.
7 could be a shortcoming of the use of a quasi-isothermal distribution
function. There are other action-based disc models available in the literature
(e.g. Binney & Vasiliev, 2023) which could be explored. As mentioned
previously, using a dynamical distribution function simply incorporates the
required asymmetry in the azimuthal component as well as necessarily linking
together the radial and azimuthal dispersions due to the requirement of
dynamical equilibrium. There could also be inconsistencies arising from this
assumption of equilibrium as it is known that the Galactic disc shows non-
equilibrium structure at the $5-10\,\mathrm{per\,cent}$ level. Any inflation
of the velocity dispersion as a result of this is not a concern as we have
anchored to tracers that will also display this inflation. The assumption of
axisymmetry could also be giving rise to similar variations. We are using the
velocity dispersion at the solar radius from a range of different azimuths but
if the velocity dispersion is varying significantly with azimuth (e.g. Gaia
Collaboration et al., 2022), the comparison between our sample and the
age–velocity dispersion results from Yu & Liu (2018) may be inappropriate.
Furthermore, our model has assumed a fixed Milky Way potential from McMillan
(2017). Whilst this potential captures many of the global features of the
Milky Way, it may not in detail be appropriate across the entirety of the
Galactic disc region considered here. In the wrong potential, it may be very
difficult to fully match the full velocity distribution of the data at every
spatial location. Reasonable variations of the potential will likely inflate
the uncertainties in our derived parameters. We should also note that although
we have inflated the Gaia astrometric uncertainties in our analysis to reflect
shortcomings of the current Gaia data processing, it is likely that future
Gaia data releases will improve the uncertainty estimates providing a better
handle on the underlying dispersions of the disc populations. This may
decrease the dispersion for the youngest populations (e.g. the $500-600$ day
period bin) but the dispersions of the oldest populations are very insensitive
to the uncertainties so we believe our measurements are reliable.
_SP-red stars_ : We found the stars in our lowest considered period bin
($80-150$ day) have significantly lower dispersions and hence lower ages than
the neighbouring $150-200$ day bin (see Table 1). This bucks the broad trend
seen in e.g. Fig. 13 and for this reason, as well as the fact that this bin
requires the largest outlier fraction of all modelled bins, we decided to
neglect these results in our period–age relation fits. Feast & Whitelock
(2000b) found a similar effect from Hipparcos data that they attribute to
short-period(SP)-red stars which contaminate the short-period end and are
kinematically more similar to the longer-period Mira variables. It is not
clear exactly what the origin of these stars is and they could represent a
different evolutionary stage to the bulk Mira variable population. Feast &
Whitelock (2000b) hypothesise they could be stars on their way to becoming
longer-period Mira variables or temporarily dimmed during their helium-shell
flash cycle (Trabucchi et al., 2017). From Gaia-2MASS colour-colour diagrams,
we did not clearly identify a distinct population of SP-red-like stars in the
short-period bin but it is likely they are present and potentially also more
weakly contaminating the $150-200$ day bin which also shows a slightly lower
$\sigma_{z}$ than perhaps expected. It is known that Mira variables in
globular clusters follow a period–metallicity relation with shorter-period
stars more prevalent in metal-poor clusters (Feast & Whitelock, 2000a). This
then suggests that the shortest period bin we considered has significant
contamination from more metal-poor objects and is not representative of the
broader solar neighbourhood samples used to calibrate the period–age
relations. However, it is then surprising that a more metal-poor population
would have a lower than expected dispersion as in both in-situ and accreted
scenarios the opposite is likely the case. More generally, our methodology
could be impacted by metallicity effects. We have already limited to O-rich
Mira variables which should preferentially remove metal-poor stars. Further
investigation is required to separate out the degeneracies between period, age
and metallicity, and a possible avenue is to consider the variation of
kinematics with unextincted colour as a metallicity proxy (e.g. Alvarez et
al., 1997).
_Hot-bottom burning_ : From equation (1), the slope of the period–luminosity
relation changes after $\mathrm{Period(days)}>400$. This hints that our O-rich
Mira variable star sample with periods above $400$ days is a mixture of hot-
bottom burning (HBB) stars and low-mass fundamental pulsators right at the end
of their lifetime (Whitelock et al., 2003; Trabucchi et al., 2019). The
balance of these two kinematically distinct populations depends on the star-
formation history (e.g. the HBB population would be reduced if there is no
recent star formation). Hence, as we are measuring the average age at a fixed
period, our result is somewhat related to the star-formation history of the
Milky Way. This mixing of HBB stars likely also broadens the period–age
relation for $\mathrm{Period(days)}>400$ (as it perhaps does the
period–luminosity relation e.g. Ita & Matsunaga, 2011), and it might address
the small discrepancy between our relation and the literature results shown in
Fig. 13. We hypothesise that the period–age relation is more universal and
reliable for periods under $400$ days.
_C-rich stars_ : Finally, a further direction is to consider the C-rich Mira
variables from Gaia. C-rich Mira variables also follow period–luminosity
relations that are typically broader than that for the O-rich Mira variables
due to circumstellar dust (Ita & Matsunaga, 2011). They also appear to trace
period–age relations (e.g. Feast et al., 2006, and evidenced in Fig. 13).
Typically they are less abundant in the Galaxy than the O-rich counterparts
(Ishihara et al., 2011) but importantly are biased towards younger ages (and
lower metallicities, e.g. Boyer et al., 2013) so present a route to better
constraining the longer-period end of the Mira variable period–age relation.
## 7 Conclusions
We have used the Gaia DR3 long-period variable candidate catalogue to produce
a calibration of the Mira variable period–age relation. Using a carefully
selected population of likely O-rich Mira variable stars, we have fitted a
series of action-based dynamical models to the stars separated by period. We
have found very good model fits for the velocity distributions of our sample
from which we have derived period–kinematic relations for the solar
neighbourhood. Comparison with an age–velocity dispersion relation for sub-
giant/red giant stars in the solar neighbourhood has allowed us to provide a
calibration of the Mira variable period–age relation.
Our derived relation agrees well with previous literature approaches using a
similar methodology and with the members of clusters with known ages. Some
theoretical models agree well with the derived relation but more recent
calibrations appear to be consistently younger at fixed period than our
relations suggest. Consideration of the age distribution of Mira variable
stars in the Galactic bar-bulge produces a consistent picture with other bar-
bulge age tracers using our relation.
This new period–age relation opens the possibility of inspecting the star
formation history and evolutionary properties of distant and/or highly-
extincted regions of our Galaxy and the Local Group. Mira variables are some
of the brightest stars in an intermediate-age population, their infrared
brightness makes them ideal tracers of dusty environments, and their high
amplitude and long periods mean they suffer low contamination. For these
reasons, in the era of JWST, Mira variables will provide us with a new window
of the evolution of the Universe.
## Data availability
All data utilised in this work are in the public domain. In the supplementary
material, we provide corner plots showing the posterior distributions of the
dynamical model parameters for each period bin.
## Acknowledgements
We thank the anonymous referee for a careful reading of the paper and
thoughtful comments that improved the presentation. JLS thanks the support of
the Royal Society (URF\R1\191555). This work has made use of data from the
European Space Agency (ESA) mission Gaia (https://www.cosmos.esa.int/gaia),
processed by the Gaia Data Processing and Analysis Consortium (DPAC,
https://www.cosmos.esa.int/web/gaia/dpac/consortium). Funding for the DPAC has
been provided by national institutions, in particular, the institutions
participating in the Gaia Multilateral Agreement. This publication makes use
of data products from the Two Micron All Sky Survey, which is a joint project
of the University of Massachusetts and the Infrared Processing and Analysis
Center/California Institute of Technology, funded by the National Aeronautics
and Space Administration and the National Science Foundation. This paper made
use of numpy (van der Walt et al., 2011), scipy (Virtanen et al., 2020),
matplotlib (Hunter, 2007), seaborn (Waskom et al., 2017), pandas (McKinney,
2010), corner (Foreman-Mackey, 2016) astropy (Astropy Collaboration et al.,
2013; Price-Whelan et al., 2018), galpy (Bovy, 2015), and Agama (Vasiliev,
2019).
## References
* Alvarez et al. (1997) Alvarez R., Mennessier M. O., Barthes D., Luri X., Mattei J. A., 1997, A&A, 327, 656
* Andriantsaralaza et al. (2022) Andriantsaralaza M., Ramstedt S., Vlemmings W. H. T., De Beck E., 2022, A&A, 667, A74
* Angus et al. (2015) Angus R., Aigrain S., Foreman-Mackey D., McQuillan A., 2015, MNRAS, 450, 1787
* Angus et al. (2020) Angus R., et al., 2020, arXiv e-prints, p. arXiv:2005.09387
* Astropy Collaboration et al. (2013) Astropy Collaboration et al., 2013, A&A, 558, A33
* Aumer & Binney (2009) Aumer M., Binney J. J., 2009, MNRAS, 397, 1286
* Aumer et al. (2016) Aumer M., Binney J., Schönrich R., 2016, MNRAS, 462, 1697
* Barbanis & Woltjer (1967) Barbanis B., Woltjer L., 1967, ApJ, 150, 461
* Baumgardt et al. (2013) Baumgardt H., Parmentier G., Anders P., Grebel E. K., 2013, MNRAS, 430, 676
* Beaulieu et al. (2001) Beaulieu S. F., Gilmore G., Elson R. A. W., Johnson R. A., Santiago B., Sigurdsson S., Tanvir N., 2001, AJ, 121, 2618
* Bensby et al. (2013) Bensby T., et al., 2013, A&A, 549, A147
* Bensby et al. (2017) Bensby T., et al., 2017, A&A, 605, A89
* Bica et al. (2008) Bica E., Bonatto C., Dutra C. M., Santos J. F. C., 2008, MNRAS, 389, 678
* Binney (2010) Binney J., 2010, MNRAS, 401, 2318
* Binney (2012) Binney J., 2012, MNRAS, 426, 1324
* Binney & Tremaine (2008) Binney J., Tremaine S., 2008, Galactic Dynamics: Second Edition. Princeton University Press
* Binney & Vasiliev (2023) Binney J., Vasiliev E., 2023, MNRAS, 520, 1832
* Bird et al. (2021) Bird J. C., Loebman S. R., Weinberg D. H., Brooks A. M., Quinn T. R., Christensen C. R., 2021, MNRAS, 503, 1815
* Blanco et al. (1984) Blanco V. M., McCarthy M. F., Blanco B. M., 1984, AJ, 89, 636
* Bland-Hawthorn & Gerhard (2016) Bland-Hawthorn J., Gerhard O., 2016, ARA&A, 54, 529
* Bonatto & Bica (2010) Bonatto C., Bica E., 2010, MNRAS, 403, 996
* Bovy (2015) Bovy J., 2015, ApJS, 216, 29
* Boyer et al. (2013) Boyer M. L., et al., 2013, ApJ, 774, 83
* Cantat-Gaudin et al. (2020) Cantat-Gaudin T., et al., 2020, A&A, 640, A1
* Catchpole et al. (2016) Catchpole R. M., Whitelock P. A., Feast M. W., Hughes S. M. G., Irwin M., Alard C., 2016, MNRAS, 455, 2216
* Chiavassa et al. (2011) Chiavassa A., et al., 2011, A&A, 528, A120
* Chiavassa et al. (2018) Chiavassa A., Freytag B., Schultheis M., 2018, A&A, 617, L1
* Clement et al. (2001) Clement C. M., et al., 2001, AJ, 122, 2587
* De Silva et al. (2015) De Silva G. M., et al., 2015, MNRAS, 449, 2604
* Dotter et al. (2010) Dotter A., et al., 2010, ApJ, 708, 698
* Eggen (1998) Eggen O. J., 1998, AJ, 115, 2435
* El-Badry et al. (2021) El-Badry K., Rix H.-W., Heintz T. M., 2021, MNRAS, 506, 2269
* Eyer et al. (2022) Eyer L., et al., 2022, arXiv e-prints, p. arXiv:2206.06416
* Feast (1963) Feast M. W., 1963, MNRAS, 125, 367
* Feast (2009) Feast M. W., 2009, in Ueta T., Matsunaga N., Ita Y., eds, AGB Stars and Related Phenomena. p. 48 (arXiv:0812.0250)
* Feast & Whitelock (2000a) Feast M., Whitelock P., 2000a, in Matteucci F., Giovannelli F., eds, Astrophysics and Space Science Library Vol. 255, Astrophysics and Space Science Library. p. 229 (arXiv:astro-ph/9911393), doi:10.1007/978-94-010-0938-6_22
* Feast & Whitelock (2000b) Feast M. W., Whitelock P. A., 2000b, MNRAS, 317, 460
* Feast & Whitelock (2014) Feast M., Whitelock P. A., 2014, in Feltzing S., Zhao G., Walton N. A., Whitelock P., eds, IAU Symposium Vol. 298, Setting the scene for Gaia and LAMOST. pp 40–52 (arXiv:1310.3928), doi:10.1017/S1743921313006182
* Feast et al. (2006) Feast M. W., Whitelock P. A., Menzies J. W., 2006, MNRAS, 369, 791
* Feast et al. (2013) Feast M. W., Menzies J. W., Whitelock P. A., 2013, MNRAS, 428, L36
* Ferraro et al. (2016) Ferraro F. R., Massari D., Dalessandro E., Lanzoni B., Origlia L., Rich R. M., Mucciarelli A., 2016, ApJ, 828, 75
* Forbes & Bridges (2010) Forbes D. A., Bridges T., 2010, MNRAS, 404, 1203
* Foreman-Mackey (2016) Foreman-Mackey D., 2016, The Journal of Open Source Software, 1, 24
* Foreman-Mackey et al. (2013) Foreman-Mackey D., Hogg D. W., Lang D., Goodman J., 2013, PASP, 125, 306
* Freeman & Bland-Hawthorn (2002) Freeman K., Bland-Hawthorn J., 2002, ARA&A, 40, 487
* Gaia Collaboration et al. (2016) Gaia Collaboration et al., 2016, A&A, 595, A1
* Gaia Collaboration et al. (2018) Gaia Collaboration et al., 2018, A&A, 616, A1
* Gaia Collaboration et al. (2021) Gaia Collaboration et al., 2021, A&A, 649, A8
* Gaia Collaboration et al. (2022) Gaia Collaboration et al., 2022, arXiv e-prints, p. arXiv:2206.06207
* Geisler et al. (2007) Geisler D., Wallerstein G., Smith V. V., Casetti-Dinescu D. I., 2007, PASP, 119, 939
* Glass & Evans (1981) Glass I. S., Evans T. L., 1981, Nature, 291, 303
* Grady et al. (2019) Grady J., Belokurov V., Evans N. W., 2019, MNRAS, 483, 3022
* Grady et al. (2020) Grady J., Belokurov V., Evans N. W., 2020, MNRAS, 492, 3128
* Groenewegen (2004) Groenewegen M. A. T., 2004, A&A, 425, 595
* Hänninen & Flynn (2002) Hänninen J., Flynn C., 2002, MNRAS, 337, 731
* Höfner & Olofsson (2018) Höfner S., Olofsson H., 2018, A&ARv, 26, 1
* Holl et al. (2018) Holl B., et al., 2018, A&A, 618, A30
* Holmberg et al. (2009) Holmberg J., Nordström B., Andersen J., 2009, A&A, 501, 941
* Huang et al. (2020) Huang C. D., et al., 2020, ApJ, 889, 5
* Hunter (2007) Hunter J. D., 2007, Computing in Science Engineering, 9, 90
* Ishihara et al. (2011) Ishihara D., Kaneda H., Onaka T., Ita Y., Matsuura M., Matsunaga N., 2011, A&A, 534, A79
* Ita & Matsunaga (2011) Ita Y., Matsunaga N., 2011, MNRAS, 412, 2345
* Kallivayalil et al. (2013) Kallivayalil N., van der Marel R. P., Besla G., Anderson J., Alcock C., 2013, ApJ, 764, 161
* Lebzelter et al. (2018) Lebzelter T., Mowlavi N., Marigo P., Pastorelli G., Trabucchi M., Wood P. R., Lecoeur-Taïbi I., 2018, A&A, 616, L13
* Lebzelter et al. (2022) Lebzelter T., et al., 2022, arXiv e-prints, p. arXiv:2206.05745
* Lindegren et al. (2012) Lindegren L., Lammers U., Hobbs D., O’Mullane W., Bastian U., Hernández J., 2012, A&A, 538, A78
* Liu et al. (2015) Liu C., et al., 2015, arXiv e-prints, p. arXiv:1510.06123
* Mackereth et al. (2019) Mackereth J. T., et al., 2019, MNRAS, 489, 176
* Maíz Apellániz (2022) Maíz Apellániz J., 2022, A&A, 657, A130
* Majewski et al. (2017) Majewski S. R., et al., 2017, AJ, 154, 94
* Marigo et al. (2022) Marigo P., et al., 2022, ApJS, 258, 43
* Marín-Franch et al. (2009) Marín-Franch A., et al., 2009, ApJ, 694, 1498
* Martig et al. (2016) Martig M., et al., 2016, MNRAS, 456, 3655
* Masseron & Gilmore (2015) Masseron T., Gilmore G., 2015, MNRAS, 453, 1855
* Matsunaga et al. (2009) Matsunaga N., Kawadu T., Nishiyama S., Nagayama T., Hatano H., Tamura M., Glass I. S., Nagata T., 2009, MNRAS, 399, 1709
* McKinney (2010) McKinney W., 2010, in Stéfan van der Walt Jarrod Millman eds, Proceedings of the 9th Python in Science Conference. pp 56 – 61, doi:10.25080/Majora-92bf1922-00a
* McMillan (2017) McMillan P. J., 2017, MNRAS, 465, 76
* Merrill (1923) Merrill P. W., 1923, ApJ, 58, 215
* Mowlavi et al. (2018) Mowlavi N., et al., 2018, A&A, 618, A58
* Nataf (2016) Nataf D. M., 2016, Publ. Astron. Soc. Australia, 33, e023
* Ortolani et al. (1999) Ortolani S., Barbuy B., Bica E., Renzini A., Marconi G., Gilmozzi R., 1999, A&A, 350, 840
* Pourbaix et al. (2003) Pourbaix D., Platais I., Detournay S., Jorissen A., Knapp G., Makarov V. V., 2003, A&A, 399, 1167
* Price-Whelan et al. (2018) Price-Whelan A. M., et al., 2018, AJ, 156, 123
* Rimoldini et al. (2019) Rimoldini L., et al., 2019, A&A, 625, A97
* Rimoldini et al. (2022) Rimoldini L., et al., 2022, Gaia DR3 documentation Chapter 10: Variability, Gaia DR3 documentation, European Space Agency; Gaia Data Processing and Analysis Consortium. Online, id. 10
* Rowell et al. (2021) Rowell N., et al., 2021, A&A, 649, A11
* Sanders & Binney (2016) Sanders J. L., Binney J., 2016, MNRAS, 457, 2107
* Sanders & Das (2018) Sanders J. L., Das P., 2018, MNRAS, 481, 4093
* Santos & Piatti (2004) Santos J. F. C. J., Piatti A. E., 2004, A&A, 428, 79
* Sharma et al. (2021) Sharma S., et al., 2021, MNRAS, 506, 1761
* Skrutskie et al. (2006) Skrutskie M. F., et al., 2006, AJ, 131, 1163
* Sloan et al. (2010) Sloan G. C., et al., 2010, ApJ, 719, 1274
* Soderblom (2010) Soderblom D. R., 2010, ARA&A, 48, 581
* Spitzer & Schwarzschild (1951) Spitzer Lyman J., Schwarzschild M., 1951, ApJ, 114, 385
* Spitzer & Schwarzschild (1953) Spitzer Lyman J., Schwarzschild M., 1953, ApJ, 118, 106
* Trabucchi & Mowlavi (2022) Trabucchi M., Mowlavi N., 2022, A&A, 658, L1
* Trabucchi et al. (2017) Trabucchi M., Wood P. R., Montalbán J., Marigo P., Pastorelli G., Girardi L., 2017, ApJ, 847, 139
* Trabucchi et al. (2019) Trabucchi M., Wood P. R., Montalbán J., Marigo P., Pastorelli G., Girardi L., 2019, MNRAS, 482, 929
* VandenBerg et al. (2013) VandenBerg D. A., Brogaard K., Leaman R., Casagrande L., 2013, ApJ, 775, 134
* van der Walt et al. (2011) van der Walt S., Colbert S. C., Varoquaux G., 2011, Computing in Science Engineering, 13, 22
* Vasiliev (2019) Vasiliev E., 2019, MNRAS, 482, 1525
* Vasiliev & Baumgardt (2021) Vasiliev E., Baumgardt H., 2021, MNRAS, 505, 5978
* Velazquez & White (1999) Velazquez H., White S. D. M., 1999, MNRAS, 304, 254
* Virtanen et al. (2020) Virtanen P., et al., 2020, Nature Methods, 17, 261
* Wang & Chen (2019) Wang S., Chen X., 2019, ApJ, 877, 116
* Waskom et al. (2017) Waskom M., et al., 2017, seaborn: v0.8.1. Zenodo, doi:10.5281/zenodo.883859
* West et al. (2015) West A. A., Weisenburger K. L., Irwin J., Berta-Thompson Z. K., Charbonneau D., Dittmann J., Pineda J. S., 2015, ApJ, 812, 3
* Whitelock et al. (2003) Whitelock P. A., Feast M. W., van Loon J. T., Zijlstra A. A., 2003, MNRAS, 342, 86
* Whitelock et al. (2008) Whitelock P. A., Feast M. W., Van Leeuwen F., 2008, MNRAS, 386, 313
* Wielen (1977) Wielen R., 1977, A&A, 60, 263
* Wilson & Woolley (1970) Wilson O., Woolley R., 1970, MNRAS, 148, 463
* Wood et al. (1999) Wood P. R., et al., 1999, in Le Bertre T., Lebre A., Waelkens C., eds, IAU Symposium Proceedings Vol. 191, Asymptotic Giant Branch Stars. p. 151
* Wyatt & Cahn (1983) Wyatt S. P., Cahn J. H., 1983, ApJ, 275, 225
* Xiang & Rix (2022) Xiang M., Rix H.-W., 2022, Nature, 603, 599
* Xiang et al. (2017) Xiang M., et al., 2017, ApJS, 232, 2
* Yu & Liu (2018) Yu J., Liu C., 2018, MNRAS, 475, 1093
* Zoccali et al. (2003) Zoccali M., et al., 2003, A&A, 399, 931
|
# Dynamic Price of Parking Service based on Deep Learning
Alejandro Luque-Cerpa Department of Computer Science and Artificial
Intelligence
University of Seville, Seville, Spain<EMAIL_ADDRESS>Miguel A.
Gutiérrez-Naranjo Department of Computer Science and Artificial Intelligence
University of Seville, Seville, Spain<EMAIL_ADDRESS>Miguel Cárdenas-Montes
Department of Fundamental Research, Centro de Investigaciones Energéticas
Medioambientales y Tecnológicas, Madrid, Spain<EMAIL_ADDRESS>
###### Abstract
The improvement of air-quality in urban areas is one of the main concerns of
public government bodies. This concern emerges from the evidence between the
air quality and the public health. Major efforts from government bodies in
this area include monitoring and forecasting systems, banning more pollutant
motor vehicles, and traffic limitations during the periods of low-quality air.
In this work, a proposal for dynamic prices in regulated parking services is
presented. The dynamic prices in parking service must discourage motor
vehicles parking when low-quality episodes are predicted. For this purpose,
diverse deep learning strategies are evaluated. They have in common the use of
collective air-quality measurements for forecasting labels about air quality
in the city. The proposal is evaluated by using economic parameters and deep
learning quality criteria at Madrid (Spain).
###### keywords:
Air Quality, Deep Learning , Convolutional Neural Networks , LSTM , U-Net
## 1 Introduction
Air pollution is one of the most critical health issues in urban areas with a
tough concern for governments and citizens. The scientific literature shows
its relation with the population health [1, 2, 3, 4, 5, 6, 7, 8, 9, 10].
Furthermore, the population growth in urban areas in the forthcoming years
will aggravate this issue.
At this point, a major contributor to the air pollution is the motor vehicles.
German Environment Agency and other studies estimate that road transportation
is responsible for about 60% of emissions of $NO_{2}$ in cities [11, 12]. For
monitoring and mitigating the adverse effect of air pollution, governments
have proposed a wide variety of actions, including: the creation of networks
of monitoring stations in the cities and outskirts, the establishing of
grading protocol scenarios, the speed limit reduction in the accesses to the
city centers, promotion of public transport, the creation of Low Emission
Zones, and prohibition of use of public parking when the air quality falls
below critical levels. In the past, some of these actions have demonstrated
their efficiency [11, 13]. However, to the best of our knowledge, a dynamic
price for public parking service as a function of the pollution level has not
been proposed yet.
For cities where it is illegal to forbid the parking during the low-quality
air episodes, this application aims at discouraging the use of public parking,
and therefore the private vehicles, during these episodes. By dynamically
adjusting the price of public parking during the high air pollution days,
restrictive actions —over the whole city or by district— can be implemented.
In the current work, diverse deep learning algorithms aimed at providing the
necessary support for dynamic pricing of public parking are proposed and
evaluated, namely, different models based on Convolutional Neural Networks,
LSTM layers and U-Time architecture are compared. Our proposals endorse
important features such as a collective behavior —the prediction is based on
measurements of multiple monitoring stations—, and measurability —both from
accuracy of prediction and economic side—. In order to fix a realistic
prediction adapted to the use of motor vehicles, days are divided in four
blocks of 6 hours each one, while the prediction focuses on the block II (from
06:00 to 12:00) and III (from 12:00 to 18:00). The blocks I (from 00:00 to
06:00) and IV (from 18:00 to 24:00) are discarded for this study due to the
low outdoor activities during those periods.
This proposal aligns with the Sustainable Development Goals (SDGs) of the
United Nations and concretely it may act as enabler for the goals Goal 3
”Ensure healthy lives and promote well-being for all at all ages” and
particularly for the Target 3.9. By 2030, substantially reduce the number of
deaths and illnesses from hazardous chemicals and air, water and soil
pollution and contamination” [14].
Concerning the data, it has been extracted from the Air Quality Monitoring
Network of Madrid [15]. This public repository offers hourly and daily data
from more than 24 monitoring stations, including three categories: suburban
(stations in parks in urban areas), traffic (term for stations affected by
traffic and close to a principal street or road), and background (urban
background station affected by both traffic and background pollution). For
this work, 12 monitoring stations have been selected.
The paper is organized as follows: In Section 2 the main background of the
work, including a description of the dataset, the deep architectures used are
presented. The Results and the Analysis are shown in Section 3. Finally, the
Conclusions are presented in Section 4.
## 2 Methods and Materials
### 2.1 Data Acquisition and Preprocessing
The datasets used in this work are obtained from Madrid’s City Council Open
Data website111https://datos.madrid.es/portal/site/egob/. They include daily
and hourly measurements of the concentration of diverse pollutants, such as
$CO$, $NO$ or $CH_{4}$. We will study the hourly data related to $NO_{2}$,
measured in $\mu$g/m3, between 2010/01 and 2019/12. The dataset contains some
missing values due to instrument failure and other reasons. In order to fill
the missing data, they are replaced with the corresponding monthly mean. A
min-max normalization method for scaling the data to the [0,1] range is also
used.
The data are obtained from 24 air quality stations distributed in 5 different
zones. Our data is collected from 12 of these stations: Escuelas Aguirre (EA)
Barrio del Pilar (BP), Plaza del Carmen (PlC), Retiro (Re), Ensanche de
Vallecas (EV), Arturo Soria (AS), Barajas (Ba), Juan Carlos I (PJCI), El Pardo
(ElP), Casa de Campo (CC), Fernández Ladreda (FL) y Farolillo (Faro). The list
of stations includes suburban, traffic, and background (Fig. 1).
Figure 1: Partial map of Madrid. The monitoring stations used in this study
are marked.
Since the goal of this work is to compute a dynamic price for the regulated
parking service during the periods where the main outdoors activities are
undertaken, as pointed above, the day is divided in four blocks: one block for
each quarter of the day. The proposal focuses on the air quality prediction
for the second (6:00-12:00) and third (12:00-18:00) blocks.
Regarding the air-quality levels, Madrid City Council through the Action
protocol for nitrogen dioxide pollution
episodes222https://www.madrid.es/UnidadesDescentralizadas/AreasUrbanas_EducacionAmbiental/
Catalogo/AirQualityPlan2011-15.pdf defines three alert levels:
* 1.
Pre-warning: when in any two stations of the same area, the 180 $\frac{\mu
g}{m^{3}}$ of $NO_{2}$ level is exceeded for two consecutive hours
simultaneously or in any three stations of the surveillance network the 180
$\frac{\mu g}{m^{3}}$ level is exceeded for three consecutive hours
simultaneously.
* 2.
Warning: when in any two stations of the same zone the 200 $\frac{\mu
g}{m^{3}}$ of $NO_{2}$ level is exceeded during two consecutive hours
simultaneously or in any three stations of the surveillance network the 200
$\frac{\mu g}{m^{3}}$ are exceeded during three consecutive hours
simultaneously.
* 3.
Alert: when any three stations in the same zone (or two if it is zone 4)
exceed 400 $\frac{\mu g}{m^{3}}$ for three consecutive hours simultaneously.
Let us remark that these pollution levels are not coherent with the air
quality of a city like Madrid: the 200 $\frac{\mu g}{m^{3}}$ of $NO_{2}$ level
is hardly reached, and the 400 $\frac{\mu g}{m^{3}}$ of $NO_{2}$ level is
never reached. Because of this, two different rules with four alert levels
each of them are proposed in this work. Instead of levels based on absolute
values, our proposal is based on annual percentiles. The first rule is the
following:
Rule I
* 1.
Pre-warning: when the 75th annual percentile is exceeded for three consecutive
hours simultaneously in any three stations of the surveillance network.
* 2.
Warning: when the 90th annual percentile is exceeded for three consecutive
hours simultaneously in any three stations of the surveillance network.
* 3.
Alert: when the 95th annual percentile is exceeded for three consecutive hours
simultaneously in any three stations of the surveillance network.
* 4.
No alert: if none of the previous conditions is satisfied.
Every 6-hours group is classified in one of these four alert levels. Since the
warning and alert level are so similar, it is expected that deep learning
algorithms will hardly differentiate them. Because of this, a second rule
—less stringent— with another four pollution levels are proposed:
Rule II
* 1.
Pre-warning: when the 50th annual percentile is exceeded for three consecutive
hours simultaneously in any three stations of the surveillance network.
* 2.
Warning: when the 75th annual percentile is exceeded for three consecutive
hours simultaneously in any three stations of the surveillance network.
* 3.
Alert: when the 95th annual percentile is exceeded for three consecutive hours
simultaneously in any three stations of the surveillance network.
* 4.
No alert: if none of the previous conditions is satisfied.
### 2.2 Convolutional Neural Networks
Convolutional Neural Networks (CNN) is one of the most popular Deep Learning
architectures [16]. They are designed to deal with pieces of information
placed on a $n$-dimensional grid and, therefore, they are profusely used to
deal with digital images, although nowadays they have applications in a wide
range of research areas including time series analysis [17] or navigation in
indoor environments [18]. Many different architectures have been defined in
the class of CNN, but all of them consist of a sequence of convolutional
layers, which deal with local environments of the grid by using the so-called
kernels together with other kind ways of processing information, which can
involve pooling-based layers, dropout and other regularization techniques.
Although CNN are usually associated with image or audio classification —2D
grid examples—- or video sequence —3D grid examples—, it can also be applied
to univariate time series analysis —1D grid examples— or multivariate time
series —2D grid examples—. The key idea in all these applications is the use
of several convolutional layers in order to abstract local semantic features
by the use of kernels which can be optimized in the training process.
In this paper, our target is to predict the alert level of pollution in a city
from a dataset which includes spatio-temporal information. The data has a
time-line dimension but also a spatial dimension due the geographical
distribution of the weather stations. Such distribution produces some kind of
correlation in the data, since close stations report close pollution levels.
In order to handle such complexity, several Deep Learning architectures have
been considered. Data have been presented as a two-dimensional matrix with one
column for each weather station and a row for each time unit. Since CNN is
adapted to data placed on a grid, one of the considered options has been to
apply this architecture to our 2D matrix.
### 2.3 LSTM
If CNN has become the standard tool for studying images with Deep Learning
techniques, Long short-term memory (LSTM) [19] architecture has become a
standard for the study of sequential information. This architecture belongs to
the so-called gated recurrent neural networks and the main motivation for
using them to create paths along time to avoid the vanishing gradient problem.
For this purpose, instead of considering simple neurons which apply an
activation function to the affine transformation of inputs, LSTM architectures
consider the so-called LSTM cells which have internal recurrence in addition
to the outer recurrence of the recurrent neural network.
The key idea on these LSTM cells is to keep information in a cell state for
later, preventing older signals from vanishing during processing. LSTM systems
can add or remove information of such cell states by structures called gates.
In LSTM, there are three types of gates: input gate, forget gate and output
gate. In 2014, Cho et al. [20] presented a new type of gated recurrent neural
network with fewer parameters than LSTM, as it lacks an output gate.
As pointed above, collected pollution data are sequentially grouped by time
units. Such a temporal dimension has been exploited in our analysis of the
data by using a LSTM architecture.
### 2.4 U-Time
In 2015, Ronneberger al [21] presented a new Deep Learning architecture for
biomedical image segmentation called U-net. The networks with this
architecture are basically CNN endowed with new features. Technically, a
contracting path and an expansive path are considered, which give to the model
the U-shaped architecture. The contracting path is a standard CNN which
includes convolutions layers, ReLU as activation functions and max pooling
operations. As usual, in the convolutional layers, the spatial information is
reduced while feature information is increased. The original part of the U-net
architecture is the expansive pathway which combines the feature and spatial
information through a sequence of up-convolutions and concatenations with
high-resolution features from the contracting path.
In 2019, Perslev et al. [22] presented a fully convolutional encoder-decoder
network called U-Time inspired by the U-Net architecture. U-Time adopts basic
concepts from U-Net for 1D time-series segmentation by mapping a whole
sequence to a dense segmentation in a single forward pass. Originally, U-Time
was used for studying time series segmentation applied to sleep staging. In
this paper, U-time is used as inspiration for our study of the air-quality in
Madrid. In a similar way to the study based on simple CNN architectures, the
U-net study also exploits the 2D representation of the data on space and time.
## 3 Results and Analysis
Since the goal of this paper is to predict the pollution level of six-hours
blocks knowing the previous pollution levels, different models and initial
conditions are considered. As pointed above, two different rules of pollution
alert levels are considered: Rule I for the 95th, 90th and 75th percentiles,
and Rule II for the 95th, 75th and 50th percentiles. Three distinct sizes of
input data are considered: one day of data per station (24x12 values), three
days of data per station (72x12 values) and a week of data per station (168x12
values).
Some considerations must be taken into account. Firstly, the classes are
highly unbalanced: most of the days, the pollution alert level is null.
Secondly, the 95th, 90th and 75th percentiles are closer together than the
95th, 75th and 50th percentiles, so we expect the models to obtain worse
results for the first pollution alert levels than for the second ones.
Finally, it is better to obtain false positives than false negatives, given
that the ultimate objective is reducing pollution.
### 3.1 Confusion Matrix based Analysis
In this work, four different pollution alert levels are defined: Pre-Warning,
Warning, Alert or No-alert. To correctly evaluate the performance of a model
predicting the pollution alert level, it is necessary to obtain the accuracy
of such model for each of these levels. The results are presented using
confusion matrices, where the performance of the different models can be
correctly visualized, and more detailed analysis is allowed. Each row of a
matrix represents the instances in an actual class, while each column
represents the instances in a predicted class.
#### 3.1.1 LSTM
In this first approach, a model based on LSTM cells is considered, namely a
model with two LSTM layers, the first one with 50 neurons and the second one
with 10 neurons. After these layers a Dropout layer [23] with index 0.05 was
included. Finally, an output layer with four neurons and softmax as activation
function is added. Cross-entropy loss has been used for training. After
training the LSTM with different data blocks, input sizes and pollution
levels, the confusion matrices are obtained. Let us remark that the classes
have been balanced before training in order to avoid overfitting.
As it is shown in Figs. 2a, 2b, and 2c; and in Figs. 4a, 4b, and 4c for the
normalized confusion matrices, for the Rule I and for the second block
(6:00-12:00), positive results are achieved. There is a significant difference
between using 24 hours, 72 hours or 168 hours of input data, obtaining better
results when using 72 hours of input data, the accuracy ranges from 63% to
80%; and when using 168 hours of input data, the accuracy ranges from 53% to
83%. In both cases, the adjacent errors —labels erroneously predicted as a
contiguous label— and the non-adjacent errors —labels predicted as a non-
contiguous label— are high. They rise up to 26%.
In a similar way, as detailed in Figs. 2d, 2e, and 2f; and in Figs. 4d, 4e,
and 4ffor the normalized confusion matrices for the Rule I and for the third
block (12:00-18:00), the results are confusing, without a obvious best
configuration, with low accuracy in the main diagonal, and large adjacent and
non-adjacent errors. It is important to remark that for both blocks the
results are not concentrated around the main diagonal, but scattered. This is
probably due to the closeness of the percentiles values, and as a consequence
of the imbalanced labels.
In the case of the pollution alert levels for the Rule II, as illustrated in
Figs. 3a, 3b, and 3c for the second block, and Figs. 3d, 3e, and 3f for the
third block; and for the normalized confusion matrices in Figs. 5a, 5b, and 5c
for the second block, and Figs. 5d, 5e, and 5f for the third block, the
results for both blocks are better than those obtained for Rule I. This stems
from the less imbalance labels under the Rule II percentiles. For the second
block and the third block, it is difficult to ascertain which configuration
produces the best results. For the second block, the accurate results —main
diagonal— range from 74% to 89% for the 72 hours configuration to the range
from 64% to 92% for the 128 hours configuration. For the third block, the
accurate results overlap: from 69% to 83% when using the configuration of 24
hours, from 49% to 91% when using the configuration of 72 hours, and from 69%
to 87% when using the configuration of 168 hours. At the same time large
adjacent errors are also produced, for instance the 28% of Pre-Warning are
incorrectly predicted as No alert when using the configuration of 72 hours.
(a) Second block. 24 hours.
(b) Second block. 72 hours.
(c) Second block. 168 hours.
(d) Third block. 24 hours.
(e) Third block. 72 hours.
(f) Third block. 168 hours.
Figure 2: Confusion matrices obtained from the LSTM model for the Rule I.
Three input sizes are considered: 24 hours, 72 hours and 168 hours of data per
station. Two blocks are considered: the second block (6:00-12:00) and the
third block (12:00-18:00). Each row represents the actual class and each
column represents the predicted class.
(a) Second block. 24 hours.
(b) Second block. 72 hours.
(c) Second block. 168 hours.
(d) Third block. 24 hours.
(e) Third block. 72 hours.
(f) Third block. 168 hours.
Figure 3: Confusion matrices obtained from the LSTM model for the Rule II.
Three input sizes are considered: 24 hours, 72 hours and 168 hours of data per
station. Two blocks are considered: the second block (6:00-12:00) and the
third block (12:00-18:00). Each row represents the actual class and each
column represents the predicted class.
(a) Second block. 24 hours.
(b) Second block. 72 hours.
(c) Second block. 168 hours.
(d) Third block. 24 hours.
(e) Third block. 72 hours.
(f) Third block. 168 hours.
Figure 4: Normalized confusion matrices obtained from the LSTM model for the
Rule I. Three input sizes are considered: 24 hours, 72 hours and 168 hours of
data per station. Two blocks are considered: the second block (6:00-12:00) and
the third block (12:00-18:00). Each row represents the actual class and each
column represents the predicted class.
(a) Second block. 24 hours.
(b) Second block. 72 hours.
(c) Second block. 168 hours.
(d) Third block. 24 hours.
(e) Third block. 72 hours.
(f) Third block. 168 hours.
Figure 5: Normalized confusion matrices obtained from the LSTM model for the
Rule II. Three input sizes are considered: 24 hours, 72 hours and 168 hours of
data per station. Two blocks are considered: the second block (6:00-12:00) and
the third block (12:00-18:00). Each row represents the actual class and each
column represents the predicted class.
#### 3.1.2 CNN
In this approach, a model of CNN networks is considered. The architecture of
the model can be summarized as follows: (1) a first convolutional layer with
16 kernels, kernel size 5 and ReLU as activation function; (2) a max-pooling
layer; (3) a second convolutional layer with 64 kernels, kernel size 3 and
ReLU as activation function; (4) another max-pooling layer; (5) A dense layer
with 20 neurons and ReLU as activation function; (6) a dropout layer with
index 0.05; (7) a final dense output layer with 20 neurons and softmax as
activation function. Cross-entropy loss has been used for training. The
confusion matrices are shown in Figs. 6, and 7; while the next comparisons are
undertaken by the normalized ones.
Following the same line of experiments, the CNN model is trained for the same
conditions that are established in the case of the LSTM model. Comparing Figs.
4a, 4b, 4c and Figs. 8a, 8b, 8c, it is shown that better results are obtained
in the case of the CNN than in the LSTM one for the second block (6:00-12:00)
and the Rule I. When using CNN model, there is a significant difference
between using 24, 72 or 168 hours of input data, obtaining better results for
greater amounts of hours. For 168 hours of input data, the accuracy ranges
from 73% to 94%. For 71 hours, the accuracy has a higher inferior accuracy
74%, but also exhibits a much lower higher accuracy 86%. For the three
configurations of number of hours, CNN outperforms the equivalent LSTM model
for the second block and the Rule I.
Comparing now Figs. 4d, 4e, 4f and Figs. 8d, 8e, 8f, we obtain considerably
better results for the Rule I and for the third block (12:00-18:00) from the
CNN model. As for the second block, there is a significant difference between
using 24, 72 or 168 hours of input data, obtaining better results for those
configurations with larger amounts of hours. The accuracy for 168 hours
configuration ranges from 80% to 98%. Besides, the CNN model shows the lowest
high-adjacent errors and non-adjacent errors than their equivalents in the
LSTM model.
Considering now the Rule II and second block, the results of LSTM (Figs. 5a,
5b, 5c) and CNN (Figs. 9a, 9b, 9c) models are difficult to ascertain a best
configuration and model. Ranges of accuracy overlapping, rawly being better
the high accuracy of CNN model —up to 95% for the configuration of 168 hours—,
but being in many cases the low accuracy worse than their counterpart
configuration at LSTM model. In this case, the exploration of diverse models
takes a great value, and the evaluation of an ensemble model could be proposed
as future work. The lowest high-errors, both adjacent and non-adjacent, are
achieved for the configuration of 168 hours, being respectively 11% for the
Alert labels predicted as Warning, and 2% for two cases.
Finally, for the Rule II and for the third block (12:00-18:00), we can observe
that the CNN model (Figs. 9d, 9e, 9f) obtains again better results than the
LSTM model (Figs. 5d, 5e, 5f). The accuracy for the CNN model for 72 hours
configuration ranges from 82% to 93%, at the same time that it keeps low the
highest errors: adjacent and non-adjacent.
(a) Second block. 24 hours.
(b) Second block. 72 hours.
(c) Second block. 168 hours.
(d) Third block. 24 hours.
(e) Third block. 72 hours.
(f) Third block. 168 hours.
Figure 6: Confusion matrices obtained from the CNN model for the Rule I. Three
input sizes are considered: 24 hours, 72 hours and 168 hours of data per
station. Two blocks are considered: the second block (6:00-12:00) and the
third block (12:00-18:00). Each row represents the actual class and each
column represents the predicted class.
(a) Second block. 24 hours.
(b) Second block. 72 hours.
(c) Second block. 168 hours.
(d) Third block. 24 hours.
(e) Third block. 72 hours.
(f) Third block. 168 hours.
Figure 7: Confusion matrices obtained from the CNN model for the Rule II.
Three input sizes are considered: 24 hours, 72 hours and 168 hours of data per
station. Two blocks are considered: the second block (6:00-12:00) and the
third block (12:00-18:00). Each row represents the actual class and each
column represents the predicted class.
(a) Second block. 24 hours.
(b) Second block. 72 hours.
(c) Second block. 168 hours.
(d) Third block. 24 hours.
(e) Third block. 72 hours.
(f) Third block. 168 hours.
Figure 8: Normalized confusion matrices obtained from the CNN model for the
Rule I and the Rule II. Three input sizes are considered: 24 hours, 72 hours
and 168 hours of data per station. Two blocks are considered: the second block
(6:00-12:00) and the third block (12:00-18:00). Each row represents the actual
class and each column represents the predicted class.
(a) Second block. 24 hours.
(b) Second block. 72 hours.
(c) Second block. 168 hours.
(d) Third block. 24 hours.
(e) Third block. 72 hours.
(f) Third block. 168 hours.
Figure 9: Normalized confusion matrices obtained from the CNN model for the
Rule II. Three input sizes are considered: 24 hours, 72 hours and 168 hours of
data per station. Two blocks are considered: the second block (6:00-12:00) and
the third block (12:00-18:00). Each row represents the actual class and each
column represents the predicted class.
#### 3.1.3 U-Time
Due to the structural needs of U-Time, we decide to only use the configuration
of 168 hours of input data for all the experiments performed with this model
(Figs. 10 and 11). U-Time is a deep autoencoder that reduces 1.920 values of
the input data into one single value of the latent vector. If we want to
benefit from the properties of U-Time, the input size must be as large as
possible. Shorter configurations are unable to pass through the U-Time model.
By analysing the results obtained from the U-Time model for the second and the
third blocks, and for the Rule I and the Rule II, it can be observed that they
are worse than those obtained from the other models. The accuracy of the
U-Time model ranges from 43% to 74% for the second block and the Rule I (Fig.
10c), from 34% to 85% for the third block and the Rule I (Fig. 10d), from 46%
to 89% for the second block and the Rule II (Fig. 11c), and from 24% to 91%
for the third block and the Rule II (Fig. 11d). The predicted values are also
scattered and not concentrated around the main diagonal. Adjacent errors grow
up to 41% for the Pre-Warning labels predicted as No-alert (Fig. 11d). And for
the non-adjacent errors, they rise up to 36% for the pre-warning labels
predicted as Alert (Fig. 10d). The CNN model clearly outperforms the U-Time
model for both blocks and rules.
(a) Second block. 168 hours.
(b) Third block. 168 hours.
(c) Second block. 168 hours. Normalized.
(d) Third block. 168 hours. Normalized.
Figure 10: Confusion matrices obtained from the U-Time model for the Rule I.
Only one input size is considered: 168 hours of data per station. Two blocks
are considered: the second block (6:00-12:00) and the third block
(12:00-18:00). Each row represents the actual class and each column represents
the predicted class.
(a) Second block. 168 hours.
(b) Third block. 168 hours.
(c) Second block. 168 hours. Normalized.
(d) Third block. 168 hours. Normalized.
Figure 11: Confusion matrices obtained from the U-Time model for the Rule II.
Only one input size is considered: 168 hours of data per station. Two blocks
are considered: the second block (6:00-12:00) and the third block
(12:00-18:00). Each row represents the actual class and each column represents
the predicted class.
### 3.2 Economic Analyses
Depending on the error direction, the parking regulatory system could
overcharge over the customers. This scenario corresponds when the predicted
labels have a higher criticality than the final true labels. Oppositely, when
the predicted labels have a lower criticality that the true labels two
pernicious effects appear. On the one hand, the charge for the use of the
parking is undercharged; and on the other hand, the deterrent effect over the
use of private motor vehicles vanish.
An additional consideration is the fairness of the prediction. A well-balanced
deep learning architecture with null economic impact in a temporal period, but
with large errors in both directions is not acceptable. This kind of error
leads to distrust of the citizens on IA applications. For this reason, a rate
for the hours of each criticality level is proposed, the application of this
rate is evaluated. The evaluation is undertaken by adding the incorrect charge
in each direction of the errors and then their absolute values added.
The prices per hour established are for the criticality levels area: 0.40,
0.60, 1.20 and 2.40 euros. Currently the price per one hour of public parking
at Madrid is 0.40 euros. In our proposal, this price is increased by 50% when
the criticality level goes up from No-alert to Pre-Warning, and then doubled
for the two next levels: Warning and Alert. The aim of doubling the price for
the two last criticality levels is to simulate the deterrent effect in the use
of private transport.
Then the economic impact —per 6 hours block— in each direction is evaluated
and their absolute values added (Tables 1 and 2). The lower the values, the
fairness of the architecture. Except for the case of the prediction for 1 day
of the criticality level in Rule II and Block II, the CNN-based architecture
produces the lowest errors and the highest fairness among the architectures
tested.
Table 1: Evaluation of the deep architectures for block II based on economic criterion. | | CNN | LSTM | UTime
---|---|---|---|---
Rule I | 1d | 872.2 | 1044.2 | NaN
3d | 354.8 | 808.6 | NaN
7d | 337.6 | 1006.6 | 816.0
Rule II | 1d | 837.8 | 708.2 | NaN
3d | 566.4 | 662.6 | NaN
7d | 493.2 | 833.4 | 1583.2
Table 2: Evaluation of the deep architectures for block III based on economic criterion. | | CNN | LSTM | UTime
---|---|---|---|---
Rule I | 1d | 246.8 | 522.8 | NaN
3d | 174.0 | 643.8 | NaN
7d | 114.0 | 593.2 | 1213.2
Rule II | 1d | 306.4 | 393.0 | NaN
3d | 187.0 | 501.8 | NaN
7d | 202.6 | 362.4 | 660.2
## 4 Conclusions
In summary, air pollution is a critical problem in densely populated areas.
Some measures have been implemented, like traffic limitations during periods
of low-quality air or forecasting systems. In our research, we conclude that
deep learning models can be useful to implement another kind of measure: some
dynamic regulated parking services that would discourage motor vehicles
parking when low-quality episodes are predicted. To achieve this objective,
three different proposals are considered: one based on LSTM, one based on CNN
and one based on the U-Time model.
While U-Time was a promising architecture, the results obtained were less than
ideal. Due to its size, the computation time needed is between three and four
times the computation time that the LSTM or the CNN models need, and the CNN
model outperforms both of them. Analysing the results obtained by these
models, we can draw some conclusions. First, it is better to use the
percentile levels 95th, 75th and 50th than the 95th, 90th and 75th. This must
be because of the fact that the last percentiles are closer together than the
first ones. Second, there is a significant difference between the results
obtained with the selected input sizes: usually it is better to use greater
input sizes. Third, it is a good sign that the amount of false negatives is
lesser than the amount of false positives. This is important for us due to the
nature of our proposal: it is more desirable to have less motor vehicles in a
regular episode, than more motor vehicles in a low-quality episode.
## 5 Future Work
In this work, the input data is taken from 12 of the 24 different air quality
stations. There is a simple reason for this: only these 12 stations have been
extracting data since 2010/01. Once there is enough data, it would be
interesting to train again the models with data taken from the 24 air quality
stations. In the case of the U-Time model, it is possible that, once there is
more data, the model could be deeper and consequently get better results.
We have decided to use three different sizes of input data: 24 hours, 72 hours
and 168 hours. It would be interesting to use greater sizes of input data,
such as 336 hours (two weeks) or even 672 hours (a month). With greater input
sizes it would be coherent to also train models with more parameters, and
because of this it would be reasonable to train CNN or LSTM models with more
layers; more filters or bigger kernels in the case of the CNN models; or more
neurons per layer in the case of the LSTM model.
Considering our goal, we have proposed a dynamic parking regulatory system
considering both the fairness and the deterrent effect in the use of private
transport, and we have defined four different alert levels based on
percentiles: No-alert, Pre-Warning, Warning and Alert. It would be interesting
to work with expert environmental scientists that could establish more useful
and realistic thresholds and pollution alert levels.
## Acknowledgements
MCM is co-funded by the Spanish Ministry of Science and Innovation for funding
support through the grant PID2020-113807RA-I00 ”SERVICIOS INNOVADORES DE
ANALISIS DE DATOS PARA EL EXPERIMENTO CMS”. MAGN is partially supported by the
Spanish Ministry of Science and Innovation through the project
PID2019-107339GB-I00 ”Advances in Computational Topology and Applications”.
## References
* [1] Linares, C., Díaz, J., Tobías, A., De Miguel, J.M., Otero, A., 2006. Impact of urban air pollutants and noise levels over daily hospital admissions in children in Madrid: a time series analysis. Int. Arch. Occup. Environ. Health 79(2), 143–152. https://doi.org/10.1007/s00420-005-0032-0
* [2] Díaz, J., García, R., Ribera, P., Alberdi, J.C., Hernández, E., Pajares, M.S., Otero, A., 1999. Modeling of air pollution and its relationship with mortality and morbidity in Madrid, Spain. Int. Arch. Occup. Environ. Health 72(6), 366–376. https://doi.org/10.1007/s004200050388
* [3] Alberdi Odriozola, J.C., Díaz Jiménez, J., Montero Rubio, J.C., Mirón Pérez, I.J., Pajares Ortíz, M.S., Ribera Rodrigues, P., 1998. Air pollution and mortality in Madrid, Spain: a time-series analysis. Int. Arch. Occup. Environ. Health 71(8) 543–549. https://doi.org/10.1007/s004200050321
* [4] Nel, A., 2005. Air Pollution-Related Illness: Effects of Particles, Science 308(5723) 804–806. https://doi.org/10.1126/science.1108752
* [5] Pope, C.A., Dockery, D.W., 2006. Health effects of fine particulate air pollution: lines that connect, J Air Waste Manag Assoc. 56(6) 709–742 https://doi.org/doi:10.1080/10473289.2006.10464485
* [6] Rückerl, R., Schneider, A., Breitner, S., Cyrys, J., Peters, A., 2011. Health effects of particulate air pollution: A review of epidemiological evidence, Inhal Toxicol. 23(10) 555–592 https://doi.org/10.3109/08958378.2011.593587
* [7] Boogaard, H., Erp, A.M., 2019. Assessing health effects of air quality actions: what’s next?, Lancet Planet Health 4(1) e4–e5 https://doi.org/10.1016/S2468-2667(18)30235-4
* [8] Balakrishnan, K. et al., 2019. The impact of air pollution on deaths, disease burden, and life expectancy across the states of India: the Global Burden of Disease Study 2017, Lancet Planet Health 3(1) e26–e39. https://doi.org/10.1016/S2542-5196(18)30261-4
* [9] Rivas, E., Santiago, J.L., Lechón, Y., Martín, F., Ariño, A., Pons, J.J., Santamaría, J.M., 2019. CFD modelling of air quality in Pamplona City (Spain): Assessment, stations spatial representativeness and health impacts valuation, Sci Total Environ 649. https://doi.org/10.1016/j.scitotenv.2018.08.315
* [10] Santamaría, J.M. et al. 2017\. Reduction of exposure of cyclists to urban air pollution, 2018. Pamplona: Programa LIFE13 ENV/ES/000417. ISBN: 978-84-947947-7-3. http://liferespira.com/libro_life_respira_ING.pdf
* [11] Borge, R., et al. 2018 Application of a short term air quality action plan in Madrid (Spain) under a high-pollution episode - Part I: Diagnostic and analysis from observations, Sci. Total Environ. 1561-1573 https://doi.org/10.1016/j.scitotenv.2018.03.149
* [12] Bannon, E., 2019. Low-Emission Zones are a success - but they must now move to zero-emission mobility, Transport and Environment September 2019.
* [13] Cárdenas-Montes, M., Evaluation of the Impact of Low-Emission Zone: Madrid Central as a Case Study, 2012.13782, 2020\.
* [14] UN General Assembly (UNGA). A/RES/70/1Transforming our world: the 2030 Agenda for Sustainable Development. Resolut 25, 1–35, 2015.
* [15] Open data Madrid, 2020, https://datos.madrid.es/portal/site/egob
* [16] Goodfellow, I., Bengio, Y. Courville, A. 2016 Deep Learning, MIT Press, http://www.deeplearningbook.org
* [17] Cárdenas-Montes, M. 2019 Forecast Daily Air-Pollution Time Series with Deep Learning. In: Pérez García H., Sánchez González L., Castejón Limas M., Quintián Pardo H., Corchado Rodríguez E. (eds) Hybrid Artificial Intelligent Systems. HAIS 2019. Lecture Notes in Computer Science, vol 11734. Springer, Cham. https://doi.org/10.1007/978-3-030-29859-3_37 2019
* [18] Muratov, M., and Kaur, S., and Szpakowicz, M. 2020 Convolutional Neural Networks Towards Arduino Navigation of Indoor Environments, arxiv preprint arXiv:2011.13893.
* [19] Hochreiter, S., Schmidhuber J. 1997 Long short-term memory. Neural Comput. 9(8), 1735-1780 https://doi.org/10.1162/neco.1997.9.8.1735
* [20] Cho, K. et al. 2014 Learning Phrase Representations using RNN Encoder-Decoder for Statistical Machine Translation. arXiv:1406.1078.
* [21] Ronneberger, O., Fischer, P., Brox, T. 2015 U-Net: Convolutional Networks for Biomedical Image Segmentation. arXiv:1505.04597
* [22] Perslev, M.. et al 2019 U-Time: A Fully Convolutional Network for Time Series Segmentation Applied to Sleep Staging, arxiv 1910.11162
* [23] Srivastava, N. et al. 2014 Dropout: A Simple Way to Prevent Neural Networks from Overfitting, J Mach Learn Res, 15(56), pp.: 1929-1958
|
# On Gravitational Stefan-Boltzmann Law and Casimir Effect in FRW Universe
A. F. Santos<EMAIL_ADDRESS>Instituto de Física,
Universidade Federal de Mato Grosso,
78060-900, Cuiabá, Mato Grosso, Brazil S. C. Ulhoa<EMAIL_ADDRESS>International Center of Physics, Instituto de Física, Universidade de
Brasília, 70910-900, Brasília, DF, Brazil Canadian Quantum Research Center,
204-3002 32 Ave Vernon, BC V1T 2L7 Canada E. P. Spaniol<EMAIL_ADDRESS>UDF Centro Universitário and Centro Universitário de Brasília UniCEUB,
Brasília, DF, Brazil. Faqir C. Khanna111Professor Emeritus - Physics
Department, Theoretical Physics Institute, University of Alberta
Edmonton, Alberta, Canada<EMAIL_ADDRESS>Department of Physics and Astronomy,
University of Victoria,
3800 Finnerty Road Victoria, BC, Canada
###### Abstract
Both Stefan-Boltzmann law and the Casimir effect, in a universe described by
the FRW metric with zero curvature, are calculated. These effects are
described by Thermo Field Dynamics (TFD). The gravitational energy-momentum
tensor is defined in the context of Teleparallel Equivalent to General
Relativity (TEGR). Each of the two effects gives a consistent prediction with
what is observed on a cosmological scale. One of the effect establishes a
minimum range for the deceleration parameter. While another leads to the
conclusion that a possible cosmological constant has a very small order of
magnitude.
## I Introduction
The introduction of temperature in the gravitational field has been
successfully implemented recently gravTFD . Thermo Field Dynamics (TFD) was
used for this purpose which is an approach that allows both a temporal
evolution of the field at finite temperature. It is an advantage over the
historical approach that associates time with temperature matsubara . A
gravitational field at finite temperature is a theory of quantum gravity since
TFD uses creation and annihilation operators. The field propagator is the
fundamental entity of the thermalization process. It is interesting to note
that such an approach associates the temperature with space such that a
universe with zero temperature will not be expected flatTFD . Absolute zero
temperature will not be natural even in flat space. Within the scope of TFD,
there is a topological structure that allows treating effects such as diverse
as the Stefan-Boltzmann law and the Casimir effect on an equal footing. An
area of intense investigation into the implications of various aspects of
quantum gravity is black hole thermodynamics. Whether in the investigation of
entropy of black holes, or in the understanding of the information paradox. It
was recently investigated how the evaporation process of a black hole
generates an entanglement between quantum fields and geometry, this yields a
modified Page curve that can have implications for several theories of quantum
gravity 4 . It has also been shown that the structure of TFD plays a key role
in this approach 5 . The TFD appears to be a promising theory of quantum
gravity.
It is necessary to thermalize the energy-momentum tensor of the field in
addition to a propagator. The standard model of gravitation is problematic. In
the construction of gravity at finite temperature, an alternative theory of
gravitation is used, Teleparallelism Equivalent to General Relativity (TEGR)
maluf . In TEGR the problem of gravitational energy is well established, as
well as other conserved quantities. As a result, gravitational entropy is
introduced as a direct consequence of Maxwell’s relationships involving
gravitational pressure. Normally this gravitational entropy may be seen as a
fundamental quantity when made equal to Hawking’s expression induces a
temperature of the black hole event horizon different from that commonly
accepted. The whole space-time has a finite temperature, not just the event
horizon of a black hole. Then there is a smooth transition from singularity to
infinity entropy . We must note that TEGR is a formulation of gravitation that
takes into account local Lorentz’s symmetry, such a dependence appears in the
field equations that are entirely equivalent to Einstein’s equations. On the
other hand, recently, proposals have emerged that attribute the local Lorentz
symmetry to the spin connection 7 ; 8 ; 9 , this line of investigation has
received some criticism and in our opinion still requires further
investigation 10 . In TEGR the conserved quantities are sensitive to the
global Lorentz transformations and that is the limit of our approach.
One of the major problems in cosmology is why there is an accelerated
expansion of the universe. Usually the explanation given is an exotic energy
known as dark energy. On the other hand, instead of looking for candidates for
such energy, alternative explanations can be tried. This last chain of thought
will be used. The more interesting features of the universe are analyzed.
There is a non-zero temperature other than zero even at the most distant point
in interstellar space. In addition, it has an observable dynamic horizon
increasing with time. Such a horizon works as a causal barrier to events
within it. Mainly this system behaves like a spherical Casimir effect. There
appears to be two associated phenomena observed in the universe: i) a thermal
radiation like Stefan-Boltzmann’s law and ii) a force a la Casimir effect
responsible for an accelerated expansion of the system. This leads us to
consider that gravitation at finite temperature explains such a phenomena.
This hypothesis is explored here.
This article is divided as follows. In section II TFD is introduced briefly.
In section III the TEGR is presented and the thermal expressions are
calculated. In section IV the energy-momentum tensor at finite temperature is
applied to the FRW universe. With this both the Stefan-Boltzmann law and the
Casimir effect for a zero curvature in the metric are calculated. Finally
conclusions are presented in the last section.
## II Thermo Field Dynamics (TFD)
A quantum field theory at finite temperature is developed by two distinct, but
equivalent, approaches: (i) the imaginary time formalism matsubara and (ii)
the real time formalism Schwinger ; Umezawa1 ; Umezawa2 ; Umezawa22 ; Khanna1
; Khanna2 . TFD is a real-time finite temperature formalism. The temperature
dependent vacuum is defined such that the vacuum expectation value of an
arbitrary operator $\displaystyle A$ agrees with the statistical average,
i.e.,
$\displaystyle\displaystyle\langle A\rangle=\langle
0(\beta)|A|0(\beta)\rangle,$ (1)
where $\displaystyle|0(\beta)\rangle$ is the thermal vacuum and
$\displaystyle\beta=\frac{1}{k_{B}T}$, with $\displaystyle T$ being the
temperature and $\displaystyle k_{B}$ the Boltzmann constant. To construct
this thermal state two elements are necessary: the doubling of the original
Hilbert space and the Bogoliubov transformation. This doubling is defined by
$\displaystyle{\cal S}_{T}={\cal S}\otimes\tilde{\cal S}$, where
$\displaystyle{\cal S}$ is the Hilbert space and $\displaystyle\tilde{\cal S}$
is the dual (tilde) space. The map between the non-tilde
$\displaystyle{A_{i}}$ and tilde $\displaystyle\tilde{A_{i}}$ operators is
given by tilde (or dual) conjugation rules. These rules are
$\displaystyle\displaystyle(A_{i}A_{j})^{\thicksim}$
$\displaystyle\displaystyle=$
$\displaystyle\displaystyle\tilde{A_{i}}\tilde{A_{j}},$
$\displaystyle\displaystyle(cA_{i}+A_{j})^{\thicksim}$
$\displaystyle\displaystyle=$ $\displaystyle\displaystyle
c^{*}\tilde{A_{i}}+\tilde{A_{j}},$
$\displaystyle\displaystyle(A_{i}^{\dagger})^{\thicksim}$
$\displaystyle\displaystyle=$
$\displaystyle\displaystyle\tilde{A_{i}}^{\dagger},$
$\displaystyle\displaystyle(\tilde{A_{i}})^{\thicksim}$
$\displaystyle\displaystyle=$ $\displaystyle\displaystyle-\xi A_{i},$ (2)
with $\displaystyle\xi=-1(+1)$ for bosons (fermions). In addition, the tilde
conjugation rules associate each operator in $\displaystyle{\cal S}$ to two
operators in $\displaystyle{\cal S}_{T}$. Considering $\displaystyle a$ as an
operator leads to
$\displaystyle\displaystyle A=a\otimes
1,\quad\quad\quad\quad\tilde{A}=1\otimes a.$ (3)
TFD and Bogoliubov transformations introduce thermal effects through a
rotation between tilde ($\displaystyle\tilde{\cal S}$) and non-tilde
($\displaystyle{\cal S}$) operators. With an arbitrary operator
$\displaystyle{\cal O}$, the Bogoliubov transformation is defined as
$\displaystyle\displaystyle\left(\begin{array}[]{cc}{\cal O}(k,\alpha)\\\
\xi\tilde{\cal O}^{\dagger}(k,\alpha)\end{array}\right)={\cal
B}(\alpha)\left(\begin{array}[]{cc}{\cal O}(k)\\\ \xi\tilde{\cal
O}^{\dagger}(k)\end{array}\right),$ (8)
where the $\displaystyle\alpha$ is called the compactification parameter
defined by $\displaystyle\alpha=(\alpha_{0},\alpha_{1},\cdots\alpha_{D-1})$
and $\displaystyle{\cal B}(\alpha)$ is
$\displaystyle\displaystyle{\cal
B}(\alpha)=\left(\begin{array}[]{cc}u(\alpha)&-w(\alpha)\\\ \xi
w(\alpha)&u(\alpha)\end{array}\right),$ (11)
with $\displaystyle u^{2}(\alpha)+\xi w^{2}(\alpha)=1$. These quantities
$\displaystyle u(\alpha)$ and $\displaystyle w(\alpha)$ are related to the
Bose distribution. For the case $\displaystyle\alpha_{0}\equiv\beta$ and
$\displaystyle\alpha_{1},\cdots\alpha_{D-1}=0$, the temperature effect is
introduced. Using such formalism, a topological quantum field theory is
considered. A topology
$\displaystyle\Gamma_{D}^{d}=(\mathbb{S}^{1})^{d}\times\mathbb{R}^{D-d}$ with
$\displaystyle 1\leq d\leq D$ is used. Here $\displaystyle D$ is the space-
time dimensions and $\displaystyle d$ is the number of compactified
dimensions. Any set of dimensions of the manifold
$\displaystyle\mathbb{R}^{D}$ can be compactified.
In the TFD formalism, all propagators are written in terms of the
compactification parameter $\displaystyle\alpha$. Here the scalar field
propagator is defined as
$\displaystyle\displaystyle G_{0}^{(AB)}(x-x^{\prime};\alpha)=i\langle
0,\tilde{0}|\tau[\phi^{A}(x;\alpha)\phi^{B}(x^{\prime};\alpha)]|0,\tilde{0}\rangle,$
(12)
where $\displaystyle\tau$ is the time ordering operator and $\displaystyle
A\,\mathrm{and}\,B=1,2$. The Bogoliubov transformation is used to write as
$\displaystyle\displaystyle\phi(x;\alpha)$ $\displaystyle\displaystyle=$
$\displaystyle\displaystyle{\cal B}(\alpha)\phi(x){\cal B}^{-1}(\alpha).$ (13)
In the thermal vacuum, which is defined as
$\displaystyle|0(\alpha)\rangle={\cal U}(\alpha)|0,\tilde{0}\rangle$, the
propagator becomes
$\displaystyle\displaystyle G_{0}^{(AB)}(x-x^{\prime};\alpha)$
$\displaystyle\displaystyle=$ $\displaystyle\displaystyle i\langle
0(\alpha)|\tau[\phi^{A}(x)\phi^{B}(x^{\prime})]|0(\alpha)\rangle,$ (14)
$\displaystyle\displaystyle=$ $\displaystyle\displaystyle
i\int\frac{d^{4}k}{(2\pi)^{4}}e^{-ik(x-x^{\prime})}G_{0}^{(AB)}(k;\alpha),$
where
$\displaystyle\displaystyle G_{0}^{(AB)}(k;\alpha)={\cal
B}^{-1}(\alpha)G_{0}^{(AB)}(k){\cal B}(\alpha),$ (15)
with
$\displaystyle\displaystyle
G_{0}^{(AB)}(k)=\left(\begin{array}[]{cc}G_{0}(k)&0\\\ 0&\xi
G^{*}_{0}(k)\end{array}\right),$ (18)
and
$\displaystyle\displaystyle G_{0}(k)=\frac{1}{k^{2}-m^{2}+i\epsilon},$ (19)
where $\displaystyle m$ is the mass. The Green function becomes
$\displaystyle\displaystyle G_{0}^{(11)}(k;\alpha)=G_{0}(k)+\xi
w^{2}(k;\alpha)[G^{*}_{0}(k)-G_{0}(k)].$ (20)
Here the physical quantities are given by the non-tilde variables, i.e.
$\displaystyle A=B=1$. In addition, $\displaystyle w^{2}(k;\alpha)$ is the
generalized Bogoliubov transformation GBT given as
$\displaystyle\displaystyle
w^{2}(k;\alpha)=\sum_{s=1}^{d}\sum_{\\{\sigma_{s}\\}}2^{s-1}\sum_{l_{\sigma_{1}},...,l_{\sigma_{s}}=1}^{\infty}(-\xi)^{s+\sum_{r=1}^{s}l_{\sigma_{r}}}\,\exp\left[{-\sum_{j=1}^{s}\alpha_{\sigma_{j}}l_{\sigma_{j}}k^{\sigma_{j}}}\right],$
(21)
where $\displaystyle\\{\sigma_{s}\\}$ denotes the set of all combinations with
$\displaystyle s$ elements and $\displaystyle k$ is the 4-momentum.
## III Teleparallel Gravity
Teleparallelism Equivalent to General Relativity (TEGR) is dynamically
equivalent to the standard theory of gravitation formulated in a Riemann
space. However, TEGR is described in terms of torsion in the Weitzenböck
space. The connection in such a space is
$\displaystyle\Gamma_{\mu\lambda\nu}=e^{a}\,_{\mu}\partial_{\lambda}e_{a\nu}\,,$
where $\displaystyle e^{a}\,_{\mu}$ is the tetrad field. It is the dynamical
variable of the theory. The relationship between the metric tensor and the
tetrad field is $\displaystyle g_{\mu\nu}=e^{a}\,_{\mu}e_{a\nu}\,.$ The tetrad
contains two important symmetries, that is the bridge between them. Lorentz
symmetry (Latin indices) and the transformation of coordinates (Greek
indices). It is interesting to note that the Weintzenböck connection is
curvature free, while the anti-symmetric part establishes the following
torsion tensor
$T^{a}\,_{\lambda\nu}=\partial_{\lambda}e^{a}\,_{\nu}-\partial_{\nu}e^{a}\,_{\lambda}\,.$
(22)
This connection is related to the Christoffel symbols by
$\Gamma_{\mu\lambda\nu}={}^{0}\Gamma_{\mu\lambda\nu}+K_{\mu\lambda\nu}\,,$
(23)
where the contortion tensor, $\displaystyle K_{\mu\lambda\nu}$, is given by
$\displaystyle\displaystyle K_{\mu\lambda\nu}$ $\displaystyle\displaystyle=$
$\displaystyle\displaystyle\frac{1}{2}(T_{\lambda\mu\nu}+T_{\nu\lambda\mu}+T_{\mu\lambda\nu})\,,$
(24)
with $\displaystyle T_{\mu\lambda\nu}=e_{a\mu}T^{a}\,_{\lambda\nu}$. The above
identity leads to the relation
$eR(e)\equiv-e(\frac{1}{4}T^{abc}T_{abc}+\frac{1}{2}T^{abc}T_{bac}-T^{a}T_{a})+2\partial_{\mu}(eT^{\mu})\,.$
(25)
The Lagrangian density for TEGR is
$\displaystyle\displaystyle\mathfrak{L}(e_{a\mu})$
$\displaystyle\displaystyle=$
$\displaystyle\displaystyle-\kappa\,e\,(\frac{1}{4}T^{abc}T_{abc}+\frac{1}{2}T^{abc}T_{bac}-T^{a}T_{a})-\mathfrak{L}_{M}$
(26) $\displaystyle\displaystyle\equiv$
$\displaystyle\displaystyle-\kappa\,e\Sigma^{abc}T_{abc}-\mathfrak{L}_{M}\;,$
where $\displaystyle\kappa=1/(16\pi)$, $\displaystyle\mathfrak{L}_{M}$ is the
Lagrangian density of matter fields and $\displaystyle\Sigma^{abc}$ is given
by
$\Sigma^{abc}=\frac{1}{4}(T^{abc}+T^{bac}-T^{cab})+\frac{1}{2}(\eta^{ac}T^{b}-\eta^{ab}T^{c})\;,$
(27)
with $\displaystyle T^{a}=e^{a}\,_{\mu}T^{\mu}$. If a derivative of eq. (26)
with respect to the tetrad field is performed, the field equation reads
$\partial_{\nu}\left(e\Sigma^{a\lambda\nu}\right)=\frac{1}{4\kappa}e\,e^{a}\,_{\mu}(t^{\lambda\mu}+T^{\lambda\mu})\;,$
(28)
where
$t^{\lambda\mu}=\kappa\left[4\,\Sigma^{bc\lambda}T_{bc}\,^{\mu}-g^{\lambda\mu}\,\Sigma^{abc}T_{abc}\right]\,,$
(29)
is the gravitational energy-momentum tensor. Such an expression is frame
dependent and to calculate its average a class of observers must be chosen,
that is, certain conditions must be imposed on the tetrad field . It is to be
noted that the skew-symmetry in $\displaystyle\Sigma^{a\lambda\nu}$ leads to
$\partial_{\lambda}\partial_{\nu}\left(e\Sigma^{a\lambda\nu}\right)\equiv
0\,.$ (30)
This is the conservation law. It is then possible to establish the energy-
momentum vector as
$P^{a}=\int_{V}d^{3}x\,e\,e^{a}\,_{\mu}(t^{0\mu}+T^{0\mu})\,,$ (31)
or with the help of eq. (28), it reads
$P^{a}=4k\,\int_{V}d^{3}x\,\partial_{\nu}\left(e\,\Sigma^{a0\nu}\right)\,.$
(32)
This is the total energy vector. It is interesting to note that it is a vector
under global Lorentz transformation which implies that energy, as the zero
component of this 4-vector, is not an invariant. In fact, it depends on the
choice of tetrad, which determines the very choice of the observer. On the
other hand the quantity is not dependent on the coordinate choice. These are
indeed desirable features for any definition of energy-momentum.
With the well-established definition of an energy-moment tensor, the first
element necessary for the application of TFD is defined. It is still necessary
to obtain a propagator for the field. Using the weak field approximation
$g_{\mu\nu}=\eta_{\mu\nu}+h_{\mu\nu},$ (33)
which in eq. (26) leads to
$\langle
e_{b\lambda},e_{d\gamma}\rangle=\Delta_{bd\lambda\gamma}=\frac{\eta_{bd}}{\kappa
q^{\lambda}q^{\gamma}}.$ (34)
This is the graviton propagator usk . Then the Green function is
$\displaystyle\displaystyle
G_{0}(x,x^{\prime})=-i\Delta_{bd\lambda\gamma}\,g^{\lambda\gamma}\eta^{bd}.$
(35)
Explicitly it is
$G_{0}(x,x^{\prime})=-\frac{i64\pi}{q^{2}}\,,$ (36)
with $\displaystyle q=x-x^{\prime}$, where $\displaystyle x$ and
$\displaystyle x^{\prime}$ are four vectors. With the weak field approximation
the gravitational energy-momentum tensor $\displaystyle t^{\lambda\mu}$
becomes
$\displaystyle\displaystyle t^{\lambda\mu}(x)$ $\displaystyle\displaystyle=$
$\displaystyle\displaystyle\kappa\Bigl{[}g^{\mu\alpha}\partial^{\gamma}e^{b\lambda}\partial_{\gamma}e_{b\alpha}-g^{\mu\gamma}\partial^{\alpha}e^{b\lambda}\partial_{\gamma}e_{b\alpha}-g^{\mu\alpha}(\partial^{\lambda}e^{b\gamma}\partial_{\gamma}e_{b\alpha}-\partial^{\lambda}e^{b\gamma}\partial_{\alpha}e_{b\gamma})$
(37)
$\displaystyle\displaystyle-2g^{\lambda\mu}\partial^{\gamma}e^{b\alpha}(\partial_{\gamma}e_{b\alpha}-\partial_{\alpha}e_{b\gamma})\Bigl{]}\,.$
For dealing with the mean of the energy-moment tensor the standard procedure
is to consider it at different points of the space and then take the limit.
This avoids divergences. Hence
$\displaystyle\displaystyle\langle t^{\lambda\mu}(x)\rangle$
$\displaystyle\displaystyle=$ $\displaystyle\displaystyle\langle
0|t^{\lambda\mu}(x)|0\rangle,$ (38) $\displaystyle\displaystyle=$
$\displaystyle\displaystyle\lim_{x^{\mu}\rightarrow
x^{\prime\mu}}4i\kappa\left(-5g^{\lambda\mu}\partial^{\prime\gamma}\partial_{\gamma}+2g^{\mu\alpha}\partial^{\prime\lambda}\partial_{\alpha}\right)G_{0}(x-x^{\prime})\,,$
where $\displaystyle\langle
e_{c}^{\,\,\,\lambda}(x),e_{b\alpha}(x^{\prime})\rangle=i\eta_{cb}\,\delta^{\lambda}_{\alpha}\,G_{0}(x-x^{\prime})$.
This average applies to any metric that is related to the linearized
Einstein’s equations. On the other hand, the validity of this expression is
restricted to stationary observers.
## IV Stefan-Boltzmann law and Casimir effect in FRW universe
The TEGR expression in the weak field approximation leads to the TFD
framework. The mean value of the energy-moment tensor becomes
$\displaystyle\displaystyle\langle
t^{\lambda\mu(AB)}(x;\alpha)\rangle=\lim_{x\rightarrow
x^{\prime}}4i\kappa\left(-5g^{\lambda\mu}\partial^{\prime\gamma}\partial_{\gamma}+2g^{\mu\alpha}\partial^{\prime\lambda}\partial_{\alpha}\right)G_{0}^{(AB)}(x-x^{\prime};\alpha).$
(39)
If we use the Casimir prescription,
$\displaystyle\displaystyle{\cal T}^{\lambda\mu(AB)}(x;\alpha)=\langle
t^{\lambda\mu(AB)}(x;\alpha)\rangle-\langle t^{\lambda\mu(AB)}(x)\rangle\,,$
(40)
then
$\displaystyle\displaystyle{\cal
T}^{\lambda\mu(AB)}(x;\alpha)=\lim_{x\rightarrow
x^{\prime}}\Gamma^{\lambda\nu}(x,x^{\prime})\overline{G}_{0}^{(AB)}(x-x^{\prime};\alpha),$
(41)
where
$\displaystyle\displaystyle\Gamma^{\lambda\nu}=4i\kappa\left(-5g^{\lambda\mu}\partial^{\prime\gamma}\partial_{\gamma}+2g^{\mu\alpha}\partial^{\prime\lambda}\partial_{\alpha}\right)\,,$
(42)
and
$\displaystyle\displaystyle\overline{G}_{0}^{(AB)}(x-x^{\prime};\alpha)=G_{0}^{(AB)}(x-x^{\prime};\alpha)-G_{0}^{(AB)}(x-x^{\prime})\,.$
(43)
It is necessary to establish the appropriate space-time geometry i.e.,
analysing the result of such expressions on cosmological scales. A homogeneous
and isotropic universe is chosen. The suitable line element is
$\displaystyle\displaystyle
ds^{2}=-dt^{2}+a\left(t\right)\left(dr^{2}+r^{2}d\theta^{2}+r^{2}\sin^{2}{\theta}d\phi^{2}\right)\,,$
(44)
which is the FRW line element of zero curvature. This metric respects the
approach used, as well as the constraints arising from the experiments. If eq.
(42) is used together with eq. (44), then
$\displaystyle\displaystyle\Gamma^{00}=\frac{i}{4\pi}\left[-3\partial^{\prime}_{0}\partial_{0}+\frac{5}{a^{2}}\left(\partial^{\prime}_{1}\partial_{1}+\frac{1}{r^{2}}\partial^{\prime}_{2}\partial_{2}+\frac{1}{r^{2}\sin^{2}{\theta}}\partial^{\prime}_{3}\partial_{3}\right)\right]$
(45)
and
$\displaystyle\displaystyle\Gamma^{11}=\frac{i}{4\pi
a^{2}}\left[5\partial^{\prime}_{0}\partial_{0}-\frac{5}{a^{2}}\left(\frac{3}{5}\partial^{\prime}_{1}\partial_{1}+\frac{1}{r^{2}}\partial^{\prime}_{2}\partial_{2}+\frac{1}{r^{2}\sin^{2}{\theta}}\partial^{\prime}_{3}\partial_{3}\right)\right]\,.$
(46)
Using these relations to calculate the energy and pressure for Stefan-
Boltzmann law and the Casimir effect according to the Bogoliubov
transformation is desired.
### IV.1 Gravitational Stefan-Boltzmann Law
To calculate the Stefan-Boltzmann law, $\displaystyle\alpha=(\beta,0,0,0)$ is
chosen, which leads to the Bogoliubov transformation
$\displaystyle\displaystyle v^{2}(\beta)=\sum_{j_{0}=1}^{\infty}e^{-\beta
k^{0}j_{0}}\,,$ (47)
where $\displaystyle\beta=\frac{1}{T}$. Then the Green function is
$\displaystyle\displaystyle\overline{G}_{0}^{(11)}(x-x^{\prime};\beta)$
$\displaystyle\displaystyle=$ $\displaystyle\displaystyle
2\sum_{j_{0}=1}^{\infty}G_{0}^{(11)}\left(x-x^{\prime}-i\beta
j_{0}n_{0}\right),$ (48)
where $\displaystyle n_{0}=(1,0,0,0)$ and the physical component
$\displaystyle(AB)=(11)$ is chosen, then
$\displaystyle\displaystyle E=\frac{32\pi^{4}}{15}T^{4}\,,$ (49)
and
$\displaystyle\displaystyle P=\frac{32\pi^{4}}{45a^{2}}T^{4}\,,$ (50)
with $\displaystyle E=\langle t^{00(11)}(x;\beta)\rangle$ and $\displaystyle
P=\langle t^{11(11)}(x;\beta)\rangle$. It is interesting to note that the
pressure is dependent on the scale factor which in turn is expanded as
$\displaystyle\displaystyle
a=1+H_{0}\left(t-t_{0}\right)-\frac{q_{0}H_{0}^{2}}{2}\left(t-t_{0}\right)^{2}\,,$
(51)
where $\displaystyle H_{0}$ and $\displaystyle q_{0}$ refer to the Hubble
constant and the deceleration parameter respectively. So when $\displaystyle
a=1$, the state equation becomes $\displaystyle P=\frac{E}{3}$, which is to be
the expected state equation for the graviton. Taking into account the relation
$\displaystyle\left(\frac{\partial P}{\partial
T}\right)_{V}=\left(\frac{\partial S}{\partial V}\right)_{T}$, the entropy
density is
$\displaystyle\displaystyle s=\frac{S}{V}=\frac{128\pi^{4}}{45a^{2}}T^{3}\,.$
(52)
Using the expansion for the scale factor above, the second time derivative of
the entropy density is
$\displaystyle\displaystyle\ddot{s}=-\frac{256\pi^{4}T^{3}}{45a^{2}}\left[\frac{\ddot{a}}{a}-3\left(\frac{\dot{a}}{a}\right)^{2}\right]\,,$
(53)
where dot means a time derivative. Here the Landau theory of second order
phase transition is involved. A divergence in the second derivative of the
entropy determines a critical quantity that characterizes the phase
transition. Here it is assumed that time is the dynamic variable. Hence
$\displaystyle s\rightarrow\infty$ implies $\displaystyle a=0$. If
$\displaystyle\tau=H_{0}\left(t-t_{0}\right)$ is defined as an auxiliary
variable, then it follows that
$\displaystyle\displaystyle\frac{q_{0}}{2}\tau^{2}-\tau-1=0\,.$ (54)
This imposes a constraint on the current deceleration parameter, such that
$\displaystyle q_{0}\geq-\frac{1}{2}\,.$ This is an interesting result
considering that the deceleration parameter is written in terms of the main
cosmological parameters. Results obtained in Feeney showed that using the
broad (truncated) Gaussian $\displaystyle q_{0}=-0.5\pm 1$, it is indeed
possible to obtain a competitive constraint on the Hubble constant. These
results are consistent with phenomenological models of the interaction rates
Pan using the latest microwave background observations from Planck 2018 and
baryon acoustic oscillations measurements.
### IV.2 Casimir effect
The Casimir effect is described in TFD with the choice
$\displaystyle\alpha=(0,i2d,0,0)$, where $\displaystyle d$ is the radius of
the outer spherical surface. This leads to the Bogoliubov transformation
$\displaystyle\displaystyle
v^{2}(d)=\sum_{l_{1}=1}^{\infty}e^{-i2dk^{1}l_{1}}.$ (55)
If the Green function is given by
$\displaystyle\displaystyle\overline{G}_{0}^{(11)}(x-x^{\prime};d)$
$\displaystyle\displaystyle=$ $\displaystyle\displaystyle
2\sum_{l_{1}=1}^{\infty}G_{0}^{(11)}\left(x-x^{\prime}-2dl_{1}n_{1}\right),$
(56)
with $\displaystyle n_{1}=(0,1,0,0)$, then
$\displaystyle\displaystyle E_{c}=-\frac{2\pi^{4}}{45d^{4}a^{4}}\,,$ (57)
and
$\displaystyle\displaystyle P_{c}=-\frac{2\pi^{4}}{15d^{4}a^{6}}\,.$ (58)
This result is obtained by choosing the physical component of Green’s function
$\displaystyle(AB)=(11)$. The same identification for the average energy-
momentum tensor, $\displaystyle E_{c}=\langle t^{00(11)}(x;d)\rangle$ and
$\displaystyle P_{c}=\langle t^{11(11)}(x;d)\rangle$. Two important features
need to be highlighted. The first is the Casimir energy and pressure are
obtained in a vacuum. A time-dependent negative pressure is consistent with an
accelerating expanding universe. The second is that Casimir pressure is
associated with the cosmological constant $\displaystyle\Lambda$. In natural
units, the pressure and the cosmological constant have the same dimension.
Thus the cosmological constant is understood as a fluid with the following
pressure
$\displaystyle p=-\frac{c^{4}\Lambda}{G}\,,$
thus this given by, in unities of the international system,
$\displaystyle\displaystyle\Lambda=\frac{2\pi^{4}G\hbar}{15d^{4}c^{3}}\,,$
(59)
in the present time. In this estimate the outer surface is used as the
observable radius of the universe, this determines $\displaystyle d$ as
$\displaystyle 10^{10}$ light years. The cosmological constant is of the order
of $\displaystyle 10^{-180}\,m^{-2}$. It is interesting to note that in an
incipient universe $\displaystyle\Lambda$ was much larger than it is today.
## V Conclusion
The Stefan-Boltzmann law and the Casimir effect are analyzed in a homogeneous
and isotropic universe. The FRW metric for zero curvature is used. From the
Stefan-Boltzmann law it is possible to understand that there is an energy and
pressure from strictly gravitational thermal radiation. The entropy density
provides for a phase transition that limits the range of the deceleration
parameter. The Casimir effect establishes a negative pressure consistent with
an accelerated expanding universe. Such a quantity when interpreted as the
observable radius of the universe leads to the conclusion that the
cosmological constant is small. It is important to note that due to its
temporal evolution, the cosmological constant played a more relevant role in a
primordial universe. When Casimir effect is established at finite temperature,
imaginary quantities are obtained. This leads to interpret that the
temperature effect in the universe is independent of the pressure exerted by
the vacuum. Perhaps both effects are linked on a smaller scale when quantum
effects are more relevant.
## Acknowledgments
This work by A. F. S. is supported by CNPq projects 308611/2017-9 and
430194/2018-8.
## References
* (1) S. C. Ulhoa, A. F. Santos, T. F. Furtado, F. C. Khanna, Advances in High Energy Physics, 2019, p. 1-6, (2019).
* (2) T. Matsubara, Prog. Theor. Phys. 14, 351 (1955).
* (3) S. C. Ulhoa, A. F. Santos and F. C. Khanna, General Relativity and Gravitation 49, 54, (2017).
* (4) G. Acquaviva, A. Iorio, M. Scholtz. Annals of Physics, 387, 317-333, (2017).
* (5) G. Acquaviva, A. Iorio and L. Smaldone, Physical Review D, 102, 106002, (2020).
* (6) J. W. Maluf, Annalen der Physik, 525, no. 5, pp. 339-357, (2013).
* (7) S. C. Ulhoa, E. P. Spaniol, R. Gomes, A. F. Santos, A. E. Santana, Advances in High Energy Physics, 2020, p. 1-9, (2020).
* (8) M. Krssak, R. J. van den Hoogen, J. G. Pereira, C. G. Böhmer and A. A. Coley, Class. Quant. Grav. 36, no.18, 183001, (2019).
* (9) E. D. Emtsova, A. N. Petrov and A. V. Toporensky, Class. Quant. Grav. 37, no.9, 095006, (2020).
* (10) M. Hohmann, L. Jarv, M. Krssak and C. Pfeifer, Phys. Rev. D 100, no.8, 084002, (2019).
* (11) J W Maluf, S C Ulhoa, J F da Rocha-Neto and F L Carneiro, Class. Quantum Grav. 37 067003 (2020).
* (12) J. Schwinger, J. Math. Phys. 2, 407 (1961); J. Schwinger, Lecture Notes Of Brandeis University Summer Institute (1960).
* (13) Y. Takahashi and H. Umezawa, Coll. Phenomena 2, 55 (1975); Int. Jour. Mod. Phys. B 10, 1755 (1996).
* (14) Y. Takahashi, H. Umezawa and H. Matsumoto, Thermofield Dynamics and Condensed States, North-Holland, Amsterdan, (1982); F. C. Khanna, A. P. C. Malbouisson, J. M. C. Malboiusson and A. E. Santana, Themal quantum field theory: Algebraic aspects and applications, World Scientific, Singapore, (2009).
* (15) H. Umezawa, Advanced Field Theory: Micro, Macro and Thermal Physics, AIP, New York, (1993).
* (16) A. E. Santana and F. C. Khanna, Phys. Lett. A 203, 68 (1995).
* (17) A. E. Santana, F. C. Khanna, H. Chu, and C. Chang, Ann. Phys. 249, 481 (1996).
* (18) F. C. Khanna, A. P. C Malbouisson, J. M. C. Malbouisson and A. E. Santana, Ann. Phys. 326, 2634 (2011).
* (19) S. C. Ulhoa, A. F. Santos and F. C. Khanna. International Journal of Theoretical Physics, p. 1995, (2017).
* (20) S. M. Feeney, D. J. Mortlock, and N. Dalmasso, Mon. Not. R. Astron. Soc. 476, 3861 (2018).
* (21) S. Pan, G. S. Sharov, and W. Yang, Phys. Rev. D 101, 103533 (2020)
|
# Binary evolution, gravitational-wave mergers and explosive transients
in multiple-populations gas-enriched globular-clusters
Mor Rozner Technion - Israel Institute of Technology, Haifa, 3200002, Israel
Hagai B. Perets Technion - Israel Institute of Technology, Haifa, 3200002,
Israel
###### Abstract
Most globular clusters (GCs) show evidence for multiple stellar populations,
suggesting the occurrence of several distinct star-formation episodes. The
large fraction of second population (2P) stars observed requires a very large
2P gaseous mass to have accumulated in the cluster core to form these stars.
Hence the first population of stars (1P) in the cluster core has had to become
embedded in 2P gas, just prior to the formation of later populations. Here we
explore the evolution of binaries in ambient 2P gaseous media of multiple-
population GCs. We mostly focus on black hole binaries and follow their
evolution as they evolve from wide binaries towards short periods through
interaction with ambient gas, followed by gravitational-wave (GW) dominated
inspiral and merger. We show this novel GW-merger channel could provide a
major contribution to the production of GW-sources. We consider various
assumptions and initial conditions and calculate the resulting gas-mediated
change in the population of binaries and the expected merger rates due to gas-
catalyzed GW-inspirals. For plausible conditions and assumptions, we find an
expected GW merger rate observable by aLIGO of the order of up to a few tens
of $\rm{Gpc^{-3}yr^{-1}}$, and an overall range for our various models of
$0.08-25.51\ \rm{Gpc^{-3}yr^{-1}}$. Finally, our results suggest that the
conditions and binary properties in the early stage of GCs could be critically
affected by gas-interactions and may require a major revision in the current
modeling of the evolution of GCs.
## 1 Introduction
Stars are thought to form following the collapse of giant molecular clouds
(GMCs), and further grow and evolve through accretion from, and interaction
with the GMC ambient gaseous environment during their early evolution, of up
to a few Myrs. Following the gas dispersal and depletion, the later long-term
evolution of stars and multiple systems is thought to be dominated by their
gas-free stellar evolution and their dynamical interactions with other stellar
companions and/or stars in the cluster. However, some environments can be
replenished with gas leading to late epochs of stellar and binary evolution of
stars embedded in gas. Already decades ago, Bahcall & Ostriker (1976) have
suggested that stellar compact objects can interact with gaseous disks around
massive black holes (active galactic nuclei; AGNs), accrete and give rise to
X-ray flarings. Ostriker (1983) suggested that stars and compact objects
embedded in AGNs disks can accrete gas from the ambient gaseous medium, grow
to Chandrasekhar mass and explode as type Ia supernovae (SNe), and later
Artymowicz et al. (1993) discussed accretion onto stars in AGN disks giving
rise to massive stars exploding as core-collapse (CC) SNe and polluting the
AGN disks.
The dynamical evolution of _b_ inary gravitating objects embedded in a large-
scale gaseous environment could be altered through gas-dynamical friction and
accretion that change their orbit and masses and potentially catalyze their
merger. We have first discussed binary evolution in gaseous media in the
context of catalyzed mergers of binary planetesimals in a protoplanetary disks
(Perets & Murray-Clay, 2011; Grishin & Perets, 2016), and later in the context
of compact object binaries in AGN disks (McKernan et al., 2012), where the
latter have been extensively studied since then (e.g. Stone et al., 2017;
McKernan et al., 2018; Roupas & Kazanas, 2019; Tagawa et al., 2020, and
references therein). Baruteau et al. (2011) explored the evolution of binary
_m_ ain-sequence stars (MS) in gas disks around massive black holes (MBHs),
suggesting they harden and merger through the interaction with the gas.
Various studies followed the evolution of pre-MS/MS binaries embedded in gas
just following their formation during the star-formation epoch of stars in
molecular clouds/young clusters, also suggesting that binaries can shrink and
merge through the process (Gorti & Bhatt, 1996; Er et al., 2009; Korntreff et
al., 2012). It was also suggested that the evolution of embedded binaries
could be driven by the formation of a circumbinary disk, which torques the
binary. The evolution of binaries in circumbinary disks have been more
extensively studied over a wide range of scales from planets, to stars and
MBHs (though typically not in the context of a large-scale gaseous
environment), but the exact evolution and even the direction of the binary
migration in such circumbinary disks are still debated (e.g. Artymowicz et
al., 1991; Artymowicz & Lubow, 1994; Bate, 2000; Tang et al., 2017; Moody et
al., 2019; Muñoz et al., 2019; Duffell et al., 2020; Muñoz et al., 2020, and
references therein).
Although the evolution of stars, binaries and compact objects embedded in
gaseous (typically AGN) disks near MBHs have been extensively studied over the
last few years, other gas-embedded stellar environments received far less
attention. Here and in a companion paper (Perets, 2022) we study the evolution
of single and binary compact-object binaries in the early gas-rich
environments that likely existed in multiple-population globular clusters
(GCs) and other young massive clusters (YMCs). We also briefly discuss other
(non compact-object - main-sequence and evolved) stars and binaries in such
environments, but postpone detailed study of the latter to future exploration.
As we discuss below, such gas-rich environments are likely to be far more
ubiquitous than AGN disks and potentially play a key role in the the
production of compact binaries, binary mergers, gravitational-waves sources
and explosive transients.
For decades, GCs were thought to host simple stellar populations formed
through a single star-formation episode. However, detailed observations over
the last decade (see e.g. Carretta et al., 2009; Bastian & Lardo, 2018, and
references therein) have shown that the vast majority of galactic GCs host
multiple stellar populations showing different light elements content. The
origins of multiple populations have been extensively studied, but no clear
solution has yet been found (see Renzini et al., 2015; Bastian & Lardo, 2018;
Gratton et al., 2019, for summaries of the scenarios and their caveats). The
current thought is that GCs experienced two or more star formation episodes,
in which second generation/population (2P) stars formed from processed (2P)
gas lost from earlier generation/population (1P) stars, and/or accreted
external gas. Kinematics show that 2P stars are more centrally concentrated
and were likely formed in the inner region of the GC where the 2P gas is
expected to have accumulated.
While the source of the 2P gas is debated, the late formation of 2P stars
require that tens up to hundreds of Myrs after their formation, 1P stars had
become embedded in a highly gas-rich environment that later produced the 2P
stars. The evolution of stars, binaries and compact objects embedded in gas
could therefore be significantly altered in such gaseous environments,
following similar processes as discussed for AGN disks and pre-MS stars
embedded in the progenitor GMCs. Such processes were little studied in the
context gas-embedded multiple-population GCs (Vesperini et al., 2010;
Maccarone & Zurek, 2012; Leigh et al., 2013, 2014; Roupas & Kazanas, 2019;
Perets, 2022, but see works by us and others on some aspects of such
evolution) which is the focus of the the study below. In particular, in this
paper, we introduce the effect of gas-catalyzed hardening (shrinkage of the
orbit) of binaries in GCs, and discuss its implications for GCs (and YMCs)
binary population and binary mergers, the production of GW sources, and the
formation of other merger products, compact binaries and explosive transient
events catalyzed by binary interactions with gas.
In section 2 we briefly discuss the gas replenishment in multiple-population
GCs. In section 3 we describe the hardening processes of binaries in globular
clusters due to gas-dynamical friction, and its relation to dynamical
hardening by stars and GW inspirals. In section 4 we introduce our results: in
subsection 4.1 we focus on the evolution of an individual binary under the
effect of gas hardening and in subsection 4.4 we estimate the expected merger
rate from the channel we proposed. In section 5 we discuss our results and
additional implications. In section 6 we summarize and conclude.
## 2 Multiple stellar populations and early gas-replenishment in GCs
As discussed above and in Perets (2022), gas could be replenished in GCs (and
YMCs) through mass lost from evolved stars and binaries and/or through
accretion of external gas onto the clusters (see a detailed review in Bastian
& Lardo, 2018).
The formation channel sets the amount of gas and hence the dynamics and
evolution of embedded stars/binaries. Given the correlation between the
fractions of 2P stars and GCs properties, it is likely that a large fraction
of 2P stars correspond to higher masses of the clusters, larger escape
velocities (Mastrobuono-Battisti & Perets, 2020), and hence larger mass of
replenished gas.
Given the observed kinematics and concentrations of 2P stars, and theoretical
models for the formation and evolution of 2P stars, it is thought that the
replenished gas is concentrated in the central part of GCs, where 2P stars are
concentrated. It is likely that the remnant angular momentum of replenished
gas gives rise to the formation of 2P in gaseous disks, rather than spherical
distribution (Bekki, 2010, 2011; Mastrobuono-Battisti & Perets, 2013, 2016).
The total mass of 2P gas in GCs is highly uncertain, but given reasonable
assumptions on the relation between the gas and the observed populations of 2P
stars in GCs one can provide an estimate the amount of replenished gas and its
density. The typical gas density in star forming regions is usually
constrained in the range $10^{2}-10^{6}\ M_{\odot}\rm pc^{-3}$ (Leigh et al.,
2014). Estimates for the 2P gas densities could be obtained from a simple
order of magnitude calculations, assuming 2P stars were formed from
replenished gas. The gas density is then $\rho_{g}\sim M_{g}/V_{\rm 2P}$ where
$M_{g}$ is the mass of the gas and $V_{\rm 2P}$ is the typical volume in which
the 2P stars reside. Following Bekki (2017), $M_{\rm 2P}\sim 10^{5}M_{\odot}$
and $\epsilon_{g}=0.3$, then $M_{g}\sim 3\times 10^{5}M_{\odot}$, where
$\epsilon$ is the star-formation efficiency. The infalling replenished gas is
likely concentrated in a compact region in the central parts of GCs, such that
the typical effective radius that encloses the 2P population is of the order
of $1\ \rm{pc}$ (Bekki, 2017). Taken together, the typical density of the
replenished gas is $\sim 3\times 10^{5}\ M_{\odot}\ \rm{pc}^{-3}$, which lies
within the expected range for gas densities in star-forming regions. From this
density, we will consider scaling to different gas masses, considering $R_{\rm
core}=1\ \rm{pc}$ and take $\rho_{g}\sim M_{\rm g}/R_{\rm core}^{3}$
accordingly. In particular, as we discuss below, the 2P gas is likely enclosed
in a disk-like configuration, in which case the expected gas densities are
higher. A priori, the binary hardening releases energy that could heat the gas
significantly, but from a crude calculation, the cooling rate is high enough
to compensate for it (see also Tagawa et al., 2020 for a similar calculation
in AGN disks). We also note that the possible production of jets could
potentially unbind gas from the disk (Soker, 2016; Tagawa et al., 2022), but
the study of this possibility is beyond the scope of the current paper.
The total amount of gas is depleted in time, due to formation of stars and/or
accretion onto stars, and later gas ejection through possible radiation
pressure processes and SNe. For simplicity we assume an exponential decay,
i.e. $\rho_{g}(t)=\rho_{g,0}\exp(-t/\tau_{\rm gas})$ and consider several
possible options for the gas lifetime, to account for uncertainties in the
possible gas-depletion processes involved.
### 2.1 Disk configuration
Gas replenishment leading to the formation of 2P stars in GCs might form a
disk-like structure in the cluster nuclei (e.g. Bekki, 2010; Mastrobuono-
Battisti & Perets, 2013).
Following Bekki (2010), we consider a flat disk, i.e. with a constant aspect
ratio. We estimate the aspect ratio by $h/r\sim c_{s}/v_{K}$ where
$v_{K}=\sqrt{G(M_{\rm gas}+M_{\star})/R_{\rm core}}$ is the typical velocity
in the central parsec. The speed of sound $c_{s}=\sqrt{k_{B}T_{\rm gas}/\mu
m_{p}}$ ranges between $0.1-10\ \rm{km/sec}$ (e.g. Bekki, 2010; Leigh et al.,
2013), in correspondence to the gas temperature $T_{\rm gas}$, such that
$c_{s}\approx 0.6\ \rm{km/sec}$ corresponds to temperature of $100\ K$, which
is the typical temperature in star formation areas, where $\mu=2.3$, and
$m_{p}$ is the proton mass. Exponential disk models were also considered
(Hénault-Brunet et al., 2015), but here we focus on simple models.
In our fiducial model, we consider $c_{s}=10\ \rm{km/sec}$, unless stated
otherwise. Then, the aspect ratio $h/r\approx 0.23$. Following Bekki (2010),
we consider a velocity dispersion of $\sigma_{\rm disk}=10\ \rm{km/sec}$ for
stars embedded in the disk. As a conservative assumption, we consider the
stellar/massive objects density in the disk to be the same as in the core i.e.
$n_{\star,disk}\approx n_{\star}=10^{5}\ \rm{pc^{-3}}$. However, it should be
noted that due to gas dynamical friction, stars will migrate and experience
inclination damping, and the effective density in the disk is expected to be
higher (e.g. Artymowicz et al., 1993; Leigh et al., 2014; Grishin & Perets,
2016).
We can estimate the volume ratio between the disk and the core volume by $\pi
R_{\rm core}^{2}h/(4\pi R_{\rm core}^{3}/3)\sim 0.75h/r$. Then, under the
assumption that all the second generation gas is concentrated in the disk, we
get a typical gas density of $\rho_{\rm g,disk}\sim 1.74\times 10^{6}\
M_{\odot}\ \rm{pc^{-3}}$. The fraction of stars in the disk will change for
thinner/thicker disks correspondingly.
The evolution of binaries in disks differs in several aspects from the
evolution in a spherical configuration. For our discussion, the major ones
are: the velocity dispersion decreases, the gas density increases and the
total number of stars contained in the disk is only the volumetric fraction of
the disk compared with the volume of the spherical core. The fraction might
change with time due to the interaction with gas.
## 3 Dynamics of binaries and their interaction with gas: binary hardening
and mergers
Binaries embedded in gas interact with it, exchange angular momentum and
energy and possibly accrete gas. These processes are quite complex; here we
focus on the interaction through gas-dynamical friction (GDF), while other
suggested processes for interaction with gas are discussed in subsec. 4.2.
Besides interaction with gas, binaries in GCs can interact with other stars
through dissipative effects such as GW inspirals or tidal evolution and
through dynamical encounters with other stars through three (or more)-body
encounters (Heggie, 1975).
The semimajor axis (SMA) of a given massive binary in a gas-enriched
environment evolves through the combined effect of the above-mentioned
processes:
$\displaystyle\frac{da_{\rm bin}}{dt}=\frac{da_{\rm bin}}{dt}\bigg{|}_{\rm
3-body}+\frac{da_{\rm bin}}{dt}\bigg{|}_{\rm GDF}+\frac{da_{\rm
bin}}{dt}\bigg{|}_{\rm GW}$ (1)
where $a_{\rm bin}$ is the binary SMA.
A priori, all the three mechanisms contribute to the evolution of the SMA.
However, in practice, each of these process dominates in a specific regime,
and can be typically neglected in other regimes. Binaries could shrink to
shorter periods (harden) due to the effect of gas-interaction or GWs inspiral,
and get harder or softer (wider) due to three-body interactions with other GC
stars. As we discuss in the following, the evolution of hard binaries is
dominated by gas-interactions at large separations and by GW-emission at small
separation, while dynamical hardening and softening through three-body
encounters (Heggie, 1975) can be neglected in these regimes. Nevertheless,
binary softening and evaporation before the gas-replenishment episode can
destroy the widest binaries in the clusters, and hence determine the largest
possible initial SMAs for binaries in the cluster at the beginning of the gas-
interaction epoch. Moreover, it could play a role in hardening binaries that
did not merge within the gas epoch.
The interaction with gas can also give rise to the formation of new wide
binaries through two-body and three-body encounters in gas (Goldreich et al.,
2002; Tagawa et al., 2020), allowing for replenishment of binaries in
clusters.
In the following we discuss these various processes, while we neglect the
effect of direct accretion onto compact objects and their growth, which is
beyond the scope of the current paper (though generally such accretion, if
effective likely further accelerates binary hardening (e.g. Roupas & Kazanas,
2019).
### 3.1 Hardening and softening through dynamical encounters with stars
Due to interactions with other stars, hard binaries tend to get harder, while
soft binaries tend to get softer (Heggie, 1975); see updated discussion and
overview of these issues in Ginat & Perets (2021a, b). Hence, in the absence
of a gaseous environment stellar dynamical hardening plays an important role
in binary evolution and in catalyzing binary mergers.
#### 3.1.1 Hard binaries
For hard binaries, the dynamical hardening rate (up to order unity corrections
calibrated usually from numerical simulations) is given by (Spitzer, 1987)
$\displaystyle\frac{da_{\rm bin}}{dt}\bigg{|}_{3-body}=-\frac{2\pi
Gn_{\star}m_{\rm pert}(2m+m_{\rm pert})a_{\rm bin}^{2}}{mv_{\infty}}$ (2)
where we consider a binary with equal mass components, $m=m_{1}=m_{2}$ and an
external perturber with mass $m_{\rm pert}$. For interactions with other
massive objects only, $n_{\star}$ and $m_{\rm pert}$ should be taken as
$n_{\bullet}$ and $\bar{m}_{\bullet}$ correspondingly.
#### 3.1.2 Soft binaries
A binary is called a soft binary if its energy is lower than
$\bar{m}\sigma^{2}$. This condition sets a critical SMA
$\displaystyle a_{\rm SH}=\frac{2Gm^{2}}{\bar{m}\sigma^{2}}\approx$ (3)
$\displaystyle\approx 200.53\rm{AU}\left(\frac{m}{10\
M_{\odot}}\right)^{2}\left(\frac{43.2\
\rm{km/sec}}{\sigma}\right)^{2}\left(\frac{0.5\ M_{\odot}}{\bar{m}}\right)$
As can be seen, massive stars tend to be hard relative to the background stars
in the cluster, due to the scaling $a_{\rm SH}\propto m^{2}/\bar{m}$. Hence,
one should define the hardness of massive binaries relative to both low mass
and high mass stars, in particular, the latter will give rise to softer
binaries. We then get the following modified expression (Quinlan, 1996; Kritos
& Cholis, 2020),
$\displaystyle a_{\rm SH,\bullet}\approx\frac{Gm}{4\sigma^{2}}\approx
1.25\rm{AU}\left(\frac{m}{10\ M_{\odot}}\right)\left(\frac{43.2\
\rm{km/sec}}{\sigma}\right)^{2}$ (4)
Soft wide binaries are prone to destruction due to encounters with other
stars. The dynamical evolution of massive binaries is dominated by
interactions with other massive stars and their number density in the core is
elevated due to mass segregation (Sigurdsson & Phinney, 1995).
As to bracket the effect of softening, we consider two possibilities. (1)
Softening is dominated by encounters with stellar BHs, where we assume the
number density of such objects to be $n_{\rm\rm b}=n_{\bullet}=10^{3}\
\rm{pc^{-3}}$, due to mass segregation to the core, where
$\bar{m}_{\bullet}=10\ M_{\odot}$ (see a discussion in Miller & Hamilton,
2002). (2) Softening is dominated by low-mass $0.5\ M_{\odot}$ mass stars, if
the cluster is not well segregated, and the $n_{\rm b}=n_{\star}=10^{5}\
\rm{pc^{-3}}$.
Hence, the typical lifetime of a soft massive binary is given by (e.g. Binney
& Tremaine, 2008),
$\displaystyle\tau_{\rm evap,massive}$
$\displaystyle\approx\frac{(m_{1}+m_{2})\sigma}{16\sqrt{\pi}n_{\rm
b}\bar{m}^{2}_{b}Ga\ln\Lambda}$ (5)
where $\ln\Lambda$ is the Coulomb logarithm and $n_{b}$ and $\bar{m}_{b}$ are
the number density and the mass of the background stars and change according
to our choice between (1) and (2). The widest binaries that survive
evaporation until the formation of second generation stars, signed as
$\tau_{\rm SG}$, taken here to be $100\ \rm{Myr}$
$\displaystyle a_{\rm widest}$ $\displaystyle=\max\left\\{a_{\rm
SH,\bullet},\frac{(m_{1}+m_{2})\sigma}{16\sqrt{\pi}n_{b}\bar{m}^{2}_{b}G\tau_{\rm
SG}\ln\Lambda}\right\\}$ (6)
For our fiducial parameters, $a_{\rm widest}=24.9\ \rm{AU}$ for the segregated
case and $200.53\ \rm{AU}$ for the non-segregated case. In principle, binaries
could soften and be disrupted via encounters during the gas-replenishment
episode, however, the GDF hardening described in the following is more
efficient at this stage. Therefore binary evaporation due to encounters sets
the stage, and determines the SMA of the widest binaries at the beginning of
the gas-enrichment stage, but can be neglected during the the time binaries
are embedded in gas.
### 3.2 Gas dynamical friction
In gas-rich environments, such as the 2P gas environment of multiple-
population GCs/YMCs (and AGN disks), GDF can play a major role in hardening.
The evolution of binaries in gaseous media has been studied over a wide range
of astrophysical scales from asteroids to MBHs (as discussed in the
introduction).
The effect of gas was suggested to be modeled mainly via several approaches.
One suggestion is the accretion of gas onto a binary forms a circumbinary
minidisk, due to accretion to the Hill sphere. In such disks, torques similar
to the ones described type I/II migration of planets in protoplanetary disks
could lead to the shrinkage of the binary SMA (e.g. Artymowicz et al., 1991;
McKernan et al., 2012; Stone et al., 2017; Tagawa et al., 2020). Such
migration leads to very efficient mergers, far more efficient than the case of
interaction dominated by GDF, as we discuss below. However, these issues are
still debated, and some hydrodynamical simulations show that such torques
might lead to outward migration (e.g. Moody et al., 2019; Duffell et al.,
2020; Muñoz et al., 2020), while other hydrodynamical studies indicate that in
thin disks one should have inward migration (Duffell et al., 2020; Tiede et
al., 2020). We do note that most studies consider initially circular orbits,
and generally follow circular orbits, while eccentric orbits could evolve
differently, with their orbital eccentricity possibly excited into very high
eccentricities, as we discuss below in the context of modeling the evolution
through GDF.
Therefore, the approach on which we focus here, considers the effects of GDF
(Ostriker, 1999). When an object has a non-zero velocity relative to the
background gas, the interaction with the gas reduces the relative velocity and
therefore hardens binaries (e.g. Escala et al., 2004; Baruteau et al., 2011).
The binary hardening induced by GDF for the circular case, with binary
components with the same mass $m_{1}=m_{2}=m$ is given by (Grishin & Perets,
2016),
$\displaystyle\frac{da_{\rm bin}}{dt}\bigg{|}_{\rm GDF}$
$\displaystyle=-\frac{8\pi G^{3/2}a_{\rm
bin}^{3/2}}{\sqrt{m_{1}+m_{2}}}\rho_{g}(t)\frac{m}{v_{\rm
rel}^{2}}f\left(\frac{v_{\rm rel}}{c_{s}}\right);$ (7) $\displaystyle f(x)$
$\displaystyle=\begin{cases}\frac{1}{2}\log\frac{1+x}{1-x}-x,\ 0<x<1,\\\
\frac{1}{2}\log\left(x^{2}-1\right)+\log\Lambda_{g},\ x>1\end{cases}$ (8)
where $f$ is a dimensionless function derived in Ostriker (1999), $v_{\rm
rel}$ is the velocity of the binary relative to the gas, taken as the
Keplerian velocity of the binary, i.e. $v_{K}=\sqrt{G(m_{1}+m_{2})/a_{\rm
bin}}$, which dominates the relative velocity throughout most of the
evolution.
Under this assumption, eq. 7 could be written as
$\displaystyle\frac{da_{\rm bin}}{dt}\bigg{|}_{\rm GDF}$
$\displaystyle=-8\pi\sqrt{\frac{Ga_{\rm
bin}^{5}}{2m}}\rho_{g}(t)f\left(\frac{v_{K}}{c_{s}}\right)$ (9)
For massive binaries, the effect of stellar hardening will be weaker than the
effect on less massive stars, as can be seen directly from eq. 2. In contrast,
the effect of gas hardening increases with mass (eq. 7). Comparison of the two
shows that hardening is dominated by gas hardening rather than stellar
hardening. Moreover, although the effect of GDF decreases as the binary
hardens, it decays more slowly than the three-body hardening, as could be seen
from the scaling $\dot{a}_{\rm hard,\star}\propto a^{2}$ and $\dot{a}_{\rm
GDF}\propto a^{3/2}$, and therefore GDF dominates the evolution over stellar-
hardening throughout the evolution. After gas depletion, three-body hardening
becomes the dominant dynamical process for wide binaries, while for
sufficiently small separations, the evolution is GWs-dominated.
### 3.3 Gravitational-wave inspiral
For stellar mass objects GW inspiral becomes important only at very small
separations, and can be neglected in regard to main-sequence (or evolved)
stellar binaries that merge before GW emission becomes important. However, GW
inspiral plays a key-role in the evolution of binaries composed of compact
objects.
For a circular binary in the quadruple approximation, the GWs inspiral rate is
given by (Peters, 1964),
$\displaystyle\frac{da}{dt}\bigg{|}_{\rm
GW}=-\frac{64G^{3}m_{1}m_{2}(m_{1}+m_{2})}{5c^{5}a^{3}}$ (10)
where $G$ is the gravitational constant and $c$ is the speed of light.
Without gas dissipation, the maximal SMA for GW merger within a Hubble time is
given by
$\displaystyle a_{max,GW}=\left(\frac{64\tau_{\rm
Hubble}G^{3}m_{1}m_{2}(m_{1}+m_{2})}{5c^{5}}\right)^{1/4}\approx$
$\displaystyle\approx 0.07\ \rm{AU}\left(\frac{m}{10\ M_{\odot}}\right)^{3/4}$
(11)
A compact binary that is driven by GDF to separations below $a_{max,GW}$ would
eventually inspiral and merge, even if it survived the gas-replenishment
stage, and would produce a GW-source.
## 4 Results
Accounting for the effects of the various processes discussed above, we can
follow the evolution of binaries in clusters during the gas epoch and assess
its outcomes. Overall we find that under plausible conditions all black-hole
binaries initially existing in the cluster inner regions that become embedded
in gas during the gas-replenishment phase could be driven to short separations
and merge within a Hubble time.
These results suggest that gas-catalyzed GW-mergers in GCs and YMCs, not
considered at all in current modeling of GCs, could serve as an important
channel for the production of GW-sources, and plays a key role in the
evolution of binaries in such clusters.
Both the GDF and GW-inspiral timescales for lower mass compact objects such as
neutron stars (NSs) and white dwarfs (WDs), are longer (as can be seen in eq.
3.3), but they are also expected to modify their semimajor axis distribution.
Here we focus on mergers of BHs, and postpone a detailed discussion of NS and
WD mergers for a follow-up paper, but we should already note that potential WD
mergers could give rise to the production of explosive events such as type Ia
supernovae from mergers of massive white-dwarfs (see also Perets, 2022), and
could produce GW sources observable by planned GW-detection space missions. NS
mergers could produce short gamma-ray bursts and aLIGO GW sources. Combined
BH-NS or BH-WD binaries with their high mass but lower mass-ratio could be
driven to mergers at intermediate timescales between highest and lowest
timescales considered here giving rise to WD/NS disruptions by the BH possibly
producing rapid faint SNe (e.g. Zenati et al., 2019, 2020; Bobrick et al.,
2022, and references therein) or short-GRBs accompanied by a potential GW
aLIGO-source. The dynamics of binaries with non-equal masses could be however
more complicated and is not explored here.
In the following we discuss our results in detail.
### 4.1 Gas-assisted GW-mergers
Figure 1: The effects of gas hardening, GWs and three-body hardening. The blue
dashed line represents the maximal SMA in which GW emission catalyzes a binary
merger within a Hubble time. We consider the evolution of a binary with masses
$m_{1}=m_{2}=10\ M_{\odot}$ and initial separation of $a_{0}=1\ \rm AU$. We
consider an exponential decaying background gas density
$\rho_{g}=\rho_{g,0}\exp(-t/\tau_{\rm gas})$ with $\rho_{g,0}=1.74\times
10^{6}\ M_{\odot}\rm{pc}^{-3}$ and $\tau_{\rm gas}=50\ \rm Myr$. Figure 2:
The combined effect of gas hardening, three-body hardening and GWs on a
binary, for different background gas masses (and corresponding gas densities).
The blue dashed line represents the maximal SMA in which GW emission catalyzes
a binary merger within a Hubble time. The solid lines corresponds to the
evolution of the SMA, starting from an initial separation of $a_{0}=1\
\rm{AU}$, and given different background densities, with an exponential
decaying gas density $\rho_{g}=\rho_{g,0}\exp(-t/\tau_{\rm gas})$ with
$\tau_{\rm gas}=50\ \rm Myr$ (that corresponds to $M_{\rm gas,0}=3\times
10^{5}\ M_{\odot}$). The velocity dispersions are calculated given the total
mass of the gas and stars. Figure 3: The combined effect of gas hardening,
three-body hardening and GWs on a binary, for different initial separations.
The blue dashed line represents the maximal SMA in which GW emission catalyzes
a binary merger within a Hubble time. The purple dashed line corresponds to
the widest binary allowed by evaporation considerations. The solid lines
corresponds to the evolution of the SMA, starting from an initial separations
of $a_{0}=1,10,100,200\ \rm{AU}$, and given an exponential decaying gas
density $\rho_{g}=\rho_{g,0}\exp(-t/\tau_{\rm gas})$ with $\tau_{\rm gas}=50\
\rm Myr$.
In Fig. 1 we compare the different hardening processes of binaries in gas-
embedded regions. As can be seen, for large separations, the evolution is
dominated by the gas hardening, while for smaller separations (at late times
after the gas depletion), three-body hardening and finally GWs dominate the
evolution. The transition between the different regimes is determined by the
gas density in the cluster, as well as stellar density. Unless stated
otherwise, we consider for our fiducial model a background of stars with
typical masses of $\bar{m}=0.5\ M_{\odot}$.
In Fig. 2 we present the evolution of binaries with an initial separation of
$a_{0}=1\rm AU$, due to GDF, for different ambient gas-densities. The gas
hardening mechanism is generally very effective and leads to binary migration
to small separations within short timescales, given a sufficiently dense
gaseous environment. As we discuss below, such gas-assisted evolution would
then give rise to high rates of GW-mergers of BH binaries, comparable with the
BH merger rates inferred from aLIGO-VIRGO-KAGRA (LVK) collaboration (Abbott et
al., 2016, 2021).
It should be noted that the gas could still dominate the evolution even after
reaching $a_{\rm GW}$, as long as the gas was not depleted and the timescale
for GWs mergers is larger than the GDF induced merger timescale. In principle,
GDF-dominated evolution might even be identified in the GW inspiral (in future
space missions) before the merger, under appropriate conditions, if GDF still
dominates the evolution in LISA frequencies.
We find that circular binaries shrink and reach final small separations,
dictated by the initial conditions, which are not sufficiently small as to
allow for GW emission alone to drive the binaries to merger even after a
Hubble time. Nevertheless, at such short period, these very-hard binaries are
more likely to merge due to dynamical encounters on the long-term compared
with the primordial population of binary-BHs, and should be appropriately
accounted for in simulations of GC stellar populations.
In Fig. 3, we introduce the evolution of binaries with different initial
separations under the combined effect of GDF, three-body hardening and GWs. It
could be seen that although the merger timescales of wider binaries are
slightly larger, all the binaries are expected to merge within a Hubble time.
Hence, the effect of the presence of gas in the initial stages is robust
across all separations and will modify the binary population. For wide enough
binaries, we enter the subsonic range. In order to avoid the discontinuity in
eq. 8, we take it as a constant in a small environment around Mach $1$ – for
$\mathcal{M}<1.01$, we consider $f(\mathcal{M})\equiv f(1.01)$, where the
widest binary we consider corresponds to $\mathcal{M}\approx 0.97$.
Figure 4: The evolution of the binary separation for different sound speeds.
We consider equal mass binaries with initial separation of $a=1\ \rm{AU}$,
masses $m=m_{1}=m_{2}=10\ M_{\odot}$ and an exponential decaying background
density with $\rho_{\rm g,0}=1.74\times 10^{6}\ M_{\odot}\rm{pc^{-3}}$. The
blue dashed line corresponds to the maximal separation from which a GWs merger
is expected. Figure 5: The effect of gas hardening on a binary, as dictated
by GDF, for different masses of binaries. The different curves correspond to
the evolution of the SMA for different binary masses, starting from an initial
separation of $a_{0}=1\ \rm{AU}$, given a background density with an
exponential decaying gas density $\rho_{g}=\rho_{g,0}\exp(-t/\tau_{\rm gas})$
with $\tau_{\rm gas}=50\ \rm Myr$ and $\rho_{g,0}=1.74\times
10^{6}M_{\odot}\rm{pc}^{-3}$.
In Fig. 5, we demonstrate the dependence of gas hardening on different binary
masses. As can be seen from eq. 7, lower-mass binaries harden over longer
timescales, due to the dependence on the mass that scales as
$\propto\sqrt{m}$, for an equal mass binary with companions $m_{1}=m_{2}=m$.
The final SMA of the binary also depends on the mass of the binary, such that
more massive binaries will attain smaller final SMAs.
In Fig. 4 we consider different sound speeds, all of them in the supersonic
regime. Higher sound speed lead to larger merger timescales, although the
results are robust and do not change steeply between the different choices of
sound speed in this regime.
### 4.2 Comparison with other gas hardening models
Heretofore, we considered gas hardening induced by GDF. However, there are
other approaches to model gas hardening.
In AGN disks, gas hardening is also modeled using processes similar to
migration models in protoplanetary disks (as was suggested in the context of
AGN disks McKernan et al., 2012; Stone et al., 2017; Tagawa et al., 2020). Gas
is captured in the Hill sphere of a binary and leads to the formation of a
circumbinary minidisk. The disk applies a torque on the binary that leads to
separation decay similar to migration type I/II in protoplanetary disks,
although there were studies that pointed out that this torque could lead to a
softening rather than hardening (Moody et al., 2019). Notwithstanding, we will
assume that the formation of a minidisk can take place in GCs and compare the
resulted hardening with our GDF model. The typical timescale for hardening due
to migration torques is given by (e.g. McKernan et al., 2012), modify to the
other type II migration equation
$\displaystyle\tau_{\rm typeII}\sim 46\
\rm{yr}\left(\frac{0.01}{\alpha}\right)\left(\frac{0.23}{h/r}\right)^{2}\left(\frac{40\rm{yr^{-1}}}{\Omega_{\rm
bin}(a_{\rm bin})}\right)$ (12)
where $\alpha$ is the Shakura Sunayev parameter, $h/r$ is the aspect ratio and
$\Omega_{\rm bin}=\sqrt{G(m_{1}+m_{2})/a_{\rm bin}^{3}}$ is the angular
frequency of the binary. We adopt values of $h/r=0.23$ and $\alpha=0.01$ as a
conservative value for the viscosity parameter of the disk. We substitute the
$\Omega_{\rm bin}$ that corresponds to a binary with a separation of $1\
\rm{AU}$. Under these assumptions, the migration timescales, which could be
used to approximate the hardening timescales, are shorter than the typical
migration timescales we derived using the GDF model. These timescales are also
shorter then the ones obtained in AGN disks (e.g. Stone et al., 2017; Tagawa
et al., 2020), as expected. We therefore expect the merger rates we derived to
be similar in this case, and even higher for the lowest gas-densities models,
where the rates were limited by slower hardening. There were more recent
studies that suggested modified migration timescales, here taken for an equal
mass binary
$\displaystyle\tau_{\rm{typeII,K}}=\frac{\Sigma_{\rm disk}}{\Sigma_{\rm
disk,min}}\tau_{\rm typeII},$ (13) $\displaystyle\Sigma_{\rm
disk,min}=\frac{\Sigma_{\rm disk}}{1+0.04K},$ $\displaystyle
K=\left(\frac{m_{1}}{m_{1}+m_{2}}\right)^{2}\left(\frac{h}{r}\right)^{-5}\alpha^{-1}$
These factors lengthen significantly the typical migration timescales, such
that for our fiducial model we expect $\tau_{\rm typeII,K}\approx 71515\
\rm{yr}$. This timescale is still much shorter than the expected timescale
calculated via the gas dynamical friction model.
Another approach to modeling gas-induced inspirals is discussed in Antoni et
al. (2019). They simulate Bondi-Hoyle-Lyttelton (BHL) supersonic flows and
derive the corresponding energy dissipation, fitted to an analytical theory.
While the overall gas hardening timescales could be comparable or shorter for
the parameters that are in our major interest, there are significant
differences in the scaling. The typical inspiral timescale is given by (eq.
$52$ in Antoni et al. (2019)),
$\displaystyle\tau_{\rm BHL}=61\ \rm{Myr}$
$\displaystyle\left(\frac{a_{0}}{AU}\right)^{0.19}\left(\frac{v_{\rm
rel}}{\rm{100\ km\times sec^{-1}}}\right)^{3.38}\times$ (14)
$\displaystyle\times\left(\frac{20\
M_{\odot}}{m_{1}+m_{2}}\right)^{1.19}\left(\frac{7.72\times 10^{7}\
\rm{cm}^{-3}}{n_{\rm gas}}\right)$
where $a_{0}$ is the initial separation of the binary and $n_{\rm{gas}}$ is
the number density of the gas, such that $\rho_{\rm gas}=n_{\rm gas}m_{p}$
where $m_{p}$ is the proton mass.
Each model for gas hardening sets a different critical initial separation from
which the binary will merge within a Hubble time. The timescales dictated both
from the type II migration and BHL mechanism are even shorter than the ones
expected by our fiducial model.
Hence, we will conclude that in all the approaches that we considered to model
gas hardening, the process is very efficient and leads to a robust rate of
mergers, that modifies significantly the binaries’ population, while the major
difference between them is the time of the merger, dictated by the different
gas hardening timescales.
### 4.3 Eccentric evolution
The evolution of binaries in a gaseous medium is significantly different for
non-circular binaries. Here we derive and solve the equations for an orbit-
averaged eccentric evolution of an initially eccentric binary embedded in gas,
but leave a more detailed discussion on the implications for the dynamical
3-body hardening of eccentric binaries to future study.
For simplicity, we will assume that the Keplerian velocity of the binary
components dominate the relative velocity to the gas, and that the gas
velocity is zero relative to the center of mass of the binary. Hence, the
relative velocity between the binary and the gas in the center of mass frame
is given by
$\displaystyle\textbf{v}_{\rm rel}=\frac{\Omega a}{2\sqrt{1-e^{2}}}\left[e\sin
f\hat{r}+(1+e\cos f)\hat{\varphi}\right]$ (15)
The orbit equations for the GDF for a binary with two equal masses are then
given by
$\displaystyle\frac{{da}}{dt}\bigg{|}_{\rm GDF}=\frac{2a^{3/2}}{m_{\rm
bin}\sqrt{Gm_{\rm bin}(1-e^{2})}}\left[F_{r}e\sin f+F_{\varphi}(1+e\cos
f)\right],$ (16) $\displaystyle\frac{{de}}{dt}\bigg{|}_{\rm
GDF}=\frac{2}{m}\sqrt{\frac{a(1-e^{2})}{Gm_{\rm bin}}}\left[F_{r}\sin
f+F_{\varphi}(\cos f+\cos E)\right]$ (17)
where $\textbf{F}_{\rm drag}=F_{r}\hat{r}+F_{\varphi}\hat{\varphi}$, $f$ is
the true anomaly and $E$ is the eccentric anomaly. The orbit-averaged
equations are given by
$\displaystyle\frac{\overline{da}}{dt}\bigg{|}_{\rm
GDF}=\frac{4F_{0}(1-e^{2})^{2}}{\pi m_{\rm
bin}\Omega^{3}a^{2}}\int_{0}^{2\pi}\frac{Idf}{(1+e\cos f)^{2}\sqrt{1+2e\cos
f+e^{2}}},$ (18) $\displaystyle\frac{\overline{de}}{dt}\bigg{|}_{\rm
GDF}=\frac{4F_{0}(1-e^{2})^{3}}{\pi m_{\rm
bin}\Omega^{3}a^{3}}\int_{0}^{2\pi}\frac{I(e+\cos f)df}{(1+e\cos
f)^{2}(1+2e\cos f+e^{2})^{3/2}}$ (19)
where $F_{0}$ is given by $\textbf{F}_{\rm drag}=F_{0}I\textbf{v}_{\rm
rel}/v_{\rm rel}^{3}$. The orbit-averaged equations for GWs are given by
$\displaystyle\frac{\overline{da}}{dt}\bigg{|}_{\rm
GW}=-\frac{64G^{3}m_{1}m_{2}(m_{1}+m_{2})}{5c^{5}a^{3}(1-e^{2})^{7/2}}\left(1+\frac{73}{24}e^{2}+\frac{37}{96}e^{4}\right),$
(20) $\displaystyle\frac{\overline{de}}{dt}\bigg{|}_{\rm
GW}=-\frac{304G^{3}em_{1}m_{2}(m_{1}+m_{2})}{15c^{5}a^{4}(1-e^{2})^{5/2}}\left(1+\frac{121}{304}e^{2}\right)$
(21)
Figure 6: The effects of gas hardening on eccentric orbit. We consider the
evolution of a binary with masses $m_{1}=m_{2}=10\ M_{\odot}$ and initial
separation of $a_{0}=1\ \rm AU$. We consider an exponential decaying
background gas density $\rho_{g}=\rho_{g,0}\exp(-t/\tau_{\rm gas})$ with
$\rho_{g,0}=1.74\times 10^{6}\ M_{\odot}\rm{pc}^{-3}$ and $\tau_{\rm gas}=50\
\rm Myr$. The solid lines correspond to semimajor axis evolution and the
dashed lines to pericenter evolution. Figure 7: The effects of gas hardening
and GWs on eccentric orbit. We consider the evolution of a binary with masses
$m_{1}=m_{2}=10\ M_{\odot}$ and initial separation of $a_{0}=1\ \rm AU$. We
consider an exponential decaying background gas density
$\rho_{g}=\rho_{g,0}\exp(-t/\tau_{\rm gas})$ with $\rho_{g,0}=1.74\times
10^{6}\ M_{\odot}\rm{pc}^{-3}$ and $\tau_{\rm gas}=50\ \rm Myr$. The solid
lines correspond to semimajor axis evolution and the dashed lines to
pericenter evolution.
In Fig. 6 and Fig. 7 we introduce the evolution of eccentric binaries. In Fig.
6, we present the evolution due only to GDF, and in Fig. 7, we also introduce
the effect of GW-emission. As can be seen, the eccentricities become extremely
high within short timescales, indicating that the pericenter shrinks
significantly. Once the pericenters are sufficiently small, the effect of GWs
becomes more significant, and the orbit shrinkage is accompanied by
eccentricity damping, and the binaries are driven into approximately circular
orbit when entering the VLK GW-bands.
Such eccentric evolution could play a key role in the evolution of the binary
populations, as eccentric binaries merge within potentially far shorter
timescales than circular binaries. We note, however, that some studies of a
circumbinary gas-disk evolution of binaries, suggest they are only excited to
moderate eccentricities Tiede et al., 2020, $\sim 0.45$. Nevertheless, if
binaries migration occurs through such processes the overall shrinkage is
rapid irrespective of the eccentricity, leading to fast migration timescale
(see previous subsec.).
We further discuss these issues, and in particular the implications for the
delay time distribution of GW sources from this channel in subsec. 4.4. We
note that the consideration of eccentric binaries gas-hardening, little
studied before should play a similarly important role in binary evolution in
AGN disks, possibly in a different manner than in cases where circumbinary-
disk evolution is assumed Samsing et al., 2020; Tagawa et al., 2021.
### 4.4 Gravitational-waves merger rate
In the following we estimate the GW mergers rate of binary black holes from
the gas-catalyzed channel studied here. We will consider old-formed GCs and
YMCs separately, given their different formation history.
In all the models we considered for gas hardening, all the binaries are
expected to merge within a Hubble time. However, different gas hardening
models suggest different merger timescales. As discussed above, our GDF models
suggest that eccentric binaries merge rapidly, and some of the hydrodynamical
studies discussed above suggest that even circular binaries merge during the
early gas phase. Since most GCs formed very early, such mergers would not be
detected by VLK, given the effectively limited lookback time. However, the
younger equivalents of GCs, so called YMCs, continue to form and generally
follow the star-formation history in the universe. Hence, mergers in such YMC
could occur sufficiently late (and hence closer by) and be detected by VLK and
the contribution of YMCs to the total VLK rate will be the dominant one for
the eccentric cases (or for all binaries, according to e.g. the circumbinary
disk migration models. It should be noted that there is an observational
evidence for gas replenishment also in YMCs (e.g. Li et al., 2016). If,
however, gas densities are lower or the binaries are initially circular/in low
eccentricity, the final SMA of the binaries could be larger, leading to longer
GW-merger time catalyzed by three-body hardening (driving the delay time
distribution to longer timescales), in which case the contribution from old
GCs would be the dominant one.
The rates as a function of the redshift change according to the geometric
structure of the 2P stars. Formation of 2P stars in disks is characterized by
lower velocity dispersions, that lead to earlier mergers, where for the case
of spherical constellation, the higher velocity dispersion leads to later
mergers.
We will start by estimating the number of mergers per cluster,
$\displaystyle N_{\rm merge}\sim f_{\rm disk}f_{\rm bin,surv}f_{\geq
20M_{\odot}}f_{\rm ret}f_{\rm merge}N_{\star}$ (22)
where $f_{\rm disk}$ is the fraction of stars that reside in the disk, $f_{\rm
bin,surv}$ is the fraction of binaries among massive stars that will survive
stellar evolution (i.e. SNe), $f_{\geq 20M_{\odot}}$ is the fraction of stars
with masses that exceed $20\ M_{\odot}$, $f_{\rm ret}$ is the retention
fraction of BHs in the cluster, $f_{\rm merge}$ is the fraction of binaries
that merge among the surviving binaries embedded in the disk and $N_{\star}$
is the number of stars in the cluster.
Following our geometrical considerations in subsec. 2.1, we set $f_{\rm disk}$
in the range $[2\%,20\%]$. However, even large fractions could be taken into
account if there is a significant capture of objects to the disk.
The binarity fraction of massive BHs is $\sim 0.7$, although even higher
values are quite plausible for the massive star progenitors of black holes
(e.g. Sana et al., 2012), stellar evolution may reduce this fraction to a
typical value of $f_{\rm bin,surv}=0.1$ (e.g. Antonini & Perets, 2012). We use
a Kroupa mass function for the cluster, such that the fraction of stars with
masses larger than $20\ M_{\odot}$ is $2\times 10^{-3}$ for non-segregated
environment, for segregated ones we take a fraction of $0.01$. The retention
fraction from the cluster is taken to be $10\%$ (e.g. Kritos & Cholis, 2020
and references therein). Taking into account the initial survival fraction of
wide binaries, we consider $f_{\rm merge}\approx 0.49-0.61$ for our fiducial
model. The lower value corresponds to massive background stars and the upper
limit to low mass background stars ($\bar{m}=0.5\ M_{\odot}$), see a
discussion below eq. 6.
Following Rodriguez et al. (2016), we consider logarithmically flat
distribution of initial SMA in the range $[10^{-2},10^{5}]\ \rm{AU}$ where the
lower limit is close to the point of stellar contact and the upper one to the
Hill radius. It should be noted that although the choice of logarithmically
flat is common, there were other choices of distribution considered, based on
observational data (see Antonini & Perets, 2012 for further discussion).
For our fiducial model, $N_{\star}=10^{5}$ and $M_{\rm cluster}=10^{5}\
M_{\odot}$.
In order to calculate the GWs merger from old GCs, we follow the calculation
of Rodriguez et al. (2016); Kritos & Cholis (2020),
$\displaystyle\mathcal{R}_{\rm old}(z)=\frac{1}{V_{c}(z)}\int_{z_{\rm
min}}^{z}\Gamma_{\rm old}(z^{\prime})n_{\rm
old}(z^{\prime})\frac{dV_{c}}{dz^{\prime}}(1+z^{\prime})^{-1}dz^{\prime}$ (23)
where $\Gamma_{\rm old}$ is the rate of mergers in old GCs, $n_{\rm old}$ is
the GCs number density, which is taken to be in the range
$[0.33,2.57]E^{3}(z)\ \rm{Mpc}^{-3}$ (Portegies Zwart & McMillan, 2000;
Rodriguez et al., 2016; Kritos & Cholis, 2020), $dV_{c}/dz$ is the comoving
volume and $(1+z)^{-1}$ accounts for the time dilation. The comoving volume is
given by (Hogg, 1999),
$\displaystyle\frac{dV_{c}}{dz}$ $\displaystyle=\frac{4\pi
c^{3}}{H_{0}^{3}E(z)}\left(\int_{0}^{z}\frac{dz^{\prime}}{E(z^{\prime})}\right)^{2},$
(24) $\displaystyle E(z)$
$\displaystyle=\sqrt{\Omega_{M}(1+z^{\prime})^{3}+\Omega_{\Lambda}}$ (25)
where $\Omega_{K}=0,\ \Omega_{M}=0.3$ and $\Omega_{\Lambda}=0.7$ (Planck
Collaboration et al., 2016).
As a conservative estimate, we take the mergers rate $\Gamma_{\rm old}$ to be
$\Gamma_{\rm old}\sim N_{\rm merge}/\tau_{\rm GC}$ where $\tau_{\rm GC}$ is
taken to be $10\ \rm{Gyr}$. In Fig. 8 we present the cumulative rate of
expected mergers in old GCs (in blue). There are two types of contributions to
the rate: eccentric binaries, such as these with initial eccentricity of $2/3$
that corresponds to the mean value of a thermal eccentricity distribution,
that will merge within short timescales, i.e. with negligible delay time.
These practically follow the star formation rate. In this case observed
contributions are likely to rise from YMCs. The second case corresponds to low
eccentricity/circular binaries, in which there will be a delay time that
corresponds to a typical time of $\sim 10^{4}\rm{Myr}$. These contributions
will be observed in old GCs.
Figure 8: The cumulative contribution to GWs rate from YMCs (in red) and old
GCs (in blue), from the gas hardening channel, as derived from the GDF. The
shaded area relates to the range of parameters. The black line relates to the
range of rates inferred by LVK. In the case of circular binaries, the rate
will be dominated by old GCs, while for eccentric it will be dominated by
YMCs.
In this case, the major contribution from our channel to currently observable
GW-sources would not originate from old GCs, but from YMCs. We define a YMC as
a cluster formed later than redshift 2 and mass $>10^{4}\ M_{\odot}$ such that
we assume for that case that the 2P formation already occurred. The formation
rate of YMCs follows the star formation rate (SFR), which enables us to write
the merger rate from YMC as (Banerjee, 2021)
$\displaystyle\mathcal{R}_{\rm young}(z)=\frac{N_{\rm mrg}}{N_{\rm
samp}}\frac{1}{2\Delta t_{\rm obs}}\frac{\int_{M_{\rm cl,low}}^{M_{\rm
cl,high}}\Phi_{\rm CLMF}(M)dM}{\int_{M_{\rm GC,low}}^{M_{\rm
GC,high}}\Phi_{\rm CLMF}(M)dM}\times$ (26)
$\displaystyle\times\frac{\int_{0}^{z}\Psi_{\rm
SFR}(z_{f})dz_{f}}{\int_{3}^{6}\Psi_{\rm SFR}(z_{f})dz_{f}}\rho_{\rm GC}$
$N_{\rm mrg}$ is the number of mergers expected in $N_{\rm samp}$ clusters,
$\Delta t_{\rm obs}=0.15\ \rm{Gyr}$ (Banerjee, 2021) is the uncertainty in the
cluster formation epoch, $\Phi_{\rm CLMF}\propto M^{-2}$ (e.g. Portegies Zwart
et al., 2010) is the cluster mass function and we consider $[M_{\rm
cl,low},M_{\rm cl,max}]=[10^{4},10^{5}]\ M_{\odot}$ as the available mass
range for YMCs and $[M_{\rm GC,low},M_{\rm GC,high}]=[10^{5},10^{6}]\
M_{\odot}$ as the typical present-day masses for GCs. $\rho_{\rm GC}$ is the
observed number density of GCs per unit comoving volume. $\Psi_{\rm SFR}(z)$
is the cosmic star formation rate, which is given by (Madau & Dickinson,
2014),
$\displaystyle\Psi_{\rm
SFR}(z)=0.01\frac{(1+z)^{2.6}}{1+[(1+z)/3.2]^{6.2}}M_{\odot}\rm{Mpc}^{-3}\rm{yr}^{-1}$
(27)
We consider $N_{\rm mrg}/N_{\rm samp}=N_{\rm merge}$ and spatial densities in
the range $[0.33,2.57]\ \rm{Mpc}^{-3}$, following (Banerjee, 2021) and
references therein. In Fig. 8, we present the cumulative rate of expected
mergers in YMCs and GCs. For YMCs, the rate follows the star formation rate
(in general, with a small correction due to the delay time – which is short),
and hence peaks in relatively low redshifts. For the eccentric case, the
dominant contribution will rise from YMC, while for circular ones the dominant
contribution is from GCs.
It should be noted that in general, there could be a non-negligible delay time
for the binaries merger. However, for all the parameters we checked for the
disk configuration, the merger timescales are extremely short and are
negligible in terms of redshifts.
The total contribution to the GWs merger rate from YMCs is in the range
$\mathcal{R}_{\rm young}\approx[0.08,25.51]\ \rm{Gpc}^{-3}\rm{yr}^{-1}$, which
intersects the expected range of LVK, i.e. $23.9_{-8.6}^{+14.3}\ \rm{Gpc^{-3}\
yr^{-1}}$ (Abbott et al., 2021), where the range is bracket by the models with
lowest and and highest rates (see Table 1).
model | $\mathcal{R}_{\rm YMC}(z\leq 1)\ [\rm{Gpc^{-3}\ yr^{-1}}]$ | model | $\mathcal{R}_{\rm YMC}(z\leq 1)\ [\rm{Gpc^{-3}\ yr^{-1}}]$
---|---|---|---
$\rho_{-}c_{s-}n_{+}$ | 0.32 | $\rho_{+}c_{s-}n_{+}$ | 2.55
$\rho_{-}c_{s-}n_{-}$ | 0.08 | $\rho_{+}c_{s-}n_{-}$ | 0.64
$\rho_{-}c_{s+}n_{+}$ | 3.28 | $\rho_{+}c_{s+}n_{+}$ | 25.51
$\rho_{-}c_{s+}n_{-}$ | 0.82 | $\rho_{+}c_{s+}n_{-}$ | 6.35
Table 1: Rates from YMCs for redshifts $z\leq 1$, for different choices of
parameters. $\rho_{\pm}$ correspond to $\rho_{\rm GC}=0.33E^{3}(z)\
\rm{Mpc}^{-3}$ and $\rho_{\rm GC}=2.57E^{3}(z)\ \rm{Mpc}^{-3}$, $c_{s\pm}$
correspond to $c_{s-}=1\ \rm{km/sec}$ and $c_{s+}=10\ \rm{km/sec}$ and
$n_{\pm}$ corresponds to high density of progenitors and low fraction of hard
binaries ($n_{+}$, segregated environment) and low density of progenitors and
high fraction of hard binaries ($n_{-}$, non-segregated environment). These
correspond also to different fractions of soft/hard binaries, see subsec.
3.1.2. Here we present the rates expected for initially eccentric binaries
(e.g. $e_{0}=0.66$)
In table 1 we present our calculated rates for different choices of
parameters. As expected, higher gas densities lead to larger merger rates and
higher sound speeds correspond to thicker disks that host more stars and hence
yield more mergers.
### 4.5 GW merger properties
Given the early epoch of gas replenishment, gas-catalyzes mergers operate on
primordial binaries in the clusters. The merging components are therefore
likely distributed similar to the primordial distribution of binary
components. However, even very wide binaries can merge in this channel
compared with only relatively close binaries merging in e.g. isolated binary
evolution channels for GW mergers. This could give rise to significant
differences in the expected masses and mass-ratios of the merger objects.
Interaction with gaseous media could excite binaries to high eccentricities,
due to the dependence of the drag force on the relative velocity between the
gas and the binary, which changes along the orbit such that the effect is the
strongest at the apocenter. Evolution of eccentric binaries hence shorten
significantly the expected merger timescales, as larger separations correspond
to small pericenters, in which GWs could dominate the evolution. In this case
GW-emission would dump the eccentricites and GW-mergers would generally be
circular in the VLK bands. However, if the combined gas-catalyzed and GW-
emission binary shrinkage is slower (e.g. for circular-orbits or lower gas-
densities), where 3-body encounters dominate the final evolution, and higher
eccentricities can be achieved for at least a small fraction of the mergers,
similar to the dynamical channels for GW-sources explored in the past.
We should remark in passing on the possibility of triples. In triples, the
outer component migrates faster than the inner binary, potentially leading to
an unstable configuration and effective chaotic three body interaction (see
e.g. a the reversed case of triples expanding due to mass-loss, leading to
similar instability in Perets & Kratter, 2012), such chaotic encounters could
give rise to eccentric mergers. This possibility and its potential
contribution will be discussed elsewhere.
## 5 Discussion
In the following we discuss our results and implications for the evolution of
binaries and singles in gas-enriched GCs.
### 5.1 Other aspects of binary evolution
As we showed, the presence of gas modifies the binary population in GCs. It
leads to an efficient merger of binaries, together with the formation of
binaries via the L2 and L3 mechanism (which was initially used to study the
formation of Kuiper-belt binaries (Goldreich et al., 2002) and recently was
applied to calculate the formation rate in AGN disks Tagawa et al., 2020).
After the gas dissipation, the initial properties of the binary, as well as
the gas, dictate the final separation, to which all the binaries with initial
separations larger than the final separations will converge.
Therefore, gas hardening leaves a significant signature on the binary
population and its properties, which sets the ground for further dynamical
processes in general and specifically for later dynamical mergers.
In addition to the contribution of the channel to the total rate of GWs, the
modification of the properties of binaries (mass, separation etc.) caused by
the gas hardening, sets unique initial conditions for the other GWs channels.
This will induce an indirect signature of the gas hardening on the expected
observed mergers. We introduced analytical results that could in principle be
plugged in as initial conditions for the later evolution of GCs, and the
dynamical channels for GW production in such environments. The binary
abundance changes due to the gas hardening, since a significant fraction of
binaries could merge, while others form. Furthermore, additional L2/L3-formed
binaries could participate and produce GW sources, beyond the primordial
binaries considered here. Nevertheless, since stars might be far more abundant
than BHs, L2/L3 processes might mostly produce mixed BH-star binaries and may
not contribute to the GW merger rate, but may form other exotic binaries such
as X-ray sources etc., and/or produce micro tidal disruption events (Perets et
al., 2016, disruption of stars by stellar black holes).
### 5.2 Implications for other gas-rich environments
The gas-catalyzed dynamics discussed here could take place in any other gas-
rich environments, with the proper scaling. While enhanced GW merger rates
were discussed in the context of AGN disks (McKernan et al., 2012; Stone et
al., 2017; Tagawa et al., 2020, and references therein), they are usually
discussed in the context of the evolution of a particular binary or the
overall BH-merger rate. However, in those cases too, the whole binary
populations, of both compact objects and stars will change their properties.
A very similar process could take place for young binaries embedded in star
formation regions Korntreff et al. (2012). In this case, the effect is limited
to a shorter timescale and compact objects might not yet have formed, and are
therefore not directly affected (but their progenitor massive stars are).
### 5.3 YMCs and very massive clusters
YMCs are still relatively little studied in the context of the production of
GW sources, although their contribution to the total estimated rate of GWs is
potentially not negligible (Portegies Zwart & McMillan, 2000; Banerjee, 2021).
In these clusters, gas can be present up to smaller redshifts, such that the
effect from the channel we suggested for GWs could potentially be observed.
Hence, their overall contribution to the currently observed merger rates in
LVG will be more significant (as can be seen also in Fig. 8). Our rate
estimates discussed below, account for both GCs and their younger
counterparts, YMCs.
### 5.4 Dynamics in gas enriched clusters
All the dynamical processes that take place in the early stages of GCs
evolution might be affected by the presence of gas, e.g. few-body dynamics.
One aspect is that wide binaries that formed during the gas epoch are
protected from evaporation by the gas hardening, as they harden within
timescales shorter than the typical evaporation/ionization timescales.
GDF could also enhance mass segregation (Indulekha, 2013; Leigh et al., 2014).
The energy dissipation leads to a change in the velocity dispersion in short
timescales, such that massive objects will fall towards the center of the
cluster. Moreover, since the more massive objects are prone to merge (as can
be seen from eq. 7, or visually from Fig. 5), the relaxation will be affected
by the modified mass function induced by the gas hardening.
#### 5.4.1 GW recoils, spins and mass-gap objects
It is possible that gas-accretion onto binaries and not only GDF (e.g Roupas &
Kazanas, 2019) could affect their evolution. In particular, sufficient
accretion might align the BH spins and orbits, especially if some circumbinary
disk forms around the binaries, in which case the GW-recoil velocity following
mergers is likely to be small, and allow a larger fraction of merged, now more
massive BHs to be retained in the cluster. This in turn would affect the later
dynamics in the clusters, and the resulting mergers in the dynamical formation
channels operating in the clusters. This could then potentially give rise to
higher fraction of BHs reaching high (even mass-gap) masses following repeated
mergers. The spin evolution and accretion, however, require more detailed
study, which is beyond the current scope.
The spin evolution of binaries will be affected by the role played by
dynamical encounters, as well as the direction of the gas relative to the
binary. In some cases, initially misaligned binaries could be aligned later
due to gas accretion, but when dynamical encounters are dominant, the spins
won’t be aligned.
### 5.5 Implications for neutron stars and white dwarfs: accretion &
explosive transients
The focus of the current paper is the merger of BHs and the production of GW
sources due to gas interactions in multiple population clusters. However, the
evolution of stars and other compact objects such as WDs and NSs could be
significantly affected in similar ways. Though some of these aspects are
discussed in a companion paper (Perets, 2022)), we postpone a detailed
exploration of these objects to a later stage, and only briefly mention
qualitatively some potentially interesting implications.
A fraction of the gas could be accreted on objects in the cluster. Gas
accretion changes the velocities of the accretors and the overall mass
function of objects in the cluster, such that there is a shift towards higher
masses (e.g. Leigh et al., 2014), that might affect the dynamical GWs channels
in clusters that operate after the gas-replenishment epoch, since we enrich
the abundance of massive objects which are likely to be the progenitors of
GWs. Stars that accrete gas could evolve into compact objects that in turn
might produce novae. Enhanced accretion in the early stages of the cluster
evolution could potentially modify the novae rates and properties (Maccarone &
Zurek, 2012) and the production of accretion-induced collapse of WDs into NSs
(Perets, 2022).
We should point out that our scenario suggests a robust merger not only of
BHs, but also of neutron stars (NSs) and white dwarfs (WDs). These mergers
might leave unique signatures. Besides their contribution to the production of
short GRBs and GW sources, binary NSs mergers are a promising channel to the
production of heavy elements via r-process (e.g. Freiburghaus et al., 1999),
and would affect the chemical evolution of the clusters.
Thermonuclear explosions of WDs could produce type Ia SNe, whether via single
degenerate channel (WD and a non-degenerate companion, Whelan & Iben, 1973) or
double degenerate (two WDs, Iben & Tutukov, 1984). Both of these channels will
be affected by the gas accretion. First, as we mentioned (Leigh et al., 2014
and references therein), the mass function will change. This is turn might
change the characteristics of the SNe and their rate. Furthermore, regardless
of the mass variation, a large fraction of the compact object binaries are
expected to merge within short timescales, which will also affect the SNe
rate.
Mergers of WDs could yield a remnant merged object with small or absent natal
kick and hence constitute another channel for NS formation. Accretion could
potentially change the retention fraction, and potentially explain the
retention problem in the formation of pulsars (Perets, 2022).
### 5.6 Constraining the parameters of the cluster
The amount and origin of gas in GCs during the formation of 2P stars are still
uncertain (Bekki, 2017). In this channel, we suggest that the amount of gas
dictates a final SMA, such that the separation distribution/GW rate could be
used to constrain the gas abundance in the cluster and its lifetime.
For sufficiently low gas densities (or lower densities following gas
depletion), gas hardening is not efficient enough to lead to a merger. In this
case, the terminal SMA of the binary will exceed $a_{\rm GW}$, such that GWs
will not be emitted without a further dissipation process. However, if the gas
remains for longer timescales, further hardening will occur. For the whole
parameter space we considered, the early stages of the hardening process are
very efficient, i.e. wide binaries harden and become hard binaries on short
timescales.
This channel of production of GWs-sources could serve as a tracer to later
star formation, as it is coupled to the gas that accompanies this formation.
The amount of gas, its decay with time are determined by the star formation
history. Since these parameters play a role in gas hardening and hence on the
final separation distribution at the end of the gas epoch, they could
potentially serve to constrain the 2P gas and star-formation phase, and may
help explain some of the differences between 1P and 2P stellar populations.
For example, we might speculate that the inferred difference between the 1P
and 2P binary fractions (e.g. Lucatello et al. (2015)) could be explained by
gas-catalyzed hardening and mergers of main-sequence stars residing in the
gaseous region. Such 1P binaries which also accrete significant mass of 2P gas
would appear and be part of the 2P populations, while outside the gas regions
binaries are not affected. In this case some of the 2P binaries preferentially
merge compared with 1P stars outside the 2P gas region, leading to an overall
smaller binary fractions.
That being said, the many uncertainties and degeneracies involved might be
challenging in directly connecting current populations with the early
conditions directly.
### 5.7 Caveats/future directions
In the following we discuss potential caveats of our model/scenario.
$\bullet$ The specific scenario for formation of 2P stars is still
unknown/debated, and hence there are large uncertainties in the amount of gas
in the cluster, its source during the different stages of evolution. Moreover,
some explanations for the different chemical composition of the so-called 2P
stars might require lower gas masses than the total mass of 2P stars. In these
cases, the phenomena we described might be somewhat suppressed, though, as we
have shown even lower gas densities could be highly effective, and will not
qualitatively change the results.
$\bullet$ The expected production rates of GW sources depend on the initial
parameters of the clusters we consider, including the gas densities, stellar
and binary populations, star formation histories etc. All of these contain
many uncertainties, which we did not directly address in this initial study,
limited to a small number of models as to provide an overall estimate to
bracket the expected GW rates from this channel. Nevertheless, all of our
models show that gas-catalyzed mergers in multiple population clusters could
produce a significant and even major contribution to the GW-merger rate, and
could play a key role in the general evolution of stars and binaries in such
clusters.
$\bullet$ The interaction of gas with binaries is complex and includes many
physical aspects. Here we assumed that the gas density in the cluster, or at
least in the region in which the binaries evolve, is spatially constant. Most
of the gas should be concentrated in the star-forming region, preferentially
towards the inner parts of the cluster. Outer parts of the cluster might be
more dilute. Future study could relax the simplified assumption of a constant
spatial density and account for a more detailed distribution of gas, stars and
binaries.
$\bullet$ We assumed that the relative velocity between the objects and the
gas is dominated by the Keplerian velocity of the binary component become
dominant. A more realistic approach, but requiring a detailed Monte-Carlo or
N-body simulation could account for the detailed velocity distribution of
binaries in the cluster.
$\bullet$ As we mentioned in subsection 5.5, objects embedded in gas could
accrete from it and change their mass over time. As a result, their dynamics
will change both in the cluster and as binaries (Roupas & Kazanas, 2019). Here
we considered constant masses throughout the evolution, and neglected the
effects of gas accretion. This is a somewhat conservative assumption, in
regard to catalysis of mergers, as more massive objects are prown to merge
even faster in gas (see eq. 7 and Fig. 5).
$\bullet$ We considered several choices for the gas depletion, assuming an
exponential decay, with a fiducial model of $50\ \rm{Myr}$ and a lifetime of
$100\ \rm{Myr}$. However, the formation epochs of stars could set different
scenarios, e.g. in which gas is abundant in the cluster for longer timescales
of $\sim 100\ \rm{Myr}$, but only intermittently (Bekki, 2017), which will
change the picture, or when several wide scale gas replenishmet episode occur
over timescale of even many hundreds of Mys or even Gyrs, as might be the case
for nuclear clusters.
$\bullet$ In our analysis we considered for simplicity only equal mass
binaries. Though we don’t expect a major change in the results, the
generalization to binaries with different masses is more complex and requires
more detailed population studies, beyond the scope of the current study.
$\bullet$ It should be noted that there were studies that suggested limited
efficiency of gas dynamical friction (e.g. Li et al., 2020; Toyouchi et al.,
2020) than considered here. A more detailed comparison is left out for further
studies.
$\bullet$ Although the initial parameters of our disk suggest a thick disk, in
later stages the disk will be thinner and finally fragment if it to enable
star formation. Hence, for these stages/initial thin disks, the gas hardening
epoch should be limited to the regime in which the disk is stable.
$\bullet$ We restrict ourselves to binaries which are not likely to be
disrupted by interactions with other stars. Further disruptions could take
place and are encapsulated in $f_{\rm bin,surv}$ (see eq. 22.
## 6 Summary
In this paper we discussed the evolution of binaries in gas-enriched
environments which likely existed at the early stage multiple-population
clusters. We showed the binary interaction with the ambient gas-environment
significantly affects their evolution and give rise to major changes in binary
population in the cluster and its properties.
Binaries interaction with gas has been extensively studied over the last few
years in the context of AGN disks. Here we show that the environments of
multiple population GCs and YMCs similarly give rise to important effects. In
particular, focusing on the production of GW sources from binary BH mergers,
we find that gas-enriched multiple population clusters could provide a
significant and possibly major contribution to the production of GW sources of
up to a few tens of Gpc-1yr-1, comparable with the GW-sources production rate
inferred by VLK for the local universe. These might even be higher once
formation of new binaries due to gas-assisted capture is considered (to be
discussed in a follow-up paper).
Moreover, we expect catalyzed mergers of other compact objects such as NSs and
WDs, and of binary main-sequence and evolved stars to give rise to the
enhanced rate of a wide range of merger outcomes, producing a range of
transient events such as supernovae, GRBs and the formation of X-ray binaries
and stellar mergers, which will be discussed elsewhere.
Furthermore, our findings on the overall evolution of binary populations are
relevant for other gas-enriched environments such as AGN disks.
Finaly, our focus here was on binary BH mergers in multiple-population cluster
environments, but we point out that the early gas enriched phase of such
clusters (which in practice is relevant to the vast majority of GCs, given
that most GCs show multiple populations) significantly affect all the stellar
and binary populations, and the overall dynamics inside GCs. Hence the current
modeling of the typical initial conditions in GCs and their evolution might
need to be fundamentally revised.
## ACKNOWLEDGMENTS
We like to thank to the anonymous referee for important comments and points
that significantly improved the manuscript. We would also like to thank
Aleksey Generozov, Johan Samsing ,Jim Fuller, Kyle Kremer, Noam Soker and
Evgeni Grishin for fruitful discussions. MR acknowledges the generous support
of Azrieli fellowship. HBP and MR acknowledge support from the from the
European Union’s Horizon 2020 research and innovation program under grant
agreement No 865932-ERC-SNeX.
## References
* Abbott et al. (2016) Abbott, B. P., et al. 2016, Phys. Rev. Lett., 116, 061102, doi: 10.1103/PhysRevLett.116.061102
* Abbott et al. (2021) Abbott, R., et al. 2021, arXiv e-prints, arXiv:2111.03606. https://arxiv.org/abs/2111.03606
* Abbott et al. (2021) Abbott, R., Abbott, T. D., Abraham, S., et al. 2021, ApJ, 913, L7, doi: 10.3847/2041-8213/abe949
* Antoni et al. (2019) Antoni, A., MacLeod, M., & Ramirez-Ruiz, E. 2019, ApJ, 884, 22, doi: 10.3847/1538-4357/ab3466
* Antonini & Perets (2012) Antonini, F., & Perets, H. B. 2012, ApJ, 757, 27, doi: 10.1088/0004-637X/757/1/27
* Artymowicz et al. (1991) Artymowicz, P., Clarke, C. J., Lubow, S. H., & Pringle, J. E. 1991, ApJ, 370, L35, doi: 10.1086/185971
* Artymowicz et al. (1993) Artymowicz, P., Lin, D. N. C., & Wampler, E. J. 1993, ApJ, 409, 592, doi: 10.1086/172690
* Artymowicz & Lubow (1994) Artymowicz, P., & Lubow, S. H. 1994, ApJ, 421, 651, doi: 10.1086/173679
* Bahcall & Ostriker (1976) Bahcall, J. N., & Ostriker, J. P. 1976, Nature, 262, 37, doi: 10.1038/262037a0
* Banerjee (2021) Banerjee, S. 2021, MNRAS, 503, 3371, doi: 10.1093/mnras/stab591
* Baruteau et al. (2011) Baruteau, C., Cuadra, J., & Lin, D. N. C. 2011, ApJ, 726, 28, doi: 10.1088/0004-637X/726/1/28
* Bastian & Lardo (2018) Bastian, N., & Lardo, C. 2018, ARA&A, 56, 83, doi: 10.1146/annurev-astro-081817-051839
* Bate (2000) Bate, M. R. 2000, MNRAS, 314, 33, doi: 10.1046/j.1365-8711.2000.03333.x
* Bekki (2010) Bekki, K. 2010, ApJ, 724, L99, doi: 10.1088/2041-8205/724/1/L99
* Bekki (2011) —. 2011, MNRAS, 412, 2241, doi: 10.1111/j.1365-2966.2010.18047.x
* Bekki (2017) —. 2017, MNRAS, 469, 2933, doi: 10.1093/mnras/stx982
* Binney & Tremaine (2008) Binney, J., & Tremaine, S. 2008, Galactic Dynamics: Second Edition
* Bobrick et al. (2022) Bobrick, A., Zenati, Y., Perets, H. B., Davies, M. B., & Church, R. 2022, MNRAS, 510, 3758, doi: 10.1093/mnras/stab3574
* Carretta et al. (2009) Carretta, E., Bragaglia, A., Gratton, R. G., et al. 2009, A&A, 505, 117, doi: 10.1051/0004-6361/200912096
* Duffell et al. (2020) Duffell, P. C., D’Orazio, D., Derdzinski, A., et al. 2020, ApJ, 901, 25, doi: 10.3847/1538-4357/abab95
* Er et al. (2009) Er, X.-Y., Jiang, Z.-B., & Fu, Y.-N. 2009, Chinese Astron. Astrophys., 33, 139, doi: 10.1016/j.chinastron.2009.03.012
* Escala et al. (2004) Escala, A., Larson, R. B., Coppi, P. S., & Mardones, D. 2004, ApJ, 607, 765, doi: 10.1086/386278
* Freiburghaus et al. (1999) Freiburghaus, C., Rosswog, S., & Thielemann, F. K. 1999, ApJ, 525, L121, doi: 10.1086/312343
* Ginat & Perets (2021a) Ginat, Y. B., & Perets, H. B. 2021a, Physical Review X, 11, 031020, doi: 10.1103/PhysRevX.11.031020
* Ginat & Perets (2021b) —. 2021b, MNRAS, 508, 190, doi: 10.1093/mnras/stab2565
* Goldreich et al. (2002) Goldreich, P., Lithwick, Y., & Sari, R. 2002, Nature, 420, 643, doi: 10.1038/nature01227
* Gorti & Bhatt (1996) Gorti, U., & Bhatt, H. C. 1996, MNRAS, 283, 566, doi: 10.1093/mnras/283.2.566
* Gratton et al. (2019) Gratton, R., Bragaglia, A., Carretta, E., et al. 2019, A&A Rev., 27, 8, doi: 10.1007/s00159-019-0119-3
* Grishin & Perets (2016) Grishin, E., & Perets, H. B. 2016, ApJ, 820, 106, doi: 10.3847/0004-637X/820/2/106
* Heggie (1975) Heggie, D. C. 1975, MNRAS, 173, 729, doi: 10.1093/mnras/173.3.729
* Hénault-Brunet et al. (2015) Hénault-Brunet, V., Gieles, M., Agertz, O., & Read, J. I. 2015, MNRAS, 450, 1164, doi: 10.1093/mnras/stv675
* Hogg (1999) Hogg, D. W. 1999, arXiv e-prints, astro. https://arxiv.org/abs/astro-ph/9905116
* Iben & Tutukov (1984) Iben, I., J., & Tutukov, A. V. 1984, ApJS, 54, 335, doi: 10.1086/190932
* Indulekha (2013) Indulekha, K. 2013, Journal of Astrophysics and Astronomy, 34, 207, doi: 10.1007/s12036-013-9175-7
* Korntreff et al. (2012) Korntreff, C., Kaczmarek, T., & Pfalzner, S. 2012, A&A, 543, A126, doi: 10.1051/0004-6361/201118019
* Kritos & Cholis (2020) Kritos, K., & Cholis, I. 2020, Phys. Rev. D, 102, 083016, doi: 10.1103/PhysRevD.102.083016
* Leigh et al. (2013) Leigh, N. W. C., Böker, T., Maccarone, T. J., & Perets, H. B. 2013, MNRAS, 429, 2997, doi: 10.1093/mnras/sts554
* Leigh et al. (2014) Leigh, N. W. C., Mastrobuono-Battisti, A., Perets, H. B., & Böker, T. 2014, MNRAS, 441, 919, doi: 10.1093/mnras/stu622
* Li et al. (2016) Li C., de Grijs R., Deng L., Geller A. M., Xin Y., Hu Y., Faucher-Giguère C.-A., 2016, Natur, 529, 502. doi:10.1038/nature16493
* Li et al. (2020) Li X., Chang P., Levin Y., Matzner C. D., Armitage P. J., 2020, MNRAS, 494, 2327. doi:10.1093/mnras/staa900
* Lucatello et al. (2015) Lucatello, S., Sollima, A., Gratton, R., et al. 2015, A&A, 584, A52, doi: 10.1051/0004-6361/201526957
* Maccarone & Zurek (2012) Maccarone, T. J., & Zurek, D. R. 2012, MNRAS, 423, 2, doi: 10.1111/j.1365-2966.2011.20328.x
* Madau & Dickinson (2014) Madau, P., & Dickinson, M. 2014, ARA&A, 52, 415, doi: 10.1146/annurev-astro-081811-125615
* Mastrobuono-Battisti & Perets (2013) Mastrobuono-Battisti, A., & Perets, H. B. 2013, ApJ, 779, 85, doi: 10.1088/0004-637X/779/1/85
* Mastrobuono-Battisti & Perets (2016) —. 2016, ApJ, 823, 61, doi: 10.3847/0004-637X/823/1/61
* Mastrobuono-Battisti & Perets (2020) —. 2020, arXiv e-prints, arXiv:2011.12292. https://arxiv.org/abs/2011.12292
* McKernan et al. (2012) McKernan, B., Ford, K. E. S., Lyra, W., & Perets, H. B. 2012, MNRAS, 425, 460, doi: 10.1111/j.1365-2966.2012.21486.x
* McKernan et al. (2018) McKernan, B., Ford, K. E. S., Bellovary, J., et al. 2018, ApJ, 866, 66, doi: 10.3847/1538-4357/aadae5
* Miller & Hamilton (2002) Miller, M. C., & Hamilton, D. P. 2002, MNRAS, 330, 232, doi: 10.1046/j.1365-8711.2002.05112.x
* Moody et al. (2019) Moody, M. S. L., Shi, J.-M., & Stone, J. M. 2019, ApJ, 875, 66, doi: 10.3847/1538-4357/ab09ee
* Muñoz et al. (2020) Muñoz, D. J., Lai, D., Kratter, K., & Miranda, R. 2020, ApJ, 889, 114, doi: 10.3847/1538-4357/ab5d33
* Muñoz et al. (2019) Muñoz, D. J., Miranda, R., & Lai, D. 2019, ApJ, 871, 84, doi: 10.3847/1538-4357/aaf867
* Ostriker (1999) Ostriker, E. C. 1999, ApJ, 513, 252, doi: 10.1086/306858
* Ostriker (1983) Ostriker, J. P. 1983, ApJ, 273, 99, doi: 10.1086/161351
* Perets (2022) Perets, H. B. 2022, arXiv e-prints, arXiv:2201.11136. https://arxiv.org/abs/2201.11136
* Perets & Kratter (2012) Perets, H. B., & Kratter, K. M. 2012, ApJ, 760, 99, doi: 10.1088/0004-637X/760/2/99
* Perets et al. (2016) Perets, H. B., Li, Z., Lombardi, James C., J., & Milcarek, Stephen R., J. 2016, ApJ, 823, 113, doi: 10.3847/0004-637X/823/2/113
* Perets & Murray-Clay (2011) Perets, H. B., & Murray-Clay, R. A. 2011, ApJ, 733, 56, doi: 10.1088/0004-637X/733/1/56
* Peters (1964) Peters, P. C. 1964, Physical Review, 136, 1224, doi: 10.1103/PhysRev.136.B1224
* Planck Collaboration et al. (2016) Planck Collaboration, Ade, P. A. R., Aghanim, N., et al. 2016, A&A, 594, A13, doi: 10.1051/0004-6361/201525830
* Portegies Zwart & McMillan (2000) Portegies Zwart, S. F., & McMillan, S. L. W. 2000, ApJ, 528, L17, doi: 10.1086/312422
* Portegies Zwart et al. (2010) Portegies Zwart, S. F., McMillan, S. L. W., & Gieles, M. 2010, ARA&A, 48, 431, doi: 10.1146/annurev-astro-081309-130834
* Quinlan (1996) Quinlan, G. D. 1996, New A, 1, 35, doi: 10.1016/S1384-1076(96)00003-6
* Renzini et al. (2015) Renzini, A., D’Antona, F., Cassisi, S., et al. 2015, MNRAS, 454, 4197, doi: 10.1093/mnras/stv2268
* Rodriguez et al. (2016) Rodriguez, C. L., Chatterjee, S., & Rasio, F. A. 2016, Phys. Rev. D, 93, 084029, doi: 10.1103/PhysRevD.93.084029
* Roupas & Kazanas (2019) Roupas, Z., & Kazanas, D. 2019, A&A, 621, L1, doi: 10.1051/0004-6361/201834609
* Samsing et al. (2020) Samsing J., Bartos I., D’Orazio D. J., Haiman Z., Kocsis B., Leigh N. W. C., Liu B., et al., 2020, arXiv, arXiv:2010.09765
* Sana et al. (2012) Sana, H., de Mink, S. E., de Koter, A., et al. 2012, Science, 337, 444, doi: 10.1126/science.1223344
* Sigurdsson & Phinney (1995) Sigurdsson, S., & Phinney, E. S. 1995, ApJS, 99, 609, doi: 10.1086/192199
* Soker (2016) Soker N., 2016, NewAR, 75, 1. doi:10.1016/j.newar.2016.08.002
* Spitzer (1987) Spitzer, L. 1987, Dynamical evolution of globular clusters
* Stone et al. (2017) Stone, N. C., Metzger, B. D., & Haiman, Z. 2017, MNRAS, 464, 946, doi: 10.1093/mnras/stw2260
* Tagawa et al. (2020) Tagawa, H., Haiman, Z., & Kocsis, B. 2020, ApJ, 898, 25, doi: 10.3847/1538-4357/ab9b8c
* Tagawa et al. (2021) Tagawa H., Kocsis B., Haiman Z., Bartos I., Omukai K., Samsing J., 2021, ApJL, 907, L20. doi:10.3847/2041-8213/abd4d3
* Tagawa et al. (2022) Tagawa H., Kimura S. S., Haiman Z., Perna R., Tanaka H., Bartos I., 2022, ApJ, 927, 41. doi:10.3847/1538-4357/ac45f8
* Tang et al. (2017) Tang, Y., MacFadyen, A., & Haiman, Z. 2017, MNRAS, 469, 4258, doi: 10.1093/mnras/stx1130
* Tiede et al. (2020) Tiede C., Zrake J., MacFadyen A., Haiman Z., 2020, ApJ, 900, 43. doi:10.3847/1538-4357/aba432
* Toyouchi et al. (2020) Toyouchi D., Hosokawa T., Sugimura K., Kuiper R., 2020, MNRAS, 496, 1909. doi:10.1093/mnras/staa1338
* Vesperini et al. (2010) Vesperini, E., McMillan, S. L. W., D’Ercole, A., & D’Antona, F. 2010, ApJ, 713, L41, doi: 10.1088/2041-8205/713/1/L41
* Whelan & Iben (1973) Whelan, J., & Iben, Icko, J. 1973, ApJ, 186, 1007, doi: 10.1086/152565
* Zenati et al. (2020) Zenati, Y., Bobrick, A., & Perets, H. B. 2020, MNRAS, 493, 3956, doi: 10.1093/mnras/staa507
* Zenati et al. (2019) Zenati, Y., Perets, H. B., & Toonen, S. 2019, MNRAS, 486, 1805, doi: 10.1093/mnras/stz316
## Appendix A Fiducial Parameters
Symbol | Definition | Fiducial Value
---|---|---
$\tau_{\rm gas}$ | gas lifetime | $50\ \rm{Myr}$
$\tau_{\rm SG}$ | formation time of SG | $100\ \rm{Myr}$
$M_{\star}$ | total mass of stars in cluster | $10^{5}\ M_{\odot}$
$M_{\rm gas}$ | gas mass in the cluster | $3\times 10^{5}\ M_{\odot}$
$\rho_{\rm g,disk}$ | initial gas density in disk | $1.74\times 10^{6}\ M_{\odot}\rm{pc^{-3}}$
$h/r$ | scale-height | $0.23$
$\sigma_{\rm disk}$ | disk velocity dispersion | $10\ \rm{km/sec}$
$\bar{m}$ | average stellar mass | $0.5\ M_{\odot}$
$n_{\star}$ | stellar density | $10^{5}\ \rm{pc^{-3}}$
$n_{\star,disk}$ | stellar density in disk | $10^{5}\ \rm{pc^{-3}}$
$c_{s}$ | sound speed | $10\ \rm km/sec$
$\log\Lambda_{g}$ | gas Coulomb logarithm | 3.1
|
# The Attractor of the Replicator Dynamic in Zero-Sum Games
Oliver Biggar
CIICADA Lab
Australian National University
Acton, ACT, Australia
<EMAIL_ADDRESS>
&Iman Shames
CIICADA Lab
Australian National University
Acton, ACT, Australia
<EMAIL_ADDRESS>
###### Abstract
In this paper we characterise the long-run behaviour of the replicator dynamic
in zero-sum games (symmetric or non-symmetric). Specifically, we prove that
every zero-sum game possesses a unique global replicator attractor, which we
then characterise. Most surprisingly, this attractor depends _only_ on each
player’s preference order over their own strategies and not on the cardinal
payoff values, defined by a finite directed graph we call the game’s
_preference graph_. When the game is symmetric, this graph is a tournament
whose nodes are strategies; when the game is not symmetric, this graph is the
game’s response graph. We discuss the consequences of our results on chain
recurrence and Nash equilibria.
## 1 Introduction
Learning in the presence of other learning agents is an increasingly
fundamental topic in modern machine learning, motivated by its role at the
core of cutting-edge techniques like _learning from self-play_ (38; 39) and
_Generative Adversarial Networks_ (18). The challenge of these systems is
analysing their collective behavior, which is where learning theory intersects
with game theory. To quote Hofbauer and Sigmund (21), “a major task of game
theory [is] to describe the dynamical outcome of model games described by
strategies, payoffs and adaptive mechanisms." That is, when agents learn
collectively, what do they learn?
[scale=1]figs/OD2
Figure 1: A zero-sum game (left) and its associated preference graph (right).
The sink component of the graph consists of all profiles other than $(a,a)$.
Under the replicator dynamic, this game has a unique attractor which is the
_content_ of the sink component (Theorem 3.3). The attractor is the union of
the strategy spaces of the subgames $\\{a,b,c\\}\times\\{b,c\\}$ and
$\\{b,c\\}\times\\{a,b,c\\}$, represented by the shaded region on the graph.
Note that the strategy space of the game is 4-dimensional, with the attractor
a 3-dimensional region on the boundary.
In online learning, the best-known approaches use variants of the
Multiplicative Weights Update algorithm (MWU) (5). To achieve the ‘no-regret’
property, these algorithms typically decrease the step size as more samples
are observed. In the long-run, as the step size becomes small, the behavior of
a collection of MWU-playing agents converges to the flow of a famous
differential equation: the _replicator dynamic_ (41; 37; 22; 44; 36). This
model was originally inspired by biological models of evolution (40), and is a
central object of study in _evolutionary game theory_ , the subfield of game
theory which focuses on dynamic processes. Since its discovery, the replicator
has been extensively analysed by biologists, mathematicians, economists and
computer scientists (22; 36; 25). Indeed, just as MWU is the flagship
algorithm in online learning, the replicator is the flagship dynamic in
evolutionary game theory (21; 36).
figs/RPS
(a) Rock-Paper-Scissors.
figs/MP
(b) Matching Pennies
Figure 2: The preference graphs (Definition 2.2) of two zero-sum games: (2(a))
Rock-Paper-Scissors (symmetric), and (2(b)) Matching Pennies (non-symmetric).
Describing the ‘dynamical outcome’ of games under the replicator dynamic (and
hence over MWU) involves answering a basic question: _to which strategy
profiles do we converge over time?_ In dynamical systems, a system’s long-run
behavior is defined by its _attractors_ (35; 27). Attractors are sets of
points which are _invariant_ (points inside the set remain there for all
time), _asymptotically stable_ (points in some neighbourhood converge to the
set) and _minimal_ (they do not contain a smaller set with the first two
properties). Attractors are fundamental objects in dynamical systems theory,
and so since the discovery of the replicator, identifying its attractors has
been a major topic of research (45; 2; 44; 22; 25; 32; 29; 26; 42). A number
of now-classical partial results are known, establishing that _pure Nash
equilibria_ (28) are always replicator attractors, and in _symmetric_ games
the related concept of _Evolutionary Stable Strategies_ (40) are likewise
attractors111In zero-sum games, a related result proves that the _time-
average_ of the replicator (and MWU) always converges to the Nash equilibrium
(20; 23; 16). However, the time-average behavior is distinct from the day-to-
day or _last-iterate_ behavior (32).. More strongly, a strategy profile is an
attractor _if and only if_ it is a _strict (pure) Nash equilibrium_ (44; 36).
However, this result only describes the simplest attractors—those which
contain only a single point. Most games—especially zero-sum games—don’t have
pure Nash equilibria, and most attractors contain more than one point!
Instead, the trajectories of the replicator in zero-sum games are typically
periodic (26), and under MWU they are often _chaotic_ (11; 12; 6).
We conclude that, despite decades of research, the attractors of the
replicator remain largely unknown, even in zero-sum games, arguably the best-
studied special case (2; 20; 21; 22; 36; 33; 30; 6; 11) (see Section 1.1).
This is what we achieve in this paper: we characterise the attractors of the
replicator dynamic in every zero-sum game. Beyond this result, our concepts
and techniques shed new light on equilibria, the graph structure of games (8;
9) and the modelling of payoffs/losses in economics and machine learning.
Characterising the replicator attractors of zero-sum games is a valuable
development, but we believe the most remarkable aspect of this result is the
form that this attractor takes. Specifically, the attractor depends _only_ on
players’ discrete preferences over their strategies, and not on the cardinal
payoff values. These ‘preferences’ are captured in a directed graph we call
the _preference graph_. In non-symmetric games this graph coincides with the
game’s _response graph_ (8; 32)222We use the name _preference graph_ to unify
the symmetric and non-symmetric cases, and because we find the name “response
graph” can be confusing. The word ‘response’ sounds like a turn-based game,
when in fact the name is actually a contraction of the “better-response
relation”, an ordering which defines player’s _preferences_., an object which
has been of increasing interest in algorithmic game theory (10; 8),
particularly in relation to the replicator (32; 29; 9). The nodes of the
preference graph are the profiles of the game, and the arcs represent which
strategies players prefer, given the strategies of the other player. As an
example, consider Figure 2(b), which shows the preference graph of the
Matching Pennies game. In this game, player 1 prefers to match the choice of
player 2, and player 2 prefers to mismatch player 1. The arc
${(T,H)}$${(H,H)}$, for example, captures the fact that, given player 2 plays
Heads, player 1 ‘prefers’ Heads over Tails. In symmetric zero-sum games, like
Rock-Paper-Scissors (Figure 2(a)), the preference graph has an even simpler
form where each arc represents the preferred option between some pair of
strategies. For example, given Rock ‘beats’ Scissors, in a match-up of Rock
and Scissors, one prefers to play Rock, hence the arc ScissorsRock.
Conceptually, the preference graph stores the underlying combinatorial
structure of the game. Most game-theoretic concepts—including the replicator
dynamic and the Nash equilibrium—are defined by cardinal payoffs, which serve
as a numerical instantiation of the underlying preference structure. Our
result shows that the choice of representation of preferences by numbers has a
transient effect: two games with different payoffs but the same preferences
have the same long-run behavior, in that their attractors are identical. One
practical consequence is that computing the attractor is easy (we can do it by
traversing the preference graph). More fundamentally, this lends our
prediction stability in the face of uncertainty in our model, an important and
rare property in game theory (43).
### 1.1 Contributions and Related Work
The main result of the paper is Theorem 3.3, which characterises the
attractors of zero-sum games. For each zero-sum game, we prove that an
attractor exists, is unique and attracts all points on the interior of the
game. This attractor is precisely the _content_ (9) (Definition 3.1) of the
preference graph’s unique (8)333This uniqueness is a zero-sum property; non-
zero-sum games, such as the $2\times 2$ Coordination game (30), may have
preference graphs with multiple sink components, and thus can have multiple
attractors under the replicator dynamic (9). _sink connected component_ ,
which is a strongly connected component with no arcs from a node inside the
component to a node outside. The sink component is a set of pure profiles; its
content is the set of _mixed_ profiles whose support contains only profiles in
this component, which is always an invariant set under the replicator. A
recent result (9) demonstrated that a replicator attractor always contains the
content of a sink component. The challenging part of our proof is showing that
the content of the sink component is also asymptotically stable in any zero-
sum game, and so is an attractor. We demonstrate stability using a potential
function argument. Our choice of function derives from the preference graph:
specifically, we use the total probability mass over all sink component
profiles. The proof then separates into two cases, reflecting two standard
types of zero-sum game: symmetric and non-symmetric. In evolutionary game
theory these are often called _single-_ and _multiple-population_ games.
Single-population games are also called _population games_ (36) or _matrix
games_ (4). The replicator has a slightly different properties in each case.
We show that in _symmetric zero-sum games_ the preference graph possesses a
special simplified form. The remainder of the symmetric case is
straightforward. The non-symmetric case is more complex, and involves
embedding the flow of the replicator on the non-symmetric zero-sum game in the
flow of the replicator on a larger symmetric zero-sum game we call the
_symmetrised game_ (Definition 3.6), restricted to the subspace of product
distributions (Theorem 3.8). This allows for a much simpler definition of the
replicator, and makes clear the dependence on the preferences.
Long-run stability of strategy profiles under the replicator is a core topic
in evolutionary game theory (40), especially with regard to classical solution
concepts, such as Nash equilibria and _evolutionarily stable strategies_. See
the textbooks (21; 44; 36) for a summary. Some particularly relevant works are
that of Akin and Losert (2), who studied zero-sum games under the replicator,
proving a crucial _volume conservation_ property, analysed in depth by
Hofbauer (20). A sequence of further results (33; 30; 26; 42; 9) used this
property to bound asymptotically stable sets of the replicator. Zeeman (45)
performed an early study of replicator attractors, suggesting that the
qualitative behaviour of the replicator can be split into a finite number of
classes; our results show that in zero-sum games the qualitative behaviour is
defined by the preference graph, of which there are a finite number.
Additionally, Ritzberger and Weibull (34) showed that when a _subgame_ is
closed under “weakly better responses", then it is an attractor under the
replicator. The preference graph is defined by the weakly better responses,
and so our results can be considered a generalisation which handles the more
complex case where the sink component is not a subgame.
Despite these results, a general negative conclusion of this line of work was
that the replicator doesn’t converge to mixed equilibria (32), and moreover no
dynamic can converge to equilibria in all games (19; 7). Indeed equilibria are
generally computationally intractable (15; 14). Consequently, the algorithmic
game theory community has increasingly shifted towards new notions of dynamic
outcome which can predict the day-to-day behavior of computational agents in
games (25). To this end, _sink chain components_ were recently proposed (32)
as the outcome of dynamic games, with the replicator used as the motivating
example. Sink chain components are a slight generalisation of attractors; when
a replicator attractor exists, it is a sink chain component (Lemma 4.2). Sink
chain components are built on a concept called _chain recurrence_ (Definition
4.1), a generalisation of periodicity which forms the foundation of the
Fundamental Theorem of Dynamical Systems (13). Crucially, chain components are
grounded in computational considerations. Informally, two points are in the
same chain component if a bounded-precision computational device cannot
determine whether they lie on the same periodic orbit (30). Thus, finding
attractors of the replicator dynamic is motivated not only by dynamical
systems but also computer science: the attractor gives us the strongest
prediction of long-run behavior which is consistent with computation.
This chain recurrence approach (32) has inspired a number of new results on
games and the replicator dynamic (30; 31; 29; 9; 8). In zero-sum games, when a
_fully-mixed_ Nash equilibria exists, the behavior is essentially
unpredictable: the sink chain component is the whole game (33; 26; 30; 31).
Under MWU, we observe chaotic behavior (11; 12) in these games. Further, when
a fully-mixed NE does not exist, (33; 26) showed that all fully-mixed strategy
profiles converge to the subgame containing the equilibrium, called the
_essential subgame_. However, this convergence is not ‘uniform’, and so the
essential subgame is typically _not_ asymptotically stable (see Section 4),
one of the defining properties of attractors and sink chain components (3).
Instead, interior profiles which are _arbitrarily close_ to the essential
subgame may move far away before returning. Consequently, like the mixed Nash
equilibrium itself, the essential subgame may not be a plausible prediction of
bounded-precision computational agents.
The story so far of predicting the replicator seems generally negative. Our
results tell a different, more positive story: we characterise the
attractor/sink chain component of the replicator, which, while larger than the
essential subgame, is the smallest ‘computationally plausible’ outcome of the
game, in the sense of chain recurrence. What’s more, being defined by discrete
preferences, the solution is somehow natural. Our shift away from Nash-based
methods also reveals new insight. Prior approaches suggested a connection
between equilibria and chain recurrence in zero-sum games (30; 26). We find
instead that chain recurrence in zero-sum games is entirely defined by the
preference graph (Corollary 4.3). The previous findings are now explained by a
non-trivial connection between equilibria and the preference graph: the
existence of a fully-mixed equilibrium implies strong connectedness of the
preference graph (Lemma 4.4).
## 2 Preliminaries
In game theory, a game is defined by a triple consisting of the _players_ ,
_strategy sets_ for each player, and _payoffs_. A combination of strategies
for each player is called a _strategy profile_ or simply a profile, and for
each profile there is a real-valued payoff to each player. In this paper we
focus on two-player games, where we denote the players by the integers 1 and 2
and their strategy sets by $S_{1}$ and $S_{2}$. The strategy names are simply
labels, so we assume $S_{1}=[n]:=\\{i\in\mathbb{N}_{0}|\ i<n\\}$ and
$S_{2}=[m]:=\\{i\in\mathbb{N}_{0}|\ i<m\\}$. The profiles are the pairs
$S_{1}\times S_{2}$. We call this an $n\times m$ game. An $n\times m$ game is
defined by a pair $(A,B)$ of matrices, representing the payoffs to players 1
and 2 respectively. We focus on _zero-sum games_ , which we represent by a
single matrix $M\in^{n\times m}$, implicitly assumed to be the payoffs for the
first player, which defines the game $(M,-M^{T})$. That is, in a profile
$(s_{1},s_{2})$, player 1 receives $M_{s_{1},s_{2}}$ and player 2 receives
$(-M^{T})_{s_{2},s_{1}}=-M_{s_{1},s_{2}}$. A _subgame_ of a game is formed by
choosing subsets $T_{1}\subseteq S_{1}$ and $T_{2}\subseteq S_{2}$ of each
player’s strategy sets and restricting the game to the profiles in
$T_{1}\times T_{2}$. We typically represent a subgame by its product set of
profiles $T_{1}\times T_{2}$.
_Symmetric_ games consist of a single set of strategies, denoted $S$, with
$|S|=n$ and a matrix $M\in^{n\times n}$. In evolutionary game theory, we think
of symmetric games as having a single population where different subtypes
evolve in competition with the other subtypes. In symmetric games, ‘strategy
profiles’ are simply individual strategies. Again, the payoffs are defined by
a matrix $M\in^{n\times n}$. A symmetric game $M$ is zero-sum if $(M,M)$
defines a zero-sum game, that is $M=-M^{T}$. This is equivalent to requiring
that $M$ is _anti-symmetric_. Thus, there is a natural one-to-one
correspondence between anti-symmetric real matrices and symmetric zero-sum
games.
A _mixed strategy_ is a distribution over a player’s strategies, and a _mixed
profile_ is an assignment of a mixed strategy to each player. We sometimes
refer to a profile as a _pure profile_ to distinguish it from a mixed profile.
If $x$ is a mixed strategy for player $i$ and $s$ is a strategy for that
player, we write $x_{s}$ for the $s$-entry of $x$. In non-symmetric games, we
denote mixed strategies by pairs $(x,y)$ where $x$ and $y$ are distributions
over the first and second player’s strategies, respectively. In symmetric
games, mixed profiles are just distributions over $S$, considered as the
‘population distribution’. The _support_ of a mixed strategy $x$ is the set of
strategies $s$ where $x_{s}$ is non-zero. We define the _support_ of a mixed
profile to be the set of profiles whose strategies are in the support of each
player’s mixed strategy. As distributions over a finite set, mixed strategies
can be naturally embedded in the standard probability simplex in Euclidean
space, by choosing some arbitrary ordering of the strategies in $S_{1}$ and
$S_{2}$. We denote these spaces by $\Delta(S_{1})$ and $\Delta(S_{2})$. The
set of mixed profiles is the product $\Delta(S_{1})\times\Delta(S_{2})$, which
we call this the _strategy space_ of the game. In a symmetric game this is
$\Delta(S)$. It also has a natural embedding in Euclidean space, so we can
talk about geometric properties of sets of mixed profiles, which we also refer
to as ‘points’ in strategy space.
The payoffs of a game extends naturally to mixed profiles. In a non-symmetric
zero-sum game $(M,-M^{T})$, the _expected payoff_ of a mixed profile $(x,y)$
is $y^{T}Mx$ to player 1 and $x^{T}(-M^{T})y=-(y^{T}Mx)$ to player 2. In a
symmetric zero-sum game, the expected payoff of a mixed profile $x$ is
$x^{T}Mx$.
### 2.1 Preference graphs
Two profiles are _$i$ -comparable_ if they differ only in the strategy of
player $i$; they are _comparable_ if they are $i$-comparable for some player
$i$. If two profiles are comparable, then there is exactly one $i$ such that
they are $i$-comparable. In symmetric games, profiles and strategies are the
same, and we define all profiles to be comparable. Up to strategic
equivalence, the game is defined by the _payoff differences_ between
comparable profiles (10). We store this in a matrix we call the _weight
matrix_ $W$ of the game.
###### Definition 2.1.
Let $M$ be a zero-sum game, and let $p$ and $q$ be comparable profiles. If $M$
is a symmetric game, then profiles and strategies are the same, and we define
$W_{p,q}$ to be the same as $M_{p,q}$. If $M$ is non-symmetric, then
$p=(p_{1},p_{2})$ and $q=(q_{1},q_{2})$ and
$W_{p,q}=\begin{cases}M_{p_{1},p_{2}}-M_{q_{1},q_{2}}&\text{$p$ and $q$ are
$1$-comparable}\\\ M_{q_{1},q_{2}}-M_{p_{1},p_{2}}&\text{$p$ and $q$ are
$2$-comparable}\end{cases}$
If $p$ and $q$ are not comparable, then $W_{p,q}$ _is undefined_.
We deliberately leave the payoff differences between incomparable profiles
undefined, so that it is clear to the reader that we will only reference $W$
when the associated profiles are comparable. Note that for any comparable
profiles $p$ and $q$, $W_{p,q}=-W_{q,p}$.
###### Definition 2.2.
Let $M$ be a zero-sum game. The _preference graph_ of $M$ is the graph whose
nodes are the profiles of the game and where there is an arc ${p}$${q}$
between profiles $p$ and $q$ if and only if they are comparable and
$W_{p,q}\leq 0$ (equivalently, $W_{q,p}\geq 0$).
While the definition is the same for symmetric and non-symmetric games, the
resultant graphs are not the same, because the weight matrix is defined
differently. In symmetric zero-sum games all profiles are comparable, so the
preference graph is a _tournament_ (a directed graph with an arc between every
pair of nodes). In non-symmetric games the preference graph is the game’s
response graph (8), which is never a tournament because not all profiles are
comparable. We think of the entries in the weight matrix as being weights on
the associated arc, as in ${p}$${q}$$\scriptstyle{|W_{p,q}|}$.
### 2.2 Dynamical Systems and the Replicator
The replicator dynamic is a continuous-time dynamical system (36; 22), defined
by an ordinary differential equation. Let $x\in\Delta(S_{1})$ and
$y\in\Delta(S_{2})$ be mixed strategies, and let $s\in S_{1}$ and $t\in S_{2}$
be pure strategies. Then, for a (non-symmetric) zero-sum game $M$ we have
###### Definition 2.3 (Non-Symmetric Zero-Sum Replicator Equation).
$\displaystyle\dot{x}_{s}$ $\displaystyle=x_{s}((My)_{s}-x^{T}My)$
$\displaystyle\dot{y}_{t}$ $\displaystyle=-y_{t}((M^{T}x)_{t}-x^{T}My)$
In a symmetric game $M$, the replicator is defined a similar way:
$\dot{x}_{s}=x_{s}((Mx)_{s}-x^{T}Mx)$
If $M$ is also zero-sum, then $x^{T}Mx=0$ (by anti-symmetry), and this reduces
to
###### Definition 2.4 (Symmetric Zero-Sum Replicator Equation).
$\dot{x}_{s}=x_{s}(Mx)_{s}$
Note that the symmetric (2.4) and non-symmetric (2.3) replicator, while
similar, are distinct equations with different properties. The solutions to
these equations define a _flow_ (36) on the strategy space of the game, which
is a function $\phi:X\times\to X$ that is a continuous group action of the
reals on $X$. We call this the (symmetric or non-symmetric) _replicator flow_.
The forward orbit of the flow from a given point is called a _trajectory_ of
the system. A set of points $Y$ is called _invariant_ under $\phi$ if
$\phi(Y,t)=Y$ for any $t\in$.
###### Definition 2.5.
Let $A$ be a compact set under a flow $\phi$ on a compact space $X$. If there
is a neighbourhood $U$ of $A$ such that
$\lim_{t\to\infty}\sup_{x\in U}\inf_{y\in A}\mathbf{d}(\phi(x,t),y)=0$
where $\mathbf{d}$ is a metric, then we say $A$ is _asymptotically stable_. If
$A$ is also invariant, we call it an _attracting set_. If $A$ is an attracting
set, and no smaller attracting set is contained within it, we call it an
_attractor_. Attracting sets of the time-reversed flow $\phi^{-1}$ we call
_repelling sets_ , and attractors of the reversed flow we call _repellors_.
There are some differences in terminology in the literature to be wary of.
What we call attracting sets are sometimes called attractors (e.g. (13; 36;
9)), in which case what we call an attractor is a _minimal attractor_.
Otherwise, our definition is the same as (36; 9). The set of points which
approach an attractor $A$ in the limit $t\to\infty$ is called the attractor’s
_basin of attraction_. We call an attractor _global_ if its basin of
attraction includes all points in $\operatorname{int}(X)$.
## 3 The Attractor of the Replicator
In this section we prove Theorem 3.3, which characterises the attractor of the
replicator in zero-sum games. We begin by noting that any mixed profile $x$
naturally defines a distribution over profiles, with $x_{p}$ denoting the mass
on a profile $p$. If the game is symmetric, profiles and strategies are the
same and so this is trivial. In a non-symmetric game, $x=(x_{1},x_{2})$ is a
pair of mixed strategies, and the distribution is defined by the _product_ ,
with the mass on a profile $p=(p_{1},p_{2})$ defined by
$x_{p}:={x_{1}}_{p_{1}}{x_{2}}_{p_{2}}$ (1)
This distribution over profiles is used to define an important concept: the
_content_ of a set of profiles.
###### Definition 3.1 (Biggar and Shames (9)).
Let $H$ be a set of profiles in a game. The _content_ of $H$, denoted
$\operatorname{content}(H)$, is the set of all mixed profiles $x$ where all
profiles in the support of $x$ are in $H$.
Equivalently, $x\in\operatorname{content}(H)$ if and only if
$x_{H}:=\sum_{h\in H}x_{h}=1$, that is, $x$ defines a distribution whose mass
is entirely distributed over profiles in $H$. The content is a union of
subgames, and so is an invariant set under the replicator (36). We will show
that the unique global attractor is the content of the unique sink component
of the preference graph. Uniqueness follows from graph structure, established
originally in (8).
###### Lemma 3.2 (Uniqueness).
The preference graph of a zero-sum game has a unique sink component.
The proof of Lemma 3.2 can be found in the appendix. Now we can prove our main
theorem.
###### Theorem 3.3 (The Attractor of the Replicator).
In a (symmetric or non-symmetric) zero-sum game $M$, the content of the unique
sink component $H$ of its preference graph is the unique global attractor of
the (respectively symmetric or non-symmetric) replicator dynamic.
###### Proof.
Proving Theorem 3.3 requires showing (i) $\operatorname{content}(H)$ is an
invariant set (_invariance_), (ii) every attracting set contains
$\operatorname{content}(H)$ (_minimality_) and (iii)
$\operatorname{content}(H)$ is asymptotically stable and its basin of
attraction is $\operatorname{int}(X)$ (_global asymptotic stability_). We
establish (i) and (ii) in Lemma 3.4.
###### Lemma 3.4 (Invariance and Minimality).
If $H$ is the sink component of the preference graph of a (symmetric or non-
symmetric) zero-sum game $M$, then $\operatorname{content}(H)$ is invariant
under the replicator. Further, for any attracting set $A$,
$\operatorname{content}(H)\subseteq A$.
Lemma 3.4 largely follows similar results in the literature, so we defer its
proof to the appendix. The challenge and main contribution of Theorem 3.3 lies
in (iii): showing asymptotic stability of $\operatorname{content}(H)$. We do
this by demonstrating that $x_{H}$, the total mass on the sink component $H$,
increases over time (Lemmas 3.5 and 3.9). That is,
$\dot{x}_{H}=\frac{\mathrm{d}}{\mathrm{d}t}x_{H}=\sum_{h\in
H}\frac{\mathrm{d}}{\mathrm{d}t}x_{h}=\sum_{h\in H}\dot{x}_{h}>0$, for any
$x\in\operatorname{int}(X)$. The function $x_{H}$ is a natural choice, because
$x_{H}$ is uniformly continuous, bounded in $[0,1]$ and $x_{H}=1$ if and only
if $x\in\operatorname{content}(H)$, so $x_{H}$ can be thought of as a metric
for the distance between $x$ and the content. Showing $\dot{x}_{H}>0$ requires
different arguments for the symmetric and non-symmetric cases. The symmetric
case is straightforward, and we complete it below (Lemma 3.5).
###### Lemma 3.5.
In a symmetric zero-sum game $M$, under the symmetric replicator (2.4), if
$x_{H}\in(0,1)$ then $\dot{x}_{H}>0$.
###### Proof.
First, $\dot{x}_{H}=\sum_{h\in H}x_{h}(Mx)_{h}$. As $M$ is anti-symmetric,
$x_{h}x_{q}M_{h,q}+x_{q}x_{h}M_{q,h}=0$ if $q\in H$. Hence,
$\dot{x}_{H}=\sum_{q\in S_{1}\times S_{2}}\sum_{h\in
H}x_{q}x_{h}M_{h,q}=\sum_{q\not\in H}\sum_{h\in H}x_{q}x_{h}M_{h,q}.$
For any $q\notin H$ and $h\in H$, as $H$ is a sink component, there is an arc
${q}$${h}$ in the preference graph. That is, $M_{h,q}>0$. Hence, all summands
are nonnegative. Moreover, due to the fact that $x_{H}\in(0,1)$, there must
exist a $q\not\in H$ and an $h\in H$ such that $x_{q}x_{h}>0$. Thus,
$\dot{x}_{H}>0$. ∎
The more challenging case of the proof is showing that $\dot{x}_{H}>0$ in non-
symmetric games (Lemma 3.9), which we will solve by embedding the flow of the
non-symmetric replicator on our non-symmetric game $M$ into a that flow of the
symmetric replicator on a larger symmetric game (Theorem 3.8) which we call
the _symmetrised game_.
###### Definition 3.6.
Let $M$ be a zero-sum game, with $M\in^{n\times m}$. The _symmetrised game_ ,
written $\mathcal{S}_{M}$, is defined as the following (nm)×(nm) matrix, which
we index by profiles $p=(p_{1},p_{2})$ and $q=(q_{1},q_{2})$:
$(\mathcal{S}_{M})_{p,q}=M_{p_{1},q_{2}}-M_{q_{1},p_{2}}.$
The strategy space of $\mathcal{S}_{M}$ is $S_{1}\times S_{2}$. That is, the
_strategy profiles_ of the original game $M$ become _strategies_ of the
symmetrised game. Mixed profiles likewise become _mixed strategies_ , using
the production distribution as in equation (1):
$x_{p}:={x_{1}}_{p_{1}}{x_{2}}_{p_{2}}$.
This construction is a natural choice, and related constructions have appeared
before in the literature on zero-sum games (17). The symmetrised game has two
important properties: (1) it is anti-symmetric, and so is a symmetric zero-sum
game, and (2) it can be viewed as an extension of the weight matrix $W$,
previously only defined on comparable profiles, to a relation over all
profiles. In particular, if $p$ and $q$ are comparable then
$(\mathcal{S}_{M})_{p,q}=W_{p,q}$. More generally,
###### Lemma 3.7.
Let $p=(p_{1},p_{2})$ and $q=(q_{1},q_{2})$ be profiles. Then:
$(\mathcal{S}_{M})_{p,q}=W_{(p_{1},q_{2}),p}+W_{(q_{1},p_{2}),p}=W_{(p_{1},q_{2}),q}+W_{(q_{1},p_{2}),q}.$
Using Theorem 3.8, we now show that the symmetrised game allows us to analyse
the flow of the replicator on $M$ by considering the flow of the replicator on
$\mathcal{S}_{M}$.
###### Theorem 3.8 (Symmetrising the Replicator Dynamic).
Let $M$ be a non-symmetric zero-sum game. Let $x=(x_{1},x_{2})$ be a mixed
profile and $p=(p_{1},p_{2})$ a pure profile. Write
$x_{p}:={x_{1}}_{p_{1}}{x_{2}}_{p_{2}}$ as in equation (1). Then, under the
non-symmetric replicator (2.3),
$\dot{x}_{p}=x_{p}(\mathcal{S}_{M}x)_{p}.$
The proofs of Lemma 3.7 and Theorem 3.8 can be found in the appendix. Notice
that the resultant expression is exactly the definition of the symmetric
replicator (2.4) on the symmetric zero-sum game $\mathcal{S}_{M}$. Using this
transformation, we can now complete the proof of Theorem 3.3 on non-symmetric
games by showing again that $x_{H}$ is increasing in $x_{H}\in(0,1)$.
###### Lemma 3.9.
In a non-symmetric zero-sum game $M$, under the non-symmetric replicator
(2.3), if $x_{H}\in(0,1)$ then $\dot{x}_{H}$ > 0.
###### Proof.
If $h=(h_{1},h_{2})$ is a pure profile, then by Theorem 3.8,
$\dot{x}_{h}=\frac{\mathrm{d}}{\mathrm{d}t}({x_{1}}_{h_{1}}{x_{2}}_{h_{2}})=x_{h}(\mathcal{S}_{M}x)_{h}$.
Because $\mathcal{S}_{M}$ is symmetric, by the same argument as in Lemma 3.5,
we can show that
$\dot{x}_{H}=\sum_{q\not\in H}\sum_{h\in H}x_{q}x_{h}(\mathcal{S}_{M})_{h,q}.$
Now pick some profiles $q\not\in H$ and $h\in H$ with $x_{q}x_{h}>0$. Since
$x_{H}>0$, at least one such pair exist. We will show that the sum above is
strictly positive. Firstly, observe that if $q$ and $h$ are comparable, the
arc ${q}$${h}$ of the preference graph of $M$ goes from $q$ to $h$ (and not
vice versa), and so $(\mathcal{S}_{M})_{h,q}=W_{h,q}>0$. Otherwise, suppose
that $q=(q_{1},q_{2})$ and $h=(h_{1},h_{1})$ are not comparable, and let
$a=(q_{1},h_{2})$ and $b=(h_{1},q_{2})$. We have the following three cases:
1. 1.
$a,b\in H$. Then the arcs ${q}$${b}$ and ${q}$${a}$ are directed towards $a$
and $b$ because $q$ is outside the sink component $H$. By Lemma 3.7,
$(\mathcal{S}_{M})_{h,q}=W_{q,a}+W_{q,b}>0$.
2. 2.
$a,b\not\in H$. Then the arcs ${a}$${h}$ and ${b}$${h}$ are directed towards
$h$ because $a$ and $b$ are outside the sink component. By Lemma 3.7,
$(\mathcal{S}_{M})_{h,q}=W_{a,h}+W_{b,h}>0$.
3. 3.
$a\in H$, $b\not\in H$ (the case $b\in H$, $a\not\in H$ is identical). The sum
$\sum_{q\not\in H}\sum_{h\in H}x_{q}x_{h}(\mathcal{S}_{M})_{h,q}$ includes the
terms $x_{q}x_{h}(\mathcal{S}_{M})_{h,q}$ and
$x_{b}x_{a}(\mathcal{S}_{M})_{a,b}$. However,
$x_{q}x_{h}=q_{1}q_{2}h_{1}h_{2}=x_{b}x_{a}$, and so
$x_{q}x_{h}(\mathcal{S}_{M})_{h,q}+x_{b}x_{a}(\mathcal{S}_{M})_{a,b}=x_{q}x_{h}((\mathcal{S}_{M})_{h,q}+(\mathcal{S}_{M})_{b,a})=x_{q}x_{h}(W_{q,a}+W_{q,b}+W_{b,q}+W_{b,h})=x_{q}x_{h}(W_{q,a}+W_{b,h})$
by Lemma 3.7. Since $a$ and $h$ are inside the sink component and $q$ and $b$
are outside, and the arcs ${q}$${a}$ and ${b}$${h}$ must be directed into the
component, so $W_{q,a}>0$ and $W_{b,h}>0$ and thus
$x_{q}x_{h}(\mathcal{S}_{M})_{h,q}+x_{b}x_{a}(\mathcal{S}_{M})_{a,b}>0$.
Overall, we conclude that $\dot{x}_{H}>0$. ∎
Finally, asymptotic stability of $\operatorname{content}(H)$ follows easily
from the fact that $x_{H}$ is uniformly continuous. Pick any
$0<\alpha<\beta<1$. Then for any $x$ with $x_{H}\in[\alpha,\beta]$,
$\dot{x}_{H}>\epsilon$ for some $\epsilon$ (uniform continuity), so after some
finite time $x_{H}>\beta$. Repeating this argument for any $\alpha,\beta$
shows that $\operatorname{content}(H)$ is asymptotically stable. ∎
## 4 Chain Recurrence and Nash Equilibria
In this section we discuss some consequences of Theorem 3.3. The first
concerns chain recurrence, which is defined by $(\epsilon,T)$-chains.
###### Definition 4.1 (Chain Recurrence, (3)).
Let $\phi$ be a flow on a compact metric space $X$, with $x$ and $y$ in $X$.
An _$(\epsilon,T)$ -chain_ from $x$ to $y$ is a finite sequence of points
$x_{1},x_{2},\dots,x_{n}$ with $x=x_{1}$ and $y=x_{n}$, and times
$t_{1},\dots,t_{n}\in[T,\infty)$ such that
$\mathbf{d}(\phi(x_{i},t_{i}),x_{i+1})<\epsilon$. If there is an
$(\epsilon,T)$-chain from $x$ to $y$ for _all_ $\epsilon>0$ and $T>0$ we say
there is a _pseudo-orbit_ from $x$ to $y$.
A point is called _chain recurrent_ if it has a pseudo-orbit to itself. Two
points are _chain equivalent_ if there are pseudo-orbits between them in both
directions, and equivalent chain recurrent points are grouped in topologically
connected components called _chain components_ (3). Reachability under pseudo-
orbits provides an ordering on the chain components, and sink chain components
are those which are minimal in this order. Sink chain components have been
increasingly studied in algorithmic game theory (30; 31; 32). A connection
between the preference graph and chain components of the replicator was
demonstrated in (32; 42; 9), with (9) using this to prove that sink chain
components always exist. However, sink chain components have not generally
been characterised. Attractors, when they exist, are sink chain components, so
our results present the first characterisation of sink chain components of
zero-sum games.
###### Lemma 4.2.
In any flow, every attractor is a sink chain component.
###### Corollary 4.3.
The content of the sink component of the preference graph is the unique sink
chain component of a zero-sum game.
The proofs of Lemma 4.2 and Corollary 4.3 can be found in the appendix. In
zero-sum games, previous state-of-the-art results on chain recurrence made use
of the Nash equilibrium (33; 30; 26). This line of inquiry established first
(1) that all points in the essential subgame (the subgame containing the
equilibrium in its interior) are contained within the sink chain component.
Secondly (2), if the essential subgame is not the whole game, all interior
starting points converge to the essential subgame. This suggests a connection
between chain recurrence and equilibria in zero-sum games. Corollary 4.3 comes
to a different conclusion; it proves that sink chain components are
characterised solely by the preference graph. Lemma 4.4 resolves this seeming
discrepancy: chain components are determined by the preference graph, but the
presence of an equilibrium in a subgame forces some structure on the induced
preference graph of that subgame—in particular, it must be strongly connected
and contained within the sink component of the whole game’s preference graph.
###### Lemma 4.4.
In a zero-sum game, any subgame with a Nash equilibrium in its interior must
(i) be contained within the sink component of the preference graph and (ii)
the subgraph which this subgame induces in the preference graph must be
strongly connected.
###### Proof.
(i) Assume the attractor is not the whole game, in which case it is on the
boundary. By Theorem 3.3, all interior points are in the basin of attraction,
and by Theorem 3.4 of (33), trajectories starting from interior points
converge to the essential subgame in the limit, and so the essential subgame
must be within the attractor. (ii) Each subgame is independent, so we can
assume we are working with the whole game and the equilibrium $x$ is fully-
mixed. For contradiction, assume the preference graph is not strongly
connected. Consequently, there is an attractor $A$ on the boundary of the
strategy space (Theorem 3.3). The point $x$ is in the basin of attraction of
$A$, so $x$ converges to $A$, which contradicts the fact that $x$ is a Nash
equilibrium, which are fixed points under the replicator (36). ∎
This Lemma rephrases our understanding of equilibria and chain recurrence,
highlighting the key role of the preference graph. As an example, Corollary
4.3 implies that sink chain component is the whole game _if and only if_ the
preference graph is strongly connected. When a fully-mixed equilibrium exists,
the preference graph must be connected, and so the sink chain component is the
whole game. Most interestingly, even though we prove Lemma 4.4 using the
replicator dynamic, this lemma is a _purely game-theoretic result_ which only
relates equilibria and the preference graph. This connection between the
preference graph and equilibria is useful for analysing games. For instance,
examining Figure 1 we find the Nash equilibrium is $((0,0.5,0.5),(0,0.5,0.5))$
(both players play $b$ and $c$ half the time), which has support
$\\{b,c\\}\times\\{b,c\\}$. As Lemma 4.4 predicts, this is within the
attractor and its induced preference graph (a 4-cycle) is strongly connected.
## 5 Conclusions and Future Work
In this paper we gave the first characterisation of the unique attractor of
the replicator in zero-sum games, thereby describing the long-term behaviour
of the dynamic in these games. As a secondary result, we have demonstrated the
importance of the preference graph as a tool for analysing game dynamics. In
particular, in the long run, only the _preferences_ dictates the outcome of
the game. This surprising conclusion has potentially significant consequences
for modelling strategic interactions. In game theory—especially evolutionary
game theory—precise knowledge of utilities is difficult to achieve; our
results show that modelling only the preferences for each player is sufficient
to characterise the attractor of the replicator dynamic.
An important goal for future research is to characterise the attractors of the
replicator in _all_ games, not just zero-sum ones. Currently, the attractors
of the replicator have been characterised in only a few classes of games, such
as potential games and $2\times n$ games without dominated strategies (9), as
well as some individual games, like the Asymmetric Cyclic Matching Pennies
game from (25). Our paper adds zero-sum games to this list of solved classes.
## References
* (1)
* Akin and Losert (1984) Ethan Akin and Viktor Losert. 1984. Evolutionary dynamics of zero-sum games. _Journal of Mathematical Biology_ 20, 3 (1984), 231–258.
* Alongi and Nelson (2007) John M Alongi and Gail Susan Nelson. 2007. _Recurrence and topology_. Vol. 85. American Mathematical Soc.
* Andrade et al. (2021) Gabriel P Andrade, Rafael Frongillo, and Georgios Piliouras. 2021\. Learning in matrix games can be arbitrarily complex. In _Conference on Learning Theory_. PMLR, 159–185.
* Arora et al. (2012) Sanjeev Arora, Elad Hazan, and Satyen Kale. 2012\. The multiplicative weights update method: a meta-algorithm and applications. _Theory of computing_ 8, 1 (2012), 121–164.
* Bailey and Piliouras (2018) James P Bailey and Georgios Piliouras. 2018. Multiplicative weights update in zero-sum games. In _Proceedings of the 2018 ACM Conference on Economics and Computation_. 321–338.
* Benaïm et al. (2012) Michel Benaïm, Josef Hofbauer, and Sylvain Sorin. 2012\. Perturbations of set-valued dynamical systems, with applications to game theory. _Dynamic Games and Applications_ 2, 2 (2012), 195–205.
* Biggar and Shames (2023a) Oliver Biggar and Iman Shames. 2023a. The graph structure of two-player games. _Scientific Reports_ 13, 1 (2023), 1833.
* Biggar and Shames (2023b) Oliver Biggar and Iman Shames. 2023b. The Replicator Dynamic, Chain Components and the Response Graph. In _International Conference on Algorithmic Learning Theory_. PMLR, 237–258.
* Candogan et al. (2011) Ozan Candogan, Ishai Menache, Asuman Ozdaglar, and Pablo A Parrilo. 2011. Flows and decompositions of games: Harmonic and potential games. _Mathematics of Operations Research_ 36, 3 (2011), 474–503.
* Cheung and Piliouras (2019) Yun Kuen Cheung and Georgios Piliouras. 2019. Vortices instead of equilibria in minmax optimization: Chaos and butterfly effects of online learning in zero-sum games. In _Conference on Learning Theory_. PMLR, 807–834.
* Cheung and Piliouras (2020) Yun Kuen Cheung and Georgios Piliouras. 2020. Chaos, extremism and optimism: Volume analysis of learning in games. _Advances in Neural Information Processing Systems_ 33 (2020), 9039–9049.
* Conley (1978) Charles C Conley. 1978\. _Isolated invariant sets and the Morse index_. Number 38. American Mathematical Soc.
* Daskalakis (2013) Constantinos Daskalakis. 2013\. On the complexity of approximating a Nash equilibrium. _ACM Transactions on Algorithms (TALG)_ 9, 3 (2013), 1–35.
* Daskalakis et al. (2009) Constantinos Daskalakis, Paul W Goldberg, and Christos H Papadimitriou. 2009. The complexity of computing a Nash equilibrium. _SIAM J. Comput._ 39, 1 (2009), 195–259.
* Freund and Schapire (1999) Yoav Freund and Robert E Schapire. 1999. Adaptive game playing using multiplicative weights. _Games and Economic Behavior_ 29, 1-2 (1999), 79–103.
* Gale et al. (1950) David Gale, Harold W Kuhn, and Albert W Tucker. 1950\. On Symmetric Games. In _Contributions to the Theory of Games (AM-24), Volume I_. Princeton University Press, 81–88.
* Goodfellow et al. (2020) Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, and Yoshua Bengio. 2020\. Generative adversarial networks. _Commun. ACM_ 63, 11 (2020), 139–144.
* Hart and Mas-Colell (2003) Sergiu Hart and Andreu Mas-Colell. 2003. Uncoupled dynamics do not lead to Nash equilibrium. _American Economic Review_ 93, 5 (2003), 1830–1836.
* Hofbauer (1996) Josef Hofbauer. 1996\. Evolutionary dynamics for bimatrix games: A Hamiltonian system? _Journal of Mathematical Biology_ 34, 5 (1996), 675–688.
* Hofbauer and Sigmund (1998) Josef Hofbauer and Karl Sigmund. 1998. _Evolutionary games and population dynamics_. Cambridge university press.
* Hofbauer and Sigmund (2003) Josef Hofbauer and Karl Sigmund. 2003. Evolutionary game dynamics. _Bulletin of the American mathematical society_ 40, 4 (2003), 479–519.
* Hofbauer et al. (2009) Josef Hofbauer, Sylvain Sorin, and Yannick Viossat. 2009\. Time average replicator and best-reply dynamics. _Mathematics of Operations Research_ 34, 2 (2009), 263–269.
* Kalies et al. (2021) William D Kalies, Konstantin Mischaikow, and Robert CAM Vandervorst. 2021. Lattice structures for attractors III. _Journal of Dynamics and Differential Equations_ (2021), 1–40.
* Kleinberg et al. (2011) Robert D Kleinberg, Katrina Ligett, Georgios Piliouras, and Éva Tardos. 2011. Beyond the Nash Equilibrium Barrier.. In _ICS_. 125–140.
* Mertikopoulos et al. (2018) Panayotis Mertikopoulos, Christos Papadimitriou, and Georgios Piliouras. 2018. Cycles in adversarial regularized learning. In _Proceedings of the Twenty-Ninth Annual ACM-SIAM Symposium on Discrete Algorithms_. SIAM, 2703–2717.
* Milnor (1985) John Milnor. 1985\. On the concept of attractor. _Communications in Mathematical Physics_ 99 (1985), 177–195.
* Nash (1951) John Nash. 1951\. Non-cooperative games. _Annals of mathematics_ (1951), 286–295.
* Omidshafiei et al. (2019) Shayegan Omidshafiei, Christos Papadimitriou, Georgios Piliouras, Karl Tuyls, Mark Rowland, Jean-Baptiste Lespiau, Wojciech M Czarnecki, Marc Lanctot, Julien Perolat, and Remi Munos. 2019. $\alpha$-rank: Multi-agent evaluation by evolution. _Scientific reports_ 9, 1 (2019), 1–29.
* Papadimitriou and Piliouras (2016) Christos Papadimitriou and Georgios Piliouras. 2016. From Nash equilibria to chain recurrent sets: Solution concepts and topology. In _Proceedings of the 2016 ACM Conference on Innovations in Theoretical Computer Science_. 227–235.
* Papadimitriou and Piliouras (2018) Christos Papadimitriou and Georgios Piliouras. 2018. From nash equilibria to chain recurrent sets: An algorithmic solution concept for game theory. _Entropy_ 20, 10 (2018), 782.
* Papadimitriou and Piliouras (2019) Christos Papadimitriou and Georgios Piliouras. 2019. Game dynamics as the meaning of a game. _ACM SIGecom Exchanges_ 16, 2 (2019), 53–63.
* Piliouras and Shamma (2014) Georgios Piliouras and Jeff S Shamma. 2014. Optimization despite chaos: Convex relaxations to complex limit sets via Poincaré recurrence. In _Proceedings of the twenty-fifth annual ACM-SIAM Symposium on Discrete Algorithms_. SIAM, 861–873.
* Ritzberger and Weibull (1995) Klaus Ritzberger and Jörgen W Weibull. 1995. Evolutionary selection in normal-form games. _Econometrica: Journal of the Econometric Society_ (1995), 1371–1399.
* Robinson (1998) Clark Robinson. 1998\. _Dynamical systems: stability, symbolic dynamics, and chaos_. CRC press.
* Sandholm (2010) William H Sandholm. 2010\. _Population games and evolutionary dynamics_. MIT press.
* Schuster and Sigmund (1983) Peter Schuster and Karl Sigmund. 1983. Replicator dynamics. _Journal of theoretical biology_ 100, 3 (1983), 533–538.
* Silver et al. (2016) David Silver, Aja Huang, Chris J Maddison, Arthur Guez, Laurent Sifre, George Van Den Driessche, Julian Schrittwieser, Ioannis Antonoglou, Veda Panneershelvam, Marc Lanctot, et al. 2016\. Mastering the game of Go with deep neural networks and tree search. _Nature_ 529, 7587 (2016), 484–489.
* Silver et al. (2018) David Silver, Thomas Hubert, Julian Schrittwieser, Ioannis Antonoglou, Matthew Lai, Arthur Guez, Marc Lanctot, Laurent Sifre, Dharshan Kumaran, Thore Graepel, et al. 2018\. A general reinforcement learning algorithm that masters chess, shogi, and Go through self-play. _Science_ 362, 6419 (2018), 1140–1144.
* Smith and Price (1973) J Maynard Smith and George R Price. 1973. The logic of animal conflict. _Nature_ 246, 5427 (1973), 15–18.
* Taylor and Jonker (1978) Peter D Taylor and Leo B Jonker. 1978. Evolutionary stable strategies and game dynamics. _Mathematical biosciences_ 40, 1-2 (1978), 145–156.
* Vlatakis-Gkaragkounis et al. (2020) Emmanouil-Vasileios Vlatakis-Gkaragkounis, Lampros Flokas, Thanasis Lianeas, Panayotis Mertikopoulos, and Georgios Piliouras. 2020. No-regret learning and mixed nash equilibria: They do not mix. _Advances in Neural Information Processing Systems_ 33 (2020), 1380–1391.
* Von Neumann and Morgenstern (1944) John Von Neumann and Oskar Morgenstern. 1944. _Theory of games and economic behavior_. Princeton university press.
* Weibull (1997) Jörgen W Weibull. 1997\. _Evolutionary game theory_. MIT press.
* Zeeman (1980) E Christopher Zeeman. 1980\. Population dynamics from game theory. In _Global theory of dynamical systems_. Springer, 471–497.
## Appendix A Proofs
###### Lemma A.1 (Lemma 3.2).
The preference graph of a zero-sum game has a unique sink component.
###### Proof.
In the non-symmetric case, the preference graph is the response graph, and the
result follows from Theorem 4.10 of (8). In the symmetric case, the preference
graph is a tournament, and all tournaments have one sink component, as they
are orientations of complete graphs. ∎
###### Lemma A.2 (Lemma 3.4).
If $H$ is the sink component of the preference graph of a (symmetric or non-
symmetric) zero-sum game $M$, then $\operatorname{content}(H)$ is invariant
under the replicator. Further, for any attracting set $A$,
$\operatorname{content}(H)\subseteq A$.
###### Proof.
(_Invariance_ :) Observe that if $x\in\operatorname{content}(H)$, then the
support of $x$ is contained in $H$, and because all mixed profiles in the
subgame $\Delta(\operatorname{support}(x))$ have the same support,
$\Delta(\operatorname{support}(x))\subseteq\operatorname{content}(H)$. It
follows that $\operatorname{content}(H)$ is a union of subgames. By Theorem
5.4.7 of (36), all subgames are invariant sets under the replicator, and
unions of invariant sets are invariant.
($\operatorname{content}(H)\subseteq A$:) By Theorems 9.1.2 and 9.1.6 of (36),
no asymptotically stable set can exist in the interior of the strategy space
of a symmetric or non-symmetric zero-sum game. Subgames have the same
properties as the whole game under the replicator so the same is true of all
subgames. Dually, no repelling set can exist in the interior of any subgame.
The remainder of the proof follows that of Theorem 5.2 of (9).
(_Claim_ : every attracting set contains a profile.) This follows by
induction, using the fact that the replicator dynamic on a subgame has the
same properties as on the whole game. In the whole game, an asymptotically
stable set intersects the boundary. This intersection with the boundary must
also be asymptotically stable in any subgame it intersects on the boundary,
and so it intersects the boundary of this smaller subgame, and so on. We
conclude that such a set contains a pure profile, the smallest possible
subgame.
(_Claim_ : every attracting set contains all profiles in $H$.) An arc
${p}$${q}$ of the preference graph is also a subgame, where only the profiles
$p$ and $q$ are in the support. The (symmetric or non-symmetric) replicator
reduces to $\dot{x}_{p}=x_{p}(1-x_{p})W_{q,p}$ on this subgame, where
$W_{q,p}\geq 0$ (for the player for which these profiles are comparable). If
$p$ is contained in an asymptotically stable set, then $q$ must also be
contained in this set, because points near $p$ move to $q$ along this arc. We
know that asymptotically stable sets contain a pure profile—by this argument
we deduce that they contain all pure profiles reachable from that one in the
preference graph. Such a set of profiles always contains the sink component
$H$.
(_Claim_ : every attracting set contains $\operatorname{content}(H)$.) Let $Y$
be a subgame, where the pure profiles in $Y$ are in an attracting set $A$. If
$Y$ is a pure profile, then all mixed profiles in $Y$ are in the set,
trivially. Now suppose for induction that all points on the boundary of $Y$
are in $A$. Suppose for contradiction that there is a point
$x\in\operatorname{int}(Y)$ that is not in $A$. The set $A\cap Y$ is
attracting in $Y$, and the boundary is contained in $A\cap Y$, but this means
that the dual repelling set of $A\cap Y$ is contained in the interior of $Y$,
but no such sets can be contained in the interior. Hence all of $Y$ is within
$A$. By induction on subgames, we find that all of $\operatorname{content}(H)$
is within every attracting set. ∎
###### Lemma A.3.
$\mathcal{S}_{M}$ is anti-symmetric.
###### Proof.
For $p=(p_{1},p_{2})$ and $q=(q_{1},q_{2})$,
$(\mathcal{S}_{M})_{p,q}=M_{p_{1},q_{2}}-M_{q_{1},p_{2}}=-(M_{q_{1},p_{2}}-M_{p_{1},q_{2}})=-(\mathcal{S}_{M})_{q,p}$.
∎
###### Lemma A.4 (Lemma 3.7).
Let $p=(p_{1},p_{2})$ and $q=(q_{1},q_{2})$ be profiles. Then the following
holds:
$(\mathcal{S}_{M})_{p,q}=W_{(p_{1},q_{2}),p}+W_{(q_{1},p_{2}),p}=W_{(p_{1},q_{2}),q}+W_{(q_{1},p_{2}),q}.$
###### Proof.
$\displaystyle W_{p,(p_{1},q_{2})}+W_{p,(q_{1},p_{2})}$
$\displaystyle=(M_{p_{1},q_{2}}-M_{p_{1},p_{2}})+(M_{p_{1},p_{2}}-M_{q_{1},p_{2}})$
$\displaystyle=M_{p_{1},q_{2}}-M_{q_{1},p_{2}}=(\mathcal{S}_{M})_{p,q}\quad\text{and}$
$\displaystyle W_{(p_{1},q_{2}),q}+W_{(q_{1},p_{2}),q}$
$\displaystyle=(M_{p_{1},q_{2}}-M_{q_{1},q_{2}})+(M_{q_{1},q_{2}}-M_{q_{1},p_{2}})$
$\displaystyle=M_{p_{1},q_{2}}-M_{q_{1},p_{2}}=(\mathcal{S}_{M})_{p,q}$
∎
###### Theorem A.5 (Theorem 3.8).
Let $(M,-M^{T})$ be a two-player zero-sum game under the replicator dynamic.
Let $(x,y)$ be a mixed profile and $p=(p_{1},p_{2})$ a pure profile. Write
$z_{p}:=x_{p_{1}}y_{p_{2}}$ as in equation (1). Then
$\dot{z}_{p}=z_{p}(\mathcal{S}_{M}z)_{p}.$
###### Proof.
The two-population replicator dynamic (written for player 1, the player 2 case
is similar) is equivalent to
$\displaystyle\dot{x}_{s}$ $\displaystyle=x_{s}((My)_{s}-x^{T}My)$
$\displaystyle=x_{s}\sum_{t\in S_{1}}x_{t}((My)_{s}-(My)_{t})$
$\displaystyle=x_{s}\sum_{t\in S_{1}}x_{t}\sum_{r\in
S_{2}}y_{r}\left(M_{s,r}-M_{t,r}\right)$ $\displaystyle=x_{s}\sum_{t\in
S_{1}}\sum_{r\in S_{2}}x_{t}y_{r}\left(M_{s,r}-M_{t,r}\right)$
$\displaystyle=x_{s}\sum_{p=(p_{1},p_{2})\in S_{1}\times
S_{2}}z_{p}\left(M_{s,p_{2}}-M_{p_{1},p_{2}}\right).$
Now we observe that for $p=(p_{1},p_{2})$,
$\displaystyle\dot{z}_{p}$
$\displaystyle=\frac{\mathrm{d}}{\mathrm{d}t}(x_{p_{1}}y_{p_{2}})$
$\displaystyle=(x_{p_{1}}y_{p_{2}})(\frac{\dot{x}_{p_{1}}}{x_{p_{1}}}+\frac{\dot{y}_{p_{2}}}{y_{p_{2}}})\quad\text{(product
rule)}$ $\displaystyle=z_{p}\left(\sum_{q\in S_{1}\times
S_{2}}z_{q}(M_{p_{1},q_{2}}-M_{q_{1},q_{2}})+\sum_{q\in S_{1}\times
S_{2}}z_{q}(M_{q_{1},q_{2}}-M_{q_{1},p_{2}})\right)\quad\text{(by above)}$
$\displaystyle=z_{p}\sum_{q\in S_{1}\times
S_{2}}z_{q}\left(M_{p_{1},q_{2}}-M_{q_{1},q_{2}}+M_{q_{1},q_{2}}-M_{q_{1},p_{2}}\right)$
$\displaystyle=z_{p}\sum_{q\in S_{1}\times S_{2}}z_{q}(\mathcal{S}_{M})_{p,q}$
$\displaystyle=z_{p}(\mathcal{S}_{M}z)_{p}.$
∎
###### Lemma A.6 (Lemma 4.2).
In any flow, every attractor is a sink chain component.
###### Proof.
We first show that all points in the attractor are chain recurrent. Attracting
sets are closed under intersection (24; 9), so an attractor cannot overlap any
other attracting set—that would contradict the minimality of the attractor.
(13) showed that points are chain recurrent if, for each attracting set $A$,
the point is contained in either $A$ or its dual repelling set $A^{*}$. The
attractor is compact, invariant, and no attracting set overlaps it, so an
attracting or repelling set must contain all points in the attractor. Hence
all points are chain recurrent.
Pseudo-orbits cannot leave attracting sets (2). Consequently, no point outside
the attractor is chain equivalent to a point inside it, and all points in the
attractor are chain equivalent, so it is a chain component. It is a _sink_
chain component because no pseudo-orbits leave the set. ∎
###### Corollary A.7 (Corollary 4.3).
The content of the sink component of the preference graph is the unique sink
chain component of a zero-sum game.
###### Proof.
By Lemma 4.2 and Theorem 3.3, the content of the sink component is a sink
chain component. Uniqueness follows for the same reason as in Theorem 3.3:
distinct sink chain components are disjoint, but every sink chain component
contains the content (9). ∎
|
Given a number of pairwise preferences of items, a common task is to rank all the items. Examples include pairwise movie ratings, New Yorker cartoon caption contests, and many other consumer preferences tasks. What these settings have in common is two-fold: a scarcity of data (it may be costly to get comparisons for all the pairs of items) and additional feature information about the items (e.g., movie genre, director, and cast). In this paper we modify a popular and well studied method, RankCentrality for rank aggregation to account for few comparisons and that incorporates additional feature information. This method returns meaningful rankings even under scarce comparisons. Using diffusion based methods, we incorporate feature information that outperforms state-of-the-art methods in practice. We also provide improved sample complexity for RankCentrality in a variety of sampling schemes.
§ INTRODUCTION
In this paper we are interested in the problem of rank aggregation from pairwise preferences under settings where the amount of data is scarce but we may have additional structural information.
For example, consider a setting where a set of pairwise comparisons on a set of $n$ movies have been collected from a set of critics and the goal is to give an overall ranking. If $n$ is large, for example, all movies released in the last two decades, it may be extremely costly to get a comparison for each of the $\binom{n}{2}$ pairs. A more realistic regime is to hope that each movie has been viewed at least once.
Standard methods of ranking suggest that the number of comparisons needed is roughly $O(n\log(n))$—when $n$ is large, even hoping for $\log(n)$ comparisons may be hopeless!
However, each movie has additional feature information $x_i\in \mathbb{R}^d$. For example, the dimensions could encapsulate the production budget, the number of A-list actors, the writer, studio, animated or live action, etc.
In general, we may suspect that these features inform the comparisons: if movies A and B have the same Oscar-winning director, and movie A beats movie C in a comparison, we may expect movie B to also perform well against movie C. In an extreme setting, even if we don't have any comparisons involving movie B, we may still hope to infer a meaningful ranking.
In this paper we focus on modifying a popular and well studied method arising in the ranking literature for this setting and demonstrate gains in the scarce setting when the number of comparisons is very small.
A common model in the literature of particular interest to us is the Bradley-Terry-Luce (BTL) model. We assume that we have $n$ items and associated to each item $i$ is a positive score $w_i$ so that the probability that $j$ is preferred to $i$ (“$j$ beats $i$”) in a comparison is
\begin{equation}
P_{ij}:=P(i \prec j) = \frac{w_j}{w_i + w_j}\label{eq:Pij},
\end{equation}
and that we see $m$ comparisons. The underlying ranking on the items is then given by the scores $w$, with an item with a larger score being ranked higher than an item with a smaller score. In the structured setting above, we may expect movies with similar features to have similar scores. Traditional methods of learning $w$ using the BTL model, e.g., maximum likelihood estimation (MLE) or spectral methods such as Rank Centrality (both discussed below), do not naturally incorporate this kind of side information.
We have two main contributions.
* Our main contribution is Algorithm <ref>, Regularized RankCentrality, in Section <ref>. We propose a novel method for regularizing the RankCentrality algorithm that returns meaningful rankings even under scarcity. Using diffusion based methods, we propose a way of incorporating feature information that is empirically competitive with other feature based methods such as RankSVM or Siamese Networks on both synthetic and real-world datasets in scarce settings. In a specific context, we provide a sample complexity result for this regularized method.
* Along the way, we discuss traditional RankCentrality and, under a natural sampling scheme extending that in <cit.>, we show an improved sample complexity bound for the RankCentrality algorithm. For example, when pairs are sampled uniformly, we improve the bound from $O(n^5 \log n)$ to $O(n\log n)$.
§ RELATED WORKS
There is an extensive amount of literature on ranking from pairwise comparisons under various models, and we refer the interested reader to the survey in <cit.>. Roughly speaking, most frameworks either fall into the parametric setting, i.e., a model such as BTL is assumed, or non-parametric where general assumptions on the pairwise comparison matrix $P$, where $P_{ij}$ is the probability that $i$ beats $j$ in a comparison, are made.
In the latter setting, several different conditions on $P$, such as stochastic transitivity and low noise described in <cit.>, or low rank as in <cit.>, and generalized low permutation rank models have been proposed (see <cit.>). All of these models include the BTL model as a specific case. Other estimators such as the Borda count and Condorcet winner (for finding the best item rather than a ranking) have been analyzed in <cit.>.
A variant of the ranking problem also falls under the category of active ranking where the comparisons that are queried are chosen by an active ranker rather than passively considered offline, see <cit.>.
A great deal of attention has been paid to the BTL model. A natural approach to this setting is to compute an estimate for $w$ using the MLE. More precisely given a set of comparisons $S = \{(i_k, j_k, y_k)\}_{k=1}^m$ where the $k$-th comparison is between items $i_k$ and $j_k$, and $y_k=0$ denotes that $i_k$ was preferred in this observation, whereas $y_k=1$ denotes that $j_k$ was preferred.
Then the MLE is given by
\begin{align}
& \argmax_{v \in \R^n} \sum_{i=1}^m -\log \left(1 + e^{(2y_k-1)(v_{j_k} - v_{i_k})} \right) \label{eq:btlmle}
\end{align}
and our estimate is $\hat{w}_i = \exp(v_i).$
We can also consider a constrained MLE where we add an additional constraint[Without loss of generality, assume $\sum_i w_i = 1$ because $P_{ij}$ is invariant to scaling $w$.], e.g., on the maximum entry of $w$, $\|w\|_{\infty} < B$, or, alternatively, we can add add an $\ell_2$ regularizer $\lambda \|v\|_2$ to the objective. The BTL-MLE in any of these formulations is a popular objective since it is convex.
We briefly review the known results on the BTL-MLE. <cit.> have shown the constrained BTL-MLE is minimax optimal for the $\ell_2$ error.
Note that low $\ell_2$ loss does not necessarily guarantee a correct recovery of a ranking. <cit.> shows that the (regularized) MLE and spectral ranking methods (discussed below) are minimax optimal for recovery of a ranking.
The critical parameter for recovery is the minimum gap between any two different BTL scores—which does not show up when one is interested in the $\ell_2$ norm only.
In the next section we discuss the class of algorithms that are the main study of this work: spectral methods and the RankCentrality algorithm.
§ SPECTRAL METHODS
We assume that we have access to a collection of $m$ independent and identically distributed pairwise comparisons $S = \{(i_k, j_k, y_k)\}_{k=1}^m$ where each $i_k < j_k\in [n]$. Furthermore we assume that each pair is i.i.d drawn: $(i,j) \sim_{\mu} \{(i,j), 1\leq i< j\leq n\}$, where $\mu$ is an unknown sampling distribution on the set of ordered pairs. Although $\mu_{ij}$ is defined for $i < j$, we assume it is understood that $\mu_{ij} = \mu_{ji}$ when $i > j$. Denote $\mumin := \min_{i<j} \mu_{ij}$ and $\mumax := \max_{i<j} \mu_{ij}$. In addition, we assume that the label is an independent Bernoulli draw, i.e.
\begin{equation*}
y_{k} = \begin{cases}
1 & \text{with probability } P_{i_kj_k}=\tfrac{w_{j_k}}{w_{i_k}+w_{j_k}} \\
0 & \text{otherwise}
\end{cases}
\end{equation*}
according to the BTL model where $(w_1, \cdots, w_n)\in \mathbb{R}_{>0}^n$ is an unknown vector of BTL-scores, i.e., $i_k\prec j_k$ with probability $P_{i_kj_k}$. Note $P_{ij} = 1- P_{ji}$. Additionally define $b := \max_{i,j} w_i/w_j$. Without loss of generality we assume that $w^T\textbf{1} = 1$, indeed scaling the weights has no effect on the comparison probabilities.
Problem. Given $S$, return $\hat{w}$, an estimator for $w$.
Consider the following matrix $Q\in \mathbb{R}^{n\times n}$, defined as
\begin{equation}
Q_{ij} :=
\begin{cases}
\mu_{ij} P_{ij} & \text{ if } i \neq j \\
1 - \sum_{\ell \neq i} \mu_{i\ell} P_{i\ell} & \text{ if } i = j
\end{cases}.
\label{eq:Q}
\end{equation}
Observe $Q_{ij}$ is the transition matrix of a time-reversible Markov chain, where the we transition from $i$ to $j$ with probability proportional to that of $i$ beating $j$ in a comparison (we refer the reader to Chapter 1 of <cit.> for background on Markov Chains), i.e., it satisfies the detailed balance equations: for all $i \neq j$, we have
\begin{equation*}
w_i Q_{ij} = \frac{\mu_{ij} w_i w_j}{w_i + w_j} = w_j Q_{ji}.
\end{equation*}
This implies the vector $w$ is the stationary distribution of $Q$, satisfying $w^TQ = w$, i.e., $w_i$ is the equilibrium probability of being in state $i$.
This motivates using the stationary distribution of an empirical estimator $\hat Q$, with $\E[\hat Q] = Q$ as an estimator $\hat{w}$ for $w$. The impatient reader can skip ahead to the next section for our choice of $\hat{Q}$.
The connection between the BTL model and time-reversible Markov chains was noticed by <cit.> where they proposed the RankCentrality algorithm for estimating $w$ under a slightly different model. In their setting, they assume they have access to a (connected) graph on $n$ vertices $G$, and for each edge in the graph they repeatedly query the associated pairwise comparison $k$ times. In the specific setting of an – graph $\mathcal{G}_{n,p}$ on $n$ vertices,
they construct an estimator $\hat{w}$
and show for $d\geq 10 C^2 \log n$ and $kd \geq 128 C^2 b^5 \log n$, setting $p = \tfrac{d}{n}$ the following bound on the error rate holds with high probability:
\[
\frac{\big\|\hat w-w\big\|_2}{\|w\|_2} \leq 8 C b^{5/2} \sqrt{\frac{\log n}{k\,d}}.
\]
(where we recall $b := \max_{i,j} w_i/w_j$). Noting that the expected number of comparisons is $O(n^2pk) = O(nkd) = O(b^5n\log(n))$ this yields a sample complexity of $O(b^5 n\log n/\epsilon^2)$ for recovering a weight vector with relative error $\epsilon$. Note that in this setting, for $\mathcal{G}_{n,p}$ to even be connected, it is important that $p$ be at least on order $\log(n)/n$, and we must at least observe $O(n\log(n))$ comparisons. In the more general setting, the sample complexity depends on the spectral gap of the graph Laplacian of $G$ ; precise dependencies have been given in <cit.>
Returning to our setting, our sampling scheme, which we refer to as independent sampling was proposed by <cit.>.
Observe that the independent sampling scheme is more natural in many applications, and in particular each observation is made independent of the other observations, which is not true of those in <cit.>.
Rajkumar and Agarwal show that if $O(\tfrac{Cn}{\varepsilon^2P_\mathrm{min}^2\mu_\mathrm{min}^2} b^3 \ln \left( \frac{n^2}{\delta}\right))$ comparisons are made
then with probability at least $1 - \delta$ (over the random draw of $m$ samples from which $\hat P$ is constructed), the score vector $\hat w$ produced by their version of the RankCentrality algorithm satisfies $\|\hat w - w \|_2 \leq \varepsilon$.
The sample complexity here scales as $O(n^5 \log n)$ since $\mu_\mathrm{min}^{-1} \geq \binom{n}{2}$, with equality achieved only when $\mu$ is uniform. In the next section we propose a different estimator from the one given in <cit.> and we are able to give a $O(n\log n)$ sample complexity bound in the case of uniform sampling.
A crucial point to note is that both <cit.> and <cit.> assume that the directed graph of comparisons, where an edge $(i,j)$ represents that $j$ beat $i$ in at least one comparison, is strongly connected. This is because the empirical estimate $\hat Q$ of the Markov transition matrix needs to be ergodic, i.e., irreducible and aperiodic, which ensures that $\hat Q$ has a unique stationary distribution. When the number of comparisons $m$ is small (i.e., $m<n\log(n)$ in the case of <cit.>), this is usually not the case and these algorithms return a default output. In particular, in the setting mentioned in the introduction where the number of comparisons are scarce, these methods will not return a useful ranking. This is a primary motivation for the work in this paper.
§.§ Warm-up: Improved Results for Independent Sampling
In this section we improve the results given in <cit.> by using a different estimator of $Q$ than the one presented there.
Recall the notation of Section <ref>. Given a dataset of comparisons $S$, define
\begin{align*}
C_{ij} = \textstyle \sum_{k=1}^m \Big( \one\{i_k = i, j_k =j, y_k=1\} \\ + \one\{i_k = j, j_k=i, y_k=0\} \Big),
\end{align*}
i.e., $C_{ij}$ is the number of comparisons between $i$ and $j$ that $j$ won.
Additionally define the empirical Markov transition matrix
\begin{equation}
\hat Q_{ij} :=
\begin{cases}
\frac{C_{ij}}{m} & \text{ if } i \neq j \\
1 - \sum_{\ell \neq i} \frac{C_{i\ell}}{m} & \text{ if } i = j
\end{cases}.
\label{eq:Qhat}
\end{equation}
By construction, $Q = \E(\hat Q)$ so $\hat Q$ is an unbiased estimator of $Q$.
Let $\hat{w}$ be the leading left eigenvector of $\hat Q$. When $\hat Q$ is ergodic, $\hat{w}$ is the unique stationary distribution of $\hat Q$.
Fix $\delta \in (0, 1)$ and $\varepsilon \in (0, 1)$. If
\[ m \geq 64b^3 n^{-1} \mumin^{-2} \varepsilon^{-2}(\mumax + n\mumax^2) \log \frac{2n}{\delta} \]
and the empirical Markov chain $\hat{Q}$ constructed as in (<ref>) is ergodic, then with probability at least $1-\delta$, we have
\[\frac{\|\hat w - w\|}{\|w\|} \leq \varepsilon.\]
A complete proof can be found in the supplementary materials. We sketch an outline of the proof here.
We first prove a result on the deviation of left eigenvectors for perturbations of ergodic row stochastic matrices, Proposition <ref> based on ideas from <cit.>. For each observation $k \in [m]$, we define a random i.i.d. matrix $Q_k$ (in terms of $i_k$, $j_k$, and $y_k$) such that $\hat Q = I + \frac{1}{m} \sum_{k=1}^m Q_k$. We can therefore write $\hat Q - Q = \sum_k Z_k$ where each $Z_k$ is an independent random matrix with $\E(Z_k) = 0$ and we can explicitly compute the matrix variance of $Z_k$ (Lemma <ref>). By using matrix Bernstein inequalities given in <cit.> we can derive a central-limit type upper bound on $P(\|\hat w - w\| > \varepsilon)$ (Theorem <ref>). Solving the resulting inequality for $m$, we get the desired result.
Because $\mumin = \mumax = \binom{n}{2}^{-1}$ when $\mu$ is uniform, we have given an $O\left(b^3\varepsilon^{-2}n\log(\tfrac{n}{\delta})\right)$ sample complexity when $\mu$ is uniform. Our argument improves upon that in <cit.> through improved matrix concentration results and a different (unbiased) estimator for $Q$.
§ REGULARIZING RANKCENTRALITY
When the number of pairwise comparison observations we have available is small, the $\hat Q_{ij}$ entries are poor estimators for $Q_{ij}$: there are $n^2 - n$ off-diagonal entries in $\hat Q$ and each observation only affects one off-diagonal entry leaving most entries zero. Furthermore, as described in the previous section, if the graph of pairwise comparisons (given by connecting any two points with an edge) is not strongly connected, may not guarantee that $\hat{Q}$ has a unique stationary distribution.
Motivated by this, we ask a natural question—when the number of pairwise comparisons is small; i.e., data is scarce (for example we have just observed one comparison per item) how can we still obtain a reasonable ranking?
Intuitively, if the items $[n]$ have some inherent structure, we can hope to exploit that structure to infer pairwise comparisons.
Since $Q_{ij} = \mu_{ij}P_{ij}$; i.e., a scaled probability of $i$ beating $j$, even if we have never seen a comparison between $i$ and $j$, it is reasonable to estimate this value by taking a weighted combination of the empirical $\hat Q_{ik}, 1\leq k\leq n$, where the choice of weights perhaps reflect some prior knowledge on the similarity between $j$ and $k$. In an extreme case—if we suspect item $j$ and $k$ would perform the same against item $i$, we may choose the weight on $\hat Q_{ik}$ to be large, and set the weights on all other $\hat Q_{ik'}, k\neq k'$ to zero.
Said more precisely, we choose a row-stochastic matrix $D$ and use the estimator $\hat{Q}D$ whose $ij$-th entry is
\begin{equation} \label{eq:QDdefn}
[\hat Q D]_{ij} = \sum_{k=1}^n D_{kj} \hat Q_{ik}
\end{equation}
How should we choose $D$? We want $\hat QD$ to be ergodic, but it should also reflect some similarity structure between the items. This prior information could take form in many ways—for example we can imagine that associated to item $i$ is a feature vector $x_i\in \mathbb{R}^d$ and intuitively items that are close together perform similarly on a comparison with some other element $j$ (see Section <ref>). An extreme case of this is assuming that the items are in clusters, and items within a cluster rank similarly (or the same). Finally, we can consider forms of $D$ that do not reflect any prior structure but do at least guarantee that $\hat QD$ is ergodic—as we will show these estimators can still perform competitively with other methods (Section <ref>). To recap, our resulting regularized RankCentrality algorithm that we will discuss in the rest of this section is given below in Algorithm <ref>.
Regularized RankCentrality algorithm
[1]
RankCentrality$n,S, D$
compute $\hat Q$ as in (<ref>)
return leading left eigenvector of $\hat Q D$
§.§ Diffusion Based Regularization
Diffusion RankCentrality leverages additional features $x_i \in \R^d$ for each of the items $i \in [n]$ being ranked. We use this to compute pairwise similarities in a manner consistent with the literature (e.g., in $t$-SNE <cit.> and diffusion maps formulated by <cit.>) so that for a fixed $i$, the similarities $D_{ik}$ are proportional to the probability density of a Gaussian centered at $x_i$. Let $D^{(\sigma)}_{ik}$, the similarity between item $i$ and $j$, be defined as
\begin{equation}
D_{ik}^{(\sigma)} := \frac{\exp\left(\frac{-\|x_i -x_k\|^2}{\sigma^2}\right)}{\sum_{l=1}^n \exp\left(\frac{-\|x_i -x_l\|^2}{\sigma^2}\right)},
\label{eq:similarityD}
\end{equation}
where $\sigma$, the kernel width, is an appropriately chosen hyperparameter. The Diffusion RankCentrality algorithm, obtained by using $D^{(\sigma)}$ in Algorithm <ref>, returns the stationary distribution of the Markov chain $\hat Q D^{(\sigma)}$.
As described in equation (<ref>), $[\hat Q D^{(\sigma)}]_{ij} = \sum_{k=1}^n D_{kj}^{(\sigma)} \hat Q_{ik}$,
i.e., the $ij$ entry is a weighted average of $\hat Q_{ik}$'s. $D_{ij}^{(\sigma)}$ is large when $x_i$ is close to $x_j$ and close to 0 when they are far apart. In particular the $\hat{Q}_{jk}$ contribute more when $j$ is close to $i$ and less otherwise.
An alternative interpretation of this procedure is given by considering the Markov chain induced by $\hat{Q}$ and contrasting it with that of $\hat{Q}D^{(\sigma)}$. Consider starting at any item $i$, and repeatedly transitioning according to $\hat{Q}$. If the number of comparisons is small, there may not even be a path from $i$ to any other item $j$. In addition, any additional comparison greatly affects the stationary distribution (i.e. the limiting distribution as we transition according to $\hat{Q}$) of $\hat{Q}$. Contrast this with the stationary distribution of $\hat{Q}D^{(\sigma)}$. By construction, $\hat{Q}D^{(\sigma)}$ will be dense (assuming each element has some neighbor that has a comparison). We can interpret the elements of $\hat{Q}D^{(\sigma)}$ as a Markov chain themselves: first, we make a sub-step (say from $i$ to $k$) according to $\hat Q$, which is based only the pairwise comparison observations, and then we make a sub-step (say from $k$ to $j$) with probability that inversely depends the distance of points to $k$. In, particular, we have imputed a series of transitions from $i$ to other elements $j$, using the underlying geometry of the points along with the pairwise comparisons. This technique is similar to that found in <cit.>, the MAGIC algorithm used in the field of single-cell RNA sequencing, where each entry in $Q$ is an extremely undersampled low integer count.
Consider the following extreme case example. Suppose the 100 points $\{x_i\}_{i=0}^{99}$ lie in 10 tight clusters with cluster $k$ being $\{x_{10k+1}, \cdots, x_{10k+9}\}$ and the clusters are spaced very far apart. Assume the BTL scores of items are constant within clusters; if items $i$ and $j$ are in the same cluster then $x_i = x_j$ and $w_i = w_j$. Set $\|x_i - x_j\| = \infty$ when $i$ and $j$ are in different clusters.
In this case, the matrix $D^{(\sigma)}$ is block diagonal: $D^{(\sigma)}_{ij} = \frac{1}{10}$ when $i$ and $j$ are in the same cluster and $D^{(\sigma)}_{ij} = 0$ otherwise.
Figure <ref> demonstrates the benefit of multiplying $\hat Q$ by $D^{(\sigma)}$. We see that a comparison between $i$ and $j$ does not just affect the $ij$ entry, but those corresponding to neighbors of $i$ and $j$. To visualize the effect of $D^{(\sigma)}$, we also show heatmaps of the 50-th powers of the transition matrices, $\hat Q$ and $\hat Q D^{(\sigma)}$. The checkered patterns in $Q$ and $QD^{(\sigma)}$ are clearly visible in $(\hat QD^{(\sigma)})^{50}$ while $\hat Q^{50}$ is still very sparse. After 50 iterations of $\hat{Q}$ vs. $\hat{Q}D^{(\sigma)}$, we see the impact of regularization, $(\hat QD^{(\sigma)})^{50}$ is far less sparse than $\hat{Q}^{50}$ and reflects a block structure that is imputing comparisons for items that have been compared less often.
Demonstrating the impact of $D^{(\sigma)}$.
The 100 items in this experiment lie in 10 equally sized tight clusters, where BTL scores are constant within clusters and the corresponding $D^{(\sigma)}$ matrix is block diagonal. The $\hat Q$ matrix was computed using 200 pairwise comparisons simulated according to the BTL model.
There are a number of different ways we could have diffused the information across the samples. We could have used $\hat Q D^{(\sigma)}$, $D^{(\sigma)}\hat Q$, or even $D^{(\sigma)} \hat Q D^{(\sigma)}$. In our empirical analysis, however, we found no significant difference in the performance of the algorithm run with these possibilities.
Finally, we note that the running time of the regularized RankCentrality algorithm is dominated by the computation of the leading eigenvector. The matrices $Q$ and $D$ are of size $n \times n$ and we can form the matrix $M = \hat Q D$ in time $O(n^3)$. We then iterate in the power method with $M$, each iteration, requiring a matrix-vector multiply takes time $O(n^2)$.
Our empirical analysis suggests that a few steps of the power method are sufficient. Furthermore, this iterative eigenvector computation on sparse matrices can be faster, than optimization procedures inherent in the MLE.
§.§ Lambda-Regularized RankCentrality
Implicitly, $D$ is chosen so that two properties are satisfied. Firstly, $\hat QD$ will be an ergodic markov chain, and secondly, as in most regularization situations, we choose $D$ to capture some inherent prior structural information we may have about $w$ apriori. In this section we ignore the second motivation and instead focus on a $D$ which just guarantees that former constraint.
In particular, given $\lambda > 0$ we consider $D_\lambda := (1-\lambda) I + \frac{\lambda}{n} \one \one^T$ as a choice of regularizer in Algorithm <ref>. Note that $\hat QD_\lambda = (1-\lambda) \hat Q + \frac{\lambda}{n} \one \one^T$, which ensures that $\hat Q D_\lambda$ is a positive row-stochastic matrix, which must be ergodic.
In particular, we can run Algorithm 1, regardless of the number of samples and we are guaranteed that $\hat Q D_\lambda$ necessarily has a unique stationary distribution. The simple nature of $D_{\lambda}$ allows us to give a precise theoretical characterization of it's performance.
In general, $\E[\hat QD_{\lambda}] = QD_{\lambda}$, but $QD_{\lambda}$ may not have the same left eigenvector as $Q$. This introduces a bias in our estimator.
How can we overcome this bias? Inspecting the form of $D_{\lambda}$, note that if $\lambda \to 0$ as $m\rightarrow \infty$ then $D_{\lambda} \rightarrow I$. The following theorem characterizes the error of this procedure of any $\lambda$ and shows that it is reasonable to take $\lambda = O(1/\sqrt{m})$.
For notational convenience, we let $\gamma := \frac{n\mumin}{2(1+\sqrt{2})b^{3/2}}$.
Note that $\gamma$ is not constant—in fact it is
Let $\lambda \in (0, \frac{\gamma}{2})$. Choose $\delta \in (0, 1)$ and $\varepsilon \in \left(2\lambda\gamma^{-1}, 1\right)$. Let $\hat w_{\lambda}$ be the output of Regularized RankCentrality run with $D=D_{\lambda}$. Then, with probability at least $1-\delta$,
\begin{equation*}
\frac{\|\hat w_{\lambda} - w\|}{\|w\|} < \ 2 \lambda \gamma^{-1} +\sqrt{\tfrac{68(1-\lambda)b^{3}(\mumax + n\mumax^2)}{n\mumin^2 m} \log\frac{2n}{\delta}},
\end{equation*}
In particular, choosing $\lambda = c/\sqrt{m}$, then with probability at least $1-\delta$, we have
\[ \frac{\|\hat{w} - w\|}{\|w\|} = O\left(\frac{b^3\log(2n/\delta)}{n\mu_{\min}m}\right). \]
We give a proof in the supplementary material under Corollary <ref>.
Our empirical experiments run with $\lambda = \eta m^{-1/2}$ for various values of $\eta$ support decaying $\lambda$ in this way. Figure <ref> demonstrates a run of $\lambda$-Regularized RankCentrality on a setting where $w = [i]_{i=1}^{200}$ and the underlying distribution on pairwise comparisons is assumed to be uniform. We compare several choices of $\lambda$ (with $\lambda = 0$ corresponding to normal RankCentrality) and the BTL MLE with an $\ell_2$ regularizer[Without such a regularizer, the BTL-MLE is underdetermined when the number of comparisons is small and cannot be solved.] on the weights (implemented using logistic regression). Note that $\eta = 1/6$ seems to perform the best and even outperforms regularizing the BTL-MLE for small sample sizes where RankCentrality may still be returning a uniform distribution.
For more details and experiments with different choices of $w$ in this setting, see Appendix <ref> in the supplementary materials.
Remark: To connect the diffusion based regularization with $\lambda$-regularization, observe that if we take $\sigma \to 0$ in the definition of $D$ in Equation <ref>, then $D\to D_0 = I_n$ (when the $x_i$'s are all distinct). The kernel width $\sigma$, therefore, determines the bias of Diffusion RankCentrality—small values of $\sigma$ only introduce a small bias in the algorithm while large values of $\sigma$ introduce considerable bias. Motivated by Theorem <ref>, to diminish this bias as $m$ increases, we can use $(1-\tfrac{1}{\sqrt{m}})I + \frac{1}{\sqrt{m}}D^{(\sigma)}$ in Diffusion RankCentrality instead of $D^{(\sigma)}$ directly. We call this Decayed Diffusion RankCentrality. In general, cross-validation could be used to choose the kernel width.
Comparing $\lambda$-Regularized RankCentrality with BTL-MLE and RankCentrality. Here $w = [i]_{i=1}^{200}$.
§ EMPIRICAL RESULTS FOR REGULARIZED RANKCENTRALITY
In this section we do a comparison of the regularized RankCentrality methods in the structured setting to standard methods for ranking on synthetic and real world datasets. The code we used along with additional plots are part of the supplementary material. Although our theoretical analyses do not make assumptions about $\mu$, our experiments focus on the case where $\mu$ is uniform.
§.§ Comparison to Scoring Functions
As discussed in Section <ref>, there is a rich literature of ranking methods, though less so for ranking data that come with features. Recall, we assume for each item $i \in [n]$ there is a vector $x_i \in \R^d$. In past work, the goal is to learn a function $f:\mathbb{R}^d\rightarrow\mathbb{R}$, presumed to be in a specified function class $\mathcal{F}$, such that $\sign(f(x_i) - f(x_j))$ predicts a comparison between item $i$ and item $j$. To learn $f$ given the dataset $S = \{(i_k, j_k, y_k)\}_{k=1}^m$, and a loss function $\ell:\mathbb{R}\times\mathbb{\R}\times \{0,1\}\rightarrow\mathbb{R}$, we can learn the empirical risk minimizer $\argmin_{f\in \mathcal{F}}\sum_{k=1}^n \ell(f(x_i), f(x_j), y_k)$. Two notable examples that focus on learning a scoring function that we compare to are RankSVM by <cit.> and Siamese network based approaches due to <cit.>.
RankSVM assumes that $\mathcal{F} = \{f: x \mapsto w^T x\}$, i.e. linear separators through the origin and choose $\ell(f(x_i), f(x_j), y) = \min(0, 1-(f(x_i)- f(x_j))(2y-1)$.
When testing RankSVM, we used it naively on the original features but also considered a kernelized version using random features, as described in <cit.> and implemented in SkLearn, <cit.>.
Note that when the loss function is the logistic loss, $\ell(f(x_i), f(x_j), y) = \log\left(\frac{\exp(f(x_j))}{\exp(f(x_i))+\exp(f(x_j))}\right)$,
we recover the MLE under the assumption that the BTL scores are given by a transformation of the features. Such an objective has been proposed several times in the literature, e.g. <cit.>.
In the extreme case $f(x_i) = \theta_i$ is the BTL-MLE.
An example of such an approach are Siamese Nets, introduced by in <cit.>.
We implemented a Siamese network using Keras (<cit.>) with two hidden dense layers, each with 20 nodes and a dropout factor of 0.1, and an output dimension of 1. Each layer in the base network used a ReLU activation. The outputs of the right network is subtracted from that of the left and a cross-entropy loss is then used.
We point out that in general both methods described above have a very different goal from what our paper proposes. Our goal is not to learn a scoring function, but instead to use the similarity information to inform the ranking process. In general, learning a scoring function can be expensive in terms of both computation, and samples. In addition, if the features do not actually inform the ranking very well, we want methods that will still learn a reasonable ranking—guaranteed by regularized RankCentrality as $m\rightarrow \infty$. We now demonstrate competitive performance of regularized RankCentrality even when the data is generated by a scoring function.
We constructed two synthetic datasets. We assume that the BTL-score is given by a continuous function of the features; i.e., there is an $f:\mathbb{R}^d\rightarrow \mathbb{R}$ so that the BTL score $w_i = f(x_i)$. This intuitively captures the idea that items which are close in space are close in rank. We consider a few examples of such functions $f$ as given below.
* In Experiment A, we generated 1600 points $\{x_i\}_{i=1}^{1600}$ chosen uniformly at random from $[0, 4]^2$, we chose $\omega_1, \omega_2, \dots, \omega_4 \in \R^2$ at random, each entry chosen independently from a Gaussian. To each $i \in [1600]$ we associate a score $w_i = \sum_{h=1}^2 \exp(\cos(5 \omega_h^T x_i)) + \sum_{h=3}^4 \exp(\omega_h^T x_i / 10)$.
* In Experiment B, we generated 1000 points $\{x_i\}_{i=1}^{1000}\in [0,4]$ chosen uniformly at random and chose $\omega \in \R$ at random from a Gaussian. To each $i \in [1000]$ we associate a score $w_i = \exp(\cos(5\omega x_i))$.
For varying of $m$, we simulated $m$ observations under the BTL-model with uniform $\mu$ and ran various algorithms that have been discussed. We recorded plotted the average Kendal-tau correlation metric (see Section <ref> in the supplementary for details) between the ranking on the synthetic scores we generated and the true ranking on the items. The results of these experiments are summarized in Figures <ref> and <ref>.
Comparison of algorithms in synthetic experiment A. Diffusion RankCentrality was run with kernel width $\sigma = 2^{-4}$.
Comparison of algorithms in synthetic experiment B. Diffusion RankCentrality was run with kernel width $\sigma = 2^{-5}$.
In Experiment A, Diffusion RankCentrality proves to be the best method when the comparisons are scarce. The impact of Diffusion RankCentrality in Experiment B is dramatic when compared to $\lambda$-regularized RankCentrality. While it is true that RankSVM with random features far outperforms other algorithms, it should not come as a surprise given that the BTL scores $w_i$, as a function of $x_i$, come from monotonic transformations of linear combinations of the basis of the RKHS used for the implementation of random Fourier Features in scikit-learn <cit.>.
In both experiments, Diffusion RankCentrality outperforms Siamese Networks.
To choose the kernel width, we ran Decayed Diffusion RankCentrality with several different choices of $\sigma$ on a validation set and chose the best one (see Figure <ref>).
Impact of kernel width on performance of Diffusion RankCentrality.
§.§ New Yorker Caption Competition
It is challenging to find real-life data sets that satisfy all of the following conditions: 1) The data is structured; i.e., has image or text features associated with the items and 2) the number of items compared is moderate to large in size.
The New Yorker Caption Competition dataset consists of a cartoon and a series of associated (supposedly) funny captions submitted by readers (see <cit.> for details on this dataset). Each week, readers vote on whether they think each caption is funny (2 points), somewhat funny(1 point) or unfunny (0 points), and the caption is assigned an average cardinal score based on these points. Included in this dataset are only two contests (#508 and #509), in which there are a large number of pairwise comparisons in addition to cardinal scores generated from user votes on a small number of items ($n = 29$ items for each contest). Each pair of items received roughly 300 comparisons and each item also received roughly 200 cardinal votes. (The associated captions and visuals of the query types are given in Figure <ref>, and Figure <ref> in the supplementary material). Run directly on this dataset, Diffusion Rank Centrality did not show an appreciable advantage since the number of items was so small and hence similarity information provided less leverage over other methods.
New Yorker Caption Competition Interface for pairwise comparisons for #508. Users were asked to click on the caption they thought was funnier.
A sample of the voting user interface presented to readers of the New Yorker Magazine for contest #651
§.§.§ Cardinal Scores model BTL-scores
We generate comparisons on a much larger set of captions for a different contest by transforming the cardinal data to infer pairwise comparisons. To determine this transformation, we used contest #508 for which we had 300 pairwise comparisons and 200 cardinal votes. For each pair of captions $i,j$ in contest #508, we compute $\hat{P}^{\text{emp}}_{ij}$, the empirical probability of item $i$ beating item $j$. In addition, we used the average empirical cardinal scores of items $i$ and $j$ denoted as $\hat{s}_i, \hat{s}_j$ we computed $\hat{P}^{\text{card}}_{ij} = \exp(\hat{s}_i)/(\exp(\hat{s}_i)+\exp(\hat{s}_j))$. In other words, we calculated the empirical probabilities implied by the cardinal scores and compared them to the empirical probabilities from the pairwise comparisons. A resulting scatterplot of the points $(\hat{P}^{\text{emp}}_{ij},\hat{P}^{\text{card}}_{ij})$ is shown in Figure <ref>. Somewhat surprisingly, this plot demonstrates that a monotonic transformation of the cardinal scores seem to model an underlying pairwise probability model fairly well—implying that up to an exponential scaling transformation, the cardinal scores determine underlying BTL scores for the captions. This seems to be an interesting non-trivial result about ranking and humor that has not been previously observed.
Scatter plot demonstrating the relationship between $\hat P^\text{emp}$ and $\hat P^\text{card}$.
§.§.§ Contest #651
Using the observations in the previous section, we chose a contest, #651, that did not have underlying pairwise comparisons but did have a large number of items all with cardinal scores. We then generated pairwise comparisons from these cardinal scores as described in Section <ref>. The cartoon associated to this contest is in Figure <ref>.
More precisely, from the captions available, we took the 400 captions (out of roughly 7000) with largest empirical average cardinal score (each caption had around 250 votes) and generated BTL weights.
We used the Universal Sentence Encoder in <cit.> to generate 512 dimensional embeddings for each of the captions (this yields the additional structural information we need for regularization). The resulting plot contrasting the methods is shown in <ref>, as before the kernel width was chosen on a validation set—in addition we used $(1-\tfrac{1}{\sqrt{m}})I + \frac{1}{\sqrt{m}}D^{(\sigma)}$ as the regularizer in Diffusion RankCentrality to debias the procedure.
In this setting, Diffusion RankCentrality performs extremely well, locking in a significantly better ranking almost immediately with few comparisons.
Test Error for various algorithms for the New Yorker Caption Competition #651 with $\sigma=.25$.
§.§ Place Pulse
Our final example involves comparisons arising from the Place Pulse dataset used in <cit.>. There were 100 images of locations in Chicago in this dataset, and a total of 5750 comparisons where MTurk workers were asked which of the two locations they thought were safer.
We used ResNetV1 <cit.> to generate features for the images of each location and broke the data up into a train, test and validation set (again used to select $\sigma$ and $\lambda$). Since we do not have an underlying ground truth ranking, we instead plot the test error in Figure <ref>.
Performance of various algorithms from the Place Pulse dataset.
Again, Diffusion RankCentrality (a non-classification based method) performed competitively matching the performance of RankSVM.
§ CONCLUSION
In this paper we provided a way to employ structure in the RankCentrality algorithm that provides meaningful results when data is scarce.
Along the way we provided a stronger sample complexity bound for a natural sampling scheme. For future work we hope to provide rigorous sample complexity bounds for diffusion based methods.
§.§.§ Acknowledgements
The first and third authors were supported by the MIDAS Challenge Grant from the University of Michigan. The first author had the initial idea and motivation for this work while at Agero, Inc., and would like to thank Michael Bell.
§.§.§ References
Missing 'biblatex' package
The bibliography requires the 'biblatex' package.
booktitleInternational Conference on Machine Learning
titleAccelerated spectral ranking
booktitleAdvances in Neural Information Processing Systems 6
titleSignature Verification using a “Siamese” Time Delay Neural
booktitleProceedings of the 22nd International Conference on
Machine learning (ICML-05)
titleLearning to rank using gradient descent
titleUniversal sentence encoder
journaltitlearXiv preprint arXiv:1803.11175
The Institute of Mathematical Statistics
titleSpectral method and regularized MLE are both optimal for
top-$K$ ranking
journaltitleAnn. Statist.
National Academy of Sciences
titleGeometric diffusions as a tool for harmonic analysis and
structure definition of data: diffusion maps
journaltitleProceedings of the National Academy of Sciences of the
United States of America
titleRecovering Gene Interactions from Single-Cell Data Using Data
booktitle2016 IEEE Conference on Computer Vision and Pattern
Recognition (CVPR)
titleDeep Residual Learning for Image Recognition
The Institute of Mathematical Statistics
titleActive ranking from pairwise comparisons and when parametric
assumptions do not help
journaltitleAnn. Statist.
Curran Associates, Inc.
booktitleAdvances in Neural Information Processing Systems 24
titleActive Ranking using Pairwise Comparisons
booktitleProceedings of the Eighth ACM SIGKDD International
Conference on Knowledge Discovery and Data Mining
seriesKDD '02
titleOptimizing Search Engines Using Clickthrough Data
Edmonton, Alberta, Canada
* Can't use 'location' + 'address'
booktitleInternational Conference on Artificial Intelligence and
titleAdaptive Sampling for Coarse Ranking
titleMatrix factorization techniques for recommender systems
given=Laurens vander,
titleVisualizing data using t-SNE
journaltitleJournal of machine learning research
titleRank centrality: Ranking from pairwise comparisons
journaltitleOperations Research
titleData from the New Yorker Caption Contest
Cambridge University Press
seriesCambridge Series in Statistical and Probabilistic
titleMarkov Chains
titleScikit-learn: Machine Learning in Python
journaltitleJournal of Machine Learning Research
Curran Associates, Inc.
booktitleAdvances in Neural Information Processing Systems 20
titleRandom Features for Large-Scale Kernel Machines
booktitleProceedings of the 31st International Conference on
Machine Learning
titleA Statistical Convergence Perspective of Algorithms for Rank
Aggregation from Pairwise Data
booktitle2018 IEEE International Symposium on Information Theory
titleLow permutation-rank matrices: Structural properties and
noisy completion
JMLR. org
titleSimple, robust and optimal ranking from pairwise comparisons
journaltitleThe Journal of Machine Learning Research
JMLR. org
titleEstimation from pairwise comparisons: Sharp minimax bounds
with topology dependence
journaltitleThe Journal of Machine Learning Research
titleUser-Friendly Tail Bounds for Sums of Random Matrices
journaltitleFoundations of Computational Mathematics
Symbol Definition
$\|\cdot\|$ unless stated otherwise, vector norms are $\ell_2$ norms, and matrix norms are operator (spectral) norms
$\gamma$ $\frac{n\mumin}{2(1+\sqrt{2})b^{3/2}}$
$w$ stationary distribution of $Q$
$\hat w$ stationary distribution of $\hat Q$
$\lambda$ regularization constant, see $D_\lambda$
$\lmax(R)$ second largest eigenvalue of matrix $R$ (because the largest eigenvalue of an irreducible Markov chain is always 1)
$\mu_{ij}$ probability that pair $(i,j)$ is observed
$\one$ vector of all one entries, usually in $\R^n$
$b$ $\max_{i,j} \frac{w_i}{w_j}$
$k$ number of comparisons per pair in sampling scheme in <cit.>
$n$ number of items being compared
$m$ number of comparisons total
$P$ pairwise preference matrix
$\hat P$ empirical comparison matrix
$Q$ true markov chain (requires knowing $P$)
$\hat Q$ empirical markov chain
$D_\lambda$ $(1-\lambda) I + \frac{\lambda}{n} \one \one^T$
Notation used in this paper.
§ CONVERGENCE OF RANKCENTRALITY
\begin{equation}
Q^{(ij)} := e_{i}e_{j}^T - e_{i} e_{i}^T,
\label{eq:Qij}
\end{equation}
and additionally
\begin{equation}
Q_k =
\begin{cases}
Q^{(j_ki_k)} & \text{ if } y_k=0 \\
Q^{(i_kj_k)} & \text{ if } y_k=1
\end{cases}.\label{eq:Qk}
\end{equation}
We see now that
\begin{equation}
\hat{Q} = I + \frac{1}{m} \sum_{k=1}^m Q_k,\label{eq:Qhatdefnsum}
\end{equation}
and for the remainder of our analysis we shall consider (<ref>) as the definition of $\hat Q$. Recall
\[Q_{ij} =
\begin{cases}
\mu_{ij} P_{ij} & \text{ if } i \neq j \\
1 - \sum_{k\neq i} \mu_{ik} P_{ik} & \text{ if } i = j
\end{cases},
\]
and observe that $\E(\hat Q) = Q$.
We begin our analysis of the RankCentrality algorithm by giving a bound on the spectral gap of the transition matrix $Q$ constructed from pairwise preferences.
The spectral gap $1-\lmax$ of $Q$ is at least $\frac{n\mumin}{2b}$, where $b = \max_{i,j} \frac{w_i}{w_j}$.
We will use the following lemma from <cit.>.
Let $Q,\pi$ and $R,\tau$ be reversible Markov chains on a finite set $[n]$ representing random walks on a graph $G=([n],E)$, i.e.
$R(i,j)=0$ and $Q(i,j)=0$ if $(i,j)\notin E$.
For $\alpha\equiv\min_{(i,j)\in E}\{\pi_iQ_{ij}/\tau_iR_{ij}\}$ and
\[
\frac{1-\lmax(Q)}{1-\lmax(R)} \geq \frac{\alpha}{\beta}
\]
We will invoke the above lemma with $R = \frac{1}{n} \one \one^T = [\frac{1}{n}]_{ij}$, $\tau = \frac{1}{n} \one = [\frac{1}{n}]_i$, $Q$ as we have defined it previously, and $\pi = w$. Observe that these define a reversible Markov chain. Since $R$ has rank 1, we have $\lmax(R) = 0$, which gives us that $1 - \lmax(Q) \geq \frac{\alpha}{\beta}$. Now we bound $\alpha$ and $\beta$.
We have
\begin{align*}
\alpha = & \ \min_{i,j} \frac{w_iQ_{ij}}{\tau_iR_{ij}} = \min_{ij} \frac{w_i \mu_{ij} \frac{w_j}{w_i + w_j}}{\frac{1}{n} \frac{1}{n}} \geq \ \min_{i,j} \frac{n^2\mumin w_iw_j}{(w_i+w_j)} \geq \frac{n^2 \mumin \min_i w_i}{2}
\end{align*}
We also see $\beta = \max_i \frac{w_i}{\tau_i} = n \max_i w_i$. Thus, $\frac{\alpha}{\beta} \geq \frac{n\mumin}{2b}$.
This bound is close to optimal when $\mu$ is uniform. Since the diagonal entries of $Q$ are each at least $1 - \frac{2}{n-1}$, we know $\frac{n-1}{2}(Q - (1-\frac{2}{n-1})I)$ is non-negative and row stochastic. By the Perron-Frobenius Theorem, the eigenvalues of $\frac{n-1}{2}(Q - (1-\frac{2}{n-1})I)$ lie in $[-1,1]$ and the eigenvalues of $Q$ must lie in $[1-\frac{4}{n-1}, 1]$. The difference between 1 and the smallest possible eigenvalue of $Q$ is only a factor of $4b$ larger than our bound on the spectral gap.
Let $Q$ be the true transition matrix as defined in (<ref>). For any ergodic Markov chain on $[n]$ with row-stochastic transition matrix $\tilde Q$ and stationary distribution $\tilde{w}$, if $\|Q - \tilde{Q}\| < \frac{n\mumin}{2 b^{3/2}}$, we have
\[ \frac{\|\tilde{w} - w \|}{\|w\|} \leq \frac{2\|\Delta\|b^{3/2}}{n\mumin - 2\|\Delta\|b^{3/2}},\]
where $\Delta = \tilde{Q} - Q$.
We begin by citing a lemma <cit.>.
For any Markov chain $\tilde Q=Q+\Delta$ with a reversible Markov chain $Q$,
let $p_t$ be the distribution of the Markov chain $\tilde Q$ when
started with initial distribution $p_0$. Then,
\begin{align*}
\frac{\left\|p_t-w\right\|}{\|w\|} \leq \rho^t\frac{\|p_0-w\|}{\|w\|}\sqrt{\frac{w_{\rm max}}{w_{\rm min}}} + \frac{1}{1-\rho}\|\Delta\|_2\sqrt{\frac{w_{\rm max}}{w_{\rm min}}}\;.
\end{align*}
where $w$ is the stationary distribution of $Q$
and $\rho=\lmax(Q)+\|\Delta\|_2\sqrt{w_{\rm max}/w_{\rm min}}$.
As before, let $b = \max_{i,j} \frac{w_i}{w_j}$. Consider the limit as $t\to\infty$:
* when $0 \leq \rho < 1$ we have $\rho^t \to 0$, and
* when the Markov chain $\tilde Q$ is irreducible we have $p_t \to \tilde w$.
In this case,
\begin{align*}
\frac{\left\|\tilde w-w\right\|}{\|w\|} \leq \frac{1}{1-\rho}\|\Delta\|_2\sqrt{b}.
\end{align*}
Recall that $1 - \lmax(Q) > \frac{n\mumin}{2 b}$ by Proposition <ref>. Now we have that $\rho < 1$ when $\|\Delta\| < \frac{n\mumin}{2b^{3/2}}$ because when this is the case, we have $\|\Delta\|\sqrt{b} < \frac{n\mumin}{2b}$ and hence $\rho \leq 1 - \frac{n\mumin}{2b} + \|\Delta\|\sqrt{b} < 1$. Assuming $\|\Delta\| < \frac{n\mumin}{2b^{3/2}}$, we have
\[\frac{\|\tilde w - w\|}{\|w\|} \leq \frac{\|\Delta\|\sqrt{b}}{\frac{n\mumin}{2b} - \|\Delta\|\sqrt{b}} = \frac{2\|\Delta\|b^{3/2}}{n\mumin - 2\|\Delta\|b^{3/2}}.\]
For transition matrices $Q$ and $\hat Q$ we define the centered transition matrices $Q'$ and $\hat Q'$ by subtracting $I$. That is, $Q' = Q - I$ and $\hat Q' = \hat Q - I$. These centered matrices $Q'$ and $\hat Q'$, as well as $Q_k$ and $Q^{(ij)}$ defined previously, have non-negative entries everywhere except on the diagonal (where they are non-positive) and their rows sum to zero. These centered matrices significantly simplify the algebra in the following computations.
The difference $Z_k := \frac{Q_k - Q'}{m}$ is bounded in norm: $\|Z_k\| < \frac{3}{m}$.
To bound $\|Q_k\|$, recall that $Q_k$ is of the form $Q^{(ij)} = (e_ie_j^T - e_i e_i)$.
Observe that $ Q^{(ij)} Q^{(ij) T} = 2e_ie_i^T$. Therefore, $\|Q_k\| \leq \sqrt{2}$. By convexity of norms, $\|Q'\| = \|\E Q_k\| \leq \E \| Q_k\| \leq \sqrt{2}$. Using the triangle inequality we get $\| Q_k - Q'\| \leq 2\sqrt{2} < 3$.
Let $Z_k = \frac{Q_k - Q'}{m}$, as before. We can bound the variance term as:
\[\sigma^2:= \max \left\{ \left \|\sum_{k=1}^m \E Z_kZ_k^*\right\|, \left \|\sum_{k=1}^m \E Z_k^*Z_k\right\| \right\} \leq \frac{3(n-1)\mumin}{m}.\]
To bound $\|\E Z_k Z_k^*\|$, we see
\[\E Z_k Z_k^* = \frac{1}{m^2} \E \left( Q_k Q_k^{T} - Q_k Q^{\prime T} - Q' Q_k^{T} + Q' Q^{\prime T} \right) = \frac{1}{m^2} \E \left( Q_k Q_k^{T} - Q' Q^{\prime T} \right).\]
We can compute these explicitly.
Begin by considering the $Q_k Q_k^{T}$ term. We know $Q^{(ij)}Q^{(ij)T} = 2e_ie_i^T$. By simple algebra, we get $\E Q_k Q_k^{T} = \sum_{i} \sum_{j\neq i} 2\mu_{ij} P_{ji} e_i e_i^T$. Therefore, $\|\E Q_k Q_k^T \| \leq \max_i \sum_{j\neq i} 2 \mu_{ij} P_{ji} \leq 2 (n-1) \mumax$.
Computing $Q'Q^{\prime T}$ is more tedious.
\begin{align*}
Q' Q^{\prime T} = & \ \left( \sum_{i\neq j} \mu_{ij} P_{ij} (e_i e_j^T - e_i e_i^T) \right)\left( \sum_{u\neq v} \mu_{uv} P_{uv} (e_v e_u^T - e_u e_u) \right) \\
= & \sum_{i\neq j, u\neq v} \mu_{ij}\mu_{uv} P_{ij} P_{uv} (e_i e_j^T e_v e_u^T - e_i e_j^T e_u e_u - e_i e_i^T e_v e_u^T + e_i e_i^T e_u e_u^T).
\end{align*}
By ignoring zero terms (notice that the first of four summands is non-zero only when $j=v$, the second when $j=u$, etc.) and re-indexing, we get
\begin{align*}
Q' Q^{\prime T} = & \left( \sum_{i\neq \ell \neq j} \mu_{i\ell} \mu_{j\ell} P_{i\ell}P_{j\ell} e_i e_j^T - \sum_{i \neq j \neq \ell} \mu_{ij} \mu_{j\ell} P_{ij} P_{j\ell} e_i e_j^T - \sum_{j \neq i \neq \ell} \mu_{i\ell} \mu_{ji} P_{i\ell} P_{ji} e_i e_j^T + \sum_{u \neq i \neq v} \mu_{iu} \mu_{iv} P_{iu} P_{iv} e_i e_i^T \right),
\end{align*}
where statements such as $i\neq \ell \neq j$ mean $i \neq \ell$ and $j \neq \ell$ (but $i$ may be equal to $j$).
This is a symmetric matrix, so its singular values are its eigenvalues. We can now invoke the Gershgorin circle theorem, a consequence of which is that $\|M\| < \max_i \sum_j |M_{ij}|$ for symmetric matrices. Therefore, $\|Q' Q^{\prime T}\| \leq 4n^2\mumax^2$. Finally, the triangle inequality gives $\|\E Z_k Z_k^* \| \leq \frac{1}{m^2} \left( 2(n-1)\mumax + 4n^2\mumax^2 \right)$.
We now turn to $ Z_k^* Z_k$. Similar to the calculations above, simple algebra gets us
\[\E Q_k^{T} Q_k = \sum_i \sum_{j\neq i} \mu_{ij}( P_{ij} + P_{ji} ) (e_ie_i^T - e_i e_j^T).\]
As before, this is a symmetric matrix and we can use the Gershgorin circle theorem to give a bound on the largest singular value of $\E Q_k^T Q_k$:
\[ \|\E Q_k^T Q_k\| \leq \max_i \sum_{j\neq i} 2\mu_{ij} \leq 2 (n-1)\mumax.\]
As before computing $Q^{\prime T} Q'$ is more tedious but gives
\begin{align*}
Q^{\prime T} Q' = & \ \sum_{i \neq j} \sum_{u \neq v} \mu_{ij} \mu_{uv} P_{ij} P_{uv} (e_j e_i^T - e_i e_i^T)(e_u e_v^T - e_u e_u^T) \\
= & \ \sum_{i \neq j} \sum_{u \neq v} \mu_{ij} \mu_{uv} P_{ij} P_{uv} (e_j e_i^T e_u e_v^T - e_j e_i^T e_u e_u^T - e_i e_i^T e_u e_v^T + e_i e_i^T e_u e_u^T) \\
= & \ \sum_{i \neq j} \sum_{v \neq i} \mu_{ij} \mu_{uv} P_{ij} P_{iv} (e_j e_v^T - e_j e_i^T - e_i e_v^T + e_i e_i^T) \\
= & \ \sum_{i \neq j} \left( \sum_{\ell \neq i; \ell \neq j} \mu_{\ell i}\mu_{\ell j} P_{\ell i}P_{\ell j} - \mu_{ji} \mu_{j\ell} P_{ji}P_{j\ell} - \mu_{i\ell} \mu_{ij} P_{i\ell} P_{ij} \right)e_i e_j^T \\
& \qquad \qquad + \sum_{i} \left( \sum_{u\neq i,v\neq i} \mu_{iu} \mu_{iv} P_{iu} P_{iv} + \sum_{\ell \neq i} \mu_{\ell i} \mu_{\ell i} P_{\ell i}P_{\ell i} \right)e_i e_i^{T}.
\end{align*}
Again, we can invoke the Gershgorin circle theorem and see that $\|Q' Q^{\prime T}\| \leq 4n^2\mumax^2$. As before, the triangle inequality gives $\|\E Z_k^* Z_k \| \leq \frac{1}{m^2} \left( 2(n-1)\mumax + 4n^2\mumax^2 \right)$.
Finally, note that $Z_k$ are not only independent but also identically distributed and hence
\[\max \left\{ \left \|\E \sum_k Z_k^* Z_k \right \|, \left \|\E \sum_k Z_k Z_k^* \right\| \right\} = m \max\left\{\|\E Z_k^*Z_k\|, \|\E Z_kZ_k^*\| \right\} \leq \frac{4(n-1)\mumax + 4n^2\mumax^2}{m}.\]
We will soon need to use the Matrix Bernstein Inequality from <cit.> and state it here as a lemma.
Consider a finite sequence $\{ \mathbf{Z}_k \}$ of independent, random matrices with dimensions $d_1 \times d_2$. Assume that each random matrix satisfies
\[
\E \; \mathbf{Z}_k = \mathbf{0}
\quad\text{and}\quad
\norm{ \mathbf{Z}_k } \leq R
\quad\text{almost surely}.
\]
\[
\sigma^2 := \max\left\{
\norm{ \sum\nolimits_k \E( \mathbf{Z}_k \mathbf{Z}_k^* ) }, \
\norm{ \sum\nolimits_k \E(\mathbf{Z}_k^* \mathbf{Z}_k) }
\right\}.
\]
Then, for all $t \geq 0$,
\[
\PP{\left( \norm{ \sum\nolimits_k \mathbf{Z}_k } \geq t \right)}
\leq (d_1 + d_2) \cdot \exp\left( \frac{-t^2/2}{\sigma^2 + Rt/3} \right).
\]
Finally, we put this all together.
Let $\hat Q$ be constructed as in (<ref>). If $\hat Q$ is ergodic and $\hat w$ is the stationary distribution of $\hat Q$, then we have (where probability is taken over the $m$ comparisons made under the BTL model and each pair is equally likely to get picked)
\[\PP\left(\frac{\|\hat w - w\|}{\|w\|} \leq \varepsilon\right) > 1 - 2n \exp \left( \frac{-\mumin^2\varepsilon^2 n m}{16b^3(1+\varepsilon)^2(\mumax + n\mumax^2)} \right).\]
Assuming $\|\Delta\| < \frac{1}{nb^{3/2}}$, by Proposition <ref> we have
\[\frac{\|\hat w - w\|}{\|w\|} \leq \frac{2\|\Delta\|b^{3/2}}{n\mumin - 2\|\Delta\|b^{3/2}}.\]
This means we want
\[\frac{2\|\Delta\|b^{3/2}}{n\mumin - 2\|\Delta\|b^{3/2}} < \varepsilon,\]
which happens when $\|\Delta\| \leq \frac{\varepsilon n\mumin}{2b^{3/2}(1 + \varepsilon)}$. Note that this is stronger than $\|\Delta\| < \frac{n\mumin}{2b^{3/2}}$, so our previous assumption will hold.
Finally, we let $t = \frac{\varepsilon n\mumin}{2b^{3/2}(1 + \varepsilon)}$ and use Lemma <ref> to get
\[ \PP \left( \frac{\|\hat w - w\|}{\|w\|} \geq \varepsilon \right) \leq \PP \left( \|\hat Q - Q\| \geq t \right) \leq -2n\exp\left( \frac{-t^2}{\sigma^2 + Rt/3} \right),\]
where we have $\sigma^2 \leq \frac{4(n-1)\mumax + 4n^2\mumax^2}{m}$ by Lemma <ref> and $R < \frac{3}{m}$ by Lemma <ref>. Therefore, we get
\begin{align*}
\PP \left( \frac{\|\hat w - w\|}{\|w\|} \geq \varepsilon \right) & \leq 2n\exp\left(\frac{-\left( \frac{\varepsilon n\mumin}{2b^{3/2}(1+\varepsilon)} \right)^2}{\frac{4(n-1)\mumax + 4n^2\mumax^2}{m} + \frac{\varepsilon n \mumin}{2mb^{3/2}(1+\varepsilon)} } \right) \\
& \leq 2n \exp \left( \frac{-\mumin^2 \varepsilon^2 n^2 m}{4b^3(1 + \varepsilon)^2\left( 2n\mumax + 4n^2\mumax^2 \right) + 2b^{3/2}\varepsilon (1 + \varepsilon)n\mumin} \right) \\
& \leq 2n \exp \left( \frac{-\mumin^2\varepsilon^2 n m}{16b^3(1+\varepsilon)^2(\mumax + n\mumax^2)} \right).
\end{align*}
Fix $\delta \in (0, 1)$ and $\varepsilon \in (0, 1)$. If
\[ m \geq 64b^3 n^{-1} \mumin^{-2} \varepsilon^{-2}(\mumax + n\mumax^2) \log \frac{2n}{\delta} \]
and the empirical Markov chain $\hat{Q}$ constructed as in (<ref>) is ergodic, then with probability at least $1-\delta$, we have
\[\frac{\|\hat w - w\|}{\|w\|} \leq \varepsilon.\]
We need
\[ \PP\left( \frac{\|\hat w - w\|}{\|w\|} \geq \varepsilon \right) \leq 2n \exp \left( \frac{-\mumin^2\varepsilon^2 n m}{16b^3(1+\varepsilon)^2(\mumax + n\mumax^2)} \right) < \delta.\]
By re-writing in terms of $m$, we see that the second inequality is true when
\[ m > 16b^3(1+\varepsilon)^2 n^{-1} \mumin^{-2} \varepsilon^{-2}(\mumax + n\mumax^2) \log \frac{2n}{\delta} . \]
The desired inequality now follows immediately from $\varepsilon < 1$ (we make this assumption for simplicity; the statement of the theorem is not very strong when $\varepsilon > 1$).
When $\mu$ is uniform and $n>4$, the above theorem requires $m > 48 b^3 \varepsilon^{-2} n \log (\frac{2n}{\delta})$. We have given an $O\left( \varepsilon^{-2} n\log \frac{n}{\delta} \right)$ upper bound on the sample complexity. This is a much better bound than in <cit.>. Their $O(\varepsilon^{-2} \mumin^{-2} n \log (\frac{n}{\delta}))$ scales as $O(\varepsilon^{-2} n^5\log(\frac{n}{\delta}))$ when $\mu$ is uniform and worse otherwise.
§ CONVERGENCE OF LAMBDA-REGULARIZED RANKCENTRALITY
This section is devoted to an analysis of the bias-variance trade-off of $\lambda$-Regularized RankCentrality. We will compare
* $\hat{\tilde w}$, the leading left eigenvector of $\hat QD_\lambda$, i.e., the output of $\lambda$-regularized RankCentrality, and
* $\tilde w$, the leading left eigenvector of $QD_\lambda$, i.e., the expected output of $\lambda$-regularized RankCentrality as $m\to\infty$,
* $w$, the leading left eigenvector of $Q$, and the expected output of RankCentrality as $m \to\infty$.
Fix $\lambda \in (0, \gamma)$. The asymptotic ($m\to \infty$) expectation of the output of the $\lambda$-Regularized RankCentrality algorithm is $\tilde w$ and the bias $ \|w - \tilde w\| / \|w\|$ can be bounded as
\[ \frac{\|w - \tilde{w}\|}{\|w\|} \leq \frac{ \lambda}{\gamma - \lambda} \]
Let $\tilde{Q} = QD_\lambda$. We now have $Q - \tilde{Q} = \lambda ( \frac{1}{n}\one \one^T - Q)$ and $\|Q - \tilde{Q}\| \leq \lambda (1 + \sqrt{2})$. Now we apply Proposition <ref> to see that
\[ \frac{\|w - \tilde{w}\|}{\|w\|} \leq \frac{2(1+\sqrt{2})\lambda b^{3/2}}{n\mumin - 2(1+\sqrt{2})\lambda b^{3/2}} = \frac{\lambda}{\gamma - \lambda}.\]
Fix $\lambda \in (0, \frac{\gamma}{2})$ and choose $\varepsilon \in ( 2\lambda\gamma^{-1}, 1)$. We construct $\hat Q$ as before and let $\tilde{\hat{w}}$ be the stationary distribution (leading left eigenvector) of $\hat Q D_\lambda$ (i.e., the output of $\lambda$-regularized RankCentrality). We have
\[
\PP\left(\frac{\|\tilde{\hat{w}} - w\|}{\|w\|} < \varepsilon \right) >
1 - 2n \exp \left( \frac{-(n\mumin \varepsilon - 4(1+\sqrt{2})b^{3/2}\lambda)^2m}{16b^3(1-\lambda)^2 \left( 4(n-1)\mumax + 4n^2 \mumax^2 \right) + 4b^{3/2}(1-\lambda)(n\mumin \varepsilon - 4b^{3/2}(1+\sqrt{2})\lambda)} \right)
\]
As we noted in the proof of Theorem <ref>, to guarantee $\|w - \tilde{\hat{w}}\|/\|w\| \leq \varepsilon$, we need $\|Q - \hat Q D_\lambda\| \leq \frac{\varepsilon n \mumin}{2(1 + \varepsilon)b^{3/2}}$. Using the triangle inequality, we have $\|Q - \hat Q D_\lambda\| \leq \|Q - QD_\lambda\| + \|QD_\lambda + \hat Q D_\lambda\|$. We showed in Proposition <ref> that $\|Q - QD_\lambda\| \leq \lambda(1+ \sqrt{2})$. So we need
\begin{align*}
\|QD_\lambda - \hat Q D_\lambda\| & \ \leq \frac{\varepsilon n \mumin}{2(1 + \varepsilon)b^{3/2}} - \lambda(1+ \sqrt{2}) \leq \frac{\varepsilon n \mumin}{4b^{3/2}} - \lambda(1+ \sqrt{2}) \\
& \ = \frac{(1+\sqrt{2})}{2}\varepsilon\gamma\ - \lambda (1+\sqrt{2}) = \frac{(1+\sqrt{2})}{2} (\varepsilon \gamma - 2\lambda)
\end{align*}
Note that this quantity is positive when $\varepsilon \in (2\lambda\gamma^{-1}, 1)$ (which is precisely the requirement in the hypothesis above). We have required that $\varepsilon < 1$ to simplify algebra; the theorem is not very useful otherwise. We now require that
\[
\|QD_\lambda - \hat Q D_\lambda\| \leq \frac{\varepsilon n \mumin}{4b^{3/2}} - \lambda(1+ \sqrt{2}).
\]
We can now invoke Lemma <ref> with $Z_k = \frac{1}{m} (Q'D_\lambda - Q_kD_\lambda) = \frac{1}{m}(1-\lambda)(Q' - Q_k)$. By our previous calculations in Lemmas <ref> and <ref>, we have the variance term $\sigma^2 \leq (1-\lambda)^2 \frac{4(n-1)\mumax + 4n^2\mumax^2}{m}$ and the norm term $R \leq (1-\lambda)\frac{3}{m}$. The resulting inequality is
\begin{align*}
\PP\left(\|QD_\lambda - \hat QD_\lambda\| \geq \frac{n\mumin \varepsilon}{4b^{3/2}} - (1+\sqrt{2})\lambda \right)
\leq 2n\exp\left( \frac{-\left( \frac{n\mumin \varepsilon}{4b^{3/2}} - (1+\sqrt{2})\lambda \right)^2}{(1-\lambda)^2 \frac{4(n-1)\mumax + 4n^2\mumax^2}{m} + \frac{1-\lambda}{m} \left( \frac{n\mumin\varepsilon}{4b^{3/2}} - (1+\sqrt{2})\lambda \right)} \right),
\end{align*}
which simplifies to the desired inequality.
Recall $\gamma = \frac{n\mumin}{2(1+\sqrt{2})b^{3/2}}$. Let $\lambda \in (0, \frac{\gamma}{2})$. Choose $\delta \in (0, 1)$ and $\varepsilon \in \left(2\lambda\gamma^{-1}, 1\right)$. If
\[ m > \frac{68(1-\lambda)b^{3}(\mumax + n\mumax^2)}{n\mumin^2 \left( \varepsilon - 2\lambda\gamma^{-1} \right)^2} \log\frac{2n}{\delta}\]
then with probability at least $1-\delta$, we have
\[ \frac{\|\tilde{\hat{w}} - w\|}{\|w\|} \leq \varepsilon. \]
As in Corollary <ref>, we need
\[
\PP\left(\frac{\|\tilde{\hat{w}} - w\|}{\|w\|} > \varepsilon \right) < \delta,\]
which we can guarantee when
\[
2n \exp \left( \frac{-(n\mumin \varepsilon - 4(1+\sqrt{2})b^{3/2}\lambda)^2m}{16b^3(1-\lambda)^2 \left( 4(n-1)\mumax + 4n^2 \mumax^2 \right) + 4b^{3/2}(1-\lambda)(n\mumin \varepsilon - 4b^{3/2}(1+\sqrt{2})\lambda)} \right) < \delta.
\]
Rewriting in terms of $m$, we see that the second inequality is true when
\begin{equation*}
m > \frac{16b^3(1-\lambda)^2 \left( 4(n-1)\mumax + 4n^2 \mumax^2 \right) + 4b^{3/2}(1-\lambda)(n\mumin \varepsilon - 4b^{3/2}(1+\sqrt{2})\lambda)}{(n\mumin \varepsilon - 4(1+\sqrt{2})b^{3/2}\lambda)^2} \log \frac{2n}{\delta}\\
\end{equation*}
The desired inequality now follows by replacing various terms in the above inequality with upper bounds for them (e.g., $(1-\lambda)^2 < 1 - \lambda$, $b^{3/2} < b^{3}$, and $\varepsilon < 1$).
Empirical evidence suggests that values of $\lambda$ larger than $\frac{\gamma}{2}$ often yield meaningful results. Future work could include bridging this gap between the theory and application.
§ EMPIRICAL RESULTS: RANKCENTRALITY AND LAMBDA-REGULARIZED RANKCENTRALITY
Our main experiments was to evaluate convergence of these algorithms with synthetic BTL scores and comparisons. We compared (unregularized) RankCentrality, $\lambda$-regularized RankCentrality (with $\lambda$ decaying as $\eta m^{-1/2}$ for different values of $\eta$, as described in Section <ref>), the BTL maximum likelihood estimation (see equation (<ref>)), and regularized BTL-MLE (using the Scikit-Learn <cit.> implementation of logistic regression).
The BTL score $w_i$ for each item $i$ was either
* assigned by choosing $v_i$ uniformly at random from $[0, 5]$ and setting $w_i = \exp(v_i)$, or
* deterministically constructed, e.g., $w_i = i$ for $i \in [200]$.
Then, for various values of $m$, we generated $m$ comparisons (first chose $m$ pairs of items, uniformly at random from all possible pairs, then drew winners with probabilities according to the BTL model) and ran each algorithm on the same set of comparisons. In each of these cases, we record the $\ell_2$ error and the Kendall's Tau correlation metric. We repeat this process of generating comparisons and evaluating algorithms for a total of 40 times and record the mean and standard error of the $\ell_2$ error and the Kendall-Tau correlation metric. The results for some of these experiments are shown in Figure <ref>.
$w\in \R^{200}$ chosen at random.
$w\in \R^{40}$ chosen at random.
Decaying $\lambda$ with a factor of $m^{-1/2}$.
§ KENDALL'S TAU-B
The Kendall-Tau correlation metric we use in our experiments is also know as Kendall's Tau-b, defined as
\begin{equation}
\tau(\alpha, \beta) = \frac{P - Q}{\sqrt{(P + Q + T) * (P + Q + U)}},
\label{eq:kendalltaub}
\end{equation}
where $P$ is the number of concordant pairs (i.e., the number of pairs $i,j$ such that the relative ordering of $\alpha_i$ and $\alpha_j$ is the same as that of $\beta_i$ and $\beta_j$), $Q$ the number of discordant pairs, $T$ the number of ties only in $\alpha$, and $U$ the number of ties only in $\beta$.
Synthetic Experiment B.
New Yorker Caption Competition #651
Place Pulse dataset.
Impact of kernel width on Diffusion RankCentrality for various datasets.
§ NEW YORKER CAPTION CONTEST
New Yorker Caption Competition Interface for pairwise comparisons for 508. Users were asked to vote for each caption.
|
$\displaystyle\leq\frac{\max_{i}\|G^{i}_{n}(x_{t},\xi_{t})\|}{\eta}\left(1+2(m-|\mathcal{I}_{t}|)\right)+\sum_{i\in\mathcal{I}_{t}}\left(\frac{\hat{\sigma}_{i}(n)\sqrt{\ln\frac{1}{\delta}}+\hat{b}_{i}}{\alpha_{t}^{i}}+\|G^{i}_{n}(x_{t},\xi_{t})\|\left|\frac{1}{\bar{\alpha}_{t}^{i}}-\frac{1}{\alpha_{t}^{i}}\right|\right)$
$\displaystyle\leq\frac{\max_{i}\|G^{i}_{n}(x_{t},\xi_{t})\|}{\eta}\left(1+2(m-|\mathcal{I}_{t}|)\right)+\sum_{i\in\mathcal{I}_{t}}\left(\frac{\hat{\sigma}_{i}(n)\sqrt{\ln\frac{1}{\delta}}+\hat{b}_{i}}{\alpha_{t}^{i}}+\|G^{i}_{n}(x_{t},\xi_{t})\|\frac{\sigma_{i}(n)\sqrt{\ln\frac{1}{\delta}}}{\bar{\alpha}_{t}^{i}\alpha_{t}^{i}}\right)$
$\displaystyle\leq\frac{L}{\eta}\left(2m+1\right),$ (42)
for
$\hat{\sigma}_{i}(n)\leq\frac{\alpha_{t}^{i}\max\|G^{i}_{n}(x_{t},\xi_{t})\|}{2\eta\sqrt{\ln\frac{1}{\delta}}}$,
$\hat{b}_{i}\leq\frac{\alpha_{t}^{i}\max\|G^{i}_{n}(x_{t},\xi_{t})\|}{2\eta}$,
and
$\sigma_{i}(n)\leq\frac{(\alpha_{t}^{i})^{2}}{2\eta\sqrt{\ln\frac{1}{\delta}}}$,
implying $\bar{\alpha}_{t}^{i}\geq\alpha_{t}^{i}/2$ and using
$\|G^{i}_{n}(x_{t},\xi_{t})\|\leq L.$ Then, if
$\min\alpha_{t}^{i}\leq\bar{c}\eta$, we have
$\sum_{i\in\mathcal{I}_{t}}\frac{1}{\alpha_{t}^{i}}\geq\frac{1}{\bar{c}\eta}=\frac{L}{l\eta}\left(2m+1\right),$
and therefore with high probability $\|B\|\leq\|A\|$. Then we get (A.5), that
implies
$\displaystyle\prod_{i\in\mathcal{I}_{t}}\alpha^{i}_{t+1}\geq\prod_{i\in\mathcal{I}_{t}}\alpha^{i}_{t}.$
(43)
Moreover, using the same reasoning, we can prove that
$\displaystyle\prod_{i\in\mathcal{I}}\alpha^{i}_{t+1}\geq\prod_{i\in\mathcal{I}}\alpha^{i}_{t}.$
(44)
for any subset of indices $\mathcal{I}\subseteq[m]$ such that
$\mathcal{I}_{t}\subseteq\mathcal{I}.$
### A.6 Lower bound on $\gamma_{t}$
Here we assume $\underline{\alpha}_{t}^{i}\geq c\eta.$ Recall that
$\displaystyle\gamma_{t}=\min\left\\{\min_{i\in[m]}\left\\{\frac{\underline{\alpha}^{i}_{t}}{2|\hat{\theta}^{i}_{t}|+\sqrt{\underline{\alpha}_{t}^{i}M_{i}}}\right\\}\frac{1}{\|g_{t}\|},\frac{1}{\hat{M}_{2}(x_{t})}\right\\}.$
where
$\displaystyle\hat{M}_{2}(x_{t})=M_{0}+{\color[rgb]{0,0,0}10}\eta\sum_{i=1}^{m}\frac{M_{i}}{\underline{\alpha}^{i}_{t}}+{\color[rgb]{0,0,0}8}\eta\sum_{i=1}^{m}\frac{(\hat{\theta}^{i}_{t})^{2}}{(\underline{\alpha}^{i}_{t})^{2}}.$
We get the lower bound by constructing a bound on both of the terms inside the
minimum.
1) We have
$\mathbb{P}\left\\{\hat{M}_{2}(x_{t})\leq\left(1+{\color[rgb]{0,0,0}10}\frac{m}{c}\right)M+{\color[rgb]{0,0,0}8}\frac{mL^{2}}{\eta
c^{2}}\right\\}\geq 1-\delta$ (Due to Lemma 6, and by definition of
$\hat{M}_{2}(x_{t})$), which implies
$\mathbb{P}\left\\{\frac{1}{\hat{M}_{2}(x_{t})}\geq\eta\left(\frac{1}{\frac{{\color[rgb]{0,0,0}8}m}{c^{2}}L^{2}+\eta(1+{\color[rgb]{0,0,0}10}\frac{m}{c})M}\right)\right\\}\geq
1-\delta.$
2) Using Lemma 6 we get $\mathbb{P}\left\\{\|g_{t}\|\leq
L_{0}+\sum_{i=1}^{m}\frac{L_{i}}{c}\right\\}\geq 1-\delta.$ Hence, we can
bound
$\mathbb{P}\left\\{\min_{i\in[m]}\left\\{\frac{\underline{\alpha}^{i}_{t}}{2|\hat{\theta}^{i}_{t}|+\sqrt{\underline{\alpha}_{t}^{i}M_{i}}}\right\\}\frac{1}{\|g_{t}\|}\geq\frac{c\eta}{(2L+\sqrt{Mc\eta})L(1+\frac{m}{c})}\right\\}\geq
1-\delta.$
Therefore,
$\mathbb{P}\left\\{\gamma_{t}\geq\frac{\eta}{2}\min\left\\{\frac{1}{\frac{{\color[rgb]{0,0,0}4}m}{c^{2}}L^{2}+\eta({\color[rgb]{0,0,0}0.5}+{\color[rgb]{0,0,0}5}\frac{m}{c})M},\frac{1}{L^{2}(\frac{1}{c}+\frac{m}{c^{2}})+0.5\sqrt{\frac{M\eta}{cL^{2}}}L^{2}(1+\frac{m}{c})}\right\\}\right\\}\geq
1-\delta,$
$\mathbb{P}\left\\{\gamma_{t}\geq\frac{\eta}{2L^{2}(1+\frac{m}{c})}\min\left\\{\frac{1}{\frac{{\color[rgb]{0,0,0}4}}{c}+\frac{{\color[rgb]{0,0,0}5}M\eta}{L^{2}}},\frac{1}{\frac{1}{c}+\sqrt{\frac{M\eta}{4cL^{2}}}}\right\\}\right\\}\geq
1-\delta.$
$\mathbb{P}\left\\{\gamma_{t}\geq\frac{c\eta}{2L^{2}(1+\frac{m}{c})}\min\left\\{\frac{1}{{\color[rgb]{0,0,0}4}+\frac{{\color[rgb]{0,0,0}5}Mc\eta}{L^{2}}},\frac{1}{1+\sqrt{\frac{Mc\eta}{4L^{2}}}}\right\\}\right\\}\geq
1-\delta.$
Finally, the bound is
$\mathbb{P}\left\\{\gamma_{t}\geq\eta C\right\\}\geq 1-\delta.$
with
$C:=\frac{c\eta}{2L^{2}(1+\frac{m}{c})}\min\left\\{\frac{1}{{\color[rgb]{0,0,0}4}+\frac{{\color[rgb]{0,0,0}5}Mc\eta}{L^{2}}},\frac{1}{1+\sqrt{\frac{Mc\eta}{4L^{2}}}}\right\\}.$
### A.7 Proof of Lemma 9
Proof From Fact 2 it follows that
$\forall x\in\mathcal{X}\leavevmode\nobreak\ \exists
s_{x}=\frac{x-x_{0}}{\|x-x_{0}\|}\in\mathbb{R}^{d}:\leavevmode\nobreak\
\langle s_{x},\nabla f^{i}(x)\rangle\geq\frac{\beta}{2R}\leavevmode\nobreak\
\leavevmode\nobreak\ \leavevmode\nobreak\ \forall
i\in\mathcal{I}_{\beta/2}(x).$
Let $\hat{x}$ be an approximately optimal point for the log barrier:
$B_{\eta}(\hat{x})-B_{\eta}(x^{*}_{\eta})\leq\eta,$ that is equivalent to:
$f^{0}(\hat{x})+\eta\sum_{i=1}^{m}-\log(-f^{i}(\hat{x}))-f^{0}(x^{*}_{\eta})-\eta\sum_{i=1}^{m}-\log(-f^{i}(x^{*}_{\eta}))\leq\eta.$
Then, for the objective function we have the following bound:
$\displaystyle
f^{0}(\hat{x})-f^{0}(x^{*}_{\eta})\leq\eta+\eta\sum_{i=1}^{m}-\log\frac{-f^{i}(x^{*}_{\eta})}{-f^{i}(\hat{x})}.$
(45)
The optimal point for the log barrier $x^{*}_{\eta}$ must satisfy the
stationarity condition
$\nabla B_{\eta}(x^{*}_{\eta})=\nabla
f^{0}(x^{*}_{\eta})+\eta\sum_{i=1}^{m}\frac{\nabla
f^{i}(x^{*}_{\eta})}{-f^{i}(x^{*}_{\eta})}=0.$
By carefully rearranging the above, we obtain
$\sum_{i\in\mathcal{I}_{\beta/2}(x^{*}_{\eta})}\frac{\nabla
f^{i}(x^{*}_{\eta})}{-f^{i}(x^{*}_{\eta})}+\sum_{i\notin\mathcal{I}_{\beta/2}(x^{*}_{\eta})}\frac{\nabla
f^{i}(x^{*}_{\eta})}{-f^{i}(x^{*}_{\eta})}=\frac{-\nabla
f^{0}(x^{*}_{\eta})}{\eta}.$
By taking a dot product of both sides of the above equation with
$s_{x}=\frac{x^{*}_{\eta}-x_{0}}{\|x^{*}_{\eta}-x_{0}\|}$, using the Lipschitz
continuity we get for $x^{*}_{\eta}$:
$\displaystyle\frac{1}{\min_{i}\\{-f^{i}(x^{*}_{\eta})\\}}\sum_{i\in\mathcal{I}_{\beta/2}(x^{*}_{\eta})}\langle\nabla
f^{i}(x^{*}_{\eta}),s_{x}\rangle\frac{\min_{i}\\{-f^{i}(x^{*}_{\eta})\\}}{-f^{i}(x^{*}_{\eta})}$
(46) $\displaystyle=\frac{\langle-\nabla
f^{0}(x^{*}_{\eta}),s_{x}\rangle}{\eta}-\sum_{i\notin\mathcal{I}_{\beta/2}(x^{*}_{\eta})}\frac{\langle\nabla
f^{i}(x^{*}_{\eta}),s_{x}\rangle}{-f^{i}(x^{*}_{\eta})}\leq\frac{mL}{\eta}.$
(47)
From the above, using Fact 2, we get
$\min\\{-f^{i}(x^{*}_{\eta})\\}\geq\frac{\eta\beta}{2mLR}.$
Hence, combining the above with (45) we get the following relation of point
$\hat{x}$ and point $x^{*}_{\eta}$ optimal for the log barrier:
$\displaystyle
f^{0}(\hat{x})-f^{0}(x^{*}_{\eta})\leq\eta+\eta\sum_{i=1}^{m}\log\frac{-f^{i}(\hat{x})}{-f^{i}(x^{*}_{\eta})}\leq\eta\left(1+m\log\left(\frac{2mLR\hat{\beta}}{\eta\beta}\right)\right).$
(48)
Next, note that the Lagrangian $\mathcal{L}(x,\lambda)$ is a convex function
over $x$ and concave over $\lambda$. Hence, for
$(x^{*}_{\eta},\lambda^{*}_{\eta}):=\left(x^{*}_{\eta},\left[\frac{\eta}{-f^{1}(x^{*}_{\eta})},\ldots,\frac{\eta}{-f^{m}(x^{*}_{\eta})}\right]^{T}\right)$
we have
$\displaystyle\mathcal{L}(x^{*}_{\eta},\lambda^{*}_{\eta})-\mathcal{L}(x^{*},\lambda^{*})\leq\mathcal{L}(x^{*}_{\eta},\lambda^{*}_{\eta})-\mathcal{L}(x^{*},\lambda^{*}_{\eta})\leq\langle\nabla_{x}\mathcal{L}(x^{*}_{\eta},\lambda^{*}_{\eta}),\lambda^{*}_{\eta}-x^{*}\rangle\leq
0.$
Expressing $\mathcal{L}(x^{*}_{\eta},\lambda^{*}_{\eta})$ and
$\mathcal{L}(x^{*},\lambda^{*})$ and exploiting the fact that $\nabla
B_{\eta}(x^{*}_{\eta})=\nabla_{x}\mathcal{L}(x^{*}_{\eta},\lambda^{*}_{\eta})=0$,
we obtain
$\mathcal{L}(x^{*}_{\eta},\lambda^{*}_{\eta})-\mathcal{L}(x^{*},\lambda^{*})=f^{0}(x^{*}_{\eta})-f^{0}(x^{*})-m\eta\leq
0.$ Consequently, we have $f^{0}(x^{*}_{\eta})-f^{0}(x^{*})\leq m\eta.$
Combining the above and (48), we get
$f^{0}(\hat{x})-\min_{x\in\mathcal{X}}f^{0}(x)\leq\eta+\eta
m\log\left(\frac{2mLR\hat{\beta}}{\eta\beta}\right)+m\eta.$
### A.8 Zeroth-order estimator properties proof
The deviation of the gradient estimators $G^{i}(x_{t},\nu)-\nabla
f^{i}_{\nu}(x_{t})$, by definition can be expressed as follows for
$i=0,\ldots,m$
$\displaystyle G^{i}(x_{t},\nu)-\nabla
f^{i}_{\nu}(x_{t})=\frac{1}{n_{t}}\sum_{j=1}^{n_{t}}\left[\underbrace{\left(d\frac{f^{i}(x_{k}+\nu
s_{tj})-f^{i}(x_{t})}{\nu}s_{tj}-\nabla
f^{i}_{\nu}(x_{t})\right)}_{v_{j}^{i}}+\underbrace{d\frac{\xi_{tj}^{i+}-\xi_{tj}^{i-}}{\nu}s_{tj}}_{u_{j}^{i}}\right],$
(49)
where the first term under the summation $v_{j}^{i}$ is dependent only on
random $s_{tj}$, however the second term is dependent on both random variables
coming from the noise $\xi^{i\pm}_{tj}$ and from the direction $s_{tj}$.
Then, using the fact that the additive noise $\xi_{tj}^{i\pm}$ is zero-mean
and independent on $s_{tj}$, we get:
$\displaystyle\mathbb{E}\left\|G^{i}_{\nu,n}(x_{t},\xi)-\nabla
f^{i}_{\nu}(x_{t})\right\|^{2}=\mathbb{E}\left\|\frac{1}{n}\sum_{j=1}^{n}v_{j}^{i}\right\|^{2}+\mathbb{E}\left\|\frac{1}{n}\sum_{j=1}^{n}u_{j}^{i}\right\|^{2}$
(50)
Using the result of Lemma 2.10 (Berahas et al., 2021), we can bound the first
part of the above expression
$\mathbb{E}\left\|\frac{1}{n}\sum_{j=1}^{n}v_{j}^{i}\right\|^{2}$:
$\displaystyle\mathbb{E}\left\|\frac{1}{n}\sum_{j=1}^{n}v_{j}^{i}\right\|^{2}\leq\frac{3d^{2}}{n}\left(\frac{\|\nabla
f^{i}(x)\|^{2}}{d}+\frac{M_{i}^{2}\nu^{2}}{4}\right).\leavevmode\nobreak\
\forall i\in\\{0,\ldots,m\\}.$ (51)
The second part $u_{j}^{i}$ is zero-mean, hence does not influence the bias.
Indeed, using the independence of $\xi^{j\pm}_{tj}$ and $s_{tj}$ we derive
$\displaystyle\mathbb{E}\sum_{j=1}^{n_{t}}u_{j}^{i}=\frac{d}{\nu}\mathbb{E}\left(\sum_{j=1}^{n_{t}}(\xi^{i+}_{tj}-\xi^{i-}_{tj})s_{tj}\right)=0.$
(52)
Its variance can be bounded as follows, using $\|s_{tj}\|=1$:
$\displaystyle\mathbb{E}\left\|\frac{1}{n}\sum_{j=1}^{n}u_{j}^{i}\right\|^{2}=\mathbb{E}\frac{d^{2}}{\nu^{2}n^{2}}\left\|\sum_{j=1}^{n}(\xi^{i+}_{tj}-\xi^{i-}_{tj})s_{tj}\right\|^{2}\leq
4\frac{d^{2}}{\nu^{2}n^{2}}\sum_{j=1}^{n}\mathbb{E}\|\xi^{i+}_{tj}\|^{2}\|s_{tj}\|^{2}\leq
4\frac{d^{2}\sigma^{2}}{\nu^{2}n}.$ (53)
From the above, and Lemma 2.10 (Berahas et al., 2021) the statement of the
Lemma follows directly.
## References
* Achiam et al. (2017) Joshua Achiam, David Held, Aviv Tamar, and Pieter Abbeel. Constrained policy optimization, 2017.
* Altman (1999) E. Altman. _Constrained Markov Decision Processes_. Chapman and Hall, 1999.
* Amani et al. (2019) Sanae Amani, Mahnoosh Alizadeh, and Christos Thrampoulidis. Linear stochastic bandits under safety constraints. _arXiv preprint arXiv:1908.05814_ , 2019.
* Arjevani et al. (2019) Yossi Arjevani, Yair Carmon, John C Duchi, Dylan J Foster, Nathan Srebro, and Blake Woodworth. Lower bounds for non-convex stochastic optimization. _arXiv preprint arXiv:1912.02365_ , 2019.
* As et al. (2022) Yarden As, Ilnura Usmanova, Sebastian Curi, and Andreas Krause. Constrained policy optimization via bayesian world models. _ArXiv_ , 2022. URL https://arxiv.org/abs/2201.09802.
* Bach and Perchet (2016) Francis Bach and Vianney Perchet. Highly-smooth zero-th order online optimization. In _Conference on Learning Theory_ , pages 257–283, 2016.
* Balasubramanian and Ghadimi (2018) Krishnakumar Balasubramanian and Saeed Ghadimi. Zeroth-order (non)-convex stochastic optimization via conditional gradient and gradient updates. In _Advances in Neural Information Processing Systems_ , pages 3455–3464, 2018.
* Berahas et al. (2021) Albert Berahas, Liyuan Cao, Krzysztof Choromanski, and Katya Scheinberg. A theoretical and empirical comparison of gradient approximations in derivative-free optimization. _Foundations of Computational Mathematics_ , 05 2021. doi: 10.1007/s10208-021-09513-z.
* Berkenkamp et al. (2016a) Felix Berkenkamp, Andreas Krause, and Angela P Schoellig. Bayesian optimization with safety constraints: safe and automatic parameter tuning in robotics. _arXiv preprint arXiv:1602.04450_ , 2016a.
* Berkenkamp et al. (2016b) Felix Berkenkamp, Angela P. Schoellig, and Andreas Krause. Safe controller optimization for quadrotors with gaussian processes. In _2016 IEEE International Conference on Robotics and Automation (ICRA)_ , pages 491–496, 2016b. doi: 10.1109/ICRA.2016.7487170.
* Berkenkamp et al. (2017) Felix Berkenkamp, Matteo Turchetta, Angela P. Schoellig, and Andreas Krause. Safe model-based reinforcement learning with stability guarantees, 2017\.
* Berkenkamp et al. (2020) Felix Berkenkamp, Andreas Krause, and Angela P. Schoellig. Bayesian optimization with safety constraints: Safe and automatic parameter tuning in robotics, 2020.
* Bubeck et al. (2017) Sébastien Bubeck, Yin Tat Lee, and Ronen Eldan. Kernel-based methods for bandit convex optimization. In _Proceedings of the 49th Annual ACM SIGACT Symposium on Theory of Computing_ , pages 72–85, 2017.
* Chen et al. (2019) Lin Chen, Mingrui Zhang, and Amin Karbasi. Projection-free bandit convex optimization. In _The 22nd International Conference on Artificial Intelligence and Statistics_ , pages 2047–2056. PMLR, 2019.
* Chow et al. (2015) Yinlam Chow, Mohammad Ghavamzadeh, Lucas Janson, and Marco Pavone. Risk-constrained reinforcement learning with percentile risk criteria. _CoRR_ , abs/1512.01629, 2015. URL http://arxiv.org/abs/1512.01629.
* Chua et al. (2018) Kurtland Chua, Roberto Calandra, Rowan McAllister, and Sergey Levine. Deep reinforcement learning in a handful of trials using probabilistic dynamics models. _CoRR_ , abs/1805.12114, 2018. URL http://arxiv.org/abs/1805.12114.
* Clevert et al. (2015) Djork-Arné Clevert, Thomas Unterthiner, and Sepp Hochreiter. Fast and accurate deep network learning by exponential linear units (elus), 2015. URL https://arxiv.org/abs/1511.07289.
* Dalal et al. (2018) Gal Dalal, Krishnamurthy Dvijotham, Matej Vecerik, Todd Hester, Cosmin Paduraru, and Yuval Tassa. Safe exploration in continuous action spaces, 2018.
* Deisenroth and Rasmussen (2011) Marc Peter Deisenroth and Carl Edward Rasmussen. Pilco: A model-based and data-efficient approach to policy search. In _Proceedings of the 28th International Conference on International Conference on Machine Learning_ , ICML’11, page 465–472, Madison, WI, USA, 2011. Omnipress. ISBN 9781450306195.
* Duchi et al. (2015) John C. Duchi, Michael I. Jordan, Martin J. Wainwright, and Andre Wibisono. Optimal rates for zero-order convex optimization: The power of two function evaluations. _IEEE Transactions on Information Theory_ , 61(5):2788–2806, 2015. doi: 10.1109/TIT.2015.2409256.
* Díaz-Francés and Rubio (2013) Eloísa Díaz-Francés and Francisco Rubio. On the existence of a normal approximation to the distribution of the ratio of two independent normal random variables. _Statistical Papers_ , 54(2):309–323, May 2013\. doi: 10.1007/s00362-012-0429-2. URL https://ideas.repec.org/a/spr/stpapr/v54y2013i2p309-323.html.
* Eriksson and Jankowiak (2021) David Eriksson and Martin Jankowiak. High-dimensional bayesian optimization with sparse axis-aligned subspaces, 2021. URL https://arxiv.org/abs/2103.00349.
* Fazlyab et al. (2019) Mahyar Fazlyab, Alexander Robey, Hamed Hassani, Manfred Morari, and George Pappas. Efficient and accurate estimation of lipschitz constants for deep neural networks. _Advances in Neural Information Processing Systems_ , 32, 2019.
* Fereydounian et al. (2020) Mohammad Fereydounian, Zebang Shen, Aryan Mokhtari, Amin Karbasi, and Hamed Hassani. Safe learning under uncertain objectives and constraints. _arXiv preprint arXiv:2006.13326_ , 2020.
* Flaxman et al. (2005) Abraham D Flaxman, Adam Tauman Kalai, and H Brendan McMahan. Online convex optimization in the bandit setting: gradient descent without a gradient. In _Proceedings of the sixteenth annual ACM-SIAM symposium on Discrete algorithms_ , pages 385–394. Society for Industrial and Applied Mathematics, 2005.
* Frazier (2018) Peter I. Frazier. A tutorial on bayesian optimization, 2018. URL https://arxiv.org/abs/1807.02811.
* Garber and Kretzu (2020) Dan Garber and Ben Kretzu. Improved regret bounds for projection-free bandit convex optimization. In _International Conference on Artificial Intelligence and Statistics_ , pages 2196–2206. PMLR, 2020.
* Hafner et al. (2021) Danijar Hafner, Timothy Lillicrap, Mohammad Norouzi, and Jimmy Ba. Mastering atari with discrete world models, 2021.
* Hansen and Ostermeier (2001) Nikolaus Hansen and Andreas Ostermeier. Completely derandomized self-adaptation in evolution strategies. _Evol. Comput._ , 9(2):159–195, 2001. URL http://dblp.uni-trier.de/db/journals/ec/ec9.html#HansenO01.
* Hazan and Luo (2016) Elad Hazan and Haipeng Luo. Variance-reduced and projection-free stochastic optimization. In _International Conference on Machine Learning_ , pages 1263–1271, 2016.
* Hinder and Ye (2018) Oliver Hinder and Yinyu Ye. A one-phase interior point method for nonconvex optimization. _arXiv preprint arXiv:1801.03072_ , 2018.
* Hinder and Ye (2019) Oliver Hinder and Yinyu Ye. A polynomial time log barrier method for problems with nonconvex constraints. _arXiv preprint: https://arxiv.org/pdf/1807.00404.pdf_ , 2019.
* Juditsky et al. (2013) Anatoli B. Juditsky, Guanghui Lan, Arkadii S. Nemirovski, and Alexander Shapiro. Stochastic Approximation approach to Stochastic Programming. Research report, LJK, 2013. URL https://hal.archives-ouvertes.fr/hal-00853911. http://www.optimization-online.org/DB_HTML/2007/09/1787.html.
* Kennedy and Eberhart (1995) James Kennedy and Russell C. Eberhart. Particle swarm optimization. In _Proceedings of the IEEE International Conference on Neural Networks_ , pages 1942–1948, 1995.
* Kirschner et al. (2019) Johannes Kirschner, Mojmir Mutnỳ, Nicole Hiller, Rasmus Ischebeck, and Andreas Krause. Adaptive and safe bayesian optimization in high dimensions via one-dimensional subspaces. _arXiv preprint arXiv:1902.03229_ , 2019.
* Koller et al. (2018) Torsten Koller, Felix Berkenkamp, Matteo Turchetta, and Andreas Krause. Learning-based model predictive control for safe exploration. In _2018 IEEE Conference on Decision and Control (CDC)_ , pages 6059–6066. IEEE, 2018.
* Lan (2020) Guanghui Lan. _First-order and Stochastic Optimization Methods for Machine Learning_. 01 2020. ISBN 978-3-030-39567-4. doi: 10.1007/978-3-030-39568-1.
* Lillicrap et al. (2015) Timothy P. Lillicrap, Jonathan J. Hunt, Alexander Pritzel, Nicolas Heess, Tom Erez, Yuval Tassa, David Silver, and Daan Wierstra. Continuous control with deep reinforcement learning, 2015. URL https://arxiv.org/abs/1509.02971.
* Luersen et al. (2004) M. Luersen, R. Le Riche, and F. A Guyon. Constrained, globalized, and bounded nelder–mead method for engineering optimization. _Struct Multidisc Optim 27_ , page 43–54, 2004. doi: 10.1007/s00158-003-0320-9.
* Mangasarian and Fromovitz (1967) O.L Mangasarian and S Fromovitz. The fritz john necessary optimality conditions in the presence of equality and inequality constraints. _Journal of Mathematical Analysis and Applications_ , 17(1):37–47, 1967. ISSN 0022-247X. doi: https://doi.org/10.1016/0022-247X(67)90163-1. URL https://www.sciencedirect.com/science/article/pii/0022247X67901631.
* Moriconi et al. (2019) Riccardo Moriconi, Marc P. Deisenroth, and K. S. Sesh Kumar. High-dimensional bayesian optimization using low-dimensional feature spaces, 2019. URL https://arxiv.org/abs/1902.10675.
* Nelder and Mead (1965) J. A. Nelder and R. Mead. A Simplex Method for Function Minimization. _The Computer Journal_ , 7(4):308–313, 01 1965\. ISSN 0010-4620. doi: 10.1093/comjnl/7.4.308. URL https://doi.org/10.1093/comjnl/7.4.308.
* Nemirovsky and Yudin (1985) A. S. Nemirovsky and D. B. Yudin. Problem complexity and method efficiency in optimization. _SIAM Review_ , 27(2):264–265, 1985. doi: 10.1137/1027074. URL https://doi.org/10.1137/1027074.
* Nocedal and Wright (2006) Jorge Nocedal and Stephen Wright. _Numerical optimization_. Springer Science & Business Media, 2006.
* Price (2019) Christopher Price. A modified nelder-mead barrier method for constrained optimization. _Numerical Algebra, Control, and Optimization_ , 11, 01 2019. doi: 10.3934/naco.2020058.
* Rasmussen and Williams (2005) Carl Edward Rasmussen and Christopher K. I. Williams. _Gaussian Processes for Machine Learning (Adaptive Computation and Machine Learning)_. The MIT Press, 2005. ISBN 026218253X.
* Ray et al. (2019) Alex Ray, Joshua Achiam, and Dario Amodei. Benchmarking Safe Exploration in Deep Reinforcement Learning. 2019\.
* Rechenberg (1989) Ingo Rechenberg. Evolution strategy: Nature’s way of optimization. In H. W. Bergmann, editor, _Optimization: Methods and Applications, Possibilities and Limitations_ , pages 106–126, Berlin, Heidelberg, 1989. Springer Berlin Heidelberg. ISBN 978-3-642-83814-9.
* Schulman et al. (2015) John Schulman, Philipp Moritz, Sergey Levine, Michael Jordan, and Pieter Abbeel. High-dimensional continuous control using generalized advantage estimation, 2015. URL https://arxiv.org/abs/1506.02438.
* Shamir (2013) Ohad Shamir. On the complexity of bandit and derivative-free stochastic convex optimization. In _Conference on Learning Theory_ , pages 3–24, 2013.
* Shapiro et al. (2009) Alexander Shapiro, Darinka Dentcheva, and Andrzej Ruszczyński. _Lectures on Stochastic Programming_. SIAM, Philadelphia, PA, USA, 2009.
* Snoek et al. (2015) Jasper Snoek, Oren Rippel, Kevin Swersky, Ryan Kiros, Nadathur Satish, Narayanan Sundaram, Md. Mostofa Ali Patwary, Prabhat, and Ryan P. Adams. Scalable bayesian optimization using deep neural networks, 2015. URL https://arxiv.org/abs/1502.05700.
* Srinivas et al. (2012) Niranjan Srinivas, Andreas Krause, Sham Kakade, and Matthias Seeger. Information-theoretic regret bounds for gaussian process optimization in the bandit setting. _IEEE Transactions on Information Theory_ , 58(5):3250–3265, May 2012. doi: 10.1109/TIT.2011.2182033.
* Storn and Price (1997) Rainer Storn and Kenneth V. Price. Differential evolution - a simple and efficient heuristic for global optimization over continuous spaces. _J. Glob. Optim._ , 11(4):341–359, 1997. URL http://dblp.uni-trier.de/db/journals/jgo/jgo11.html#StornP97.
* Sui et al. (2015a) Yanan Sui, Alkis Gotovos, Joel Burdick, and Andreas Krause. Safe exploration for optimization with gaussian processes. In Francis Bach and David Blei, editors, _Proceedings of the 32nd International Conference on Machine Learning_ , volume 37 of _Proceedings of Machine Learning Research_ , pages 997–1005, Lille, France, 07–09 Jul 2015a. PMLR. URL https://proceedings.mlr.press/v37/sui15.html.
* Sui et al. (2015b) Yanan Sui, Alkis Gotovos, Joel Burdick, and Andreas Krause. Safe exploration for optimization with gaussian processes. In _International Conference on Machine Learning_ , pages 997–1005, 2015b.
* Sutton et al. (2000) Richard S Sutton, David McAllester, Satinder Singh, and Yishay Mansour. Policy gradient methods for reinforcement learning with function approximation. In S. Solla, T. Leen, and K. Müller, editors, _Advances in Neural Information Processing Systems_ , volume 12. MIT Press, 2000. URL https://proceedings.neurips.cc/paper/1999/file/464d828b85b0bed98e80ade0a5c43b0f-Paper.pdf.
* Usmanova et al. (2019) Ilnura Usmanova, Andreas Krause, and Maryam Kamgarpour. Safe convex learning under uncertain constraints. In _The 22nd International Conference on Artificial Intelligence and Statistics_ , pages 2106–2114, 2019.
* Usmanova et al. (2020) Ilnura Usmanova, Andreas Krause, and Maryam Kamgarpour. Safe non-smooth black-box optimization with application to policy search. In _Learning for Dynamics and Control_ , pages 980–989, 2020.
* Vaswani et al. (2021) Sharan Vaswani, Benjamin Dubois-Taine, and Reza Babanezhad. Towards Noise-adaptive, Problem-adaptive Stochastic Gradient Descent. working paper or preprint, November 2021. URL https://hal.archives-ouvertes.fr/hal-03456663.
* Wabersich and Zeilinger (2021) Kim P. Wabersich and Melanie N. Zeilinger. A predictive safety filter for learning-based control of constrained nonlinear dynamical systems, 2021.
|
# Self-Translate-Train: A Simple but Strong Baseline
for Cross-lingual Transfer of Large Language Models
Ryokan Ri Shun Kiyono Sho Takase
SB Intuitions
<EMAIL_ADDRESS>
###### Abstract
Cross-lingual transfer is a promising technique for utilizing data in a source
language to improve performance in a target language. However, current
techniques often require an external translation system or suffer from
suboptimal performance due to over-reliance on cross-lingual generalization of
multi-lingual pretrained language models. In this study, we propose a simple
yet effective method called Self-Translate-Train. It leverages the translation
capability of a large language model to generate synthetic training data in
the target language and fine-tunes the model with its own generated data. We
evaluate the proposed method on a wide range of tasks and show substantial
performance gains across several non-English languages.
Self-Translate-Train: A Simple but Strong Baseline
for Cross-lingual Transfer of Large Language Models
Ryokan Ri and Shun Kiyono and Sho Takase SB Intuitions
<EMAIL_ADDRESS>
## 1 Introduction
Cross-lingual transfer is a technique to solve tasks in a target language by
leveraging training data from the other languages (Pikuliak et al., 2021).
This has been increasingly feasible with the rise of multilingual pre-trained
models, which are trained on multilingual corpora and capture commonalities
across languages (Conneau et al., 2020; Xue et al., 2021; Scao et al., 2022).
Capable multilingual models can perform tasks in a target language without
being trained on task-specific data in that language, which is known as zero-
shot cross-lingual transfer (Artetxe and Schwenk, 2019; Chen et al., 2021).
This technique is expected to reduce the disparity between high-resource and
low-resource languages.
Cross-lingual transfer is also exhibited in large language models (LLMs),
which refers to auto-regressive language models with billion-scale parameters
that are trained with a massive amount of text data (Brown et al., 2020;
Touvron et al., 2023). A common approach for zero-shot cross-lingual transfer
is fine-tuning the model with supervised training data available in a source
languages, mostly English, and then applying the model to the target language
(Chen et al., 2024; Shaham et al., 2024). However, we argue that this approach
does not fully elicit the model’s cross-lingual capability as the model has no
clue the input language at the test time. To achieve better cross-lingual
performance, we let the model teach itself how to solve the task in the target
language. In our proposed method, Self-Translate-Train, we produce the target
language’s translation of the training data leveraging the strong capability
of the LLM to generate text, and train the LLM with its own generated
translation.
Figure 1: An overview of Self-Translate-Train. An LLM translates training data
to the target language and then fine-tuned on its own generated data.
We evaluate Self-Translate-Train with several tasks including question
answering, text-pair classification, and mathematical reasoning across
multiple languages. Our experiments show that Self-Translate-Train
consistently improves the performance of baselines given a multilingual
capability of the LLMs. Our results indicate that we can achieve better cross-
lingual performance by correctly elicit the model’s translation capability,
which encourages further exploration of how to better utilize the model’s
cross-lingual capability.
## 2 Related Work
### 2.1 Cross-lingual Transfer Learning
There are two main approaches to transfer task knowledge across languages:
data transfer and model transfer (Pikuliak et al., 2021).
Data transfer translates the source language data to the target language. In
the Translate-test approach, models are trained on source language data and at
inference time, the task inputs are translated into the source language
(Conneau et al., 2018; Asai et al., 2018). Although the training stage is
simple, it incurs additional translation costs at inference time. The
Translate-train approach, on the other hand, translates the training data and
the resulting model is used to predict the target language data directly
(Conneau et al., 2018). Data transfer is quite effective in terms of
performance (Hu et al., 2020), but one drawback is its requirement of
additional translation systems.
Model transfer alleviates the need for translation systems by using
multilingual pretrained models, which are trained on a large amount of data
from multiple languages and capture the commonality between languages. These
models can be fine-tuned on task-specific data in a single source language and
generalize to solve the task in other languages (Pires et al., 2019; Mulcaire
et al., 2019; Conneau et al., 2020), eliminating the need for translation
systems.
Our approach, Self-Translate-Train, leverages LLMs’ translation and cross-
lingual generalization capabilities. It combines the advantages of data
transfer and model transfer by using explicit training signals in the target
language while eliminating the need for external translation systems.
### 2.2 Self-Improvement of LLMs
LLMs have demonstrated remarkable text generation capabilities, which has been
leveraged to generate training data for various purposes (Li et al., 2023b;
Lee et al., 2024). The generated data can be used to further specialize the
LLM itself for downstream applications, without requiring an extensive
collection of additional data. This process can be viewed as a form of self-
improvement (Bai et al., 2022; Huang et al., 2023; Sun et al., 2023; Li et
al., 2023a).
Self-Translate-Train is also a self-improvement approach to specialize the LLM
to a target language by translating the source language data to the target
language.
## 3 Self-Translate-Train
Our framework focuses on fine-tuning LLMs on a small amount of data for a
specific task. Let the training corpus in a source language, say English, be
$\mathcal{D}_{\text{src}}=\\{(\mathbf{x}_{\text{src}}^{i},\mathbf{y}_{\text{src}}^{i})\\}_{i=1}^{N}$,
where $\mathbf{x}$ is the input and $\mathbf{y}$ is the output. In a typical
cross-lingual transfer setting, the model is fine-tuned only on
$\mathcal{D}_{\text{src}}$ and expected to generalize to target languages.
### Translated Synthetic Data
Given the LLM’s translation capability, we can let it translate the training
corpus into a synthetic corpus in the target language
$\mathcal{D}_{\text{tgt}}$. The synthetic data can be added to
$\mathcal{D}_{\text{src}}$ to achieve a better generalization to the target
language.
The translation can be performed in various ways depending on the model’s
capabilities or available resources. In this paper, we experiment with the
few-shot prompting technique (Section 4.4).
### Code-switched Synthetic Data
The generated data has an interesting aspect: each synthetic instance has a
corresponding instance in the original dataset with the same semantics. We can
exploit this to further synthesize data by generating code-switched instances
where the input and output are in different languages.
We pair the original and translated instances to construct
$\mathcal{D}_{\text{cs}}=\\{(\mathbf{x}_{\text{src}}^{i},\mathbf{y}_{\text{tgt}}^{i})\\}_{i=1}^{N}\bigcup\\{(\mathbf{x}_{\text{tgt}}^{i},\mathbf{y}_{\text{src}}^{i})\\}_{i=1}^{N}$.
When the task output is natural language, we manually translate the prompt
“Please answer in {{ tgt }}.” into the target language, and add it to the
input $\mathbf{x}$.
## 4 Experimental Setups
To verify the effectiveness of Self-Translate-Train, we conduct extensive
experiments on multiple tasks and languages.
### 4.1 Task and Datasets
We present a list of datasets for experiments in Table 1. For each task, an
English dataset is used for training and a multilingual dataset for
evaluation. To make the computational cost feasible, we use a 10,000-sample
subset of the training data for SQuAD and MultiNLI.
Task | Training | Evaluation
---|---|---
QA | SQuAD (Rajpurkar et al., 2016) | XQuAD (Artetxe et al., 2020)
Classification | MultiNLI (Williams et al., 2018) | XNLI (Conneau et al., 2018)
Math | GSM8k (Cobbe et al., 2021) | MGSM (Shi et al., 2023)
Table 1: List of datasets for experiments. The details are described in
Section A.1.
### 4.2 Languages
We conducted evaluation on four languages: German (de), Russian (ru), Thai
(th), and Chinese (zh). German is a Germanic language, which is
phylogenetically close to English and expected to show better cross-lingual
transfer, while Russian, Thai, and Chinese are from different language
families. In particular, Thai is a low-resource language with a different
script from English, which is expected to be more challenging for cross-
lingual transfer.
### 4.3 Language Models
Our main experiments use Llama2-7B (Touvron et al., 2023), a public LLM.
Although 90% of its pretraining corpus is English, the model has a
multilingual capability (e.g., Table 2) from the remaining fraction of
multilingual data.
### 4.4 Synthetic Data Generation
Recent LLMs are known to exhibit a translation capability without much task-
specific data Briakou et al. (2023). In our experiments, we elicit the
translation capability of the LLMs via few-shot in-context learning (Brown et
al., 2020).
To construct few-shot translation samples, we sample eight pairs from the
train or validation splits of the multilingual datasets, where instances
across languages form parallel data. The translation was performed for each
field individually, e.g., for GSM8k, we translated the question and answer
separately. The prompt template simply alternates the source and target text
prepended with the language tag (Section A.2).
An important step to ensure the quality of the synthetic data is to filter out
the low-quality data (the details in Section A.3). To remove under- or over-
translation (Tu et al., 2016), we filter out texts with an extreme source-
target length ratio. Also, to address the repetition problem (Holtzman et al.,
2020), we set the max number of tokens for generation and filter out the
translation that does not end with the EOS token. With the translations from
Llama2-7B, this process removes around 10% of the data for most languages and
around 50% for Thai due to the model’s limited generation quality.
To provide the sense of the translation quality, we report the BLEU score
(Papineni et al., 2002) measured by the parallel data constructed from
questions in the MGSM test set in Table 2. Overall, the translation quality is
sufficiently high except for Thai. As we will see in Section 5.1, this poses a
challenge for cross-lingual transfer to Thai.
Model | de | ru | th | zh
---|---|---|---|---
Llama2-7B | 37.1 | 27.2 | 1.9 | 29.4
Table 2: BLEU scores from the MGSM test set. The configuration of BLEU is
described in Section A.4.
### 4.5 Fine-tuning
All the tasks are cast as text generation tasks, where the LLM is given the
inputs as a prompt and generate the answer. Fine-tuning is conducted with
causal language modeling loss, computed only for output tokens. We use LoRA
(Hu et al., 2022), a parameter-efficient tuning technique, to reduce
computational cost.
We use AdamW (Loshchilov and Hutter, 2019) and the cosine learning rate
schedule for optimization, training with a batch size of 64 for 1,000 steps.
For each setting, we conduct six runs with two learning rates (5e-5 and 3e-4)
and different random seeds, reporting summarization statistics of the top four
runs based on validation accuracy to remove runs with optimization failure.
See Section A.5 for other hyperparameters.
## 5 Results
### 5.1 Main Results
| MGSM | XQuAD | XNLI
---|---|---|---
| de | ru | th | zh | de | ru | th | zh | de | ru | th | zh
$\mathcal{D}_{\text{src}}$ | | 30.1
---
$\pm$0.4
| 25.0
---
$\pm$0.7
| 8.1
---
$\pm$0.7
| 21.1
---
$\pm$1.7
| 60.3
---
$\pm$0.8
| 49.3
---
$\pm$0.4
| 34.5
---
$\pm$1.0
| 66.3
---
$\pm$0.8
| 79.7
---
$\pm$0.4
| 76.9
---
$\pm$0.1
| 53.7
---
$\pm$0.9
| 74.1
---
$\pm$0.2
$+\mathcal{D}_{\text{tgt}}$ | * | 36.4
---
$\pm$1.3
* * | 34.0
---
$\pm$1.7
* | 7.7
---
$\pm$3.4
* | 27.1
---
$\pm$1.2
* | 61.7
---
$\pm$0.9
* | 57.8
---
$\pm$0.8
* * | 46.4
---
$\pm$1.3
* * | 77.7
---
$\pm$0.4
* * | 81.6
---
$\pm$0.7
* * | 78.5
---
$\pm$0.6
* | 56.3
---
$\pm$1.5
* | 78.5
---
$\pm$0.3
*
$+\mathcal{D}_{\text{tgt}}+\mathcal{D}_{\text{cs}}$ | * | 35.9
---
$\pm$1.3
* * | 34.5
---
$\pm$2.4
* | 10.4
---
$\pm$1.6
* | 28.8
---
$\pm$1.9
* * | 62.2
---
$\pm$0.9
* * | 58.0
---
$\pm$0.6
* * | 46.2
---
$\pm$1.7
* * | 77.3
---
$\pm$0.7
* * | 81.6
---
$\pm$0.5
* | 78.4
---
$\pm$1.2
* | 58.9
---
$\pm$1.5
* * | 77.6
---
$\pm$0.6
*
Table 3: Results on multilingual evaluation datasets. Scores are marked with ∗
if its improvement is statistically significant ($p<0.05$ in Welch’s t-test)
compared to the baseline $\mathcal{D}_{\text{src}}$. The significant and
highest score in each column is marked in bold.
(a) de
(b) th
(c) zh
Figure 2: Accuracy in the MGSM dataset with different model sizes of Llama2.
As the baseline, we fine-tune the LLM with the source language dataset
$\mathcal{D}_{\text{src}}$. To ensure a fair comparison, we augment
$\mathcal{D}_{\text{src}}$ with the eight target language samples used for
few-shot translation (Section 4.4). We then compare the baseline with the
models fine-tuned on the data generated from Self-Translate-Train
($\mathcal{D}_{\text{src}}$ and $\mathcal{D}_{\text{cs}}$) in Table 3.
First, Self-Translate-Train is indeed an effective method;
$+\mathcal{D}_{\text{tgt}}$ almost consistently outperforms the baseline
$\mathcal{D}_{\text{src}}$. The only exception is Thai (th), where there is no
significant improvement. This is likely due to the low translation quality of
the model in Thai (Table 2).
The effectiveness of code-switching dataset is limited. When we add
$\mathcal{D}_{\text{cs}}$ to $\mathcal{D}_{\text{tgt}}$, there is no
significant improvement from adding $\mathcal{D}_{\text{tgt}}$ only ($p<0.05$
in Welch’s t-test). This indicates that the code-switching data does not
provide additional information for the model to generalize in the task.
### 5.2 Does the model size matter?
The size of the language model can influence both its ability to generalize
across languages and the quality of its translations, which in turn may impact
the effectiveness of Self-Translate-Train. We compare the performance of
Llama2 with different sizes, i.e., 7B, 13B, and 65B, on the math task in de,
th, and zh (Figure 2).
The larger model size generally tends to perform better, and the improvement
from Self-Translate-Train remains consistent across different model sizes. In
Thai (th), we did not observe a significant improvement in the 7B model, but
do in the larger models (13B and 65B), likely due to their better translation
quality. The 7B model has a low Thai translation BLEU score of 1.9 (Table 2),
while the 13B and 65B models have BLEU scores of 5.1 and 12.0, respectively.
The improvement in Thai (th) with the 70B model is the most significant (+19.8
average points). This implies that Self-Translate-Train is particularly
effective when the model struggles with generalizing across the source and
target languages but can still generate reasonable translations.
## 6 Conclusion
We introduced Self-Translate-Train, a method to improve cross-lingual transfer
performance by generating synthetic training data in the target language. We
validated its effectiveness on various tasks and languages, demonstrating
substantial performance gains across several non-English languages. Self-
Translate-Train is effective when the zero-shot cross-lingual transfer
performance is suboptimal and the model can generate reasonable translations.
Self-Translate-Train neither requires external translation systems nor
intensive additional data collection, making it a simple yet effective method
for cross-lingual transfer. We encourage practitioners to try this approach as
an improved baseline for cross-lingual transfer of LLMs.
Our research also shows that relying solely on the model’s generalization
capability may be suboptimal, and there is a better way to elicit the cross-
lingual capability of the model. We hope this work encourages further
exploration of how to better utilize the model’s cross-lingual capability.
## 7 Limitations
Our experiments are conducted on a modern type of LLM, an autoregressive
Transformer decoder, and centered around the Llama2 model families (Touvron et
al., 2023). Although we further validate our method in Appendix B, the
effective of our proposed method is uncertain when applied to other types of
LLMs developed in the future.
Our method is based on the assumption that the model can generate reasonable
translations in the target language. This may be challenging when the task
inputs are long or complex. One solution is to split the input into smaller
segments, as we have done with the SQuAD dataset (Section A.2).
Finally, when the task requires generating long and natural text, the quality
of the generated translation matters more. If the translation quality is low,
the model outputs may degrade due to translation errors or unnaturalness. The
application of our method on more challenging tasks requires further
investigation.
## References
* Artetxe et al. (2020) Mikel Artetxe, Sebastian Ruder, and Dani Yogatama. 2020. On the cross-lingual transferability of monolingual representations. In _Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics_ , pages 4623–4637, Online. Association for Computational Linguistics.
* Artetxe and Schwenk (2019) Mikel Artetxe and Holger Schwenk. 2019. Massively multilingual sentence embeddings for zero-shot cross-lingual transfer and beyond. _Transactions of the Association for Computational Linguistics_ , 7:597–610.
* Asai et al. (2018) Akari Asai, Akiko Eriguchi, Kazuma Hashimoto, and Yoshimasa Tsuruoka. 2018. Multilingual extractive reading comprehension by runtime machine translation. _ArXiv_ , abs/1809.03275.
* Bai et al. (2022) Yuntao Bai, Saurav Kadavath, Sandipan Kundu, Amanda Askell, John Kernion, Andy Jones, Anna Chen, Anna Goldie, Azalia Mirhoseini, Cameron McKinnon, Carol Chen, Catherine Olsson, Christopher Olah, Danny Hernandez, Dawn Drain, Deep Ganguli, Dustin Li, Eli Tran-Johnson, E Perez, Jamie Kerr, Jared Mueller, Jeff Ladish, J Landau, Kamal Ndousse, Kamilė Lukošiūtė, Liane Lovitt, Michael Sellitto, Nelson Elhage, Nicholas Schiefer, Noem’i Mercado, Nova Dassarma, Robert Lasenby, Robin Larson, Sam Ringer, Scott Johnston, Shauna Kravec, Sheer El Showk, Stanislav Fort, Tamera Lanham, Timothy Telleen-Lawton, Tom Conerly, Tom Henighan, Tristan Hume, Sam Bowman, Zac Hatfield-Dodds, Benjamin Mann, Dario Amodei, Nicholas Joseph, Sam McCandlish, Tom B. Brown, and Jared Kaplan. 2022. Constitutional ai: Harmlessness from ai feedback. _ArXiv_ , abs/2212.08073.
* Briakou et al. (2023) Eleftheria Briakou, Colin Cherry, and George Foster. 2023. Searching for needles in a haystack: On the role of incidental bilingualism in PaLM’s translation capability. In _Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)_ , pages 9432–9452, Toronto, Canada. Association for Computational Linguistics.
* Brown et al. (2020) Tom Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared D Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, Sandhini Agarwal, Ariel Herbert-Voss, Gretchen Krueger, Tom Henighan, Rewon Child, Aditya Ramesh, Daniel Ziegler, Jeffrey Wu, Clemens Winter, Chris Hesse, Mark Chen, Eric Sigler, Mateusz Litwin, Scott Gray, Benjamin Chess, Jack Clark, Christopher Berner, Sam McCandlish, Alec Radford, Ilya Sutskever, and Dario Amodei. 2020. Language Models are Few-Shot Learners. In _Advances in Neural Information Processing Systems_ , volume 33, pages 1877–1901. Curran Associates, Inc.
* Chen et al. (2021) Guanhua Chen, Shuming Ma, Yun Chen, Li Dong, Dongdong Zhang, Jia Pan, Wenping Wang, and Furu Wei. 2021. Zero-shot cross-lingual transfer of neural machine translation with multilingual pretrained encoders. In _Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing_ , pages 15–26, Online and Punta Cana, Dominican Republic. Association for Computational Linguistics.
* Chen et al. (2024) Pinzhen Chen, Shaoxiong Ji, Nikolay Bogoychev, Andrey Kutuzov, Barry Haddow, and Kenneth Heafield. 2024. Monolingual or multilingual instruction tuning: Which makes a better alpaca. In _Findings of the Association for Computational Linguistics: EACL 2024_ , pages 1347–1356, St. Julian’s, Malta. Association for Computational Linguistics.
* Chen et al. (2023) Yang Chen, Chao Jiang, Alan Ritter, and Wei Xu. 2023. Frustratingly easy label projection for cross-lingual transfer. In _Findings of the Association for Computational Linguistics: ACL 2023_ , pages 5775–5796, Toronto, Canada. Association for Computational Linguistics.
* Cobbe et al. (2021) Karl Cobbe, Vineet Kosaraju, Mohammad Bavarian, Mark Chen, Heewoo Jun, Lukasz Kaiser, Matthias Plappert, Jerry Tworek, Jacob Hilton, Reiichiro Nakano, Christopher Hesse, and John Schulman. 2021. Training verifiers to solve math word problems. _ArXiv_ , abs/2110.14168.
* Conneau et al. (2020) Alexis Conneau, Kartikay Khandelwal, Naman Goyal, Vishrav Chaudhary, Guillaume Wenzek, Francisco Guzmán, Edouard Grave, Myle Ott, Luke Zettlemoyer, and Veselin Stoyanov. 2020. Unsupervised Cross-lingual Representation Learning at Scale. In _Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics_ , pages 8440–8451. Association for Computational Linguistics.
* Conneau et al. (2018) Alexis Conneau, Ruty Rinott, Guillaume Lample, Adina Williams, Samuel Bowman, Holger Schwenk, and Veselin Stoyanov. 2018. XNLI: Evaluating cross-lingual sentence representations. In _Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing_ , pages 2475–2485, Brussels, Belgium. Association for Computational Linguistics.
* Holtzman et al. (2020) Ari Holtzman, Jan Buys, Li Du, Maxwell Forbes, and Yejin Choi. 2020. The curious case of neural text degeneration. In _8th International Conference on Learning Representations, ICLR 2020, Addis Ababa, Ethiopia, April 26-30, 2020_. OpenReview.net.
* Hu et al. (2022) Edward J. Hu, Yelong Shen, Phillip Wallis, Zeyuan Allen-Zhu, Yuanzhi Li, Shean Wang, Lu Wang, and Weizhu Chen. 2022. Lora: Low-rank adaptation of large language models. In _The Tenth International Conference on Learning Representations, ICLR 2022, Virtual Event, April 25-29, 2022_. OpenReview.net.
* Hu et al. (2020) Junjie Hu, Sebastian Ruder, Aditya Siddhant, Graham Neubig, Orhan Firat, and Melvin Johnson. 2020. XTREME: A massively multilingual multi-task benchmark for evaluating cross-lingual generalisation. In _Proceedings of the 37th International Conference on Machine Learning_ , volume 119 of _Proceedings of Machine Learning Research_ , pages 4411–4421. PMLR.
* Huang et al. (2023) Jiaxin Huang, Shixiang Gu, Le Hou, Yuexin Wu, Xuezhi Wang, Hongkun Yu, and Jiawei Han. 2023. Large language models can self-improve. In _Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing_ , pages 1051–1068, Singapore. Association for Computational Linguistics.
* Kim and Rush (2016) Yoon Kim and Alexander M. Rush. 2016. Sequence-level knowledge distillation. In _Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing_ , pages 1317–1327, Austin, Texas. Association for Computational Linguistics.
* Lee et al. (2024) Jinhyuk Lee, Zhuyun Dai, Xiaoqi Ren, Blair Chen, Daniel Cer, Jeremy R. Cole, Kai Hui, Michael Boratko, Rajvi Kapadia, Wen Ding, Yi Luan, Sai Meher Karthik Duddu, Gustavo Hernández Abrego, Weiqiang Shi, Nithi Gupta, Aditya Kusupati, Prateek Jain, Siddhartha R. Jonnalagadda, Ming-Wei Chang, and Iftekhar Naim. 2024. Gecko: Versatile text embeddings distilled from large language models. _ArXiv_ , abs/2403.20327.
* Li et al. (2023a) Xian Li, Ping Yu, Chunting Zhou, Timo Schick, Luke Zettlemoyer, Omer Levy, Jason Weston, and Mike Lewis. 2023a. Self-alignment with instruction backtranslation. _CoRR_ , abs/2308.06259.
* Li et al. (2023b) Zhuoyan Li, Hangxiao Zhu, Zhuoran Lu, and Ming Yin. 2023b. Synthetic data generation with large language models for text classification: Potential and limitations. In _Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing_ , pages 10443–10461, Singapore. Association for Computational Linguistics.
* Loshchilov and Hutter (2019) Ilya Loshchilov and Frank Hutter. 2019. Decoupled weight decay regularization. In _7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, May 6-9, 2019_. OpenReview.net.
* Mulcaire et al. (2019) Phoebe Mulcaire, Jungo Kasai, and Noah A. Smith. 2019. Polyglot contextual representations improve crosslingual transfer. In _Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers)_ , pages 3912–3918, Minneapolis, Minnesota. Association for Computational Linguistics.
* Papineni et al. (2002) Kishore Papineni, Salim Roukos, Todd Ward, and Wei-Jing Zhu. 2002. Bleu: a method for automatic evaluation of machine translation. In _Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics_ , pages 311–318, Philadelphia, Pennsylvania, USA. Association for Computational Linguistics.
* Pikuliak et al. (2021) Matús̆ Pikuliak, Marián Simko, and Mária Bieliková. 2021. Cross-lingual learning for text processing: A survey. _Expert Systems with Applications_ , 165:113765.
* Pires et al. (2019) Telmo Pires, Eva Schlinger, and Dan Garrette. 2019. How multilingual is multilingual BERT? In _Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics_ , pages 4996–5001, Florence, Italy. Association for Computational Linguistics.
* Post (2018) Matt Post. 2018. A call for clarity in reporting BLEU scores. In _Proceedings of the Third Conference on Machine Translation: Research Papers_ , pages 186–191, Brussels, Belgium. Association for Computational Linguistics.
* Rajpurkar et al. (2016) Pranav Rajpurkar, Jian Zhang, Konstantin Lopyrev, and Percy Liang. 2016. SQuAD: 100,000+ questions for machine comprehension of text. In _Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing_ , pages 2383–2392, Austin, Texas. Association for Computational Linguistics.
* Scao et al. (2022) Teven Le Scao, Angela Fan, Christopher Akiki, Ellie Pavlick, Suzana Ili’c, Daniel Hesslow, Roman Castagn’e, Alexandra Sasha Luccioni, François Yvon, Matthias Gallé, Jonathan Tow, Alexander M. Rush, Stella Biderman, Albert Webson, Pawan Sasanka Ammanamanchi, Thomas Wang, Benoît Sagot, Niklas Muennighoff, Albert Villanova del Moral, Olatunji Ruwase, Rachel Bawden, Stas Bekman, Angelina McMillan-Major, Iz Beltagy, Huu Nguyen, Lucile Saulnier, Samson Tan, Pedro Ortiz Suarez, Victor Sanh, Hugo Laurenccon, Yacine Jernite, Julien Launay, Margaret Mitchell, Colin Raffel, Aaron Gokaslan, Adi Simhi, Aitor Soroa Etxabe, Alham Fikri Aji, Amit Alfassy, Anna Rogers, Ariel Kreisberg Nitzav, Canwen Xu, Chenghao Mou, Chris C. Emezue, Christopher Klamm, Colin Leong, Daniel Alexander van Strien, David Ifeoluwa Adelani, Dragomir R. Radev, Eduardo Gonz’alez Ponferrada, Efrat Levkovizh, Ethan Kim, Eyal Natan, Francesco De Toni, Gérard Dupont, Germán Kruszewski, Giada Pistilli, Hady ElSahar, Hamza Benyamina, Hieu Trung Tran, Ian Yu, Idris Abdulmumin, Isaac Johnson, Itziar Gonzalez-Dios, Javier de la Rosa, Jenny Chim, Jesse Dodge, Jian Zhu, Jonathan Chang, Jorg Frohberg, Josephine L. Tobing, Joydeep Bhattacharjee, Khalid Almubarak, Kimbo Chen, Kyle Lo, Leandro von Werra, Leon Weber, Long Phan, Loubna Ben Allal, Ludovic Tanguy, Manan Dey, Manuel Romero Muñoz, Maraim Masoud, María Grandury, Mario vSavsko, Max Huang, Maximin Coavoux, Mayank Singh, Mike Tian-Jian Jiang, Minh Chien Vu, Mohammad A. Jauhar, Mustafa Ghaleb, Nishant Subramani, Nora Kassner, Nurulaqilla Khamis, Olivier Nguyen, Omar Espejel, Ona de Gibert, Paulo Villegas, Peter Henderson, Pierre Colombo, Priscilla Amuok, Quentin Lhoest, Rheza Harliman, Rishi Bommasani, Roberto L’opez, Rui Ribeiro, Salomey Osei, Sampo Pyysalo, Sebastian Nagel, Shamik Bose, Shamsuddeen Hassan Muhammad, Shanya Sharma, S. Longpre, Somaieh Nikpoor, S. Silberberg, Suhas Pai, Sydney Zink, Tiago Timponi Torrent, Timo Schick, Tristan Thrush, Valentin Danchev, Vassilina Nikoulina, Veronika Laippala, Violette Lepercq, Vrinda Prabhu, Zaid Alyafeai, Zeerak Talat, Arun Raja, Benjamin Heinzerling, Chenglei Si, Elizabeth Salesky, Sabrina J. Mielke, Wilson Y. Lee, Abheesht Sharma, Andrea Santilli, Antoine Chaffin, Arnaud Stiegler, Debajyoti Datta, Eliza Szczechla, Gunjan Chhablani, Han Wang, Harshit Pandey, Hendrik Strobelt, Jason Alan Fries, Jos Rozen, Leo Gao, Lintang Sutawika, M Saiful Bari, Maged S. Al-Shaibani, Matteo Manica, Nihal V. Nayak, Ryan Teehan, Samuel Albanie, Sheng Shen, Srulik Ben-David, Stephen H. Bach, Taewoon Kim, Tali Bers, Thibault Févry, Trishala Neeraj, Urmish Thakker, Vikas Raunak, Xiang Tang, Zheng-Xin Yong, Zhiqing Sun, Shaked Brody, Y Uri, Hadar Tojarieh, Adam Roberts, Hyung Won Chung, Jaesung Tae, Jason Phang, Ofir Press, Conglong Li, Deepak Narayanan, Hatim Bourfoune, Jared Casper, Jeff Rasley, Max Ryabinin, Mayank Mishra, Minjia Zhang, Mohammad Shoeybi, Myriam Peyrounette, Nicolas Patry, Nouamane Tazi, Omar Sanseviero, Patrick von Platen, Pierre Cornette, Pierre Franccois Lavall’ee, Rémi Lacroix, Samyam Rajbhandari, Sanchit Gandhi, Shaden Smith, Stéphane Requena, Suraj Patil, Tim Dettmers, Ahmed Baruwa, Amanpreet Singh, Anastasia Cheveleva, Anne-Laure Ligozat, Arjun Subramonian, Aur’elie N’ev’eol, Charles Lovering, Daniel H Garrette, Deepak R. Tunuguntla, Ehud Reiter, Ekaterina Taktasheva, Ekaterina Voloshina, Eli Bogdanov, Genta Indra Winata, Hailey Schoelkopf, Jan-Christoph Kalo, Jekaterina Novikova, Jessica Zosa Forde, Xiangru Tang, Jungo Kasai, Ken Kawamura, Liam Hazan, Marine Carpuat, Miruna Clinciu, Najoung Kim, Newton Cheng, Oleg Serikov, Omer Antverg, Oskar van der Wal, Rui Zhang, Ruochen Zhang, Sebastian Gehrmann, Shachar Mirkin, S. Osher Pais, Tatiana Shavrina, Thomas Scialom, Tian Yun, Tomasz Limisiewicz, Verena Rieser, Vitaly Protasov, Vladislav Mikhailov, Yada Pruksachatkun, Yonatan Belinkov, Zachary Bamberger, Zdenvek Kasner, Zdeněk Kasner, Amanda Pestana, Amir Feizpour, Ammar Khan, Amy Faranak, Ananda Santa Rosa Santos, Anthony Hevia, Antigona Unldreaj, Arash Aghagol, Arezoo Abdollahi, Aycha Tammour, Azadeh HajiHosseini, Bahareh Behroozi, Benjamin Ayoade Ajibade, Bharat Kumar Saxena, Carlos Muñoz Ferrandis, Danish Contractor, David M. Lansky, Davis David, Douwe Kiela, Duong Anh Nguyen, Edward Tan, Emi Baylor, Ezinwanne Ozoani, Fatim Tahirah Mirza, Frankline Ononiwu, Habib Rezanejad, H.A. Jones, Indrani Bhattacharya, Irene Solaiman, Irina Sedenko, Isar Nejadgholi, Jan Passmore, Joshua Seltzer, Julio Bonis Sanz, Karen Fort, Lívia Dutra, Mairon Samagaio, Maraim Elbadri, Margot Mieskes, Marissa Gerchick, Martha Akinlolu, Michael McKenna, Mike Qiu, Muhammed Ghauri, Mykola Burynok, Nafis Abrar, Nazneen Rajani, Nour Elkott, Nourhan Fahmy, Olanrewaju Samuel, Ran An, R. P. Kromann, Ryan Hao, Samira Alizadeh, Sarmad Shubber, Silas L. Wang, Sourav Roy, Sylvain Viguier, Thanh-Cong Le, Tobi Oyebade, Trieu Nguyen Hai Le, Yoyo Yang, Zach Nguyen, Abhinav Ramesh Kashyap, Alfredo Palasciano, Alison Callahan, Anima Shukla, Antonio Miranda-Escalada, Ayush Kumar Singh, Benjamin Beilharz, Bo Wang, Caio Matheus Fonseca de Brito, Chenxi Zhou, Chirag Jain, Chuxin Xu, Clémentine Fourrier, Daniel Le’on Perin’an, Daniel Molano, Dian Yu, Enrique Manjavacas, Fabio Barth, Florian Fuhrimann, Gabriel Altay, Giyaseddin Bayrak, Gully Burns, Helena U. Vrabec, Iman I.B. Bello, Isha Dash, Ji Soo Kang, John Giorgi, Jonas Golde, Jose David Posada, Karthi Sivaraman, Lokesh Bulchandani, Lu Liu, Luisa Shinzato, Madeleine Hahn de Bykhovetz, Maiko Takeuchi, Marc Pàmies, María Andrea Castillo, Marianna Nezhurina, Mario Sanger, Matthias Samwald, Michael Cullan, Michael Weinberg, M Wolf, Mina Mihaljcic, Minna Liu, Moritz Freidank, Myungsun Kang, Natasha Seelam, Nathan Dahlberg, Nicholas Michio Broad, Nikolaus Muellner, Pascale Fung, Patricia Haller, R. Chandrasekhar, Renata Eisenberg, Robert Martin, Rodrigo Canalli, Rosaline Su, Ruisi Su, Samuel Cahyawijaya, Samuele Garda, Shlok S Deshmukh, Shubhanshu Mishra, Sid Kiblawi, Simon Ott, Sinee Sang-aroonsiri, Srishti Kumar, Stefan Schweter, Sushil Pratap Bharati, Tanmay Laud, Théo Gigant, Tomoya Kainuma, Wojciech Kusa, Yanis Labrak, Yashasvi Bajaj, Y. Venkatraman, Yifan Xu, Ying Xu, Yu Xu, Zhee Xao Tan, Zhongli Xie, Zifan Ye, Mathilde Bras, Younes Belkada, and Thomas Wolf. 2022. Bloom: A 176b-parameter open-access multilingual language model. _ArXiv_ , abs/2211.05100.
* Shaham et al. (2024) Uri Shaham, Jonathan Herzig, Roee Aharoni, Idan Szpektor, Reut Tsarfaty, and Matan Eyal. 2024. Multilingual instruction tuning with just a pinch of multilinguality. _ArXiv_ , abs/2401.01854.
* Shi et al. (2023) Freda Shi, Mirac Suzgun, Markus Freitag, Xuezhi Wang, Suraj Srivats, Soroush Vosoughi, Hyung Won Chung, Yi Tay, Sebastian Ruder, Denny Zhou, Dipanjan Das, and Jason Wei. 2023. Language models are multilingual chain-of-thought reasoners. In _The Eleventh International Conference on Learning Representations, ICLR 2023, Kigali, Rwanda, May 1-5, 2023_. OpenReview.net.
* Sun et al. (2023) Zhiqing Sun, Yikang Shen, Qinhong Zhou, Hongxin Zhang, Zhenfang Chen, David D. Cox, Yiming Yang, and Chuang Gan. 2023. Principle-driven self-alignment of language models from scratch with minimal human supervision. In _Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, NeurIPS 2023, New Orleans, LA, USA, December 10 - 16, 2023_.
* Touvron et al. (2023) Hugo Touvron, Louis Martin, Kevin R. Stone, Peter Albert, Amjad Almahairi, Yasmine Babaei, Nikolay Bashlykov, Soumya Batra, Prajjwal Bhargava, Shruti Bhosale, Daniel M. Bikel, Lukas Blecher, Cristian Cantón Ferrer, Moya Chen, Guillem Cucurull, David Esiobu, Jude Fernandes, Jeremy Fu, Wenyin Fu, Brian Fuller, Cynthia Gao, Vedanuj Goswami, Naman Goyal, Anthony S. Hartshorn, Saghar Hosseini, Rui Hou, Hakan Inan, Marcin Kardas, Viktor Kerkez, Madian Khabsa, Isabel M. Kloumann, A. V. Korenev, Punit Singh Koura, Marie-Anne Lachaux, Thibaut Lavril, Jenya Lee, Diana Liskovich, Yinghai Lu, Yuning Mao, Xavier Martinet, Todor Mihaylov, Pushkar Mishra, Igor Molybog, Yixin Nie, Andrew Poulton, Jeremy Reizenstein, Rashi Rungta, Kalyan Saladi, Alan Schelten, Ruan Silva, Eric Michael Smith, R. Subramanian, Xia Tan, Binh Tang, Ross Taylor, Adina Williams, Jian Xiang Kuan, Puxin Xu, Zhengxu Yan, Iliyan Zarov, Yuchen Zhang, Angela Fan, Melanie Kambadur, Sharan Narang, Aurelien Rodriguez, Robert Stojnic, Sergey Edunov, and Thomas Scialom. 2023. Llama 2: Open foundation and fine-tuned chat models. _ArXiv_ , abs/2307.09288.
* Tu et al. (2016) Zhaopeng Tu, Zhengdong Lu, Yang Liu, Xiaohua Liu, and Hang Li. 2016. Modeling coverage for neural machine translation. In _Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)_ , pages 76–85, Berlin, Germany. Association for Computational Linguistics.
* Williams et al. (2018) Adina Williams, Nikita Nangia, and Samuel Bowman. 2018. A broad-coverage challenge corpus for sentence understanding through inference. In _Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers)_ , pages 1112–1122, New Orleans, Louisiana. Association for Computational Linguistics.
* Xue et al. (2021) Linting Xue, Noah Constant, Adam Roberts, Mihir Kale, Rami Al-Rfou, Aditya Siddhant, Aditya Barua, and Colin Raffel. 2021. mT5: A massively multilingual pre-trained text-to-text transformer. In _Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies_ , pages 483–498, Online. Association for Computational Linguistics.
## Appendix A The Details of Experimental Setups
### A.1 Tasks and Datasets
We provide the details of the datasets introduced in Section 4.1.
#### A.1.1 Question Answering (QA)
SQuAD (Rajpurkar et al., 2016) is an English QA dataset created from Wikipedia
articles as training data. Given a question and a passage, the task is to
extract the answer from the passage. Evaluation is conducted with XQuAD
(Artetxe et al., 2020), which consists of translation of SQuAD into multiple
languages.
Context: Architecturally, the school has a Catholic character. Atop the
Main Building’s gold dome is a golden statue of the Virgin Mary. Immediately
in front of the Main Building and facing it, is a copper statue of Christ
with arms upraised with the legend "Venite Ad Me Omnes". Next to the Main
Building is the Basilica of the Sacred Heart. Immediately behind the basilica
is the Grotto, a Marian place of prayer and reflection. It is a replica of the
grotto at Lourdes, France where the Virgin Mary reputedly appeared to Saint
Bernadette Soubirous in 1858. At the end of the main drive (and in a direct line
that connects through 3 statues and the Gold Dome), is a simple, modern stone
statue of Mary.
Question: To whom did the Virgin Mary allegedly appear in 1858 in Lourdes France?
---
Saint Bernadette Soubirous
Figure 3: An input and output example of the SQuAD dataset.
#### A.1.2 Text-Pair Classification
We also evaluate our method on cross-lingual text-pair classification tasks.
The MultiNLI dataset (Williams et al., 2018) involves determining the logical
relationship between a premise sentence and a hypothesis sentence. XNLI
(Conneau et al., 2018) is a multilingual NLI dataset for evaluation.
Premise: Conceptually cream skimming has two basic dimensions - product and geography.
Hypothesis: Product and geography are what make cream skimming work.
What is their logical relation? Entailment, Neutral or Contradition.
---
Neutral
Figure 4: An input and output example of the MultiNLP dataset.
#### A.1.3 Mathematical Reasoning
GSM8k (Cobbe et al., 2021) is an English dataset of 8.5K high-quality grade
school math problems. Each problem is annotated with a solution that shows the
mathematical steps required to reach the final answer. As the evaluation
dataset, we use MGSM (Shi et al., 2023), a multilingual version of the GSM8k
dataset.
The LLM is trained to generate the step-by-step solution to math problems. The
answer is extracted from the LLM output as the final digits, and the accuracy
is calculated based on the exact match of the extracted answer and the ground
truth.
Natalia sold clips to 48 of her friends in April,
and then she sold half as many clips in May.
How many clips did Natalia sell altogether in April and May?
---
Natalia sold 48/2 = clips in May.
Natalia sold 48+24 = 72 clips altogether in April and May.
#### 72
Table 4: An input and output example of the GSM8k dataset.
### A.2 Prompt Format for LLM Translation
To translate training data using a LLM (Section 4.4), we employed the
following prompt template for each task. The template simply consists of the
source text and target text prepended with the language tag. The text is
surrounded by backticks and the LLM starts generating the target text the open
backtick until the close backtick is found.
{% for sample in few_shot_samples %}
en: ‘{{ sample.data_field }}‘
{{ target_language }}: ‘{{ sample.data_field }}‘
{% endfor %}
en: ‘{{ data_field }}‘
{{ target_language }}: ‘
---
Figure 5: Prompt format for LLM translation.
The SQuAD dataset annotates answer spans in the context passages. We translate
the annotations using the mark-then-translate approach (Chen et al., 2023). We
mark the answer span in the context passage with the tokens “<answer>” and
“</answer>”, translate the marked text, and then extract the translated answer
span from the translated context. Note that in this case, the few-shot samples
are also marked with the answer span.
The context passages in the SQuAD dataset are relatively long, and it is
challenging for the LLM with a limited context window to fit the entire few-
shot samples and the source text. To address this issue, we split the context
into sentences using the spaCy library111https://spacy.io/ and translate them
separately, i.e., the few-shot samples and the source text are sentences.
### A.3 Data Filtering for Synthetic Data
We remove pairs where the target length is less than one-third or more than
three times the source length. The text length is heuristically determined to
account for character length differences between languages. For example,
phonogram-based text (e.g., English) has much more characters than ideograph-
based text (e.g., Chinese). We set normalization factors where English,
German, Thai, and Russian characters count as 1, and Chinese characters as 3.
We also filter out incomplete translations which are typically produced by
repetitive generation. We set the maximum number of tokens for generation
(Table 5) and remove the outputs not ending with the token indicating the end
of the translation, in our case, the backtick character used in the prompt
format.
| Data Field | Max Number of Tokens
---|---|---
SQuAD (Rajpurkar et al., 2016) | context | 512
question | 256
MultiNLI (Williams et al., 2018) | premise | 256
hypothesis | 256
GSM8k (Cobbe et al., 2021) | question | 512
answer | 512
Table 5: Maximum number of tokens set for generating translations.
### A.4 Assessing the Translation Quality
To evaluate the translation quality, we use the BLEU score (Papineni et al.,
2002) measured by the parallel data constructed from questions in the MGSM
test set. The translation is performed in few-shot in-context learning with 8
translation samples constructed from the train set of the MGSM dataset. The
BLEU score is calculated using the SacreBLEU library (Post,
2018)222https://github.com/mjpost/sacrebleu. As the tokenizer option, we use
“13a” for de and ru, “flores101” for th, and “zh” for zh.
Table 6 shows the BLEU scores from the LLMs evaluated in this paper. The
result of Qwen1.5-1.8B is discussed in Appendix B, and gpt-3.5-turbo-0125 in
Section C.3.
Model | de | ru | th | zh
---|---|---|---|---
Llama2-7B | 37.1 | 27.2 | 1.9 | 29.4
Llama2-13B | 41.3 | 33.4 | 5.1 | 34.3
Llama2-70B | 45.6 | 41.9 | 12.0 | 42.4
Qwen1.5-1.8B | 21.9 | 11.3 | 1.5 | 41.3
gpt-3.5-turbo-0125 | 48.0 | 44.6 | 23.1 | 47.2
Table 6: BLEU scores from the MGSM test set.
### A.5 Hyper-parameters for Fine-tuning
We provide the hyper-parameters used for fine-tuning the LLMs in Table 7.
Hyper-parameter | Value | Hyper-parameter | Value
---|---|---|---
Batch size | 64 | Adam $\epsilon$ | 1e-8
Number of steps | 1,000 | Adam $\beta_{1}$ | 0.9
Learning rate | [5e-5, 3e-4] | Adam $\beta_{2}$ | 0.999
LR Scheduler | Cosine | Weight decay | 0.1
Warmup ratio | 0.05 | |
Table 7: Hyper-parameters used for fine-tuning the LLMs.
## Appendix B Results from Qwen1.5-1.8B
To increase the robustness of the results, we also conducted experiments with
Qwen1.5-1.8B 333https://qwenlm.github.io/blog/qwen1.5/. While the model is
mainly trained on Chinese and English data, it is also constructed with the
multilingual use cases in mind.
| gsm8k
---|---
| de | ru | th | zh
$\mathcal{D}_{\text{src}}$ | | 8.0
---
$\pm$0.3
| 6.5
---
$\pm$0.4
| 2.8
---
$\pm$0.5
| 23.3
---
$\pm$0.5
$+\mathcal{D}_{\text{tgt}}$ | * | 18.7
---
$\pm$1.8
* * | 15.0
---
$\pm$1.2
* | 3.4
---
$\pm$1.8
| 21.4
---
$\pm$1.2
$+\mathcal{D}_{\text{cs}}$ | * | 17.1
---
$\pm$1.3
* * | 14.9
---
$\pm$0.5
* | 2.7
---
$\pm$0.8
| 23.2
---
$\pm$0.9
Table 8: Results on the MGSM dataset with Qwen1.5-1.8B. Scores are marked with
∗ if its improvement is statistically significant ($p<0.05$ in Welch’s t-test)
compared to the baseline $\mathcal{D}_{\text{src}}$. The significant and
highest score in each column is marked in bold.
We observe that the results are consistent with the main experiments: Self-
Translate-Train is effective when the zero-shot cross-lingual transfer
performance is suboptimal and the model can generate reasonable translations.
The performance is improved by adding the target language data
$+\mathcal{D}_{\text{tgt}}$. However, when the translation quality is poor as
in Thai (1.5 BLEU score in Table 6), the improvement is not observed.
Additionally, Qwen1.5-1.8B seems to have good cross-lingual capability between
English and Chinese, as indicated by the high BLEU score (41.3 in Table 6).
With this, tuning on the source language data alone is sufficient to achieve
high performance.
## Appendix C Frequently Asked Questions
In this section, we discuss questions that are outside the scope of the main
topic of this paper but are somewhat relevant and may be of interest to
readers.
### C.1 Does Self-Translate-Train improve the performance in the source
language?
The performance somtimes improves, given the task is challenging and the
translation quality is sufficiently high.
Table 9 shows the results on the English test set with Llama2-7B. The
performance improves in the MGSM dataset when adding the synthetic data from
de, ru, and zh. The Thai language does not show the improvement possibly due
to the low translation quality.
However, the improvement is not observed in the XQuAD and XNLI datasets. This
might be because the task performance is already high with the source language
data alone, and the synthetic data does not provide additional information to
improve the performance.
| MGSM | XQuAD | XNLI
---|---|---|---
| de | ru | th | zh | de | ru | th | zh | de | ru | th | zh
$\mathcal{D}_{\text{src}}$ | | 37.8
---
$\pm$0.8
| 70.2
---
$\pm$0.5
| 88.2
---
$\pm$1.3
$+\mathcal{D}_{\text{tgt}}$ | * | 42.6
---
$\pm$1.9
* * | 41.8
---
$\pm$1.3
* | 40.0
---
$\pm$1.9
* | 42.7
---
$\pm$1.0
* | 69.1
---
$\pm$0.7
| 69.6
---
$\pm$0.5
| 70.0
---
$\pm$0.5
| 69.9
---
$\pm$0.4
| 89.0
---
$\pm$0.4
| 88.3
---
$\pm$0.7
| 88.6
---
$\pm$1.0
| 89.3
---
$\pm$0.3
$+\mathcal{D}_{\text{cs}}$ | * | 42.7
---
$\pm$0.7
* * | 42.9
---
$\pm$0.7
* | 39.8
---
$\pm$1.6
| 40.8
---
$\pm$2.0
| 69.5
---
$\pm$0.5
| 69.6
---
$\pm$0.3
| 69.8
---
$\pm$0.5
| 68.7
---
$\pm$0.5
| 88.7
---
$\pm$0.2
| 88.5
---
$\pm$1.0
| 88.0
---
$\pm$0.7
| 88.7
---
$\pm$0.6
Table 9: Results on the Englihs test set with Llama2-7B. Scores are marked
with ∗ if its improvement is statistically significant ($p<0.05$ in Welch’s
t-test) compared to the baseline $\mathcal{D}_{\text{src}}$. The significant
and highest score in each column is marked in bold.
### C.2 Does the synthetic data alone improve the performance in the target
language?
Yes, but adding the source language data is more effective. Table 10 shows the
results with Llama2-7B on the multilingual evaluation datasets. Tuning on the
synthetic data alone ($\mathcal{D}_{\text{tgt}}$) improves the performance in
the target language, but the improvement is not as significant as adding the
synthetic data to the source language data ($+\mathcal{D}_{\text{tgt}}$). In
practice, we recommend using the synthetic data in combination with the
original data to achieve the best performance.
| MGSM | XQuAD | XNLI
---|---|---|---
| de | ru | th | zh | de | ru | th | zh | de | ru | th | zh
$\mathcal{D}_{\text{src}}$ | | 30.1
---
$\pm$0.4
| 25.0
---
$\pm$0.7
| 8.1
---
$\pm$0.7
| 21.1
---
$\pm$1.7
| 60.3
---
$\pm$0.8
| 49.3
---
$\pm$0.4
| 34.5
---
$\pm$1.0
| 66.3
---
$\pm$0.8
| 79.7
---
$\pm$0.4
| 76.9
---
$\pm$0.1
| 53.7
---
$\pm$0.9
| 74.1
---
$\pm$0.2
$D_{\text{tgt}}$ | * | 32.1
---
$\pm$0.6
* * | 30.4
---
$\pm$1.8
* | 8.7
---
$\pm$0.6
| 24.6
---
$\pm$2.5
* | 62.5
---
$\pm$0.4
* * | 57.4
---
$\pm$1.1
* * | 44.3
---
$\pm$0.8
* * | 75.4
---
$\pm$1.1
* * | 81.5
---
$\pm$0.6
* | 77.9
---
$\pm$0.7
| 53.8
---
$\pm$1.7
* | 77.1
---
$\pm$0.5
*
$+\mathcal{D}_{\text{tgt}}$ | * | 36.4
---
$\pm$1.3
* * | 34.0
---
$\pm$1.7
* | 7.7
---
$\pm$3.4
* | 27.1
---
$\pm$1.2
* | 61.7
---
$\pm$0.9
* | 57.8
---
$\pm$0.8
* * | 46.4
---
$\pm$1.3
* * | 77.7
---
$\pm$0.4
* * | 81.6
---
$\pm$0.7
* * | 78.5
---
$\pm$0.6
* | 56.3
---
$\pm$1.5
* | 78.5
---
$\pm$0.3
*
Table 10: Results on the Englihs test set with Llama2-7B with the setting of
tuning the synthetic data alone $\mathcal{D}_{\text{tgt}}$.
### C.3 Is tuning on $\mathcal{D}_{\text{tgt}}$ generated by another model
still effective?
Yes, if the other model can generate reasonable translations. Such approach
can be seen as the Translate-train approach (Section 2.1) or sequence
distillation from a teacher model (Kim and Rush, 2016).
As an upper-bound experiment using the math task, we fine-tune Llama2-7B on
the synthetic data generated by gpt-3.5-turbo-0125 from the OpenAI
API444https://openai.com/index/openai-api/, which produces high-quality
translations across the languages explored in this paper (Table 6).
| MGSM
---|---
| de | ru | th | zh
$\mathcal{D}_{\text{src}}$ | | 30.1
---
$\pm$0.4
| 25.0
---
$\pm$0.7
| 8.1
---
$\pm$0.7
| 21.1
---
$\pm$1.7
$+\mathcal{D}_{\text{tgt}}$ | * | 37.4
---
$\pm$1.0
* * | 35.9
---
$\pm$1.2
* * | 28.5
---
$\pm$1.4
* * | 33.2
---
$\pm$1.1
*
$+\mathcal{D}_{\text{cs}}$ | * | 38.6
---
$\pm$1.5
* * | 34.4
---
$\pm$0.5
* * | 28.7
---
$\pm$1.3
* * | 34.8
---
$\pm$1.3
*
Table 11: Results on multilingual evaluation datasets with Llama2-7B tuned on
the synthetic data generated by gpt-3.5-turbo-0125.
Adding the synthetic data generated by gpt-3.5-turbo-0125 improves performance
across languages.
However, the outputs from external models are often restricted in its
usage555For example, Term of use of OpenAI API (January 31, 2024) restricts
the usage of the outputs for training a model that competes with the API
(https://openai.com/policies/terms-of-use/). Meta Llama 3 License (April 18,
2024) prohibits using the outputs to improve any other large language model
(https://llama.meta.com/llama3/license/)., while the method explored in this
paper can be used with the model at hand without other resources.
Additionally, our interest in this paper is rather to explore the cross-
lingual potential of the model itself and how to better utilize it.
|
# Inverse Design of Nonlinear Metasurfaces for Sum Frequency Generation
Neuton Li ARC Centre of Excellence for Transformative Meta-Optical Systems
(TMOS), Department of Electronic Materials Engineering, Research School of
Physics, The Australian National University, Canberra, ACT 2600, Australia,
<EMAIL_ADDRESS>Jihua Zhang ARC Centre of Excellence for Transformative
Meta-Optical Systems (TMOS), Department of Electronic Materials Engineering,
Research School of Physics, The Australian National University, Canberra, ACT
2600, Australia<EMAIL_ADDRESS>Songshan Lake Materials Laboratory,
Dongguan, Guangdong 523808, P. R. China Dragomir N. Neshev ARC Centre of
Excellence for Transformative Meta-Optical Systems (TMOS), Department of
Electronic Materials Engineering, Research School of Physics, The Australian
National University, Canberra, ACT 2600, Australia<EMAIL_ADDRESS>Andrey A. Sukhorukov ARC Centre of Excellence for Transformative Meta-Optical
Systems (TMOS), Department of Electronic Materials Engineering, Research
School of Physics, The Australian National University, Canberra, ACT 2600,
Australia<EMAIL_ADDRESS>
###### Abstract
Sum frequency generation (SFG) has multiple applications, from optical sources
to imaging, where efficient conversion requires either long interaction
distances or large field concentrations in a quadratic nonlinear material.
Metasurfaces provide an essential avenue to enhanced SFG due to resonance with
extreme field enhancements with an integrated ultrathin platform. In this
work, we formulate a general theoretical framework for multi-objective
topology optimization of nanopatterned metasurfaces that facilitate high-
efficiency SFG and simultaneously select the emitted direction and tailor the
metasurface polarization response. Based on this framework, we present novel
metasurface designs showcasing ultimate flexibility in transforming the
outgoing nonlinearly generated light for applications spanning from imaging to
polarimetry. For example, one of our metasurfaces produces highly polarized
and directional SFG emission with an efficiency of over
$0.2\text{\,}{\mathrm{cm}}^{2}\text{\,}{\mathrm{GW}}^{-1}$ in a
$10\text{\,}\mathrm{nm}$ signal operating bandwidth.
## 1 Introduction
Sum-frequency generation (SFG) is a fundamentally important second-order
nonlinear process with many applications ranging from wavelength conversion of
optical sources [1] and infrared imaging [2, 3, 4, 5] to nonlinear polarimetry
[6]. This phenomenon arises from induced polarizations in the medium, and in
general, it can be observed in the presence of strong optical fields.
Efficient second-order nonlinear frequency conversion, such as the SFG,
traditionally requires long interaction lengths in bulky nonlinear crystals.
As a result, only certain crystalline orientations and input polarisations can
satisfy the phase-matching conditions for producing sizable nonlinear effects.
This limits the types of polarization transformations and the directionality
of emission that are possible in nonlinear bulk crystals.
Recent advances in nanotechnologies have facilitated the development of ultra-
thin single-layer dielectric metasurfaces, where optical nano-resonators can
enhance and tailor the nonlinear interactions with functionalities beyond the
capabilities of traditional bulky crystals [7, 8, 9, 10, 11]. To generate
optical resonances, previous metasurface designs have often relied on semi-
analytical approaches in the limiting cases of Mie-type modes for individual
nanoresonators [12, 13, 14, 15, 16, 17, 18], or bound state in the continuum
resonances [19, 20, 21, 22, 23, 24, 25, 26]. The angular-dependent properties
of nonlocal metasurfaces could also be utilized to tune the nonlinear
interactions over a range of wavelengths [27]. There is ongoing research on
the enhancement of SFG in metasurfaces with resonances at non-degenerate
wavelengths [28, 29, 30, 31]. Non-planar structures with broken 3D symmetries
have been identified for designing effective nonlinear susceptibility response
[32, 33]. Furthermore, several studies have considered inverse-design or
machine learning approaches to optimize for strong resonances [34, 35, 36, 37,
38]. The resulting geometries have highly counterintuitive nanostructured
geometries, which prove superior to conventional designs in their respective
applications.
However, these examples fall short of controlling nonlinear generation beyond
the mere enhancement of conversion efficiency, which is only a part of the
advantages that metasurfaces offer. For many practical applications such as
imaging, it is often important to consider the input and output polarizations
as well as the directional distribution of the generated emission. Control of
the nonlinear polarization and the regulation of diffraction orders while
maintaining high overall efficiency still remain a major challenge in the
field. High-efficiency nonlinear conversion generally requires strong field
enhancements in the nonlinear region, which can be characterized by the
Q-factor of the resonance, its matching with the pulse bandwidth, and the
overlap between high Q-factor modes. The task of engineering overlapping high
Q-factor resonances with a prescribed bandwidth in materials at non-degenerate
wavelengths is difficult in itself. The additional objectives of engineering
emission polarization and direction expand the complexity of the task even
further.
In this manuscript, we address the aforementioned engineering challenges by
developing a novel inverse-design framework for optimizing metasurfaces that
enables the simultaneous enhancement of the SFG efficiency over a desired
pulse bandwidth, tailoring the polarization transformation matrix, and
increasing the emission directionality in a multi-objective manner. Our
resulting metasurfaces account for the intrinsic $\hat{\chi}^{(2)}$ of the
material and combine it with the optimized structural geometries to form an
effective $\hat{\chi}^{(2)}_{\text{eff}}$ of the device [Fig. 1(a)], enabling
ultimate control of nonlinear interactions. Thereby, we broaden the
flexibility and increase the benefits of using nonlinear metasurfaces, from
wavelength conversion to imaging and other applications.
Fig. 1: (a) Schematic of the possible processes that can occur through SFG.
The input fields $\mathbf{E}_{1}$ and $\mathbf{E}_{2}$ can have arbitrary
polarization states, which then generate the SFG field $\mathbf{E}_{3}$. Each
diffracted order of the SFG field can have an independent effective
$\chi^{(2)}_{\text{eff}}$ associated with it. By optimization of the
structure, each of these $\chi^{(2)}_{\text{eff}}$ can be tailored as desired.
(b) The energy levels of the SFG process with each field's respective
frequencies.
## 2 Methods
### 2.1 Adjoint Optimization of SFG
Beginning from the perspective of classical nonlinear optics, a polarization
field is induced in the medium that has a nonlinear relationship with applied
electric field [39, 40]
$\mathbf{P}=\varepsilon_{0}\quantity[\chi^{(1)}:\mathbf{E}+\chi^{(2)}:\mathbf{EE}+\chi^{(3)}:\mathbf{EEE}+\dots].$
(1)
This polarization then generates a nonlinear current that excites the fields
at harmonic frequencies. Our work focuses on the second-order process, so we
explicitly write this nonlinear current at the sum-frequency $\omega_{3}$ as
$J_{3,i}(\mathbf{r})=-i\omega_{3}\varepsilon_{0}\xi(\mathbf{r})\sum_{jk}\chi^{(2)}_{ijk}E_{1,j}(\mathbf{r})E_{2,k}(\mathbf{r})$
(2)
for input fields $\mathbf{E}_{1}$ and $\mathbf{E}_{2}$ at the frequencies
$\omega_{1}$ and $\omega_{2}$, respectively [Fig. 1(b)]. The implementation of
an additional position-dependent parameter $\xi(\mathbf{r})$ allows the
strength of the $\chi^{(2)}$ interaction to be modified simultaneously with
the value of the refractive index in the presence or absence of material [Fig.
2(a)]. We assume the non-depleted pump approximation, where the driving fields
will not suffer a significant loss of intensity as it propagates due to low
enough conversion efficiency. Then the fields $\mathbf{E}_{1}$ and
$\mathbf{E}_{2}$ are found as solutions of linear Maxwell's equations, and
then $\mathbf{E}_{3}$ at $\lambda_{3}$ may be obtained by solving the
inhomogeneous wave equation with the current from Eq. (2).
We aim to maximize the value of an objective function that depends on the
generated sum-frequency field, which in turn depends on the medium parameters
($p$) and on the incident fields:
$T(\mathbf{E}_{3})=T(\mathbf{E}_{3}(p,\mathbf{E}_{1},\mathbf{E}_{2})).$ (3)
We choose the optimization parameter as $p=\xi({\mathbf{r}})$ and define the
permittivity at each position of the design space as
$\varepsilon(\mathbf{r})=\varepsilon_{c}+\xi({\mathbf{r}})(\varepsilon_{d}-\varepsilon_{c}),\leavevmode\nobreak\
0\leq\xi({\mathbf{r}})\leq 1,$ (4)
where $\varepsilon_{d}$ and $\varepsilon_{c}$ are the permittivity for the
patterned and cladding materials, respectively. Thereby, $p=0$ or $p=1$
corresponds to the absence or presence of a nonlinear material at a particular
spatial location according to Eqs. (2) & (4), while the values of $0<p<1$ are
used during the intermediate stages of the optimization process as we discuss
in the following paragraph.
Fig. 2: (a) Depiction of topology optimization. The material derivative at
each discrete point is calculated through a series of forward and adjoint
simulations. (b) Forward simulations of sum-frequency generation, and (c)
corresponding adjoint simulations. The forward fields are labelled
$\mathbf{E}_{i}$, and the adjoint fields are labelled $\mathbf{v}_{j}$ for
wavelength $i,j$.
We formulate the adjoint optimization approach to nonlinear metasurfaces
following the general principles developed in Refs. [41, 42, 43, 44, 45, 46].
Previously, nonlinear cavities [47] and metasurfaces [35, 38] were optimized
to maximize the total second-harmonic generation. Here, we focus on the SFG
process, which includes three fields with different spatial distributions,
adding further complexity to the optimisation problem. Furthermore, we perform
optimization for SFG radiation over targeted diffraction orders, rather than
just the total converted radiation in all directions, as well as the tailored
nonlinear polarization transformations. For this purpose, we aim to
computationally efficiently calculate the derivatives $dT/dp$ simultaneously
for all spatial points in the nonlinear layer, which then allows for the fast
gradient descent optimization of the metasurface patterns as sketched in Fig.
2(a). Instead of repeating the forward calculation for material variations at
every spatial point [Fig. 2(b)], we identify the adjoint linear problems at
the three wavelengths [Fig. 2(c)], which solutions allow the calculation of
objective function derivative at all spatial locations through the overlaps
between the fields at the three non-degenerate wavelengths and each of their
corresponding adjoint fields. Detailed mathematical expressions for the
material derivative and its formulation can be found in Supplementary S1. In
the main manuscript, we summarize the ways in which the necessary adjoint
electric fields are obtained.
For optimization of sum-frequency emission in the outward far-field and on the
surface $\bf{\Omega}$, we can define the objective function through the
complex amplitude $a$ of a particular wave or mode with a field $({\bf
E}_{3,f},{\bf H}_{3,f})$. This mode can be of any form, including plane waves,
Gaussian beams, vortices, and other beam shapes, in combination with any
polarization structure. Then, the mode amplitude can be defined as
$a_{3}=\frac{1}{N_{3}}\iint_{\Omega}{\bf n}\cdot\left[{\bf E_{3}}\times{\bf
H}_{3,b}-{\bf E}_{3,b}\times{\bf H_{3}}\right]d\Omega\,,$ (5)
where n is the unit normal vector outward the surface $\Omega$, $({\bf
E}_{3,b},{\bf H}_{3,b})$ is a direction-reversed wave from $({\bf
E}_{3,f},{\bf H}_{3,f})$, and the normalisation coefficient is
$N_{3}=\iint_{\Omega}{\bf n}\cdot\left[{\bf E}_{3,f}\times{\bf H}_{3,b}-{\bf
E}_{3,b}\times{\bf H}_{3,f}\right]d\Omega$.
We now consider the most common case, in which all materials are reciprocal
with symmetric permittivity and permeability, and we only have the electric
field-induced electric current source. Then, the adjoint field
$\mathbf{v}_{E,3}$ at $\lambda_{3}$ is a result of linear scattering from the
metasurface for a source whose input wave is $({\bf E}_{3,b}/N_{3},{\bf
H}_{3,b}/N_{3})$.
The adjoint fields $\mathbf{v}_{E,1,2}$ at $\lambda_{1,2}$ are obtained by
solving Maxwell's equations with a current source
$\displaystyle{\bf J}_{v,1}$ $\displaystyle=$ $\displaystyle{\bf
L}_{1}^{T}{\bf v}_{E,3}\,,$ (6) $\displaystyle{\bf J}_{v,2}$ $\displaystyle=$
$\displaystyle{\bf L}_{2}^{T}{\bf v}_{E,3}\,,$ (7)
where
$\displaystyle
L_{1,ij}=-i\varepsilon_{0}\omega_{3}\xi(\mathbf{r})\sum_{k}\chi^{(2)}_{ijk}E_{2,k}\,,$
(8) $\displaystyle
L_{2,ik}=-i\varepsilon_{0}\omega_{3}\xi(\mathbf{r})\sum_{j}\chi^{(2)}_{ijk}E_{1,j}\,.$
(9)
These currents only exist where $\xi(\mathbf{r})\chi_{ijk}^{(2)}$ is non-zero,
i.e. in the nonlinear material. We note that these currents have a dependency
on $\mathbf{v}_{E,3}$, and so it must be obtained before solving the
aforementioned equations.
Finally, the objective function gradient is
$\begin{split}\frac{dT}{d\xi}=\frac{da_{3}}{d\xi}=&-i\varepsilon_{0}\omega_{3}\sum_{ijk}\chi^{(2)}_{ijk}E_{1,j}E_{2,k}\text{v}_{E,3,i}\\\
&-i\sum_{q}\omega_{q}{\bf
v}_{E,q}^{T}(\varepsilon_{q,d}-\varepsilon_{q,c}){\bf E}_{q},\end{split}$ (10)
where indices $q=1,2$ are for $\lambda_{1,2}$. This equation allows the
calculation of derivatives for any functions $T(a_{3})$ using a chain rule.
For example, for the maximization of SFG light power into a particular mode,
we can set $T=\absolutevalue{a_{3}}^{2}$, and obtain
$\frac{d|a_{3}|^{2}}{d\xi}=2\,\text{Re}\left\\{a_{3}^{\ast}\frac{da_{3}}{d\xi}\right\\}.$
(11)
In total, a set of six simulations is required to calculate the gradient using
the above equations for an arbitrarily large number of spatial positions
determined by the computational grid: three forward simulations to calculate
${\bf E}_{q}$, and three adjoint simulations for ${\bf v}_{E,q}$ at the three
non-degenerate wavelengths $\lambda_{q}$ with $q=1,2,3$, respectively.
### 2.2 Optimization for Multiple Polarizations in SFG
We now discuss the methodology for multi-objective optimization by extending
the results in the previous section that were formulated assuming fixed input
waves ${\bf E}_{1}$ and ${\bf E}_{2}$. Of particular interest is to
simultaneously tailor the sum-frequency polarization states for multiple
combinations of different input polarizations. Then, we can define a figure of
merit as
${\cal F}\left(\\{a_{3}^{(m)}({\bf E}_{1}^{(m)},{\bf
E}_{2}^{(m)})\\}_{m=1,\ldots,M}\right),$ (12)
where $m$ enumerates different input wave combinations and $a_{3}^{(m)}$ are
the sum-frequency amplitudes of the chosen polarization and spatial mode
profiles. Then, the derivative can be found using a chain rule
$\begin{split}\frac{d{\cal F}}{d\xi}=&\sum_{m=1}^{M}\frac{\partial{\cal
F}}{\partial a_{3}^{(m)}}\frac{da_{3}^{(m)}({\bf E}_{1}^{(m)},{\bf
E}_{2}^{(m)})}{d\xi}\\\ &+\sum_{m=1}^{M}\frac{\partial{\cal F}}{\partial
a_{3}^{\ast(m)}}\frac{da_{3}^{\ast(m)}({\bf E}_{1}^{(m)},{\bf
E}_{2}^{(m)})}{d\xi},\end{split}$ (13)
where the derivatives on the right-hand-side are determined using the adjoint
formulation in Eq. (10).
In the most general case, the input polarizations for $\lambda_{1,2}$ can each
be decomposed into pairs of orthogonal polarization states. Each of $M=4$
combinations of input waves can generate a different nonlinear current. Then,
a total of 14 unique simulations are required for the material derivative to
fully capture all possible input and output polarization combinations,
including 8 forward (2 at $\lambda_{1,2}$ and 4 at $\lambda_{3}$) and 6
adjoint (2 at each wavelength) simulations.
Furthermore, our method can also optimize diffraction outputs by specifying
the adjoint source waves' $k$-vectors. Diffraction into particular orders can
be enhanced (suppressed) by increasing (decreasing) the corresponding mode
overlap $a_{3}$, respectively.
### 2.3 Numerical implementation
We iteratively update the function $\xi(\mathbf{r})$ at all spatial locations
inside the quadratically nonlinear material. At each iteration, the material
permittivity is updated via gradient descent, or another gradient-based
method, at each discretized point in the domain [44, 48]. For a single-layer
metasurface design, we consider the pattern to be independent of the
longitudinal coordinate $z$. The optimization concludes when the FOM no longer
increases from one iteration to the next or the set maximum number of
iterations has been reached.
In our optimization, we employ several different techniques to ensure that the
free-form structures converge to a state that is both physical and realisable.
We introduce an increasing binarisation factor as the optimization progresses
that eventually forces the final pattern to be either material or air [49,
50]. The patterns are also periodically blurred with a Gaussian filter that
ensures the minimum feature is larger than limitations imposed by fabrication
precision [51, 52]. Importantly, an artificial absorption coefficient is added
onto the material, which will prevent convergence to structures that have
arbitrarily large quality factors (Q factor) [47, 53]. This non-radiative
decay rate defines the finite bandwidth of the SFG process, which is a
practical consideration for future experimental verification and distinguishes
our optimisation algorithms from other works purely relying on high-quality
factor resonances. This artificial absorption coefficient is later removed for
the analysis of the actual metasurfaces.
## 3 Results
Fig. 3: (a) Schematic of an SFG polarizing metasurface. The state of
$\mathbf{E}_{2}$ is fixed, while the state for $\mathbf{E}_{1}$ is
unpolarized. The SFG light from the optimized metasurface is polarized in the
$\ket{V}$, with the higher diffraction orders also being suppressed. (b)
optimized pattern of the metasurface. (c) Poincaré sphere representation of
all input states of $\mathbf{E}_{1}$ being transformed into the same SFG
output ($S_{1}$, corresponding to $\ket{H}$) state. (d) SFG efficiency from an
unpolarized source for different diffraction orders. (e) Signal wavelengths
sweep of SFG efficiency spectra for a pump wavelength fixed at
$1350\text{\,}\mathrm{nm}$, and (f) pump wavelengths sweep of SFG efficiency
spectra for a signal wavelength fixed at $1550\text{\,}\mathrm{nm}$.
In general, we can optimize for any combination of polarizations and elements
in the effective $\chi^{(2)}$ tensor, as discussed above. However, in this
work, we tackle a simpler problem that nevertheless still highlights the
strength of our optimization scheme. Specifically, we consider a signal having
any transverse polarization while the pump has a fixed polarization. Such
functionality may be beneficial for the operation of upconversion infrared
imaging, where the image can have any arbitrary polarization and the pump is a
light source of fixed polarization. For our examples, we have a signal
wavelength $\lambda_{1}=$1550\text{\,}\mathrm{nm}$$ and a pump wavelength
$\lambda_{2}=$1350\text{\,}\mathrm{nm}$$ (resulting in $\lambda_{3}=721.6$
nm). The transformation of the signal state into the SFG state can be
expressed in the form
$\mathbf{E}_{3}=\mathbf{M}\quantity(\mathbf{E}_{2},\chi^{(2)},\varepsilon)\cdot\mathbf{E}_{1}.$
(14)
The fields $\mathbf{E}_{q}$ represent the orthogonal complex amplitudes
$\quantity(E_{q}^{x},E_{q}^{y})$ in the transverse components basis. We can
make a connection of Eq. (14) to the Jones vectors in polarization optics in
the following way. Let $\mathbf{M}$ represent the nonlinear scattering matrix
analogous to the Jones matrix in linear materials. Individual elements of
$\mathbf{M}$ can be interpreted as complex scattering amplitudes. For an
unpatterned film, $\mathbf{M}$ is constrained by the crystal orientation and
is set by the elements of $\chi^{(2)}$ and the input state of
$\mathbf{E}_{2}$. One such case is provided in Supplementary S3.1 for
unpatterned nonlinear film. Then $\mathbf{E}_{1}$ and $\mathbf{E}_{3}$ can be
interpreted as the input and output Jones vectors of the system, respectively.
Now, with the ability to pattern structures into the nonlinear film, the
permittivity $\varepsilon$ is no longer uniform across the domain but instead
can be engineered to achieve the desired transformation of $\mathbf{M}$.
In all the examples, the nonlinear material is indium gallium phosphide
(InGaP) that is of (100) crystalline orientation and $300\text{\,}\mathrm{nm}$
thick. The film is resting on the fused silica substrate, with a
$$900\text{\,}\mathrm{nm}$\times$900\text{\,}\mathrm{nm}$$ unit cell. See
Supplementary S2 for the details regarding InGaP parameters. The
electromagnetic simulations are performed using the commercial COMSOL
Multiphysics software package suite. Each simulation takes approximately 100
seconds when using a discretization of $40\text{\,}\mathrm{nm}$. The resulting
fields are exported to the MATLAB programming language, where we implement our
inverse-design optimization.
### 3.1 Polarizing Nonlinear Metasurface
At the SFG wavelength, multiple diffraction orders exist due to the
periodicity of the metasurface. In infrared imaging applications, the higher
diffraction orders should be ideally suppressed so that the majority of the
SFG light is propagating in the zeroth order. In previous works [5, 30], it
has been a challenge to suppress these higher-order propagating modes.
Simultaneously, with our method, we can also optimize diffraction outputs by
specifying the adjoint source waves' $k$-vectors. This is a significant
advance compared to previous inverse-design approaches for quadratically
nonlinear metasurfaces [35, 38] where only total second-harmonic conversion
but not directionality could be optimized.
We denote the zeroth order of the scattering matrix as $\mathbf{M}_{0}$, whose
elements can be optimized individually. In the first example, the signal is an
unpolarized state while the pump at $\lambda_{2}$ is polarized in the
$\ket{V}$ state. We intend for the SFG zeroth order diffraction to also be
polarized in the $\ket{V}$ state [Fig. 3(a)]. In this scheme, the metasurface
can be considered to be a nonlinear polarizer by taking an unpolarized input
and polarizing it into the $\ket{V}$ state at the SFG output. The figure of
merit is formulated as ${\cal
F}=\frac{1}{2}\quantity(\absolutevalue{\bra{V_{3}}\mathbf{M}_{0}\ket{H_{1}}}^{2}+\absolutevalue{\bra{V_{3}}\mathbf{M}_{0}\ket{V_{1}}}^{2})$,
and its derivative is calculated with Eq. (13). We note that unpolarized light
can be decomposed into equal powers of any pair of orthogonal states, which
are the $\ket{V}$ and $\ket{H}$ pair in our modelling. In this notation, the
states $\ket{\psi_{1}}$ denote the input signal state, while $\bra{\phi_{3}}$
denote the outgoing SFG state.
The optimization converges to a design that is highly non-trivial, as shown in
Fig. 3(b). We calculate the expected transformation of input unpolarized light
and find that it is almost fully converted into the $\ket{V}$ state at the SFG
wavelength, as depicted on the Poincaré sphere in Fig. 3(c). The numerical
values of $\mathbf{M}_{0}$ matrix elements can be found in Supplementary S3.2,
where further analysis is also provided. In Fig. 3(d), we show the predicted
SFG conversion efficiency into all the propagating diffraction orders, of
which the zeroth order comprises approximately 80% of the total SFG light (see
Supplementary S4.1 for precise values). Therefore, the optimized metasurface
directs the vast majority of SFG light perpendicular to the surface, which
greatly benefits imaging applications.
We now determine the frequency bandwidth by fixing the wavelength of the pump
at $\lambda_{2}=$1350\text{\,}\mathrm{nm}$$ and calculating the SFG efficiency
as an unpolarized signal wavelength is swept around the operating wavelength
of $\lambda_{1}=$1550\text{\,}\mathrm{nm}$$ [Fig. 3(e)]. The transmission into
the desired $\ket{V}$ state is significantly larger than the $\ket{H}$ state
and reaches a maximum SFG efficiency of
$0.25\text{\,}{\mathrm{cm}}^{2}\text{\,}{\mathrm{GW}}^{-1}$, which is more
than 3 orders of magnitude larger than an unpatterned film of the same
nonlinear material and thickness. This performance is echoed when we fix the
signal $\lambda_{1}=$1550\text{\,}\mathrm{nm}$$ and sweep the pump wavelengths
instead [Fig. 3(f)]. The full-width at half maximum (FWHM) conversion
efficiency for the signal is approximately $20\text{\,}\mathrm{nm}$, while it
is approximately $3\text{\,}\mathrm{nm}$ for the pump, representing a
reasonably large operating bandwidth suitable for efficient conversion with
short optical pulses. From these plots, it is evident that there are
resonances present within the metasurfaces at the input wavelengths
$\lambda_{1}$ and $\lambda_{2}$ that greatly enhance the SFG process. We
provide the field distributions and further analysis in the Supplementary S5.
### 3.2 Polarization Independent Nonlinear Metasurface
Fig. 4: (a) Schematic of an SFG waveplate metasurface. The state of
$\mathbf{E}_{2}$ is fixed, while the state for $\mathbf{E}_{1}$ is
unpolarized. The SFG light from the optimized metasurface is rotated in the
$\ket{V}$, with the higher diffraction orders also being suppressed. (b)
optimized pattern of the metasurface. (c) Poincaré sphere representation of
all input states of $\mathbf{E}_{1}$ being transformed into different SFG
output states while preserving their orthogonality. (d) SFG efficiency from an
unpolarized source for different diffraction orders. (e) Signal wavelengths
sweep of SFG efficiency spectra for a pump wavelength fixed at
$1350\text{\,}\mathrm{nm}$, and (f) pump wavelengths sweep of SFG efficiency
spectra for a signal wavelength fixed at $1550\text{\,}\mathrm{nm}$. The
dashed orange curves show the ratio of singular values (SV) for each plot at
their respective wavelengths.
In this example, we focus on a metasurface whose SFG enhancement is
independent of the polarization of $\lambda_{1}$. A metasurface that has this
property can enhance SFG conversion efficiency for all input polarization
states equally. This is particularly useful for imaging, where the source is
typically unpolarized or partially polarized. Such an upconverted image will
preserve the relative amplitudes of the original image, even if there are
spatial variations in the polarization. This is in contrast to previous
demonstrations of SFG imaging, where the metasurfaces typically rely on
polarization-sensitive resonant modes of bound states in the continuum [54].
Therefore, for this case, the target $\mathbf{M}_{0}$ matrix is close to
unitary after scaling, or its singular values are close to equality. The FOM
for this case is
${\cal F}=s_{2}\,,$ (15)
where $s_{1}$ and $s_{2}$ are the ordered singular values of $\mathbf{M}_{0}$,
and for this demonstration, the pump polarization at $\lambda_{2}$ is fixed at
$\ket{V}$. By maximising $s_{2}$, which is always defined as the smaller of
the two singular values, we are effectively increasing the SFG enhancement of
the worst-performing polarization input state. For a unitary matrix, the ratio
of the singular values must be $s_{2}/s_{1}=1$.
The optimized metasurface is again a non-trivial free-form pattern, as shown
in Fig. 4(b). For this metasurface, we then calculate the transformation of
various input states at the signal wavelength and plot them on a Poincaré
sphere [Fig. 4(c)]. We see that the metasurface imparts a rotation of the
eigenstates to the input states during the SFG process. We provide further
analysis of the matrix $\mathbf{M}_{0}$ in Supplementary S3.3. Importantly,
the eigenstates are nearly orthogonal, which indicates that the transformation
is indeed near unitary after scaling. The SFG is primarily channelled into the
zeroth diffraction order for an unpolarized signal [Fig. 4(d)], with reduced
light leakage into higher order diffraction modes (see Supplementary S4.2 for
numerical values).
We calculate the SFG efficiency for a range of unpolarized input signal
wavelengths and pump wavelength fixed at
$\lambda_{2}=$1350\text{\,}\mathrm{nm}$$ and find a maximum in average
efficiency of
$6\text{\times}{10}^{-3}\text{\,}{\mathrm{cm}}^{2}\text{\,}{\mathrm{GW}}^{-1}$
[Fig. 4(e)]. Again in this example, the FWHM conversion efficiency for the
signal is approximately $20\text{\,}\mathrm{nm}$, while it is approximately
for $10\text{\,}\mathrm{nm}$ for the pump. The SFG efficiency for two
orthogonal polarizations of $\ket{H}$ and $\ket{V}$ are almost equal. On the
right axis, we show the ratio of singular values that reaches a maximum of
0.8, reasonably close to the ratio of 1 for a truly unitary transformation.
This analysis is repeated for a fixed signal ($\lambda_{1}$ =
$1550\text{\,}\mathrm{nm}$) and varying pump wavelengths [Fig. 4(f)], with the
maximum efficiency peaking at $\lambda_{2}$ = $1350\text{\,}\mathrm{nm}$, as
expected. Therefore, the transformation of input light into SFG output from
the metasurface can be considered to be nearly polarization-independent. The
preservation of polarization information leads to the ability to perform
upconversion polarimetric imaging [6] with greater efficiency than previously
reported.
### 3.3 Structure of Resonances in the Optimised Metasurfaces
Finally, we perform linear simulations to inspect the resonances that we
expect to be present in the metasurfaces at
$\lambda_{1},\lambda_{2},\lambda_{3}$ (Supplementary S5.1). Because the two
metasurfaces presented in this work in Sec. 3.1 and 3.2 above are optimized
for different functionalities, they also have different resonant
characteristics. For the nonlinear polarizer metasurface, only $\ket{H}$
produces a resonance at $\lambda_{1}$, while for the polarization-independent
nonlinear metasurface, resonances appear for both $\ket{H}$ and $\ket{V}$
close to $\lambda_{1}$. As expected for both nonlinear metasurfaces, only
$\ket{V}$ produces a resonance at the pump wavelength because the polarization
at $\lambda_{2}$ was fixed. We note that the $Q$-factors are on the order of
70 to 300 for the different wavelengths, which allows for a reasonably broad
imaging bandwidth and for SFG with short optical pulses. We also provide field
enhancement distributions for the metasurfaces at all the interacting
wavelengths in Supplementary S5.2.
## 4 Conclusions
We have developed a novel method of multi-objective optimization of nonlinear
frequency mixing processes in metasurfaces. Our method allows for the
simultaneous control of polarizations and directionality and maximizes
efficiency across a target bandwidth for SFG processes, which is beyond what
is possible with conventional design strategies. We present a computationally
efficient implementation based on adjoint formulation and demonstrate two
essential examples of metasurfaces that enhance SFG either for one signal
input polarization or for all input polarizations. In both cases, the SFG is
emitted in the forward direction while the higher diffraction orders are
suppressed.
This method can be naturally extended to optimize metasurfaces that explore
other nonlinear phenomena such as four-wave mixing, spontaneous parametric
down-conversion, and high harmonic generation with greater efficiency. In the
future, nonlinear metasurfaces that exhibit more complicated characteristics
will be enabled by sophisticated optimization algorithms, widening the
spectrum of potential functionalities.
This work was supported by the Australian Research Council (CE200100010).
All authors have accepted responsibility for the entire content of this
manuscript and approved its submission.
The authors state no conflict of interest.
The datasets generated during and/or analyzed during the current study are
available from the corresponding author upon reasonable request.
## References
* [1] Benjamin Doughty, Lu Lin, Uvinduni I Premadasa and Ying-Zhong Ma ``Considerations in upconversion: A practical guide to sum-frequency generation spectrometer design and implementation'' In _Biointerphases_ 17.2, 2022, pp. 021201 DOI: 10.1116/6.0001817
* [2] Saher Junaid, Jan Tomko, Mykhaylo P Semtsiv, Jan Kischkat, W Ted Masselink, Christian Pedersen and Peter Tidemand-Lichtenberg ``Mid-infrared upconversion based hyperspectral imaging'' In _Optics Express_ 26.3 Optica Publishing Group, 2018, pp. 2203–2211 DOI: 10.1364/OE.26.002203
* [3] Ashik A.., Callum F O’Donnell, S Chaitanya Kumar, M Ebrahim-Zadeh, P Tidemand-Lichtenberg and C Pedersen ``Mid-infrared upconversion imaging using femtosecond pulses'' In _Photonics Research_ 7.7 Optica Publishing Group, 2019, pp. 783–791 DOI: 10.1364/PRJ.7.000783
* [4] Adam M Hanninen, Richard C Prince, Raul Ramos, Maksim V Plikus and Eric O Potma ``High-resolution infrared imaging of biological samples with third-order sum-frequency generation microscopy'' In _Biomedical Optics Express_ 9.10 Optica Publishing Group, 2018, pp. 4807–4817 DOI: 10.1364/BOE.9.004807
* [5] Maria del Rocio Camacho-Morales, Davide Rocco, Lei Xu, Valerio F Gili, Nikolay Dimitrov, Lyubomir Stoyanov, Zhonghua Ma, Andrei Komar, Mykhaylo Lysevych, Fouad Karouta, Alexander A Dreischuh, Hark Hoe H Tan, Giuseppe Leo, Costantino De Angelis, Chennupati Jagadish, Andrey E Miroshnichenko, Mohsen Rahmani and Dragomir N Neshev ``Infrared upconversion imaging in nonlinear metasurfaces'' In _Advanced Photonics_ 3.3, 2021, pp. 036002 DOI: 10.1117/1.AP.3.3.036002
* [6] Zhanghang Zhu, Di Zhang, Fei Xie, Junjun Ma, Jiaxin Chen, Shengchao Gong, Wei Wu, Wei Cai, Xinzheng Zhang, Mengxin Ren and Jingjun Xu ``Nonlinear polarization imaging by parametric upconversion'' In _Optica_ 9.11 Optica Publishing Group, 2022, pp. 1297–1302 DOI: 10.1364/OPTICA.471177
* [7] Kirill L Koshelev, Pavel Tonkaev and Yuri S Kivshar ``Nonlinear chiral metaphotonics: a perspective'' In _Advanced Photonics_ 5.6, 2023, pp. 064001 DOI: 10.1117/1.AP.5.6.064001
* [8] ``Nonlinear Meta-Optics'' London: CRC Press, 2020 DOI: 10.1201/b22515
* [9] Gustavo Grinblat ``Nonlinear Dielectric Nanoantennas and Metasurfaces: Frequency Conversion and Wavefront Control'' In _ACS Photonics_ 8.12 American Chemical Society, 2021, pp. 3406–3432 DOI: 10.1021/acsphotonics.1c01356
* [10] L.. Huang, L. Xu, D.. Powell, W.. Padilla and A.. Miroshnichenko ``Resonant leaky modes in all-dielectric metasystems: Fundamentals and applications'' In _Phys. Rep._ 1008, 2023, pp. 1–66 DOI: 10.1016/j.physrep.2023.01.001
* [11] Arseniy I. Kuznetsov, Mark L. Brongersma, Jin Yao, Mu Ku Chen, Uriel Levy, Din Ping Tsai, Nikolay I. Zheludev, Andrei Faraon, Amir Arbabi, Nanfang Yu, Debashis Chanda, Kenneth B. Crozier, Alexander V. Kildishev, Hao Wang, Joel K.. Yang, Jason G. Valentine, Patrice Genevet, Jonathan A. Fan, Owen D. Miller, Arka Majumdar, Johannes E. Fröch, David Brady, Felix Heide, Ashok Veeraraghavan, Nader Engheta, Andrea Alù, Albert Polman, Harry A. Atwater, Prachi Thureja, Ramon Paniagua-Dominguez, Son Tung Ha, Angela I. Barreda, Jon A. Schuller, Isabelle Staude, Gustavo Grinblat, Yuri Kivshar, Samuel Peana, Susanne F. Yelin, Alexander Senichev, Vladimir M. Shalaev, Soham Saha, Alexandra Boltasseva, Junsuk Rho, Dong Kyo Oh, Joohoon Kim, Junghyun Park, Robert Devlin and Ragip A. Pala ``Roadmap for Optical Metasurfaces'' In _ACS Photonics_ , 2024 DOI: 10.1021/acsphotonics.3c00457
* [12] Giuseppe Marino, Carlo Gigli, Davide Rocco, Aristide Lemaître, Ivan Favero, Costantino De Angelis and Giuseppe Leo ``Zero-Order Second Harmonic Generation from AlGaAs-on-Insulator Metasurfaces'' In _ACS Photonics_ 6.5 American Chemical Society, 2019, pp. 1226–1231 DOI: 10.1021/acsphotonics.9b00110
* [13] Yuri Kivshar and Andrey Miroshnichenko ``Meta-Optics with Mie Resonances'' In _Optics and Photonics News_ 28.1 Optica Publishing Group, 2017, pp. 24–31 DOI: 10.1364/OPN.28.1.000024
* [14] Elizaveta V Melik-Gaykazyan, Sergey S Kruk, Rocio Camacho-Morales, Lei Xu, Mohsen Rahmani, Khosro Zangeneh Kamali, Aristeidis Lamprianidis, Andrey E Miroshnichenko, Andrey A Fedyanin, Dragomir N Neshev and Yuri S Kivshar ``Selective Third-Harmonic Generation by Structured Light in Mie-Resonant Nanoparticles'' In _ACS Photonics_ 5.3 American Chemical Society, 2018, pp. 728–733 DOI: 10.1021/acsphotonics.7b01277
* [15] Rocio Camacho-Morales, Mohsen Rahmani, Sergey Kruk, Lei Wang, Lei Xu, Daria A Smirnova, Alexander S Solntsev, Andrey Miroshnichenko, Hark Hoe Tan and Fouad Karouta ``Nonlinear generation of vector beams from AlGaAs nanoantennas'' In _Nano letters_ 16.11 ACS Publications, 2016, pp. 7191–7197
* [16] Daria Smirnova, Alexander I Smirnov and Yuri S Kivshar ``Multipolar second-harmonic generation by Mie-resonant dielectric nanoparticles'' In _Physical Review A_ 97.1 American Physical Society, 2018, pp. 13807 DOI: 10.1103/PhysRevA.97.013807
* [17] Kristina Frizyuk, Irina Volkovskaya, Daria Smirnova, Alexander Poddubny and Mihail Petrov ``Second-harmonic generation in Mie-resonant dielectric nanoparticles made of noncentrosymmetric materials'' In _Physical Review B_ 99.7 American Physical Society, 2019, pp. 75425 DOI: 10.1103/PhysRevB.99.075425
* [18] Rocio Camacho-Morales, Mohsen Rahmani, Sergey Kruk, Lei Wang, Lei Xu, Daria A Smirnova, Alexander S Solntsev, Andrey Miroshnichenko, Hark Hoe Tan, Fouad Karouta, Shagufta Naureen, Kaushal Vora, Luca Carletti, Costantino De Angelis, Chennupati Jagadish, Yuri S Kivshar and Dragomir N Neshev ``Nonlinear Generation of Vector Beams From AlGaAs Nanoantennas'' In _Nano Letters_ 16.11 American Chemical Society, 2016, pp. 7191–7197 DOI: 10.1021/acs.nanolett.6b03525
* [19] Polina P Vabishchevich, Sheng Liu, Michael B Sinclair, Gordon A Keeler, Gregory M Peake and Igal Brener ``Enhanced Second-Harmonic Generation Using Broken Symmetry III–V Semiconductor Fano Metasurfaces'' In _ACS Photonics_ 5.5 American Chemical Society, 2018, pp. 1685–1690 DOI: 10.1021/acsphotonics.7b01478
* [20] Luca Carletti, Sergey S Kruk, Andrey A Bogdanov, Costantino De Angelis and Yuri Kivshar ``High-harmonic generation at the nanoscale boosted by bound states in the continuum'' In _Physical Review Research_ 1.2 American Physical Society, 2019, pp. 23016 DOI: 10.1103/PhysRevResearch.1.023016
* [21] Aravind P Anthur, Haizhong Zhang, Ramon Paniagua-Dominguez, Dmitry A Kalashnikov, Son Tung Ha, Tobias W W Maß, Arseniy I Kuznetsov and Leonid Krivitsky ``Continuous Wave Second Harmonic Generation Enabled by Quasi-Bound-States in the Continuum on Gallium Phosphide Metasurfaces'' In _Nano Letters_ 20.12 American Chemical Society, 2020, pp. 8745–8751 DOI: 10.1021/acs.nanolett.0c03601
* [22] C.. Fang, Q.. Yang, Q.. Yuan, L.. Gu, X.. Gan, Y. Shao, Y. Liu, G.. Han and Y. Hao ``Efficient Second-Harmonic Generation from Silicon Slotted Nanocubes with Bound States in the Continuum'' In _Laser Photon. Rev._ 16.5, 2022, pp. 2100498 DOI: 10.1002/lpor.202100498
* [23] George Zograf, Kirill Koshelev, Anastasiia Zalogina, Viacheslav Korolev, Richard Hollinger, Duk-Yong Choi, Michael Zuerch, Christian Spielmann, Barry Luther-Davies, Daniil Kartashov, Sergey V Makarov, Sergey S Kruk and Yuri Kivshar ``High-Harmonic Generation from Resonant Dielectric Metasurfaces Empowered by Bound States in the Continuum'' In _ACS Photonics_ 9.2 American Chemical Society, 2022, pp. 567–574 DOI: 10.1021/acsphotonics.1c01511
* [24] X.. Zhang, L.. He, X. Gan, X.. Huang, Y.. Du, Z.. Zhai, Z. Li, Y.. Zheng, X.. Chen, Y.. Cai and X.. Ao ``Quasi-Bound States in the Continuum Enhanced Second-Harmonic Generation in Thin-Film Lithium Niobate'' In _Laser Photon. Rev._ 16.9, 2022, pp. 2200031 DOI: 10.1002/lpor.202200031
* [25] Z. Zheng, L. Xu, L.. Huang, D. Smirnova, P.. Hong, C.. Ying and M. Rahmani ``Boosting second-harmonic generation in the LiNbO3 metasurface using high-Q guided resonances and bound states in the continuum'' In _Phys. Rev. B_ 106.12, 2022, pp. 125411 DOI: 10.1103/PhysRevB.106.125411
* [26] Radoslaw Kolkowski, Tommi K Hakala, Andriy Shevchenko and Mikko J Huttunen ``Nonlinear nonlocal metasurfaces'' In _Applied Physics Letters_ 122.16, 2023, pp. 160502 DOI: 10.1063/5.0140483
* [27] Hui Jiang, Kaili Sun, Yuechen Jia, Yangjian Cai, Uriel Levy and Zhanghua Han ``Tunable Second Harmonic Generation with Large Enhancement in A Nonlocal All-Dielectric Metasurface Over A Broad Spectral Range'' In _Advanced Optical Materials_ n/a.n/a John Wiley & Sons, Ltd, 2024, pp. 2303229 DOI: https://doi.org/10.1002/adom.202303229
* [28] Sheng Liu, Polina P Vabishchevich, Aleksandr Vaskin, John L Reno, Gordon A Keeler, Michael B Sinclair, Isabelle Staude and Igal Brener ``An all-dielectric metasurface as a broadband optical frequency mixer'' In _Nature Communications_ 9.1, 2018, pp. 2507 DOI: 10.1038/s41467-018-04944-9
* [29] Qingchen Yuan, Liang Fang, Hanlin Fang, Juntao Li, Tao Wang, Wanqi Jie, Jianlin Zhao and Xuetao Gan ``Second Harmonic and Sum-Frequency Generations from a Silicon Metasurface Integrated with a Two-Dimensional Material'' In _ACS Photonics_ 6.9 American Chemical Society, 2019, pp. 2252–2259 DOI: 10.1021/acsphotonics.9b00553
* [30] Rocio Camacho-Morales, Lei Xu, Haizhong Zhang, Son Tung Ha, Leonid Krivitsky, Arseniy I Kuznetsov, Mohsen Rahmani and Dragomir Neshev ``Sum-Frequency Generation in High-Q GaP Metasurfaces Driven by Leaky-Wave Guided Modes'' In _Nano Letters_ 22.15 American Chemical Society, 2022, pp. 6141–6148 DOI: 10.1021/acs.nanolett.2c01349
* [31] Davide Rocco, Rocio Camacho Morales, Lei Xu, Attilio Zilli, Vincent Vinel, Marco Finazzi, Michele Celebrano, Giuseppe Leo, Mohsen Rahmani, Chennupati Jagadish, Hoe Tan, Dragomir Neshev and Costantino De Angelis ``Second order nonlinear frequency generation at the nanoscale in dielectric platforms'' In _Advances in Physics: X_ 7.1 Taylor & Francis, 2022, pp. 2022992 DOI: 10.1080/23746149.2021.2022992
* [32] D Rocco, C Gigli, L Carletti, G Marino, M A Vincenti, G Leo and C De Angelis ``Vertical Second Harmonic Generation in Asymmetric Dielectric Nanoantennas'' In _IEEE Photonics Journal_ 12.3, 2020, pp. 1–7 DOI: 10.1109/JPHOT.2020.2988502
* [33] Carlo Gigli, Giuseppe Marino, Alberto Artioli, Davide Rocco, Costantino De Angelis, Julien Claudon, Jean-Michel Gérard and Giuseppe Leo ``Tensorial phase control in nonlinear meta-optics'' In _Optica_ 8.2 Optica Publishing Group, 2021, pp. 269–276 DOI: 10.1364/OPTICA.413329
* [34] Lei Xu, Mohsen Rahmani, Yixuan Ma, Daria A Smirnova, Khosro Zangeneh Kamali, Fu Deng, Yan Kei Chiang, Lujun Huang, Haoyang Zhang, Stephen Gould, Dragomir N Neshev and Andrey E Miroshnichenko ``Enhanced light–matter interactions in dielectric nanostructures via machine-learning approach'' In _Advanced Photonics_ 2.2, 2020, pp. 026003 DOI: 10.1117/1.AP.2.2.026003
* [35] C. Sitawarin, W.. Jin, Z. Lin and A.. Rodriguez ``Inverse-designed photonic fibers and metasurfaces for nonlinear frequency conversion [Invited]'' In _Phot. Res._ 6.5, 2018, pp. B82–B89 DOI: 10.1364/PRJ.6.000B82
* [36] Tyler W. Hughes, Momchil Minkov, Ian A D Williamson and Shanhui Fan ``Adjoint Method and Inverse Design for Nonlinear Nanophotonic Devices'' In _ACS Photonics_ 5.12 American Chemical Society, 2018, pp. 4781–4787 DOI: 10.1021/acsphotonics.8b01522
* [37] Lakshmi Raju, Kyu-Tae Lee, Zhaocheng Liu, Dayu Zhu, Muliang Zhu, Ekaterina Poutrina, Augustine Urbas and Wenshan Cai ``Maximized Frequency Doubling through the Inverse Design of Nonlinear Metamaterials'' In _ACS Nano_ 16.3 American Chemical Society, 2022, pp. 3926–3933 DOI: 10.1021/acsnano.1c09298
* [38] Sander A Mann, Heedong Goh and Andrea Alù ``Inverse Design of Nonlinear Polaritonic Metasurfaces for Second Harmonic Generation'' In _ACS Photonics_ 10.4 American Chemical Society, 2023, pp. 993–1000 DOI: 10.1021/acsphotonics.2c01342
* [39] Mark Hillery ``An Introduction to the Quantum Theory of Nonlinear Optics'', 2009 DOI: 10.2478/v10155-010-0094-8
* [40] R.. Boyd ``Nonlinear Optics'' San Diego: Academic Press, 2020 DOI: 10.1016/C2015-0-05510-1
* [41] J.. Jensen and O. Sigmund ``Topology optimization for nano-photonics'' In _Laser Photon. Rev._ 5.2, 2011, pp. 308–321 DOI: 10.1002/lpor.201000014
* [42] C.. Lalau-Keraly, S. Bhargava, O.. Miller and E. Yablonovitch ``Adjoint shape optimization applied to electromagnetic design'' In _Opt. Express_ 21.18, 2013, pp. 21693–21701 DOI: 10.1364/OE.21.021693
* [43] A… Niederberger, D.. Fattal, N.. Gauger, S.. Fan and R.. Beausoleil ``Sensitivity analysis and optimization of sub-wavelength optical gratings using adjoints'' In _Opt. Express_ 22.11, 2014, pp. 12971–12981 DOI: 10.1364/OE.22.012971
* [44] S. Molesky, Z. Lin, A.. Piggott, W.. Jin, J. Vučković and A.. Rodriguez ``Inverse design in nanophotonics'' In _Nat. Photon._ 12.11, 2018, pp. 659–670 DOI: 10.1038/s41566-018-0246-9
* [45] R.. Christiansen, J. Michon, M. Benzaouia, O. Sigmund and S.. Johnson ``Inverse design of nanoparticles for enhanced Raman scattering'' In _Opt. Express_ 28.4, 2020, pp. 4444–4462 DOI: 10.1364/OE.28.004444
* [46] R.. Christiansen and O. Sigmund ``Inverse design in photonics by topology optimization: tutorial'' In _J. Opt. Soc. Am. B_ 38.2, 2021, pp. 496–509 DOI: 10.1364/JOSAB.406048
* [47] Z. Lin, X.. Liang, M. Loncar, S.. Johnson and A.. Rodriguez ``Cavity-enhanced second-harmonic generation via nonlinear-overlap optimization'' In _Optica_ 3.3, 2016, pp. 233–238 DOI: 10.1364/OPTICA.3.000233
* [48] M… Elsawy, S. Lanteri, R. Duvigneau, J.. Fan and P. Genevet ``Numerical Optimization Methods for Metasurfaces'' In _Laser Photon. Rev._ 14.10, 2020, pp. 1900445 DOI: 10.1002/lpor.201900445
* [49] Conner Ballew, Gregory Roberts, Tianzhe Zheng and Andrei Faraon ``Constraining Continuous Topology Optimizations to Discrete Solutions for Photonic Applications'' In _ACS Photonics_ 10.4 American Chemical Society, 2023, pp. 836–844 DOI: 10.1021/acsphotonics.2c00862
* [50] Zhaoyi Li, Raphaël Pestourie, Zin Lin, Steven G Johnson and Federico Capasso ``Empowering Metasurfaces with Inverse Design: Principles and Applications'' In _ACS Photonics_ 9.7 American Chemical Society, 2022, pp. 2178–2192 DOI: 10.1021/acsphotonics.1c01850
* [51] Evan W Wang, David Sell, Thaibao Phan and Jonathan A Fan ``Robust design of topology-optimized metasurfaces'' In _Optical Materials Express_ 9.2 Optica Publishing Group, 2019, pp. 469–482 DOI: 10.1364/OME.9.000469
* [52] Soumyashree S Panda, Hardik S Vyas and Ravi S Hegde ``Robust inverse design of all-dielectric metasurface transmission-mode color filters'' In _Optical Materials Express_ 10.12 Optica Publishing Group, 2020, pp. 3145–3159 DOI: 10.1364/OME.409186
* [53] Xiangdong Liang and Steven G Johnson ``Formulation for scalable optimization of microcavities via the frequency-averaged local density of states'' In _Optics Express_ 21.25 Optica Publishing Group, 2013, pp. 30812–30841 DOI: 10.1364/OE.21.030812
* [54] R. Camacho-Morales, D. Rocco, L. Xu, V.. Gili, N. Dimitrov, L. Stoyanov, Z.. Ma, A. Komar, M. Lysevych, F. Karouta, A. Dreischuh, H.. Tan, G. Leo, C. De Angelis, C. Jagadish, A.. Miroshnichenko, M. Rahmani and D.. Neshev ``Infrared upconversion imaging in nonlinear metasurfaces'' In _Adv. Photon._ 3.3, 2021, pp. 036002 DOI: 10.1117/1.AP.3.3.036002
|
# Magnetically Induced Schrödinger Cat States: The Shadow of a Quantum Space
Partha Nandia<EMAIL_ADDRESS>Nandita Debnathb<EMAIL_ADDRESS>Subhajit Kalac<EMAIL_ADDRESS>A. S. Majumdard<EMAIL_ADDRESS>aInstitute
of Theoretical Physics, University of Stellenbosch, Stellenbosch-7600, South
Africa.
bDepartment of Physics, University of Calcutta, Kolkata 700009, India.
cDepartment of Physics, Indian Institute of Technology Guwahati, Guwahati
781039, Assam, India.
dS. N. Bose National Centre for Basic Sciences, JD Block, Sector III, Salt
Lake, Kolkata 700106, India.
###### Abstract
Schrödinger cat states, which are superpositions of macroscopically distinct
states, are potentially critical resources for upcoming quantum information
technologies. In this paper, we introduce a scheme to generate entangled
Schrödinger cat states in a non-relativistic electric dipole system situated
on a two-dimensional plane, along with an external potential and a uniform
strong magnetic field perpendicular to the plane. Additionally, our findings
demonstrate that this setup can lead to the phenomenon of collapse and revival
of entanglement for a specific range of our model parameters.
## I Introduction
In quantum theory, the transition between the microscopic and macroscopic
worlds is one of the less-understood features Zuk . Such a transition plays a
direct role in the realm of quantum measurements. In an ideal measurement
paradigm, the interaction of macroscopic equipment and a microscopic system
yields entanglement and a superposed quantum state with both macroscopic and
microscopic components pn2 . Schrödinger was the first to highlight the
physical subtleties of this kind of superposition by replacing the macroscopic
part of the system by a “cat”, in order to illustrate a dramatic superposition
of “states” of both alive and dead cats, that should, in practice, be
distinguished macroscopically par . The superposition of macroscopically
different quantum states, generically referred to as non-classical Schrödinger
Cat State (SCS) par ; 2 ; bose , is crucial for understanding the conceptual
underpinnings of quantum physics, especially with reference to wave function
collapse models pnr ; pr ; hr ; hr1 . In recent years, the advancement of
quantum technologies has brought into sharp focus the utility of several
quantum phenomena such as photon anti-bunching Rov , sub-Poissonian statistics
mar and squeezing mtw , along with the dynamics of SCS.
The success of quantum information theory and its potential applications
qinform ; qinf that significantly outperform their classical equivalents have
recently sparked a renewed interest in the generation of non-classical states
such as SCS. Several applications of cat states have been suggested in the
realm of quantum information qinfor , quantum metrology qmeter , quantum
teleportation qteleport , and quantum error correction schemes qerror1 ;
qerror2 . Besides, the concept of decoherence between two superposed quantum
objects, or the quantum-to-classical transition, can be studied using the SCS
as a platform. In quantum optics, a superposition of two diametrically
opposite coherent states $|\pm\alpha>$ with large value of $\mid\alpha\mid$,
can be interpreted as a quantum superposition of two macroscopically distinct
states, i.e., a Schrödinger cat-like state tw ; legg . However, due to decay
of their interference properties, it is extremely difficult to detect such
states in practice egg . Nonetheless, the universality of SCS enables it to be
realized in a wide variety of physical arenas such as nonlinear quantum optics
Sb , quantum dot systems kaku , superconducting cavities kau , Bose Einstein
condensates (BEC) goenner and quantization of weak gravity rbm ; wolf ; sch .
A fascinating direction of research in recent years has been the mechanism for
the natural generation of SCS in some specific condensed matter systems sch1 ;
sch2 .
Schrödinger cat states with entanglement based protocols provide a novel
technique to explore short-distance quantum physics in a non-relativistic
domain when there is a magnetic dipole interaction background ed . At
extremely short distances, the space-time structure needs to be “granular” in
order to account for both gravity and quantum uncertainty ein . A viable
approach towards quantum gravity is through quantizing space-time itself in ,
rather than the construction of an effective field theory of gravity. This
approach is an active area of research on quantum gravity, commonly referred
to as non-commutative geometry go ; pe . The fundamental goal is to derive
classical geometry from a suitable limit of a non-commutative algebra. Though
such a proposal may appear as ad-hoc pek , the physical justification for such
a non-commutative space-time is strong since it provides a solution to the
geometric measurement problem near the Planck scale.
Non-commutative geometry appears naturally in various non-relativistic planar
systems. For instance, it occurs using the lowest Landau-level (LLL)
projection to study the behavior of charged particles in a strong magnetic
field ek . Further, the incompressibility of fractional quantum Hall fluids
qhefluids has a strong connection to the emergence of a non-commutative
geometry in which the fundamental Planck length is substituted by the magnetic
length. Non-commutative space-time forms an alternative paradigm for studying
the behavior of relativistic anyonic systems in interaction with the ambient
electromagnetic field vpn ; Rabin . Additionally, non-commutative properties
of real-space coordinates in the presence of the Berry curvature k produce
skew scattering by a non-magnetic impurity without relativistic spin-orbit
interactions in a condensed matter system. Non-commutative space provides a
paradigm for describing the behavior of the quantum to classical transition
under the influence of decoherence fg1 ; fg2 , which is relevant for
implementation of quantum information protocols. From an experimental
standpoint, there have been efforts in search of evidence of possible non-
commutative effect manifestations in cosmology and high-energy physics kk ; c
; d . A testable framework has been suggested in low-energy experiments in the
arena of quantum Hall effect qhe ; e .
The motivation for the present study is to investigate whether multi-component
entangled non-classical SCS could be produced in deformed quantum space, where
non-commutativity arises naturally in an easily accessibly low energy physical
system. In this article, we investigate the phenomenology of a two-particle
electric dipole model with an additional harmonic interaction and a strong
background magnetic field, with motion constrained to the plane perpendicular
to the field. Such a system may be considered as a toy version of a real
Excitonic dipole set-up exci . By exploring the high magnetic field limit, we
reveal the emergence of planar non-commutative space as a natural consequence.
Furthermore, we establish the deformed Heisenberg algebra as the origin of
multi-component entangled SCS in this system. Moreover, we quantify the degree
of entanglement of our SCS, and show that the phemomenon of collapse and
revival of entanglement R1 ; yueberly ; R2 occurs in this system under the
influence of the harmonic potential.
The organization of our paper is as follows. The interacting two particle
electric dipole system is introduced in Section 2, showing how classical non-
commutative space appears in the presence of a very strong, constant, uniform
magnetic field. Then, in Section 3, we move on to the quantum picture, where
intricacies of the system dynamics are revealed, in context of mapping between
two reference frames. Section 4 discusses how our model with a harmonic
oscillator potential that is dependent only on one spatial variable is able to
generate entangled multi-component Schrödinger cat states. In Section 5, we
compute the degree of entanglement in the generated SCS system and demonstrate
that it exhibits the phenomenon of entanglement collapse and revival. Section
6 is reserved for concluding remarks and discussions.
## II Two-Particle System: Classical picture
We begin by considering a pair of non-relativistic, oppositely charged
particles with equal mass $m$ moving on the plane subjected to a constant
magnetic field $B$ along the $z$ axis (ignoring Coulomb and radiation
effects). In component form, $x_{i}$ and $y_{i}$ $(i=1,2)$ correspondingly
represents the positive and negative charge coordinates. The $z$ coordinate
can be suppressed since the dynamics of the system is confined in a plane.
Standard Lagrangian in C.G.S. units is used to define the system as follows
Dunne ; bag ; pn :
$\displaystyle
L=\frac{1}{2}m(\dot{x}_{i}^{2}+\dot{y}_{i}^{2})+\frac{eB}{2c}\epsilon_{ij}(x_{j}\dot{x}_{i}-y_{j}\dot{y}_{i})$
$\displaystyle-\frac{K_{0}}{2}(x_{i}-y_{i})^{2}-V({x_{1}});~{}~{}~{}i,j=1,2$
(1)
where $c$ is the speed of light in vacuum and $K_{0}$ is the spring constant
corresponding to the harmonic interaction between the two oppositely charged
particles. This model is constructed in the spirit of the “2D excitonic dipole
model” exmol ; emg ; dipex , wherein $m$ can be realized by the effective mass
of the “electron-hole” pair in some specific cases where the magnitude of the
effective mass of electrons and holes can be considered as approximately same
and the Fermi velocity provides an upper bound for its characteristic velocity
in a real physical solid state system. Note that the first term of the above
Lagrangian (1) represents the kinetic term of the charges and the second term
represents their interaction with the external magnetic field $\vec{B}$. We
use a rotationally symmetric gauge to define the vector potential $\vec{A}$
satisfying the equation $\vec{\nabla}\times\vec{A}=B\hat{z}$. The third term
is the harmonic interaction between the two charges, and finally, the fourth
term describes the additional interaction of the positive charge with an
impurity in the $x_{1}$ direction. The limit of a strong magnetic field $B$
and small mass $m$ such as $\frac{m}{eB}\rightarrow 0$ is of interest here, in
which the kinetic energy term becomes negligible in the Lagrangian (1) BD .
Thus, we may approximate the dynamics by the effective Lagrangian,
$L_{0}=\frac{eB}{2c}\epsilon_{ij}(x_{j}\dot{x}_{i}-y_{j}\dot{y}_{i})-V_{0}(x_{i},y_{i})$
(2)
where $V_{0}(x_{i},y_{i})=\frac{K_{0}}{2}(x_{i}-y_{i})^{2}+V({x_{1}})$.
The Lagrangian equations of motion of the co-ordinates of the positive and
negatively charged particles are given by,
$\dot{x}_{i}=\frac{c}{eB}\epsilon_{ij}\frac{\partial V_{0}}{\partial
x_{j}},~{}~{}\dot{y}_{i}=-\frac{c}{eB}\epsilon_{ij}\frac{\partial
V_{0}}{\partial y_{j}}$ (3)
Since our effective Lagrangian (2) is in first-order form, the effective
Hamiltonian of the model is given by
$H=V_{0}(x_{i},y_{i})$ (4)
In order to show the equivalence between the Lagrangian and Hamiltonian
formalism fj ; Rb , we consider Hamilton’s equations of motion:
$\dot{x}_{i}=\\{{x_{i},H}\\}={\\{x_{i},V_{0}(x_{i},y_{i})}\\}$ (5)
$\dot{y}_{i}=\\{{y_{i},H}\\}={\\{y_{i},V_{0}(x_{i},y_{i})}\\}$ (6)
The nontrivial symplectic structure can readily be obtained now by comparing
the Lagrangian equations of motion (3) with the form of Hamilton’s equations
of motion ($\ref{a1},\ref{a2}$) to yield the following brackets:
$\\{x_{i},x_{j}\\}=\frac{c}{eB}\epsilon_{ij};~{}~{}\\{y_{i},y_{j}\\}=-\frac{c}{eB}\epsilon_{ij};~{}~{}\\{y_{i},x_{j}\\}=0$
(7)
The canonical spatial translation generators for individual charged particles
are given by
$P_{x_{i}}=\frac{eB}{c}\epsilon_{ij}x_{j};~{}~{}P_{y_{i}}=-\frac{eB}{c}\epsilon_{ij}y_{j}$
(8)
Using the above expressions and the nontrivial symplectic structures between
the position co-ordinates (7), it can be checked that the momentum co-
ordinates also satisfy a nontrivial symplectic bracket, given by
$\displaystyle\\{P_{x_{i}},P_{x_{j}}\\}=\frac{eB}{c}\epsilon_{ij};~{}~{}\\{P_{y_{i}},P_{y_{j}}\\}=-\frac{eB}{c}\epsilon_{ij};~{}~{}$
$\displaystyle\\{x_{i},P_{x_{j}}\\}=\\{y_{i},P_{y_{j}}\\}=\delta_{ij}$ (9)
## III Quantum dynamics
In this section, we discuss the quantum theory of our non-relativistic two-
particle model at the strong magnetic field limit by elevating the phase space
variables to the level of quantum operators. We obtain the nontrivial or
unusual commutation brackets between the position operators given by:
$[\hat{x}_{i},\hat{x}_{j}]=il^{2}_{B}\epsilon_{ij};~{}~{}[\hat{y}_{i},\hat{y}_{j}]=-il^{2}_{B}\epsilon_{ij};~{}~{}[\hat{x}_{i},\hat{y}_{j}]=0;~{}~{}i,j=1,2$
(10)
with $l_{B}=\sqrt{\frac{\hbar c}{eB}}$ known as the magnetic quantum length
scale. Likewise, the other nontrivial phase space non-commutative algebras are
given as
$[\hat{P}_{x_{i}},\hat{P}_{x_{j}}]=i\frac{\hbar^{2}}{l^{2}_{B}}\epsilon_{ij};~{}~{}[\hat{P}_{y_{i}},\hat{P}_{y_{j}}]=-i\frac{\hbar^{2}}{l^{2}_{B}}\epsilon_{ij},$
(11)
$[\hat{x}_{i},\hat{P}_{x_{j}}]=[\hat{y}_{i},\hat{P}_{y_{j}}]=i\hbar\delta_{ij};$
(12)
It may be observed that in this case, neither the coordinates nor the momentum
operators commute pnb2 . However, the operators
$\hat{P}_{i}=\hat{P}_{x_{i}}+\hat{P}_{y_{i}}=\frac{eB}{c}\epsilon_{ij}(\hat{x}_{j}-\hat{y}_{j}),$
(13)
can act as proper (commutative) translation generators, so that they satisfy
the following commutation relations:
$[\hat{x}_{i},\hat{x}_{j}]=il^{2}_{B}\epsilon_{ij};~{}~{}[\hat{P}_{i},\hat{P}_{j}]=0;~{}~{}[\hat{x}_{i},\hat{P}_{j}]=i\hbar\delta_{ij},$
(14)
which represents a non-commutative Heisenberg algebra (NCHA) in two
dimensions. In this instance, the operator-valued Hamiltonian of the effective
system is given by
$\hat{H}=\frac{K_{0}}{2}(\hat{x}_{i}-\hat{y}_{i})^{2}+V(\hat{x}_{1})$ (15)
A more conventional setting of this Hamiltonian in terms of the commutative
translation generator $\hat{P}_{i}$ is as follows:
$\hat{H}=\frac{1}{2m_{B}}\hat{P}^{2}_{i}+V(\hat{x}_{1});~{}~{}~{}i=1,2$ (16)
where $m_{B}=\frac{e^{2}B^{2}}{c^{2}K_{0}}$ is the effective mass of the
reduced two-particle system. It turns out to be instructive to introduce the
pair of canonical variables:
$\hat{R}_{i}=\frac{\hat{x}_{i}+\hat{y}_{i}}{2};~{}~{}\hat{P}_{i}=\frac{eB}{c}\epsilon_{ij}(\hat{x}_{j}-\hat{y}_{j});~{}~{}i,j=1,2$
(17)
where $\hat{R}_{i}$ is the centre of mass coordinate and $\hat{P}_{i}$ is the
corresponding canonical momentum of our two-particle system. They satisfy the
usual Heisenberg commutation relations as
$[\hat{R}_{i},\hat{R}_{j}]=0;~{}~{}[\hat{P}_{i},\hat{P}_{j}]=0;~{}~{}~{}[\hat{R}_{i},\hat{P}_{j}]=i\hbar\delta_{ij}$
(18)
However, it is worth noting that the centre of mass position coordinates may
also satisfy non-commutative Heisenberg algebra (NCHA) if the two particles
are assumed to have different masses (For further details, see Appendix A).
Even, if the two particles have the same mass, but their position coordinates
satisfy NCHA with different non-commutativity parameters, in that case also,
the centre of mass position coordinates can give rise to a non-commutative
algebra.
Note further, that since the dynamics of the composite system is realized in
terms of the coordinates of the positively charged particle, the information
of the negatively charged particle is completely suppressed in the equations
(14, 16), but it is incorporated into the expression of commuting momentum
operators. The extended Heisenberg algebra of the type as considered in
eq.(14) has the important property: it is realizable in terms of commutative
usual phase space variables (17) as
$\hat{x}_{1}=Ad_{\hat{U}}(\hat{R}_{1});~{}~{}\hat{P}_{1}=Ad_{\hat{U}}(\hat{P}_{1})$
(19)
$\hat{x}_{2}=Ad_{\hat{U}^{\dagger}}(\hat{R}_{2});~{}~{}\hat{P}_{2}=Ad_{\hat{U}^{\dagger}}(\hat{P}_{2})$
(20)
where we have made use of the fact of adjoint action:
$Ad_{\hat{U}}(\hat{A})=\hat{U}\hat{A}\hat{U}^{\dagger}$ with a quasi unitary
operator $\hat{U}$:
$\hat{U}=\exp[(-\frac{il^{2}_{B}}{2\hbar^{2}})\hat{P}_{1}\hat{P}_{2}],$ (21)
as it does not act unitarily on the entirely non-commutative phase space.
We can observe from the aforementioned equations (19, 20) that the non-
commutative phase space commutation algebra (14) can be simulated in terms of
commutative phase space variables (canonical variables) i.e. the centre of
mass coordinates as
$\hat{x}_{i}=\hat{R}_{i}-\frac{c}{2eB}\epsilon_{ij}\hat{P}_{j},~{}~{}~{}~{}i,j=1,2$
(22)
It may be noted that this transformation is not canonical because it changes
the commutation brackets. This transformation has occasionally been called a
Darboux map n2 or Bopp’s shift hm2 which is of relevance in the Bohmian
interpretation of non-commutative quantum mechanics pbom . Furthermore, this
transformation with an explicit dependence on the deformation parameter,
allows us to convert the Hamiltonian in NC space into a modified Hamiltonian
in commutative equivalent space. It follows that if we are able to solve the
spectrum of the system Hamiltonian in commutative equivalent space, we can
also obtain the spectrum of the system in primitive non-commutative space,
though the states in both situations are not the same. We will discuss how the
aforementioned maps aid in the extraction of non-classical cat states in the
next section.
## IV Preparation of Schrödinger Cat states
Using the formalism presented in the previous section, we are now in a
position to investigate the main goal of this work, viz., how we might
naturally prepare Schrödinger’s Cat states. To do so, we first consider a
particular Hamiltonian with a harmonic oscillator potential in the
$\hat{x}_{1}$ direction, given by
$\hat{H}\rightarrow\hat{H}_{NC}=\frac{\hat{P}^{2}_{1}}{2m_{B}}+\frac{\hat{P}^{2}_{2}}{2m_{B}}+V(\hat{x}_{1}),$
(23)
where $V(\hat{x}_{1})=\frac{1}{2}K\hat{x}^{2}_{1}$ and
$m_{B}=\frac{e^{2}B^{2}}{c^{2}K_{0}}$. The corresponding time dependent
Schrödinger equation is:
$i\hbar\frac{\partial}{\partial t}|\psi(t)>_{NC}=\hat{H}_{NC}|\psi(t)>_{NC}$
(24)
Note that, because of the non-commutativity of this theory, it is impossible
to construct simultaneous eigenstates with noncommutative coordinates, which
makes it difficult to define a local probability density for the wave-function
that corresponds to a particular state $|\psi(t)>_{NC}$ pbom1 . However, this
issue can be bypassed by using the interpretation mentioned in pbom1 , or by
using the coherent states formulation of noncommutative quantum mechanics with
the help of the Voros product g .
In our present case, it can be easily observed that the system Hamiltonian
mentioned above can be rewritten as,
$\hat{H}_{NC}=\hat{U}\hat{H}_{CM}\hat{U}^{\dagger},$ (25)
with
$\hat{H}_{CM}=\frac{\hat{P}^{2}_{1}}{2m_{B}}+\frac{\hat{P}^{2}_{2}}{2m_{B}}+V(\hat{R}_{1}),$
(26)
where we have used the fact that
$V(\hat{x}_{1})=V(\hat{U}\hat{R}_{1}\hat{U}^{\dagger})=\hat{U}V(\hat{R}_{1})\hat{U}^{\dagger}$.
Here $\hat{H}_{CM}$ is the unitarily equivalent form of the system Hamiltonian
expressed in terms of the Center of Mass coordinates, whereas the
$\hat{H}_{NC}$ represents the system Hamiltonian written in terms of the
positively charged particle coordinates. We can readily recognize that
$V(\hat{R}_{1})=\frac{1}{2}K\hat{R}^{2}_{1}$, where $K$ is the spring constant
of the impurity interaction faced by the positive charge in the $\hat{x}_{1}$
direction only. Accordingly, the Schrödinger equation (24) transforms as
follows:
$i\hbar\frac{\partial}{\partial t}|\psi(t)>_{CM}=\hat{H}_{CM}|\psi(t)>_{CM}$
(27)
where $|\psi(t)>_{CM}=\hat{U}^{\dagger}|\psi(t)>_{NC}.$ The ground state of
the unitarily equivalent Hamiltonian ($\hat{H}_{CM}$) is now represented as
$|\psi_{0}>_{CM}=|0>\otimes[d_{+}|+k_{2}>+d_{-}|-k_{2}>],$ (28)
where $|d_{+}|^{2}$ and $|d_{-}|^{2}$ denote the probability of finding the
free particle in $|+k_{2}>$ and $|-k_{2}>$ states respectively, $|0>$
represents the ground state of the 1D harmonic oscillator system with
$\hat{a}_{1}$ and ${\hat{a}_{1}}^{\dagger}$ representing the corresponding
annihilation and creation operators respectively, satisfying the following
algebra:
$[\hat{a}_{1},\hat{a}^{\dagger}_{1}]=\mathbb{I};~{}~{}~{}~{}\hat{a}_{1}=\frac{m_{B}\omega_{B}\hat{R}_{1}+i\hat{P}_{1}}{\sqrt{2m_{B}\omega_{B}\hbar}};~{}~{}~{}~{}~{}\hat{a}_{1}|0>=0,$
(29)
with $\omega_{B}=\sqrt{\frac{K}{m_{B}}}$, and $|\pm k_{2}>$ corresponds to the
right and left moving free particle’s momentum state respectively, which
satisfies:
$\hat{P}_{2}|\pm k_{2}>=\pm P_{2}|\pm k_{2}>;~{}~{}~{}~{}P_{2}=\hbar k_{2}$
(30)
The state vector corresponding to the non-commutative phase space (or in terms
of the positively charged particle coordinates) is given by
$|\psi_{0}>_{NC}=\hat{U}|\psi_{0}>_{CM},$ (31)
$|\psi_{0}>_{NC}$ can be expressed as,
$\displaystyle|\psi_{0}>_{NC}=(\exp[(-\frac{il^{2}_{B}}{2\hbar^{2}})\hat{P}_{1}\otimes\hat{P}_{2}])$
$\displaystyle[|0>\otimes(d_{+}|+k_{2}>+d_{-}|-k_{2}>)]$ (32)
which leads to
$\displaystyle|\psi_{0}>_{NC}=d_{+}([\exp(-\frac{il^{2}_{B}k_{2}}{2\hbar}\hat{P}_{1})]|0>)\otimes|+k_{2}>$
$\displaystyle+d_{-}([\exp(\frac{il^{2}_{B}k_{2}}{2\hbar}\hat{P}_{1})]|0>)\otimes|-k_{2}>$
(33)
On substituting $l^{2}_{B}=\frac{\hbar c}{eB}$ in the above equation, we
arrive at-
$\displaystyle|\psi_{0}>_{NC}=d_{+}([\exp[{(-i\frac{ck_{2}}{2eB})}\hat{P}_{1}]]|0>)\otimes|+k_{2}>$
$\displaystyle+d_{-}([\exp[(i\frac{ck_{2}}{2eB})\hat{P}_{1}]]|0>)\otimes|-k_{2}>$
(34)
Now, for a harmonic oscillator potential, the momentum operator $\hat{P}_{1}$
can be written as-
$\hat{P}_{1}=i\sqrt{\frac{m_{B}\omega_{B}\hbar}{2}}({\hat{a}_{1}}^{\dagger}-\hat{a}_{1})$
(35)
Putting the above expression in equation (33), we obtain,
(36)
It follows that the above state vector (36) may also be written in the form of
a superposition of single-component coherent states as
$\displaystyle\scalebox{0.8}{$|\psi_{0}>_{NC}=d_{+}|+\alpha>\otimes|+k_{2}>+d_{-}|-\alpha>\otimes|-k_{2}>$},$
(37)
wherein $|\pm\alpha>=e^{\pm\alpha({\hat{a}_{1}}^{\dagger}-\hat{a}_{1})}|0>$
with $\alpha=\frac{ck_{2}}{2eB}\sqrt{\frac{m_{B}\omega_{B}\hbar}{2}}$ are
real-valued coherent states (or a displacement of the vacuum) that belong to
the subset of the over complete space of usual complex parameter valued
coherent states bom1 .
Here it may be worthwhile to mention a property of the coherent state
$|\pm\alpha>$: the dimensionless parameter $\alpha$ may be rewritten as
$\alpha=\frac{1}{2}P_{2}({\frac{K}{K_{0}}})^{1/4}\sqrt{\frac{c}{2eB\hbar}}=\xi
k_{2}l_{B}$ (38)
with $\xi=\frac{1}{2}(\frac{K}{4K_{0}})^{\frac{1}{4}}$ . A coherent state
$|\alpha>$ can have an arbitrarily large amplitude, and hence, the energy of a
macroscopic harmonic oscillator scv can be approximated by the energy of a
one-dimensional quantum mechanical HO by suitably choosing $\mid\alpha\mid$ to
be arbitrarily large. For large enough $\mid\alpha\mid$ values, $|+\alpha>$
and $|-\alpha>$ correspond to macroscopically distinguishable states and may
be labelled as ‘(+) (alive)’ and ‘(-) (dead)’ sv ; gl . In this sense, we can
regard the above state (37) as an entangled SCS, holding
$\mid\alpha\mid\sqrt{h}$ fixed with finite value in the classical limit v ; s
. Accordingly, one may consider $|\pm\alpha>$ to be “classical-like” states,
but their coherent superposition is endowed with non-classical properties. In
fact, this type of Schrödinger cat states have been generated by pulsed
stimulation of atomic Rydberg wave packets vn .
In the primitive non-commutative phase space, we may rewrite the state vectors
(36) in the following concise way:
$\displaystyle|\psi_{0}>_{NC}=\mathcal{N}[|+\alpha;+k_{2}>+e^{i\phi}|-\alpha;-k_{2}>];~{}~{}$
$\displaystyle|\pm\alpha;\pm k_{2}>=|\pm\alpha>\otimes|\pm k_{2}>,$ (39)
with an arbitrary phase factor ($\phi$) and normalization constant
$\mathcal{N}$. For the aforementioned reason, the states $|\pm\alpha>$ may be
considered to be “macroscopic” like states with the same amplitude but
opposite in phase. (in the present case, the $\mid\alpha\mid$ parameter is not
arbitrary, but is defined in terms of the spring constants, magnetic field and
electric charge). However, their superposition (39) has several non-classical
characteristics op . Particularly, for the relative phase factor
$e^{i\phi}=\pm 1$, we get even and odd cat states that have been well-studied
in the literature 2 ; bose . Moreover, it is evident from (39) that the
coherent states and the free particle states are entangled: when the coherent
state parameter has a positive sign, the free particle state is right-moving.
On the other hand, the free particle state is left-moving when the coherent
state parameter has a negative sign. Therefore, $|\psi_{0}>_{NC}$ is an
entangled Schrödinger cat state containing the coherent superposition cohsup1
; cohsup2 of two states that are diametrically opposite to one another.
Since a momentum eigenstate is an idealization kop , we consider a more
realistic scenario in which the system’s motion in the commutative phase space
is localized within a specific length scale $\sigma$ along the $\hat{R}_{2}$
direction. In this case, we generalize the notion of free particle states to a
propagating Gaussian state given by
$|\psi_{G}>=\sqrt{\frac{\sigma}{\sqrt{\pi}}}\int_{-\infty}^{+\infty}e^{-\frac{\sigma^{2}}{2}(k_{2}-k_{0})^{2}}|k_{2}>dk_{2}$
(40)
where $\sigma$ is the width and $k_{0}$ is the peak momentum of the wave
packet. Now, following the prescription of (28), we can write the composite
state of the particle, when the dynamics of the system are realized in terms
of the centre of mass coordinates, as
$|\psi_{0}>_{CM}=|0>\otimes|\psi_{G}>$ (41)
Accordingly, we can generalize the notion of a two-component cat state (39) to
(42)
which describes a multi-component entangled Schrödinger cat state rop1 where
each component is specified through the momentum eigenvalues. Such a state is
highly non-classical, which can be verified through the corresponding Wigner
function rop1 . Thus, in the presence of a strong magnetic field background,
one may successfully prepare a Schrödinger Cat State utilizing a non-
relativistic electric dipole model, where non-commutativity plays an important
role. It may be reiterated here that we explore the system in terms of the
positively charged particle coordinates.
## V Collapse and revival of entanglement of SCS
In this section, we will begin by investigating the degree of entanglement of
the SCS state $|\psi>_{NC}$. In order to do so, we first write down the
corresponding density matrix given by
$\displaystyle\hat{\rho}_{NC}=(\sqrt{\frac{\sigma}{\sqrt{\pi}}})^{2}\int_{-\infty}^{+\infty}\int_{-\infty}^{+\infty}[|\alpha(k_{2})>_{A}<\alpha(k_{2}^{\prime})|]$
$\displaystyle\otimes[|k_{2}>_{B}<k_{2}^{\prime}|]e^{\frac{-\sigma^{2}}{2}(k_{2}-k_{0})^{2}}e^{\frac{-\sigma^{2}}{2}(k_{2}^{\prime}-k_{0})^{2}}dk_{2}dk_{2}^{\prime}$
(43)
where the subscripts $A$ and $B$ denote two distinct subsections of our
bipartite system, one of which is associated with coherent states and the
other with momentum eigenstates, each of which corresponds to two distinct
degrees of freedom in the non-commutative plane. Since $|\psi>_{NC}$ is a
composite pure state, the entanglement between the coherent states and free
particle states can be quantified in terms of the von-Neumann entropy given by
$S=-Tr_{A}[\hat{\rho}_{red}~{}ln(\hat{\rho}_{red})]$ (44)
where the reduced density matrix is defined as
$\displaystyle\hat{\rho}_{red}=\text{Tr}_{B}[\hat{\rho}_{NC}]$
$\displaystyle=\frac{\sigma}{\sqrt{\pi}}\int_{-\infty}^{+\infty}[|\alpha(k_{2})>_{A}<\alpha(k_{2})|]e^{-\sigma^{2}(k_{2}-k_{0})^{2}}dk_{2}$
(45)
with
$\text{Tr}(\hat{\rho}_{red})=\frac{\sigma}{\sqrt{\pi}}\int_{-\infty}^{+\infty}e^{-\sigma^{2}(k_{2}-k_{0})^{2}}dk_{2}=1$
(46)
For the present purpose, it suffices to compute the purity function purity ,
given by
$\displaystyle\text{P}(\alpha)=\text{Tr}(\hat{\rho}^{2}_{red})=\sum_{n}<n|\hat{\rho}^{2}_{red}|n>$
$\displaystyle=\sum_{m}\sum_{n}<n|\hat{\rho}_{red}|m><m|\hat{\rho}_{red}|n>$
(47)
After a little algebra, one obtains
$\displaystyle<n|\hat{\rho}_{red}|m>=\frac{\sigma}{\sqrt{\xi^{2}l_{B}^{2}+\sigma^{2}}}\frac{1}{\sqrt{n!}\sqrt{m!}}e^{(-\sigma^{2}k_{0}^{2})}$
$\displaystyle(\frac{\xi
l_{B}}{2\sigma^{2}})^{n+m}\frac{\partial^{n+m}}{\partial
k^{n+m}_{0}}(e^{\frac{\sigma^{4}k_{0}^{2}}{\xi^{2}l_{B}^{2}+\sigma^{2}}})$
(48)
By inserting equation (V) into (V) it follows that
$\displaystyle\scalebox{0.9}{$\text{P}(\xi_{0};l_{B})=(\frac{1}{1+\xi^{2}_{0}})e^{(-2\sigma^{2}k_{0}^{2})}[e^{(\frac{\sigma^{2}k_{0}^{2}}{1+\xi_{0}^{2}})}(e^{\frac{\xi^{2}_{0}}{2\sigma^{2}}\overset{\leftarrow}{\frac{\partial}{\partial
k_{0}}}\overset{\rightarrow}{\frac{\partial}{\partial
k_{0}}}})e^{(\frac{\sigma^{2}k_{0}^{2}}{1+\xi_{0}^{2}})}]$};$ (49)
where $\xi_{0}=\frac{\xi l_{B}}{\sigma}$. The above expression can be
rewritten (see Appendix B) as
$\text{P}(\xi_{0};l_{B})=\frac{1}{\sqrt{{1+2\xi^{2}_{0}}}}$ (50)
In Figure 1, we plot the Purity function versus the parameter $\xi_{0}$. It
can be observed that the purity function reduces from unit value (separable or
disentangled state) with increase of the parameter $\xi_{0}$, indicating
increment of entanglement in the system for higher values of $\xi_{0}$ (or
lower values of the width of the wave packet $\sigma$). We consider the
quantum length scale $l_{B}=1.483\times 10^{-8}$m, and vary the width of the
wave-packet in the range of $O(10^{-11}\to 10^{-6})$. Different $l_{B}$ values
displayed in the figure may originate due to the variation of the magnetic
length scale with different accessible magnetic fields in the laboratory.
Figure 1: The Purity function is plotted against the dimensionless factor
$\xi_{0}$ which varies inversely with the width of the wave-packet $\sigma$.
Plots for several choices of the quantum length scale are displayed.
It may be noted that if we just assume $\xi_{0}<<1$ with $\xi\sim 1$ which
implies that $l_{B}<<\sigma$, i.e, the width of the Gaussian packet ($\sigma$)
is large enough compared to the magnetic quantum length scale such that we can
ignore $\xi_{0}$, then it leads to the unit value of the purity function, or
in other words, the collapse of the entanglement in the state. On the other
hand, we can make the states entangled by choosing $\sigma$ comparable to the
magnetic length scale $l_{B}$ where $\text{P}(\xi_{0};B)$ becomes less than
unity. More interestingly, the revival of the entanglement state can occur, if
one considers a time-dependent regime. Let us recall from the definition of
$\xi$, that it basically depends on the coupling strength $K$ of the
“impurity” interaction.
The dynamic behaviour of impurities in materials is known to lead to time-
varying spring interaction Dj1 ; Dj6 . Such dynamical nature of the coupling
has been studied in the literature in the context of several physical systems
such as in optical lattices Dj5 , and extensively in the domain of quantum
electronic transport Dj ; Dj3 ; Dj4 . Let us now, consider that the spring
“constant” $K$ is a slowly varying periodic function of time due to some
external effects, with the time-variation given by
$K(t)=K\text{cos}^{4}\omega_{d}t=K\text{cos}^{4}\theta(t)$ (51)
which clearly indicates
$\xi(t)=\frac{1}{2}(\frac{K\text{cos}^{4}\omega_{d}t}{4K_{0}})^{(1/4)}$ and
$\xi_{0}(t)=\frac{\xi(t)l_{B}}{\sigma}$. Hence, the purity function gets
modified to,
$\text{P}(\xi_{0};l_{B})=\frac{1}{\sqrt{1+2{\xi^{2}_{0}(t)}}}$ (52)
From the above equation, it follows that the Purity function is periodic. It
may be noted that even if the $\sigma$ is comparable to the magnetic length
scale $l_{B}$, disentanglement occurs for
$t_{d}=\frac{\pi}{2\omega_{d}},\frac{3\pi}{2\omega_{d}},\frac{5\pi}{2\omega_{d}},\frac{7\pi}{2\omega_{d}},.....$
with a separation of $\frac{\pi}{\omega_{d}}$ time period between two
successive collapses. For the rest of the time interval, the states are
entangled. This distinguishing feature is known as the collapse and revival of
entanglement in the literature rop2 . In Figure 2, we plot the Purity function
versus the periodic parameter $\theta(t)$ for several different values of the
wavepacket width $\sigma$. It is clearly seen that the magnitude of
entanglement revival increases more for narrower wavepackets.
Figure 2: Time evolution of the Purity function is plotted against the
parameter $\theta(t)$, for various widths of the wavepackets $\sigma$.
Here it needs to be mentioned that in order to observe entanglement revival of
the states, it is required to choose $\sigma$ of the order comparable to that
of the magnetic length scale $l_{B}$ or less, as
$-1\leq\text{cos}\omega_{d}t\leq+1$. On the other hand, if we choose $\sigma$
to be much larger than $l_{B}$, then the additional term in the denominator of
Eq.(52) can be completely negligible which will take us back again to the
situation of the entanglement collapse, viz. $\text{P}(\xi_{0};B)\sim 1$.
Instances of the phenomenon of entanglement collapse and revival have been
pointed out earlier in the literature predominantly in the context of the
Jaynes-Cummings model for optical systems R2 ; rop2 . Here we furnish a
striking example of entanglement collapse and revival in the context of an
excitonic dipole in a condensed matter system.
## VI Conclusions
To summarize, in this work, we have considered a composite two-particle planar
dipole system in the presence of a strong constant and uniform magnetic field,
in which two oppositely charged particles interact via harmonic interaction,
in addition to an impurity interaction experienced by the positively charged
particle. Our system may be regarded as a toy version of excitonic dipole
models that can be realized in some specific direct band gap semiconductors
exciband2 ; exciband3 ; exciband1 having the conduction band minimum for
electrons and the valence band maximum for holes both located at the same
point of the Brillouin zone, where the effective mass of electrons and holes
can be quite similar in magnitude. This typically arises due to specific band
structures and symmetries of materials. The additional interaction could arise
from intrinsic features such as defects or impurities, as well as from
external influences like an external electric field or strain in the material
e4 .
In our analysis, we have first addressed the classical picture in the context
of our system’s Lagrangian formulation which is the most natural in a strong
magnetic field limit. Using symplectic analysis of this first-order
Lagrangian, we have specified the canonical/Weyl-Moyal type deformed NC
classical phase space to be an intrinsic part of our model. Next, we have
explored the quantum mechanical description of our model by elevating all the
phase space variables to the level of Hermitian operators. The spatial and
momentum sectors of individual charged particles obey a non-commutative
deformed algebra. Here, the non-commutativity emerges as a natural consequence
of placing two oppositely charged particles in a strong constant background
magnetic field. The square of the magnetic length scale acts as the effective
non-commutative parameter.
We have presented a physical interpretation of the mapping from the deformed
phase space to the usual commutative phase space. The non-commutative phase
space represents the system Hamiltonian written in terms of the positively
charged particle coordinates, while the standard quantum mechanical phase
space is more suitable for describing our system in terms of the composite
system’s centre of mass coordinates. The dynamics can, therefore, be analyzed
in terms of non-commuting variables or, alternatively, using phase space
transformations, in terms of commuting variables. In literature, non-
commutativity has been often introduced by hand for a single point particle,
thus ruling out any physicality of commutative phase-space variables in such
cases. However, in the present case, non-commutativity emerges naturally,
thereby giving a physical meaning to the commutative phase-space variables.
Determining the Hamiltonian’s ground state in the commutative phase space
allows us to express the quantum state in the non-commutative phase space as a
superposition of two diametrically opposite coherent states, entangled with
momentum eigenstates. This reveals the emergence of entangled and two-
component as well as multi-component Schrödinger Cat States (SCS) in our
system.
Furthermore, we have estimated the magnitude of entanglement in the system of
multicomponent entangled cat states. By utilizing the purity function, we
demonstrate that the effective non-commutative parameter ($l_{B}^{2}$) is
responsible for the entanglement. We show that when the width of the Gaussian
wave packet ($\sigma$) significantly exceeds the minimal length scale
($l_{B}$), the entangled cat states undergo collapse. Conversely, when
$\sigma$ is comparable to the nonzero magnetic length scale $l_{B}$, the
entanglement can be observed. Moreover, we show that if time-dependent
impurity potential is chosen, entanglement revival and collapse occurs
periodically. So notably, within the same formalism, we observe the phenomenon
of collapse and revival of entanglement in the non-commutative plane in the
time-dependent regime with a suitable choice of the $\sigma$ parameter for the
revival case, while the collapse is completely controlled by the nodes of the
periodic function involved in the impurity interaction.
Before concluding, it may be noted that spin-orbit interactions in solid-state
systems introduce electronic band curvature, leading to the emergence of Berry
curvature in momentum space. Such Berry curvature modifies the usual phase
space symplectic structure of Bloch electrons xio ; xio2 . In light of non-
commutative quantum mechanics, our present analysis can be extended to include
investigations on the possible emergence of Schrödinger cat states in solid
state systems involving the 2D excitonic Coulomb problem with the Berry
curvature of the electron’s and the hole’s Bloch states Be1 ; Be2 ; Be3 . This
may open up a new window to experimentally observe quantum superposition for
“macroscopic” states.
## VII Acknowledgements
PN and ND acknowledge support from S.N. Bose National Centre for Basic
Sciences where this work was initiated. PN would also like to thank the
Institute of Theoretical Physics, Stellenbosch University for providing
postdoctoral funds during the period when a major part of this work was
completed. We thank Biswajit Chakraborty, Debasish Chatterjee, Ananda Dasgupta
and Frederik G. Scholtz for some fruitful discussions. ASM acknowledges
support from the Project No. DST/ICPS/QuEST/2018/98 from the Department of
Science and Technology, Government of India.
## VIII Appendix A
Here we present a manifestation of the non-commutativity of the centre of mass
coordinates arising in the case of two oppositely charged particles with
different masses $m_{+}$ and $m_{-}$ representing the masses of positive and
negatively charged particles respectively. The corresponding centre of mass
(CM) coordinates of the above-discussed system is-
$\displaystyle\hat{R}_{i}=\frac{m_{+}\hat{x}_{i}+m_{-}\hat{y}_{i}}{m_{+}+m_{-}};$
$\displaystyle~{}\hat{P}_{i}=\hat{P}_{x_{i}}+\hat{P}_{y_{i}}=\frac{eB}{c}\epsilon_{ij}(\hat{x}_{j}-\hat{y}_{j});~{}~{}i,j=1,2$
(53)
Now, utilizing the results obtained from equation (10), the commutation
brackets between the CM coordinates can be obtained in the following form-
$[\hat{R}_{i},\hat{R}_{j}]=\frac{m_{+}^{2}-m_{-}^{2}}{(m_{+}+m_{-})^{2}}il_{B}^{2}\epsilon_{ij};~{}~{}i,j=1,2$
(54)
clearly indicating the non-commutativity between the CM position coordinates
with
$\theta=\frac{m_{+}^{2}-m_{-}^{2}}{(m_{+}+m_{-})^{2}}il_{B}^{2}\epsilon_{ij}$
being the effective non-commutativity parameter. However, it is
straightforward to check that the other two commutation brackets remain
preserved.
$[\hat{P}_{i},\hat{P}_{j}]=0;~{}~{}[\hat{R}_{i},\hat{P}_{j}]=i\hbar\delta_{ij}$
(55)
It may be noted that the order of magnitude of the non-commutativity between
the CM position coordinates is much lesser compared to that of the position
coordinates of the individual constituent particles. This is simply because
$l_{B}^{2}$ itself is very small due to the strong magnetic field limit, the
presence of the additional mass factor reduces the whole effective non-
commutativity parameter $\theta$ to a much smaller value.
Now, let us introduce the relative coordinate system:
(56)
The commutation relations satisfied by the relative coordinates are given by
(57)
It is evident that the relative position coordinates commute as we have
considered two oppositely charged particles on a non-commutative space (it has
been shown earlier pmho , that the non-commutativity of a charged particle
differs from its antiparticle and also from any other particle of opposite
charge by the sign). On the other hand, the coordinates of relative momenta
give rise to a nontrivial commutation algebra with a reduced order of
magnitude from that of the individual constituent particle’s momentum
coordinates.
It may be further noted that the position coordinates of the centre of mass
and the position coordinates of the relative motion are not independent,
rather they obey the relation given by
$[\hat{R}_{i},\hat{r}_{j}]=-i{l_{B}^{2}}\epsilon_{ij};~{}~{}i,j=1,2$ (58)
So, clearly, there is a connection between the motion of the centre of mass
and the relative motion of the composite system in the non-commutative space.
This helps us to reduce the two-body problem completely to a one-body problem
for the internal motion in non-commutative space using the CM coordinates of
the composite system where the information of the negatively charged particle
is solely hidden/encoded within the CM momenta giving rise to a standard
commutative algebra.
## IX Appendix B
Here we provide a derivation for the expression of the purity function. We
begin with the expression of the reduced density matrix of the equation (V)
and the expression of the coherent state $|\alpha(k_{2})>$ and definition of
the Purity function from the equation(V),
$\text{P}(\alpha)=\sum_{l}\sum_{s}<l|\hat{\rho}_{red}|s><s|\hat{\rho}_{red}|l>$
(59)
The coherent state can be expressed as
$|\alpha(k_{2})>=e^{-\frac{\alpha^{2}}{2}}e^{\alpha\hat{a}_{1}^{\dagger}}e^{-\alpha\hat{a}_{1}}|0>=e^{-\frac{\alpha^{2}}{2}}e^{\alpha\hat{a}_{1}^{\dagger}}|0>$
$<l|\alpha(k_{2})>=<l|e^{-\frac{\alpha^{2}}{2}}\sum_{n=0}^{\infty}\frac{\alpha^{n}}{\sqrt{n!}}|n>=e^{-\frac{\alpha^{2}}{2}}\frac{\alpha^{l}}{\sqrt{l!}}$
(60)
Similarly,
$<\alpha(k_{2})|s>=e^{-\frac{\alpha^{2}}{2}}\frac{\alpha^{s}}{\sqrt{s!}}$.
Plugging this into the equation (59), one gets
$<l|\hat{\rho}_{red}|s>=\frac{\sigma}{\sqrt{\pi}}\int_{-\infty}^{+\infty}e^{-\alpha^{2}}\frac{(\alpha)^{l+s}}{\sqrt{l!}\sqrt{s!}}e^{-\sigma^{2}(k_{2}-k_{0})^{2}}dk_{2}$
(61)
Now substituting, $\alpha(k_{2})=\beta k_{2}$, where $\beta=\xi l_{B}$, we
get-
(62) (63)
$=\frac{\sigma}{\sqrt{\pi}}\frac{\beta^{l+s}}{\sqrt{l!}\sqrt{s!}}e^{-\sigma^{2}k_{0}^{2}}\frac{1}{(2\sigma^{2})^{l+s}}\frac{\partial^{l+s}}{\partial
k_{0}^{l+s}}\\{\sqrt{\frac{\pi}{\beta^{2}+\sigma^{2}}}e^{\frac{\sigma^{4}k_{0}^{2}}{\beta^{2}+\sigma^{2}}}\\}$
(64)
Using the above expressions in the purity function, we get,
(65)
Performing the summations, we are led to
$\text{P}(\alpha(k_{2}))=\frac{\sigma^{2}}{\beta^{2}+\sigma^{2}}e^{(-2\sigma^{2}k_{0}^{2})}[e^{\frac{\sigma^{4}k_{0}^{2}}{\beta^{2}+\sigma^{2}}}(e^{\frac{\beta^{2}}{2\sigma^{4}}\overset{\leftarrow}{\frac{\partial}{\partial
k_{0}}}\overset{\rightarrow}{\frac{\partial}{\partial
k_{0}}}})e^{\frac{\sigma^{4}k_{0}^{2}}{\beta^{2}+\sigma^{2}}}]$ (66)
Now, replacing $\xi_{0}=\frac{\xi l_{B}}{\sigma}$, we arrive at-
$\text{P}(\xi_{0};l_{B})=(\frac{1}{1+\xi^{2}_{0}})e^{(-2\sigma^{2}k_{0}^{2})}[e^{(\frac{\sigma^{2}k_{0}^{2}}{1+\xi_{0}^{2}})}(e^{\frac{\xi^{2}_{0}}{2\sigma^{2}}\overset{\leftarrow}{\frac{\partial}{\partial
k_{0}}}\overset{\rightarrow}{\frac{\partial}{\partial
k_{0}}}})e^{(\frac{\sigma^{2}k_{0}^{2}}{1+\xi_{0}^{2}})}]$ (67)
Next, we obtain a compactified form of
$[e^{\frac{\sigma^{2}k_{0}^{2}}{1+\xi_{0}^{2}}}(e^{\frac{\xi_{0}^{2}}{2\sigma^{2}}\overset{\leftarrow}{\frac{\partial}{\partial
k_{0}}}\overset{\rightarrow}{\frac{\partial}{\partial
k_{0}}}}e^{\frac{\sigma^{2}k_{0}^{2}}{1+\xi_{0}^{2}}}]$. For that, let us
consider the following integral:
$\int_{-\infty}^{+\infty}e^{-bs^{2}+2sk_{0}}ds=e^{\frac{k_{0}^{2}}{b}}\int_{-\infty}^{+\infty}e^{-b(s+\frac{k_{0}}{b})^{2}}ds=e^{\frac{k_{0}^{2}}{b}}\sqrt{\frac{\pi}{b}}$
(68)
From the expression of $e^{\frac{\sigma^{2}k_{0}^{2}}{1+\xi_{0}^{2}}}$, it
follows that-
$e^{\frac{\sigma^{2}k_{0}^{2}}{1+\xi_{0}^{2}}}=\sqrt{\frac{1+\xi_{0}^{2}}{\sigma^{2}\pi}}\int_{-\infty}^{+\infty}e^{-\frac{(1+\xi_{0}^{2})}{\sigma^{2}}s^{2}+2sk_{0}}ds$
(69)
Therefore,
$[e^{\frac{\sigma^{2}k_{0}^{2}}{1+\xi_{0}^{2}}}(e^{\frac{\xi_{0}^{2}}{2\sigma^{2}}\overset{\leftarrow}{\frac{\partial}{\partial
k_{0}}}\overset{\rightarrow}{\frac{\partial}{\partial
k_{0}}}})e^{\frac{\sigma^{2}k_{0}^{2}}{1+\xi_{0}^{2}}}]$
$\displaystyle=\scalebox{0.9}{$\frac{1+\xi_{0}^{2}}{\sigma^{2}\pi}\int_{-\infty}^{+\infty}e^{-\frac{(1+\xi_{0}^{2})}{\sigma^{2}}s^{2}+2sk_{0}}ds(e^{\frac{\xi_{0}^{2}}{2\sigma^{2}}\overset{\leftarrow}{\frac{\partial}{\partial
k_{0}}}\overset{\rightarrow}{\frac{\partial}{\partial
k_{0}}}})\int_{-\infty}^{+\infty}e^{-\frac{(1+\xi_{0}^{2})}{\sigma^{2}}s^{\prime
2}+2s^{\prime}k_{0}}ds^{\prime}$}$
[where we have used the relation ${e^{a{\frac{\partial}{\partial
k_{0}}}}{e^{bk_{0}}}}=e^{ab}e^{bk_{0}}$]
$=\sqrt{\frac{1+\xi_{0}^{2}}{\sigma^{2}\pi}}\int_{-\infty}^{+\infty}e^{-\frac{(1+\xi_{0}^{2})}{\sigma^{2}}s^{2}+2sk_{0}}e^{\frac{(\sigma^{2}k_{0}+\xi_{0}^{2}s)^{2}}{\sigma^{2}(1+\xi_{0}^{2})}}ds$
$=\sqrt{\frac{1+\xi_{0}^{2}}{\sigma^{2}\pi}}e^{\frac{\sigma^{2}k_{0}^{2}}{1+\xi_{0}^{2}}}\int_{-\infty}^{+\infty}e^{-\frac{(1+2\xi_{0}^{2})}{\sigma^{2}(1+\xi_{0}^{2})}s^{2}+2k_{0}\frac{(1+2\xi_{0}^{2})}{(1+\xi_{0}^{2})}s}ds$
After performing some suitable steps, we get the final simplified form as
$[e^{\frac{\sigma^{2}k_{0}^{2}}{1+\xi_{0}^{2}}}(e^{\frac{\xi_{0}^{2}}{2\sigma^{2}}\overset{\leftarrow}{\frac{\partial}{\partial
k_{0}}}\overset{\rightarrow}{\frac{\partial}{\partial
k_{0}}}})e^{\frac{\sigma^{2}k_{0}^{2}}{1+\xi_{0}^{2}}}]=\frac{1+\xi_{0}^{2}}{\sqrt{1+2\xi_{0}^{2}}}e^{2\sigma^{2}k_{0}^{2}}$
(70)
Now after plugging the above result (70) in equation (67), the expression of
the Purity function reduces to
$\text{P}(\xi_{0};l_{B})=\frac{1}{\sqrt{{1+2\xi^{2}_{0}}}}$ (71)
## References
* (1) W. H. Zurek, “Decoherence and the Transition from Quantum to Classical,” Phys. Today 36-44 1991, Quantum Theory of Measurement, edited by J. A. Wheeler and W. H. Zurek Princeton U.P. Princeton, 1983!, pp. 152-167.
* (2) S. Haroche and J. M. Raimond, in Cavity Quantum Electrodynamics, edited by P. Berman (Academic Press, New York, 1994), p. 123.
* (3) E. Schrödinger, Naturwissenschaften 23, 807, 823, 844(1935)
* (4) A. J. Leggett, “Schrödinger’s Cat and Her Laboratory Cousins,” Contemp. Phys. 25, 583-598 1984
* (5) C. C. Gerry and P. L. Knight, Quantum superpositions and Schrödinger cat states in quantum optics, Am. J. Phys. 65, 964 (1997)
* (6) R. Penrose, On gravity’s role in quantum state reduction, General Relativity and Gravitation 28, 581 (1996).
* (7) A. Vinante, R. Mezzena, P. Falferi, M. Carlesso, and A. Bassi, Improved Noninterferometric Test of Collapse Models Using Ultracold Cantilevers, Physical Review Letters 119, 110401 (2017).
* (8) B. Helou, B. J. J. Slagmolen, D. E. McClelland, and Y. Chen, LISA pathfinder appreciably constrains collapse models, Physical Review D 95, 084054 (2017).
* (9) Angelo Bassi, Kinjalk Lochan, Seema Satin, Tejinder P. Singh, and Hendrik Ulbricht, “Models of wave-function collapse, underlying theories, and experimental tests”, Rev. Mod. Phys. 85, 471
* (10) H. J. Kimble, M. Dagenais, and L. Mandel, Phys. Rev. Lett. 39, 691 (1977).
* (11) R. Short and L. Mandel, Phys. Rev. Lett. 51, 384 (1983).
* (12) R. E. Slusher, L. W. Hollberg, B. Yurke, J. C. Mertz, and J. F. Valley, Phys. Rev. Lett. 55, 2409 (1985).
* (13) M.A. Nielsen and I.L. Chuang, Quantum Computation and Quantum Information, Cambridge University Press, Cambridge, U.K. (2009).
* (14) D. Bouwmeester, A. Ekert and A. Zeilinger, The Physics of Quantum Information (Springer, Berlin, 2000).
* (15) A. Gilchrist, Kae Nemoto, W.J. Munro, T.C. Ralph, S. Glancy, Samuel. L. Braunstein and G.J. Milburn, Schrödinger cats and their power for quantum information processing, J. Opt. B: Quantum Semiclass. Opt. 6, S828 (2004).
* (16) K. Gietka, Squeezing by critical speeding up: Applications in quantum metrology, Phys. Rev. A 105, 042620 (2022).
* (17) Hung Do, Robert Malaney Jonathan Green, Teleportation of a Schrödinger’s-Cat State via Satellite-based Quantum Communications, arXiv:1911.04613v1 [quant-ph] 11 Nov 2019.
* (18) David S. Schlegel, Fabrizio Minganti, Vincenzo Savona, Quantum error correction using squeezed Schrödinger cat states, Phys.Rev.A 106 (2022) 2, 022431
* (19) Jacob Hastrup and Ulrik Lund Andersen, All-optical cat-code quantum error correction, Phys. Rev. Research 4, 043065(2022)
* (20) W. Schleich, J. P. Dowling, R.J. Horowicz, and S. Varro, in New Frontiers in Quantum Optics and Quantum Electrodynamics, edited by A. Barut (Plenum, New York, 1990)
* (21) The notion of interference between macroscopically distinguishable states has been promoted most prominently by A. Leggett, in Proceedings of the Internal Symposium on Foundations of Quantum Mechanics in the Light of New Technology, edited by S. Kamefuchi (Physical Society of Japan, Tokyo,1983).
* (22) G.J. Milburn and D. F. Walls, Phys.Rev. A 3S, 1087 (1988)
* (23) E. C. G. Sudarshan, Equivalence of semiclassical and quantum mechanical descriptions of statistical light beams, Phys. Rev. Lett. 10, 277 (1963).
* (24) M. Cosacchi, J. Wiercinski, T. Seidelmann, M. Cygorek, A. Vagov, D. E. Reiter, and V. M. Axt, On-demand generation of higher-order Fock states in quantum-dot-cavity systems, Phys. Rev. Research 2, 033489 (2020).
* (25) C. Navau, S. Minniberger, M. Trupke, and A. Sanchez, Levitation of superconducting microrings for quantum magnetomechanics, Physical Review B 103, 174436 (2021).
* (26) B. Li, W. Qin, Y. F. Jiao, C. L. Zhai, X. W. Xu, L. M. Kuang, H. Jing, Optomechanical Schrödinger cat states in a cavity Bose-Einstein condensate, Fundamental Research, 3, 15 (2023).
* (27) J. Foo, R. B. Mann, and M. Zych, Schrödinger’s cat for de Sitter spacetime, Classical and Quantum Gravity 38, 115010 (2021)
* (28) C. Marletto and V. Vedral, Gravitationally Induced Entanglement between Two Massive Particles is Sufficient Evidence of Quantum Effects in Gravity, Physical Review Letters 119, 240042 (2017)
* (29) M. Christodoulou and C. Rovelli, On the possibility of laboratory evidence for quantum superposition of geometries, Physics Letters, Section B: Nuclear, Elementary Particle and High Energy Physics 792, 64 (2019).
* (30) J.-Q. Liao, J.-F. Huang, and L. Tian, Generation of macroscopic Schrödinger-cat states in qubit-oscillator systems, Phys. Rev. A 93, 033853 (2016).
* (31) F.-X. Sun, S.-S. Zheng, Y. Xiao, Q. Gong, Q. He, and K. Xia, Remote Generation of Magnon Schrödinger Cat State via Magnon-Photon Entanglement, Phys. Rev. Lett. 127, 087203 (2021).
* (32) R. J. Marshman, S. Bose, A. Geraci, and A. Mazumdar, “Entanglement of Magnetically Levitated Massive Schrödinger Cat States by Induced Dipole Interaction,” arXiv:2304.14638.
* (33) S. Doplicher, K. Fredenhagen and J. Roberts, Commun. Math.Phys. 172 (1995) 187; Phys. Lett. B331 (1994) 39
* (34) H. Snyder, Quantized space-time, Phys. Rev. 71 (1) (1947) 38-41
* (35) A. Connes, Noncommutative geometry and reality, J. Math. Phys. 36, 6194 (1995); R. Szabo, Phys.Rep. 378 (2003) 207; E. Akofor, A. P. Balachandran and A. Joseph, arXiv:0803.4351 (hep-th).
* (36) R. Banerjee, B. Chakraborty, S. Ghosh, P. Mukherjee and S. Samanta, “Topics in Noncommutative Geometry Inspired Physics.” , Found Phys 39, 1297 (2009).
* (37) M. Chaichian, P. P. Kulish, K. Nishijima, and A. Tureanu, On a Lorentz-invariant interpretation of noncommutative space-time and its implications on noncommutative QFT, Phys. Lett. B 604, 98 (2004).
* (38) S. Girvin, cond-mat/9907002; N. Macris and S. Ouvry, J. Phys. A 35, 4477 (2002).
* (39) L. Susskind, The Quantum Hall fluid and noncommutative Chern-Simons theory, 2001, [arXiv:hep-th/0101029]
* (40) R. Jackiw and V.P. Nair, Phys. Rev. D43 (1991) 1933.
* (41) N. Banerjee, R. Banerjee and S. Ghosh, “Relativistic theory of free anyon revisited” Phys. Rev. D54, 1719 (1996).
* (42) H. Ishizuka, N. Nagaosa, “Noncommutative quantum mechanics and skew scattering in ferromagnetic metals”, Phys. Rev. B 96, 165202 (2017)
* (43) IB Pittaway and FG Scholtz, “Quantum interference on the non-commutative plane and the quantum-to-classical transition”, J.Phys.A 56 (2023) 16, 165303
* (44) D Trinchero and FG Scholtz, “Pinhole interference in three-dimensional fuzzy space” Annals Phys. 450 (2023) 169224
* (45) G. Amelino-Camelia, L. Doplicher, S. Nam and Y.-S. Seo, Phys. Rev. D 67, 085008 (2003)
* (46) I. Hinchliffe and N. Kersting, Int. J. Mod. Phys. A 19, 179 (2004)
* (47) G. Amelino-Camelia, G. Mandanici and K. Yoshida, J. High Energy Phys. 01, 037 (2004)
* (48) Z. Dong and T. Senthil, Non-commutative field theory and composite Fermi Liquids in some quantum Hall systems, Phys. Rev. B 102, 205126 (2020)
* (49) S. Hellerman and M. Van Raamsdonk, Quantum Hall physics = noncommutative field theory, JHEP 2001 (10), 039, arXiv:hep-th/0103179.
* (50) K. Cong, G. T. Noe II, and J. Kono, Excitons in magnetic fields, in Encyclopedia of Modern Optics (Second Edition) (Elsevier, Oxford, 2018) pp. 63-81
* (51) J. Gea-Banacloche, “Collapse and revival of the state vector in the Jaynes-Cummings model: An example of state preparation by a quantum apparatus”, Phys. Rev. Lett. 65, 3385 (1990).
* (52) Ting Yu and J.H. Eberly, Finite-Time Disentanglement via Spontaneous Emission, Phys. Rev. Lett. 93, 140404 (2004).
* (53) Saha, P., Majumdar, A. S., Singh, S., Nayak, N., 2010, “Collapse and revival of atomic entanglement in an intensity dependent Jaynes-Cummings interaction”, Int. J. Quant. Information, 8, 1397-1409
* (54) G.V.Dunne, R.Jackiw, C.A.Trugenberger, Phys.Rev D41(1990) 661
* (55) D. Bigatti, L. Susskind, Magnetic fields, branes and noncommutative geometry, Phys. Rev. D 62 (2000) 066004
* (56) P. Nandi, S. Sahu, S. K. Pal , Nucl. Phys. B 971 (2021) 115511
* (57) J. Zhou, W.-Y. Shan, W. Yao, and D. Xiao, Phys. Rev. Lett. 115, 166803 (2015).
* (58) A. Chernikov, T. C. Berkelbach, H. M. Hill, A. Rigosi, Y. Li, O. B. Aslan, D. R. Reichman, M. S. Hybertsen, and T. F. Heinz, Phys. Rev. Lett. 113, 076802 (2014)
* (59) Cong K, Noe G T II, Kono J. Excitons in Magnetic Fields. Oxford: Elsevier, 2018, p.63-81
* (60) E. Cobanera, P. Kristel, and C. Morais Smith, Phys. Rev. B 93, 245422 (2016).
* (61) L. D. Faddeev and R. Jackiw, “Hamiltonian reduction of unconstrained and constrained systems,” Phys. Rev. Lett. 60, 1692-1694 (1988)
* (62) R. Banerjee, H. J. Rothe and K. D. Rothe, Phys. Lett.B 462 (1999) 248-251, R. Banerjee, The commutativity principle and lagrangian symmetries, arXiv:hep-th/0001087
* (63) S. Biswas, P. Nandi, B. Chakraborty, Phys.Rev.A, 102 (2020) 2, 022231.
* (64) H.J. Rothe, K.D. Rothe, Classical and Quantum Dynamics of Constrained Hamiltonian Systems, World Scientific, Singapore, 2010.
* (65) Bertolami, O.; Rosa, J.G.; de Aragao, C.M.L.; Castorina, P.; Zappala, D. Noncommutative gravitational quantum well. Phys. Rev. D 2005, 72, 025010.
* (66) G. D. Barbosa and N. Pinto-Neto, “Noncommutative quantum mechanics and Bohm’s ontological interpretation”, Phys. Rev. D 69, 065014 (2004).
* (67) G.D.Barbosa, J. High Energy Phys. 0305 (2003) 024
* (68) P. Basu, B. Chakraborty and F. G. Scholtz, J. Phys. A 44 285204 (2011).
* (69) W.M. Zhang, D. H. Feng, and R. Gilmore, “Coherent states: Theory and some applications”, Rev. Mod. Phys., vol. 62, pp. 867-927, Oct 1990.
* (70) S. Mancini, D. Vitali, and P. Tombesi, Phys. Rev. Lett. 80, 688, 1998.
* (71) R.L. de Matos Filho and W. Vogel, “Even and odd coherent states of the motion of a trapped ion,” Phys. Rev. Lett. 76, 608-611 (1996).
* (72) C. Gerry and P. L. Knight, Introductory Quantum Optics (Cambridge University Press, Cambridge, 2004).
* (73) B. Yurke and D. Stoler, Generating quantum mechanical superpositions of macroscopically distinguishable states via amplitude dispersion, Phys. Rev. Lett. 57, 13 (1986).
* (74) A. Mecozzi and P. Tombesi, Phys. Rev. Lett. 58, 1055 (1987)
* (75) M. W. Noel and C. R. Stroud, Jr., in “Coherence and Quantum Optics VII,” edited by J. Eberly, L. Mandel, and E. Wolf (Plenum, New York, to be published).
* (76) J. Janszky, P. Domokos, and P. Adam, Coherent states on a circle and quantum interference, Phys. Rev. A 48, 2213 (1993).
* (77) R. Mirman, Analysis of the Experimental Meaning of Coherent Superposition and the Nonexistence of Superselection Rules, Phys. Rev. D 1, 3349 (Published: 15 June 1970).
* (78) Mark R. Dowling, Stephen D. Barlett, Terry Rudolph, and Robert W. Spekkens, Observing a coherent superposition of an atom and a molecule, arXiv:quant-ph/0606128v2 11 Dec 2006.
* (79) Binayak Dutta Roy, “Elements of Quantum Mechanics”, New Age Science (April 15, 2009)
* (80) S. Bose, K. Jacobs, and P. L. Knight, Preparation of nonclassical states in cavities with a moving mirror, Phys. Rev. A 56, 4175 (1997).
* (81) Gerardo Adesso, Alessio Serafini, and Fabrizio Illuminati, Entanglement, Purity, and Information Entropies in Continuous Variable Systems, arXiv:quant-ph/0506049v1 6 Jun 2005.
* (82) Brown, L.S.: Quantum motion in a Paul trap. Phys. Rev. Lett 66, 527 (1991).
* (83) Crefeld, C.E., Platero, G.: ac-driven localization in a two-electron quantum dot molecule. Phys. Rev. B 65, 113304 (2002).
* (84) Zeng, H.: Quantum-state control in optical lattices. Phys. Rev. A 57, 388 (1997).
* (85) Thouless D J, Quantization of particle transport, Phys. Rev. B 27, 6083 (1983).
* (86) Burmeister, G., Maschke, K.: Scattering by time-periodic potentials in one dimension and its infuence on electronic transport. Phys. Rev. B 57, 13050 (1998).
* (87) Li, W., Reichl, L.E.: Transport in strongly driven heterostructures and bound-state-induced dynamic resonances. Phys. Rev. B 62, 8269 (2000).
* (88) S. Das and G. S. Agarwal, J. Phys. B 42, 141003 (2009).
* (89) Jagdeep Shah , “Excitons in Semiconductor Nanostructures” (Springer, 1999).
* (90) E. J. Sie, J. W. McIver, Y.-H. Lee, L. Fu, J. Kong, and N. Gedik, Optical Stark effect in 2D semiconductors, Proc. SPIE Int. Soc. Opt. Eng. 9835, 129 (2016).
* (91) Thomas Mueller and Ermin Malic, Exciton physics and device application of two-dimensional transition metal dichalcogenide semiconductors, npj 2D Materials and Applications (2018)2:29.
* (92) G. Bir and G. Pikus, Symmetry and Strain-induced Effects in Semiconductors, A Halsted Press Book (Wiley, New York, 1974).
* (93) Xiao, D., J. Shi, and Q. Niu, 2005, “Berry phase correction to electron density of states in solids,” Phys. Rev. Lett. 95, 137204,
* (94) D. Xiao, M.-C. Chang and Q. Niu, “Berry phase effects on electronic properties”, Rev. Mod. Phys. 82, 1959-2007 (2010).
* (95) W. Yao and Q. Niu, “Berry Phase Effect on the Exciton Transport and on the Exciton Bose-Einstein Condensate” Phys. Rev. Lett. 101, 106401 (2008)
* (96) J. Zhou, W.-Y. Shan, W. Yao, and D. Xiao, “Berry Phase Modification to the Energy Spectrum of Excitons” Phys. Rev. Lett. 115, 166803 (2015).
* (97) A. A. Allocca, D. K. Efimkin, and V. M. Galitski, Fingerprints of Berry phases in the bulk exciton spectrum of a topological insulator, Phys. Rev. B 98, 045430 (2018).
* (98) P.-M. Ho, H.-C. Kao, Phys. Rev. Lett. 88 (2002) 151602.
|
# Discovering Block Structure in Networks
Rudy Arthur
###### Abstract
A generalization of modularity, called block modularity, is defined. This is a
quality function which evaluates a label assignment against an arbitrary block
pattern. Therefore, unlike standard modularity or its variants, arbitrary
network structures can be compared and an optimal block matrix can be
determined. Some simple algorithms for optimising block modularity are
described and applied on networks with planted structure. In many cases the
planted structure is recovered. Cases where it is not are analysed and it is
found that strong degree-correlations explain the planted structure so that
the discovered pattern is more ‘surprising’ than the planted one under the
configuration model. Some well studied networks are analysed with this new
method, which is found to automatically deconstruct the network in a very
useful way for creating a summary of its key features.
[inst1]organization=University of Exeter, Department of Computer
Science,addressline=Stocker Rd, city=Exeter, postcode=EX4 4PY, country=UK
## 1 Introduction
Modularity [1] is a function which takes a label assignment on the nodes of a
network and returns a score evaluating how effectively the label assignment
partitions the network into non-overlapping communities. Given a network with
(weighted) adjacency matrix $A_{ij}$ and the labelling function $c(i)$ mapping
nodes, $i$, to community labels, modularity is defined as
$Q_{\text{Newman}}=\frac{1}{2E}\sum_{ij}\left(A_{ij}-\gamma\frac{k_{i}k_{j}}{2E}\right)\delta(c(i),c(j)).$
(1)
$\sum_{ij}A_{ij}=2E$ is the total number of edges, $k_{i}$ is the degree of
the node $i$ and $\gamma$ is the so-called resolution parameter [2], set to
$1$ throughout this work. The sum over $A_{ij}$ measures the fraction of
within community edges in the observed network and the sum over degrees gives
the fraction of within community edges expected under a degree preserving
randomization of the network, known as the configuration model [3]. Despite
some well-known issues identifying small communities (the so-called resolution
limit [4]) modularity maximization is the basis for a number of very popular
community detection algorithms e.g. [5, 6].
Modularity has been extended in a number of ways to identify communities in
structured networks. Various authors [7, 8, 9] have given a definition of
modularity appropriate for community detection in bipartite networks and [10]
gives a definition of modularity appropriate for community detection in the
unipartite projection of bipartite networks. A definition of modularity
appropriate for finding multi-core-periphery structure is given in [11]. [12]
defines “anti-modularity” for finding sets of unconnected nodes, see also [13]
which uses modularity in a similar way as [12] to find approximately bipartite
node sets. The original paper defining modularity [1] has, at time of writing,
over 11000 citations, so clearly modularity is an important and well used tool
for community detection in networks.
Figure 1: Same network with two different node labellings, indicated by
colours. (a) emphasises community structure and (b) bipartite structure.
Most work identifies a target structure, like non-overlapping communities or
core and periphery sets, and aims to find a label assignment that maximises
modularity or one of its variants. However, consider Figure 1. The same
network is shown with two different labellings to emphasise either its
approximate community structure or its approximate bipartite structure. While
either of these labellings might be useful in different problem settings, this
paper tries to answer the question of which one is ‘best’ in the sense of
‘least expected under the configuration model’.
Section 2 generalises modularity to arbitrary network structures which are
specified by a block matrix $B$. Section 3 describes algorithms to optimise
this generalised modularity at fixed $B$ and then to optimise $B$ itself.
These algorithms are applied to synthethic and real networks in Sections 3 and
4. Section 5 summarises the results and suggests some directions for future
work.
## 2 Block Modularity
Consider a network with $N$ nodes labelled into $N_{B}$ groups or ‘blocks’.
Define the matrix of modularity $Q$ as the $N_{B}\times N_{B}$ matrix with
elements
$Q_{ab}=\sum_{ij}\left(A_{ij}-\gamma\frac{k_{i}k_{j}}{2E}\right)\delta(c(i),a)\delta(c(j),b)$
(2)
The standard modularity, equation 1, is equal to the trace of this matrix
divided by $2E$. Following [5], it is convenient to change the sum over nodes
in equation 2 to a sum over blocks. Defining
$\displaystyle\Sigma_{ab}$ $\displaystyle=\sum_{i\in a,j\in b}A_{ij}$ (3)
$\displaystyle T_{a}$ $\displaystyle=\sum_{i\in a}k_{i}$ (4)
lets us re-write equation 2 as
$Q_{ab}=\Sigma_{ab}-\gamma\frac{T_{a}T_{b}}{2E}$ (5)
When there are more connections between $a$ and $b$ than the configuration
model would predict $Q_{ab}>0$ and when there are fewer $Q_{ab}<0$. Thus, the
sign and magnitude of $Q_{ab}$ is a measure of how ‘surprising’ the edge
density between node sets $a$ and $b$ is, relative to the configuration model.
Large positive values correspond to an unexpected excess and large negative
values to a deficit.
Define block modularity as
$Q(B)=\frac{1}{2E}\sum_{ab}Q_{ab}B_{ab}$ (6)
Here $B$ is a $N_{B}\times N_{B}$ matrix with entries equal to $\pm 1$. To
gain some intuition it is helpful to consider the block matrices
$B_{0}=\begin{pmatrix}1&-1\\\ -1&1\\\ \end{pmatrix}\text{ and
}B_{1}=\begin{pmatrix}-1&1\\\ 1&-1\\\ \end{pmatrix}$
With nodes split into two blocks, with labels 0 and 1, $Q(B_{0})$ will be
large when there is an excess of edges within node sets and a deficit of edges
between them - this is the usual non-overlapping two community structure.
$Q(B_{1})$ is the opposite, large when there is an excess of edges between
node sets and a deficit within them. Therefore $Q(B_{1})$ will be large for
networks which are bipartite or approximately so.
Other modularity formulations can be recovered by taking different values for
the block matrix $B$. The standard equation 1, can be recovered with
$B_{ab}=\delta_{ab}$ for example. Other formulations can be recovered by
substituting the corresponding block pattern. If we split the label $a$ into a
parity label $x_{a}$ and community label $c_{a}$ then a modularity definition
suitable for a bipartite graph made of multiple communities, after [8], can be
obtained using
$B_{x_{a},c_{a};x_{b},c_{b}}=\delta_{c_{a}c_{b}}(1-\delta_{x_{a}x_{b}})$.
Similarly, the multi-core-periphery modularity of [11] can be recovered with
$B_{x_{a},c_{a};x_{b},c_{b}}=\delta_{c_{a}c_{b}}(x_{a}+x_{b}-x_{a}x_{b})$. In
contrast to most other definitions of modularity, we will use $B$ matrices
with values $\\{+1,-1\\}$ rather than $\\{1,0\\}$. This is to give equal
weight to excesses and deficits of connections between blocks.
Figure 2: Allowed $2\times 2$ and $3\times 3$ block patterns, after [14].
The work of [14] is the most similar to the above. They provide limits on the
types of block structure the configuration model can detect. By enforcing
symmetry and the identity
$\sum_{a}\Sigma_{ab}=\sum_{a}\frac{T_{a}T_{b}}{2E}$ (7)
certain types of block matrix are forbidden. For example, it is not possible
to simultaneously have an excess within and between all node sets. Defining
black cells as ones where $Q_{ab}>0$ and white cells where $Q_{ab}<0$,
equation 7 is equivalent to forbidding completely white or black rows or
columns so, in particular, a $2\times 2$ core-periphery block pattern is
forbidden. The allowed $2\times 2$ and $3\times 3$ patterns are shown in
figure 2.
The maximum value of standard modularity is 1 [15]. To understand the maximum
value of $Q(B)$, let us follow [16] and consider a block matrix $B$ with $+1$
on the diagonal and $-1$ elsewhere for a network of $N_{B}$ cliques with the
canonical labelling. The diagonal terms contribute
$\frac{1}{N_{B}}\left(1-\frac{1}{N_{B}}\right)$
and the off-diagonals give
$-\frac{1}{N_{B}^{2}}$
So that
$Q(B)_{max}=\frac{N_{B}}{N_{B}}\left(1-\frac{1}{N_{B}}\right)+\frac{N_{B}(N_{B}-1)}{N_{B}^{2}}=2-\frac{2}{N_{B}}$
(8)
which converges to $2$ for large $N_{B}$. The value is $2$ instead of the
standard modularity bound of $1$ due to counting the deficits as well as the
excesses. Like standard modularity, the upper bound is achieved only in the
limit of very large $N_{B}$ thus, like standard modularity, $Q(B)$ will tend
to be larger for higher $N_{B}$. In this work we compare $Q(B)$ at fixed
$N_{B}$, comparing values at different $N_{B}$ should be done cautiously.
For any fixed $B$, optimising equation 6 will find the label assignment on the
nodes that best matches that structure. For example $B_{6}$ is the block
pattern of 3 isolated communities. $Q(B_{6})$ for the network in Figure 1 will
be large given labels shown in 1(a). $B_{1}$ is the block pattern of a
bipartite network. $Q(B_{1})$ for the same network will be large for the label
assignment shown in 1(b). Given there are a finite number of allowed block
patterns we can find the optimal label assignment for every $B$ and compare
all of the maximised values of $Q(B)$ for different $B$. The block matrix that
gives the maximum $Q(B)$ score is the least expected under the configuration
model and therefore represents the structure in the network which can be least
well explained by degree correlation. In the following sections we will give
some examples that show how this optimal structure matrix can be useful in
characterising networks.
## 3 Algorithms for Finding Block Patterns
Algorithm 1 $\text{Label Swap}(B,T_{0}=0.01,k_{max}=1000)$
$\text{moves}\leftarrow 1$
$k\leftarrow-1$
Let $\vec{n}$ be the list of nodes
Let $c(n)$ be the label of node $n$
while $\text{moves}>0$ or $k<k_{max}$ do
$\text{moves}\leftarrow 0$
$k\leftarrow k+1$
$T=T_{0}\left(\frac{k_{max}-k}{k_{max}}\right)^{2}$
Randomly shuffle the list of nodes $\vec{n}\leftarrow\vec{n}^{\prime}$
for $n$ in $\vec{n}^{\prime}$ do
for $a$ in block labels where $a\neq c(n)$ do
Compute $dQ(a)$, the change in $Q(B)$ when $c(n)\leftarrow a$
end for
$dq\leftarrow\text{max}\left(dQ(a)\right)$
$c_{max}\leftarrow\text{argmax}\left(dQ(a)\right)$
if $dq>0$ then
$\text{moves}\leftarrow\text{moves}+1$
$c(n)\leftarrow c_{max}$
else if $k<k_{max}$ and $r<\exp(dq/T)$ then
$c(n)\leftarrow c_{max}$
end if
end for
end while
return $Q(B)$
For fixed $B$ and some initial labelling of the nodes $c(i)$, the optimisation
algorithm 1 finds a labelling with high $Q(B)$. The algorithm performs
simulated annealing with quadratic cooling, swapping node labels to increase
$Q(B)$, where moves that decrease $Q(B)$ are allowed at higher temperature.
$r$ is a random number and typical parameters are $T_{0}=0.1$, $k_{max}=100$.
The algorithm 1 performs better than a greedy method (setting $T=0$,
$k_{max}=0$) in most cases.
Figure 3: (a) A tripartite network with the optimal labelling under
$Q(B_{8})$, (b) the values of $Q(B_{i})$ for all $3\times 3$ block patterns,
the optimum is achieved at $B_{8}$.
Figure 3 (a) shows a tripartite network. Running algorithm 1 using each of the
allowed $3\times 3$ block patterns in Figure 2 gives the optimal values of
$Q(B_{i})$ shown in Figure 3 (b). The maximum $Q(B)$ is achieved for
$B=B_{8}$, which is the block pattern corresponding to the tripartite
structure of the network, with the optimal labelling for $Q(B_{8})$ indicated
by the colours in Figure 3 (a).
Figure 4: (a) A core-periphery network (pattern $B_{5}$) with the optimal
labelling under $Q(B_{7})$, (b) the values of $Q(B_{i})$ for all $3\times 3$
block patterns.
It is not always the case that the block pattern ‘planted’ in the network is
the one recovered by optimising $Q(B)$. Figure 4 (a) shows a network
consisting a core-periphery with an isolated communtiy, which corresponds to
block pattern $B_{5}$. The same process is applied and the optimal $Q(B)$ is
achieved with $B=B_{7}$ rather than $B_{5}$. This is further analysed in the
next section and A where it is shown that, due to the high degrees of the core
nodes, connections between core nodes are expected under a degree preserving
randomisation. This means that patterns with bipartite structure (as in
$B_{7}$) are favoured since they admit a labelling that is ‘more surprising’.
Figure 5: Left: The network, the block pattern used in its construction and
the optimal labelling under the optimal $B$ identified on the right. Right:
$Q(B)$ for every allowed $4\times 4$ block pattern.
To generate more complicated networks with planted block structures, for
$N_{B}$ blocks, $n$ nodes per block, construct an $N_{B}\times N_{B}$ density
matrix $P$, where $P_{ab}$ is the probability of a link between nodes in block
$a$ and block $b$. Figure 5 shows two networks where $N_{B}=4$, $n=10$ and $P$
is constructed by replacing black elements with $0.8$ and white elements with
$0.1$ in the corresponding block pattern. This is just the Stochastic Block
Model (SBM), see e.g. the review of [17]. In this work the SBM is only used to
generate networks with interesting block structure, but the SBM has a close
relationship with modularity maximisation. It was shown in [18] that
maximising the modularity is equivalent to finding the maximum likelihood
estimate of the parameters of a particular type of SBM called the planted
partition model. The approach taken here of optimising equation 6 likely has
some relationship with maximum likelihood estimation of the parameters of some
SBM, we will discuss connections to the SBM further in Section 5.
The top row of Figure 5 shows a network which consists of an isolated, dense
community, loosely connected to a tripartite network. Running Algorithm 1 for
all possible $4\times 4$ block patterns to find the optimal labelling for each
gives the result on the right. The planted structure is recovered as the
structure corresponding to the maximum $Q(B)$.
It is appropriate at this stage to look at the performance of Algorithm 1.
This algorithm, at any value of $T_{0}$, can get stuck at local maxima, much
like other modularity maximisation algorithms [5]. Slower cooling schedules
(larger $k_{max}$) typically find better maxima. For the network shown in the
bottom of Figure 5 optimising $Q(B)$ with the block partition shown in the
figure on the left, using the greedy algorithm ($T_{0}=0$) we find $31/100$
runs achieve the optimal label assignment. Using the annealing algorithm
($T_{0}=0.01$, $k_{max}=100$) this goes up to $49/100$. These numbers are
broadly representative of other networks and block patterns, with annealing
finding the optimum a factor of $\sim 2-5$ times more often in most cases.
Thus it is recommended (and implemented in Figure 5 and elsewhere) to run
Algorithm 1 a number of times, $N_{r}$, and choose the run with the largest
value of $Q(B)$. $N_{r}=20$ is found to be sufficient in the cases considered
in this work.
Again, it is not always the case that the planted structure is recovered. The
bottom row of Figure 5 shows a network constructed as an isolated community,
loosely connected to a core-periphery network, where the core has two distinct
peripheries. The optimal $B$ and label assignments are not the canonical ones
implied by the block matrix used to construct the network, a number of other
block patterns admit labellings with higher $Q(B)$. The largest $Q(B)$ is
found with the pattern shown the on right of Figure 5. The nearly isolated
community is recovered exactly but instead of the ‘core double periphery’
pattern there is a very small core connected to one periphery which itself
forms one half of a nearly bipartite pair.
To show explicitly how and why core-periphery structure can ‘vanish’ under
block modularity, consider a fully connected clique of $M$ nodes, all sharing
the label $0$, connected to a periphery of $qM$ nodes, labelled $1$. To
satisfy the consistency condition, equation 7, also add a disconnected clique
of $bM$ nodes. In A it is shown that $Q(B_{5})$, the optimal block modularity
for the planted structure is greater than $Q(B_{7})$ only if $b>q^{2}$.
The reqirement $b>q^{2}$ means that if the periphery is large (high $q$) or if
the core-periphery makes up the majority of the network (low $b$) then the
core-periphery block pattern can be a sub-optimal description of the network
structure under the configuration model. Ultimately, this condition derives
from squaring the degree sum of the core nodes. Since these nodes have very
high degree, under a degree preserving randomisation it is not unlikely that
they are connected to each other. In the network since only a subset of these
‘core’ edges are intra-block connections the high connectedness in the core is
expected and doesn’t contribute to the ‘surprise’ measured by $Q(B)$. From
this we can conclude that, as well as the conditions given by equation 7, for
core-periphery structure to exist (under the configuration model) requires
either a relatively small periphery or that the core-periphery only forms a
relatively small part of the overall network. The result is that core-
periphery structure, even if explicitly planted, can give lower $Q(B)$ than
using a block pattern where the core is missing. The result is analogous to
[14], certain block patterns may intuitively appear to be optimal, but under
the configuration model they are not, due to the importance of degree
correlations.
Algorithm 2 Simulated Annealing($N_{B},T_{0}=0.01,k_{max}=100$)
$B=I_{N_{B}\times N_{B}}$
$Q=\text{Label Swap}(B)$
while $T>0$ do
$T=T_{0}\left(\frac{k_{max}-k}{k_{max}}\right)^{2}$
do
For a random element of $B$
$B_{ij}\leftarrow-B_{ij}$
if $i\neq j$ then
$B_{ji}\leftarrow-B_{ji}$
end if
while equation 7 is false
$Q_{n}=\text{Label Swap}(B)$
if $Q_{n}>Q$ or $r<\exp((Q_{n}-Q)/T)$ then
$Q=Q_{n}$
else
Undo flip
end if
end while
Figure 6: Left: The network and the block pattern used in its construction
(which is also the optimal block pattern identified by Algorithm 2). The
network is labelled according to to the block pattern found by Algorithm 2,
shown on the right. Right: Shows the block pattern found by Algorithm 1 and
$Q(B)$ as a function of temperature.
The number of allowed block patterns grows quite rapidly with $N_{B}$ and an
exhaustive search becomes infeasible. Inspired by the similarity of these
block matrices to spin systems, Algorithm 2 is a simulated annealing approach
to finding the optimal block pattern. Figure 6 shows the final block pattern
and label set found by this algorithm for $T_{init}=0.01$ and $1000$ steps of
quadratic cooling. In practice the final results of algorithms 1 and 2 are
fairly insensitive to the exact cooling scheme, starting temperature and
number of cooling steps. Figure 6 shows that algorithm 2 recovers an
equivalent block pattern to the one used to generate the network, where a
permutation of the labels turns the pattern on the right into the one on the
left. The optimal labelling for this block pattern $B$ is the one expected
based on the planted structure. Algorithm 2, like algorithm 1, can become
stuck at local maxima. The same solution - repeated, independent runs - is
used to alleviate this problem. Algorithms 2 and 1 will find optimal label
assignments and block patterns but are somewhat inefficient. In this work the
intention is to understand $Q(B)$ rather than find the best possible algorithm
to optimise it, so algorithmic improvements are left for future work.
## 4 Real Networks
Figure 7: The optimal block pattern, labelling and value of $Q(B)$ for the
‘Southern Women’ network for $N_{B}=\\{2,3,4,5\\}$. The map of colours to
block indices is {0:blue, 1:orange, 2:green, 3:red, 4:purple, 5:brown}
In this section algorithm 2 will be run on some well known empirical networks
obtained from the KONECT database [19]. Since the optimal number of blocks is
unknown, the algorithm is run for a range of $N_{B}$. The first network to be
analysed is the Southern Women network [20], a bipartite network consisting of
18 women (labelled 1 through 18) and 14 events (labelled 19 through 32).
Typical analysis of this network [21] identifies 2 or more communities of
women and 2 or more classes of event. The block optimisation is summarised in
Figure 7. For $N_{B}=2$ the algorithm identifies the (exact) bipartite
structure of the network. $N_{B}=3$ demonstrates some interesting behaviour;
the optimal labelling does not include any labels for the third block. A block
pattern with lower $N_{B}$ can outperform a higher $N_{B}$ pattern. Algorithms
1 and 2 discover this fact by returning an optimal partition that only uses a
subset of the allowed node labels.
$N_{B}=4$ splits the events into two groups (red and orange) and splits the
women into two groups (red and blue) with green events predominantly attended
by red women and blue women attending orange events. $N_{B}=5$ refines this
picture, with 3 classes of event and 2 communities of women. Blue women attend
orange and purple events, while red women attend green and purple events. This
is also shown by the block matrix. The purple events bridge the two separate
bipartite communities. This network has been analysed in great detail, e.g. in
[21], and it is remarkable that this kind of structure can be detected fairly
automatically by optimising $Q(B)$.
Figure 8: The optimal block pattern, labelling and value of $Q(B)$ for the
Karate Club network with $N_{B}=\\{3,6\\}$. The map of colours to block
indices is {0:blue, 1:orange, 2:green, 3:red, 4:purple, 5:brown}. The bottom
shows a representation of the $6\times 6$ block matrix as a network.
Another commonly studied network is Zachary’s karate club [22], a social
network split into two communities due to a dispute between the club’s
instructor (node 33) and administrator (node 1). Using $N_{B}=2$ recovers the
usual two community pattern, Figure 8 shows results for $N_{B}=\\{3,6\\}$.
Higher $Q(B)$ can be obtained for higher $N_{B}$ but there is a balance to
strike between optimisation and interpretability of the block matrix. It is
also the case that optimisation of larger block patterns is slower and more
prone to get stuck in local minima.
The $3\times 3$ pattern is a community centered around the administrator and a
bipartite community of the other members, where the blue nodes have very few
ties with anyone other than the instructor. The optimal $6\times 6$ pattern is
shown in the middle panel. For large $N_{B}$ it can be hard to interpret the
block matrix from the binary pattern alone, so this pattern is visualised as a
network at the bottom of the figure, where a clearer picture emerges. Blocks 2
and 0 are two independent communities which only interact with the
administrator’s faction (block 1). The instructor’s faction, block 5,
interacts with a ‘loyalist’ block (3) and another block (4) which retains some
ties to the administrator. Important to note is the fact that unless a block
has a self loop it should not be considered a ‘community’. For example, the
nodes in block 3 have very few connections to each other, members of this
group are only connected to each other only via the instructor’s faction.
Figure 9: The optimal block pattern, labelling and value of $Q(B)$ for the
Dolphin network with $N_{B}=\\{2,3,4,5\\}$. The map of colours to block
indices is {0:blue, 1:orange, 2:green, 3:red, 4:purple}. Squares are female
dolphins, circles are male and triangles are unknown. The inset figure shows
the network diagram corresponding to the optimal block pattern, see Figure 8.
The final example is Figure 9 which shows the dolphin social network from
[23], which is suggested to be an example of a multiple core and periphery
structure in [14]. The optimal block patterns for $N_{B}\in\\{2,3,4\\}$ are
$N_{B}$ isolated communities with the labelling shown. For $N_{B}=5$ the
pattern is a little more interesting, the all female group 1 when $N_{B}=4$
splits in two and loses some members to the mixed red group. Blocks 1 and 3
now form two cores connected to a shared periphery, block 4. Unlike Figure 4
where a planted core-core-periphery structure could not be detected, as
discussed in 3, if a core-periphery pattern forms a small enough subset of the
total network, as here, it can be detected. This kind of structure is also
seen in Figure 8, where block 1 is the shared periphery of 0 and 2.
## 5 Conclusions
This paper presents a generalization of modularity, called block modularity,
which provides a framework for detecting arbitrary structure in networks.
Because different structures are evaluated with the same quality function,
these structures can be compared and an optimal one identified, meaning fewer
assumptions need to be made about the network. Searching for particular
patterns, like community structure or core-periphery, is still a useful thing
to do, however block modularity and an algorithm like 2 can approach the
network ‘blind’ and discover potentially interesting and unexpected
structures.
Section 4 shows that optimal block patterns can help create a ‘narrative’
account of a complex network. They can also be counter-intuitive, in
particular it seems that core-periphery structure is sometimes elusive. The
configuration model is a fairly powerful null model and degree correlations
‘explain’ a lot of network structure. It would be interesting future research
to explore in general what circumstances core-periphery networks may be better
explained by other block patterns, under the configuration model.
To apply this method on very large networks (web pages, social networks),
faster optimisation algorithms are required. Algorithm 1 requires $O(N^{2})$
operations per temperature increment and both algorithms can become trapped in
local minima. Direct enumeration of all block patterns is preferable where
possible, though seems to only be practical for $N_{B}<5$. Ideally, like the
Louvain and Leiden methods [5, 6], algorithms could be developed which are not
only fast, but which do not require the number of blocks $N_{B}$ to be
specified in advance, allowing a completely blind approach.
This approach shares many of the problems of standard modularity maximisation.
For example the resolution limit identified by [4] as preventing the detection
of small communities. There is also the fact that ‘modular’ partitions can be
found even for random networks [7]. As emphasised in [24] and [25], modularity
maximisation is not an inferential approach, like maximum likelihood
estimation of a SBM, but is descriptive. The label assignment, and here also
the structure matrix $B$, found by modularity maximisation answers the
question of which sets of nodes, in a specific network, have more or fewer
inter-connections than a degree preserving randomisation would predict. When
applied to any network, even a random one, the maximum modularity score may be
low but something will be found, since we are asking for a description of that
network.
The aim of this paper is not to add to the zoo of community detection methods.
[26] has performed a thorough study of numerous different community detection
algorithms and their relative performance. [27] has shown that the very
general SBM framework subsumes many different ‘mesoscopic pattern extraction’
problems, including community detection by modularity maximisation. Following
[18] or [27] it is likely that the maximisation of equation (6) has some
relationship with the SBM. Understanding these connections and comparing the
outputs of the methods described here (or some refinement of them) against
modern SBM techniques would be interesting future work. This paper proposes a
way to unify the many different modularity-like functions in the literature;
demonstrates some unexpected and interesting consequences for multi-core-
periphery networks and shows how block matrices provide a nice ‘summary’ of a
network. Recent critism of modularity based clustering [24, 25] raises many
interesting points about the drawbacks of the method and the misunderstanding
of its results in applications. However, while modularity remains a popular
approach to structure detection I hope that the unifying framework described
here will help researchers and practitioners to better understand these
methods and use them appropriately.
## Appendix A Core-periphery
Consider a fully connected clique of $M$ nodes, all sharing the label $0$,
connected to a periphery, labelled $1$, of $qM$ nodes and a disconnected
clique of $bM$ nodes.
$\Sigma=\begin{pmatrix}M^{2}&qM^{2}&0\\\ qM^{2}&0&0\\\
0&0&bM^{2}\end{pmatrix}$
$T=\begin{pmatrix}(1+q)M^{2}&qM^{2}&bM^{2}\end{pmatrix}$
$2E=M^{2}((1+q)+(b+q))$
Gives
$Q=M^{2}\left(\begin{pmatrix}1&q&0\\\ q&0&0\\\
0&0&b\end{pmatrix}-\begin{pmatrix}(1+q)^{2}&q(1+q)&b(1+q)\\\
q(1+q)&q^{2}&bq\\\ b(1+q)&qp&b^{2}\end{pmatrix}\frac{1}{(b+q)+(1+q)}\right)$
The matrix element $Q_{11}$
$M^{2}\left(-\frac{q^{2}}{(1+q)+(b+q)}\right)$
is always negative. The matrix elements $Q_{01},Q_{10}$
$M^{2}\left(q-\frac{q(1+q)}{(1+q)+(b+q)}\right)$
are always positive, as long as
$(b+q)>0$
which is always true, since $b$ and $q$ are positive and non-zero for non-
trivial networks.
The interesting term is $Q_{00}$ which is
$M^{2}\left(1-\frac{(1+q)^{2}}{(1+q)+(b+q)}\right)$
This is positive if
$b>q^{2}$
and otherwise negative. If $Q_{00}$ is negative then a bipartite block pattern
with
$\begin{pmatrix}B_{00}&B_{01}\\\
B_{10}&B_{11}\end{pmatrix}=\begin{pmatrix}-1&1\\\ 1&-1\end{pmatrix}$
will give higher $Q(B)$ than the ‘core-periphery’ block pattern.
## References
* [1] M. E. Newman, Modularity and community structure in networks, Proceedings of the national academy of sciences 103 (23) (2006) 8577–8582.
* [2] R. Lambiotte, J.-C. Delvenne, M. Barahona, Laplacian dynamics and multiscale modular structure in networks, arXiv preprint arXiv:0812.1770 (2008).
* [3] M. E. Newman, The structure and function of complex networks, SIAM review 45 (2) (2003) 167–256.
* [4] S. Fortunato, M. Barthelemy, Resolution limit in community detection, Proceedings of the national academy of sciences 104 (1) (2007) 36–41.
* [5] V. D. Blondel, J.-L. Guillaume, R. Lambiotte, E. Lefebvre, Fast unfolding of communities in large networks, Journal of statistical mechanics: theory and experiment 2008 (10) (2008) P10008.
* [6] V. A. Traag, L. Waltman, N. J. Van Eck, From louvain to leiden: guaranteeing well-connected communities, Scientific reports 9 (1) (2019) 1–12.
* [7] R. Guimera, M. Sales-Pardo, L. A. N. Amaral, Modularity from fluctuations in random graphs and complex networks, Physical Review E 70 (2) (2004) 025101.
* [8] M. J. Barber, Modularity and community detection in bipartite networks, Physical Review E 76 (6) (2007) 066102.
* [9] T. Murata, Detecting communities from bipartite networks based on bipartite modularities, in: 2009 International Conference on Computational Science and Engineering, Vol. 4, IEEE, 2009, pp. 50–57.
* [10] R. Arthur, Modularity and projection of bipartite networks, Physica A: Statistical Mechanics and its Applications 549 (2020) 124341.
* [11] S. Kojaku, N. Masuda, Finding multiple core-periphery pairs in networks, Physical Review E 96 (5) (2017) 052313.
* [12] L. Chen, Q. Yu, B. Chen, Anti-modularity and anti-community detecting in complex networks, Information Sciences 275 (2014) 293–313.
* [13] H. Li, C. Zhao, Y. Liu, X. Zhang, Anomaly detection by discovering bipartite structure on complex networks, Computer Networks 190 (2021) 107899.
* [14] S. Kojaku, N. Masuda, Core-periphery structure requires something else in the network, New Journal of physics 20 (4) (2018) 043012.
* [15] U. Brandes, D. Delling, M. Gaertler, R. Gorke, M. Hoefer, Z. Nikoloski, D. Wagner, On modularity clustering, IEEE transactions on knowledge and data engineering 20 (2) (2007) 172–188.
* [16] L. Danon, A. Diaz-Guilera, J. Duch, A. Arenas, Comparing community structure identification, Journal of statistical mechanics: Theory and experiment 2005 (09) (2005) P09008.
* [17] C. Lee, D. J. Wilkinson, A review of stochastic block models and extensions for graph clustering, Applied Network Science 4 (1) (2019) 1–50.
* [18] M. E. Newman, Community detection in networks: Modularity optimization and maximum likelihood are equivalent, arXiv preprint arXiv:1606.02319 (2016).
* [19] J. Kunegis, Konect: the koblenz network collection, in: Proceedings of the 22nd international conference on world wide web, 2013, pp. 1343–1350.
* [20] A. Davis, B. B. Gardner, M. R. Gardner, Deep South: A social anthropological study of caste and class, Univ of South Carolina Press, 2009.
* [21] M. G. Everett, S. P. Borgatti, The dual-projection approach for two-mode networks, Social networks 35 (2) (2013) 204–210.
* [22] W. W. Zachary, An information flow model for conflict and fission in small groups, Journal of anthropological research 33 (4) (1977) 452–473.
* [23] D. Lusseau, K. Schneider, O. J. Boisseau, P. Haase, E. Slooten, S. M. Dawson, The bottlenose dolphin community of doubtful sound features a large proportion of long-lasting associations, Behavioral Ecology and Sociobiology 54 (4) (2003) 396–405.
* [24] L. Peel, T. P. Peixoto, M. De Domenico, Statistical inference links data and theory in network science, Nature Communications 13 (1) (2022) 6794.
* [25] T. P. Peixoto, Descriptive vs. inferential community detection in networks: pitfalls, myths, and half-truths, arXiv preprint arXiv:2112.00183 (2021).
* [26] A. Ghasemian, H. Hosseinmardi, A. Clauset, Evaluating overfit and underfit in models of network community structure, IEEE Transactions on Knowledge and Data Engineering 32 (9) (2019) 1722–1735.
* [27] J.-G. Young, G. St-Onge, P. Desrosiers, L. J. Dubé, Universality of the stochastic block model, Physical Review E 98 (3) (2018) 032309.
|
11institutetext: Institute of Mathematics, Pedagogical University of Cracow,
Podchorazych 2, Cracow, Poland
11email<EMAIL_ADDRESS><EMAIL_ADDRESS>
https://matematyka.up.krakow.pl/
# On diagrams accompanying reductio ad absurdum proofs in Euclid’s Elements
book I. Reviewing Hartshorne and Manders
Piotr Błaszczyk 11 0000-0002-3501-3480
Anna Petiurenko 11 0000-0002-0196-6275
###### Abstract
Exploring selected reductio ad absurdum proofs in Book 1 of the Elements, we
show they include figures that are not constructed. It is squarely at odds
with Hartshorne’s claim that “in Euclid’s geometry, only those geometrical
figures exist that can be constructed with ruler and compass”.
We also present diagrams questioning Manders’ distinction between exact and
co-exact attributes of a diagram, specifically, a model of semi-Euclidean
geometry which satisfies straightness of lines and equality of angles and does
not satisfy the parallel postulate.
###### Keywords:
Impossible figuresEuclidean parts Semi-Euclidean plane.
## 1 Introduction
Euclid’s propositions are of two kinds: constructions and demonstrations. I.1
and I.32 are model examples: the first requires the construction of an
equilateral triangle, the second – a demonstration that angles in a triangle
sum up to “two right angles” ($\pi$, in short). Yet, the proof of I.32
includes the construction of a parallel through a point. Indeed, each
proposition includes a construction part (_kataskeuē_) which introduces
auxiliary lines exploited in the proof (_apodeixis_). In this paper, we focus
on diagrams accompanying reductio ad absurdum proofs, as they undermine the
common understanding of Euclid’s diagram.
The 20th-century Euclid scholarship grows in two trends: mathematical and
historical-philological. For diagrams, the first school examines them only as
straightedge and compass constructions, while the second seeks to show they
convey some mathematical information beyond construction requirements. We
challenge both approaches with specific diagrams.
R. Hartshorne [9] develops a coherent reading of the Elements, Books I–IV
focused on tacit axioms, non-defined concepts, or relations and interprets
them in the system of Hilbert axioms. He does not find Euclidean diagrams
problematic, misleading, or competing with a logical account of geometry; on
constructions, though, he writes: “The constructive approach pervades Euclid’s
Elements. There is no figure in the entire work that cannot be constructed
with ruler and compass […] in Euclid’s geometry only those geometrical figures
exists that can be constructed with ruler and compass” ([9], 18–19).
We discuss two examples undermining Hartshorne’s claim: the figure
accompanying proposition I.7 and one implied by the proof of I.27. The first
is not, and indeed, cannot be constructed, as assumptions of the proposition
introduce an inconsistent object. The non-constructive mode of the second
figure is related to the requirement “being produced to infinity” inherent in
the definition of parallel lines. These two figures are by no means
incidental, as the first props the SSS theorem (I.8), and the second brings us
to the core of the Euclid system.
J. Ferreirós provides a concise picture of the second school: “The original
geometry of Euclid lacks the means to derive its theorems by pure logic, but
it presents us with a most interesting and fruitful way of proving results by
diagrams” ([6], 132). In this vein, K. Manders [13] introduced a distinction
between exact and co-exact information inferred from a diagram that got
widespread renown among scholars exercising diagrammatic approach. Respective
definitions read: “Co-exact attributes are those conditions which are
unaffected by some range of every continuous variation of a specified diagram;
paradigmatically, that one region includes another […], or the existence of
intersection points such as those required in Euclid I.1 […]. Exact attributes
are those which, for at least some continuous variation of the diagram, obtain
only in isolated cases; paradigmatically, straightness of lines or equality of
angles […]. Exact attributes […] are unstable under perturbation of a diagram
([13], 92–93).111[13], p. 92 presents parallelism as an exact attribute.
Instead of ‘continuous variation of a diagram’, we introduce a global
perspective, meaning specifications of the plane on which a diagram lies.
Accordingly, we examine Euclid diagram I.1 on various Cartesian planes showing
that the existence of the intersection of circles involved depend on
characteristics of a plane. Regarding exact attributes, we present a model of
a semi-Euclidean plane that does not affect straightness of lines or equality
of angles but affects parallelism (especially I.29). Both counterexamples meet
the scheme: without touching a diagram but changing assumptions on the space
hosting it, we get different results concerning co-exact (intersection of
circles) and exact (parallelism) attributes. Depending on assumptions
concerning space, the same (from the diagrammatic perspective) circles meet or
not, and the same straight lines are parallel or not.
As for co-exact attributes, we provide an analysis of proposition I.6 that
undermines Mandres’ interpretation of inequality in terms of part-whole.
For the most part, our arguments exploit an interpretation of greater-than
relation. It affects our account of the deductive structure of the Elements,
some existential claims, Manders’ interpretation of greater-than in terms of
part-whole, and his claim regarding exact attributes. Euclid’s arguments
exploring that relation proceed reductio ad absurdum mode. Significantly, out
of eleven indirect proofs in Book I, ten, specifically I.6, 7, 27, 29, employ
greater-than relation.222The others are I.14, 19, 25, 26, 39, 40.
## 2 Euclid and Hilbert construction tools
Hartshorne ([9], 102) introduces term Hilbert construction tools, meaning
transportation of line segments and angles. Hilbert axioms justify these
operations; besides, they also state the uniqueness of respective line
segments and angles.333[8], pp. 597–602 provide a concise account of Hilbert
axioms. C1 and C4, respectively, decree construction tools. Diagrams drawn up
with both tools are acquired using the first alone; it suggests Euclid’s
straightedge and compass are more effective. We follow that clue to contrast
Euclid and Hilbert approaches.
### 2.1 Transportation of segments. I.1–3
I.1 To construct an equilateral triangle on the given line AB.444English
translation by Fitzpatrick [7], diagrams after [10].
Given that $AB=a$, point C, the third vertex of the wanted triangle is an
intersection of circles $(A,a)$ and $(B,a)$. In tables like the one below, we
lay out points resulting from intersections of straight lines and
circles.555The idea of such tables originates from [14].
$(A,a),(B,a)$
---
$C$
I.2 To place a straight-line at point A equal to the given straight-line BC
[$b$].
On the line-segment AB, we construct an equilateral triangle ABD with side a;
the below diagram depicts its shadow in grey. Point G is the intersection of
the circle $(B,b)$ and the half-line $DB^{\rightarrow}$. Now, DG represents
the sum of line-segments $a,\,b$. Circle $(D,a+b)$ intersects the half-line
$DA^{\rightarrow}$ at point L. Due to the Common Notions (CN, in short) 3, AL
proves to be equal $b$.
$(B,b),\,\,{DB}^{\rightarrow}$ | $(D,a+b),\,\,{DA}^{\rightarrow}$
---|---
$G$ | $L$
Owning to I.1-2, $b$ is placed at A in a precise position. Drawing circle
(A,b), one can choose any other position at will, and that is the substance of
I.3.
I.3 To cut off a straight-line equal to the lesser C [$b$] from the greater
AB.
Line-segment $b$ is transported to A into position AL; the above diagram
depicts the shadow of that construction; let $Ab$ be its symbolic
representation. The intersection of circle $(A,b)$ and line AB determines E
such that ${AE=b}$.
$Ab$ | $(A,b),\,\,AB$
---|---
| $E$
Summing up, due to I.1–3, one can transport any line segment to any point and
position. An equilateral triangle is a tool to this end, while the existence
of circle-circle and circle-line intersection points are taken for granted.
The Euclid system requires a circle-circle or circle-line axiom, both finding
grounds in Postulates 1–3 that introduce straight-edge and compass. Logically,
these two tools reduce to compass alone (vide Mohr-Mascheroni theorem), yet,
throughout the ages, the economy of diagrams prevailed and no one questioned
the rationale for Euclid’s instruments. There are, however, models of the
Hilbert system that do not satisfy the circle-circle axiom. Moreover,
Hartshorne shows ([9], 373) that I.1 does not hold in the Hilbert system of
absolute geometry. Thus, already at the very first proposition of the
Elements, we observe that Euclid and Hilbert’s systems follow alternative
deductive tracks. Therefore one cannot simply merge Hilbert’s axioms with
Euclid’s arguments.
### 2.2 Transportation of angles. I.22–23
In I.22, Euclid builds a triangle from three given line segments.666[8], p.
173 observes it is equivalent to the circle-circle axiom.
The below table presents $D,E$ as random- and $G,K$ as intersection- points.
| ${Db}$ | $(D,b),\,DE$ | $Da$ | $Gc$ | $(D,a)$, $(G,c)$
---|---|---|---|---|---
$D$, $E$ | | $G$ | | | $K$
I.23, angle transportation, rests on triangle construction as follows: on
sides of the given angle, Euclid builds a triangle, transports its sides to
$A$, $G$, obtaining another triangle. By the SSS, $\triangle CDE=\triangle
AFG$, hence $\angle KCL=\angle FAG$.
${Ab}$ | $(A,b),\,AB$ | $Aa$ | $Gc$ | $(A,a)$, $(G,c)$
---|---|---|---|---
| $G$ | | | $F$
## 3 Euclid’s vs Hilbert’s deduction: SAS to SSS
Throughout propositions I.1–34, equality means congruence, whether applied to
line segments, angles, or triangles. In I.5–8, showing the SSS theorem, Euclid
assumes I.4, Common Notions, and characteristics of the greater-than relation.
The proof of I.4 (SAS criterion) relies on the ad hoc rule: two straight-lines
can not encompass an area. The diagram depicts an area encircled by the base
$EF$ of the triangle and a curve with ends $E,F$. By contrast, Hilbert axioms
guarantee a unique line through points $E,\,F$ and diagram I.4 has no grounds.
I.5 Let ABC be an isosceles triangle. I say that the angle ABC is equal to
ACB.
The construction part is simple: F is taken at random on the half-line
$AB^{\rightarrow}$, then G such that $AF=AG$ is determined on the half-line
$AC^{\rightarrow}$.
$AB^{\rightarrow}$ | $(A,a+b),\,\,AC^{\rightarrow}$
---|---
$F$ | $G$
(i) Now, due to SAS, $\triangle GAB=\triangle FAC$. Thus $FC=BG$ and
$\beta=\angle AGB=\angle AFC=\beta^{\prime},$ $\gamma=\angle ABG=\angle
ACF=\gamma^{\prime}.$
(ii) Again by SAS, $\triangle BFC=\triangle BGC$, and
$\delta=\angle CBG=\angle BCF=\delta^{\prime}.$
(iii) By CN 3, $\gamma-\delta=\gamma^{\prime}-\delta^{\prime}$. Since
$\alpha=\gamma-\delta,\ \ \gamma^{\prime}-\delta^{\prime}=\alpha^{\prime},$
the equality $\alpha=\alpha^{\prime}$ holds. $\Box$
I.6 Let ABC be a triangle having the angle ABC equal to the angle ACB. I say
that side AB is also equal to side AC.
The proof reveals assumptions in no way conveyed through definitions or
axioms. At first, it is the trichotomy law for line segments. Let $AB=b$,
$AC=c$ (see Fig. 1). To reach a contradiction Euclid takes: if $b\neq c$, then
$b<c$ or $b>c$. Tacitly he assumes that exactly one of the conditions holds
$b<c,\ \ \ b=c,\ \ \ b>c.$
Figure 1: Proof of I.6 schematized.
Let $b>c$. Then the construction follows: “let DB, equal to the lesser AC,
have been cut off from the greater AB”. However, given that angles at $B$ and
$C$ are equal, then $AB=c$, and the cutting off “the lesser AC from the
greater AB” cannot be carried out. On the other hand, if $AB=b$ and $b>c$, the
triangle $ABC$ is not isosceles, and angles at $B$, $C$ are not equal.
Throughout the proof, thus, the diagram changes its metrical characteristic
and cannot meet the assumptions of the proposition. Contrary to Euclid’s
claim, $D$ is a random point on $AB$, rather than introduced via the following
construction
$(B,c),AB$
---
$D$
Now, by SAS, the equality $\triangle DBC=\triangle ACB$ holds, and Euclid
concludes the lesser to the greater. The very notion is absurd.
This time, the trichotomy law applies to triangles. The contradiction
$\triangle DBC=\triangle ACB\ \ \&\ \ \triangle DBC<\triangle ACB$
occurs against the rule: For triangles, exactly one of the following
conditions holds
$\triangle_{1}<\triangle_{2},\ \ \triangle_{1}=\triangle_{2},\ \ \
\triangle_{1}>\triangle_{2}.$
Figure 2: Elements, I.7 – letters $a,b$ added
I.7 On the segment-line $AB$, two segment lines cannot meet at a different
point on the same side of AB.
The proof, atypically, includes no construction as the thesis explicitly
states the impossibility of configuration depicted by the accompanying
diagram. To get a contradiction, Euclid assumes there are two points $C,D$
such that $AC=a=AD$ and $BC=b=BD$ (see Fig. 2). Both triangles $\triangle ACD$
and $\triangle BCD$ are isosceles and share the common base CD. Angles at
their bases are equal, $\alpha=\alpha^{\prime}$ and $\beta=\beta^{\prime}$.
Due to visual evidence, at the vertex $C$, the inequality $\alpha>\beta$
holds, while at $D$, $\beta^{\prime}>\alpha^{\prime}$.777[3] expounds the term
visual evidence in a bigger context. Thus, $\beta^{\prime}>\beta$ and, as
stated earlier, $\beta^{\prime}=\beta$. The very thing is impossible –
clearly, because exactly one of the conditions holds
$\beta^{\prime}<\beta,\ \ \beta^{\prime}=\beta,\ \ \beta^{\prime}>\beta.$
That proof assumes the trichotomy rule for angles and transitivity of greater-
than relation. By modern standards, it is, thus, a total order.888Euclid
applies the phrase “is much greater than” when referring to the transitivity.
In I.8, Euclid literally states the SSS criterion. Since the proof relies on a
superposition of triangles, we propose the following paraphrase: If two
triangles share a common side and have other corresponding sides equal, then
their corresponding angles will also be equal. In I.9–12, it is employed in
that form as Euclid considers two equal triangles on both sides of the common
side. Proof of that modification of I.8 effectively reduces to I.7.
### 3.1 Greater-than and Common Notions
Through §§ 10–11 of [9], Hartshorne seeks to prove Euclid’s propositions
I.1–34 within the Hilbert system, except I.1 and I.23, as they rely on the
circle-circle axiom. He observes that “Euclid’s definitions, postulates, and
common notions have been replaced by the undefined notions, definitions, and
axioms” in the Hilbert system. Commenting on Euclid’s proofs of I.5–8,
Hartshorne writes: “Proposition I.5 and its proof is ok as they stand. […]
every step of Euclid’s proof can be justified in a straightforward manner
within the framework of a Hilbert plane. […] Looking at I.6 […] we have not
defined the notion of inequality of triangles. However, a very slight change
will give a satisfactory proof. […] I.7 […] needs some additional
justification […] which can be supplied from our axioms of betweenness […].
For I.8, (SSS), we will need a new proof, since Euclid’s method of
superposition cannot be justified from our axioms” ([9], 97–99).
The above comparison between Euclid’s and Hilbert’s axiomatic approach
simplifies rather than expounds. Euclid implicitly adopts greater-than
relation between line segments, angles, and triangles as primitive concepts;
similarly to addition and subtraction (a lesser from the greater). In the
previous section, we have shown that he takes transitivity and the trichotomy
law being self-evident. Further characteristics one can recover from his
theory of magnitudes developed in Book V – the only part of Euclid’s geometry
hardly discussed by Hartshorne (see [9], 166–167). Here is a brief account.
Euclidean proportion (for which we adopt the 17th-century symbol $::$) is a
relation between two pairs of geometric figures (megethos) of the same kind,
triangles being of one kind, line segments of another kind, angles of yet
another. Magnitudes of the same kind form an ordered additive semi-group
$\mathfrak{M}=(M,+,<)$ characterized by the five axioms given below ([4], §
3).
1. E1
$(\forall{a,b\in M})(\exists{n\in{\mathbb{N}}})(na>b)$.
2. E2
$(\forall{a,b\in M})(\exists{c\in M})(a>b\Rightarrow a=b+c)$.
3. E3
$(\forall{a,b,c\in M})(a>b\Rightarrow{a+c>b+c})$.
4. E4
$(\forall{a\in M})(\forall{n\in{\mathbb{N}}})(\exists{b\in M})(nb=a)$.
5. E5
$(\forall{a,b,c\in M})(\exists{d\in M})(a:b::c:d),\ \ \mbox{where}\ \
na=\underbrace{a+a+...+a}_{n-times}$.
Clearly, E1–E3 provide extra characteristics of the greater-than relation.
A modern interpretation of Common Notions is simple: CN 1 justifies the
transitivity of congruence of line segments, triangles, and angles, CN 2 and 3
– addition and subtraction in the following form
$a=a^{\prime},\ b=b^{\prime}\Rightarrow a+b=a^{\prime}+b^{\prime},\ \
a-a^{\prime}=b-b^{\prime}.$
The famous CN 5, Whole is greater than the part, allows an interpretation by
the formula $a+b>a$ ([3], 73–76).
In the Hilbert system, the greater-than relation is defined through the
concept of betweeness and refers only to line segments and angles ([9], 85,
95); similarly, addition of line segments and angles are introduced by
definitions ([9], 168, 93). Then counterparts of Euclid’s axioms E2, E3, CN
1–3 are proved as theorems.
Here is a sample argument based on inequalities and its Hilbert-style
counterpart. In I.29, Euclid proves the thesis: When a line falls across
parallel lines $l,\,p$, equality of angles obtains $\alpha=\beta$ (see Fig.
3). For, if they are not equal, one of the angles is greater, suppose
$\alpha>\beta$. Then (implicitly by E3),
$\alpha>\beta\Rightarrow\alpha+\alpha^{\prime}>\beta+\alpha^{\prime}.$
Since $\alpha+\alpha^{\prime}=\pi$, angles $\beta,\,\alpha^{\prime}$ satisfy
the requirement of the parallel axiom, i.e., $\beta+\alpha^{\prime}<\pi$ and
straight lines $l,\,p$ meet, contrary to initial assumption.
Figure 3: Elements, I.29 schematized (left). Hartshorne’s version (right)
In contrast, Hartshorne’s proof of I.29 rests on the axiom stating there is
exactly one line through the point $A$ parallel to $p$ (see Fig. 3). Then, if
$\alpha\neq\beta$, a line $l^{\prime}$ through $A$ making angle $\beta$ with
$n$, by I.27, is parallel to $p$ – it contradicts the uniqueness of a parallel
line through $A$. Euclid’s proof, thus, implies an intersection point of $l$
and $p$, Hartshorne’s – a second parallel line to $p$.
## 4 Existence
### 4.1 Existence via Hilbert axioms
Tables in sections § 1–2 expound grounds for introducing points, namely: a
point is (1) an end of a given line segment, (2) a random point on a line
segment, a straight line, a half-line, or a circle, (3) a circle-circle,
circle-line, or line-line intersection point. Their existence is covered by
Hilbert axioms of Incidence, Betweenness, and Pasch, while circle-circle and
circle-line intersection points require the circle-circle axiom. Propositions
I.7 and I.27 bring in other cases: (4) a random point on the plane, (5) a
vertex of a triangle that exists owning to the definition of parallel lines.
Hilbert’s characterization of a plane does not explain case I.7: the axiom on
the existence of three non-colinear points allows to introduce triangle $ABC$
(see Fig. 2), yet, there are no grounds for point $D$ in the Hibert system.
Similarly, Euclid assumes point $D$ while none of the above rules (1)–(3)
guarantee its existence, and indeed, the very proposition does not include a
construction part. Commenting on I.7, Hartshorne points out the implicit
argument on betweenness, but does not report the suspicious status of $D$ (see
[9], 96, 99).
Ad I.27. Until proposition I.29, Euclid’s arguments do not rely on the
parallel postulate, yet, in I.27, aiming to show $AB\parallel CD$, given that
$\angle AEF=\angle EFD$ (see Fig. 4), he invokes definition of parallel lines:
“Parallel lines are straight-lines which, being in the same plane, and being
produced to infinity in each direction, meet with one another in neither” (I
def. 23).
Figure 4: Elements, I.27 (left). Triangle assumed in the proof (right).
The proof proceeds reductio ad absurdum mode and starts with the claim: “if
not, being produced, AB and CD will certainly meet together”. Suppose, thus,
$AB$ and $CD$ are not parallel and meet in $G$, then, in triangle $EFG$, the
external angle $\angle AEF$ is equal to the internal and opposite angle
$\angle EFD$, but by I.16, $\angle AEF$ is also greater than $\angle EFD$.
Hence, $\angle AEF=\angle EFD$ and $\angle AEF>\angle EFD$. The very thing is
impossible.
The rationale for point $G$ lies in the definition of parallel lines rather
than in construction with straightedge and compass. $G$ is not, and cannot be
determined as an intersection of lines $AB$ and $CD$: straight lines $AB$ and
$CD$ do not meet in either absolute or Euclidean geometry.
Summing up, points $D$, in I.7, and $G$, in I.27, constitute impossible
figures that can not be constructed with a straightedge and compass. Similar
arguments apply to the figure I.39 and those in Book III accompanying
propositions 2, 10, 13, 16, 23, 24. Generally, such impossible figures reveal
Euclid’s struggle with foundational problems: I.39 aims to link the theory of
equal figures and parallel lines, those in Book III – making the foundations
of trigonometry.
### 4.2 Existence and co-exact attributes
Euclid proposition I.1 is a model example for the proponents of diagrammatic
thinking. The enunciation of the proposition introduces point $C$ depicted on
the accompanying diagram through the following phrase “the point C, where the
circles cut one another” (see Fig. 5). It is not the case one has to read off
from the diagram that involved circles intersect. The circle-circle axiom is
not explicitly assumed, but it is not the only tacit supposition of the
Elements. Modern geometry studies how it relates to other axioms and how it
affects the characterization of a plane. The key result in that respect
states: if ${\mathbb{F}}\times{\mathbb{F}}$ is a Cartesian plane over an
ordered field $({\mathbb{F}},+,\cdot,0,1,<)$, the field is closed under the
square operation iff the circle-circle axiom is satisfied on
${\mathbb{F}}\times{\mathbb{F}}$ ([9], 144).
Figure 5: Circles on non-Euclidean and Euclidean planes
Let us consider now Manders’ claim on co-exact attribute as it explicitly
refers to I.1. It reads: “Co-exact attributes are those conditions which are
unaffected by some range of every continuous variation of a specified diagram
[…] or the existence of intersection points such as those required in Euclid
I.1 (which is unaffected no matter how the circles are to some extent
deformed)” ([13], 92).
The key concepts here are deformation and continuous variation. Manders
suggest reference to topological deformations, yet does not provide any
details. Taken literally, they require specification of the plane on which a
diagram lays. Instead of any deformation, we consider diagram I.1 on three
planes: ${\mathbb{Q}}\times{\mathbb{Q}}$, ${\mathbb{R}}\times{\mathbb{R}}$,
and ${\mathbb{R}}^{*}\times{\mathbb{R}}^{*}$ (see Fig.
5).999${\mathbb{R}}^{*}$ is the set of hyperreals (see § 6 below); circles
with radius $1$ have centers at $(0,0)$ and $(0,1)$. The second and third are
models of the Euclidean planes as fields of real and hyperreal numbers are
closed under the square root operation, while the first is not a Euclidean
plane. Due to calculations, we get to know that on the plane
${\mathbb{Q}}\times{\mathbb{Q}}$ the intersection point of circles does not
exist. However, from the cognitive perspective, one and the same diagram
represents circles in these three various mathematical contexts.101010As for
${\mathbb{Q}}\times{\mathbb{Q}}$, we get sure that circles look like
continuous objects due to a technical result such as [15]. Thus, with no
deformation of the diagram, but by switching from one mathematical context to
another we get or not that the intersection point of the circles exists.
## 5 Inequalities and co-exact attributes
In his analysis of proposition I.6, Manders interprets Euclid’s argument
$\triangle DBC$ is smaller than $\triangle ABC$ in terms of part-whole
relation read-off from the diagram ([13], 109–110). On another occasion, he
explicitly writes: “A strict inequality may be reduced […] to a proper-part
relationship in the diagram” ([13], 112).
In [3], we provide a detailed analysis of Common Notion 5 showing it does not
reduce to part-whole relation. In short, logical analysis gives the following
formula for CN5: $a+b>a$, where $a+b$ stands for the whole and $b$ – for its
part, and $a+b,\,a,\,b$ have to be of the same kind (triangles or angles).
Given that interpretation, triangle $ADB$ is not a Euclidean part of triangle
$ACB$; similarly, the gray angle is not a Euclidean part of the angle $CAB$,
as represented in Fig. 6.
Figure 6: Non-Eculidean parts
Euclid’s I.32 sets another challenge for the diagrammatic philosophy. Fereirós
reads it out as follows: “Most of the proof steps involve exact information,
dealing with equalities of angles, and they depend on previous theorems (I.31,
I.29) and common notions 2. Only two steps involve co-exact information,
bringing in attributions read from the diagram”, namely, that $\angle ACD$
covers $\angle AGE$ and $\angle EGD$ “is co-exact because it has to do with
part-whole relation and it is not affected by deformation” ([6], 135–136).
Here is our interpretation (see Fig. 7). Euclid transports angle $\alpha$ to
point $C$, and draws $CE$, which, by I.27, is parallel to $AB$. Hence, by
I.29, $\angle ECD=\beta$, and angles at $C$ sum up to “two right angles”,
$\gamma+\alpha+\beta=\pi$.
Figure 7: Proof of I.32 schematized.
Now, $CE$ lies inside angle $ACD$ because, by I.17, $\gamma+\alpha<\pi$. It is
not the case, thus, that the only way to get that information is to read it
from the diagram.
## 6 Existence meets Inequalities
### 6.1 Semi-Euclidean plane
In this section, we present a model o semi-Euclidean plane, i.e., a plane in
which angles in a triangle sum up to $\pi$ yet the parallel postulate fails.
[9], p. 311, introduces that term, but the very idea originates in Max Dehn’s
1900 [5], § 9, which built such a model owning to a non-Archimedean
Pythagorean field. Dehn explored a non-Euclidean field introduced already in
Hilbert [11], § 12.111111See also [9], § 18. Example 18.4.3 expounds Dehn’s
model. We employ the Euclidean field of hyperreal numbers. On the Cartesian
plane over hyperreals, the circle-circle and circle-line intersection axioms
are satisfied, meaning one can mirror Euclid’s straightedge and compass
constructions. To elaborate, let us start with the introduction of the
hyperreal numbers.
An ordered field $({\mathbb{F}},+,\cdot,0,1,<)$ is a commutative field
together with a total order that is compatible with sums and products. In such
a field, one can define the following subsets of ${\mathbb{F}}$:
1. $\mathbb{L}=\\{x\in{\mathbb{F}}:(\exists n\in{\mathbb{N}})(|x|<n)\\}$,
2. $\Psi=\\{x\in{\mathbb{F}}:(\forall n\in{\mathbb{N}})(|x|>n)\\}$,
3. $\Omega=\\{x\in{\mathbb{F}}:(\forall n\in{\mathbb{N}})(|x|<\tfrac{1}{n})\\}$.
They are called limited, infinite, and infinitely small numbers, respectively.
Here are some relationships helpful to pursue our arguments.
1. $(\forall x,y\in\Omega)(x+y\in\Omega,xy\in\Omega)$,
2. $(\forall x\in\Omega)(\forall y\in\mathbb{L})(xy\in\Omega)$,
3. $(\forall x\neq 0)(x\in\Omega\Leftrightarrow\ x^{-1}\in\Psi)$.
To clarify our account, let us observe the following equality $\Omega=\\{0\\}$
is a version of the well-known Archimedean axiom. Since real numbers form the
biggest Archimedean field, every field extension of
$({\mathbb{R}},+,\cdot,0,1,<)$ includes positive infinitesimals.
Let $\mathcal{U}$ be a non-principal ultrafilter on ${\mathbb{N}}$. The set of
hyperreals is defined as a reduced product
${\mathbb{R}}^{*}={\mathbb{R}}^{{\mathbb{N}}}/{\mathcal{U}}$. Sums, products,
and the ordered are introduced pointwise. A reader can take for granted that
the field of hyperreals $({\mathbb{R}}^{*},+,\cdot,0,1,<)$ extends real
numbers, hence, includes infinitesimals and infinite numbers; moreover, it is
closed under the square root operation (see [2], [1]). Fig. 8 represents in a
schematized way a relationship between ${\mathbb{R}}$ and ${\mathbb{R}}^{*}$,
as well as between $\mathbb{L}$, $\Psi$, and $\Omega$.
Figure 8: The line of real numbers and its extension to hyperreals
Due to the proposition 16.2 ([9], 144), the Cartesian plane over the field of
hyperreals is a model of Euclidean plane, with straight lines and circles
given by equations $ax+by+c=0$, $(x-a)^{2}+(y-b)^{2}=r^{2}$, where
$a,b,c,r\in{\mathbb{R}}^{*}$; angles between straight lines are defined as in
the Cartesian plane over the field of real numbers. Specifically, on the plane
${\mathbb{R}}^{*}\times{\mathbb{R}}^{*}$, angles in triangles sum up to $\pi$.
Parallel lines are of the form $y=mx+b$ and $y=mx+c$, while a perpendicular to
the line $y=mx+b$ is given by the formula $y=-\frac{1}{m}x+d$.
Now, take us a subspace $\mathbb{L}\times\mathbb{L}$ of the plane
${\mathbb{R}}^{*}\times{\mathbb{R}}^{*}$. On that plane, circles are defined
by analogous formula, namely $(x-a)^{2}+(y-b)^{2}=r^{2}$, where
$a,b,r\in\mathbb{L}$, while every line in $\mathbb{L}\times\mathbb{L}$ is of
the form $l\cap\mathbb{L}\times\mathbb{L}$, where $l$ is a line in
${\mathbb{R}}^{*}\times{\mathbb{R}}^{*}$. Since we want plane
$\mathbb{L}\times\mathbb{L}$ include lines such as $y_{1}=\varepsilon x$,
where $\varepsilon\in\Omega$, it has also to include the perpendicular
$y_{2}=\frac{-1}{\varepsilon}x$, but $\frac{-1}{\varepsilon}\notin\mathbb{L}$.
Formula $l\cap\mathbb{L}\times\mathbb{L}$, where $l=ax+by+c$ and
$a,b,c\in{\mathbb{R}}^{*}$ guarantees the existence of the straight line
$y_{2}$ in $\mathbb{L}\times\mathbb{L}$. Finally, the interpretation of angle
is the same as in the model ${\mathbb{R}}^{*}\times{\mathbb{R}}^{*}$.
Figure 9: Perpendicular lines with infinitesimal and infinitely large slopes
Explicit checking shows that the model characterized above satisfies all
Hilbert axioms of non-Archimdean plane geometry plus the circle-circle and
line-circle axioms, except parallel axiom; the more general theorem concerning
Hilbert planes also justifies our model, namely [9], p. 425, theorem, 43.7
(a).
Figure 10: Non-Euclidean plane $\mathbb{L}\times\mathbb{L}$ (left) vs.
Euclidean plane ${\mathbb{R}}^{*}\times{\mathbb{R}}^{*}$ (right)
With regard to parallel lines, let us consider the horizontal line $y=1$ and
two specific lines through $(0,0)$, $y_{1}=\varepsilon x,y_{2}=\delta x$,
where $\varepsilon,\delta\in\Omega$. (see Fig. 10). Since
$\Omega\mathbb{L}\subset\Omega$, the following inclusions hold
$y_{1},y_{2}\subset\mathbb{L}\times\Omega$. In other words, values of maps
$y_{1}(x),y_{2}(x)$ are infinitesimals, given that $x\in\mathbb{L}$. The same
obtains for any line of the form $y=\mu x$, with $\mu\in\Omega$. Since there
are infinitely many infinitesimals, there are infinitely many lines through
$(0,0)$ not intersecting the horizontal line $y=1$.
Since every triangle in $\mathbb{L}\times\mathbb{L}$ is a triangle in
${\mathbb{R}}^{*}\times{\mathbb{R}}^{*}$, it follows that angles in a triangle
on the plane $\mathbb{L}\times\mathbb{L}$ sum up to $\pi$ (see Fig. 11).
Figure 11: Triangles in Euclidean plane
${\mathbb{R}}^{*}\times{\mathbb{R}}^{*}$ and its subspace
$\mathbb{L}\times\mathbb{L}$
### 6.2 Exact attributes on the semi-Euclidean plane
Let us go back to Manders’ claim “Exact attributes are those which, for at
least some continuous variation of the diagram, obtain only in isolated cases;
paradigmatically, straightness of lines or equality of angles (neither of
which survive any except exceptional types of deformation, no matter how
small)”.
In Fig. 10, points $A=(\frac{1}{\varepsilon},1)$, $B=(\frac{1}{\delta},1)$ are
intersections of $y_{1},y_{2}$ with the horizontal line $y=1$ on the plane
${\mathbb{R}}^{*}\times{\mathbb{R}}^{*}$. Vertical line $x=0$ falls on $y=1$
and $y_{1}$ making internal angles less than $\pi$. Switching from
${\mathbb{R}}^{*}\times{\mathbb{R}}^{*}$ to $\mathbb{L}\times\mathbb{L}$, we
do not modify straightness of lines or equality of angles as far as it regards
the diagram on $\mathbb{L}\times\mathbb{L}$, however, there is nothing in the
diagram which informs us whether they intersect or not. Knowing plane
characteristics, we can infer they intersect, given that being on
${\mathbb{R}}^{*}\times{\mathbb{R}}^{*}$. The clue is the diagrammatic
philosophy does not impose any restrictions on Euclid’s clause “being produced
to infinity”. As focused on a diagram, it does not have means to decide
whether lines meet beyond a diagram.
Possibly, the claim on straightness of lines or equality of angles is designed
to exclude hyperbolic representations of lines and angles. Indeed, in the
Poincare model, straight lines are circles and angles are determined between
tangents; in the Klein model, straight lines are straight, while angles are
retrieved from the Poincare model. Both models then change the diagrammatic
representation of straight lines and angles. In the plane
$\mathbb{L}\times\mathbb{L}$, straight-lines are usual straight-lines, angles
are usual angles; moreover, triangles are Euclidean, there are also squares
and usual perpendicular lines. Yet, there are no instruments in the
diagrammatic tool-box to reveal a non-Euclidean character of that plane. It is
because diagrammatic perspective is local, while the phenomenon of parallelism
requires a global perspective.
## References
* [1] Błaszczyk, P.: Galileo’s paradox and numerosities. Zagadnienia Filozoficzne w Nauce 70, 73–107 (2021).
* [2] Błaszczyk, P.: A Purely Algebraic Proof of the Fundamental Theorem of Algebra. AUPC 8, 6–22 (2016); https://didacticammath.up.krakow.pl/article/view/3638
* [3] Błaszczyk, P. Mrówka, K., Petiurenko, A.: Decoding Book II of the Elements. AUPC 12, 39–88 (2020); https://didacticammath.up.krakow.pl/article/view/8462
* [4] Błaszczyk, P., Petiurenko, A.: Euclid’s proportion revised. AUPC 11, 37–61 (2019). https://didacticammath.up.krakow.pl/index.php/aupcsdmp/article/view/6901
* [5] Dehn, M.: Legendre’schen Sätze über die Winkelsumme im Dreieck. Mathematische Annalen 53(3), 404–439 (1900).
* [6] Ferreirós, J.: Mathematical Knowledge and the Interplay of Practices. Princeton University Press, Princeton (2016).
* [7] Fitzpatrick, R.: Euclid’s Elements of Geometry translated by R. Fiztpatrick (2007); http://farside.ph.utexas.edu/Books/Euclid/Elements.pdf
* [8] Greenberg, M.: Euclidean and non-Euclidean Geometries. Freeman, N York (2008).
* [9] Hartshorne, R.: Geometry: Euclid and Beyond. Springer, New York (2000).
* [10] Heiberg, J.: Euclidis Elementa. Vol. I. Teubneri, Lipsiae (1883).
* [11] Hilbert, D.: Grundlagen der Geometrie. Festschrift Zur Feier Der Enthüllung Des Gauss-Weber-Denkmals in Göttingen. Teubner, Leipzig (1899), 1–92.
* [12] Hilbert, D. Über den Zahlbegriff. Jahresbericht der Deutschen Mathematiker-Vereinigung 8, 180–184 (1900).
* [13] Manders, K.: The Euclidean Diagram. In: P. Mancosu (ed.): Philosophy of Mathematical Practice, Oxford University Press, Oxford (2008), 80–136.
* [14] Martin, G.: Geometric Constructions. Springer, Berlin (1998).
* [15] Tan, L.: The group of rational points on the unit circle. Mathematics Magazine 69(3), 162–171 (1996).
|
KDK collaboration
# Precision measurement of 65Zn electron-capture decays with the KDK
coincidence setup
L. Hariasz P.C.F. Di Stefano<EMAIL_ADDRESS>M. Stukel Department of
Physics, Engineering Physics & Astronomy, Queen’s University, Kingston,
Ontario K7L 3N6, Canada B.C. Rasco K.P. Rykaczewski Physics Division, Oak
Ridge National Laboratory, Oak Ridge, Tennessee 37831, USA N.T. Brewer
Physics Division, Oak Ridge National Laboratory, Oak Ridge, Tennessee 37831,
USA Joint Institute for Nuclear Physics and Application, Oak Ridge National
Laboratory, Oak Ridge, Tennessee 37831, USA R.K. Grzywacz Physics Division,
Oak Ridge National Laboratory, Oak Ridge, Tennessee 37831, USA Joint
Institute for Nuclear Physics and Application, Oak Ridge National Laboratory,
Oak Ridge, Tennessee 37831, USA Department of Physics and Astronomy,
University of Tennessee, Knoxville, Tennessee 37996, USA E.D. Lukosi
Department of Nuclear Engineering, University of Tennessee, Knoxville,
Tennessee 37996, USA Joint Institute for Advanced Materials, University of
Tennessee, Knoxville, Tennessee 37996, USA D.W. Stracener Physics Division,
Oak Ridge National Laboratory, Oak Ridge, Tennessee 37831, USA M. Mancuso F.
Petricca Max-Planck-Institut für Physik, Munich D-80805, Germany J. Ninkovic
P. Lechner MPG Semiconductor Laboratory, Munich D-80805, Germany
###### Abstract
65Zn is a common calibration source, moreover used as a radioactive tracer in
medical and biological studies. In many cases, $\gamma$-spectroscopy is a
preferred method of 65Zn standardization, which relies directly on the
branching ratio of $J\pi(\mbox{${}^{65}$Zn})=5/2^{-}\rightarrow
J\pi(\mbox{${}^{65}$Cu})=5/2^{-}$ via electron capture (EC∗). We measure the
relative intensity of this branch to that proceeding directly to the ground
state (EC0) using a novel coincidence technique, finding
$\mbox{$I_{\text{EC}^{0}}$}/\mbox{$I_{\text{EC}^{*}}$}=\mbox{$0.9684\pm
0.0018$}$. Re-evaluating the decay scheme of 65Zn by adopting the commonly
evaluated branching ratio of $\mbox{$I_{\beta^{+}}$}=1.4271(7)\%$ we obtain
$\mbox{$I_{\text{EC}^{*}}$}=\mbox{$(50.08\pm 0.06)\%$}$, and
$\mbox{$I_{\text{EC}^{0}}$}=\mbox{$(48.50\pm 0.06)\%$}$. The associated
$1115\text{\,}\mathrm{keV}$ gamma intensity agrees with the previously
reported NNDC value, and is now accessible with a factor of $\sim$2 increase
in precision. Our re-evaluation removes reliance on the deduction of this
gamma intensity from numerous measurements, some of which disagree and depend
directly on total activity determination. The KDK experimental technique
provides a new avenue for verification or updates to the decay scheme of 65Zn,
and is applicable to other isotopes.
###### Contents
1. I Introduction
2. II Methods
1. II.1 Apparatus and Visible Features
2. II.2 Physical Processes and Likelihood
3. III Results
4. IV Conclusions
5. A Components of the 65Zn Decay Scheme and Re-evaluated Branching Ratios
6. B Likelihood details
1. B.1 Coincidence sorting
2. B.2 Physical quantities
7. C Systematic errors
## I Introduction
65Zn is a common $\gamma$-ray calibration source [1], and for nearly a century
has been applied in the fields of medicine and biology as a radioactive tracer
[2, 3, 4]. It has been applied in various studies [5, 6] including an
investigation of potential orbital-modulation effects on decay constants [7].
In many applications, $\gamma$-ray spectroscopy is a convenient avenue for
activity determination of 65Zn, which is an emitter of essentially
monoenergetic $\gamma$ rays ($1115\text{\,}\mathrm{keV}$) associated with some
of its electron capture (EC) decays. This technique relies on knowledge of the
absolute $1115\text{\,}\mathrm{keV}$ intensity (fraction of decays that emit
$1115\text{\,}\mathrm{keV}$ $\gamma$s), available from existing decay scheme
evaluations such as those by the Laboratoire National Henri Becquerel (LNHB)
[8] or National Nuclear Data Center (NNDC) [9]. Though both evaluations report
$\mbox{$\mathrm{I}($1115\text{\,}\mathrm{keV}$)$}\sim 50\%$ with relative
errors of $\sim 0.2\%$, values (Table 1) deviate by $\sim 0.4\%$ between the
two sources. These evaluations combine dedicated measurements from the
Euromet-721 exercise [10, 11] with other reported values (e.g. from Refs. [12,
13]). All such determinations of absolute intensity are directly reliant on
activity measurements, which may require various corrections; though the
commonly used $4\pi\beta\gamma$ technique is influenced by low-energy X and
Auger radiation when used with EC-decaying nuclides [14], such corrections
have only been applied in some 65Zn studies [11].
| $\mathrm{I}($1115\text{\,}\mathrm{keV}$)$ (%)
---|---
LNHB (2006) [8] | $50.22\pm 0.11$
NNDC (2010) [9] | $50.04\pm 0.10$
Table 1: Absolute intensities of the 65Zn $1115\text{\,}\mathrm{keV}$
$\gamma$-ray, as reported in decay scheme evaluations of the Laboratoire
National Henri Becquerel and National Nuclear Data Center.
In this work, we present a novel measurement and resulting alternate, precise
determination of $\mathrm{I}($1115\text{\,}\mathrm{keV}$)$ through re-
evaluation of the decay scheme of 65Zn. Using data from the KDK experiment
[15, 16], $1115\text{\,}\mathrm{keV}$-producing electron capture decays (EC∗)
of 65Zn are distinguished from those proceeding directly to the ground state
(EC0) of its Cu daughter. The KDK collaboration has recently employed this
technique to obtain the first measurement of the exceedingly rare EC0 decay of
40K [17, 18]. The measurement of this work obtains a ratio of 65Zn
intensities, $\rho=\mbox{$I_{\text{EC}^{0}}$}/\mbox{$I_{\text{EC}^{*}}$}$. The
decay scheme of 65Zn, evaluated with our measurement, is displayed in Fig. 1.
Figure 1: The decay scheme of 65Zn. Branching ratios are calculated combining
our measurement of
$\rho=\mbox{$I_{\text{EC}^{0}}$}/\mbox{$I_{\text{EC}^{*}}$}$ with the shown,
commonly evaluated value of $I_{\beta^{+}}$ [8, 9] and assuming these three
branches complete the decay scheme
($1=\mbox{$I_{\text{EC}^{*}}$}+\mbox{$I_{\text{EC}^{0}}$}+\mbox{$I_{\beta^{+}}$}$).
Gamma energies are from Ref. [19] ($1115.539(2)\text{\,}\mathrm{keV}$), and
adopted $\gamma$ levels [9] ($770.6(2)\text{\,}\mathrm{keV}$). Intermediate
$\gamma$ intensities are from Ref. [10]. Total half-life is from Ref. [9], and
$\text{Q}_{\text{EC}^{0}}$ is from AME 2020 [20].
## II Methods
The KDK apparatus was designed for the discrimination between electron capture
branches proceeding through an excited daughter state to those transitioning
directly to the ground state. Below, we briefly outline the experimental
setup, which has been fully characterized elsewhere [16], with focus on
features visible in the 65Zn data and subsequently all physical processes
relevant for obtaining a measurement of $I_{\text{EC}^{0}}$/
$I_{\text{EC}^{*}}$.
### II.1 Apparatus and Visible Features
An inner Silicon Drift Detector (SDD) tags X-rays accompanying the source EC
decays of interest, as shown in Fig. 2. Due to its excellent resolution (FWHM
of $200\text{\,}\mathrm{eV}$ at $8\text{\,}\mathrm{keV}$), the Cu Kα
($8\text{\,}\mathrm{keV}$) and Kβ ($8.9\text{\,}\mathrm{keV}$) X-rays of
interest are easily distinguishable. In the same region, the majority of
remaining counts are attributed to K Augers from the same electron captures,
and a flat background originating primarily from the source $\beta^{+}$
branch. Below the signal (Cu) peaks, a small contribution from Zn Kα X-ray
fluorescence ($8.6\text{\,}\mathrm{keV}$) is included when fitting SDD
spectra, as shown further. Moreover, due to the low noise threshold of the
detector, L-shell X-rays (Cu, or fluoresced Zn) are visible near
$0.9\text{\,}\mathrm{keV}$.
Figure 2: SDD spectrum obtained with the 65Zn source over 1.4 days. The main
features of interest are Cu Kα and Kβ X-rays at 8 and 9 keV corresponding to
65Zn electron captures. Some L-shell X-rays are visible below
$1\text{\,}\mathrm{keV}$. A fit to SDD data distinguishes components near the
K X-ray peaks in Fig. 5.
The 65Zn source and SDD are centered inside a large, outer Modular Total
Absorption Spectrometer (MTAS) as depicted in Fig. 3. MTAS is an extremely
efficient NaI(Tl) $\gamma$-tagger, originally designed to study
$\beta$-strength distributions of fission products [21, 22]. For the 65Zn
$\gamma$ of interest ($1115\text{\,}\mathrm{keV}$), MTAS boasts a $\sim 98\%$
tagging efficiency [16].
Figure 3: Schematic of the Silicon Drift Detector
($\mathcal{O}(${\mathrm{mm}}^{2}$)$) and source placed inside the Modular
Total Absorption Spectrometer ($\mathcal{O}(${\mathrm{m}}^{2}$)$; cross-
section displayed). The latter contains a central module along with inner,
middle and outer rings, each containing NaI(Tl) volumes coupled to PMTs.
Several processes are visible in 65Zn MTAS spectra, such as those in Fig. 4.
Fully collected $1115\text{\,}\mathrm{keV}$ $\gamma$s from 65Zn EC∗ decays
form the most-prominent peak, though others originating from the natural MTAS
background along with source+background and source+source sum-peaks are
visible. The data is fit with calibrated, simulated spectra of 65Zn $\gamma$s
and $\beta^{+}$ in MTAS, along with a measured MTAS background and
convolutions. The shape of each spectrum is fixed, while the integral is
allowed to vary. Such spectral analysis verifies simulation methods used to
obtain MTAS efficiencies, and has been used to explore the source-SDD geometry
(Ref. [16] and App. B.2).
Figure 4: MTAS events of the 65Zn run in a
$2\text{\,}\mathrm{\SIUnitSymbolMicro s}$ coincidence window with a SDD
trigger. The dominant component is the $1115\text{\,}\mathrm{keV}$ $\gamma$
spectrum, associated with the 65Zn EC decays of interest. The measured MTAS
background, and its convolution with the $1115\text{\,}\mathrm{keV}$ spectrum
are included. The $\beta^{+}$ spectrum and a $\gamma+\gamma$ convolution
provide additional contributions. Simulated $\gamma$ and $\beta^{+}$ spectra
are calibrated to data in energy and resolution.
SDD data, including the K X-ray electron-capture signal of interest, are
categorized by coincidence with MTAS, generally using a
$t^{\prime}=\mbox{$2\text{\,}\mathrm{\SIUnitSymbolMicro s}$}$ nominal
coincidence window (CW). Details of coincidence characterization are available
in our previous work [16]. The sorted SDD data is used to inform the parameter
of interest, $\rho$, as described in the following section.
### II.2 Physical Processes and Likelihood
The main result is obtained through a simultaneous minimization of the Baker-
Cousins likelihood [23] on coincident and anti-coincident SDD spectra,
$-\ln\mathcal{L}=\sum_{i}\left\\{f_{i}(\bm{\theta})-n_{i}+n_{i}\ln\left[\frac{n_{i}}{f_{i}(\bm{\theta})}\right]\right\\},$
(1)
which compares the total observed events ($n_{i}$) in a bin ($i$) to the
corresponding model-predicted events ($f_{i}$). The parameters $\bm{\theta}$
include that of interest, $\rho$, along with various fixed and free terms
pertaining to the spectra.
Such a simultaneous fit to data sorted using a
$2\text{\,}\mathrm{\SIUnitSymbolMicro s}$ CW is shown in Fig. 5. The main
features are the Gaussian Cu Kα and Kβ peaks of fixed means near 8 and
$9\text{\,}\mathrm{keV}$. An additional X-ray peak near
$8.6\text{\,}\mathrm{keV}$ corresponds to fluoresced source Zn Kαs. Lastly, an
ad hoc component consists of a wide Gaussian term of free mean and width
describing Cu K Augers, and a flat term attributed primarily to source
$\beta^{+}$. The shape of all components is shared across coincident and anti-
coincident spectra, whereas the integral of each can vary.
Figure 5: A fit to 65Zn SDD data sorted using a
$2\text{\,}\mathrm{\SIUnitSymbolMicro s}$ coincidence window with MTAS. Cu Kα
and Kβ peaks from 65Zn electron captures inform
$\rho=\mbox{$I_{\text{EC}^{0}}$}/\mbox{$I_{\text{EC}^{*}}$}$. Fluoresced Zn Kα
and a continuous component consisting of Augers and flat background complete
the fit.
To obtain $\rho$, all processes affecting SDD signal detection and coincident
categorization must be considered. To assess the former, we consider expected
true, detected signal counts from the EC∗ branch ($\sigma^{*}$) along with
those from the EC0 branch ($\sigma^{0}$),
$\displaystyle\sigma^{*}$
$\displaystyle=\mathcal{N}\mbox{$I_{\text{EC}^{*}}$}P_{K}^{*}\omega_{K}(1-\eta_{\gamma})\eta$
$\displaystyle\sigma^{0}$
$\displaystyle=\mathcal{N}\mbox{$I_{\text{EC}^{0}}$}P_{K}^{0}\omega_{K}\eta.$
(2)
Both expressions above require the production of a Cu K X-ray and its
successful detection in the SDD. To first order, the total source decays
throughout data collection ($\mathcal{N}$), fluorescence probability
($\omega_{K}$), and SDD tagging efficiency of Cu K X-rays ($\eta$) do not need
to be known to inform $\rho$,
$\displaystyle\rho$
$\displaystyle=\frac{\mbox{$I_{\text{EC}^{0}}$}}{\mbox{$I_{\text{EC}^{*}}$}}$
$\displaystyle=\frac{\sigma^{0}}{\sigma^{*}}\frac{P_{K}^{*}}{P_{K}^{0}}(1-\eta_{\gamma}).$
(3)
However, K-shell capture probabilities ($P_{K}$) differ between the two
electron capture branches, and in the case of EC∗\- originating signal, we
require that the associated de-excitation gamma is not also seen in the SDD
($1-\eta_{\gamma}$). K-capture probabilities are obtained from BetaShape V2.2
[24], and $\eta_{\gamma}$ is obtained combining a measured geometric
efficiency with simulations (App. B.2). Above, internal conversion, internal
pair formation, and internal Bremsstrahlung have been neglected as these sub-
dominant processes are negligible at our precision.
The expected true signal counts are related to the observed, sorted signal
counts (those in the Cu K X-ray peaks of Fig. 5) by accounting for any process
affecting coincidence sorting. One such effect stems from the imperfect MTAS
$\gamma$-tagging efficiency $\epsilon$, which primarily affects sorting of
EC∗-originating signal. This efficiency has been characterized for multiple
coincidence windows by scaling measured 54Mn efficiencies in energy through
simulations, and correcting for deadtime [16], yielding a value of
$97.93(6)\%$ at the $2\text{\,}\mathrm{\SIUnitSymbolMicro s}$ CW.
Moreover, any process detectable in MTAS can occur in spurious coincidence
with the Cu SDD signal. The natural MTAS background rate is accounted for by
considering a Poisson probability $P^{0}_{\mathrm{BG}}=\exp(-\beta\bar{t})$ of
no natural MTAS background events within the effective coincidence window
$\bar{t}$, along with an analagous probability of no MTAS-detected source
$1115\text{\,}\mathrm{keV}$ $\gamma$s ($P^{0}_{\gamma}$). Expected MTAS
background coincidence rates have been obtained elsewhere [16], and
$P^{0}_{\gamma}$ is obtained in App. B. An additional source of coincidences
with source $\beta^{+}$ has been explored and deemed negligible. All spurious
coincidences are proportional to the $\mathcal{O}($\mathrm{\SIUnitSymbolMicro
s}$)$ CW, and, most significantly, place some EC0-originating signal in the
coincident spectrum. Neglecting these terms thus tends to underestimate
$\rho$, with increasingly dramatic effects at larger CWs.
With the above MTAS gamma efficiency and spurious coincidence considerations,
expected coincident ($\Sigma^{*}$) and anti-coincident ($\Sigma^{0}$) signal
counts are obtained,
$\displaystyle\Sigma^{*}$
$\displaystyle=\frac{\nu}{1+\rho^{\prime}}\biggl{[}\epsilon+(1-P^{0}_{\mathrm{BG}}P^{0}_{\gamma})(1-\epsilon+\rho^{\prime})\biggr{]}$
$\displaystyle\Sigma^{0}$
$\displaystyle=\frac{\nu}{1+\rho^{\prime}}P^{0}_{\mathrm{BG}}P^{0}_{\gamma}(1-\epsilon+\rho^{\prime}),$
(4)
with more detail available in App. B. Above, $\nu=\sigma^{*}+\sigma^{0}$ are
total signal events and $\rho^{\prime}=\sigma^{0}/\sigma^{*}$ is introduced
for simplicity ($\rho^{\prime}\propto\rho$ via Eq. (3)). These expressions for
expected signal counts are inserted directly into the model $f$ of Eq. (1) as
the integrals of Gaussian Cu K X-rays, such that $\rho$ is obtained directly
from the fit along with its statistical error.
## III Results
With the likelihood method described above, we obtain
$\rho=\mbox{$I_{\text{EC}^{0}}$}/\mbox{$I_{\text{EC}^{*}}$}=\mbox{$0.9684\stackrel{{\scriptstyle\mathrm{stat}}}{{\pm}}0.0013\stackrel{{\scriptstyle\mathrm{syst}}}{{\pm}}0.0013$},$
(5)
using the $2\text{\,}\mathrm{\SIUnitSymbolMicro s}$ dataset. This result, and
those obtained using 1 and $4\text{\,}\mathrm{\SIUnitSymbolMicro s}$ CWs are
shown in Fig. 6. Our values of $I_{\text{EC}^{0}}$/$I_{\text{EC}^{*}}$ agree
across coincidence windows, and we note that these measurement uncertainties
are correlated. We also report values derived from branching ratios in
existing NNDC [9] and LNHB [8] evaluations.
Figure 6: Measurements of
$\rho=\mbox{$I_{\text{EC}^{0}}$}/\mbox{$I_{\text{EC}^{*}}$}$ obtained at three
different coincidence windows (red points) are compared to expected values
based on most-recent LNHB [8] and NNDC [9] evaluations. The uncertainties on
our measurements are correlated.
The systematic errors considered in this analysis generally fall into two
categories: (1) physical limitations (finitely known quantities such as
tagging efficiency; App. B.2) and (2) spectral characteristics (e.g. fit
range). The former (1) are accounted for analytically using equations of Sec.
II.2. All physical parameters are varied simultaneously to account for
covariances, and are assumed to follow Gaussian distributions. The latter (2)
are gauged by performing numerous fits, such as that in Fig. 5, while randomly
varying the binning and fit range. The final systematic error of $1.3\times
10^{-3}$ sums those of category (1) and (2) in quadrature, and within rounding
happens to be equivalent to the statistical error. We find that the physical
systematics completely dominate any spectral effects, with the leading
contribution stemming from the error on MTAS $\gamma$-tagging efficiency.
Details of these considerations are provided in App. C, and contributions of
individual sources of systematic error are summarized in Table 6.
We re-evaluate the decay scheme of 65Zn (Fig. 1) by combining our result of
$I_{\text{EC}^{0}}$/ $I_{\text{EC}^{*}}$ with the commonly evaluated branching
ratio $\mbox{$I_{\beta^{+}}$}=1.421(7)\%$ [9, 8], and assuming these three
branches complete the decay scheme. Our result is not sensitive to internal
conversion and pair formation, internal Bremsstrahlung, de-excitation through,
or EC to the intermediate $770\text{\,}\mathrm{keV}$ Cu level. The evaluated
$I_{\text{EC}^{*}}$ is thus equivalent to the absolute
$1115\text{\,}\mathrm{keV}$ $\gamma$ intensity.
With our re-evaluation, we find an improvement in sensitivity to the
$1115\text{\,}\mathrm{keV}$ branching ratio of almost a factor of 2 relative
to that obtained by the Euromet exercise [10, 11]. We compare this result of
$\mbox{$\mathrm{I}($1115\text{\,}\mathrm{keV}$)$}=50.08(6)\%$ to various most-
recent measurements in Fig. 7, with values listed in Table 3. An alternate re-
evaluation combining the result of this work with a theoretical
$\mbox{$I_{\text{EC}^{0}}$}/\mbox{$I_{\beta^{+}}$}$ ratio from Betashape V2.2
[24] yields highly consistent results (Table 2).
Figure 7: Selected measured absolute intensities of the 65Zn
$1115\text{\,}\mathrm{keV}$ gamma-ray (blue points) along with the re-
evaluation of this work (dark red line). Individual measurements [25, 13, 12]
and a collective result from Euromet [11, 10] are shown, with values available
in Table 3. Additional measurements are listed in the evaluations of Refs. [8,
9].
## IV Conclusions
We have re-determined the absolute emission intensity of
$1115\text{\,}\mathrm{keV}$ $\gamma$s originating from 65Zn electron-capture
decays, improving the available precision by a factor of $\sim$2, with a
result that agrees with the existing National Nuclear Data Center evaluation.
This improvement stems from precise determination of a ratio of electron-
capture decay intensities, which does not require precise knowledge of source
activity. As in other experiments, the efficiency of our $\gamma$-tagger is a
limiting systematic, though this precisely known quantity yields a systematic
error roughly equivalent to our statistical uncertainty. Additionally, our
determination is the first involving the total EC0 branching ratio of 65Zn.
Our successful measurement is analogous to an existing result for 40K, and may
be applied to other nuclides which decay through multiple modes of electron
capture. The unique detector configuration of the KDK experiment provides a
precise result available for use in the decay scheme evaluation of the
familiar and vastly used 65Zn.
###### Acknowledgements.
We thank Xavier Mougeot and Sylvain Leblond of LNHB for input on 65Zn. Work
was performed at Oak Ridge National Laboratory, managed by UT-Battelle, LLC,
for the U.S. Department of Energy under Contract DE- AC05-00OR22725. The
United States Government retains and the publisher, by accepting the article
for publication, acknowledges that the United States Government retains a non-
exclusive, paid-up, irrevocable, world-wide license to publish or reproduce
the published form of this manuscript, or allow others to do so, for United
States Government purposes. The Department of Energy will provide public
access to these results of federally sponsored research in accordance with the
DOE Public Access Plan (http://energy.gov/downloads/doe- public-access-plan).
US support has also been supplied by the Joint Institute for Nuclear Physics
and Applications, and by NSF grant EAR-2102788. Funding in Canada has been
provided by NSERC through a SAPIN grant, as well as by the Faculty of Arts and
Science of Queen’s University, and by the McDonald Institute.
## References
* Hoppes and Schima [1984] D. D. Hoppes and F. J. Schima, _Nuclear data for the efficiency calibration of Germanium spectrometer systems: measurements from the laboratories of the International Committee for Radionuclide Metrology $\alpha-,\beta-$ and $\gamma-$ Spectrometry Group_, Tech. Rep. NBS Special Publication 626 (National Bureau of Standards, Washington, DC, USA, 1984).
* Banks _et al._ [1954] T. E. Banks, R. Tupper, R. W. E. Watts, and A. Wormall, Estimation of zinc in biological specimens: Determination of Zinc-65, Nature 173, 348 (1954).
* Hunt _et al._ [2002] J. R. Hunt, L. K. Johnson, and B. O. Juliano, Bioavailability of zinc from cooked Philippine milled, undermilled, and brown rice, as assessed in rats by using growth, bone zinc, and zinc-65 retention, Journal of Agricultural and Food Chemistry 50, 5229 (2002).
* Sugita _et al._ [2014] R. Sugita, N. I. Kobayashi, A. Hirose, K. Tanoi, and T. M. Nakanishi, Evaluation of _in vivo_ detection properties of 22Na, 65Zn, 86Rb, 109Cd and 137Cs in plant tissues using real-time radioisotope imaging system, Physics in Medicine & Biology 59, 837 (2014).
* Lucconi _et al._ [2013] G. Lucconi, G. Cicoria, D. Pancaldi, F. Lodi, C. Malizia, S. Fanti, S. Boschi, and M. Marengo, Use of 65Zn as a tracer for the assessment of purification in the 68Ga-DOTANOC synthesis, Applied Radiation and Isotopes 80, 27 (2013).
* Korinko and Tosten [2013] P. S. Korinko and M. H. Tosten, Analysis of zinc 65 contamination after vacuum thermal process, Journal of Failure Analysis and Prevention 13, 389 (2013).
* Pommé _et al._ [2016] S. Pommé, H. Stroh, J. Paepen, R. Van Ammel, M. Marouli, T. Altzitzoglou, M. Hult, K. Kossert, O. Nähle, H. Schrader, _et al._ , On decay constants and orbital distance to the Sun – part III: beta plus and electron capture decay, Metrologia 54, 36 (2016).
* Bé _et al._ [2006] M.-M. Bé, V. Chisté, C. Dulieu, E. Browne, C. Baglin, V. Chechev, N. Kuzmenko, R. Helmer, F. Kondev, D. MacMahon, and K. Lee, _Table of Radionuclides_, Monographie BIPM-5, Vol. 3 (Bureau International des Poids et Mesures, Pavillon de Breteuil, F-92310 Sèvres, France, 2006).
* Browne and Tuli [2010] E. Browne and J. Tuli, Nuclear data sheets for A = 65, Nuclear Data Sheets 111, 2425 (2010).
* Bé _et al._ [2005] M.-M. Bé, M. N. Amiot, C. Bobin, M. C. Lepy, J. Plagnard, J. M. Lee, K. B. Lee, T. S. Park, A. Luca, M. Sahagia, A. M. Razdolescu, L. Grigorescu, Y. Sato, Y. Hino, K. Kossert, R. Klein, M. H. K. Schneider, H. Schrader, P. Dryak, J. Sochorova, P. Kovar, P. Auerbach, M. Havelka, T. Altzitzoglou, A. Iwahara, M. A. L. Da Silva, J. U. Delgado, C. J. Da Silva, L. Johansson, S. Collins, and A. Stroak, _Activity measurements and gamma emission intensities determination in the decay of Zn 65_, Tech. Rep. CEA-R-6081 (CEA Saclay, Dept. des Technologies du Capteur et du Signal, Gif-sur-Yvette, France, 2005).
* Bé [2006] M.-M. Bé, Activity measurements and determination of gamma-ray emission intensities in the decay of 65Zn, Applied Radiation and Isotopes 64, 1396 (2006).
* Iwahara _et al._ [2005] A. Iwahara, M. A. L. da Silva, A. E. Carvalho Filho, E. M. de Oliveira Bernardes, and J. U. Delgado, Determination of disintegration rates and $\gamma$-ray emission probabilities of 65Zn and 241Am, Applied Radiation and Isotopes 63, 107 (2005).
* Luca _et al._ [2003] A. Luca, M.-N. Amiot, and J. Morel, Determination of half-life and photon emission probabilities of 65Zn, Applied Radiation and Isotopes 58, 607 (2003).
* Funck and Larsen [1983] E. Funck and A. N. Larsen, The influence from low energy x-rays and auger electrons on 4$\pi$$\beta$-$\gamma$ coincidence measurements of electron-capture-decaying nuclides, The International Journal of Applied Radiation and Isotopes 34, 565 (1983).
* Di Stefano _et al._ [2020] P. C. F. Di Stefano, N. Brewer, A. Fijalkowska, Z. Gai, K. C. Goetz, R. Grzywacz, D. Hamm, P. Lechner, Y. Liu, and E. Lukosi, The KDK (potassium decay) experiment, in _Journal of Physics: Conference Series_, Vol. 1342 (IOP Publishing, 2020) p. 012062, xV International Conference on Topics in Underground and Astroparticle Physics, June 24-28, 2017 (Sudbury, ON, Canada).
* Stukel _et al._ [2021] M. Stukel, B. Rasco, N. Brewer, P. Di Stefano, K. Rykaczewski, H. Davis, E. Lukosi, L. Hariasz, M. Constable, P. Davis, K. Dering, A. Fijałkowska, Z. Gai, K. Goetz, R. Grzywacz, J. Kostensalo, J. Ninkovic, P. Lechner, Y. Liu, M. Mancuso, C. Melcher, F. Petricca, C. Rouleau, P. Squillari, L. Stand, D. Stracener, J. Suhonen, M. Wolińska-Cichocka, and I. Yavin, A novel experimental system for the KDK measurement of the 40K decay scheme relevant for rare event searches, NIM A 1012, 165593 (2021).
* Stukel _et al._ [2023] M. Stukel, L. Hariasz, P. C. F. Di Stefano, B. C. Rasco, K. P. Rykaczewski, N. T. Brewer, D. W. Stracener, Y. Liu, Z. Gai, C. Rouleau, J. Carter, J. Kostensalo, J. Suhonen, H. Davis, E. D. Lukosi, K. C. Goetz, R. K. Grzywacz, M. Mancuso, F. Petricca, A. Fijałkowska, M. Wolińska-Cichocka, J. Ninkovic, P. Lechner, R. B. Ickert, L. E. Morgan, P. R. Renne, and I. Yavin, Rare 40K decay with implications for fundamental physics and geochronology, Physical Review Letters 131, 052503 (2023).
* Hariasz _et al._ [2023] L. Hariasz, M. Stukel, P. C. F. Di Stefano, B. C. Rasco, K. P. Rykaczewski, N. T. Brewer, D. W. Stracener, Y. Liu, Z. Gai, C. Rouleau, J. Carter, J. Kostensalo, J. Suhonen, H. Davis, E. D. Lukosi, K. C. Goetz, R. K. Grzywacz, M. Mancuso, F. Petricca, A. Fijałkowska, M. Wolińska-Cichocka, J. Ninkovic, P. Lechner, R. B. Ickert, L. E. Morgan, P. R. Renne, and I. Yavin, First observation of the ground-state electron-capture of 40K, Physical Review C 108, 014327 (2023).
* Helmer and van der Leun [2000] R. Helmer and C. van der Leun, Recommended standards for $\gamma$-ray energy calibration (1999), NIM A 450, 35 (2000).
* Wang _et al._ [2021] M. Wang, W. Huang, F. Kondev, G. Audi, and S. Naimi, The AME 2020 atomic mass evaluation (II). Tables, graphs and references*, Chinese Physics C 45, 030003 (2021).
* Rasco _et al._ [2015] B. C. Rasco, A. Fijałkowska, M. Karny, K. P. Rykaczewski, M. Wolińska-Cichocka, K. C. Goetz, R. K. Grzywacz, C. J. Gross, K. Miernik, and S. V. Paulauskas, Multiple $\gamma$ emission of the 137Xe $2849-2850$ keV levels studied with the modular total absorption spectrometer (MTAS), in _Proceedings of the Conference on Advances in Radioactive Isotope Science (ARIS2014)_ (JPS Journals, 2015) p. 030018.
* Karny _et al._ [2016] M. Karny, K. P. Rykaczewski, A. Fijałkowska, B. C. Rasco, M. Wolińska-Cichocka, R. K. Grzywacz, K. C. Goetz, D. Miller, and E. F. Zganjar, Modular total absorption spectrometer, NIM A 836, 83 (2016).
* Baker and Cousins [1984] S. Baker and R. D. Cousins, Clarification of the use of chi-square and likelihood functions in fits to histograms, NIM Physics Research 221, 437 (1984).
* Mougeot [2017] X. Mougeot, BetaShape: A new code for improved analytical calculations of beta spectra, _European Physical Journal Web of Conferences_ , European Physical Journal Web of Conferences European Physical Journal Web of Conferences, 146, 12015 (2017), nD2016: International Conference on Nuclear Data for Science and Technology, September 11-15, 2016 (Bruges, Belgium).
* Schötzig [1990] U. Schötzig, Photon emission probabilities of 44Ti, 65Zn, 88Y, 89Sr, 147Pm, 204Ti and 210Pb, NIM A 286, 523 (1990).
* Kibédi _et al._ [2008] T. Kibédi, T. W. Burrows, M. B. Trzhaskovskaya, P. M. Davidson, and C. W. Nestor, Evaluation of theoretical conversion coefficients using BrIcc, NIM A 589, 202 (2008).
* Bambynek _et al._ [1977] W. Bambynek, H. Behrens, M. H. Chen, B. Crasemann, M. L. Fitzpatrick, K. W. D. Ledingham, H. Genz, M. Mutterer, and R. L. Intemann, Orbital electron capture by the nucleus, Reviews of Modern Physics 49, 77 (1977).
* Colgate and Gilbert [1953] S. A. Colgate and F. C. Gilbert, Electron-positron annihilation in flight, Physical Review 89, 790 (1953).
* Schönfeld and Janßen [1996] E. Schönfeld and H. Janßen, Evaluation of atomic shell data, NIM A 369, 527 (1996).
## Appendix A Components of the 65Zn Decay Scheme and Re-evaluated Branching
Ratios
The decay scheme of 65Zn (Fig. 1) can be constructed either wholly
empirically, or with a combination of measurements and theoretical values.
Prior to this work, the purely empirical evaluation relied on intensity
measurements of $1115\text{\,}\mathrm{keV}$ and $511\text{\,}\mathrm{keV}$
$\gamma$s associated with 65Zn.
The $1115\text{\,}\mathrm{keV}$ $\gamma$ intensity is equivalent to the
branching ratio to the $1115\text{\,}\mathrm{keV}$ Cu state
($I_{\text{EC}^{*}}$) assuming a simplified scheme where (1) the intermediate
344, 770 keV $\gamma$ emissions are negligible, and (2) internal conversion is
negligible. The order of precision for $I_{\text{EC}^{*}}$ from both previous
evaluations [8, 9] is $\sim\mathcal{O}(10^{-3})$, and the
$\mathcal{O}(10^{-5})$ intensities of intermediate $\gamma$s [10] along with
the $\mathcal{O}(10^{-4})$ internal conversion process (BrIcc program [26])
can be neglected. Radiative electron capture (REC) may accompany both the EC0
and EC∗ 65Zn decays, resulting in internal Bremsstrahlung emission. However,
for these allowed transitions, the frequency of REC relative to non-REC decays
is of the order of $10^{-4}$ [27] and can be neglected in the determination of
$I_{\text{EC}^{*}}$.
The branching ratio through $\beta^{+}$ disintegration ($I_{\beta^{+}}$) is
informed from measured $511\text{\,}\mathrm{keV}$ intensities obtained taking
annihilation-in-flight [28] into account. The decay scheme is then built with
$1=\mbox{$I_{\text{EC}^{*}}$}+\mbox{$I_{\text{EC}^{0}}$}+\mbox{$I_{\beta^{+}}$},$
(6)
which yields the final required branching ratio ($I_{\text{EC}^{0}}$). To re-
evaluate the scheme with the result of this work, the above assumption of
unitarity is maintained and the measured parameter
$\rho=\mbox{$I_{\text{EC}^{0}}$}/\mbox{$I_{\text{EC}^{*}}$}$ is inserted,
$1=\mbox{$I_{\text{EC}^{*}}$}(1+\rho)+\mbox{$I_{\beta^{+}}$}.$ (7)
Above, $\mbox{$I_{\beta^{+}}$}=1.4271(7)\%$ [8, 9] can be fixed to obtain
$I_{\text{EC}^{*}}$ and subsequently $I_{\text{EC}^{0}}$. We assume this
$I_{\beta^{+}}$ value is effectively uncorrelated with the EC branching
ratios. Alternatively, $\rho$ may be combined with a theoretical value of
$\rho_{B}=\mbox{$I_{\text{EC}^{0}}$}/\mbox{$I_{\beta^{+}}$}=33.4(7)$ from
Betashape V2.2 [24],
$1=\mbox{$I_{\text{EC}^{*}}$}\left(1+\rho+\rho\frac{1}{\rho_{B}}\right),$ (8)
which yields consistent branching ratios. These re-evaluations are compared to
existing ones in Table 2. Both the main and alternate methods above reduce the
uncertainty on 65Zn EC branching ratios by a factor of 2 compared to existing
evaluations [8, 9]. The re-evaluated intensity of the
$1115\text{\,}\mathrm{keV}$ $\gamma$ (=$I_{\text{EC}^{*}}$) is compared to
other measurements in Table 3 and Fig. 7. Calculations with our measurement
remain insensitive to internal conversion and intermediate gamma emissions.
Radiative electron capture does not affect our measurement of $10^{-3}$
precision. Internal Bremsstrahlung photons are detectable in both the SDD and
MTAS, and can create false positives and negatives. All such effects are
suppressed by $\mathcal{O}(10^{-4})$ relative to each radiative electron
capture branch. In the next generation of precision measurements, internal
Bremsstrahlung and other low-order effects will contribute.
| LNHB (2006) | NNDC (2010) | This work (main) | This work (alt.)
---|---|---|---|---
$I_{\text{EC}^{*}}$ (%) | $50.23(11)$ | $50.04(10)$ | $50.08(6)$ | $50.06(5)$
$I_{\text{EC}^{0}}$ (%) | $48.35(11)$ | $48.54(7)$ | $48.50(6)$ | $48.48(5)$
$I_{\beta^{+}}$ (%) | $1.421(7)$ | $1.421(7)$ | $1.421(7)$ | $1.45(3)$
Table 2: Branching ratios of 65Zn decays from existing evaluations reported by the LNHB [8] and NNDC [9] along with the re-evaluation of this work. The main re-evaluation combines the listed, adopted value of $I_{\beta^{+}}$ with the KDK measurement of $\mbox{$I_{\text{EC}^{0}}$}/\mbox{$I_{\text{EC}^{*}}$}=\mbox{$0.9684\pm 0.0018$}$. The alternate re-evaluation combines the KDK measurement with the theoretical ratio of $\mbox{$I_{\text{EC}^{0}}$}/\mbox{$I_{\beta^{+}}$}=33.4(7)$ from Betashape V2.2 [24], leading to a larger error on the resulting $I_{\beta^{+}}$ value relative to the $\sim 4$ times more-precise, measured $I_{\beta^{+}}$. Source | $I($1115\text{\,}\mathrm{keV}$)$ (%) | Rel. error (%)
---|---|---
Schötzig (1990) [25] | $50.2$ | $0.8$
Luca _et al._ (2003) [13] | $49.76$ | $0.42$
Iwahara _et al._ (2005) [12] | $49.71$ | $0.33$
Bé (2006) [11, 10] | $50.21$ | $0.20$
This work | $50.08$ | $0.12$
Table 3: Absolute intensities of the $1115\text{\,}\mathrm{keV}$ $\gamma$-ray
of 65Zn from existing measurements along with the value deduced from the decay
scheme re-evaluation of this work (Fig. 1 and App. A).
## Appendix B Likelihood details
The likelihood fit performed in Fig. 5 consists of 4 distinct spectral
components in both the coincident and anti-coincident spectra. Counts in the
Gaussian Cu Kα and Kβ peaks corresponding to 65Zn electron captures are used
to inform $\rho$ as discussed in the main text, with some details available
below. Symbol definitions are retained from the main text, where applicable.
### B.1 Coincidence sorting
Expected coincident signal counts consist of the terms,
$\Sigma^{*}=\overbrace{\sigma^{*}\epsilon}^{\text{\mbox{EC${}^{*}$},
{\hbox{\gamma}}
detected}}+\overbrace{\sigma^{*}(1-\epsilon)(1-P^{0}_{\mathrm{BG}}P^{0}_{\gamma})}^{\text{\mbox{EC${}^{*}$},
{\hbox{\gamma}} missed , BG
coincidence}}+\overbrace{\sigma(1-P^{0}_{\mathrm{BG}}P^{0}_{\gamma})}^{\text{\mbox{EC${}^{0}$},
BG coincidence}},$ (9)
which contain EC∗-originating events whose gamma-ray was successfully tagged
by MTAS, those where the gamma was missed occurring in coincidence with
another event in MTAS, and lastly EC0-originating events which occurred with
an event in MTAS. With the substitution $\sigma^{*}=\nu/(1-\rho^{\prime})$,
this expression yields Eq. (4) in the main text. The expression for expected
anti-coincident signal counts $\Sigma^{0}$ is the complement to the above such
that all events are accounted for:
$\Sigma^{*}+\Sigma=\sigma^{*}+\sigma^{0}=\nu$.
The primary source of spurious coincidences is the natural MTAS background of
rate $\beta$, corrected for via
$P^{0}_{\mathrm{BG}}=e^{-\beta\bar{t}},$ (10)
which is the probability of no occurrences within an average coincidence
window $\bar{t}$. Additionally, there is a rate of EC∗ events which are not
detected in the SDD, though the associated $\gamma$-ray is detected in MTAS in
coincidence with SDD signal. Such events have an expected rate
$\mathcal{R}_{*}=A\mbox{$I_{\text{EC}^{*}}$}(1-\eta_{*})\epsilon,$ (11)
where $A$ is the source activity and a lack of SDD X-ray detection is ensured
via
$(1-\eta_{*})=P_{K}^{*}(1-\omega_{K}\eta)+(1-P_{K}^{*}).$ (12)
At the order of this correction, the $<1\%$ probability of $\gamma$
interaction with the SDD (App. B.2) is negligible. This additional source of
spurious coincidences has a 0-event probability within $\bar{t}$ of
$P^{0}_{\gamma}=\exp(-R_{*}\bar{t})$. Altogether, the probability of any (1+)
event(s) from $\beta$ or $\mathcal{R}_{*}$ within the timescale $\bar{t}$ is
$(1-P^{0}_{\mathrm{BG}}P^{0}_{\gamma})$.
Expected MTAS background counts $\beta\bar{t}$ have been measured directly
[16], and are far larger than those from $\mathcal{R}_{*}$, as shown in Table
4. To obtain $\mathcal{R}_{*}$, known values for K-capture [24] and
fluorescence [29] probabilities are used, along with an assumed partial EC∗
activity $A\mbox{$I_{\text{EC}^{*}}$}$ from calculated source activity [16]
and $\mbox{$I_{\text{EC}^{*}}$}\sim 0.5$ with an assigned error of 10%. MTAS
$\gamma$ and SDD X-ray tagging-efficiencies have been measured elsewhere [16],
wherein average coincidence windows $\bar{t}$ are obtained from reported
quantities $\beta\bar{t},\beta$.
CW ($\mathrm{\SIUnitSymbolMicro s}$) | $\beta\bar{t}$ ($10^{-2}$) | $\mathcal{R}_{*}\bar{t}$ $(10^{-3}$)
---|---|---
1 | $0.74(1)$ | $0.7(1)$
2 | $1.25(1)$ | $1.1(2)$
4 | $2.27(2)$ | $2.1(3)$
Table 4: Expected counts leading to spurious coincidences of SDD signal with
MTAS events for a coincidence window $\bar{t}$. The natural MTAS background
($\beta\bar{t}$) dominates the effect of $\gamma$s from EC∗ events which were
missed by the SDD ($\mathcal{R}_{*}\bar{t}$).
The effect of neglecting spurious coincidences is depicted in Fig. 8. Without
this correction, results for $\rho$ are directly anti-correlated with
coincidence window. Applying the $\mathcal{O}(\text{CW})$ corrections for such
coincidences resolves the unphysical behaviour.
Figure 8: Obtaining
$\rho=\mbox{$I_{\text{EC}^{0}}$}/\mbox{$I_{\text{EC}^{*}}$}$ of 65Zn while
neglecting expected spurious coincidences of signal with background (black
squares) produces an unphysical, linear dependence on coincidence window. The
corrected results (red circles) are consistent across CWs.
### B.2 Physical quantities
Beyond coincidence sorting, any processes affecting relative production or SDD
detection of EC0 and EC∗ events must be accounted for. As per Eq. (3), $\rho$
is directly dependent on relative K-shell capture probabilities
$P_{K}^{*},P_{K}^{0}$, and SDD efficiency of tagging gammas $\eta_{\gamma}$.
K-capture probabilities from Betashape V2.2 [24] yield
$P_{K}^{0}/P_{K}^{*}=1.00690(44)$. Due to the high precision of the
measurement, this seemingly small variation in relative probabilities is the
second-most-dominant source of systematic error, as discussed further. It is
notable that previously reported [8] values from a 1996 evaluation [29] yield
a consistent ratio of $P_{K}^{0}/P_{K}^{*}=1.0067(27)$.
The SDD $\gamma$-tagging efficiency is obtained from a mixture of measurement
and simulations. The SDD K X-ray tagging efficiency obtained elsewhere [16] is
assumed to be equivalent to the geometric efficiency of the SDD; 8-9 keV Cu
X-rays penetrate the dead layers of the SDD without escaping the detector, and
all are likely to escape the source, as same-energy Auger electrons are
readily visible in the SDD (Fig. 5). Two terms are considered,
$\eta_{\gamma}=\eta\eta_{\gamma}^{i},$ (13)
where $\eta=21.5(11)\%$ is the geometric SDD efficiency [16], and
$\eta_{\gamma}^{i}$ is the probability that the $\gamma$ interacts with (does
not escape) the SDD volume.
Simulations are not relied upon for total SDD tagging efficiencies due to the
uncertainty in sub-mm, geometric source-SDD modelling. As such, only the
probability of $\gamma$ interaction with the volume it passes through is
obtained through a simulation of 10 million events, yielding
$\eta_{\gamma}^{i}=1.522(6)\%$. The SDD $\gamma$-tagging efficiency is then
$\eta_{\gamma}=0.327(17)\%$.
## Appendix C Systematic errors
Systematic errors are accounted for in two separate groups: physical sources
(1) discussed in App. B.2, and spectral characteristics (2) chosen prior to
the likelihood fit.
Every physical source of error is included in expressions for expected
coincident and anti-coincident signal counts of Eq. (4). Combined with Eq.
(3), $\rho^{\prime}=\sigma^{0}/\sigma^{*}$ and spurious coincidence
expressions (Eqs. (10)–(12)), $\rho$ is related to all physical parameters.
Values of all such parameters are listed in Table 5.
Parameter | Value | Source
---|---|---
$P_{K}^{*}$ (%) | $87.497(28)$ | Betashape V2.2 [24]
$P_{K}^{0}$ (%) | $88.101(26)$ | Betashape V2.2 [24]
$\omega_{K}$ (%) | $45.4(4)$ | Ref. [29]
$\eta$ (%) | $21.5(11)$ | Ref. [16]
$\epsilon$ (%) | $97.93(6)$ | Ref. [16]†
$\beta\bar{t}$ | $0.0125(1)$ | Ref. [16]†
$\beta$ (kHz) | $2.63951(15)$ | Ref. [16]†
$A\mbox{$I_{\text{EC}^{*}}$}$ (kBq) | $0.268(29)$ | App. B.2
$\eta_{\gamma}^{i}$ (%) | $1.522(6)$ | App. B.2
Table 5: Values and errors of physical parameters which contribute to the
overall systematic error. †Value for $2\text{\,}\mathrm{\SIUnitSymbolMicro s}$
coincidence window, with others available within Ref. [16].
All physical parameters, such as the MTAS $\gamma$-tagging efficiency, are
assumed to follow Gaussian distributions with a width corresponding to their
68% CL error. The effect of this efficiency on $\rho$ is gauged by sampling
1,000 values from the efficiency distribution, and obtaining the resulting
difference in obtained $\rho$ values. The width of the latter distribution
corresponds to the systematic error due to MTAS $\gamma$-tagging efficiency,
which is found to be $1.2\times 10^{-3}$.
To account for covariances, all physical parameters are varied simultaneously
over 10,000 iterations, resulting in a distribution of $\rho$ variation as
shown in Fig. 9. The spread of this distribution, $1.33\times 10^{-3}$, is
equivalent to the total systematic error of category (1).
Figure 9: Variation in $\rho$ induced from 10,000 instances of randomly
varying all physical parameters used to obtain the result. The standard
deviation of this distribution ($1.3\times 10^{-3}$) corresponds to the
systematic error on $\rho$. Binned counts are normalized to total iterations.
Remaining sources of systematics are of category (2), and contain the choice
of binning and fit range. Bin widths were considered between $[10,30]$ eV,
ensuring X-ray peaks remain easily distinguishable. The low end of the fit
range is constrained by the validity of the ad hoc model, as the wide Gaussian
describing primarily Auger counts (of three different energy ranges) is not
sufficient to describe data at much lower energies than pictured in Fig. 5
(data at lower energies is shown in Fig. 2). Both the low and high energy cuts
must encompass the ad hoc background in a region not dominated by X-rays. The
sets of low and high cuts considered are $[7.2,7.5]\text{
keV}\times[10.5,14.0]\text{ keV}$.
The effect of the binning is gauged by performing fits while randomly varying
the bin width from a uniform distribution bounded as described above. The
resulting distribution of $\rho$ values has a width of $\mathcal{O}(10^{-5})$.
The effect of varying the fit range is similarly small. The overall systematic
error of category (2) is equivalent to the width of the $\rho$ distribution
obtained varying all parameters of this category simultaneously. This width is
$3\times 10^{-5}$, which is negligible relative to the systematic error of
category (1). The total systematic error is obtained summing those of both
categories in quadrature to obtain $1.33\times 10^{-3}$.
A summary of systematic errors from individual sources (of both categories) is
displayed in Table 6. The dominant source of error is the MTAS
$\gamma$-tagging efficiency, which defines the order of the systematic error.
Though varying spectral characteristics can affect the ad hoc model in the
fit, any effects on $\rho$ are minimal as the signal Cu peaks of interest
dominate in counts. The statistical error of $1.25\times 10^{-3}$ is of the
same order as the overall systematic error. Added in quadrature, these two
quantities yield the overall error on $\rho$ of $1.83\times 10^{-3}$.
Source | Systematic error
---|---
MTAS $\gamma$-tagging efficiency ($\epsilon$) | $1.2\times 10^{-3}$
K-capture probabilities ($P_{K}^{*},P_{K}^{0}$) | $4.2\times 10^{-4}$
Partial $1115\text{\,}\mathrm{keV}$ activity ($A\mbox{$I_{\text{EC}^{*}}$}$) | $2.5\times 10^{-4}$
Natural MTAS backgrounds ($\beta\bar{t},\beta$) | $2.1\times 10^{-4}$
SDD-source geometric efficiency ($\eta$) | $1.8\times 10^{-4}$
Expected $\gamma$ interaction with SDD ($\eta_{\gamma}^{i}$) | $1.3\times 10^{-5}$
K fluorescence probability ($\omega_{K}$) | $1.8\times 10^{-6}$
|
Fit range | $5.2\times 10^{-5}$
Binning | $1.1\times 10^{-5}$
Table 6: Effects of individual sources of systematic error on $\rho$. The
statistical error is 0.0013. Sources of error stemming from physical
limitations (top group) have the associated symbol(s) in parentheses as used
in the text. Spectral characteristics (bottom group) have a sub-dominant
contribution to the overall systematic error of $1.33\times 10^{-3}$ obtained
in the text.
|
# Improving weakly supervised sound event detection with self-supervised
auxiliary tasks
###### Abstract
While multitask and transfer learning has shown to improve the performance of
neural networks in limited data settings, they require pretraining of the
model on large datasets beforehand. In this paper, we focus on improving the
performance of weakly supervised sound event detection in low data and noisy
settings simultaneously without requiring any pretraining task. To that
extent, we propose a shared encoder architecture with sound event detection as
a primary task and an additional secondary decoder for a self-supervised
auxiliary task. We empirically evaluate the proposed framework for weakly
supervised sound event detection on a remix dataset of the DCASE 2019 task 1
acoustic scene data with DCASE 2018 Task 2 sounds event data under 0, 10 and
20 dB SNR. To ensure we retain the localisation information of multiple sound
events, we propose a two-step attention pooling mechanism that provides a
time-frequency localisation of multiple audio events in the clip. The proposed
framework with two-step attention outperforms existing benchmark models by
22.3 %, 12.8 %, 5.9 % on 0, 10 and 20 dB SNR respectively. We carry out an
ablation study to determine the contribution of the auxiliary task and two-
step attention pooling to the SED performance improvement.111The code is
publicly released..
Index Terms: sound event detection, self-supervised learning, pooling function
## 1 Introduction
Sound Event Detection (SED) aims to determine the presence, nature and
temporal location of sound events in audio signals. Many SED algorithms rely
on strongly labelled data [1, 2, 3] for training to perform accurate event
detection and localisation. However, producing strongly labelled data for SED
is quite expensive in terms of the expertise, time and human resources
required for the annotation. This has led to the creation of weakly labelled
sound event detection dataset like Audioset [4] which contains audio clip
level annotations without the corresponding onset and offset times of the
audio events.
The weakly supervised sound event detection was first formulated as a
Multiple-Instance Learning (MIL) problem [5, 6] with the recent emergence of
Neural MIL. In Neural MIL, the first half of the network (segmentation
network) produces temporal predictions which are then aggregated by the second
half of the network (classification network) usually a pooling operator to
produce audio clip level predictions. The benefit of such formulation is,
along with detecting audio events in the clip, it provides insight into time
level localisation of those sound events in the audio clip. Since then, recent
works have focused on improving the model architecture of the segmentation
network [7, 8, 9] and developing better pooling methods [10, 11, 12, 13, 14].
However, few works have focused on how sound event detection models perform in
either limited data or noisy settings let alone in both of them.
The noisy data also affects the training of networks for sound event
detection. Specifically, the deep CNN architectures [15, 16] currently used to
provide benchmark performance for different speech and audio tasks [17]
require large labelled clean datasets to train on and when considered in a
noisy environment the performance is known to deteriorate [10]. The two
general learning strategies used as solutions are transfer learning and
multitask learning which were recently utilised for sound event detection [18,
19, 20]. However, in the multitask learning setup, it’s assumed you have
richly annotated labels for all the tasks. We investigate a counterpart of
this where only weak labels are available without any labels for the secondary
task. For this setting, we propose a self-supervised auxiliary task that will
be jointly trained with the primary task of sound event detection. The
auxiliary task is chosen to be the reconstruction of log Mel spectrogram of
audio and we show how the auxiliary task denoises internal representations and
improves network performance in noisy settings.
In all, in this paper, we address the challenge of training sound event
detection models in noisy (domestic or environmental) and limited data
settings. To that effort, we make two-fold contributions. First, identify
appropriate self-supervised auxiliary task for sound event detection in noisy
settings and demonstrate performance benefits to the same. Second, develop a
two step attention pooling mechanism that improves time-frequency localisation
of audio events and indirectly improves sound event detection performance in
noisy settings. We perform all the experiments on a standard noisy sound event
detection dataset remix [10] and release the code publicly.
Figure 1: Our proposed self-supervised learning assisted framework for weakly
supervised sound event detection. (A) The general architecture with shared
encoder and multiple decoder branches. Shared encoder, primary decoder,
auxiliary decoder is represented by $g$, $g_{2}$, $g_{4}$ respectively (B)
shows the two step attention pooling function used for primary decoder. (C)
The attention mechanism used for frequency and time attention in two step
attention pooling along different axis. (D) The CNN architecture used for
shared encoder and auxiliary decoder. The last layer is either class or
reverse convolution for encoder and decoder respectively.
## 2 Related work
A prominent recent work [10] analysed the performance of different model
architectures (segmentation network and pooling functions) under different
Signal to Noise Ratio (SNR) for sound event detection and localisation. The
paper showed that the segmentation network of type ‘VGG-like’ CNN performed
best for audio tagging and variability in performance resulted from the choice
of pooling methods with not a clear winning pooling method across different
SNR. Specifically, Global Attention Pooling outperformed other pooling methods
on some SNR and metrics, while Global Weighted Rank Pooling (GWRP) results in
the best performance on others. Still, the work on sound event detection
performance in limited data and noisy settings is sparse.
Though the various type of multitask learning methods have been greatly
explored for vision and natural language processing (NLP) tasks [21], it has
not been utilised by the audio community. Most of the works in multitask
learning for SED focus on jointly training SED with another strongly labelled
task like Sound Source Localisation (SSL) [18] or Acoustic Scene
Classification (ASC) [19, 22, 23]. A combination of multitask learning and
self-supervised learning is shown to improve performance on speech and audio
tasks [20]. However, the work uses large scale speech datasets like
Librispeech [24] as pretasks to pretrain the networks using self-supervised
learning and does not analyse the effect of noise (domestic or environmental)
on sound event detection performance.
## 3 Methodology
This section contains the details of the proposed approach for SED,
segmentation mapping network $g_{1}$, classification mapping network $g_{2}$,
and the auxiliary time-frequency reconstruction auxiliary task. The
architecture is depicted in figure 1
### 3.1 Self-supervised Learning formulation for SED
Let the raw audio be represented by $X=\\{{x_{i}}\\}_{i=1}^{T}$ where each
$x_{i}\in\mathbb{R}$ is a frame in the audio clip. We extract time-frequency
features for each audio, let them be represented by
$\hat{X}=\\{{\hat{x}_{i}}\\}_{i=1}^{T}$ where each
$\hat{x_{i}}\in\mathbb{R}^{d}$, $d\in\mathbb{Z}$ corresponds to frame in the
audio clip. In practice, d are the number of mel bins obtained after computing
the spectrogram. As per MIL formulation, we can represent each sample in
dataset as a bag $B_{j}=(\\{\hat{x}_{i}\\}_{i=1}^{T},y)|_{j=0}^{N}$ where
$y\in\mathbb{R}^{C}$ is the weak label, N are the number of samples and C are
the number of audio events. The primary task in our self-supervised framework
is SED. The segmentation mapping $g_{1}(.)$ of SED also acts as a shared
encoder for the auxiliary task. The shared encoder maps the feature set
$\\{{\hat{x}_{i}}\\}_{i=1}^{T}$ to $Z=\\{z_{i}\\}_{i=1}^{T}$ where
$z_{i}\in\mathbb{R}^{C\times F\times T}$. The second part of SED task is
network which classifies $\\{z_{i}\\}_{i=1}^{T}$ to $P=\\{p_{i}\\}_{i=1}^{C}$
where $P\in\mathbb{R}^{C}$. The network learns a mapping $g_{2}$ which maps
each audio events time-frequency segmentation to corresponding presence
probabilities of $c^{th}$ event known as $p_{k}$
$g_{1}:\hat{X}\mapsto Z\quad\quad g_{2}:Z\mapsto P$ (1)
The auxiliary self-supervised task chosen needs to help in learning robust
representations which generalise to noisy settings without requiring
additional labels. This will impact not only the learned internal
representation but also downstream sound event detection and localisation
performance. Inorder to achieve that we choose auxiliary task as
reconstruction of extracted time-frequency features for audio. By having time-
frequency reconstruction auxiliary task we hypothesise the network will learn
representations which retain audio event information better [25, 26]. We use
an auto-encoder structure for reconstruction where the encoder is shared with
the primary task of SED. If $g_{3}(.)$ is encoder mapping for reconstruction
task, we now represent $g_{1}(.)=g_{3}(.)=g(.)$ as the shared segmentation
mapping function. The second part of auxiliary task, is a decoder network
which learns a mapping $g_{4}$ such that $g_{4}:Z\mapsto\bar{X}$ where
$\bar{X}$ is the reconstructed time-frequency representation. Specifically
$\\{\bar{x_{i}}\\}_{i=1}^{T}=g_{4}(\\{z_{i}\\}_{i=1}^{T})$. Here the learned
mapping function $g_{4}(.)$ should satisfy:
$g_{4}^{-1}(g(.))=g^{-1}(g_{4}(.))=I$ (2)
To learn the function mappings satisfying primary SED task, let the objective
function be $\mathcal{L}_{1}$. To enforce the constraint of auxiliary task,
let the objective function be $\mathcal{L}_{2}$ where the aim is to minimise
the difference between T-F representation $\\{\hat{x_{i}}\\}_{i=1}^{T}$ and
predicted time-frequency representation $\\{\bar{x_{i}}\\}_{i=1}^{T}$ of audio
clip. If the learnable parameters are W = $[w,w_{2},w_{4}]$ and
$w,w_{2},w_{4}$ corresponding to $g(.),g_{2}(.),g_{4}(.)$ respectively, then
the optimisation problem can be framed in terms of these weights W over all
data points as:
$\underset{W}{\text{min}}\;\mathcal{L}_{1}(P,y|w,w_{4})+\alpha\mathcal{L}_{2}(\\{\bar{x_{i}}\\}_{i=1}^{T},\\{\hat{x_{i}}\\}_{i=1}^{T}|w,w_{2})$
(3)
The parameter alpha ($\alpha$) accounts for scale difference between losses
$\mathcal{L}_{1}$ and $\mathcal{L}_{2}$. It helps in adjusting the
contribution of auxiliary task relative to the primary task in learning
weights.
### 3.2 Shared encoder and auxiliary task decoder network
The segmentation mapping function (shared encoder) converts the time-frequency
audio input into a T-F representation for each of the audio events. The time-
frequency feature extracted for audio here is log Mel spectrogram as it has
shown to provide better performance [27, 28, 17]. We choose a CNN based
architecture similar to ‘VGG-like’ [10] for both shared encoder and auxiliary
task decoder. The shared encoder has CNN based network consists of 8 blocks of
2D Convolution, BatchNorm and ReLU with an Average Pool after every 2 blocks.
Having a common encoder helps the network to learn a shared representation by
exploiting the similarity across SED and T-F reconstruction and enables the
network to generalise better on our original task. We use a hard parameter
sharing framework to reduce the risk of overfitting [29] to limited samples.
Table 1: Weakly supervised sound event detection performance across different SNR Network | SNR 20 dB | SNR 10 dB | SNR 0 dB
---|---|---|---
encoder | pooling | aux. | micro-p | macro-p | AUC | micro-p | macro-p | AUC | micro-P | macro-p | AUC
VGGish | GAP | ✗ | 0.5067 | 0.6127 | 0.9338 | 0.4291 | 0.5390 | 0.9144 | 0.3295 | 0.4093 | 0.8694
VGGish | GMP | ✗ | 0.5390 | 0.5186 | 0.8497 | 0.5263 | 0.5023 | 0.8422 | 0.4640 | 0.4441 | 0.8189
VGGish | GWRP | ✗ | 0.7018 | 0.7522 | 0.9362 | 0.6538 | 0.7129 | 0.9265 | 0.5285 | 0.6084 | 0.8985
VGGish (dil.) | AP | ✗ | 0.7391 | 0.7586 | 0.9279 | 0.6740 | 0.7404 | 0.9211 | 0.5714 | 0.6341 | 0.9014
VGGish | 2AP | ✓ | 0.7829 | 0.7645 | 0.9390 | 0.7603 | 0.7486 | 0.9343 | 0.6986 | 0.6892 | 0.9177
The decoder of the auxiliary-task takes $Z=\\{z_{i}\\}_{i=1}^{T}$ as input and
reconstructs it to $\\{\bar{x_{i}}\\}_{i=1}^{T}$. The decoder consists of CNN
based network for combining the intermediate time-frequency representations
obtained for each audio event to an audio level time-frequency representation.
The architecture closely follows the common encoder structure in reverse order
consisting of 8 blocks of 2D Convolution, BatchNorm and ReLU with Average Pool
after every two blocks with a decreasing number of filters.
### 3.3 Primary decoder
The primary decoder is not CNN based, instead, it is a pooling operator to
satisfy MIL formulation. The choice of pooling operator has a significant
performance effect on both the SED and each audio events intermediate time-
frequency representation obtained. Global max pooling and global average
pooling results in underestimate and overestimate the audio event’s temporal
presence respectively, and to overcome this problem dynamic poolings were
proposed [10, 12, 14]. However, the developed pooling mechanisms still lacks
the granularity in temporal predictions and does not provide frequency
localisation which might be used to further disambiguate sound events. Also,
the standard attention pooling [14] is known to be unstable with cross-entropy
usually used for multi-class setup in practice.
We propose a two-step attention pooling mechanism to covert each audio events
segmentation maps $\\{z_{i}\\}_{i=1}^{T}$ into audio level predictions $P$.
The first step in the two step attention pooling takes
$Z=\\{z_{i}\\}_{i=1}^{T}$ as input. This undergoes two independent learned
linear transformation to produce classification and attention output
respectively. The attention output is squashed to ensure its valid probability
distribution. Mathematically, the attention output $Z_{a_{1}}$ and
classification output $Z_{c_{1}}$ are:
$Z_{a_{1}}=\frac{e^{\sigma(ZW_{a_{1}}^{T}+b_{a_{1}})}}{\sum_{i=1}^{F}e^{\sigma(ZW_{a_{1}}^{T}+b_{a_{1}})}}\quad
Z_{c_{1}}=(ZW_{c_{1}}^{T}+b_{c_{1}})$ (4)
This is followed by a weighted combination of classification output
$Z_{c_{1}}$ by attention weights $Z_{a_{1}}$:
$Z_{p_{1}}=\sum_{i=0}^{F}Z_{c_{1}}\cdot Z_{a_{1}}$ (5)
The time level attention is similar to frequency (first step) attention except
it operates along time axis:
$Z_{a_{2}}=\frac{e^{\sigma(Z_{p_{1}}W_{a_{2}}^{T}+b_{a_{2}})}}{\sum_{t=1}^{T}e^{\sigma(Z_{p_{1}}W_{a_{2}}^{T}+b_{a_{2}})}}\quad
Z_{p_{1}}=(ZW_{c_{2}}^{T}+b_{c_{2}})$ (6)
$Z_{p_{2}}=\sum_{t=0}^{T}Z_{c_{2}}\cdot Z_{a_{2}}$ (7)
where $Z_{p_{2}}\in[0,1]$ and denotes the presence probability of each sound
event in the audio clip. Figure 1 subsection c, provides an overview of a
single attention step. In relation to figure, $Z_{a}$, $Z_{c}$, $Z_{p}$ are
the outputs after attention matrix, classification matrix and $P(.)$
respectively in the first stage and second stage depending on subscript. By
breaking the attention into two steps, it makes the pooling more interpretable
by answering the questions of what frequency bins and what time steps
contributes to which audio events by visualising normalised attention weights
$Z_{a_{1}},Z_{a_{2}}$ and output $Z_{p_{1}},Z_{p_{2}}$. Also, the sigmoid
($\sigma$) ensures the attention output stays between 0 to 1 and avoids
unstable training for multilabel training with cross-entropy in practice.
## 4 Experiments
### 4.1 Dataset
To study the effect of noise in limited data settings, we form a noisy dataset
by mixing DCASE 2019 Task 1 of Acoustic scene classification [30] and DCASE
2018 Task 2 of General purpose Audio tagging [31]. The DCASE 2019 Task 1
provides background sounds (noise) recorded from a variety of real world
scenes in which the sounds from DCASE 2019 Task 2 are randomly embedded [10].
To ensure the noise conditions are natural, diverse and challenging, we use
the new DCASE 2019 Task 1 instead of DCASE 2018 as used in [10]. The 2019
variant extends the TUT Urban Acoustic Scenes 2018 with the other 6 cities to
a total of 12 large European cities. This results in 32000 audio clips with
8000 audio clips for each 20,10,0 dB SNR where each audio clip is of 10 secs
with background noise and three random audio events (out of total 41) in it.
### 4.2 Set up
The raw data is converted to time-frequency representation by applying FFT
with a window size of 2048 and an overlap of 1024 between windows. This is
followed by applying Mel filter banks with 64 bands and converting them to log
scale to obtain log Mel spectrogram. The network architecture used is
described in section 3.2. The entire network is trained end-to-end with a
batch size of 24 and learning of 1e-3 using Adam optimiser [32]. The code and
setup is publicly
released222https://github.com/soham97/MTL_Weakly_labelled_audio_data.
## 5 Results
### 5.1 Sound event detection
We evaluate our self-supervision assisted architecture and pooling method
against different baselines, benchmark architectures and pooling methods [14,
10]. Table 1 shows weakly supervised sound event detection performance across
different SNR of 20,10, and 0 dB. The important evaluation metric here under
consideration is micro precision (micro-p), as it uses global counts of true
positives, false negatives and false positives for metric computation against
macro precision which does simple unweighted averaging disregarding class-
imbalance. The VGGish (dil.) encoder here indicates VGGish architecture but
with dilated/atrous convolutions known to provide benchmark performance for
sound event detection, [14]. The VGGish encoder with reconstruction based
auxiliary task and two step attention pooling outperforms the existing
benchmark of atrous attention pooling [14] on SNR 20, 10 and 0 dB by 5.9%,
12.8% and 22.3% respectively. Apart from improving performance, by breaking
the attention into two steps, it allows for the intermediate use of sigmoid
which helps in ensuring the outputs don’t overflow above 1 during training.
Table 2: Ablation study to determine auxiliary task contribution auxiliary task | SNR 20 dB | SNR 10 dB | SNR 0 dB
---|---|---|---
$\alpha$ = 0.0 | 0.7772 | 0.7430 | 0.6937
$\alpha$ = 0.001 | 0.7829 | 0.7603 | 0.6986
$\alpha$ = 0.1 | 0.7637 | 0.7428 | 0.6792
### 5.2 Ablation study for auxiliary task contribution
We perform an ablation study to determine the contribution of reconstruction
auxiliary task and two step attention pooling towards the total performance
improvement. As described in Section 3.1, the total loss is:
$\displaystyle\mathcal{L}=\mathcal{L}_{1}(P,y|w,w_{4})+\alpha\mathcal{L}_{2}(\\{\bar{x_{i}}\\}_{i=1}^{T},\\{\hat{x_{i}}\\}_{i=1}^{T}|w,w_{2})$
(8)
By changing the value of $\alpha$ before training, we can adjust the
contribution of the auxiliary task to primary sound event detection. When
$\alpha=0.0$, the network has no contribution from the reconstruction
auxiliary task during training and it can be used to evaluate the performance
of two step attention pooling. In terms of micro-precision, the two step
attention pooling outperforms existing benchmark of atrous AP (row 4) from
table 1 on SNR 20, 10 and 0 dB by 5.2%, 10.2% and 21.4% respectively. By
adding the auxiliary task contribution with a relative weightage of
$\alpha=0.001$, an additional improvement of 0.7%, 2.3% and 0.7% is observed.
This indicates that two step attention has a prominent contribution in
improving the performance of sound event detection in limited data and noisy
settings, with additional performance gains from the auxiliary task. When
$\alpha$ is increased to 0.01, the performance compared to $\alpha=0.001$ is
decreased. This suggests that the auxiliary task’s loss contribution starts to
overpower the primary SED task’s loss contribution rather than improving
generalisation.
Table 3: Two top and worst performing sound events- SNR 0 dB model | aux. | bus | cowbell | gong | meow
---|---|---|---|---|---
Atrous + AP | ✗ | 0.2 | 0.781 | 0.692 | 0.583
VGGish + 2AP | ✗ | 0.572 | 0.921 | 0.643 | 0.483
VGGish + 2AP | ✓ | 0.627 | 0.94 | 0.663 | 0.532
### 5.3 SED performance on specific audio events
For almost all audio events, our proposed architectures have the best
precision scores against GMP, GAP, GWRP, Atrous across all SNR = 0, 10, 20.
Particularly, for audio events like ‘Bass drum’, ‘bus’, ‘double bass’,
‘cowbell’ the architecture outperforms other models by a large margin as shown
in table 3. However, the proposed model struggles in audio events like
‘gong’,‘chime’ and ‘meow’ where the attention pooling with dilated convolution
encoder performs better [14]. This indicates using atrous or dilated
convolutions helps in detecting audio events whose energy is spread wide in
the temporal domain. This can be incorporated into our current architecture by
replacing the linear convolutions in the shared encoder with dilated
convolutions. Further analysis and event-specific results are available in the
long version of paper 333https://arxiv.org/pdf/2008.07085.pdf and skipped due
to space constraints.
Figure 2: Visualisation of two step attention pooling and reconstruction
decoder outputs. Subplot 1 depicts the scaled log Mel spectrogram of an audio
clip. Subplot 2 is the output of the reconstruction auxiliary task. Subplots
3,4 and 5 are attention weights for the three most probable audio events in
the audio clip. Subplot 6 is the output of the first step attention pooling.
Subplot 7 and 8 is the attention weight and output of second step attention
pooling respectively. The y-axis in subplot 1-4 corresponds to Mel-bins and
sound events in subplot 5-6. The x-axis in subplot 1-7 corresponds to time and
sound events in subplot 8
### 5.4 Interpretable visualisation of audio events
Apart from improved performance, using two step attention pooling provides a
way to localise each audio event present in the audio clip along with both the
time and frequency axis. To illustrate this, we pick a random example with SNR
20 dB and show the end to end visualisation of the two step attention pooling
mechanism in figure 2. The audio under consideration has three events
occurring in it: telephone ringing, cello playing and cat meowing, with
outdoor environmental background noise. Subplot 2 in figure 2 depicts the
reconstructed Mel spectrogram of the audio clip. From the subplot, we can see
that the decoder is not only able to reconstruct the audio events clearly but
it is also denoising the log Mel spectrogram retaining the key elements of
three audio events. A future extension of work is to jointly train sound
source separation along with weakly supervised SED by using the auxiliary task
reconstruction output.
## 6 Conclusions
This paper proposes assisted self-supervised task for improving sound event
detection in limited data and noisy settings. The architecture consists of
sound event detection as a primary task with two-step attention pooling as a
primary decoder and time-frequency representation reconstruction as an
auxiliary task. We empirically evaluate the proposed framework for multi-label
weakly supervised sound event detection, on a remix DCASE 2019 and 2018
dataset under 0, 10 and 20 dB SNR. The proposed self-supervised auxiliary task
framework with two-step attention outperforms existing benchmark models by
22.3 %, 12.8 %, 5.9 % on 0, 10 and 20 dB SNR respectively. The ablation study
carried out indicates the majority of performance improvement is associated
with two step attention pooling with secondary performance improvement from
self-supervised auxiliary task. Furthermore, by using two step attention, we
can easily visualise the sound event presence along both time-frequency axis.
The code is publicly released.
## References
* [1] D. Stowell, D. Giannoulis, E. Benetos, M. Lagrange, and M. D. Plumbley, “Detection and classification of acoustic scenes and events,” _IEEE Transactions on Multimedia_ , vol. 17, no. 10, pp. 1733–1746, 2015\.
* [2] E. Çakır, G. Parascandolo, T. Heittola, H. Huttunen, and T. Virtanen, “Convolutional recurrent neural networks for polyphonic sound event detection,” _IEEE/ACM Transactions on Audio, Speech, and Language Processing_ , vol. 25, no. 6, pp. 1291–1303, 2017.
* [3] A. Mesaros, T. Heittola, and T. Virtanen, “Tut database for acoustic scene classification and sound event detection,” in _2016 24th European Signal Processing Conference (EUSIPCO)_ , 2016, pp. 1128–1132.
* [4] J. F. Gemmeke, D. P. W. Ellis, D. Freedman, A. Jansen, W. Lawrence, R. C. Moore, M. Plakal, and M. Ritter, “Audio set: An ontology and human-labeled dataset for audio events,” in _2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)_ , 2017, pp. 776–780.
* [5] T. G. Dietterich, R. H. Lathrop, and T. Lozano-Pérez, “Solving the multiple instance problem with axis-parallel rectangles,” _Artif. Intell._ , vol. 89, no. 1–2, p. 31–71, Jan. 1997. [Online]. Available: https://doi.org/10.1016/S0004-3702(96)00034-3
* [6] A. Kumar and B. Raj, “Audio event detection using weakly labeled data,” in _Proceedings of the 24th ACM International Conference on Multimedia_ , ser. MM ’16. New York, NY, USA: Association for Computing Machinery, 2016, p. 1038–1047. [Online]. Available: https://doi.org/10.1145/2964284.2964310
* [7] S.-Y. Tseng, J. Li, Y. Wang, F. Metze, J. Szurley, and S. Das, “Multiple instance deep learning for weakly supervised small-footprint audio event detection,” in _Proc. Interspeech 2018_ , 2018, pp. 3279–3283. [Online]. Available: http://dx.doi.org/10.21437/Interspeech.2018-1120
* [8] A. Kumar and B. Raj, “Deep cnn framework for audio event recognition using weakly labeled web data,” 2017, arXiv preprint, https://arxiv.org/abs/1707.02530.
* [9] Y. Xu, Q. Kong, W. Wang, and M. D. Plumbley, “Large-scale weakly supervised audio classification using gated convolutional neural network,” in _2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)_ , 2018, pp. 121–125.
* [10] Q. Kong, Y. Xu, I. Sobieraj, W. Wang, and M. D. Plumbley, “Sound event detection and time–frequency segmentation from weakly labelled data,” _IEEE/ACM Trans. Audio, Speech and Lang. Proc._ , vol. 27, no. 4, p. 777–787, Apr. 2019. [Online]. Available: https://doi.org/10.1109/TASLP.2019.2895254
* [11] S.-Y. Chou, J.-S. R. Jang, and Y.-H. Yang, “Framecnn : A weakly-supervised learning framework for frame-wise acoustic event detection and classification,” 2017, technical report, DCASE.
* [12] B. McFee, J. Salamon, and J. P. Bello, “Adaptive pooling operators for weakly labeled sound event detection,” _IEEE/ACM Trans. Audio, Speech and Lang. Proc._ , vol. 26, no. 11, p. 2180–2193, Nov. 2018. [Online]. Available: https://doi.org/10.1109/TASLP.2018.2858559
* [13] T. Su, J. Liu, and Y. Yang, “Weakly-supervised audio event detection using event-specific gaussian filters and fully convolutional networks,” in _2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)_ , 2017, pp. 791–795.
* [14] Z. Ren, Q. Kong, J. Han, M. D. Plumbley, and B. W. Schuller, “Attention-based atrous convolutional neural networks: Visualisation and understanding perspectives of acoustic scenes,” in _ICASSP 2019 - 2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)_ , 2019, pp. 56–60.
* [15] K. Simonyan and A. Zisserman, “Very deep convolutional networks for large-scale image recognition,” in _International Conference on Learning Representations_ , 2015.
* [16] K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” _2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)_ , pp. 770–778, 2016.
* [17] S. Hershey, S. Chaudhuri, D. P. W. Ellis, J. F. Gemmeke, A. Jansen, C. Moore, M. Plakal, D. Platt, R. A. Saurous, B. Seybold, M. Slaney, R. Weiss, and K. Wilson, “Cnn architectures for large-scale audio classification,” in _International Conference on Acoustics, Speech and Signal Processing (ICASSP)_ , 2017. [Online]. Available: https://arxiv.org/abs/1609.09430
* [18] W. Xue, Y. Tong, C. Zhang, G.-H. Ding, X. He, and B. Zhou, “Sound event localization and detection based on multiple doa beamforming and multi-task learning,” in _INTERSPEECH_ , 2020.
* [19] K. Imoto, N. Tonami, Y. Koizumi, M. Yasuda, R. Yamanishi, and Y. Yamashita, “Sound event detection by multitask learning of sound events and scenes with soft scene labels,” _ICASSP 2020 - 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)_ , pp. 621–625, 2020.
* [20] T. Lee, T. Gong, S. Padhy, A. Rouditchenko, and A. Ndirango, “Label-efficient audio classification through multitask learning and self-supervision,” _ArXiv_ , vol. abs/1910.12587, 2019.
* [21] Y. Zhang and Q. Yang, “A survey on multi-task learning,” 2018, preprint arXiv, https://arxiv.org/abs/1707.08114.
* [22] N. Tonami, K. Imoto, M. Niitsuma, R. Yamanishi, and Y. Yamashita, “Joint analysis of acoustic events and scenes based on multitask learning,” _2019 IEEE Workshop on Applications of Signal Processing to Audio and Acoustics (WASPAA)_ , pp. 338–342, 2019.
* [23] H. L. Bear, I. Nolasco, and E. Benetos, “Towards joint sound scene and polyphonic sound event recognition,” in _INTERSPEECH_ , 2019.
* [24] V. Panayotov, G. Chen, D. Povey, and S. Khudanpur, “Librispeech: An asr corpus based on public domain audio books,” in _2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)_ , 2015, pp. 5206–5210.
* [25] E. M. Grais and M. D. Plumbley, “Single channel audio source separation using convolutional denoising autoencoders,” in _2017 IEEE Global Conference on Signal and Information Processing (GlobalSIP)_ , 2017, pp. 1265–1269.
* [26] D. Stowell and R. E. Turner, “Denoising without access to clean data using a partitioned autoencoder,” 2015, preprint arXiv, https://arxiv.org/abs/1509.05982.
* [27] K. J. Piczak, “Environmental sound classification with convolutional neural networks,” in _2015 IEEE 25th International Workshop on Machine Learning for Signal Processing (MLSP)_ , 2015, pp. 1–6.
* [28] J. Salamon and J. P. Bello, “Deep convolutional neural networks and data augmentation for environmental sound classification,” _IEEE Signal Processing Letters_ , vol. 24, no. 3, pp. 279–283, 2017.
* [29] J. Baxter, “A bayesian/information theoretic model of learning to learn via multiple task sampling,” in _Machine Learning_ , 1997, pp. 7–39.
* [30] A. Mesaros, T. Heittola, and T. Virtanen, “A multi-device dataset for urban acoustic scene classification,” in _Proceedings of the Detection and Classification of Acoustic Scenes and Events 2018 Workshop (DCASE2018)_ , November 2018, pp. 9–13. [Online]. Available: https://arxiv.org/abs/1807.09840
* [31] E. Fonseca, M. Plakal, F. Font, D. P. W. Ellis, X. Favory, J. Pons, and X. Serra, “General-purpose tagging of freesound audio with audioset labels: Task description, dataset, and baseline,” in _Proceedings of the Detection and Classification of Acoustic Scenes and Events 2018 Workshop (DCASE2018)_ , November 2018, pp. 69–73. [Online]. Available: https://arxiv.org/abs/1807.09902
* [32] D. P. Kingma and J. Ba, “Adam: A method for stochastic optimization,” 2017, preprint arXiv, https://arxiv.org/abs/1412.6980.
|
# Mixing across stable density interfaces in forced stratified turbulence
Miles M. P. Couchman1<EMAIL_ADDRESS>Stephen M. de Bruyn Kops2 Colm-cille
P. Caulfield1,3 1Department of Applied Mathematics and Theoretical Physics,
University of Cambridge, Cambridge CB3 0WA, UK 2Department of Mechanical and
Industrial Engineering, University of Massachusetts Amherst, Amherst, MA
01003, USA 3Institute for Energy and Environmental Flows, University of
Cambridge, Cambridge CB3 0EZ, UK
###### Abstract
Understanding how turbulence enhances irreversible scalar mixing in density-
stratified fluids is a central problem in geophysical fluid dynamics. While
isotropic overturning regions are commonly the focus of mixing analyses, we
here investigate whether significant mixing may arise in anisotropic
statically-stable regions of the flow. Focusing on a single forced direct
numerical simulation of stratified turbulence, we analyze spatial correlations
between the vertical density gradient $\partial\rho/\partial z$ and the
dissipation rates of kinetic energy $\epsilon$ and scalar variance $\chi$, the
latter quantifying scalar mixing. The domain is characterized by relatively
well-mixed density layers separated by sharp stable interfaces that are
correlated with high vertical shear. While static instability is most
prevalent within the mixed layers, much of the scalar mixing is localized to
the intervening interfaces, a phenomenon not apparent if considering local
static instability or $\epsilon$ alone. While the majority of the domain is
characterized by the canonical flux coefficient
$\Gamma\equiv\chi/\epsilon=0.2$, often assumed in ocean mixing
parameterizations, extreme values of $\chi$ within the statically-stable
interfaces, associated with elevated $\Gamma$, strongly skew the bulk
statistics. Our findings suggest that current parameterizations of turbulent
mixing may be biased by undersampling, such that the most common, but not
necessarily the most significant, mixing events are overweighted. Having
focused here on a single simulation of stratified turbulence, it is hoped that
our results motivate a broader investigation into the role played by stable
density interfaces in mixing, across a wider range of parameters and forcing
schemes representative of ocean turbulence.
## 1 Introduction
In a density-stratified fluid, turbulence enhances the rate at which scalars
are irreversibly diffused throughout the flow, a process of great interest in
a variety of geophysical, environmental and industrial settings (e.g. Fernando
(1991)). Of particular importance is characterizing the role of turbulence in
the vertical transport of heat within the ocean, a crucial mechanism for
driving the required upwelling of cold bottom waters to maintain the ocean’s
vertical stratification profile and to complete global circulation currents
(Wunsch & Ferrari, 2004). Turbulence in the ocean generates dynamically
relevant motions on the order of millimeters, which cannot currently be
resolved in numerical circulation models and must therefore be parameterized,
with the choice of parameterization found to influence future climate
projections strongly (Whalen et al., 2020). Considerable observational,
numerical and theoretical work has thus been focused on developing more
accurate and universal mixing models which account for the wide range of
turbulent processes observed in different flow regimes within the ocean
(Caulfield, 2020).
The rate at which turbulence mixes a non-uniform density field is often
defined in terms of an appropriately-averaged vertical density flux
$B\equiv\left\langle\rho^{\prime}w^{\prime}\right\rangle$, where
$\rho^{\prime}$ and $w^{\prime}$ denote fluctuations in density and vertical
velocity away from the mean flow, respectively. If $B$ is to be used as a
robust indicator of irreversible mixing, it is critical that measurements of
$B$ are averaged over sufficiently large spatial volumes or time intervals, in
order to isolate irreversible diffusive processes from reversible stirring
motions (Villermaux, 2019). Stirring, occurring on relatively large scales,
may be thought of as the adiabatic rearrangement of fluid parcels of different
density induced by the underlying turbulence, which in principle is
reversible. Hence, a pointwise measurement of $B$ would not be a sufficient
indicator that irreversible mixing had occurred, as the sign of $B$ could
subsequently switch direction yielding a net flux of zero. Thus, we here use
the term mixing to refer specifically to the diffusive transport of density
across gradients that have been enhanced by such macroscopic stirring motions,
irreversibly leading the system toward a state of greater homogenization. In
order to isolate only irreversible contributions to mixing, Lorenz (1955)
introduced the concept of an available potential energy (APE). APE quantifies
the difference between a system’s current potential energy and its minimum
background potential energy (BPE) that could be achieved if fluid parcels were
adiabatically sorted into their most stable configuration. For a Boussinesq
fluid, Winters et al. (1995) demonstrated that irreversible mixing may be
described as the conversion of APE into BPE, with a system’s BPE increasing in
time as it homogenizes. Generalizing this mixing framework to compressible
flows, Tailleux (2009) argued that the mixing of a thermally-stratified fluid
should most rigorously be defined as the conversion of APE into internal
energy, which in the Boussinesq limit then exactly matches the generation of
BPE.
Given a variety of sampling limitations involved with collecting turbulence
data within the ocean, it is exceedingly difficult to perform the averaging
required to extract the irreversible component of density fluxes from direct
observational measurements of $B$. Therefore, a number of indirect methods
have been proposed that infer such fluxes from more readily available
quantities, which may be computed locally (Gregg et al., 2018). Two such
quantities, associated with what is conventionally referred to as turbulent
microstructure, are the dissipation rates of kinetic energy $\epsilon$ and
scalar variance $\chi$,
$\epsilon=\frac{\nu}{2}\left(\frac{\partial u^{\prime}_{i}}{\partial
x_{j}}+\frac{\partial u^{\prime}_{j}}{\partial
x_{i}}\right)^{2},\quad\chi=\frac{g^{2}\kappa}{\rho_{0}^{2}N^{2}}\left(\frac{\partial\rho^{\prime}}{\partial
x_{i}}\frac{\partial\rho^{\prime}}{\partial x_{i}}\right),$ (1)
representing the rates at which viscosity $\nu$ and molecular diffusivity
$\kappa$ smooth gradients in the turbulent velocity $\mathbf{u^{\prime}}$ and
density $\rho^{\prime}$ fields, respectively. In equation (1), $g$ denotes the
gravitational acceleration, $\rho_{0}$ a reference background density and
$N=\sqrt{\left(-g/\rho_{0}\right)\partial\overline{\rho}/\partial z}$ the
buoyancy frequency, defined by an appropriately averaged ambient density
gradient $\partial\overline{\rho}/\partial z$ against which the turbulence
acts. The quantities $\epsilon$ and $\chi$ are intimately related to the
irreversible processes associated with the conversion of kinetic energy and
available potential energy into internal energy, respectively, as is further
described by Caulfield (2021). In particular, for the class of direct
numerical simulation considered here, characterized by an imposed uniform
background stratification $N^{2}_{0}$, Howland et al. (2021) demonstrated that
$\chi$ computed using $N^{2}=N_{0}^{2}$ in equation (1) provides an excellent
approximation to the destruction rate of APE and is therefore a good measure
of local irreversible mixing.
As discussed by Ivey et al. (2018), $\chi$ also arguably provides the most
robust method for estimating irreversible mixing from oceanographic
measurements, since $\left\langle
B\right\rangle\simeq\left\langle\chi\right\rangle$ in steady-state provided
that averaging is performed over sufficiently long times and large volumes so
that reversible processes and transport terms are negligible (Osborn & Cox,
1972). Importantly, $\chi$ is both directly proportional to the scalar
diffusivity $\kappa$ and sign-definite, providing a robust local measure of
the irreversible fluxes associated with molecular diffusion, which does not
require the averaging of the density flux $B$ needed to filter our reversible
local stirring motions in the turbulent flow. Due to a scarcity of $\chi$
measurements, however, $\epsilon$ is more commonly used to infer mixing
following the method of Osborn (1980), which requires the introduction of a
flux coefficient $\Gamma\equiv\chi/\epsilon$ to prescribe the fraction of
turbulent kinetic energy that leads to irreversible mixing, as opposed to
being directly dissipated by viscosity. A constant value $\Gamma=0.2$ is
commonly assumed when estimating global patterns of oceanic mixing (MacKinnon
et al., 2017), which has been found to be in agreement with tracer release
experiments (Gregg et al., 2018). However, there is significant evidence
suggesting that $\Gamma$ varies appreciably in different flow regimes
(Caulfield, 2021) and so a clear physical picture has not yet emerged
explaining why $\Gamma=0.2$ is a reasonable assumption.
In the absence of measurements of $\epsilon$ or $\chi$, mixing locations are
primarily inferred from the presence of unstable overturns in vertical density
profiles, as proposed by Thorpe (1977). Assuming that the vertical extent of
an overturn is correlated with the Ozmidov length
$L_{O}=\sqrt{\epsilon/N^{3}}$, $\epsilon$ may be inferred from the measurement
of overturns which can then be converted into a flux via $\Gamma$. However,
this assumed correlation between the vertical overturning scale and $L_{O}$ is
not always robust, as has recently been discussed, for example, by Ivey et al.
(2018), Ijichi et al. (2020) and Mashayek et al. (2021). Using a forced direct
numerical simulation (DNS) similar to that considered here, Taylor et al.
(2019) quantified the errors associated with the indirect flux estimates of
Osborn & Cox (1972), Osborn (1980) and Thorpe (1977) by sparsely sampling
vertical profiles of the computational domain in order to mimic oceanographic
measurements.
Spatio-temporal intermittency in stratified turbulence greatly reduces the
applicability of classical turbulence modeling assumptions, including the
common assumption of log-normal distributions for $\epsilon$ and $\chi$ (de
Bruyn Kops, 2015). Cael & Mashayek (2021) found that global ocean measurements
of $\epsilon$ were not well approximated by an assumed log-normal distribution
but instead had a skewed right tail, indicating that a small number of extreme
events dominated the bulk statistics. By considering local correlations
between direct ocean measurements of $\epsilon$ and $\chi$, Couchman et al.
(2021) further emphasized the importance of extreme events, finding that while
the majority of the sampled domain was characterized by the canonical flux
coefficient $\Gamma=0.2$, isolated mixing events containing the largest $\chi$
were not reflected by a corresponding local increase in $\epsilon$, yielding a
dramatic increase in $\Gamma$.
Vertical layering is also known to be a canonical feature of stratified
turbulent flows, with the density field often forming ‘staircases’ of deep,
relatively well-mixed layers separated by thin interfaces with strong
gradients (Caulfield, 2021). For sufficiently stratified environments,
vertical shearing induced by the decoupling of horizontal and vertical motions
in such a layered structure becomes an important mechanism for triggering
instability and the ensuing generation of turbulence (Lilly, 1983; Billant &
Chomaz, 2001). Parameterizations of mixing based on simple domain averages are
thus unlikely to be accurate as rare extreme events and spatial heterogeneity
within the flow will be missed, a potential cause of the highly-scattered
mixing statistics currently reported throughout the literature (Gregg et al.,
2018).
In an attempt to classify such intermittency in an automatic, yet robust and
interpretable manner, Portwood et al. (2016) devised an algorithm for
splitting a snapshot from a forced DNS into three dynamically distinct
regions: quiescent regions, intermittent layers and turbulent patches. These
regions were distinguished by an increasing degree of local overturning, as
determined by computing the fraction of unstable points $\partial\rho/\partial
z>0$ within an extended neighbourhood. Local overturning fractions and
dissipation rates $\epsilon$ were found to be strongly correlated, in
agreement with the arguments of Thorpe (1977). For the relatively large filter
sizes used to segment the domain, on the order of a buoyancy length
$L_{B}=2\pi u_{h}/N$ (where $u_{h}$ denotes a characteristic horizontal
velocity scale), distributions of $\chi$ associated with each region were also
found to be correlated with $\epsilon$, although the finer spatial
distributions of $\epsilon$ and $\chi$ within each region, and the resulting
flux coefficient $\Gamma$, were not considered.
Motivated by the automated flow segmentation of Portwood et al. (2016) in
terms of unstable local density gradients $\partial\rho/\partial z>0$, and the
observation of Couchman et al. (2021) that, within the ocean, extreme events
in $\chi$ are not necessarily correlated with those in $\epsilon$, we here
analyze spatial mixing distributions within a computational domain by
considering local correlations between $\epsilon$, $\chi$ and
$\partial\rho/\partial z$. In particular, we wish to probe whether overturning
alone provides a robust indicator for local mixing, or if significant mixing
as revealed by $\chi$ might occur in other regions that would seem
inconspicuous based on consideration of only $\epsilon$ or
$\partial\rho/\partial z$.
In line with the previous investigations of Portwood et al. (2016) and Taylor
et al. (2019), we consider a forced DNS of stratified turbulence using the
methodologies presented in Almalkie & de Bruyn Kops (2012). In $\S$2, we
summarize the DNS dataset considered here and highlight the presence of a
(previously-unreported) robust vertically-aligned vortex generated by the
forcing scheme, that injects energy into the domain at large scales and
induces vertical layering in the surrounding flow. In $\S$3, we then consider
pointwise correlations between $\partial\rho/\partial z$, $\epsilon$, $\chi$
and the flux coefficient $\Gamma$, which suggests that mixing occurs not only
in overturning regions, but also in areas of local static stability. In $\S$4,
we move beyond pointwise statistics to consider extended mixing structures
within the flow, highlighting two ways in which local static instability in
the density gradient fails to be a sufficient indicator of mixing: within the
vortex a lateral density gradient is correlated with the majority of $\chi$,
and outside the vortex extreme values of $\chi$ are localized to relatively
‘sharp’ stable density interfaces at the bounding edges between overturning
layers. Finally, in $\S$5, we summarize our results and discuss implications
for parameterizing turbulent mixing within the ocean.
## 2 Summary of DNS dataset
We consider a statistically-steady, forced DNS of stratified turbulence from
the simulation campaign originally reported by Almalkie & de Bruyn Kops
(2012), and subsequently analyzed by Portwood et al. (2016) and Taylor et al.
(2019). Using a characteristic root-mean-square horizontal velocity scale
$u_{h}$, length scale $L$, and background buoyancy frequency $N_{0}$, the non-
hydrostatic Boussinesq approximation of the Navier-Stokes equations may be
written in the following dimensionless form:
$\nabla\cdot\mathbf{u}=0,\quad\frac{\partial\mathbf{u}}{\partial
t}+\mathbf{u}\cdot\nabla\mathbf{u}=-\left(\frac{2\pi}{Fr}\right)^{2}\rho\hat{\mathbf{z}}-\nabla
p+\frac{\nabla^{2}\mathbf{u}}{Re}+\mathcal{F},\quad\frac{\partial\rho}{\partial
t}+\mathbf{u}\cdot\nabla\rho-w=\frac{\nabla^{2}\rho}{RePr}.$ $None$
The governing equations ($None$) are numerically integrated using a
pseudospectral technique in a triply-periodic domain, as detailed by Almalkie
& de Bruyn Kops (2012). The dimensionless parameters governing the flow are
the Prandtl number $Pr=\nu/\kappa$, Froude number $Fr_{h}=2\pi u_{h}/(N_{0}L)$
and Reynolds number $Re_{h}=u_{h}L/\nu$. The density field satisfies
${\color[rgb]{0,0,0}\rho\left(\boldsymbol{x},t\right)=\rho_{0}(1-N_{0}^{2}z/g)+\rho^{\prime}\left(\boldsymbol{x},t\right)},$
(3)
where $\rho_{0}(1-N_{0}^{2}z/g)$ defines a time-independent, linear background
density gradient characterized by a reference density $\rho_{0}$ and an
imposed constant background buoyancy frequency $N_{0}$. Density perturbations
$\rho^{\prime}$ away from this linear background state satisfy the periodic
boundary conditions and are used to compute $\chi$ in equation (1). The
imposed constant background buoyancy frequency $N_{0}$ is used as the
characteristic ‘appropriately-averaged’ buoyancy frequency $N$ required to
compute $\chi$ in equation (1), as is widely considered the natural choice
when quantifying irreversible mixing in numerical simulations with an imposed
background stratification (see e.g. Shih et al. (2005); Maffioli et al.
(2016); Garanaik & Venayagamoorthy (2019); Portwood et al. (2019)). By
explicitly computing the available potential energy (APE) of a triply-periodic
domain with an imposed uniform background stratification $N_{0}$, Howland et
al. (2021) confirmed that normalizing $\chi$ by $N_{0}$ indeed provides an
excellent approximation to the true irreversible mixing rate as computed
through changes in the system’s APE. The forcing term $\mathcal{F}$ is
governed by the deterministic scheme denoted ‘Rf’ in Rao & de Bruyn Kops
(2011), which adds energy to horizontal motions larger than 1/8th of the
horizontal box size so as to match a target kinetic energy spectrum at small
wavenumbers. We consider a simulation characterized by $Pr=1$, $Fr_{h}=2.23$
and $Re_{h}=1271$, in a domain of size $2\pi\times 2\pi\times\pi$ with
$4096\times 4096\times 2048$ grid points, resulting in a grid spacing of
$\Delta\approx L_{K}/2$, with $L_{K}$ denoting the Kolmogorov length scale.
For reference, the characteristic buoyancy Reynolds number of the simulation
is $Re_{b}=\left\langle\epsilon\right\rangle/\nu N_{0}^{2}=50$. We consider a
single snapshot of the flow in time and all figures are displayed on grids
that have been sparsed by a factor of eight in each dimension.
Figure 1: Large-scale characteristics of the velocity and density fields
within the computational domain. a) The vertical average of the dissipation
rate $\epsilon$ at gridpoint $(x,y)$, normalized by the domain average. The
whole domain is shown, with gridpoints sparsed by a factor of eight in each
dimension. The dashed circle highlights a region of elevated $\epsilon$,
coinciding with a vortex in the velocity field, as is further examined in
Figure 3. A vertical slice of the domain at $y=200$ is considered in Figure 4.
b) The horizontal velocity normal to the plane for a vertical slice at
constant $y$ passing through the center of the dashed circle in panel a),
revealing a vertically-aligned vortex rotating counterclockwise. The grid has
been shifted in $x$ relative to panel a) so as to center the vortex. The
buoyancy length $L_{B}$ and Taylor length $L_{T}\approx 25L_{K}$ (where
$L_{K}$ is the Kolmogorov length) are marked for reference, as were the filter
sizes used in the segmentation analysis of Portwood et al. (2016). Green
contours mark stable interfaces in the density field, as shown in panel d). c)
The radially-averaged angular velocity $u_{\theta}$ and vertical component of
vorticity $\omega_{z}$ as a function of the distance $r$ from the center of
the vortex. The dashed line at $r\approx 80$ corresponds to the dashed circle
plotted in panel a). d) The density field corresponding to the vertical slice
of velocity plotted in panel b). The green lines in panels b) and d)
illustrate contours at the minimum values of the histogram of the density
field in panel e), as are marked with vertical green lines, delineating the
interfaces between relatively well-mixed density layers in the flow
surrounding the vortex.
The main characteristics of the dataset are summarized in Figure 1. The
vertically-averaged dissipation rate $\epsilon$ (Figure 1a) reveals a dominant
patch of elevated turbulence that is generated by a large-scale vertically-
aligned vortex in the velocity field, rotating counter-clockwise (Figure 1b).
Radial averages, centered on the vortex, of the angular velocity $u_{\theta}$
and vertical component of vorticity $\omega_{z}=\partial v/\partial x-\partial
u/\partial y$ are plotted in Figure 1c, indicating a Rankine-type vortex that
is approximately characterized by rigid body rotation at small radii $r$ from
the vortex core, followed by a transition to roughly irrotational flow at
larger $r$. It is important to note that such a description characterizes the
radially-averaged flow, and that smaller-scale vortical motions will still
certainly be present in the turbulent patch surrounding the vortex. A series
of horizontal currents traveling in alternating directions are found to
emanate from the vortex, characterized by a vertical scale on the order of a
buoyancy length $L_{B}$. In Figure 1d, we plot the vertical slice of the
density field that corresponds to the velocity field shown in Figure 1b,
highlighting an analogous vertically-layered structure outside of the vortex,
with relatively well-mixed density layers separated by sharp, stable
interfaces. The approximate locations of these density interfaces (delineated
by green contours) correspond to minima in the histogram of $\rho$ (Figure
1e), which highlights a strong perturbation of the density field away from its
uniform background gradient. Superimposing these density contours on the
velocity field in Figure 1b highlights that the sheared interfaces in the
velocity field are strongly correlated with the stable interfaces in the
density field. This correlation is further demonstrated in Supplementary Video
1, where rotations of the slices in Figures 1b,d around the center of the
vortex are shown. In $\S$4, we demonstrate that these interfaces,
characterized by both high shear and a strong statically-stable density
gradient, are critically important for the mixing generated outside of the
vortex.
The spontaneous formation of a persistent vortex is a key, yet previously
unreported feature of the forcing scheme of Rao & de Bruyn Kops (2011) used to
generate statistically-steady turbulence. In particular, the identification of
the vortex in Figure 1 provides insight into how the segmentation results of
Portwood et al. (2016) (see their Figure 2c), who used an identical forcing
scheme, are related to the background flow field. Specifically, the roughly
cylindrical patch of most vigorous turbulence detected by Portwood et al.
(2016), using a filter of size $L_{B}$, extends across the entire vertical
domain and almost certainly corresponds to an analogous vortical structure in
their DNS. Similarly, their ‘intermittent layers’ are primarily composed of
horizontal offshoots from the central vertically-aligned turbulent patch, and
are shaped by a similar pattern to the sheared velocity interfaces observed in
Figure 1b. As it is now evident that Portwood et al. (2016) have broadly
identified such a vortex to be a turbulent hotspot, a goal of this study is to
perform a finer analysis of mixing patterns both within and outside of the
vortex, in order to determine how patterns in the small-scale turbulent
microstructure, as described by $\epsilon$ and $\chi$, are related to the
larger-scale layered structure of the flow.
## 3 Pointwise statistics conditioned on local density gradient
Motivated by the flow segmentation of Portwood et al. (2016) in terms of the
local fraction of overturning $\partial\rho/\partial z>0$, we first consider
how the pointwise distributions of $\epsilon$, $\chi$ and
$\Gamma=\chi/\epsilon$ depend on the magnitude of $\partial\rho/\partial z$,
for both statically stable and unstable points, as shown in Figure 2. For
illustration, in Figure 2a we split the distribution of $\partial\rho/\partial
z$ into three regions: two tails containing 10$\%$ by volume of the most
stable and unstable points (coloured blue and red, respectively), and the
remaining 80$\%$ of the intermediate values (green). For such a division, we
then consider the distributions of $\epsilon$, $\chi$ and $\Gamma$ within each
region, as shown in Figures 2b-d. Although the distribution characterizing the
bulk of the domain (green) is centered around the canonical flux coefficient
$\Gamma=0.2$ (see Figure 2d), such points contain the lowest $\chi$ (Figure
2c) and are thus not of primary importance for the total mixing arising within
the computational domain. Instead, it is the extreme tails of the
$\partial\rho/\partial z$ distribution that must be considered, containing the
most significant values of $\chi$. While both the blue and red tails contain
elevated but similar distributions of $\epsilon$, they may be distinguished by
their asymmetry in $\chi$; the stable tail (blue) contains disproportionately
elevated $\chi$ as compared to $\epsilon$, and therefore some of the highest
values of $\Gamma$ within the domain.
Figure 2: Pointwise mixing statistics of the DNS data, conditioned by the
local vertical density gradient. a) Histogram of the perturbed density
gradient $\partial\rho^{\prime}/\partial z$ normalized by the magnitude of the
imposed uniform background gradient, with values greater than one indicating
local overturning. The distribution is split into three regions, by assigning
a fixed volume (here $10\%$) to each tail. Panels b)-d) illustrate the
distributions of $\epsilon$, $\chi$ and $\Gamma$, respectively, for the whole
domain (black) and the subdomains encompassed by the coloured regions in panel
a). Circles mark the median values of each distribution, and the dashed line
in panel d) indicates the canonical flux coefficient $\Gamma=0.2$ for
reference. Panels e)-g) illustrate how the medians of the respective
distributions in panels b)-d) vary with the tail volume selected in panel a).
The circles mark the medians for the segmented distributions shown in panels
b)-d) for the case of $10\%$ tail volume. For panels e) and f), the fraction
of each quantity ($\epsilon$, $\chi$) contained within each tail relative to
the entire domain is also indicated. The dashed diagonal lines mark what would
be expected for a uniform distribution of each quantity throughout the domain.
In Figures 2e-g, we then analyze how the medians of the $\epsilon$, $\chi$ and
$\Gamma$ distributions change as a function of the volume contained within the
blue and red tails, and additionally plot the relative contributions of these
tails to the domain total. Comparing the right panels of Figures 2e and 2f
reveals that $\chi$ is far more dominated by extreme events than $\epsilon$,
in agreement with the analysis of oceanographic data by Couchman et al.
(2021). For instance, when each tail contains $10\%$ volume, the stable (blue)
and unstable (red) tails each contain approximately $20\%$ of the total
$\epsilon$ in the domain, but $45\%$ and $30\%$ of the total $\chi$,
respectively. Furthermore, while the contributions to $\epsilon$ from both
tails is roughly equal, the contribution to $\chi$ from the stable tail is
always roughly $50\%$ greater than for the unstable tail. While $\Gamma=0.2$
may thus be a suitable approximation for the bulk of the domain, it may here
not be relied upon for capturing the most extreme events in $\chi$, which
dominate the bulk mixing statistics.
Additionally, the statistics in Figure 2 suggest that local instability may
not be a sufficient indicator for mixing, given the significance of the blue
stable tail. However, we note that such a conclusion cannot definitively be
drawn from the _pointwise_ distributions of $\partial\rho/\partial z$, as such
a distribution provides no information about the extended spatial environment
around each point. For example, in regions of fully-developed turbulence that
might emerge after the collapse of a shear-induced billow, there is likely a
random mixture of neighbouring unstable and stable points in close proximity
(roughly a $50\%$ mixture as identified by Portwood et al. (2016) in their
most turbulent patches), and so points within the red and blue tails of Figure
2a could be direct neighbours in space. Therefore, in $\S$4, we extend our
pointwise analysis by identifying spatially extended and coherent stable
regions, which appear to take the form of ‘interfaces’ with enhanced density
gradients. We then assess the significance of these non-overturning structures
to the overall mixing statistics.
Figure 3: Mixing patterns within the vortex. Vertical averages of the a)
dissipation rate of kinetic energy $\epsilon$, b) dissipation rate of scalar
variance $\chi$, c) vertical velocity $w$, and d) density perturbations
$\rho^{\prime}$, for the vortex region delineated in Figure 1a. The outer
green circle in panels a)-d) coincides with the black circle in Figure 1a. The
gray curve in panel d) delineates the sharp lateral gradient separating
regions of positive and negative $\rho^{\prime}$ and is found to be correlated
with the spiral distribution of $\chi$ observed in panel b). Radial averages
of e) the azimuthal velocity $u_{\theta}$, f) $\epsilon$, g) $\chi$ and h)
$\Gamma$, as a function of the distance $r$ from the center of the vortex,
with shading denoting the standard deviation around the radial mean. In panel
h),
$\left\langle\Gamma\right\rangle_{r}=\left\langle\chi\right\rangle_{r}/\left\langle\epsilon\right\rangle_{r}$
is the ratio of the red lines in panels f) and g). The vertical green dashed
lines in panels e)-h) mark the radial locations of the maximum and first zero
of the radially-averaged azimuthal velocity, and correspond to the radii of
the green dashed circles in panels a)-d).
## 4 Extended mixing structures
We now consider coherent spatial distributions of the microstructure
quantities $\epsilon$ and $\chi$, and their relation to the large-scale flow
patterns observed in Figure 1, by focusing on mixing structures arising both
within and outside of the vortex. We first perform a closer examination of
mixing within the vortex, as shown in Figure 3. Vertical averages of
$\epsilon$, $\chi$, $w$ and $\rho^{\prime}$ are plotted in Figures 3a-d,
respectively, highlighting clear differences in the spatial distributions of
$\epsilon$ and $\chi$. Such differences are further illustrated in Figures
3e-h, which show the respective radial distributions of the azimuthal velocity
$u_{\theta}$, $\epsilon$, $\chi$ and $\Gamma$ with respect to the vortex core.
These radial distributions illustrate that the inner section of the vortex,
characterized by roughly rigid-body rotation, is well-mixed and contains the
largest values of $\epsilon$ despite having minimal scalar diffusion rates
$\chi$. This observation is consistent with the density field shown in Figure
1d, where initially horizontal contours of constant density (green lines) are
strongly deflected toward the vertical before reaching the center of the
vortex, resulting in a vertically-extended core of roughly constant density
(seen predominantly in the vertical interval $25\lesssim z\lesssim 175$).
Conversely, the majority of $\chi$ is found outside the core at radii where
the angular velocity begins to decay, and is distributed in a roughly spiral
pattern (Figure 3b). Examination of the vertically-averaged perturbed density
field $\rho^{\prime}$ (Figure 3d) reveals the presence of a strong lateral
density gradient, induced by the alternating upwelling of dense fluid and
downwelling of lighter fluid within the vortex as a function of $r$.
Superimposing the position of this lateral gradient (gray) onto the
distribution of $\chi$ in Figure 3b reveals that this gradient is strongly
correlated with the spiral distribution of the most intense $\chi$. While the
vortex was identified by Portwood et al. (2016) to be a patch of vigorous
turbulence with elevated $\epsilon$ due to its generation of significant local
vertical overturning, our analysis suggests that much of the mixing within the
vortex, as quantified by $\chi$, instead results from diffusion across a
strong lateral gradient in the perturbed density field.
Outside of the vortex, the vertical homogeneity of the velocity and density
fields collapses, forming a vertically layered structure. In Figure 4, we
consider a vertical $(x,z)$ slice of the domain at position $y=200$ in Figure
1a, in order to understand how this large-scale layering pattern gives rise to
mixing at the microscale. Motivated by the significance of the stable tail
(blue) in the pointwise distribution of $\partial\rho/\partial z$ in Figure
2a, and the observation of horizontally-extended stable filaments of
$\partial\rho/\partial z$ in Figure 4a, we examine whether such structures
contribute substantially to mixing in the layered flow surrounding the vortex.
To isolate these stable filaments, we apply a Gaussian filter to the density
field with standard deviation $\sigma\approx 6L_{K}$ (corresponding to 2 grid
points in Figure 4), where $L_{K}$ denotes the Kolmogorov length scale. The
intent of such a filter is to isolate spatially-coherent stable structures
from patchy overturning regions that would contain a random assortment of
stable and unstable neighbouring points. We note that our filter length is on
the order of $10L_{K}$ as suggested by Kuo & Corrsin (1971) for removing
internal intermittency. Further, it is significantly finer than the Taylor
length $L_{T}\approx 25L_{K}$, which was the smallest filter size considered
by Portwood et al. (2016) in their identification of ‘intermittent layers’,
allowing us to examine the importance of finer-scale structures within the
flow.
Having filtered the density field (Figure 4b), we then extract the most stable
density structures by considering points in the bottom (most stable) $q$
percent of the filtered $\partial\rho/\partial z$ distribution. For
illustration, we here extract structures comprised of the most stable $q=15\%$
of points (Figure 4c), and in Appendix A demonstrate the effect of changing
this percentage. The green contours from Figures 1b,d are overlaid on Figure
4c, demonstrating that the extracted filaments correspond to segments of the
sharp interfaces separating relatively well-mixed layers in the density field.
Importantly, Figure 4d highlights that the concentration of locally-overturned
points (the segmentation indicator used by Portwood et al. (2016)) is greatest
in the regions between these stable interfaces. In Figures 4e,f, we again
highlight that these stable interfaces are also roughly correlated with
regions of high vertical shear in the layered velocity field, as are generated
by the vortex. Corresponding slices of the dissipation rates of kinetic energy
$\epsilon$ and scalar variance $\chi$ are shown in Figures 4g,h, respectively.
The spatial distribution of $\epsilon$ is seen to be much more diffuse than
that of $\chi$, with extreme values of $\chi$ being primarily concentrated
within thin filamentary structures such as those identified in Figure 4c.
Crucially, there are many examples of locations in the flow (see green
crosses, Figures 4g,h) where the stable interfaces identified in Figure 4c
contain highly-elevated local signatures of $\chi$ without a proportional
local increase in $\epsilon$. Such an observation thus raises the question as
to whether these stable interfaces contribute significantly to the total
mixing within the domain, in addition to the mixing occurring in more
conventionally-studied isotropic overturning regions (such as the large
overturn located in the vicinity of $(x,y)=(450,125)$ in Figure 4).
Figure 4: Characteristics of the layered flow outside the vortex. Vertical
slices at $y=200$ in Figure 1a of: a) the density field $\partial\rho/\partial
z$ normalized by the magnitude of the imposed background gradient; b) the
result of applying a Gaussian filter with a standard deviation of two
gridpoints ($\approx 6L_{K}$) to the density field in panel a); c) extracted
stable filaments from panel b) obtained by retaining points in the bottom
(most-stable) $15\%$ of the filtered density distribution; d) the local
fraction of unstable overturned points computed with a filter size
corresponding to the Taylor length ($L_{T}\approx 25L_{K}$), following the
method of Portwood et al. (2016); e-f) the $x$ and $y$ horizontal components
of velocity, respectively; and g-h) the dissipation rates of kinetic energy
$\epsilon$ and scalar variance $\chi$, respectively. The stable filaments from
panel c) are overlaid on panels d)-f) for reference. The contours coloured
green in panels c-f) and white in panels g-h) correspond to those plotted in
Figures 1b,d, marking the stable interfaces separating relatively well-mixed
density layers. Green crosses in panels g-h) indicate examples of locations
where $\chi$ is locally high, due to the presence of a stable density
interface, without a corresponding increase in local $\epsilon$.
We address the question of whether the identified stable filaments contribute
substantially to the total mixing occuring within the domain in Figure 5,
where we consider the relative contributions of both the vortex and the
isolated stable interfaces to the domain totals of $\epsilon$ and $\chi$. In
agreement with Portwood et al. (2016), despite occupying less than $10\%$ of
the domain volume, the vortex contributes approximately a third of the entire
domain’s $\epsilon$ and $\chi$ (red bars, Figure 5a). Outside of the vortex,
however, it is the stable interfaces that play a key role in the overall
mixing, contributing
$\frac{\textrm{(\% in interface) $\cap$ (\% outside vortex)}}{(\textrm{\%
outside vortex) }}=\frac{26\%}{26\%+40\%}=39\%$ (4)
of the total $\chi$ outside the vortex, despite appearing unremarkable based
on their much smaller contribution to $\epsilon$
($11\%/\left(11\%+55\%\right)=17\%$). Figure 4d thus highlights a key
conclusion of this study: while the concentration of overturned points is most
prevalent within the well-mixed density layers, relatively thin stable
interfaces between such relatively deep layers, which are also correlated with
high vertical shear, yield a crucial component of the bulk scalar mixing rate
$\chi$. In particular, Figures 5b,c highlight that while these interfaces may
be strongly distinguished by their distributions of $\chi$, where the median
values differ by almost an order of magnitude, they are virtually
indistinguishable based on their distributions of $\epsilon$. This mismatch
between the spatial distributions of $\epsilon$ and $\chi$ results in
significantly elevated $\Gamma$ within the interfaces, well above the
canonical value $\Gamma=0.2$ (Figure 5d). It thus appears crucial to consider
the independent information provided by the distributions of $\epsilon$ and
$\chi$ within a domain when quantifying mixing, particularly for identifying
the locations of the most extreme scalar mixing events.
Figure 5: Mixing contributions of the vortex and stable interfaces. a)
Fractional contributions to the domain total of the dissipation rates of
kinetic energy $\epsilon$ and scalar variance $\chi$, from within the vortex
(red), stable interfaces outside the vortex (dark blue), as illustrated in
Figure 4c, and the rest of the domain outside both the vortex and interfaces
(light blue). Histograms of b) $\epsilon$, c) $\chi$ and d) $\Gamma$ outside
the vortex (black), further split according to whether points are contained
within a stable interface (dark blue) or not (light blue). Dashed lines
indicate median values of the respective distributions.
## 5 Discussion
We have considered local correlations between the vertical density gradient
$\partial\rho/\partial z$ and the dissipation rates of kinetic energy
$\epsilon$ and scalar variance $\chi$ in order to characterize the spatial
distributions of mixing within a forced direct numerical simulation of
density-stratified turbulence. The forcing scheme is found to generate a
vertically-aligned vortex within the domain, largely explaining the
concentrated ‘patch’ region of vigorous turbulence reported by Portwood et al.
(2016). Outside of the vortex, the flow is characterized by a layered density
profile, with thin, highly stable interfaces separating relatively well-mixed
layers. While a mixing analysis based solely on the identification of local
overturning would deem the well-mixed layers to be of primary importance, as
in the identification of ‘intermittent layers’ with elevated $\epsilon$ by
Portwood et al. (2016), we have demonstrated that a significant fraction of
$\chi$ is localized to the edges of such layers, within the _stable_
intervening interfaces. Notably, these interfaces appear unremarkable if
looking at $\epsilon$ alone (see Figure 5b), emphasizing the importance of
$\chi$ as an independent indicator of local mixing. A number of other studies
have also highlighted that significant mixing rates may be found in regions
devoid of local overturning, emphasizing the importance of considering other
mixing mechanisms present within stratified flows. For instance, by
considering a different class of forced direct numerical simulations to those
analyzed here, Basak & Sarkar (2006) demonstrated that horizontal shear is
able to generate a complex pattern of vorticies which efficiently mix the
density field without local overturning. A striking experimental demonstration
of dye being transported across stationary, highly-stable density interfaces
has been demonstrated by Oglethorpe et al. (2013), where a “scouring” rather
than overturning dynamic generates the mixing. As the flux coefficient
$\Gamma=\chi/\epsilon$ has been found to strongly depend on the time history
of a turbulent event (Mashayek et al., 2021), it would be instructive to now
consider the time evolution and formation of the stable interfaces identified
in our study, characterized by strongly elevated $\Gamma$. For instance, as
the density interfaces are correlated with regions of high vertical shear, it
is conceivable that they might be remnants of the previous collapse of shear-
induced billows that are now only visible in signals of $\chi$ but not
$\epsilon$, as coined ‘fossil turbulence’ by Nasmyth (1970).
Our findings have two potential implications for the parameterization of ocean
mixing. Firstly, our analysis highlights the importance of adequately sampling
rare, yet extreme mixing events in a turbulent flow, as was also recently
discussed by Cael & Mashayek (2021). In agreement with the analysis of
oceanographic data by Couchman et al. (2021), Figure 2d demonstrates that
although the majority of the domain indeed appears to be well characterized by
the canonical flux coefficient $\Gamma=0.2$, significantly elevated $\Gamma$
is associated with the most extreme events in $\chi$, events that are not
reflected by a corresponding local increase in $\epsilon$. Given the current
relative sparsity of measurements within the ocean, mixing parameterizations
may thus be biased toward the most commonly measured events, which are not
necessarily the most significant. Secondly, even with perfect sampling,
different proxies for mixing are likely to yield contrasting predictions for
the amount and spatial distribution of mixing within the highly-anisotropic
layered flow considered here. For example, if measurements of $\chi$ were not
available, the stable filaments at the edges of the overturning layers (Figure
4d) would appear unremarkable, as they appear locally quiescent based on their
density gradient and are not correlated with any discernible increase in
$\epsilon$. Further, given the strong spatial variability of $\Gamma$ within
the vertically-layered flow (see Figure 5d), it is unclear what value of
$\Gamma$ should be used in the method of Osborn (1980) if trying to infer a
flux from values of $\epsilon$ measured directly by a microstructure profiler
or derived from a Thorpe overturning analysis.
As discussed by Caulfield (2021), an accurate parameterization of the flux
coefficient $\Gamma$ is likely to depend on multiple dimensionless groups
characterizing the underlying flow, such as the buoyancy Reynolds number
$Re_{b}$, Froude number $Fr$, and Prandtl number $Pr$. For instance, DNS
studies have demonstrated that bulk-averages of $\Gamma$ decrease with
increasing $Pr$ (Salehipour et al., 2015) and decreasing $Fr$ (Maffioli et
al., 2016). A promising future direction of inquiry would be to try and
rationalize such variations in $\Gamma$ in terms of differences in the
prevalence and structure of smaller-scale extreme events within the flow, such
as analyzing changes in the morphology of the stable filaments considered
here. It would also be instructive to extend our analysis to simulations of
decaying turbulence which also develop layered structures (de Bruyn Kops &
Riley, 2019), in order to establish whether the spatial distribution of mixing
events observed here changes significantly in forced versus unforced
scenarios.
Finally, following Portwood et al. (2016) and typical oceanographic
measurements, we have here primarily relied upon the local density gradient
$\partial\rho/\partial z$ to inform our analysis of spatial mixing patterns.
However, there are likely more optimal flow variables, or linear combinations
thereof, that could lead to a more robust segmentation of the turbulent domain
into distinct regimes. For example, one could imagine constructing more
insightful indicator functions of mixing from components of the velocity
gradient tensor $\partial u_{i}/\partial x_{j}$, as suggested by de Bruyn Kops
et al. (2019). Applying data-driven techniques, such as unsupervised
clustering or dimensionality reduction, to the wealth of observational,
experimental and numerical stratified turbulence data currently available has
the potential to discover automatically optimal mixing indicators free of
human bias. Such an analysis would hopefully further our understanding of the
dominant mixing mechanisms present in different flow regimes, along with their
prevalence, guiding the search for a more universal and accurate mixing
parameterization.
[Acknowledgements] This research used resources of the Oak Ridge Leadership
Computing Facility at the Oak Ridge National Laboratory, which is supported by
the Office of Science of the U.S. Department of Energy under Contract No. DE-
AC05-00OR22725. SdeBK was supported under U.S. Office of Naval Research Grant
number N00014-19-1-2152. For the purpose of open access, the authors have
applied a Creative Commons Attribution (CC BY) licence to any Author Accepted
Manuscript version arising from this submission.
[Declaration of interests]The authors report no conflict of interest.
## Appendix A Thresholding of stable filaments
The stable filaments (black) plotted in Figure 4c were extracted by
identifying points within the bottom (most stable) $q=15\%$, by volume, of the
Gaussian-filtered distribution of the density gradient $\partial\rho/\partial
z$ (Figure 4b). We here briefly consider how changing this thresholding
percentage $q$ influences the characteristics of the extracted stable
structures.
In Figure 6, we plot the stable structures that are identified by varying the
percentage $q$ from $5\%$ to $30\%$. As $q$ is increased, meaning that more
points in the stable tail of the filtered $\partial\rho/\partial z$
distribution are considered, the identified stable structures are found to
grow primarily in the horizontal direction, tracing out more of the stable
interfaces identified by the green contours from Figures 1b,d. Figure 6 thus
highlights that the magnitude of the vertical density gradient along such
stable contours is not uniform, with certain segments having stronger
gradients (as identified by using a smaller $q$) and thus being characterized
by a larger local $\chi$.
It is also natural to consider how the mixing statistics presented in Figure
5a depend on the thresholding percentage $q$. In Figure 7, considering only
the computational domain outside of the vortex, we plot the percent
contribution of the extracted interfaces to $\epsilon$ and $\chi$, as a
function of $q$. The points at $q=15\%$ correspond to the statistics presented
in Figure 5a, noting that in Figure 7 the percent contributions are normalized
by the domain total outside the vortex, and not the entire domain including
the vortex as in Figure 5a. Figure 7 demonstrates that over a wide range of
threshold percentages $q$, the identified stable filaments always contribute
over twice the amount of $\chi$ as compared to $\epsilon$.
Figure 6: Extracted stable filaments as a function of the thresholding
percentage $q$. Filaments (black) are identified as the union of points within
the bottom (most stable) $q\%$ of the filtered density distribution (see
discussion in $\S$4). The filaments detected using values of $q$ between $5\%$
and $30\%$ are shown in panels a)-f), respectively. Panel c) corresponds to
Figure 4c.
Figure 7: The percent contributions of $\epsilon$ and $\chi$ contained within
the stable interfaces identified in Figure 6, normalized by the domain totals
outside of the vortex, as a function of the thresholding percentage $q$. The
statistics at $q=15\%$ correspond to those presented in Figure 5a, noting that
in Figure 5a the contributions are normalized by the total domain including
the vortex.
## References
* Almalkie & de Bruyn Kops (2012) Almalkie, S. & de Bruyn Kops, S.M. 2012 Kinetic energy dynamics in forced, homogeneous, and axisymmetric stably stratified turbulence. Journal of Turbulence 13 (29).
* Basak & Sarkar (2006) Basak, S. & Sarkar, S. 2006 Dynamics of a stratified shear layer with horizontal shear. Journal of Fluid Mechanics 568, 19–54.
* Billant & Chomaz (2001) Billant, P. & Chomaz, J.-M. 2001 Self-similarity of strongly stratified inviscid flows. Physics of Fluids 13 (6), 1645–1651.
* de Bruyn Kops & Riley (2019) de Bruyn Kops, S.M. & Riley, J.J. 2019 The effects of stable stratification on the decay of initially isotropic homogeneous turbulence. Journal of Fluid Mechanics 860, 787–821.
* Cael & Mashayek (2021) Cael, B.B. & Mashayek, A. 2021 Log-skew-normality of ocean turbulence. Physical Review Letters 126 (22), 224502\.
* Caulfield (2020) Caulfield, C.P. 2020 Open questions in turbulent stratified mixing: Do we even know what we do not know? Physical Review Fluids 5 (11), 110518.
* Caulfield (2021) Caulfield, C.P. 2021 Layering, instabilities, and mixing in turbulent stratified flows. Annual Review of Fluid Mechanics 53, 113–145.
* Couchman et al. (2021) Couchman, M.M.P., Wynne-Cattanach, B., Alford, M.H., Caulfield, C.P., Kerswell, R.R., MacKinnon, J.A. & Voet, G. 2021 Data-driven identification of turbulent oceanic mixing from observational microstructure data. Geophysical Research Letters 48 (23), e2021GL094978.
* de Bruyn Kops (2015) de Bruyn Kops, S.M. 2015 Classical turbulence scaling and intermittency in stably stratified Boussinesq turbulence. Journal of Fluid Mechanics 775, 436–463.
* de Bruyn Kops et al. (2019) de Bruyn Kops, S.M., Saunders, D.J., Rietman, E.A. & Portwood, G.D. 2019 Unsupervised machine learning to teach fluid dynamicists to think in 15 dimensions. arXiv:1907.10035 .
* Fernando (1991) Fernando, H.J.S. 1991 Turbulent mixing in stratified fluids. Annual Review of Fluid Mechanics 23, 455–493.
* Garanaik & Venayagamoorthy (2019) Garanaik, A. & Venayagamoorthy, S.K. 2019 On the inference of the state of turbulence and mixing efficiency in stably stratified flows. Journal of Fluid Mechanics 867, 323–333.
* Gregg et al. (2018) Gregg, M.C., D’Asaro, E.A., Riley, J.J. & Kunze, E. 2018 Mixing efficiency in the ocean. Annual Review of Marine Science 10, 443–473.
* Howland et al. (2021) Howland, C.J., Taylor, J.R. & Caulfield, C.P. 2021 Quantifying mixing and available potential energy in vertically periodic simulations of stratified flows. Journal of Fluid Mechanics 914.
* Ijichi et al. (2020) Ijichi, T., St. Laurent, L., Polzin, K.L. & Toole, J.M. 2020 How variable is mixing efficiency in the abyss? Geophysical Research Letters 47 (7), e2019GL086813.
* Ivey et al. (2018) Ivey, G.N., Bluteau, C.E. & Jones, N.L. 2018 Quantifying diapycnal mixing in an energetic ocean. Journal of Geophysical Research: Oceans 123 (1), 346–357.
* Kuo & Corrsin (1971) Kuo, A.Y.S. & Corrsin, S. 1971 Experiments on internal intermittency and fine-structure distribution functions in fully turbulent fluid. Journal of Fluid Mechanics 50 (2), 285–319.
* Lilly (1983) Lilly, D.K. 1983 Stratified turbulence and the mesoscale variability of the atmosphere. Journal of Atmospheric Sciences 40 (3), 749–761.
* Lorenz (1955) Lorenz, E.N. 1955 Available potential energy and the maintenance of the general circulation. Tellus 7 (2), 157–167.
* MacKinnon et al. (2017) MacKinnon, J.A., Zhao, Z., Whalen, C.B. & Others 2017 Climate process team on internal wave–driven ocean mixing. Bulletin of the American Meteorological Society 98 (11), 2429–2454.
* Maffioli et al. (2016) Maffioli, A., Brethouwer, G. & Lindborg, E. 2016 Mixing efficiency in stratified turbulence. Journal of Fluid Mechanics 794.
* Mashayek et al. (2021) Mashayek, A., Caulfield, C.P. & Alford, M.H. 2021 Goldilocks mixing in oceanic shear-induced turbulent overturns. Journal of Fluid Mechanics 928, A1.
* Nasmyth (1970) Nasmyth, P.W. 1970 Oceanic turbulence. PhD thesis, University of British Columbia.
* Oglethorpe et al. (2013) Oglethorpe, R.L.F., Caulfield, C.P. & Woods, A.W. 2013 Spontaneous layering in stratified turbulent Taylor–Couette flow. Journal of Fluid Mechanics 721.
* Osborn (1980) Osborn, T.R. 1980 Estimates of the local rate of vertical diffusion from dissipation measurements. Journal of Physical Oceanography 10 (1), 83–89.
* Osborn & Cox (1972) Osborn, T.R. & Cox, C.S. 1972 Oceanic fine structure. Geophysical Fluid Dynamics 3 (4), 321–345.
* Portwood et al. (2019) Portwood, G.D., de Bruyn Kops, S.M. & Caulfield, C.P. 2019 Asymptotic dynamics of high dynamic range stratified turbulence. Physical Review Letters 122 (19), 194504.
* Portwood et al. (2016) Portwood, G.D., de Bruyn Kops, S.M., Taylor, J.R., Salehipour, H. & Caulfield, C.P. 2016 Robust identification of dynamically distinct regions in stratified turbulence. Journal of Fluid Mechanics 807.
* Rao & de Bruyn Kops (2011) Rao, K.J. & de Bruyn Kops, S.M. 2011 A mathematical framework for forcing turbulence applied to horizontally homogeneous stratified flow. Physics of Fluids 23 (6), 065110.
* Salehipour et al. (2015) Salehipour, H., Peltier, W.R. & Mashayek, A. 2015 Turbulent diapycnal mixing in stratified shear flows: the influence of prandtl number on mixing efficiency and transition at high reynolds number. Journal of Fluid Mechanics 773, 178–223.
* Shih et al. (2005) Shih, L.H., Koseff, J.R., Ivey, G.N. & Ferziger, J.H. 2005 Parameterization of turbulent fluxes and scales using homogeneous sheared stably stratified turbulence simulations. Journal of Fluid Mechanics 525, 193–214.
* Tailleux (2009) Tailleux, R. 2009 On the energetics of stratified turbulent mixing, irreversible thermodynamics, boussinesq models and the ocean heat engine controversy. Journal of Fluid Mechanics 638, 339–382.
* Taylor et al. (2019) Taylor, J.R., de Bruyn Kops, S.M., Caulfield, C.P. & Linden, P.F. 2019 Testing the assumptions underlying ocean mixing methodologies using direct numerical simulations. Journal of Physical Oceanography 49 (11), 2761–2779.
* Thorpe (1977) Thorpe, S.A. 1977 Turbulence and mixing in a Scottish loch. Philosophical Transactions of the Royal Society of London A 286 (1334), 125–181.
* Villermaux (2019) Villermaux, E. 2019 Mixing versus stirring. Annual Review of Fluid Mechanics 51 (1), 245–273.
* Whalen et al. (2020) Whalen, C.B., de Lavergne, C., Garabato, A.C.N., Klymak, J.M., Mackinnon, J.A. & Sheen, K.L. 2020 Internal wave-driven mixing: governing processes and consequences for climate. Nature Reviews Earth & Environment 1 (11), 606–621.
* Winters et al. (1995) Winters, K.B., Lombard, P.N., Riley, J.J. & D’Asaro, E.A. 1995 Available potential energy and mixing in density-stratified fluids. Journal of Fluid Mechanics 289, 115–128.
* Wunsch & Ferrari (2004) Wunsch, C. & Ferrari, R. 2004 Vertical mixing, energy, and the general circulation of the oceans. Annual Review of Fluid Mechanics 36, 281–314.
|
MRI]Materials Research Institute, The Pennsylvania State University,
University Park, PA 16802 MTSE]Department of Materials Science and
Engineering, The Pennsylvania State University, University Park, PA 16802
MTSE]Department of Materials Science and Engineering, The Pennsylvania State
University, University Park, PA 16802 [ICDS]Institute for Computational and
Data Sciences, The Pennsylvania State University, University Park, PA 16802
# Crystal Growth Characterization of WSe2 Thin Film Using Machine Learning
Isaiah A. Moses [ Chengyin Wu [ Wesley F. Reinhart<EMAIL_ADDRESS>[
###### Abstract
Materials characterization remains a labor-intensive process, with a large
amount of expert time required to post-process and analyze micrographs. As a
result, machine learning has become an essential tool in materials science,
including for materials characterization. In this study, we perform an in-
depth analysis of the prediction of crystal coverage in WSe2 thin film atomic
force microscopy (AFM) height maps with supervised regression and segmentation
models. Regression models were trained from scratch and through transfer
learning from a ResNet pretrained on ImageNet and MicroNet to predict
monolayer crystal coverage. Models trained from scratch outperformed those
using features extracted from pretrained models, but fine-tuning yielded the
best performance, with an impressive 0.99 $R^{2}$ value on a diverse set of
held-out test micrographs. Notably, features extracted from MicroNet showed
significantly better performance than those from ImageNet, but fine-tuning on
ImageNet demonstrated the reverse. As the problem is natively a segmentation
task, the segmentation models excelled in determining crystal coverage on
image patches. However, when applied to full images rather than patches, the
performance of segmentation models degraded considerably, while the regressors
did not, suggesting that regression models may be more robust to scale and
dimension changes compared to segmentation models. Our results demonstrate the
efficacy of computer vision models for automating sample characterization in
2D materials while providing important practical considerations for their use
in the development of chalcogenide thin films.
##### Keywords:
WSe2 thin film, Crystal coverage, Machine learning, Semantic Segmentation,
Transfer learning, Materials characterization
## 1 Introduction
Great advances are being made in the synthesis of two-dimensional materials
(2D)1, 2, 3, since the successful isolation of graphene in 20044. The
transition metal dichalcogenides (TMD) is a major class of 2D materials that
have gained much attention due to their interesting properties and potential
for applications in areas including electric and optoelectronic, energy, and
sensing3, 5. A number of synthesis methods, including mechanical exfoliation3,
powder vaporization6, 7, pulsed laser deposition8, chemical vapor deposition
(CVD), and metal organic chemical vapor deposition (MOCVD)9, 10, 11, 12 are
being deployed in a bid to improve both the quality and scalability of the
grown TMDs. Associated with the materials synthesis is the need for an
efficient characterization technique to determine the various features of the
samples, ranging from the basic crystal qualities to the determination of the
properties and potential applications of the materials13, 14.
Atomic force microscopy (AFM) is a scanning probe microscopy that is widely
applied in 2D materials characterization due to its versatile capability in
electrical, mechanical, chemical, thermal, electrochemical, and topological
characterization of samples15, 16, 17. The topological mode of the AFM is
crucial in determining the quality and properties of a sample as it is used to
produce an AFM image from which several characteristics, including crystal
coverage, domain size, shape and thickness, and nucleation density can be
determined18, 19, 10, 20, 21. Given the fundamental role the information from
the AFM image analysis plays in determining the grown sample’s quality, even
before further characterization to determine their properties and potential
applications, the fidelity and efficiency of the analysis are of major
priority in the workflow to accelerate the 2D materials qualitative and
quantitative synthesis and exploration.
The conventional approach to AFM image analysis, such as with manual image
correction in ImageJ22, is prone to inconsistencies that inherently arise from
human errors. Beyond potential inaccuracies, this laborious and time-consuming
process can become a bottleneck in the materials discovery process. Therefore,
the deployment of a method that minimizes human interference and could be
applied to thousands of images in a matter of seconds is a necessity for a
high throughput synthesis and characterization of TMDs. The application of
machine learning (ML) for AFM image analysis provides an alternative that
eliminates the limitation posted by the manual analysis23.
A number of studies have been reported on the deployment of ML models to the
AFM image analysis. Among them are the segmentation of the molecular resolved
AFM images23, classification of quasi-planar molecules that spans relevant
structural and compositional moieties in organic chemistry based on AFM
images24, identification of self-organized nanostructures25, extraction of
molecule graphs of samples from AFM images26, atomic structure recovery from
AFM images27, and quantitative analysis of MoS2 thin film micrographs.28
Crucial to the determination of the quality of the materials synthesis is the
domain size and thickness, and surface coverage29, 12, 18, 19, an isolation of
the grown crystal from the substrate on which it is grown.
The crystal coverage is a basic metric that indicates the extent to which the
thin film has grown on the substrate. A rapid and automated determination of
the crystal coverage can enhance materials synthesis as the growth parameters
can be optimized based on this figure of merit. In our present study,
convolutional regression models are developed to be deployed in determining
the crystal coverage of 2D WSe2 grown using MOCVD9. Additionally, robust
semantic segmentation models30, 31, 32, 33, 34 which give a pixel-wise
classification of the grown samples AFM images, as either belonging to the
substrate or the crystals, are trained. Our models exhibit excellent results
with $R^{2}$ exceeding 0.99 in quantification of the crystal coverage in held-
out test samples.
Furthermore, we have systematically evaluated the efficacy of different
transfer learning schemes, namely feature extraction and fine-tuning. We also
include the effects of different pretraining domains, specifically materials
micrographs compared to miscellaneous everyday objects. Our results have some
important and counter-intuitive implications on the practical implementation
of these computer vision models in materials characterization workflows.
## 2 Method
### 2.1 Dataset
Figure 1: WSe2 samples used in the study showing the growth parameters space.
$T$ is the growth chamber inner temperature, $P$ is the pressure, and time is
the growth time. Multiple micrographs are obtained for each sample, so there
are fewer unique conditions than images in our study. Bolder circles indicates
more samples at the same point. Some samples in the test set occupy unique
points in the parameter space, such as the samples at the lowest $T$. Figure
2: Sample AFM images of WSe2 thin film in our dataset. Data ingested in our
workflow have already been preprocessed by other software and include
dimensional scale bar, color scale, and text annotations.
The WSe2 AFM data used in this research were grown by Eichfeld et al.9 and
stored in the Lifetime Sample Tracking (LiST), a database hosted by the 2D
Crystal Consortium (2DCC).35 The 52 WSe2 thin film samples were synthesized
using the metal-organic chemical vapor deposition (MOCVD) technique. The
samples were grown at various conditions, including the growth time, chamber
inner temperature, and pressure (Fig. 1), resulting in significant variations
in the morphological features of the AFM micrographs obtained. Additionally,
different imaging conditions were employed for the samples, with
characterization obtained at the centers and edges of the wafer and at
different resolutions. This resulted in a total of 221 micrographs from the 52
grown samples.
The micrographs were preprocessed by another software which performed
flattening, inserted a colorbar, and annotated the images with text labels and
a scale bar. These images were finally stored as TIF files such as those shown
in Fig. 2. One important consequence of this choice is that our models were
trained not on height maps, but on height-normalized images. That is, the
relationship between pixel intensity and the original height measurement was
different within each image. The same was true of the length scale, where
pixels represented different sample area within each image. We believe this
better represents the practical use case for these models compared to
carefully controlled height and length scales.
Figure 3: Sample images with their lightness histograms demonstrating how
segmentation could be performed on the basis of a bimodal lightness
distributions (one for foreground, one for background). This assumption is
often violated due to imperfect flattening, texture, or artifacts.
The figure of merit for these thin film sample is the monolayer coverage,
which can be computed from an AFM height map according to the fraction of
pixels in the foreground compared to the overall image. This essentially
reduces the problem to a segmentation task, which has many possible solutions.
One simple method to perform binary segmentation (i.e., foreground/background
separation) is to define a lightness threshold (corresponding to a height
threshold) based on the assumption of a mixture of approximately Normal
distributions for each height range of interest (such as background and
foreground). Each image was cropped to only the AFM micrograph portion (no
padding, annotations, color bar, scale bar, etc.), a lightness histogram was
prepared, and a threshold value was selected based on an assumed bimodal
distribution, as shown in Fig. 3. Choosing this threshold produces a binary
mask for each image; these thresholds were chosen and masks evaluated manually
for each micrograph. This labeling procedure resulted in 221 image-mask pairs,
from which the monolayer coverage was computed by counting the number of
pixels above the lightness threshold (i.e., masked).
#### 2.1.1 Augmentation
A dataset consisting of only 221 images might be insufficient to effectively
train a robust ML model. Therefore, in this study, we utilized image patching,
a common data augmentation technique to generate additional data points with
greater variance in image characteristics, thus creating a more diverse
dataset for deep learning model training. We utilized the random transforms
implemented in torch-vision from the Pytorch library36 to generate the image
patches, with a final patch height and width of $224\times 224$ for regression
models and $512\times 512$ for the segmentation models. Each patch had an
equal and independent chance of being flipped vertically, horizontally,
$0-360^{\circ}$ rotation, $0.5-2.0\times$ rescale, and random crop within the
rescaled image. An example of this procedure is shown in Fig. 4. Because this
random transformation could result in out of bounds pixels, we rejected any
patch that did not fall entirely within the original image. We repeated this
sampling until 10 valid patches were obtained for each image.
Figure 4: Data augmentation and image patching schematic scheme. The original
AFM image (top left) is thresholded to produce a mask (top right). Random
image patches are jointly taken from both the image and mask to yield new
(image, mask, coverage) sets where all patches are of size $224\times 224$ but
represent different portions of the original image.
### 2.2 Regression Models
Figure 5: Schematic of different transfer learning paradigms. Feature
extraction is a scheme that only modifies the trainable weights in the fully
connected layer (or other shallow model) while leaving the pre-trained weights
in the convolutional layers unchanged. In fine tuning, all the trainable
weights from the pre-trained model are adjusted to improve the model’s fit to
the new task. In end-to-end learning, the entire model is trained from
scratch, without any knowledge transfer.
We consider two variants of the ML task: regression (predicting the coverage
label directly from the image) and segmentation (predicting the binary mask
and then computing the coverage from the mask). Within the regression task, we
further consider three training paradigms: training from scratch using end-to-
end learning (i.e., with randomly initialized weights), transfer learning by
fine-tuning (i.e., intializing the model with pretrained weights), and
transfer learning by feature extraction (i.e., training a shallow model to
predict target label with pretrained convolutional filters).
For all the regression models, Adam optimizer, ReLU activation function, and
mean squared error (MSE) loss functions were used. 10% of the data samples,
grown under different growth parameters than the rest of the data and/or
obtained under different imaging conditions, were held out to determine how
well the models generalize to out of distribution data (Table 1).
Additionally, about 80% and 10% were used for the training and validation,
respectively.
We started by training a small Convolutional Neural Network from scratch
(CNNsc). The architecture of the CNNsc network was optimized using Bayesian
hyperparameter tuning implemented in the ax-platform package37 which leverages
a Gaussian-process-based Bayesian optimization38.. After each of the
convolutional layers, a max pooling and ReLU activation function were applied
to downsize the feature maps and extract the most important features, and
introduce non-linearity, respectively. This network was deliberately
simplified compared to the pretrained models to evaluate whether fewer
trainable weights would be more robust in extrapolating to the test domain.
We also explored the application of pretrained models, specifically ResNet18
architecture pre-trained on ImageNet39 and MicroNet40 datasets, to predict the
coverage of WSe2 thin films. We chose ResNet18 as it is among the shallowest
standard computer vision architectures available today, which we felt was
important given our low data volume. The features were extracted from the
average pool layer of the pretrained models, given 512 features. MLP models
were then built to learn the crystal coverage from the image features obtained
from the ResNet18 pretrained on the ImageNet and MicroNet. The MLP models are
hereafter referred to as MLP-I and MLP-M, respectively. MLP model
hyperparameters were tuned using ax-platform as in the case of the CNNsc.
For completeness, we also employed the fine-tuning paradigm of transfer
learning. This allowed us to assess the performance of these pre-trained
models in our specific context and evaluate their potential for accurate thin
film coverage prediction. The pretrained models’ classifiers were replaced
with 2 FC layers of 512 and 100 neurons and an output layer. Between the 2 FC
layers is a ReLU activation function to introduce non-linearity and a dropout
of 0.25 to minimize over-fitting. Sigmoid activation function was additionally
placed before the output layer to ensure only values between 0.0 and 1.0
(range of coverage values) are predicted. The models were then tuned with our
data to learn the crystal coverage. The fine-tuning were carried out for the
ResNet18 pretrained on the ImageNet (CNN-I) and another on the MicroNet
(CNN-M).
### 2.3 Segmentation Models
Separately from the regression task, we attempt to solve the problem using
segmentation models to work natively with the binary mask. Similar to the
regression models, encoders pretrained on MicroNet by Stuckner et al. 40 were
used. In their report, they found ResNeXT,41 SE,42 Inception,43 and
EfficientNet44 encoder architectures to give better performances.
Additionally, Unet45 and Unet++46 decoders were found to outperform others.
Specifically, SE_ResNeXt-50_32x4d and SE_ResNeXt-101_32x4d encoders pretrained
on MicroNet coupled with Unet++ decoders gave, on the average, the best
intersection over union (IoU) accuracy for models trained on the full sets of
2 different SEM images (nickel-based superalloys and environmental barrier
coatings). We therefore used SE_ResNeXt-50_32x4d and SE_ResNeXt-101_32x4d
encoders pretrained on MicroNet coupled with Unet++ decoders in our study.
These segmentation models are termed SEG50 and SEG101, respectively.
In order for us to compare the performance of the segmentation and regression
models from the same pretrained architectures, we have additionally trained
segmentation models based on the ResNet18 pretrained encoder and using the
Unet++ decoder. Both encoders pretrained on the ImageNet and MicroNet were
used, and termed SEG18-I and SEG18-M, respectively. The Adam optimizer, 1e-4
learning rate, and a batch size of 6 were used on the training. We utilized an
early stopping after 30 epochs of training without further improvement on the
IoU accuracy of validation set, while the loss function was a weighted sum of
balanced cross entropy (BCE) and dice loss with a 70% weighting towards BCE.
## 3 Results and Discussion
### 3.1 Regression Models
#### 3.1.1 Training from Scratch
Figure 6: (a) is the CNN architecture built from scratch (CNNsc) showing the
convolutional (Conv1, Conv2, Conv3, and Conv4), the pooling (MaxPool), and
fully connected layers (FC1, FC2, FC3), as well as the feature maps and
channel sizes for each of the convolution layer and the neurons connecting the
FC layers. (b) is the root mean squared error value (RMSE) on the train and
validation (val) data against the learning iteration (epochs). (c) is the
parity plot of the predicted and target coverage. The $R^{2}$ and RMSE values
in (c) are for the test set.
The architecture of the CNNsc network found by hyperparameter tuning consisted
of four convolutional layers and three fully connected (FC) layers (Fig. 6).
The kernel size was 5 with a stride of 1 and zero padding. This model was
trained to minimize the MSE loss between the target and the predicted
coverage. A stochastic behavior is observed in the learning resulting in the
fluctuation in losses with the training iterations both for the training and
validation set (Fig. 6(b)). The random initialization of the weights might
have resulted in such behavior. To obtain an optimally trained model, the
model was set to stop once the minimal obtainable value of the training and
validation loss was achieved. This results in the model’s performance with
train, validation, and test set RMSE of 0.022, 0.062, and 0.042, respectively
(Fig. 6(c) and Table 1). These correspond to $R^{2}$ values of 0.995, 0.961,
and 0.982 for train, validation, and test, respectively. Only a few scattered
points were observed in the validation and test parity plots, indicating a
minimal over-fitting.
#### 3.1.2 Feature Extraction
Figure 7: (a), (c), and (e) are the root mean squared error value (RMSE) on
the train and validation (val) data against the learning iteration (epochs)
for the multilayer perceptron model (MLP) trained with features extracted
using the ResNet18 pretrained on the ImageNet data (MLP-I), MLP trained with
features extracted using the ResNet18 pretrained on the MicroNet data (MLP-M),
and for the CNN built from scratch without on-the-fly data augmentation
(CNNsc*), respectively. (b), (d), and (f) are the parity plots of the
predicted and target coverage corresponding to (a), (c), and (e),
respectively. The $R^{2}$ and RMSE values in (b), (d), and (f) are for the
test set.
The MLP architectures were tuned (with an objective of minimizing the
validation loss) to yield 2 hidden layers with (120, and 84) neurons in the
MLP-I and MLP-M. The trained MLP-I exhibited an $R^{2}$ value of 0.873 on the
test set (Fig. 7 and Table 1). MLP-M performs better than the MLP-I, though
still slightly worse than the CNNsc. A better performance observed in the
MLP-M than the MLP-I might be due to the proximity of the data for the
pretraining and our data; MicroNet consists of gray scale micrographs while
ImageNet is made up of the macroscale color images of natural objects. The
features extracted from the former may therefore be more relevant in learning
our image features than those from the latter.
The superior performance of the CNNsc may be due to its smaller size or its
on-the-fly data augmentations; random rotations and flips were applied to the
data while training. To verify if the data augmentations applied to the CNNsc
made a significant difference to the model performance, we trained the same
architecture of CNN with the same hyperparameters without the augmentations
(CNNsc*). The result shows that the augmentations indeed significantly enhance
the performance of the CNNsc (Fig. 7 and Table 1). Overfitting is observed to
set in soon after the first few epochs of training on data without
augmentation. The model accurately predicts the coverage for the train set but
a worse performance than both MLP-I and MLP-M is observed in the validation
and test sets.
However, the on-the-fly augmentation cannot be readily applied in the feature
extraction case as data are not seen by the model more than once. The closest
we can get to the on-the-fly augmentation is to obtain different features for
the rotated and horizontal and vertically flipped images, then training the
MLP model on all of these at once. We also tried average pooling on these
variants as input to the model rather than trying to learn a many-to-one
mapping. Both of these approaches gave worse performance compared to the
vanilla MLP models, with the augmentation giving the $R^{2}$ values of 0.86
for the MLP-I and 0.93 for the MLP-M, while the pooling strategy was worse.
These results underscores a fundamental difference in the static augmentation
of the data for the MLP models and the on-the-fly augmentation for the CNN
models.
#### 3.1.3 Fine-Tuning
Figure 8: (a), and (c) are the root mean squared error value (RMSE) on the
train and validation (val) data against the learning iteration (epochs) for
the fine-tuned ResNet18 pretrained on the ImageNet data (CNN-I), the fine-
tuned ResNet18 pretrained on the MicroNet data (CNN-M), respectively. (b) and
(d) are the parity plots of the predicted and target coverage corresponding to
(a) and (c), respectively. The $R^{2}$ and RMSE values in (b) and (d) are for
the test set.
Finally, we examined the fine-tuning of the pretrained model to predict the
crystal coverage. This approach needs to be explored especially because of our
observation of the significant impact data augmentation has on CNN model
performance. Fine-tuning is carried out for the ResNet18 pretrained on the
ImageNet and another on the MicroNet. These models are termed CNN-I and CNN-M,
respectively. As observed in the CNNsc, capturing the grokking effect is
important in obtaining the optimally trained model; the training and
validation losses were closely monitored, and the training halted once the
minimal obtainable validation loss is reached. The validation loss associated
with the grokking point was determined by an initial training for the models
for a few thousands epochs. The performance of the CNN-I and CNN-M are quite
similar, with CNN-I giving a marginally better result. Both have accurate
predictions on the validation and test set with $R^{2}$ value of 0.99 (see
Fig. 8 and Table 1).
Interestingly, while a significantly better performance is observed from
features extracted from the model pretrained on MicroNet than that from the
ImageNet, the fine-tuning shows the reverse. This means that the filters
pretrained on the MicroNet extract much useful features from the AFM than that
pretrained on the ImageNet. However, the latter scenario seems to provide more
generic image features in which case fine-tuning on sufficient target data has
yielded a better result. A nearly non-existent over-fitting, even on the held-
out test data is noteworthy. The excellent performance of CNN-I and CNN-M
underscores the advantage of not just the transfer learning but also the data
augmentations used with CNN at combating the over-fitting and producing models
that have been accurately trained on our target data which shares generic
features learned from larger data sets used for the pretraining.
#### 3.1.4 Summary of Regression Results
The results of all the regression models have been compiled in Table 1. While
comparable performance on training data can be obtained by all three learning
paradigms, their test performance vary substantially. Fine-tuning yielded the
best results in this regard, followed by training from scratch, then feature
extraction. However, this seems to have been largely a result of on-the-fly
data augmentation, as our ablation study showed that removing this from the
trained-from-scratch CNNsc led to a nearly triple test RMSE, making it the
worst model. Unfortunately, this approach could not be applied to the feature
extraction strategy to improve its performance. Between the two pretraining
domains, there was no clear winner; ImageNet gave better performance in fine-
tuning, while MicroNet was superior in feature extraction. This is not an
obvious result and may warrant further investigation regarding the nature of
the pretrained filters.
Table 1: RMSE and $R^{2}$ values for the predicted coverage on the train, validation (val), and test sets for models trained from scratch and through transfer learning. CNNsc and CNNsc* are the CNN trained from scratch with and without on the fly data augmentation, respectively. MLP-I and MLP-M are the MLP trained using the features extracted with ResNet18 architecture pretrained on ImageNet and MicroNet, respectively. CNN-I and CNN-M are the fine-tuning models of the ResNet18 architecture pretrained on ImageNet and MicroNet, respectively. The best performance in each row is shown in bold, including ties and near-ties. | From scratch | Feature extraction | Fine Tuning
---|---|---|---
RMSE | CNNsc | CNNsc* | MLP-I | MLP-M | CNN-I | CNN-M
train | 0.018 | 0.013 | 0.012 | 0.023 | 0.013 | 0.022
val | 0.039 | 0.120 | 0.098 | 0.047 | 0.021 | 0.030
test | 0.041 | 0.121 | 0.101 | 0.054 | 0.029 | 0.035
$R^{2}$ | CNNsc | CNNsc* | MLP-I | MLP-M | CNN-I | CNN-M
train | 0.997 | 0.998 | 0.998 | 0.995 | 0.998 | 0.995
val | 0.984 | 0.855 | 0.904 | 0.978 | 0.995 | 0.991
test | 0.979 | 0.818 | 0.873 | 0.963 | 0.989 | 0.984
### 3.2 Segmentation Models
Figure 9: (a) and (b) are the parity plots of coverage predicted, using the segmentation model pretrained on the MicroNet, and the target coverage. The encoder for the SEG50 and SEG101 models are SE_ResNeXt-50_32x4d and SE_ResNeXt-101_32x4d, respectively. (c) are sample images, the corresponding ground truth mask, and the predicted mask by the SEG50 and SEG101 models. The $R^{2}$ and RMSE values are for the test set. Table 2: The RMSE, $R^{2}$, and IOU values on the train, validation (val), and test data sets for the segmentation models. SEG18-I and SEG18-M uses ResNet18 pretrained on ImageNet and MicroNet, respectively. SE_ResNeXt-50_32x4d (SEG50) and SE_ResNeXt-101_32x4d (SEG101) encoder are pretrained on MicroNet data. | RMSE | $R^{2}$ | Average IoU (%)
---|---|---|---
| train | val | test | train | val | test | train | val | test
SEG18-I | 0.017 | 0.021 | 0.022 | 0.997 | 0.995 | 0.994 | 89$\pm{21}$ | 88$\pm{26}$ | 90$\pm{13}$
SEG18-M | 0.028 | 0.043 | 0.020 | 0.992 | 0.977 | 0.995 | 87$\pm{22}$ | 87$\pm{27}$ | 90$\pm{14}$
SEG50 | 0.007 | 0.024 | 0.020 | 0.999 | 0.993 | 0.995 | 92$\pm{19}$ | 90$\pm{23}$ | 92$\pm{9}$
SEG101 | 0.013 | 0.020 | 0.025 | 0.998 | 0.997 | 0.992 | 90$\pm{18}$ | 89$\pm{25}$ | 90$\pm{13}$
We now reframe the task as a binary segmentation, where crystal (foreground)
is separated from the substrate (background) and then counted to obtain the
crystal coverage. SE_ResNeXt-50_32x4d and SE_ResNeXt-101_32x4d encoders
pretrained on MicroNet coupled with Unet++ decoders are termed SEG50 and
SEG101, respectively. While ResNet18 encoder pretrained on the ImageNet and
another on the MicroNet with both coupled with the Unet++ are termed SEG18-I,
and SEG18-M, respectively. As this is natively a segmentation problem, it is
not surprising that these models can achieve excellent performance; all the
segmentation models all have marginal improvement over the regression models
as shown in Table 2. To be specific, the best model from the regression
models, CNN-I (Fig. 8 and Table 1) exhibits a test RMSE of 0.029, whereas
SEG18-M and SEG50 both obtain 0.020 RMSE.
Based on the patches of the images, it seems that segmentation models provide
higher performance in determining the crystal coverage than regression models.
Additionally, segmentation models offer the advantage of giving impressive
performances even with a much smaller data set for training40, 47, 48 since
each pixel is in effect a training data point. In our present study, the total
image patches used in the segmentation models are half of that used in the
regression models.
In addition to the coverage value determination, segmentation models provide
pixel-wise classification of the image, classifying each pixel in the AFM
images of WSe2 samples as either belonging to the substrate or the crystal.
This has some additional utility in determining not only how much crystal is
present, but its location in the micrograph. The intersect over union (IoU)
metric shows high performance even on the pixel-level classification task,
with 92% (SEG50) and 90% (SEG101) IoU on held-out test images. It is worth
noting that similar performances are observed on both the train and test sets,
indicating low memorization. This level of generalization, despite the held
out test set samples being grown at different conditions and/or obtained at
different imaging conditions, underscores the potential of the models to
produce reliable results in practical applications.
### 3.3 Inference on Full Images
Figure 10: The original (whole) images in the hold-out test set (first rows),
and the pixel-wise classification, as either belonging to crystal or substrate
(second rows) obtained from the SEG50 model. The intersection over union (IoU)
accuracy for each image is given below the classification. Figure 11: Coverage
analysis and segmentation of the original (whole) test images. Results
obtained using the segmentation models, SEG50 and SEG101, and the best
regression models, CNN-I and CNN-M are shown. The S/No. corresponds to the
image # shown in Fig. 10.
The test set discussed in the previous sections is based on patches created
from the full image test set. However, it is important to characterize the
held-out test set in its original full image format, as this is the real
measure of the practical value of our trained models. For this test, we are
using SEG50 and SEG101 and only the best regression models: CNN-I and CNN-M.
While SEG50 gives best performance on the held-out test set among the
segmentation models, SEG101 and SEG18-I give similar results (Table 2).
The full images were padded such that they match the exact multiple of model
training patch size, $224\times 224$ and $512\times 512$, for regression and
segmentation, respectively, or the last row/column is lost. The tiles (with
the same sizes as those used in training the models) are then obtained from
the full images and the coverage and segmentation are predicted using the
trained models. For CNN-I and CNN-M, the predicted coverage for each tiles is
multiplied by the size of the tile in order to obtain the number of pixels
with the value above the threshold for crystal. The pixel values above the
threshold are added for all the tiles from the same full image. The crystal
coverage of a given full image is then obtained by dividing the sum of the
number of pixels above the crystal threshold from all the tiles by the size of
the full image (the total number of pixels in the full image). Meanwhile for
the SEG50 and SEG101, the resulted segmented tiles are concatenated and the
artificial padding added is removed. The coverage label is then obtained based
on the concatenated segmentation mask.
For the 23 held-out test images which were grown with growth parameters and/or
obtained at different imaging conditions than the train and validation sets
(Table 1), the performance of the models are not as good as on the patched
images for any model. The regression models are at least 30% worse while the
segmentation models are at least four times worse – this means that the
regression models outperform the segmentation models in practice despite worse
test performance on image patches (Figures 10 and 11). The results obtained
from SEG50 are mostly consistent with the results on image patches with an
average IoU accuracy of 86% compared to 92%. Except for a few cases such as
the image #6, 15,and 19, less than 10% error are typical for both the coverage
and the IoU.
In contrast, the SEG101 performed quite poorly, despite being a similar
architecture compared to SEG50, which is surprising because both models give
comparable performance on the patched images. The fact that SEG101 gave the
best result on the first 4 images, which are the same size but different from
the rest of the test set, provides the clue as to why the model performs
poorly on most of the images as well as the SEG50’s lower accuracy on the full
images compared to the patches. Creating the tiles for the full image
inference requires processing that could result in the loss of some parts of
the original images. The resizing involved in the patches created for training
the models is also inevitably not exactly the same as that for the tiles. The
sensitivity of the different models to the different image processing and the
image morphological features have therefore resulted in the observed variation
in the model performances. Also worthy of note is the fact that significant
variations in the segmentation model performances have been observed depending
on the encoder and/or decoder architecture.40
Overall, the results on full images show an important distinction between the
training protocol and real-world application of CNNs. Deep CNNs such as SEG101
may not be robust in practical micrograph analysis despite excellent
performance even on held-out test data due to the image augmentation scheme.
Meanwhile, even though the calculation of crystal coverage is natively a
segmentation problem, the regression models perform well on the full images,
suggesting that they may be more robust to changes of scale, dimension, or
other factors compared to the segmentation models.
## 4 Conclusion
In this study, we conduct a comprehensive analysis of crystal coverage (the
proportion of the substrate covered with grown crystal) in WSe2 thin film
atomic force microscopy (AFM) micrographs using regression and segmentation
models. Regression models were trained to predict the monolayer crystal
coverage from image patches. Models were trained from the scratch and using
transfer learning from ResNet pretrained on ImageNet and MicroNet. MicroNet
consists of grayscale micrographs while ImageNet is made up of the macroscale
color images of natural objects. For transfer learning, both feature
extraction and fine-tuning approaches were used.
Our analysis revealed that the CNN models trained from the scratch outperforms
MLP models trained on features extracted from the pretrained models, while
fine-tuning gave the best performance with up to 0.99 $R^{2}$ value on the
held-out test set. Interestingly, while a significantly better performance is
observed from features extraction using MicroNet than that from the ImageNet,
the fine-tuning shows the reverse. This means that the filters pretrained on
the MicroNet extract more useful features from the AFM than that pretrained on
the ImageNet. However, the latter scenario seems to provide more generic image
features in which case fine-tuning on sufficient target data has yielded a
better result.
Beyond the prediction of crystal coverage over entire patches, segmentation
models provide pixel-wise classification of the image, classifying each pixel
in the AFM images of WSe2 samples as either belonging to the substrate or the
crystal. This has some additional utility in determining not only how much
crystal is present, but its location in the micrograph. Based on the patches
of the images, the segmentation models provide higher performance in
determining the crystal coverage than regression models. The intersection over
union (IoU) metric shows high performance even on the pixel-level
classification task, with up to 92% IoU on held-out test images.
The results on full images show an important distinction between the training
protocol and real-world application of the models. Contrary to the results
from image patches, the regression models performed better than the
segmentation models at predicting the monolayer crystal coverage of the full
images of the held-out test set, giving the $R^{2}$ values of 0.98 and 0.90,
respectively, from the best models. The average IoU on the full held-out test
images reduced to 86% from the 92% obtained for the patch images. Our finding
suggests that the regression models may be more robust to changes of scale,
dimension, or other factors compared to the segmentation models. Overall,
these results highlight the efficacy of machine learning for automated, high-
throughput sample characterization, demonstrating its potential for
accelerating the high-throughput development of chalcogenides for
technological applications. At the same time, it provides practical guidelines
for implementing standard computer vision workflows in real-world materials
characterization applications.
## Acknowledgments
This study is based upon research conducted at The Pennsylvania State
University Two-Dimensional Crystal Consortium – Materials Innovation Platform
(2DCC-MIP) which is supported by NSF cooperative agreement DMR-2039351.
## Data Availability
The raw data required to reproduce these findings are available to download
from Ref.35
## References
* Gupta et al. 2015 Gupta, A.; Sakthivel, T.; Seal, S. Recent development in 2D materials beyond graphene. _Progress in Materials Science_ 2015, _73_ , 44–126
* Mas-Balleste et al. 2011 Mas-Balleste, R.; Gomez-Navarro, C.; Gomez-Herrero, J.; Zamora, F. 2D materials: to graphene and beyond. _Nanoscale_ 2011, _3_ , 20–30
* Lv et al. 2015 Lv, R.; Robinson, J. A.; Schaak, R. E.; Sun, D.; Sun, Y.; Mallouk, T. E.; Terrones, M. Transition metal dichalcogenides and beyond: synthesis, properties, and applications of single-and few-layer nanosheets. _Accounts of chemical research_ 2015, _48_ , 56–64
* Novoselov et al. 2004 Novoselov, K. S.; Geim, A. K.; Morozov, S. V.; Jiang, D.-e.; Zhang, Y.; Dubonos, S. V.; Grigorieva, I. V.; Firsov, A. A. Electric field effect in atomically thin carbon films. _science_ 2004, _306_ , 666–669
* Choi et al. 2017 Choi, W.; Choudhary, N.; Han, G. H.; Park, J.; Akinwande, D.; Lee, Y. H. Recent development of two-dimensional transition metal dichalcogenides and their applications. _Materials Today_ 2017, _20_ , 116–130
* Huang et al. 2014 Huang, J.-K.; Pu, J.; Hsu, C.-L.; Chiu, M.-H.; Juang, Z.-Y.; Chang, Y.-H.; Chang, W.-H.; Iwasa, Y.; Takenobu, T.; Li, L.-J. Large-area synthesis of highly crystalline WSe2 monolayers and device applications. _ACS nano_ 2014, _8_ , 923–930
* Lin et al. 2014 Lin, Y.-C.; Lu, N.; Perea-Lopez, N.; Li, J.; Lin, Z.; Peng, X.; Lee, C. H.; Sun, C.; Calderin, L.; Browning, P. N.; others Direct synthesis of van der Waals solids. _Acs Nano_ 2014, _8_ , 3715–3723
* Grigoriev et al. 2012 Grigoriev, S.; Fominski, V. Y.; Gnedovets, A.; Romanov, R. Experimental and numerical study of the chemical composition of WSex thin films obtained by pulsed laser deposition in vacuum and in a buffer gas atmosphere. _Applied Surface Science_ 2012, _258_ , 7000–7007
* Eichfeld et al. 2015 Eichfeld, S. M.; Hossain, L.; Lin, Y.-C.; Piasecki, A. F.; Kupp, B.; Birdwell, A. G.; Burke, R. A.; Lu, N.; Peng, X.; Li, J.; others Highly scalable, atomically thin WSe2 grown via metal–organic chemical vapor deposition. _ACS nano_ 2015, _9_ , 2080–2087
* Zhang et al. 2016 Zhang, X.; Al Balushi, Z. Y.; Zhang, F.; Choudhury, T. H.; Eichfeld, S. M.; Alem, N.; Jackson, T. N.; Robinson, J. A.; Redwing, J. M. Influence of carbon in metalorganic chemical vapor deposition of few-layer WSe 2 thin films. _Journal of Electronic Materials_ 2016, _45_ , 6273–6279
* Kang et al. 2015 Kang, K.; Xie, S.; Huang, L.; Han, Y.; Huang, P. Y.; Mak, K. F.; Kim, C.-J.; Muller, D.; Park, J. High-mobility three-atom-thick semiconducting films with wafer-scale homogeneity. _Nature_ 2015, _520_ , 656–660
* Kim et al. 2017 Kim, H.; Ovchinnikov, D.; Deiana, D.; Unuchek, D.; Kis, A. Suppressing nucleation in metal–organic chemical vapor deposition of MoS2 monolayers by alkali metal halides. _Nano letters_ 2017, _17_ , 5056–5063
* Lin et al. 2018 Lin, Y.-C.; Jariwala, B.; Bersch, B. M.; Xu, K.; Nie, Y.; Wang, B.; Eichfeld, S. M.; Zhang, X.; Choudhury, T. H.; Pan, Y.; others Realizing large-scale, electronic-grade two-dimensional semiconductors. _ACS nano_ 2018, _12_ , 965–975
* Lin et al. 2016 Lin, Z.; McCreary, A.; Briggs, N.; Subramanian, S.; Zhang, K.; Sun, Y.; Li, X.; Borys, N. J.; Yuan, H.; Fullerton-Shirey, S. K.; others 2D materials advances: from large scale synthesis and controlled heterostructures to improved characterization techniques, defects and applications. _2D Materials_ 2016, _3_ , 042001
* Rugar and Hansma 1990 Rugar, D.; Hansma, P. Atomic force microscopy. _Physics today_ 1990, _43_ , 23–30
* Giessibl 2003 Giessibl, F. J. Advances in atomic force microscopy. _Reviews of modern physics_ 2003, _75_ , 949
* Zhang et al. 2018 Zhang, H.; Huang, J.; Wang, Y.; Liu, R.; Huai, X.; Jiang, J.; Anfuso, C. Atomic force microscopy for two-dimensional materials: A tutorial review. _Optics Communications_ 2018, _406_ , 3–17
* Cohen et al. 2020 Cohen, A.; Patsha, A.; Mohapatra, P. K.; Kazes, M.; Ranganathan, K.; Houben, L.; Oron, D.; Ismach, A. Growth-etch metal–organic chemical vapor deposition approach of WS2 atomic layers. _ACS nano_ 2020, _15_ , 526–538
* Cun et al. 2019 Cun, H.; Macha, M.; Kim, H.; Liu, K.; Zhao, Y.; LaGrange, T.; Kis, A.; Radenovic, A. Wafer-scale MOCVD growth of monolayer MoS 2 on sapphire and SiO 2. _Nano Research_ 2019, _12_ , 2646–2652
* Li et al. 2021 Li, T.; Guo, W.; Ma, L.; Li, W.; Yu, Z.; Han, Z.; Gao, S.; Liu, L.; Fan, D.; Wang, Z.; others Epitaxial growth of wafer-scale molybdenum disulfide semiconductor single crystals on sapphire. _Nature Nanotechnology_ 2021, _16_ , 1201–1207
* Xiang et al. 2020 Xiang, Y.; Sun, X.; Valdman, L.; Zhang, F.; Choudhury, T. H.; Chubarov, M.; Robinson, J. A.; Redwing, J. M.; Terrones, M.; Ma, Y.; others Monolayer MoS2 on sapphire: an azimuthal reflection high-energy electron diffraction perspective. _2D Materials_ 2020, _8_ , 025003
* Abràmoff et al. 2004 Abràmoff, M. D.; Magalhães, P. J.; Ram, S. J. Image processing with ImageJ. _Biophotonics international_ 2004, _11_ , 36–42
* Borodinov et al. 2020 Borodinov, N.; Tsai, W.-Y.; Korolkov, V. V.; Balke, N.; Kalinin, S. V.; Ovchinnikova, O. S. Machine learning-based multidomain processing for texture-based image segmentation and analysis. _Applied Physics Letters_ 2020, _116_ , 044103
* Carracedo-Cosme et al. 2021 Carracedo-Cosme, J.; Romero-Muñiz, C.; Pérez, R. A deep learning approach for molecular classification based on AFM images. _Nanomaterials_ 2021, _11_ , 1658
* Gordon et al. 2020 Gordon, O. M.; Hodgkinson, J. E.; Farley, S. M.; Hunsicker, E. L.; Moriarty, P. J. Automated searching and identification of self-organized nanostructures. _Nano Letters_ 2020, _20_ , 7688–7693
* Oinonen et al. 2022 Oinonen, N.; Kurki, L.; Ilin, A.; Foster, A. S. Molecule graph reconstruction from atomic force microscope images with machine learning. _MRS Bulletin_ 2022, _47_ , 895–905
* Alldritt et al. 2020 Alldritt, B.; Hapala, P.; Oinonen, N.; Urtev, F.; Krejci, O.; Federici Canova, F.; Kannala, J.; Schulz, F.; Liljeroth, P.; Foster, A. S. Automated structure discovery in atomic force microscopy. _Science advances_ 2020, _6_ , eaay6913
* Moses and Reinhart 2023 Moses, I. A.; Reinhart, W. F. Quantitative Analysis of MoS $\\_2$ Thin Film Micrographs with Machine Learning. _arXiv preprint arXiv:2310.07816_ 2023,
* Tang et al. 2023 Tang, S.; Grundmann, A.; Fiadziushkin, H.; Wang, Z.; Hoffmann-Eifert, S.; Ghiami, A.; Debald, A.; Heuken, M.; Vescan, A.; Kalisch, H. Migration-Enhanced Metal–Organic Chemical Vapor Deposition of Wafer-Scale Fully Coalesced WS2 and WSe2 Monolayers. _Crystal Growth & Design_ 2023, _23_ , 1547–1558
* Holm et al. 2020 Holm, E. A.; Cohn, R.; Gao, N.; Kitahara, A. R.; Matson, T. P.; Lei, B.; Yarasi, S. R. Overview: Computer vision and machine learning for microstructural characterization and analysis. _Metallurgical and Materials Transactions A_ 2020, _51_ , 5985–5999
* Zhao et al. 2023 Zhao, P.; Wang, Y.; Jiang, B.; Wei, M.; Zhang, H.; Cheng, X. A new method for classifying and segmenting material microstructure based on machine learning. _Materials & Design_ 2023, _227_ , 111775
* Baskaran et al. 2020 Baskaran, A.; Kane, G.; Biggs, K.; Hull, R.; Lewis, D. Adaptive characterization of microstructure dataset using a two stage machine learning approach. _Computational Materials Science_ 2020, _177_ , 109593
* Kim et al. 2020 Kim, H.; Inoue, J.; Kasuya, T. Unsupervised microstructure segmentation by mimicking metallurgists’ approach to pattern recognition. _Scientific Reports_ 2020, _10_ , 17835
* Gupta et al. 2020 Gupta, S.; Banerjee, A.; Sarkar, J.; Kundu, M.; Sinha, S. K.; Bandyopadhyay, N.; Ganguly, S. Modelling the steel microstructure knowledge for in-silico recognition of phases using machine learning. _Materials Chemistry and Physics_ 2020, _252_ , 123286
* 35 Moses, I. A.; Chengyin, W.; Reinhart, W. F. Evaluating Transfer Learning Strategies for WSe2 Thin Film Micrograph Analysis. https://m4-2dcc.vmhost.psu.edu/list/data/RVJkDr8j1RPU
* Paszke et al. 2019 Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; others Pytorch: An imperative style, high-performance deep learning library. _Advances in neural information processing systems_ 2019, _32_
* 37 Bakshy, E.; Balandat, M.; Kashin, K. Open-sourcing Ax and BoTorch: New AI tools for Adaptive Experimentation. _URL https://ai. facebook. com/blog/open-sourcing-ax-and-botorch-new-ai-tools-for-adaptive-experimentation_
* Snoek et al. 2012 Snoek, J.; Larochelle, H.; Adams, R. P. Practical Bayesian Optimization of Machine Learning Algorithms. 2012
* Deng et al. 2009 Deng, J.; Dong, W.; Socher, R.; Li, L.-J.; Li, K.; Fei-Fei, L. ImageNet: A large-scale hierarchical image database. 2009 IEEE Conference on Computer Vision and Pattern Recognition. 2009; pp 248–255
* Stuckner et al. 2022 Stuckner, J.; Harder, B.; Smith, T. M. Microstructure segmentation with deep learning encoders pre-trained on a large microscopy dataset. _npj Computational Materials_ 2022, _8_ , 200
* Xie et al. 2017 Xie, S.; Girshick, R.; Dollár, P.; Tu, Z.; He, K. Aggregated residual transformations for deep neural networks. Proceedings of the IEEE conference on computer vision and pattern recognition. 2017; pp 1492–1500
* Hu et al. 2018 Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. Proceedings of the IEEE conference on computer vision and pattern recognition. 2018; pp 7132–7141
* Szegedy et al. 2017 Szegedy, C.; Ioffe, S.; Vanhoucke, V.; Alemi, A. Inception-v4, inception-resnet and the impact of residual connections on learning. Proceedings of the AAAI conference on artificial intelligence. 2017
* Tan and Le 2019 Tan, M.; Le, Q. Efficientnet: Rethinking model scaling for convolutional neural networks. International conference on machine learning. 2019; pp 6105–6114
* Ronneberger et al. 2015 Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, October 5-9, 2015, Proceedings, Part III 18. 2015; pp 234–241
* Zhou et al. 2019 Zhou, Z.; Siddiquee, M. M. R.; Tajbakhsh, N.; Liang, J. Unet++: Redesigning skip connections to exploit multiscale features in image segmentation. _IEEE transactions on medical imaging_ 2019, _39_ , 1856–1867
* Akers et al. 2021 Akers, S.; Kautz, E.; Trevino-Gavito, A.; Olszta, M.; Matthews, B. E.; Wang, L.; Du, Y.; Spurgeon, S. R. Rapid and flexible segmentation of electron microscopy data using few-shot machine learning. _npj Computational Materials_ 2021, _7_ , 187
* Azimi et al. 2018 Azimi, S. M.; Britz, D.; Engstler, M.; Fritz, M.; Mücklich, F. Advanced steel microstructural classification by deep learning methods. _Scientific reports_ 2018, _8_ , 2128
|
# Data-driven classification of low-power communication signals by an
unauthenticated user using a software-defined radio
Tarun Rao Keshabhoina1 and Marcos M. Vasconcelos2 1T. R. Keshabhoina is with
the Department of Electrical Engineering and the Commonwealth Cyber
Initiative, Virginia Tech, USA. Email<EMAIL_ADDRESS>M. Vasconcelos is
with the Department of Electrical and Computer Engineering, FAMU-FSU College
of Engineering, Florida State University, USA. Email<EMAIL_ADDRESS>
###### Abstract
Many large-scale distributed multi-agent systems exchange information over
low-power communication networks. In particular, agents intermittently
communicate state and control signals in robotic network applications, often
with limited power over an unlicensed spectrum, prone to eavesdropping and
denial-of-service attacks. In this paper, we argue that a widely popular low-
power communication protocol known as LoRa is vulnerable to denial-of-service
attacks by an unauthenticated attacker if it can successfully identify a
target signal’s bandwidth and spreading factor. Leveraging a structural
pattern in the LoRa signal’s instantaneous frequency representation, we relate
the problem of jointly inferring the two unknown parameters to a
classification problem, which can be efficiently implemented using neural
networks.
## I Introduction
Multi-agent robotic systems are used in various modern applications, including
industrial automation, agriculture, and environmental monitoring [1, 2]. In
these systems, autonomous robots work together to accomplish a common goal,
such as monitoring an environment or cooperatively completing a task. In such
systems, coordination, and communication among the robots are critical to
their success. Each robot must be aware of the state and actions of the other
robots in the system to coordinate their actions and achieve their goals. For
example, in an agricultural monitoring system, each robot may be responsible
for monitoring a different field area, and they must coordinate their
movements to ensure that the entire field is covered. Therefore, communication
among the robots must be reliable, even in challenging scenarios such as
remote or outdoor environments, which are subject to disruption by obstacles
or malicious interference. Protecting such networks against denial-of-service
attacks is of paramount importance to prevent service disruption and economic
loss.
Figure 1: Block diagram for the communication scenario herein: two legitimate
agents communicate a signal represented by $X$, an attacker observes a
correlated signal $\tilde{X}$, with the intent to emit a jamming signal $J$.
A Low Power Wide Area Network (LPWAN) protocol LoRaWAN (Long Range Wide Area
Network) offers long-range and low-power communication capabilities well-
suited to multi-agent robotic systems [3]. Additionally, LoRaWAN supports
creating large-scale networks with multiple nodes, making it an ideal solution
for coordinating the activities of large groups of robots communicating
intermittently. While LoRaWAN is one of the most robust and resilient low-
power communication protocols, it is still vulnerable to a class of denial-of-
service attacks known as jamming.
A jamming attack follows the diagram in Fig. 1: a transmitting agent, Tx,
sends a signal to a receiving agent, Rx; the transmitted signal is intercepted
by an attacker using a software-defined radio unit; The attacker then infers
two private parameters used for communication between Tx and Rx, and
subsequently sends a jamming signal to interfere with the transmitted signal
at the receiver.
### I-A Related Work
Wireless communication protocols transmit over the air, which makes them
vulnerable to interference from any radio transmitter within their vicinity.
This fundamental aspect of shared media in wireless networks has made way for
extensive research in the wireless jamming domain [4, 5, 6]. Energy-
constrained jamming methodologies attempt to block the channel in reaction to
transmission activity to save power. Herein, we discuss such a reactive
jamming strategy for LoRa PHY. Securing communication systems and improving
performance in the presence of intelligent jammers [7, 8] is the motivation to
this work.
Numerous studies have examined the throughput and performance of ultra-narrow
band (UNB) and spread spectrum-based technologies in the unlicensed
Industrial, Scientific, and Medical (ISM) band [9, 10]. Amongst these, a
comprehensive study of PHY layer vulnerabilities, countermeasures and security
features of LoRaWAN are presented in [11], and its authors also provide a
brief overview of jamming methodologies for LoRa. Long-range transmissions on
LoRa are susceptible to several attack strategies such as replay attacks,
wormhole attacks, and compromising network key information, in addition to
jamming [11].
LoRa’s medium access control (MAC) layer design introduces many configurable
parameters that affect its service reliability. An in-depth explanation of
such parameters, and their resulting performance tradeoffs are presented in
[12]. Choices of these parameters, driven by service requirements, also play a
role in the PHY layer encoding of signals, having implications on the
approaches adopted by intelligent jammers.
When signals from one packet are $6\mathrm{dB}$ stronger than another, it goes
on to be demodulated, leaving the weaker packet to be discarded (this is the
so-called channel capture effect) [13]. Building on the concept, the authors
of [14] have shown that LoRa can be jammed using commercially available
hardware. Herein, they induce collisions on the channel, by flooding it with
numerous packets of identical parameter choices. A more advanced technique,
targetting the symbol demodulation process in LoRa was explored in [15],
introducing the idea of jamming chirps. They revealed that LoRa receivers
cannot distinguish between a well-synchronized jamming chirp and a legitmate
chirp.
LoRa was found vulnerable to interference when two packets employ the same
configuration of two parameters known as the Bandwidth ($BW$) and Spreading
Factor ($SF$). The symbol demodulation process in LoRa involves two steps:
first, dechirping, and then, FFT (Fast Fourier Transform). Symbols are
determined by identifying peaks within the FFT. When interfering packets
utilize the same $BW$ and $SF$, this can cause multiple indiscernible peaks in
the FFT, leading to symbol errors [16].
Contemporary work in LoRa jamming exploit this property, and an empirical
analysis of the approach is discussed in [17]. While they prove the
effectiveness of this strategy, they also make a hard assumption.
Particularly, that the jammer has apriori knowledge of the target signal’s
$BW$ and $SF$ choices, neccessary for generating the jamming chirps. However,
these parameters are generally not available to adversarial agents, which are
unathenticated users of the network.
In this paper, we take a step further, exploring how an adversary may employ a
simple neural network implementation to estimate this information and jam LoRa
signals reactively, without such assumed knowledge. We provide numerical
results on the detection and identification of the $BW$ and $SF$ parameters
from observed signals. Then, we quantify the robustness of our model by
evaluation on a wide range of signal-to-noise ratio (SNR) levels of signals.
The rest of this paper is organized as follows. Section II introduces LoRa PHY
and the chirp spread spectrum. Section III describes system architecture.
Section IV describes our proposed feature extraction technique. Section V
describes the architecture of the neural network classifier. Section VI
presents our simulation results and discusses our system’s performance.
Finally, Section VII concludes the paper and outlines future research
directions.
## II Signal description
Figure 2: A chirp signal with BW = 125 KHz, and SF = 7 in continuous time
(left), discrete time (middle), and its spectrogram (right).
LoRa PHY is a pass band modulation technique that uses chirp spread spectrum
(CSS) to modulate digital information onto a carrier wave. In CSS, a chirp is
a signal whose instantaneous frequency increases or decreases linearly as a
function of time.
In LoRa, each transmitted symbol is mapped into a chirp. The bandwidth ($BW$)
and spreading factor ($SF$) are the most critical parameters defining a LoRa
chirp. The $BW$ corresponds to the range of frequencies of the channel
occupied by the chirp, and the $SF$ determines the number of bits transmitted
in a symbol. Each symbol carries $SF$ bits (i.e., values ranging from $0$ to
$2^{SF}-1$). The joint choice of $SF$ and $BW$ determines the data rate of the
communication link. Following [18], in this section, we describe the CSS
modulation.
A fundamental characteristic of the LoRa chirp is its cyclically shifted
frequency. Wherein, the frequency incrementally rises from the initial
frequency in discrete steps. Upon reaching the highest frequency, it wraps
around to the lowest frequency and continues its ascent until it cycles back
to the initial frequency. The chirp encodes information by adjusting its
starting frequency according to its symbol value, $s_{n}$.
Consider the transmission of a sequence of symbols $\mathbf{s}:=\\{s_{n}\\}$.
Each symbol carries $SF$ bits, denoted by a vector
$\mathbf{w}_{n}=(w_{n,0},\ldots,w_{n,SF-1})$, where $w_{n,b}\in\\{0,1\\}$,
$b\in\\{0,\ldots,SF-1\\}$. A new symbol is transmitted every $T_{s}$ seconds,
corresponding to a chirp signal’s duration in time. The value of the symbol
$s_{n}$ is given by
$s_{n}=\sum_{b=0}^{SF-1}w_{n,b}\times 2^{b}.$ (1)
Since $s_{n}$ can take on $2^{SF}$ distinct values, the channel bandwidth is
divided into $2^{SF}$ discrete levels. Each of these levels signifies the
starting frequency for a specific symbol value.
Therefore, the chirp completes $2^{SF}$ discrete steps throughout its
duration, in cycling back to its initial frequency. For a chosen bandwidth,
$BW$, each step lasts for a duration of $T=1/BW$ seconds, adding up to the
entire symbol duration $T_{s}$. Thus, $SF$ determines the number of steps, and
$BW$ determines the time period of each step, collectively defining the symbol
duration, $T_{s}=2^{SF}/BW$.
Let $f_{c}$ denote the channel’s center frequency. The $n$-th transmitted
symbol, $s_{n}$, is mapped into a chirp signal $c_{n}(t)\in\mathbb{C}$ given
by
$c_{n}(t)=\frac{1}{\sqrt{2^{SF}}}\exp\big{\\{}j\big{(}2\pi
f_{n}(t)\big{)}t\big{\\}},\ \ t\in[0,T_{s}]$ (2)
where,
$f_{n}(t)=f_{c}+\mathrm{mod}\big{(}s_{n}+t\times
BW,2^{SF}\big{)}\times\frac{BW}{2^{SF}}-\frac{BW}{2},$ (3)
and $\mathrm{mod}(\xi,2^{SF})$ is the remainder of the division of $\xi$ by
$2^{SF}$.
In LoRa $SF\in\\{7,8,9,10,11,12\\}$. It is customary to represent a chirp in
discrete-time using $2^{SF}\times f_{s}/BW$ samples indexed by $k$, where
$f_{s}$ is the sampling frequency and $f_{s}/BW$ is the oversampling factor.
Letting $t=k/f_{s}$, we obtain:
$c_{n}(k)=\frac{1}{\sqrt{2^{SF}}}\exp\bigg{\\{}j2\pi\times\Big{(}f_{c}+\\\
\mod\big{(}s_{n}+k\times\frac{BW}{f_{s}},2^{SF}\big{)}\times\frac{BW}{2^{SF}}-\frac{BW}{2}\Big{)}\bigg{\\}}\times
k,$ (4)
where $k=\\{0,1,2,\ldots,(2^{SF}\times f_{s}/BW)-1\\}.$ Figure 2 shows a chirp
in continuous time, in discrete time and in its time-frequency representation.
## III System description
Traditionally, jamming in the physical layer corresponds to adding white
Gaussian noise (AWGN) to the transmitted signal. Such naïve strategies are
ineffective in LoRa communications. Due to that resiliency to AWGN, LoRa has
also been referred to as a secure communication protocol. However, it has been
shown by [17] that LoRa is vulnerable to jamming using a chirp-type waveform.
Generating the chirp-type waveform to cause destructive interference requires
the knowledge of $BW$ and $SF$.
The LoRaWAN specification fixes the choice of these parameters to a finite set
of $18$ combinations ($BW\in\\{125\mathrm{kHz},\ 250\mathrm{kHz},\
500\mathrm{kHz}\\}$ and $SF\in\\{7,8,9,10,11,12\\}$). These parameters are
agreed by the legitimate communicating parties, but are not readily available
to a jamming adversary. Hence, the jammer needs to estimate this information
from an observed signal.
Figure 3: Block diagram for a reactive jammer in a communication system that
uses CSS modulation.
Figure 3 shows the block diagram of the data pipeline used by a reactive LoRa
jammer. Each component of this system is described in the following
subsections.
### III-A Data batch preprocessing block
The SDR captures signals in real time and outputs a stream of In-phase and
Quadrature (IQ) samples of indefinite length. On the other hand, our neural
network classifier operates on data batches of finite size. The preprocessor
block collects data flowing in from the SDR into a matrix of appropriate size
for processing in the subsequent blocks.
The SDR is tuned to the channel of interest and configured to a sampling rate
of $1\mathrm{MHz}$. Due to the Shannon-Nyquist Theorem, a minimum sampling
rate of $1\mathrm{MHz}$ is required since the maximum $BW$ in LoRa is
$500\mathrm{KHz}$. A lower sampling rate might result in distortion from
aliasing, and higher rates imply higher demand for computational resources.
Therefore, the SDR generates a noisy IQ stream $\tilde{X}$ of discrete-time
samples to the host PC. The preprocessor block parses this stream of complex
values into smaller signal blocks and reshapes them into a matrix of
dimensions $B\times M$. Where $B$ represents batch size and $M$ represents
length of the signal segment.
Determining the proper block length $M$ is crucial, as it must contain enough
samples to distinguish the LoRa configurations reliably. If the block length
is too small, the signal is truncated and information is lost. If the block
length is too large, the the neural network processing introduces latency.
Hence it must be as small as possible yet carry enough signal information.
We have empirically determined that the ideal block length must span two LoRa
symbols for the longest configuration. The longest configuration in LoRa is
$BW=125\mathrm{KHz}$, and $SF=12$, resulting in a symbol duration of
$T_{s}=2\times 2^{12}/125000$ seconds. For a sampling frequency of
$1\mathrm{MHz}$, we obtain an over-sampling factor of $8$, resulting in
$2\times 2^{12}\times 8=65,536$ of samples. Therefore, we fix the block length
to $M=65,600$.
### III-B Feature Extraction
The feature extraction block employs an algorithm based on the instantaneous
frequency (IF), which leads to a compact representation of LoRa signal
sequences. Such representation accentuates features related to the
identification of $BW$ and $SF$. The algorithm first transforms the signal
vectors from the time domain to the frequency domain and tracks the
instantaneous frequency of the signal over time. In the frequency domain, any
pair of LoRa signals corresponding to different configurations appear
distinctly different. The algorithm takes in a batch of signal blocks from the
preprocessor block, $V$, and applies the algorithm described in Section IV to
produce a matrix $F$ of IF vectors.
Our goal is to infer the parameters $SF$ and $BW$. One influences the duration
of the chirp, and the other affects both the duration and frequency sweep
range in the chirp. Our approach to feature extraction here is to characterize
the instantaneous frequency of the signal, describing the evolution of the
frequency in the signal with time. Through this representation, we can observe
both the range of the frequencies swept and the time elapsed for each sweep,
enabling simultaneous estimation of $SF$ and $BW$.
### III-C Chirp classifier
The chirp classifier block uses a neural network (NN) to identify the
transmitted chirp signal. Our model is trained using a dataset of IF vectors
labeled with their corresponding $BW$ and $SF$ configurations. Once trained,
this block receives an IF vector and performs a soft-decision classification
of $BW$ and $SF$ in a vector $C$ of probabilities for each of the $18$
possible signal configurations. This information passed to the chirp-generator
block.
In the context of classifying LoRa signals based on their features, it is
important to note that the relationship between these features and their
respective classifications is non-linear. NNs can learn complex relationships
and patterns in data, making them suitable for tasks like classifying signals
with intricate or non-linear relationships between their features and
categories. With proper training and a sufficiently rich architecture, NNs can
provide accurate signal classification even at extremely low levels of SNR. We
will discuss the NN architecture in more detail on Section V.
### III-D Chirp generator
The chirp generator block is responsible for utilizing the inferred $BW$ and
$SF$ to generate a stream of discrete-time IQ values for the jamming chirps,
denoted by $J$. The IQ stream should be sent to the SDR, which uses a Digital
to Analog Converter (DAC) that converts them from discrete-time to a
corresponding continuous-time signal. Once converted to analog, the SDR can
adjust the signal to the channel’s center frequency for transmission. The
resulting signal would represent a chirp with the same $BW$ and $SF$ as the
target signal, leading to interference at the receiver.
LoRa uses a two-step demodulation procedure: the first is known as dechirping,
followed by an FFT. The dechirping operation multiplies the sampled signal
with a base down chirp of the same $BW$ and $SF$. The resulting signal has a
constant frequency, which matches the chirp’s initial frequency. Then, from
its FFT, we identify the bin index of this frequency, determining the encoded
symbol’s value. Under this demodulation scheme, when two signals of the same
$BW$ and $SF$ configuration interfere at the receiver, they result in multiple
indiscernible peaks in the FFT step. Such interference deceives the receiver
into misidentifying the original symbol. This misidentification leads to
symbol demodulation errors, resulting in packet drops, effectively jamming the
signal.
With the knowledge of $BW$ and $SF$ we can generate chirp signals using Eq. 4.
However, the chirp’s polarity (upchirp or downchirp), the symbol value, and
the arrival time influence the effectiveness of interference with the target
signal. Considering these factors, the authors of [17] introduced three
effective methods to jam LoRa signals when $BW$ and $SF$ are known, which can
be implemented in the chirp generator block, summarized as follows:
* •
Identical chirps: A simple approach is to continuously repeat the same symbol
in sequence. By transmitting continuously, we avoid sudden shifts across
demodulation windows. Any delays and time offsets only affect the initial
frequency of the chirp and still result in demodulation errors. This method is
lightweight because it does not require strict time synchronization.
* •
Consecutive downchirps: This method targets the Start Frame Delimiter (SFD)
symbol of LoRa packets, which is a base downchirp that marks the beginning of
the packet header. From transmitting base downchirps consecutively, the
receiver is tricked into making errors in identifying the legitimate SFD,
resulting in incorrect packet parsing and leading to packet drops.
* •
Synchronized chirps: This method is considered to be the most effective
jamming strategy in LoRa [15, 17]. It involves transmitting random symbols
that perfectly align with the demodulation window at a receiver. This is made
possible by estimating and compensating the Carrier Frequency Offset (CFO) and
the Sampling Time Offset (STO), as in a legitmate LoRa demodulator. The
synchronized chirps method requires strict synchronization and additional
computing, however, it is the most effective and difficult to detect method
known to date.
In conjunction with the inferred parameters, the chosen method defines the
sequence of jamming chirps to be transmitted. The IQ values corresponding to
this sequence is streamed from the chirp generator block to the SDR at a fixed
rate. Consequently, the SDR transmits this waveform over the air to jam the
target signal at the receiver. This strategy shows that it is possible to jam
LoRa signals of unknown $BW$ and $SF$ configurations by an unauthenticated
agent.
Figure 4: Feature representations for various LoRa Configurations
## IV Feature extraction
In this section, we identify a pattern in the data, also known as feature,
that aids in distinguishing one category from another. To that end, we compute
the instantaneous frequency of the signal. Considering this feature, we can
retain information about the range of frequencies swept, and their sweep rate
simultaneously, which are directly related to our two parameters of interest,
$BW$ and $SF$.
Here, we follow a two-step procedure to computing the instantaneous frequency:
a Short Term Fourier Transform (STFT) followed by Instantaneous Frequency (IF)
estimation.
### IV-A Short Term Fourier Transform (STFT)
Given the inherent time-varying nature of frequency in a chirp signal, we
employ the STFT on each input signal segment [19]. A given signal segment is
further subdivided into overlapping windows, each consisting of $W=128$
samples, with an overlap of $L=64$ samples. Subsequently, an FFT is executed
on these windows. This operation obtains the power distribution across all the
frequencies in the channel bandwidth, $BW$ as the signal evolves in time, as
follows:
$Q[k,m]=\sum_{n=0}^{W-1}x[n+mL]w[n]e^{-j2\pi nk/W},$ (5)
where $Q[k,m]$ is the STFT coefficient at frequency bin $k$ and time index
$m$, $x$ is the input signal segment, and $w[n]$ is the Hann window function
[20].
### IV-B Instantaneous Frequency Estimation
Unlike stationary signals where the spectral properties are constant, the
frequency of a chirp signal varies linearly with time [21]. For such signals,
we must compute the instantaneous frequency instead of frequency. The
instantaneous frequency is a time-varying parameter related to the average of
the frequencies present in the signal as it evolves in time [22].
From the STFT operation in Eq. 5, we obtain the energy distribution over all
frequency bins for every time-step. We use this energy distribution to compute
a weighted average of the frequencies at each time-step, obtaining the
instantaneous frequency of the signal, as follows:
$f_{inst}(m)=\frac{\sum_{k=1}^{K}P(k,m)f(k,m)}{\sum_{k=1}^{K}P(k,m)},$ (6)
where $f_{inst}(m)$ is the instantaneous frequency at the time index $m$,
$f(k,m)$ is the peak frequency at frequency index $k$ and time index $m$, and
$P(k,m)$ is the power spectral density, computed as $P(k,m)=|Q[k,m]|^{2}$.
## V Neural network architecture
LoRa nodes operate under power constraints (typically from $10\mathrm{dBm}$ to
$20\mathrm{dBm}$) and often transmit over long communication distances
(typically from $10^{3}$m to $10^{4}$m). As a result, LoRa signals are often
received at low SNR, sometimes even below the noise floor. Identifying and
distinguishing such signals reliably demand a classifier model with high noise
tolerance and discriminative power.
Neural networks have been extensively used for signal classification in
wireless communications, spanning applications such as channel sensing,
interference detection and spectrum management [23, 24, 25]. Central to their
efficacy in these applications is their inherent ability to model non-linear
relationships between parameters and noisy data [26].
Figure 4 illustrates the feature representations corresponding to different
$BW$ and $SF$ configurations. The first three sub-figures show the case of
fixed $BW$, and the last three figures illustrate the case of fixed $SF$.
These graphs indicate that changes in $BW$ and $SF$ result in clearly distinct
waveforms. Additionally, the characterization based on the IF of these
waveforms makes the task of distinguishing signals of different configurations
much simpler by converting the the problem of estimating $BW$ and $SF$ into a
signal classification problem.
For this classification task, we use a feed-forward neural network as
illustrated in Figure 5. The model features two hidden layers with 16 neurons
each, and an output layer of 18 neurons, as specified in Table I. The input is
the IF vector, where $t$ is the time index. To classify an IF vector into one
of the $18$ categories, we use a softmax function in the output layer to
obtain a probability distribution on the likelihood of each class given the
observed data. Consider an output of the final layer,
$Z=[z_{1},z_{2},\dots,z_{18}]$ of $18$ real numbers, the softmax function,
$S(\cdot)$, is defined as:
$S(z_{i})=\frac{e^{z_{i}}}{\sum_{j=1}^{18}e^{z_{j}}},\ \ i=1,\dots,18.$ (7)
Figure 5: Neural network architecture used in our system. TABLE I: Key Attributes of the Neural Network Architecture Attribute | Description
---|---
Input | Flatten Layer
Hidden Layer 1 | Dense (16 units, tanh activation)
Hidden Layer 2 | Dense (16 units, tanh activation)
Regularization | Dropout (0.5 rate)
Output Layer | Dense (18 units, softmax activation)
Loss Function | Categorical Cross-Entropy
Optimizer | Adam
Evaluation Metric | Classification Accuracy
## VI Simulation Results
We use synthetic datasets of LoRa signals, creating separate datasets for
training and validation.111The data and code for all the simulations and
numerical experiments in this paper are available at https://github.com/MINDS-
code/jammingSDR.git. Here, the noisy signals are generated according to an
Additive White Gaussian Noise (AWGN) model producing signal data at diverse
SNR levels. Our training dataset has $10$ SNR levels, ranging from $0$ to
$20\mathrm{dB}$. For each of the $18$ configurations, we have generated $50$
signal files. Thus, the training dataset contains a total of $9000$ entries.
Our validation dataset has a broader SNR range, from $-15$ to $20\mathrm{dB}$,
leading to $18$ SNR levels in total. Here, we have generated $20$ signal files
for each case, leading to a total of $6480$ entries. We found out that if the
training dataset included signals with SNR below zero, the classification
performance of the of the NN is severely degraded.
Figure 6: Classification accuracies against SNR with a 95% confidence
interval: (1) overall, (2) by spreading factor, and (3) by bandwidth.
Consider a clean signal, denoted by $X$, subjected to AWGN denoted by $Z$ as
follows:
$\tilde{X}=X+Z,$ (8)
where $\tilde{X}$ is the resulting noisy signal. The power level of $Z$ is
determined by the desired SNR level. We obtain confidence intervals on results
by repeating the experiment $30$ times. Addionally, we experiment with fixed
$BW$ and $SF$ choices, observing their influence on classification
performance. Figure 6 (left) illustrates the classifier’s overall accuracy in
relation with SNR. Classification accuracy starts at around 12% for
$-15\mathrm{dB}$ SNR. The accuracy increases sharply and saturates at
$-5\mathrm{dB}$ SNR. Figure 6 (middle) illustrates the classifier’s accuracy
as a function of SNR for the different possible $SF$ configurations. Each
curve differs from the others by their saturation points and the accuracy
levels they can reach. The curve for $SF$ 12 reaches saturation the earliest
and at the highest accuracy level, succeeded by $SF$ 11, with subsequent
configurations following in descending order. We observe that for a fixed SNR
level, higher SF choices yield consistently higher accuracy scores. The mean
classification accuracy improved with an increase in $SF$ from 7 to 12. This
trend results from LoRa’s spreading waveform, where $SF$ determines the sweep
rate of the chirp. A higher $SF$ leads to a longer chirp duration resulting in
a more elongated and discernible frequency trajectory over time. With more
samples constituting the waveform, identification becomes more precise,
improving classification accuracy.
Figure 6 (right) illustrates the classifier’s accuracy as a function of SNR
for three $BW$ configurations: $125\mathrm{KHz}$, $250\mathrm{KHz}$, and
$500\mathrm{KHz}$. Before reaching saturation, the $125\mathrm{KHz}$ curve
shows a higher accuracy compared to the other two. Meanwhile, the classifier’s
accuracy for the $250\mathrm{KHz}$ curve is consistenlty higher than for
$500\mathrm{KHz}$.
After reaching saturation, the $500\mathrm{KHz}$ curve exhibits higher
accuracy over the other two. However, beyond this point, all three
configurations deliver high classification accuracy. Thus, despite the high
accuracy of the $500\mathrm{KHz}$ curve post saturation, the real
differentiator lies in their points of saturation. The earlier the saturation,
the lower the minimum SNR needed to classify the signal reliably. Thus, the
order in which the curves saturate imply that lower $BW$ configurations yield
better detection.
The mean classification accuracy saturates later for higher $BW$ choices from
$125\mathrm{KHz}$ to $500\mathrm{KHz}$. With a wider $BW$, the signal’s
frequency changes on a broader range in a reduced period. This rapid shifting
causes the instantaneous frequency vectors to become too closely spaced,
making it more challenging for the classifier to distinguish them.
The choice of $BW$ and $SF$ in LoRa is motivated by the application’s quality
of service requirements. However, in practice, nodes switch between several
parameter choices to save power and optimize throughput. Therefore, when
jamming or extensive interference is a concern, legitimate nodes must consider
switching to faster $BW$ and $SF$ choices. Our results conclude that, to avoid
detection by unauthorized agents, legitmate LoRa nodes must opt for lower $SF$
choices and higher $BW$ choices whenever possible.
## VII Conclusions and future work
Many large-scale multi-agent systems rely on LPWAN protocols. Amongst these,
LoRaWAN has found widespread adoption, due to its energy efficiency, long
range, and use of unlicensed spectrum. However, it is succeptible to cyber-
attacks, including eavesdropping and jamming. In this paper, we explored the
vulnerability of LoRa to signal jamming.
A survey of related literature revealed that LoRa is vulnerable to jamming
with a particular chirp type signal. However, generating such signals require
the knowledge of the bandwidth, and spreading factor of the target LoRa
signal. We argue that this information is shared amongst legitimate parties
but unavailable to an unautheticated adversarial agent. In this work, we
presented the high-level design of a practical jammer, that makes use of a
neural network classifier for estimating these parameters by eavesdropping and
reactively emits jamming chirps.
Leveraging a structural pattern in LoRa’s signal waveform, we relate the
problem of estimating these parameters to a signal classification task. To
that end, we proposed a feature extraction method that computes the
instantaneous frequency of signals, enhancing features pertinent to
identifying $BW$ and $SF$ configurations. Then we trained a feedforward neural
network classifier on a dataset LoRa signals to learn these characteristics
for predictive analysis. Our results indicate that the classifier begins to
reliably estimate these parameters for signals stronger than $-5\mathrm{dB}$
SNR. Additionally, we analyzed detection performance at various configurations
of $BW$ and $SF$. Ultimately revealing that, to hinder such detection,
legitimate users of LoRa must use lower $SF$ and higher $BW$.
Directions for future work include experimenting this classifier on a dataset
of real signals captured using a software radio to provide a real-world
validation of this analysis, and an end-to-end implementation of the proposed
jammer to explore real-time performance of the design.
## References
* [1] Z. H. Ismail, N. Sariff, and E. G. Hurtado, “A survey and analysis of cooperative multi-agent robot systems: challenges and directions,” _Applications of Mobile Robots_ , pp. 8–14, 2018.
* [2] A. A. A. Rasheed, M. N. Abdullah, and A. S. Al-Araji, “A review of multi-agent mobile robot systems applications.” _International Journal of Electrical & Computer Engineering (2088-8708)_, vol. 12, no. 4, 2022.
* [3] F. Adelantado, X. Vilajosana, P. Tuset-Peiro, B. Martinez, J. Melia-Segui, and T. Watteyne, “Understanding the limits of LoRaWAN,” _IEEE Communications magazine_ , vol. 55, no. 9, pp. 34–40, 2017.
* [4] Y. Zou, J. Zhu, X. Wang, and L. Hanzo, “A survey on wireless security: Technical challenges, recent advances, and future trends,” _Proceedings of the IEEE_ , vol. 104, no. 9, pp. 1727–1765, 2016.
* [5] W. Xu, W. Trappe, Y. Zhang, and T. Wood, “The feasibility of launching and detecting jamming attacks in wireless networks,” in _Proceedings of the 6th ACM international symposium on Mobile ad hoc networking and computing_ , 2005, pp. 46–57.
* [6] Y. Liu, H.-H. Chen, and L. Wang, “Physical layer security for next generation wireless networks: Theories, technologies, and challenges,” _IEEE Communications Surveys & Tutorials_, vol. 19, no. 1, pp. 347–376, 2016.
* [7] X. Zhang and M. M. Vasconcelos, “Robust remote estimation over the collision channel in the presence of an intelligent jammer,” in _61st Conference on Decision and Control (CDC)_. IEEE, 2022, pp. 5472–5479.
* [8] ——, “Robust one-shot estimation over shared networks in the presence of denial-of-service attacks,” _arXiv:2302.14689_ , 2023.
* [9] B. Reynders, W. Meert, and S. Pollin, “Range and coexistence analysis of long range unlicensed communication,” in _23rd International Conference on Telecommunications (ICT)_. IEEE, 2016, pp. 1–6.
* [10] N. Naik, “LPWAN technologies for IoT systems: choice between ultra narrow band and spread spectrum,” in _International systems engineering symposium (ISSE)_. IEEE, 2018, pp. 1–8.
* [11] H. Ruotsalainen, G. Shen, J. Zhang, and R. Fujdiak, “LoRaWAN physical layer-based attacks and countermeasures, a review,” _Sensors_ , vol. 22, no. 9, p. 3127, 2022.
* [12] D. Magrin, M. Capuzzo, A. Zanella, and M. Zorzi, “A configurable mathematical model for single-gateway LoRaWAN performance analysis,” _IEEE Transactions on Wireless Communications_ , vol. 21, no. 7, pp. 5049–5063, 2021\.
* [13] A. Rahmadhani and F. Kuipers, “When LoRaWAN frames collide,” in _Proceedings of the 12th International Workshop on Wireless Network Testbeds, Experimental Evaluation & Characterization_, 2018, pp. 89–97.
* [14] E. Aras, N. Small, G. S. Ramachandran, S. Delbruel, W. Joosen, and D. Hughes, “Selective jamming of LoRaWAN using commodity hardware,” in _Proceedings of the 14th EAI International Conference on Mobile and Ubiquitous Systems: Computing, Networking and Services_ , 2017, pp. 363–372.
* [15] N. Hou, X. Xia, and Y. Zheng, “Jamming of LoRa PHY and Countermeasure,” in _IEEE Conference on Computer Communications (INFOCOM)_ , 2021, pp. 1–10.
* [16] C. Goursaud and J.-M. Gorce, “Dedicated networks for IoT: PHY/MAC state of the art and challenges,” _EAI endorsed transactions on Internet of Things_ , 2015.
* [17] N. Hou, X. Xia, and Y. Zheng, “Jamming of LoRa PHY and countermeasure,” _ACM Transactions on Sensor Networks_ , vol. 19, no. 4, pp. 1–27, 2023.
* [18] L. Vangelista, “Frequency shift chirp modulation: The LoRa modulation,” _IEEE signal processing letters_ , vol. 24, no. 12, pp. 1818–1821, 2017.
* [19] J. Allen, “Short term spectral analysis, synthesis, and modification by discrete fourier transform,” _IEEE Transactions on Acoustics, Speech, and Signal Processing_ , vol. 25, no. 3, pp. 235–238, 1977.
* [20] A. Testa, D. Gallo, and R. Langella, “On the processing of harmonics and interharmonics: Using Hanning window in standard framework,” _IEEE Transactions on Power Delivery_ , vol. 19, no. 1, pp. 28–34, 2004.
* [21] B. Boashash, “Estimating and interpreting the instantaneous frequency of a signal. I. Fundamentals,” _Proceedings of the IEEE_ , vol. 80, no. 4, pp. 520–538, 1992.
* [22] ——, “Estimating and interpreting the instantaneous frequency of a signal. II. Algorithms and applications,” _Proceedings of the IEEE_ , vol. 80, no. 4, pp. 540–568, 1992.
* [23] A. Fehske, J. Gaeddert, and J. H. Reed, “A new approach to signal classification using spectral correlation and neural networks,” in _International Symposium on New Frontiers in Dynamic Spectrum Access Networks (DySPAN)_. IEEE, 2005, pp. 144–150.
* [24] A. H. Oveis, A. Capria, A. L. Saverino, and M. Martorella, “Convolutional neural network for joint communication and radar signals classification,” in _24th International Radar Symposium (IRS)_. IEEE, 2023, pp. 1–10.
* [25] W. Si, C. Wan, and Z. Deng, “An efficient deep convolutional neural network with features fusion for radar signal recognition,” _Multimedia Tools and Applications_ , vol. 82, no. 2, pp. 2871–2885, 2023.
* [26] S. Sharma, S. Sharma, and A. Athaiya, “Activation functions in neural networks,” _Towards Data Science_ , vol. 6, no. 12, pp. 310–316, 2017.
|
# Exceptional points in dielectric spheroid
Evgeny Bulgakov Kirensky Institute of Physics Federal Research Center KSC SB
RAS 660036 Krasnoyarsk Russia Reshetnev Siberian State University of Science
and Technology, 660037, Krasnoyarsk, Russia Konstantin Pichugin Kirensky
Institute of Physics Federal Research Center KSC SB RAS 660036 Krasnoyarsk
Russia Almas Sadreev Kirensky Institute of Physics Federal Research Center
KSC SB RAS 660036 Krasnoyarsk Russia
(August 27, 2024)
###### Abstract
Evolution of resonant frequencies and resonant modes as dependent on the
aspect ratio is considered in a dielectric high index spheroid. Because of
rotational symmetry of the spheroid the solutions are separated by the
azimuthal index $m$. By the two-fold variation of a refractive index and the
aspect ratio we achieve exceptional points (EPs) at which the resonant
frequencies and resonant modes are coalesced in the sectors $m=0$ for both TE
and TM polarizations and $m=1$.
## I Introduction
Optical properties of dielectric particle is described by resonant frequencies
and corresponding resonant modes. The most famous case is a dielectric sphere
whose resonant modes and frequencies were first considered by Stratton [1].
The solutions in the form quasi-normal modes (QNMs) leaking from the sphere
were considered in Refs. [2, 3]. The frequencies of these solutions resonances
are complex because of coupling of the dielectric particle with the radiation
continuum and can be considered as the eigenvalues of the non-Hermitian
Hamiltonian [4, 5, 6, 7]. Material losses as well as thermal fluctuations [8]
of dielectric particle can considerably contribute into the imaginary part of
complex resonant frequencies through complex refractive index. Non-Hermitian
phenomena drastically alters the behavior of a system compared to its
Hermitian counterpart describing the closed system. The best example of such a
difference is the avoided resonant crossing (ARC) because of coupling of a
particle with the radiation continuum [9, 10, 11, 12, 13]. In turn the ARC can
emerge to singularities, bound states in the continuum at which the imaginary
part of resonance turns to zero [14, 15] that gives rise to collapse of Fano
resonance and exceptional points (EPs). The last is remarkable by that complex
frequencies become degenerate and the eigenmodes coalesce [16, 17, 18]. Early
experiments on microwave coupled resonators revealed the peculiar topology of
eigenvalue surfaces near exceptional points for encircling of EP [19].
Many works on EPs and their applications are associated with parity-time (PT)
symmetric optical systems with a balanced gain and loss. In that case, EPs can
be easily found by tuning a single parameter, namely, the amplitude of the
balanced gain and loss [20, 21, 22, 23, 24]. Since it is not always easy or
desirable to keep a balanced gain and loss in an optical system there is of
significant interest to explore EPs and their applications in non-PT-symmetric
optical systems. Currently, there exist studies concerning EPs for resonant
states in extended periodic dielectric structures sandwiched between two
homogeneous half-spaces [25, 26, 27, 28], dual-mode planar optical waveguides
[29] and plasmonic waveguide [30], layered structures [31, 32, 33], two
infinitely long dielectric cylinders [34, 35, 36, 37, 38] and even single rod
with deformed cross-section [39, 40, 41, 36, 42]. As for compact dielectric
resonators we distinguish the only study of EPs in compact coated dielectric
sphere [43].
In the present paper we consider similar compact elementary dielectric
resonator such as a spheroid in which EPs can be achieved by two-fold
variation of aspect ratio and refractive index. Although the spheroid allows
the solution due to separation of variables in spheroidal coordinate system
[44, 45], analytical expressions for solutions are too cumbersome. We use
software package COMSOL Multiphysics which allows to obtain numerically the
complex resonant frequencies and corresponding resonant modes of particle of
arbitrary shape embedded into the radiation continuum by use of perfectly
absorbing boundary conditions.
## II Evolution of resonant frequencies in spheroid
A rotational symmetry of spheroid implies that the azimuthal index $m$ is
preserved. That allows to calculate the resonant frequencies and resonant
eigenmodes separately in each sector $m$ and calculate EM field configurations
as series over the orbital momenta outside spheroid [44]
$\overrightarrow{E}^{(m)}(\overrightarrow{r})=\sum_{l=1}^{\infty}[a_{lm}\overrightarrow{M}_{lm}(\overrightarrow{r})+b_{lm}\overrightarrow{N}_{lm}(\overrightarrow{r})]$
(1)
where $\overrightarrow{M}_{lm}$ and $\overrightarrow{N}_{lm}$ are the
spherical harmonics [1]. In what follows we consider the sectors $m=0$ and
$m=1$.
The sector $m=0$ is simplified compared to the sector $m=1$ because of
separation of TE and TM modes. Figure 1 presents evolution of complex TE
resonant frequencies with variation of the equatorial radius $R_{\bot}$
relative to the polar radius $R_{z}$ from oblate silicon spheroid
$R_{z}=0.4R_{\bot}$ to prolate spheroid $R_{z}=1.6R_{\bot}$. $k$ is the wave
number, and $R=(R_{z}R_{\bot}^{2})^{1/3}$ is the mean radius that equalizes
volumes of sphere and spheroid. For the reader convenience we split the
frequency range in Figure 1 into two parts. The insets show the QNMs of a
sphere.
Figure 1: Evolution of complex TE resonant frequencies in silicon spheroid
with permittivity $\epsilon=12$ for variation of aspect ratio of polar $R_{z}$
and equatorial $R_{\bot}$ radii in the sector $m=0$. Wave patterns show
azimuthal component of electric field $|E_{\phi}|$ of the Mie resonant modes
in sphere at points marked by closed circles where integers above the insets
notify the orbital momentum $l$ and the radial index $n$. ’x’ marks the case
of oblate spheroid with $R_{z}=0.4R_{\bot}$ while ’+’ marks the case of
prolate spheroid with $R_{z}=1.6R_{\bot}$. Figure 2: Evolution of resonant
frequencies and resonant modes labelled as 1 and 2 in Fig. 1 versus ratio of
radii $R_{z}$ and $R_{\bot}$.
In Figure 2 we demonstrate a phenomenon of avoided crossing of resonances
marked as 1 and 2 in Figure 1 which is the result of interaction of the dipole
QNM with the octuple QNM [46]. There is a general belief that a homogeneous
spherical dielectric body represents the ideal case, so that any perturbation
of shape of sphere can only degrade the resonance (the imaginary part
increases or the $Q$-factor decreases). Lai et al [8, 46] have shown this,
however, provided that imaginary part of the spherical QNM is small enough.
For the QNMs with low $Q$-factor their frequencies deviate from the complex
eigenfrequencies of sphere linearly [4].
This anomalous behavior of the low-$Q$ resonances can be comprehend if to
refer to the series over spherical harmonics (1). For the TE polarization we
have
$E_{\phi}=\sum_{l}a_{l0}M_{l0}^{\phi}$ (2)
where $l=1,3,5,\ldots$ if $E_{\phi}$ is even relative to $z\rightarrow-z$ and
$l=2,4,6,\ldots$ if $E_{\phi}$ is odd. Once a sphere transforms into spheroid
the orbital momentum $l$ is not preserved. Figure LABEL:TEm0coef shows as new
multipole radiation channels are opened with this transformation.
Figure 3: Evolution of multipole coefficients in series (1) for evolution of
resonant modes $l,n$ shown in Fig. 1.
Let us consider some of resonances shown Figure 1. For variation of the polar
radius $R_{z}$ the lowest mode shown by black line goes through the Mie dipole
mode $1,0$ of a sphere with the frequency $kR=0.862+0.0414i$. As seen from the
first subplot of Figure 3 at this moment the only radiation channel is given
by the coefficient $a_{10}$. The resonant widths of the Mie resonant modes
fast fall down with the orbital momentum $l$ and grow with the radial index
$n$ [47]. As a result, when a sphere is deformed, the fast decaying dipole
channel is weakening at the cost of linear arising of the next slower decay
octuple channel $l=3$ in accordance to Eq. (2). These comprehensive
considerations were issued by Lai et al [46]. Respectively the resonant width
is decreased as shown in Figure 1 by black line. However, there are exceptions
from this rule, for example, the QNMs $l=2,n=1$ and $l=2,n=0$ (The last column
of subplots in Figure 3). In both cases the same slower decaying radiation
channels with $l=4$ and $l=6$ are attaching to the quadruple channel with
$l=2$ for deviation from a sphere. Nevertheless the behavior of resonant
widths is dramatically different as seen from Figure 1. For the radial quantum
$n=0$ we observe a degradation of the quadruple QNM, while for $n=1$ we
observe the opposite behavior. That shows the importance of the radial indices
for resonant widths [47].
Let us consider also the resonances evolving with the Mie resonances with
higher orbital momentum, octuple resonance $3,0$ with the frequency
$kR=1.629+0.0042i$ shown by green line in Figure 1. Corresponding evolution of
multipole coefficients is shown in Figure 3 in subplot labelled $3,0$. In
contrast to previous dipole and quadruple resonances the high-$Q$ decaying
octuple resonance is substituted by the fast decaying dipole resonance $1,0$.
As a result we observe an increase of resonant width in Figure 1 for
transformation of sphere into spheroid. Other subplot $4,0$ in Figure 3 shows
the same result.
We omit analysis of the TM resonances shown in Figure 4 because of a
similarity with the case of the TE resonances except that the series (1) for
magnetic field are given by the coefficients $b_{l0}$ with the same sequence
for $l=1,3,5,\ldots$ for the even solutions of magnetic field $H_{\phi}$ and
$l=2,4,6,\ldots$ for the odd solutions relative to $z\rightarrow-z$.
Figure 4: Evolution of complex TM resonant frequencies. Wave patterns show
azimuthal component of magnetic field $|H_{\phi}|$ of the Mie resonant modes.
’x’ marks $R_{z}=0.4R_{\bot}$ and ’+’ marks $R_{z}=1.6R_{\bot}$. The inset
shows behavior of multipolar coefficients on the aspect ratio.
As a result we have similar rules for resonant widths. The Mie TM dipole and
quadrupole resonances yield to spheroid resonances in the $Q$ factor in
contrast to the Mie resonances with higher orbital momenta. However there is
an exception for the resonance $3,1$ which have no minimal resonant width at
$R_{z}=R_{\bot}$. The inset in Figure 4 shows that in the prolate spheroid we
have extremely large contribution of the spherical harmonic $l=5$ compared to
the dipole harmonic $l=1$ that suppresses emission from the prolate spheroid.
Figure 5: Evolution of resonant frequencies for traversing from the oblate
spheroid $R_{z}=0.4R_{\bot}$ (pluses) through a sphere (closed circles) to the
prolate spheroid $R_{z}=1.6R_{\bot}$ (crosses) in the sector $m=1$. Titles
above the insets indicate the orbital momentum $l$ and radial index $n$ (the
number of radial nodal circles). The TE/TM modes are presented by the
azimuthal components $|E_{\phi}|$/$|H_{\phi}|$.
The sector $m=1$ is destined to show that the phenomena of ARCs exist in the
other sectors of the azimuthal index $m$, in particular $m=1$ as demonstrated
in Figure 5. Moreover one can observe the same tendency of degradation of the
high-$Q$ QNMs and, visa versa, enhancement of the $Q$-factor for the low-$Q$
QNMs for deformation of sphere.
## III Exceptional points.
The sector $m=0$ demonstrates EPs separately for each polarization. Figure 6
shows numerous examples of avoided crossing of TE modes highlighted by open
circles.
Figure 6: ARCs of TE QNMs for evolution of sphere into spheroid in the sector
$m=0$.
It is interesting that the ARC phenomena are observed only for the oblate
spheroids below $R_{z}/R_{\bot}=1/2$. The behavior of QNMs is presented in
Figure 7 which shows as the modes are exchanging for variation of the aspect
ratio of spheroid.
Figure 7: Evolution of selected resonant frequencies and resonant modes
labelled as 1 and 2 in Fig. 6 vs ratio of radii $R_{z}$ and $R_{\bot}$. The
insets show the azimuthal component $|E_{\phi}|$ of corresponding resonant
modes at points marked by closed circles.
As shown in Figure 6 (b) the ARCs are complemented by strong enhancement of
the $Q$-factor in an agreement with numerous considerations in different
dielectric resonators [11, 48, 49, 50].
What is remarkable, the oblate spheroid demonstrates numerous EPs for the two-
fold variation of the permittivity and the aspect ratio for both sectors $m=0$
and $m=1$. Figure LABEL:fig80 shows the behavior of QNMs with the aspect ratio
at $\epsilon=17.2$ in the sector $m=0$. One can see that inside the areas
highlighted by open circles two QNMs coalesce into the one QNM.
Figure 8: Evolution of resonant frequencies and resonant modes versus
$R_{z}/R_{\bot}$ at $\epsilon=17.2$ in the sector $m=0$. Open circles
highlight EPs. The left one at $R_{z}/R_{\bot}=0.292,\epsilon=17.2$ and the
right one at $R_{z}/R_{\bot}=0.304,\epsilon=18.4$. The insets show the
$|E_{\phi}|$ profiles of TE QNMs at points marked by closed circles.
Such a behavior of resonances close to the EP behavior was observed in
different dielectric structures [40, 29, 36, 37, 43].
Figure 9: Encircling of EPs shown by open circles in Fig. LABEL:fig80. (a) and
(b) Encircling separate EPs. (c) Encircling of both EPs. Insets show the
component $E_{\phi}$ of resonant mode.
In order to be convinced that there are indeed the EPs we encircle the EP
points shown open circles in Figure LABEL:fig80 by three ways. In the first
case the rectangular contour encircles only the left EP at the point
$R_{z}/R_{\bot}=0.292,\epsilon=17.2$ as shown in Figure 9 (a). Respectively,
in the second case the contour encircles the right EP point
$R_{z}/R_{\bot}=0.304,\epsilon=18.4$ as shown in Figure 9 (b). At last, we
present also the case of encircling of both EPs shown in Figure 9 (c). In all
cases we encircle EPs counterclockwise.
Let us consider the first case shown in Figure 9 (a) where encircling starts
with point $R_{z}/R_{\bot}=0.32,\epsilon=17$ marked by open circle in the
inset of Figure. In the first downward path we decrease the aspect ratio at
the same permittivity reaching the point till
$R_{z}/R_{\bot}=0.27,\epsilon=17$ marked by cross. In the complex plane this
path maps into sharp trajectory shown by dot-dashed blue line that features
high response of resonant frequency on shape of spheroid. Respectively the
resonant mode demonstrate sharp change of the resonant mode. In the next
horizontal path we slightly increase the permittivity from $\epsilon=17$ till
$\epsilon=17.8$ of the oblate spheroid with the same shape and reach the point
$R_{z}/R_{\bot}=0.27,\epsilon=17.8$ marked by square in the inset. In the
complex plane this path maps into monotonic descent of resonant frequency by
law $(kR)^{2}\epsilon\approx C$ or $kR\approx\sqrt{C/17}(1-\Delta\epsilon/2)$.
That linear part of trajectory is plotted by solid blue line in Figure 9 (a).
The resonant mode presented by the insets at staring and finishing points also
does not show visible changes. The third upward part of rectangular contour
goes from the point marked by square $R_{z}/R_{\bot}=0.27,\epsilon=17.8$ to
the point marked by star $R_{z}/R_{\bot}=0.32,\epsilon=17.8$ maps into sharp
trajectory shown by blue dash line. However the resonant mode is not changing
that is related to far distance between the left EP and the path as distinct
from the first downward path from circle to cross. By doing so we closed the
rectangular contour however as the resonant frequency as the resonant mode are
interchanged as was first demonstrated by Dembowskii et al in a microwave
metallic resonator [19]. And only the second encircling of the left EP
restores the resonant mode as demonstrated in Figure 9 (a) by red lines.
The right EP $R_{z}/R_{\bot}=0.304,\epsilon=18.4$ is expected to give rise to
the same features. However as shown in Figure 9 (b) counterclockwise
encircling of this EP demonstrates clockwise behavior of the resonant
frequency and mode opposite to the case of counterclockwise encircling of the
left EP. That is related to that the signs of winding numbers of neighboring
EPs arising after crossing of two lines in the complex plane are opposite each
other [51, 52]. Figure 9 (c) presents graphical evidence for that. The one
encircling of both EPs restores the resonant modes of each resonance.
Next, we show an existence of EPs in the sector $m=1$ too in which the QNMs
with mixed polarizations can be excited by plane wave incident along the
z-axis as different from the case $m=0$ [1]. The first example of evolution of
the QNMs (only the component $E_{\phi}$ is presented) and their complex
eigenfrequencies in the sector $m=1$ is presented in Figure 10.
Figure 10: Evolution of resonant frequencies and resonant modes marked as 1
and 2 in Fig. 5 (sector $m=1$) versus ratio of radii $R_{z}$ and $R_{\bot}$
around EPs at $\epsilon=12$. The EP is given by the point
$\epsilon=12,R_{z}/R_{\bot}=0.84,kR=1.25$.
The EPs occur for precise two-fold tuning of the aspect ratio $R_{z}/R_{\bot}$
and the refractive index of spheroid that is challengeable experimentally.
However there is a way to show EPs by encircling the EP through which resonant
eigenmodes are interchanged [19]. Figure 11 demonstrates as for encircling of
the EPs in plane $R_{\bot}/R_{z}$ and $\epsilon$ one of resonant modes
restores only after encircling by $4\pi$. It is clear that the same refers to
the multipole coefficients $a_{l1}$ and $b_{l1}$ as shown in Figure 3.
Figure 11: Evolution of the field patterns $E_{y}$ for encircling the EPs
$\epsilon=12,R_{z}/R_{\bot}=0.84$ marked by star. Figure 12: Evolution of the
expansion coefficients $a_{l1}$ (TE modes) and $b_{l1}$ (TM modes) for
encircling the EP shown in Fig. 11.
There are also many other EPs with higher frequencies. One example of the EP
is presented in Figures 13 and 14.
Figure 13: Evolution of resonant frequencies and resonant modes marked as 3
and 4 in Fig. 5 (sector $m=1$) versus ratio of radii $R_{z}$ and $R_{\bot}$
around EPs at $\epsilon=12$. The EP is given by the point
$\epsilon=12.46,R_{z}/R_{\bot}=1.11,kR=2.56$. Figure 14: Evolution of the
field patterns $E_{y}$ for encircling the EPs highlighted in Fig. 13.
## IV Summary and conclusions
It seems reasonable that resonances of any dielectric particle shaped
differently from a sphere yield to the Mie resonances of sphere by the
$Q$-factors because the surface of sphere is minimal. However as Lai et al [8,
46] have shown that is truth only for those resonances whose imaginary part is
small enough. We present numerous examples which confirm this rule and give
comprehensible insight by demonstration of multipole radiation channels for
evolution of a sphere into spheroid. However we also show exceptions from this
rule.
However the main objective of the present paper was demonstration of EPs in a
spheroid that has fundamental significance because of compactness of these
dielectric resonators. Moreover, evolution of expansion coefficients in Fig6
demonstrate multipole conversion for encircling of EPs and what is the most
remarkable this evolution has a period $4\pi$. In the photonic system, the
appearance of EPs can be exploited to a broad range of interesting
applications, including lasing [53], asymmetric mode switching [29],
nonreciprocal light transmission [54, 55], enhancement of the spontaneous
emission [56] and ultrasensitive sensing [57].
###### Acknowledgements.
The work was supported by Russian Foundation for Basic Research projects No.
19-02-00055.
## References
* Stratton [1941] J. A. Stratton, _Electromagnetic theory_ , edited by L. A. DuBridge (McGraw-Hill Book Company, Inc., 1941).
* Conwell _et al._ [1984] P. Conwell, P. Barber, and C. Rushforth, Resonant spectra of dielectric spheres, J. Opt. Soc. Am. A 1, 62 (1984).
* Lai _et al._ [1990a] H. M. Lai, P. T. Leung, K. Young, P. W. Barber, and S. C. Hill, Time-independent perturbation for leaking electromagnetic modes in open systems with application to resonances in microdroplets, Phys. Rev. A 41, 5187 (1990a).
* Ching _et al._ [1998] E. S. C. Ching, P. T. Leung, A. M. van den Brink, W. M. Suen, S. S. Tong, and K. Young, Quasinormal-mode expansion for waves in open systems, Rev. Mod. Phys. 70, 1545 (1998).
* Leung _et al._ [1998] P. T. Leung, W. M. Suen, C. P. Sun, and K. Young, Waves in open systems via a biorthogonal basis, Phys. Rev. E 57, 6101 (1998).
* Okołowicz _et al._ [2003] J. Okołowicz, M. Płoszajczak, and I. Rotter, Dynamics of quantum systems embedded in a continuum, Phys. Rep. 374, 271 (2003).
* Lalanne _et al._ [2018] P. Lalanne, W. Yan, K. Vynck, C. Sauvan, and J.-P. Hugonin, Light interaction with photonic and plasmonic resonances, Laser & Photonics Reviews 12, 1700113 (2018).
* Lai _et al._ [1990b] H. M. Lai, P. T. Leung, and K. Young, Limitations on the photon storage lifetime in electromagnetic resonances of highly transparent microdroplets, Phys. Rev. A 41, 5199 (1990b).
* Heiss [2000] W. D. Heiss, Repulsion of resonance states and exceptional points, Phys. Rev. E 61, 929 (2000).
* Rotter and Sadreev [2005] I. Rotter and A. F. Sadreev, Zeros in single-channel transmission through double quantum dots, Phys. Rev. E 71, 046204 (2005).
* Wiersig [2006] J. Wiersig, Formation of long-lived, scarlike modes near avoided resonance crossings in optical microcavities, Physical Review Letters 97, 10.1103/physrevlett.97.253901 (2006).
* Bernier _et al._ [2018] N. R. Bernier, L. D. Tóth, A. K. Feofanov, and T. J. Kippenberg, Level attraction in a microwave optomechanical circuit, Phys. Rev. A 98, 10.1103/physreva.98.023841 (2018).
* Park _et al._ [2018] K.-W. Park, S. Moon, H. Jeong, J. Kim, and K. Jeong, Non-hermiticity and conservation of orthogonal relation in dielectric microcavity, J. Physics Communications 2, 075007 (2018).
* Friedrich and Wintgen [1985] H. Friedrich and D. Wintgen, Interfering resonances and bound states in the continuum, Phys. Rev. A 32, 3231 (1985).
* Sadreev [2021] A. F. Sadreev, Interference traps waves in an open system: bound states in the continuum, Rep. Progr. Phys. 84, 055901 (2021).
* Heiss and Sannino [1990] W. D. Heiss and A. L. Sannino, Avoided level crossing and exceptional points, J. Phys. A: Math. and Gen. 23, 1167 (1990).
* Heiss [1999] W. Heiss, Phases of wave functions and level repulsion, The European Physical Journal D - Atom., Mol. and Opt. Phys. 7, 1 (1999).
* Eleuch and Rotter [2013] H. Eleuch and I. Rotter, Width bifurcation and dynamical phase transitions in open quantum systems, Phys. Rev. E 87, 10.1103/physreve.87.052136 (2013).
* Dembowski _et al._ [2001] C. Dembowski, H.-D. Graf, H. L. Harney, A. Heine, W. D. Heiss, H. Rehfeld, and A. Richter, Experimental observation of the topological structure of exceptional points, Phys. Rev. Lett. 86, 787 (2001).
* Brandstetter _et al._ [2014] M. Brandstetter, M. Liertzer, C. Deutsch, P. Klang, J. Schöberl, H. E. Türeci, G. Strasser, K. Unterrainer, and S. Rotter, Reversing the pump dependence of a laser at an exceptional point, Nature Communications 5, 10.1038/ncomms5034 (2014).
* Longhi [2017] S. Longhi, Parity-time symmetry meets photonics: A new twist in non-hermitian optics, EPL (Europhysics Letters) 120, 64001 (2017).
* Feng _et al._ [2017] L. Feng, R. El-Ganainy, and L. Ge, Non-hermitian photonics based on parity–time symmetry, Nature Photonics 11, 752 (2017).
* Oezdemir _et al._ [2019] Ş. K. Oezdemir, S. Rotter, F. Nori, and L. Yang, Parity–time symmetry and exceptional points in photonics, Nature Materials 18, 783 (2019).
* Miri and Alù [2019] M.-A. Miri and A. Alù, Exceptional points in optics and photonics, Science 363, 7709 (2019).
* Zhen _et al._ [2015] B. Zhen, C. W. Hsu, Y. Igarashi, L. Lu, I. Kaminer, A. Pick, S.-L. Chua, J. Joannopoulos, and M. Soljačić, Spawning rings of exceptional points out of dirac cones, Nature 525, 354 (2015).
* Kamiński _et al._ [2017] P. M. Kamiński, A. Taghizadeh, O. Breinbjerg, J. Mørk, and S. Arslanagić, Control of exceptional points in photonic crystal slabs, Opt. Lett. 42, 2866 (2017).
* Abdrabou and Lu [2018] A. Abdrabou and Y. Y. Lu, Exceptional points of resonant states on a periodic slab, Phys. Rev. A 97, 10.1103/physreva.97.063822 (2018).
* Abdrabou and Lu [2020] A. Abdrabou and Y. Y. Lu, Exceptional points of bloch eigenmodes on a dielectric slab with a periodic array of cylinders, Physica Scripta 95, 095507 (2020).
* Ghosh and Chong [2016] S. N. Ghosh and Y. D. Chong, Exceptional points and asymmetric mode conversion in quasi-guided dual-mode optical waveguides, Scientific Reports 6, 10.1038/srep19837 (2016).
* Min _et al._ [2020] S. Y. Min, J. Y. Kim, S. Yu, S. G. Menabde, and M. S. Jang, Exceptional points in plasmonic waveguides do not require gain or loss, Phys. Rev. Applied 14, 10.1103/physrevapplied.14.054041 (2020).
* Feng _et al._ [2013] L. Feng, X. Zhu, S. Yang, H. Zhu, P. Zhang, X. Yin, Y. Wang, and X. Zhang, Demonstration of a large-scale optical exceptional point structure, Optics Express 22, 1760 (2013).
* Gomis-Bresco _et al._ [2019] J. Gomis-Bresco, D. Artigas, and L. Torner, Transition from dirac points to exceptional points in anisotropic waveguides, Phys. Rev. Research 1, 033010 (2019).
* Popov _et al._ [2019] V. Popov, S. Tretyakov, and A. Novitsky, Brewster effect when approaching exceptional points of degeneracy: Epsilon-near-zero behavior, Phys. Rev. B 99, 10.1103/physrevb.99.045146 (2019).
* Ryu _et al._ [2012] J.-W. Ryu, S.-Y. Lee, and S. W. Kim, Analysis of multiple exceptional points related to three interacting eigenmodes in a non-hermitian hamiltonian, Phys. Rev. A 85, 10.1103/physreva.85.042101 (2012).
* Kullig _et al._ [2018a] J. Kullig, C.-H. Yi, M. Hentschel, and J. Wiersig, Exceptional points of third-order in a layered optical microdisk cavity, New Journal of Physics 20, 083016 (2018a).
* Yi _et al._ [2019] C.-H. Yi, J. Kullig, M. Hentschel, and J. Wiersig, Non-hermitian degeneracies of internal–external mode pairs in dielectric microdisks, Photonics Research 7, 464 (2019).
* Huang _et al._ [2019] Y. Huang, Y. Shen, and G. Veronis, Non-PT-symmetric two-layer cylindrical waveguide for exceptional-point-enhanced optical devices, Optics Express 27, 37494 (2019).
* Abdrabou and Lu [2019] A. Abdrabou and Y. Y. Lu, Exceptional points for resonant states on parallel circular dielectric cylinders, J. Opt. Soc. Am. B 36, 1659 (2019).
* Unterhinninghofen _et al._ [2008] J. Unterhinninghofen, J. Wiersig, and M. Hentschel, Goos-hänchen shift and localization of optical modes in deformed microcavities, Phys. Rev. E 78, 10.1103/physreve.78.016201 (2008).
* Kullig and Wiersig [2016] J. Kullig and J. Wiersig, Perturbation theory for asymmetric deformed microdisk cavities, Phys. Rev. A 94, 10.1103/physreva.94.043850 (2016).
* Kullig _et al._ [2018b] J. Kullig, C.-H. Yi, and J. Wiersig, Exceptional points by coupling of modes with different angular momenta in deformed microdisks: A perturbative analysis, Phys. Rev. A 98, 10.1103/physreva.98.023851 (2018b).
* Jiang and Xiang [2019] T. Jiang and Y. Xiang, Perturbation model for optical modes in deformed disks, Phys. Rev. A 99, 10.1103/physreva.99.023847 (2019).
* Jiang and Xiang [2020] T. Jiang and Y. Xiang, Perfectly-matched-layer method for optical modes in dielectric cavities, Phys. Rev. A 102, 10.1103/physreva.102.053704 (2020).
* Asano and Yamamoto [1975] S. Asano and G. Yamamoto, Light scattering by a spheroidal particle, Appl. Opt. 14, 29 (1975).
* Barber and Yeh [1975] P. Barber and C. Yeh, Scattering of electromagnetic waves by arbitrarily shaped dielectric bodies, Appl. Opt. 14, 2864 (1975).
* Lai _et al._ [1991] H. M. Lai, C. C. Lam, P. T. Leung, and K. Young, Effect of perturbations on the widths of narrow morphology-dependent resonances in mie scattering, J. Opt. Soc. Am. B 8, 1962 (1991).
* Lam _et al._ [1992] C. C. Lam, P. T. Leung, and K. Young, Explicit asymptotic formulas for the positions, widths, and strengths of resonances in mie scattering, J. Opt. Soc. Am. B 9, 1585 (1992).
* Rybin _et al._ [2017] M. V. Rybin, K. L. Koshelev, Z. F. Sadrieva, K. B. Samusev, A. A. Bogdanov, M. F. Limonov, and Y. S. Kivshar, High-Q supercavity modes in subwavelength dielectric resonators, Phys. Rev. Lett. 119, 243901 (2017).
* Chen _et al._ [2019] W. Chen, Y. Chen, and W. Liu, Multipolar conversion induced subwavelength high-q kerker supermodes with unidirectional radiations, Laser & Photonics Reviews 13, 1900067 (2019).
* Bulgakov _et al._ [2021] E. Bulgakov, K. Pichugin, and A. Sadreev, Mie resonance engineering in two disks, MDPI Photonics 8, 49 (2021).
* Shvartsman and Freund [1994] N. Shvartsman and I. Freund, Vortices in random wave fields: Nearest neighbor anticorrelations, Phys. Rev. Lett. 72, 1008 (1994).
* Berggren _et al._ [2002] K.-F. Berggren, A. Sadreev, and A. Starikov, Crossover from regular to irregular behavior in current flow through open billiards, Phys. Rev. E 66, 10.1103/physreve.66.016218 (2002).
* Feng _et al._ [2014] L. Feng, Z. J. Wong, R.-M. Ma, Y. Wang, and X. Zhang, Single-mode laser by parity-time symmetry breaking, Science 346, 972 (2014).
* Feng _et al._ [2011] L. Feng, M. Ayache, J. Huang, Y.-L. Xu, M.-H. Lu, Y.-F. Chen, Y. Fainman, and A. Scherer, Nonreciprocal light propagation in a silicon photonic circuit, Science 333, 729 (2011).
* Laha _et al._ [2020] A. Laha, S. Dey, H. K. Gandhi, A. Biswas, and S. Ghosh, Exceptional point and toward mode-selective optical isolation, ACS Photonics 7, 967 (2020).
* Pick _et al._ [2017] A. Pick, B. Zhen, O. Miller, C. W. Hsu, F. Hernandez, A. Rodriguez, M. Soljačić, and S. Johnson, General theory of spontaneous emission near exceptional points, Optics Express 25, 12325 (2017).
* Chen _et al._ [2017] W. Chen, Özdemir, G. Zhao, J. Wiersig, and L. Yang, Exceptional points enhance sensing in an optical microcavity, Nature 548, 192 (2017).
|
# Sparse Regression for Extreme Values
Andersen<EMAIL_ADDRESS>[ Minjie<EMAIL_ADDRESS>[
Department of Statistics, Rice University
Genevera I. Allen<EMAIL_ADDRESS>[ Department of Electrical and
Computer Engineering, Rice University,
Department of Computer Science, Rice University,
Department of Statistics, Rice University,
Department of Pediatrics-Neurology, Baylor College of Medicine,
Jan and Dan Duncan Neurological Research Institute, Texas Children’s Hospital
###### Abstract
We study the problem of selecting features associated with extreme values in
high dimensional linear regression. Normally, in linear modeling problems, the
presence of abnormal extreme values or outliers is considered an anomaly which
should either be removed from the data or remedied using robust regression
methods. In many situations, however, the extreme values in regression
modeling are not outliers but rather the signals of interest; consider traces
from spiking neurons, volatility in finance, or extreme events in climate
science, for example. In this paper, we propose a new method for sparse high-
dimensional linear regression for extreme values which is motivated by the
Subbotin, or generalized normal distribution, which we call the extreme value
linear regression model. For our method, we utilize an $\ell_{p}$ norm loss
where $p$ is an even integer greater than two; we demonstrate that this loss
increases the weight on extreme values. We prove consistency and variable
selection consistency for the extreme value linear regression with a Lasso
penalty, which we term the Extreme Lasso, and we also analyze the theoretical
impact of extreme value observations on the model parameter estimates using
the concept of influence functions. Through simulation studies and a real-
world data example, we show that the Extreme Lasso outperforms other methods
currently used in the literature for selecting features of interest associated
with extreme values in high-dimensional regression.
62J05,
62J07,
62P10,
62P05,
linear regression,
sparse modeling,
extreme values,
Subbotin distribution,
generalized normal distribution,
###### keywords:
[class=MSC]
###### keywords:
and
###### Contents
1. 1 Introduction
2. 2 Regression for Extreme Values
1. 2.1 Sparse Extreme Value Regression
3. 3 Theoretical Results
1. 3.1 Consistency of the Extreme Lasso
1. 3.1.1 Sub-Gaussian Errors
2. 3.1.2 Subbotin Error
2. 3.2 Influence of Extreme Values
4. 4 Empirical Investigations
1. 4.1 Linear Model Simulation Study
2. 4.2 Mixture Model Simulation Study
3. 4.3 Real Data Investigation: Calcium Imaging
4. 4.4 Real Data Investigation: Climatology
5. 5 Discussion
6. A Proofs for Section 3
1. A.1 Lemma 3.3
2. A.2 Theorem 3.1
3. A.3 Theorem 3.2
4. A.4 Lemma 3.5
5. A.5 Lemma 3.7
6. A.6 Theorem 3.3
7. A.7 Theorem 3.4
8. A.8 Theorem 3.5
7. B Full Tabular Results
1. B.1 Linear Model Simulation Study
2. B.2 Mixture Model Simulation Study
## 1 Introduction
When applying linear regression models, one often encountered issue is the
presence of rare extreme values, defined here as abnormally large magnitude
observations. This can occur in the form of outliers in the response variable
as well as in the form of highly influential points in the predictor
variables. Historically, statisticians have tried to develop methods to ignore
or dampen the effects of outliers in data sets when doing a linear regression
analysis. Metrics such as residual analysis, Cook’s distance, and DFFIT can be
used to identify and possibly remove outliers from the data set [31]. New
regression methods have also been developed to handle outliers in response
variables as well. For example, robust regression [15] has been used in many
different applications, and much work has been to done to show theoretical
asymptotic performance in the presence of outliers [14, 32]. More recently,
several have studied robust regression procedures for high-dimensional data
[23, 42].
However, in certain contexts, the important information in the response
variable that we want to model or predict is in the rare, abnormally large
magnitude observations. For these types of applications, rather than wanting
to remove outliers or use robust regression methods, we instead want to focus
on these extreme values when fitting models to the data. For example, in
neuroscience, calcium imaging data collected contains measurements of
fluorescence traces of neurons in the imaged brain [43]; the signal that is
important in this situation is the occurrences of neuron firing, indicated by
large positive spikes in the fluorescence trace. Extreme value regression
models are often used as well in climatology to measure the rate and strength
of extreme climate or weather events [21], or in finance to predict periods of
high volatility of asset prices [7]. Their potential usage has also been
studied in spectroscopy analysis and signal processing[25].
Several different possible approaches to the problem of high-dimensional
regression for extreme values have been used in various fields. Sparse
regression methods based on classical extreme value theory utilize a
generalized linear model framework. The extreme values above a predetermined
threshold in a response variable are specified to follow a distribution, such
as the Gumbel, whose parameters are a linear function of the predictor
variables and which determine the frequency and magnitude of the extreme
values [3, 30]. Another regression model commonly applied to model extreme
values in the high-dimensional setting is sparse quantile regression,
specifically applied to a very high or very low quantile [16]. These types of
models use a weighted absolute deviation loss function in order to find the
expected value of a response variable at a particular quantile. Extensions to
high-dimensional sparse $\ell_{1}$ quantile regression have also been studied
extensively [5, 22]. These types of regression methods have shown to be
effective for finding features which are correlated to larger magnitude values
of a response variable when there is ample data to create a reliable model. In
the types of applications we are considering, though, the extreme values tend
to be very rare for a typical set of observations. Because of this, it is
unclear how the desired quantile should be chosen based on the number and
magnitude of the extreme events. The rarity of the extreme values can also
cause the results from the regression model to be numerical unstable due to
the lack of adequate data to get accurate estimates and to sensitivity to
choice of quantile at the extremes. Additionally, quantile regression will not
be as useful in the situation when the response variable of interest has both
positive and negative extreme values, as the model by construction will
upweight the impact of one side of the extreme values while heavily
downweighting the other. Thus, quantile regression potentially restricts us to
focusing only on some of the extreme values while essentially ignoring others.
One other widely-used approach for modeling extreme values involves pre-
processing the data via some type of thresholding algorithm, keeping only the
observed values of each variable which are above either a static or dynamic
threshold and zeroing out the others. Examples of this in different fields
include spike calling or deconvolution in neuroscience [36] or Otsu’s method
in image processing [4]. After these algorithms have been applied to the data,
typical high-dimensional regression methods are then applied to the data. In
general, thresholding data can help in regression analysis for extreme values
by removing any influence from non-extreme values. However, this type of
filtering is not necessarily desirable in all situations. Thresholding
approaches by their nature binarize the observations of a variable in to
extreme and non-extreme categories, whereas in some cases it may make more
sense to smooth the transition from extreme to non-extreme values if it is not
clear where the boundary between the two should lie. Also, the addition of an
extra data pre-processing step can potentially lead to less precise estimates
from the following regression analysis, since any errors made in the former
will propagate to the latter regression step.
In this paper, we explore a different potential approach to tackle the problem
of modeling and predicting extreme values. Our approach to this problem is to
increase the relative weight of larger magnitude losses compared to regular
ordinary least squares. Conceptually, this problem is analogous to increasing
the power of the Gaussian kernel function, which leads to the generalized
normal distribution [35]. Thus, we base our method on $\ell_{p}$ norm
regression, which uses a general $p$ norm for regression rather than the
ordinary $\ell_{2}$ norm. This is a method which has been well-studied as a
whole in the past in the statistics literature [26, 28]. However, much of the
effort in previous research has been focused on showing that $\ell_{p}$-norm
regression can be more robust to outliers [13, 34] by using a norm between 0
and 1. On the other hand, we are interested in using this type of regression
model to create a method which is more sensitive to extreme values in the
response by using norms larger than the squared error loss, i.e. when $p>2$.
By doing this, we skew the regression results toward finding the relationships
with extreme values in a response variable without disregarding potentially
useful observations that could otherwise be ignored by thresholding or
substantially downweighted by quantile regression. We also analyze the
theoretical influence of extreme value observations on our proposed regression
model as well as the finite sample performance guarantees of the estimation
procedure. While general theoretical properties of $\ell_{p}$ norm regression
have been examined in previous literature [18], the performance with respect
to regression for extreme values when $p>2$ for $\ell_{p}$ norm regression has
not been well-studied; this particular situation presents its own unique
theoretical and practical challenges, which we will investigate in this paper.
The rest of the paper is organized as follows. Section 2 introduces and
characterizes the extreme value linear regression method and presents the
algorithm used for parameter estimation. We then prove consistency and
sparsistency results in section 3. Lastly, in section 4, we investigate the
performance of the extreme value linear regression through simulation and real
data studies.
## 2 Regression for Extreme Values
Let $\mathbf{X}\in\mathbb{R}^{n\times p}$ be a data matrix of predictor
variables and $\mathbf{y}\in\mathbb{R}^{n}$ be a corresponding vector of
responses. For our particular problem, we would like to find features in
$\mathbf{X}$ that are correlated with the extreme values of $\mathbf{y}$. (For
simplicity, we presume without loss of generality that each of the variables
are centered and scaled.) In this paper, we consider the context of a linear
data generating model, which will be the focus of the theory presented in
section 3 and the empirical investigations of section 4. Here, we assume that
the data are generated from a simple linear process:
$\mathbf{y}_{i}=\mathbf{X}_{i}\mathbf{\boldsymbol{\beta}}^{*}+\epsilon_{i},\,\epsilon\text{
i.i.d.}.$
In order to produce large magnitude extreme values in the observed response
$\mathbf{y}_{i}$ from this model, either some of the corresponding predictors
at the observed time $\mathbf{X}_{i}$ need to be large in magnitude relative
or some of the parameters in $\mathbf{\boldsymbol{\beta}}^{*}$ need to be
large in magnitude.
Note that we assume that the errors $\epsilon_{i}$ are independently and
identically distributed, but do not necessarily assume that they follow a
Gaussian distribution. In section 3, we will study both the cases where
$\epsilon$ follows a Gaussian distribution and where $\epsilon$ follows a
generalized normal, or Subbotin, distribution [35]. The generalized normal
distribution is defined as
$f(\boldsymbol{\epsilon})=\frac{\gamma}{2\sigma\Gamma(1/\gamma)}e^{-\left(\frac{|\epsilon|}{\sigma}\right)^{\gamma}}$
for scale parameter $\sigma>0$ and shape parameter $\gamma>0$. When
$\gamma=2$, the generalized normal distribution will be equivalent to a
Gaussian distribution, while when $\gamma>2$ the generalized normal
distribution will have a thinner tail compared to a Gaussian. Thus, we are
specifically interested in studying the case where the generalized normal
distribution with $\gamma>2$ as a potential error distribution of the data
generating model, as this relatively discourages the presence of extremely
large residuals in the regression model estimate when compared to a Gaussian
error distribution.
To get estimates of the parameters of the model above, we propose to use the
extreme value linear regression model, which is characterized by the
$\ell_{\gamma}$-norm regression for $\gamma>2$. The foundation for this method
is a generalized linear model applied to the generalized normal distribution
as described above. It follows naturally from the Gaussian case that
estimating the parameters of the generalized normal distribution for a
particular value of $\gamma$ is analogous to minimizing an $\ell_{\gamma}$
norm regression model loss function [26], which is of the form
$\mathcal{L}(\mathbf{y},\mathbf{X},\hat{\mathbf{\boldsymbol{\theta}}})=\frac{1}{\gamma
N}\|\mathbf{y}-\mathbf{X}\hat{\mathbf{\boldsymbol{\theta}}}\|_{\gamma}^{\gamma}$
where $\gamma$ corresponds to the shape parameter in the generalized normal
distribution. As follows from above, we are particularly interested in the
case of $\ell_{\gamma}$ norm regression for $\gamma>2$.
Figure 1: Loss functions for ordinary linear regression, extreme
$\ell_{\gamma}$ norm regression, and quantile regression.
To demonstrate the differences between the different regression methods
discussed in Section 1, we show the respective loss functions for each in
Figure 1. Specifically, we show the extreme linear regression loss function
for $\gamma=4,6,$ and $8$ and the loss for quantile regression at the 0.5 and
0.99 quantiles are shown. Comparing the different methods, we see the
advantage that the $\gamma$-th power error loss has over the other two loss
functions. Relative to the squared error loss function, the extreme value loss
function puts much less weight on very small residuals. However, as the
magnitude of a residual increases, the weight given by the extreme value loss
function grows exponentially compared to the squared error loss function. In
particular, this means that the extreme linear regression will reduce the
presence of abnormally large residuals, which occur when there is an extreme
value in the response variable which is not captured by the estimate from the
model. Thus, the extreme linear regression will find parameter estimates for
the model which better predict the occurrences of the extreme values of a
response variable. Quantile regression, on the other hand, proportionally
increases the relative weight of extreme values by adjusting the weights of an
absolute value loss function using linear constants. However, since the loss
function only grows linearly, the weight of extreme values compared to
relatively large but non-extreme values in the response variable will be small
for the quantile regression loss function compared to the extreme linear
regression loss function. Additionally, quantile regression can only put
increasing weights on either positive or negative residuals in the regression
estimate and not both, meaning that it is not suitable in the case where a
response variable has both positive and negative extreme values.
In terms of the impact to the weight of observations in a regression model,
our extreme linear regression model functions most similarly to the
$\epsilon$-invariant loss used in SVM regression [9] and to the heterogeneous
noise regression models [33]. Both of these methods can also be used to
substantially decrease the weight of smaller magnitude residuals compared to
the larger magnitude ones; this is accomplished by the $\epsilon$-invariant
loss by setting the loss for all residuals below a selected magnitude to be 0,
while the heterogeneous noise regression models can be used to increase the
weight of the observations which are large with respect to either the
predictors or the response variable. However, the extreme linear regression
model has some notable advantages compared to these techniques; it is not
sensitive to the choice of a thresholding hyperparameter as is the case for
SVM regression, and it does not require estimation of extra parameters as in
the heterogeneous noise model in order to achieve the desired effect for this
particular application.
### 2.1 Sparse Extreme Value Regression
In high-dimensional regression problems, automatic feature selection
techniques are used to obtain sparse solutions. In many contexts, this is done
by adding a sparsity-inducing regularization penalty. In the case of ordinary
linear regression, this leads to the penalized squared error loss function.
Applying the same idea to the extreme value linear regression model gives the
loss function:
$\min_{\mathbf{\boldsymbol{\beta}}}\frac{1}{2N}\|\mathbf{y}-\mathbf{X}\mathbf{\boldsymbol{\beta}}\|_{\gamma}^{\gamma}+\lambda\mathcal{P}(\beta).$
The form of the extreme value $\ell_{\gamma}$ norm loss function permits the
usage of any type of regularization penalty that can be applied to the
ordinary linear regression case. For example, one can employ more complex
penalties such as SCAD [39] or MCP [44], or specify a more specific structure
with penalties such as the Fused Lasso [37], Group Lasso [10], or Exclusive
Lasso [8].
Input : $\mathbf{y}\in\mathbb{R}^{N\times 1}\mathbf{X}\in\mathbb{R}^{N\times
p}$, $\lambda\geq 0,\gamma>2,\delta>0,0<\alpha<1.$
Initialize: $\mathbf{\boldsymbol{\beta}}^{(0)}=\boldsymbol{0}_{p}$
while
_$\frac{1}{N}\|\boldsymbol{\beta}^{(r)}-\boldsymbol{\beta}^{(r-1)}\|_{1}\geq\delta$_
do
(1) Find gradient $\nabla g(\mathbf{\boldsymbol{\beta}})$ and optimal step
size $t_{r}$ via backtracking:
(a) Set $t_{r}=1$.
(b) Calculate $\nabla
g^{(r)}(\mathbf{\boldsymbol{\beta}}^{(r)})=-\gamma\mathbf{X}^{T}[|y-\mathbf{X}\boldsymbol{\beta}^{(r)}|^{\circ(\gamma-1)}\circ\text{sgn}(y-\mathbf{X}\boldsymbol{\beta}^{(r)})]$
(c) Repeat:
(i)
$\mathbf{z}=\text{prox}_{\lambda*t_{r}\mathcal{P}}(\boldsymbol{\beta}^{(r)}-t_{r}g^{(r)}(\mathbf{\boldsymbol{\beta}}^{(r)}))$
(ii) $t_{r}=\alpha t_{r}$
until $g(\mathbf{z})\leq g(\boldsymbol{\beta}^{(r)})-\nabla
g(\boldsymbol{\beta}^{(r)})^{T}(\boldsymbol{\beta}^{(r)}-\mathbf{z})+\frac{1}{2t_{r}}\|\mathbf{z}-\boldsymbol{\beta}^{(r)}\|_{2}^{2}$
(2) Update $\mathbf{\boldsymbol{\beta}}^{(r+1)}=\mathbf{z}.$
(3) Update $r=r+1.$
end while
return _$\hat{\mathbf{\boldsymbol{\beta}}}=\boldsymbol{\beta}^{(r)}$_.
Algorithm 1 Regularized Extreme Regression Algorithm with Backtracking
Similar to the Lasso and other penalized ordinary linear regression models,
the objective function for the penalized extreme linear regression can be
decomposed in to the sum of two convex functions, the residual norm and the
penalty terms. Thus, a proximal gradient descent algorithm can be used to
estimate $\hat{\mathbf{\boldsymbol{\beta}}}$. Algorithmic convergence
properties of proximal gradient descent algorithms for penalized linear
regression have been well-studied in recent literature. Notably, it has been
shown that the proximal gradient algorithm is guaranteed to converge to a
minimum. Additionally, because the $\ell_{\gamma}$ loss function is convex for
$\gamma>2$, if the regularization penalty is also convex, then the algorithm
is guaranteed to converge to a global solution [29]. Algorithm 1 gives the
general outline of the computational methodology. As a practical
consideration, one can choose the $\gamma$ parameter for the regression model
either by a priori preference or by using a stability selection criteria,
which chooses the $\gamma$ value for the regression model that provides the
most consistent estimates of the sparse feature set based on a bootstrapping
procedure.
## 3 Theoretical Results
In this section, we present theoretical results for the performance of the
sparse extreme value regression method introduced previously. Specifically, we
focus our studies on the Extreme Lasso, i.e. the extreme value $\ell_{\gamma}$
norm estimator with an $\ell_{1}$ regularization penalty. We note that, for
the following results, we assume $\gamma$ to be a fixed parameter rather than
a parameter to estimate. Our analysis below is separated in to two parts.
First, we derive high-dimensional and finite-sample performance guarantees for
the Extreme Lasso estimator, showing that it is consistent and variable
selection consistent under two different error distributions appropriate for
the generalized normal. We then study our method with respect to the concept
of influence functions, a statistic to measure the effect of infinitesimal,
pointwise contamination of the covariates and response variable on the
resulting regression coefficients. Specifically, we formulate the influence
function of the Extreme Lasso regression model and use this to demonstrate
that the Extreme Lasso method is more heavily skewed toward selecting features
associated with extreme values compared to the ordinary Lasso regression
method. Formal proofs for all of the statements in Section 3 can be found in
the Appendix.
### 3.1 Consistency of the Extreme Lasso
We now present theoretical results for consistency and model selection
consistency of the Extreme Lasso. Our results bear similarity to existing
results for the consistency of Lasso-regularized M-estimators; the main
difference between the results presented here and those in previous works lies
in the distributional assumptions of the errors. Specifically, our
contribution lies in deriving concentration bounds for sub-Weibull and sub-
Gamma random variables. Consider the linear data generating model:
$\displaystyle\mathbf{y}_{i}=\mathbf{x}_{i}^{T}\mathbf{\boldsymbol{\beta}}^{*}+\epsilon_{i},\,\epsilon\text{
i.i.d.}.$
The Extreme Lasso regression thus solves the optimization problem:
$\displaystyle\operatorname*{minimize}_{\beta}\sum_{i=1}^{n}|y_{i}-\mathbf{x}_{i}^{T}\operatorname{\boldsymbol{\beta}}|^{\gamma}+\lambda\|\operatorname{\boldsymbol{\beta}}\|_{1}$
For simplicity, we consider the case when $\gamma$ is an even integer. The
problem can now be written as:
$\displaystyle\operatorname*{minimize}_{\beta}\sum_{i=1}^{n}(y_{i}-\mathbf{x}_{i}^{T}\operatorname{\boldsymbol{\beta}})^{\gamma}+\lambda\|\operatorname{\boldsymbol{\beta}}\|_{1}$
Define
$\mathcal{L}(\operatorname{\boldsymbol{\beta}})=\frac{1}{n}\sum_{i=1}^{n}\ell(\mathbf{x}_{i}^{T}\operatorname{\boldsymbol{\beta}}-y_{i}).$
Clearly, $\mathcal{L}$ belongs to the family of M-estimators, whose properties
have been widely studied in literature; in particular, Negahban et al. [27],
Loh et al. [23], and Loh and Wainwright [24] have established the consistency
of M-estimators in the high-dimensional setting. Thus, we can apply the ideas
and theories for high-dimensional M-estimators from these papers to the
Extreme Lasso case to obtain the results for the regularized extreme value
linear regression.
We first state the previous results regarding the consistency and variable
selection consistency for general robust M-estimators which we use below. In
the literature, Negahban et al. [27] established consistency for high-
dimensional M-estimators:
###### Lemma 3.1 (Estimation Consistency [27]).
Suppose $\mathcal{L}$ satisfies the Restricted Strong Convexity (RSC)
condition with curvature $\kappa_{\mathcal{L}}$ and
$\displaystyle\lambda\geq
2\|\nabla\mathcal{L}(\operatorname{\boldsymbol{\beta}}^{*})\|_{\infty}.$
Then $\hat{\operatorname{\boldsymbol{\beta}}}$ exists and satisfies the
bounds:
$\displaystyle\|\hat{\operatorname{\boldsymbol{\beta}}}-\operatorname{\boldsymbol{\beta}}^{*}\|_{2}$
$\displaystyle\leq\frac{3\sqrt{s}}{\kappa_{\mathcal{L}}}\lambda$
where $s=|\text{supp}(\operatorname{\boldsymbol{\beta}}^{*})|$, i.e.,
$\|\operatorname{\boldsymbol{\beta}}^{*}\|_{0}$.
Note that Lemma 3.1 corresponds to Theorem 1 in Negahban et al. [27] assuming
that the restricted strong convexity (RSC) holds with tolerance parameter
$\tau_{\mathcal{L}}=0$. Also, here we consider $\ell_{1}$ penalty and
$\Psi(\mathcal{M})=\sqrt{s}$. Similarly, Loh et al. [23] established model
selection consistency, also known as sparsistency, for high-dimensional robust
M-estimators:
###### Lemma 3.2 (Model Selection Consistency [19]).
Suppose the following conditions hold:
(1) $\ell$ satisfies RSC.
(2) $\ell$ satisfies irrepresentability.
Let $\kappa_{\text{IC}}$ denote the compatibility constant defined in Lee et
al. [19]. Then, for any
$\frac{4\kappa_{\text{IC}}}{\tau}\|\nabla\mathcal{L}(\operatorname{\boldsymbol{\beta}}^{*})\|_{\infty}<\lambda<\frac{\kappa_{\mathcal{L}}^{2}}{2L}\big{(}2\sqrt{s}+\frac{\sqrt{s}}{\kappa_{\text{IC}}}\frac{\tau}{2}\big{)}^{-2}\frac{\tau}{\kappa_{\text{IC}}}$,
the optimal solution to an M-estimator problem is unique and model selection
consistent: $\hat{\beta}\in M$.
Further, if
$\min_{a\in\mathcal{S}}|\beta_{a}^{*}|>\frac{2}{\kappa_{\mathcal{L}}}\big{(}\sqrt{s}+\frac{\tau}{4}\frac{\sqrt{s}}{\kappa_{\text{IC}}}\big{)}\lambda$,
then the estimator is also sign consistent:
$\text{sign}(\hat{\operatorname{\boldsymbol{\beta}}}_{\mathcal{S}})=\text{sign}(\operatorname{\boldsymbol{\beta}}^{*}_{\mathcal{S}})$.
Lemma 3.2 refers to Theorem 3.4 in Lee et al. [19]. The finite constant
$\kappa_{\text{IC}}$ is the compatibility constant between the irrepresentable
term and $\rho^{*}$. $\tau$ is the constant in the irrepresentable condition.
Since we consider the $\ell_{1}$-norm, i.e., $\rho=\|\cdot\|_{1}$, we have
$k_{\rho}=\sqrt{s}$ and $k_{\rho^{*}}=1$ in the theorem. $L$ is a constant
such that
$\|\nabla^{2}\ell(\operatorname{\boldsymbol{\beta}})-\nabla^{2}\ell(\operatorname{\boldsymbol{\beta}}^{*})\|_{2}\leq
L\|\operatorname{\boldsymbol{\beta}}-\operatorname{\boldsymbol{\beta}}^{*}\|_{2}$.
Note in the Lasso problem, it can be shown that $L=0$; hence there is no upper
bound for $\lambda$. In the Extreme Lasso case, in general we have $L\neq 0$
and there is an upper bound for $\lambda$.
Importantly, the results from both Lemma 3.1 and Lemma 3.2 are entirely
deterministic. Thus, we can guarantee that, under certain conditions, the
extreme value linear regression with the Lasso penalty will provide consistent
estimates of the true parameters of the model. Additionally, both Lemma 3.1
and Lemma 3.2 suggest that the key ingredients for statistical consistency are
the boundedness of $\|\nabla\mathcal{L}(\beta^{*})\|_{\infty}$, which
ultimately determines the rate of convergence of
$\hat{\operatorname{\boldsymbol{\beta}}}$ to
$\operatorname{\boldsymbol{\beta}}^{*}$ and the local RSC condition. Notice
that when $\ell$ is the squared error loss, we get the same consistency and
model selection consistency rate for the Lasso regression problem:
$\displaystyle\|\nabla\mathcal{L}(\beta^{*})\|_{\infty}=\frac{1}{n}\|\mathbf{X}^{T}(y-\mathbf{X}\operatorname{\boldsymbol{\beta}}^{*})\|_{\infty}=\|\mathbf{X}^{T}\epsilon\|_{\infty}/n.$
On the other hand, for the Extreme Lasso case, i.e.
$\ell(\mathbf{x}_{i}^{T}\operatorname{\boldsymbol{\beta}}-y_{i})=(y_{i}-x_{i}^{T}\operatorname{\boldsymbol{\beta}})^{\gamma}$,
we have:
$\displaystyle\|\nabla\mathcal{L}(\beta^{*})\|_{\infty}=\gamma\cdot\frac{1}{n}\|\mathbf{X}^{T}\epsilon^{\circ(\gamma-1)}\|_{\infty}.$
To establish complete results for consistency and model selection consistency
for the Extreme Lasso, we first build a concentration bound for the quantity
$\|\nabla\mathcal{L}(\beta^{*})\|_{\infty}$, i.e.,
$\gamma\cdot\frac{1}{n}\|\mathbf{X}^{T}\epsilon^{\circ(\gamma-1)}\|_{\infty}.$
To do this, we first need to build a tail bound for $\epsilon_{i}^{\gamma-1}$,
which will differ under different distributional assumptions on the covariates
and error terms in the linear model. These assumptions on the distributional
properties will come into play in verifying that the inequality and the RSC
condition hold with high probability under the prescribed sample size scaling.
We can then combine the tail bound results with Lemma 3.1 and Lemma 3.2 to
derive full results. Below, we present tail bounds for
$\frac{1}{n}\|\mathbf{X}^{T}\epsilon^{\circ(\gamma-1)}\|_{\infty}$ under two
different distribution assumptions on the error $\epsilon$.
#### 3.1.1 Sub-Gaussian Errors
We first assume that $\epsilon_{i}$ follows a sub-Gaussian distribution, and
we construct a tail bound for a sub-Gaussian random variable raised to a
power.
###### Lemma 3.3 (Tail Bound for Sub-Gaussian Raised to a Power).
For sub-Gaussian random variable $\mathbf{Q}$, we have
$\displaystyle\mathbb{P}(|\mathbf{Q}|^{\gamma-1}\geq t)\leq
2\exp\bigg{\\{}-\frac{t^{2/(\gamma-1)}}{2\sigma^{2}}\bigg{\\}}.$
Under ordinary least squares, i.e. when $\gamma=2$, we get the usual sub-
Gaussian tail bound; when $\gamma=3$, $\mathbf{Q}^{2}$ follows a sub-
exponential distribution. When $\gamma\geq 4$, as we have for the Extreme
Lasso, $\mathbf{Q}^{\gamma-1}$ is neither sub-Gaussian nor sub-exponential.
Instead, in this situation the tail bound will follow what is known in the
literature as a sub-Weibull distribution [17, 38], which we define below.
###### Definition 3.1 (Sub-Weibull Variables).
A random variable $\mathbf{Z}$ is said to be sub-Weibull of order $\alpha>0,$
denoted as sub-Weibull($\alpha$), if
$\|\mathbf{Z}\|_{\psi_{\alpha}}<\infty,\quad\text{ where
}\psi_{\alpha}(x):=\exp\left(x^{\alpha}\right)-1\quad\text{ for }x\geq 0.$
Based on this definition, it follows that if $\mathbf{Z}$ is sub-Weibull
$(\alpha),$ then
$\mathbb{P}(|\mathbf{Z}|\geq t)\leq
2\exp(-\frac{t^{\alpha}}{\|\mathbf{Z}\|_{\psi_{\alpha}}^{\alpha}}),\text{ for
all }t\geq 0.$
In the Extreme Lasso problem, since $\epsilon_{i}$ is sub-Gaussian, we have
$\mathbb{P}(|\epsilon_{i}|^{\gamma-1}\geq t)\leq
2\exp\bigg{\\{}-\frac{t^{2/(\gamma-1)}}{2\sigma^{2}}\bigg{\\}}$, which means
$\epsilon_{i}^{\gamma-1}$ is sub-Weibull, i.e.,
$\|\epsilon_{i}^{\gamma-1}\|_{\psi_{2/(\gamma-1)}}<\infty$. In the literature,
Kuchibhotla and Chakrabortty [17] established concentration inequalities
related to sub-Weibull random variables. We apply the results and build a tail
bound for
$\|\sum_{i=1}^{n}\mathbf{x}_{i}\epsilon_{i}^{\gamma-1}\|_{\infty}/n$, i.e.,
$\|\mathbf{X}^{T}\epsilon^{\circ(\gamma-1)}\|_{\infty}/n$ by making the
substitution $\mathbf{Z}=\epsilon_{i}^{\gamma-1}$. Note that by Negahban et
al. [27], restricted strong convexity (for M-estimators) with respect to the
$\ell_{2}$-norm is equivalent to the restricted eigenvalues condition (for the
Lasso estimator).
###### Lemma 3.4 (Concentration Bound for Sum of Sub-Weibull Random Variables
[17]).
Consider the Lasso estimator for linear regression case. Suppose there exists
$0<\alpha\leq 2$, and $\gamma,K_{n,p}>0$ such that
$\max\bigg{\\{}\|X_{i}\|_{M,\psi_{\alpha}},\|\epsilon_{i}\|_{\psi_{\gamma}}\bigg{\\}}\leq
K_{n,p}\hskip 14.22636pt\text{for all}\hskip 5.69054pt1\leq i\leq n.$
Also suppose $n\geq 2$, $k\geq 1$ and the covariance matrix $\Sigma_{n}$
satisfies $\lambda_{\min}(\Sigma_{n})\geq K_{n,s}$. Then, with probability at
least $1-3(np)^{-1}$,
$\left\|\frac{1}{n}\sum_{i}^{n}X_{i}\epsilon_{i}\right\|_{\infty}\leq
7\sqrt{2}\sigma_{n,p}\sqrt{\frac{\log(np)}{n}}+\frac{C_{\tau}K_{n,p}^{2}(\log(2n))^{1/\tau}(2\log(np))^{1/\tau}}{n}$
where $\frac{1}{\tau}=\frac{1}{\alpha}+\frac{1}{\gamma}$.
###### Theorem 3.1 (Consistency for Sub-Gaussian Error).
Given the Extreme Lasso program with regularization parameter
$\lambda_{n}=2\gamma\big{(}7\sqrt{2}\sigma_{n,p}\sqrt{\frac{\log(np)}{n}}+$
$\frac{C_{\tau}K_{n,p}^{2}(\log(2n))^{1/\tau}(2\log(np))^{1/\tau}}{n}\bigg{)}$,
then with probability at least $1-3(np)^{-1}$, any optimal solution
$\hat{\beta}$ satisfies the bounds:
$\displaystyle\|\hat{\operatorname{\boldsymbol{\beta}}}-\operatorname{\boldsymbol{\beta}}^{*}\|_{2}$
$\displaystyle\leq\frac{6\sqrt{s}}{\kappa_{\mathcal{L}}}\cdot\gamma\bigg{(}7\sqrt{2}\sigma_{n,p}\sqrt{\frac{\log(np)}{n}}+\frac{C_{\tau}K_{n,p}^{2}(\log(2n))^{1/\tau}(2\log(np))^{1/\tau}}{n}\bigg{)}.$
where $\tau=2/(\gamma-1)$.
###### Theorem 3.2 (Model Selection Consistency for Sub-Gaussian Error).
Consider the Extreme Lasso program with sub-Gaussian error. Assume that the
loss $\ell$ satisfies Restricted Strong Convexity and covariance matrices
satisfy irrepresentability. Consider the family of regularization parameters
$\lambda=\frac{4\kappa_{\text{IC}}}{\tau}\cdot\gamma\bigg{(}7\sqrt{2}\sigma_{n,p}\sqrt{\frac{\log(np)}{n}}+\frac{C_{\tau}K_{n,p}^{2}(\log(2n))^{1/\tau}(2\log(np))^{1/\tau}}{n}\bigg{)},$
then the following properties holds with probability greater than
$1-3(np)^{-1}$:
(i) The Lasso has a unique solution with support contained within $S$, i.e.
$S(\hat{\beta})\subset S(\beta^{*})$.
(ii) If $\min_{a\in
S}|\beta^{*}_{a}|>(\frac{\tau}{\kappa_{\text{IC}}}\cdot\frac{1}{4}+1)\cdot\frac{2\sqrt{s}}{\kappa_{\mathcal{L}}}\cdot\frac{4\kappa_{\text{IC}}}{\tau}\cdot\gamma\bigg{[}7\sqrt{2}\sigma_{n,p}\sqrt{\frac{\log(np)}{n}}+\frac{C_{\tau}K_{n,p}^{2}(\log(2n))^{1/\tau}(2\log(np))^{1/\tau}}{n}\bigg{]}$
, the lasso estimator is also sign consistent:
$\text{sign}(\hat{\beta}_{S})=\text{sign}(\beta^{*}_{S})$.
Applying the result of Theorem 3.1 for $\gamma=2$, we can achieve the usual
consistency rate of $\sqrt{k\log p/n}$ for the ordinary squared error Lasso
loss function under the constraint
$K_{\varepsilon,r}(\log(np))^{-1/2}(\log(2n))^{1/2}=o\left(n^{1/2}\right)$
Note that the probability of the bound being satisfied approaches 1 as
$n\to\infty$, and thus the bound is proportional $\log(np)$ instead of the
usual $\log p$. By setting the probability to be $1-O(p^{-1})$, the usual
Lasso rate $\sqrt{k\log p/n}$ can be recovered.
#### 3.1.2 Subbotin Error
In the following section, we assume that $\epsilon$ follows a Subbotin
distribution, i.e., $\epsilon\sim\text{Subbotin}(\gamma)$. We study this
particular distributional assumption as the Extreme Lasso problem is
equivalent to minimizing the negative log-likelihood of the Subbotin
distribution plus the regularization penalty. To see this, recall the
likelihood of Subbotin distribution:
$\displaystyle f_{Y}(\mathbf{y};\mathbf{X};\operatorname{\boldsymbol{\beta}})$
$\displaystyle=c_{1}\prod_{i=1}^{n}\exp\bigg{[}-|y_{i}-\mathbf{x}_{i}^{T}\operatorname{\boldsymbol{\beta}}|^{\gamma}\bigg{]}$
$\displaystyle=c_{1}\exp\bigg{[}-\sum_{i=1}^{n}|y_{i}-\mathbf{x}_{i}^{T}\operatorname{\boldsymbol{\beta}}|^{\gamma}\bigg{]}$
Thus, the negative log-likelihood,
$\ell(\mathbf{x}_{i}^{T}\operatorname{\boldsymbol{\beta}}-y_{i})\propto\sum_{i=1}|y_{i}-\mathbf{x}_{i}^{T}\operatorname{\boldsymbol{\beta}}|^{\gamma}$,
corresponds to the loss function in the Extreme Lasso problem. Similar to
before, our goal is to build a tail bound for
$\|\mathbf{X}^{T}\epsilon^{\circ(\gamma-1)}\|_{\infty}/n$. To do this, we
first observe that $\epsilon_{i}^{\gamma}$ follows a Gamma distribution.
###### Lemma 3.5 (Change of Variables).
Suppose $\mathbf{Z}\sim$ Subbotin($\alpha$), where $\alpha$ is an even
integer, then
$\mathbf{Y}=\mathbf{Z}^{\alpha}\sim Gamma(\frac{1}{\alpha},1).$
Thus, by Lemma 3.5, we have
$\epsilon_{i}^{\gamma}\sim\text{Gamma}(\frac{1}{\theta},1)$. Hence,
$\epsilon_{i}^{\gamma}$ follows a Gamma distribution and can be bounded by
sub-Gamma tail bounds in literature [6]. These results are stated in Lemma 3.6
and used to derive the results for Theorem 3.3 and Theorem 3.4 below.
###### Lemma 3.6 (Concentration Bound for Sub-Gamma Random Variables [6]).
If $\mathbf{Z}\sim$ Gamma($\alpha,\beta$), then we have:
$\displaystyle\mathbb{P}[\mathbf{Z}-\mathbb{E}\mathbf{Z}]\geq\sqrt{2\gamma
t}+ct]\leq e^{-t}\hskip 14.22636pt$
where $\gamma=\alpha\beta^{2}$, $c=\beta$. We call that $\mathbf{Z}$ is sub-
Gamma with $(\gamma,c)$.
###### Lemma 3.7 (Concentration Bound for Sum of Sub-Gamma Random Variables).
If $\mathbf{Z}\sim$ Gamma($\alpha,\beta$), then with probability at least
$1-c_{1}\exp(-c_{2}\log p)$,
$\left\|\frac{1}{n}\sum_{i}^{n}X_{i}\epsilon_{i}\right\|_{\infty}\leq\sqrt{\frac{\log
p}{n}}\bigg{[}2\sqrt{\frac{2}{\gamma}}+\sqrt{\frac{\log p}{n}}\bigg{]}$
###### Theorem 3.3 (Consistency for Subbotin Error).
Given the Extreme Lasso program with regularization parameter
$\lambda_{n}=2\gamma\sqrt{\frac{\log
p}{n}}\bigg{[}2\sqrt{\frac{2}{\gamma}}+\sqrt{\frac{\log p}{n}}\bigg{]}$, then
with probability at least $1-c_{1}\exp(-c_{2}\log p)$, any optimal solution
$\hat{\beta}$ satisfies the bounds:
$\displaystyle\|\hat{\operatorname{\boldsymbol{\beta}}}-\operatorname{\boldsymbol{\beta}}^{*}\|_{2}$
$\displaystyle\leq\frac{6\sqrt{s}}{\kappa_{\mathcal{L}}}\gamma(\sqrt{\frac{\log
p}{n}}\bigg{[}2\sqrt{\frac{2}{\gamma}}+\sqrt{\frac{\log p}{n}}\bigg{]}).$
###### Theorem 3.4 (Model Selection Consistency for Subbotin Error).
Consider the Extreme Lasso program with Subbotin distributed error. Assume
that the loss $\ell$ satisfies Restricted Strong Convexity and the covariance
matrices satisfy irrepresentability. Consider the family of regularization
parameters $\lambda=\frac{4\kappa_{\text{IC}}}{\tau}\gamma\sqrt{\frac{\log
p}{n}}\bigg{[}2\sqrt{\frac{2}{\gamma}}+\sqrt{\frac{\log p}{n}}\bigg{]}$, then
the following properties holds with probability greater than
$1-c_{1}\exp(-c_{2}\log p)$:
(i) The Lasso has a unique solution with support contained within $S$, i.e.
$S(\hat{\beta})\subset S(\beta^{*})$.
(ii) If $\min_{a\in
S}|\beta^{*}_{a}|>(\frac{\tau}{\kappa_{\text{IC}}}\cdot\frac{1}{4}+1)\cdot\frac{2\sqrt{s}}{\kappa_{\mathcal{L}}}\cdot\frac{4\kappa_{\text{IC}}}{\tau}\gamma\sqrt{\frac{\log
p}{n}}\bigg{[}2\sqrt{\frac{2}{\gamma}}+\sqrt{\frac{\log p}{n}}\bigg{]}$, the
lasso estimator is also sign consistent:
$\text{sign}(\hat{\beta}_{S})=\text{sign}(\beta^{*}_{S})$.
Note that Gaussian distribution is equivalent to the Subbotin distribution
when $\theta=2$. Thus, in the case where $\epsilon_{i}$ is a Gaussian random
variable, we have by Lemma 3.5 that $\epsilon_{i}^{2}$ is
Gamma($\frac{1}{2},1)$. Hence, $\epsilon_{i}^{2}$ is sub-Gamma with
$(\frac{1}{2},1)$. Suppose that $\|\mathbf{X}_{j}\|_{\infty}\leq 1$, we then
have $\mathbf{X}_{j}^{T}\epsilon$ is a sub-Gamma$(n/2,1)$ random variable.
Thus, it follows from Lemma 3.6 that, in this particular case, we have:
$\displaystyle\mathbb{P}\big{(}\mathbf{X}_{j}^{T}\epsilon-\mathbb{E}[\mathbf{X}_{j}^{T}\epsilon]\geq
2\sqrt{nt}+t\big{)}\leq e^{-t}.$
However, if we instead use known Lasso results for $\epsilon$ with sub-
Gaussian tail bounds and set $t=\sigma\sqrt{\frac{c\log p}{n}}$, then we have:
$\displaystyle\mathbb{P}\big{(}|\mathbf{X}_{j}^{T}\epsilon|/n\geq{t}\big{)}\leq
2e^{-\frac{nt^{2}}{2\sigma^{2}}}.$
In effect, the sub-Gamma tail bound has an extra term compared to the sub-
Gaussian bound. This can be seen when comparing the result of Theorem 3.3 and
Theorem 3.4 to the Lasso consistency rate derived using sub-Gaussian tail
bounds. Specifically, there is an extra factor of $\frac{\log p}{n}$ in the
consistency rate result from Theorem 3.3 and Theorem 3.4 compared to the
regular Lasso consistency rate. This is to be expected given that the sub-
Gamma is generally a weaker distributional assumption compared to the sub-
Gaussian. However, this does show that the bound for Theorem 3.3 and Theorem
3.4 is not necessarily tight for any particular values of $\theta$.
### 3.2 Influence of Extreme Values
Here, we demonstrate that our Extreme Lasso estimator is better at selecting
features associated with extreme values than the regular Lasso estimator. We
do this by utilizing the concept of influence functions, which have been
previously proposed in the regression literature as a method for analyzing and
quantifying the effect of outliers in data on statistical estimators [12].
However, in previous works, the influence functions have generally been used
in order to demonstrate the robustness of a regression estimator to the
outlier observations. In our case, we consider the opposite direction, where
we show that the Extreme Lasso estimator is more sensitive to the extreme
values and hence tends to select features associated with extreme values more.
To do this, we show that the value of influence function of the Extreme Lasso
is greater than the Lasso, suggesting that our proposed estimator is affected
more by extreme values.
We follow closely the approach by Wang et al. [40]. Denote as
$\delta_{\mathbf{Z}}$ the point mass probability distribution at a fixed point
$\mathbf{z}=\left(\mathbf{x}_{0},y_{0}\right)^{T}\in\mathbb{R}^{p+1}$. Given
the distribution $F$ of $(\mathbf{x},y)$ in $\mathbb{R}^{p+1}$ and proportion
$\epsilon\in(0,1)$, the mixture distribution of $F$ and $\delta_{\mathbf{Z}}$
is $F_{\epsilon}=(1-\epsilon)F+\epsilon\delta_{\mathbf{Z}}$. Let
$\displaystyle\boldsymbol{\beta}_{0}^{*}$
$\displaystyle=\operatorname*{arg\,min}_{\beta}\left[\left\\{\int\left(\|y-\mathbf{X}^{T}\boldsymbol{\beta}\|^{\gamma}\right)\mathrm{d}F\right\\}+\sum_{j=1}^{p}p_{\lambda_{j}}\left(\left|\beta_{j}\right|\right)\right]$
and
$\displaystyle\boldsymbol{\beta}_{\epsilon}^{*}$
$\displaystyle=\operatorname*{arg\,min}_{\beta}\left[\left\\{\int\left(\|y-\mathbf{X}^{T}\boldsymbol{\beta}\|^{\gamma}\right)\mathrm{d}F_{\epsilon}\right\\}+\sum_{j=1}^{p}p_{\lambda_{j}}\left(\left|\beta_{j}\right|\right)\right].$
For the Lasso and Extreme Lasso,
$p_{\lambda_{j}}\left(\left|\beta_{j}\right|\right)=|\beta_{j}|$. For an
exponential-type estimator, the influence function at a point
$\mathbf{z}\in\mathbb{R}^{p+1}$ is defined as
$\operatorname{IF}\left(\mathbf{z};\boldsymbol{\beta}_{0}^{*}\right)=\lim_{\epsilon\rightarrow
0^{+}}\left(\boldsymbol{\beta}_{\epsilon}^{*}-\boldsymbol{\beta}_{0}^{*}\right)/\epsilon,$
as long as the limit exists. We use this definition to derive the specific
form of the influence function for the Extreme Lasso; the result is shown
below in Theorem 3.5.
###### Theorem 3.5 (Influence Function of Extreme Lasso).
For the penalized extreme value regression estimators with
$\ell_{\gamma}$-norm loss, the $j$th element of the influence function
$\operatorname{IF}_{j}\left(\mathbf{z};\boldsymbol{\beta}_{0}^{*}\right)$ has
the following form:
$\operatorname{IF}_{j}\left(\mathbf{z};\boldsymbol{\beta}_{0}^{*}\right)\\\
=\begin{cases}0,&\text{ if }\beta_{0j}^{*}=0,\\\ -\Gamma_{j}\left\\{-\gamma
r_{0}^{\gamma-1}x_{0}+v_{2}\right\\},&\text{otherwise,}\end{cases}$
where $\Gamma_{j}$ denotes the $j$th row of
$\left\\{A\left(\gamma_{0}\right)-B\right\\}^{-1},r_{0}=y_{0}-$
$\mathbf{x}_{0}^{T}\boldsymbol{\beta}_{0}^{*}$,
$\displaystyle v_{2}$
$\displaystyle=\left\\{p_{\lambda_{1}}^{\prime}\left(\left|\beta_{01}^{*}\right|\right)\operatorname{sign}\left(\beta_{01}^{*}\right),\ldots,p_{\lambda_{d}}^{\prime}\left(\left|\beta_{0d}^{*}\right|\right)\operatorname{sign}\left(\beta_{0d}^{*}\right)\right\\}^{T},$
$\displaystyle B$
$\displaystyle=\operatorname{diag}\left\\{p_{\lambda_{1}}^{\prime\prime}\left(\left|\beta_{01}^{*}\right|\right),\ldots,p_{\lambda_{d}}^{\prime\prime}\left(\left|\beta_{0d}^{*}\right|\right)\right\\},$
and
$A(\gamma)=\int\mathbf{x}\mathbf{x}^{T}\gamma(\gamma-1)\left(y-\mathbf{x}^{T}\boldsymbol{\beta}_{0}^{*}\right)^{\gamma-2}\times\mathrm{d}F(\mathbf{x},y).$
One important implication of this result is that the Extreme Lasso with
$\ell_{\gamma}$ regression is more sensitive to features containing extreme
values, as formally stated below in Corollary 1.
###### Corollary 1.
The influence function of the Extreme Lasso with $\gamma>2$ is greater than
the influence function of Lasso.
Corollary 1 can be shown by using a direct comparison with the Lasso influence
function, i.e. the case where $\gamma=2$. Specifically, we have:
$\displaystyle\frac{\operatorname{IF}\left((x,y);T_{\text{Extreme}},F_{\beta_{j}}\right)}{\operatorname{IF}\left((x,y);T_{\text{Lasso}},F_{\beta_{j}}\right)}$
$\displaystyle=\frac{\gamma
r_{0}^{\gamma-1}x_{0}-v_{2}}{2r_{0}x_{0}-v_{2}}\cdot\frac{A(2)-B_{1}}{A(\gamma)-B_{1}}.$
Recall that $\beta_{0}^{*}$ is the coefficient of fitting the data without
extreme values. Hence, if $x_{0}$ is an influential point,
$r_{0}=y_{0}-\mathbf{x}_{0}^{T}\beta_{0}^{*}$ is sufficiently large, which
means in this case that $r_{0}^{\gamma-1}\gg r_{0}$ for $\gamma>2$. Hence,
$\frac{\operatorname{IF}\left((x,y);T_{\text{Extreme}},F_{\beta_{j}}\right)}{\operatorname{IF}\left((x,y);T_{\text{Lasso}},F_{\beta_{j}}\right)}>1$,
i.e., the influence function evaluated at $\gamma>2$ is greater than evaluated
at $\gamma=2$. Thus, the Extreme Lasso will be more likely to select features
associated with large magnitude values of $\mathbf{x}$ and $y$ compared to the
ordinary Lasso regression.
## 4 Empirical Investigations
Below, we analyze the performance of sparse extreme value linear regression
below on two sets of simulations studies and two real-world case studies.
### 4.1 Linear Model Simulation Study
We first study the performance of our method on a simulation study with data
generated from the linear model as described in section 2, i.e.
$\mathbf{y}_{i}=\mathbf{X}_{i}\mathbf{\boldsymbol{\beta}}^{*}+\epsilon_{i}$.
We let $\boldsymbol{\epsilon}\overset{iid}{\sim}Gamma(\alpha,\beta)$ (using
the rate parameterization) before centering such that $\bar{\epsilon}=0.$ The
predictor matrices $\mathbf{X}$ contain $n=1000$ observations and $p=750$
features. The columns of the matrix are generated as AR(1) processes with
variance 1 and a cross-correlation with one other column of $\rho=0.9$. We
then add large positive extreme values to the columns at known observation
points; these are different for each column. The true parameter vector
$\mathbf{\boldsymbol{\beta}}^{*}$ is set to have 10 randomly selected nonzero
entries. Our goal is to recover the full non-zero support of
$\mathbf{\boldsymbol{\beta}}^{*}$ without recovering false positives. We
analyze four different varying simulation specifications:
1. 1.
The signal to noise ratio of the extreme events relative to baseline noise,
which we denote as $\tau$.
2. 2.
The number of extreme events added to each of the columns of $\mathbf{X}$.
3. 3.
The distribution of the errors $\boldsymbol{\epsilon}$.
4. 4.
The number of dimensions $p$, holding the number of observations and parameter
sparsity level constant.
We compare the Extreme Lasso regression model, as defined in section 3, with
the ordinary Lasso, $\ell_{1}$ quantile regression, and Lasso regression after
preprocessing the data using data-driven thresholding. We fit the Extreme
Lasso regression model, as defined in section 3, using $\gamma=4$ and
$\gamma=6$. For $\ell_{1}$ quantile regression, we find parameter estimates at
the 0.5, 0.9, 0.99, and 0.999 quantiles. Data-driven thresholding is done by
using the adaptive CUSUM method [41] to identify extreme values in the
response variable and removing any data which does not correspond to those
observed extreme values. The number of variables for all methods is selected
via approximate oracle sparsity tuning. We use 4 replications for each
scenario. The results for each of the simulations studies are shown below
using average F-1 scores along with the standard deviations across all
replications. The full results, which include F-1 scores, true positive rates,
and false positive rates for each of the simulations, as well as comparisons
with different regularization penalties for the extreme value linear
regression and ordinary linear regression models, can be found in the
Appendix.
##### Scenario 1: Magnitudes of Extreme Values in Response
Here, we change the size of the signal to noise ratio, comparing
$\tau=6,7,11$, and $15$. The results are shown in Table 1. When the signal to
noise ratio of the extreme values is not sufficiently large, none of the
methods are able to select the correct features. Similarly, if the signal to
noise ratio is large enough, all of the methods except quantile regression are
able to pick out the correct features. However, we see that there is a fairly
large window of $\tau$ values in which the Extreme Lasso is able to find the
correct features while ordinary linear regression and thresholding fail.
Table 1: Average F-1 scores, changing relative extreme value magnitudes for the linear model. | $\tau$ = 6 | $\tau$ = 7 | $\tau$ = 11 | $\tau$ = 15
---|---|---|---|---
ExLasso ($\gamma=4$) | 0.196 (0.1382) | 0.209 (0.1778) | 0.875 (0.05) | 0.938 (0.0481)
ExLasso ($\gamma=6$) | 0.296 (0.1416) | 0.782 (0.0894) | 0.85 (0.0577) | 0.938 (0.0481)
Lasso | 0.2 (0.1414) | 0.225 (0.15) | 0.3 (0) | 0.938 (0.0481)
Median | 0.149 (0.0357) | 0.301 (0.2087) | 0.529 (0.0626) | 0.44 (0.1056)
Q0.9 | 0.149 (0.0357) | 0.127 (0.0429) | 0.185 (0.1239) | 0.147 (0.0508)
Q0.99 | 0.095 (0.0394) | 0.09 (0.0194) | 0.102 (0.0355) | 0.111 (0)
Q0.999 | 0.132 (0.1028) | 0.219 (0.2222) | 0.328 (0.2583) | 0.321 (0.1821)
Threshold | 0.028 (0.0556) | 0 (0) | 0 (0) | 0.893 (0.1056)
##### Scenario 2: Number of Extreme Events in Response
We now vary the number of extreme value events $E$ from 1 to 4, with $\tau=6$.
Results are shown in Table 2. As we observed above, in the case of one extreme
event at $\tau=6$, none of the methods do well. When there is more than one
extreme event though, the Extreme Lasso is able to pick out the correct
features. None of the other methods are able to perform nearly as well when we
increase the number of extreme value events in this case, with only a slight
improvement in performance at $E=4$ compared to $E=1$.
Table 2: Average F-1 scores, changing number of extreme events for the linear model. | E = 1 | E = 2 | E = 3 | E = 4
---|---|---|---|---
ExLasso ($\gamma=4$) | 0.875 (0.05) | 0.225 (0.05) | 0.79 (0.0838) | 0.913 (0.0857)
ExLasso ($\gamma=6$) | 0.85 (0.0577) | 0.788 (0.2022) | 0.779 (0.1447) | 0.85 (0.1291)
Lasso | 0.3 (0) | 0.36 (0.1925) | 0.339 (0.0773) | 0.325 (0.05)
Median | 0.529 (0.0626) | 0.513 (0.059) | 0.472 (0.1155) | 0.457 (0.1337)
Q0.9 | 0.185 (0.1239) | 0.301 (0.0809) | 0.311 (0.157) | 0.414 (0.1092)
Q0.99 | 0.102 (0.0355) | 0.107 (0.0053) | 0.099 (0.0048) | 0.126 (0.0376)
Q0.999 | 0.328 (0.2583) | 0.232 (0.0992) | 0.334 (0.0793) | 0.445 (0.2531)
Threshold | 0 (0) | 0 (0) | 0 (0) | 0.075 (0.15)
##### Scenario 3: Error Distributions
In this scenario, we change the distribution of the added errors by changing
the rate parameter of the pre-centered Gamma distribution from which they are
generated. By decreasing the rate parameter, we increase the variance of the
errors and thus increase the probability of the presence of added errors with
magnitudes that are approximately as large as the true extreme events
themselves. We study the cases where $\beta=0.33,0.2,0.125$, and $0.083$ at
$\tau=11$. We can see from Table 3 that, starting from the baseline scenario
with $\beta=0.33$, the increasing rate parameter significantly affects the
Extreme Lasso in terms of accuracy compared to the other methods. This is not
surprising, since we would expect the Extreme Lasso to be more sensitive to
large errors that are not actually true signal. However, even in the scenario
with the largest error variance, the Extreme Lasso still outperform all of the
others. Thus, even in the presence of potentially large residuals, the Extreme
Lasso is still a preferable method compared to the others.
Table 3: Average F-1 scores, changing error distribution for the linear model. | $\beta$ = 0.33 | $\beta$ = 0.2 | $\beta$ = 0.125 | $\beta$ = 0.083
---|---|---|---|---
ExLasso ($\gamma=4$) | 0.875 (0.05) | 0.8 (0.1155) | 0.625 (0.1258) | 0.275 (0.2217)
ExLasso ($\gamma=6$) | 0.85 (0.0577) | 0.75 (0.1732) | 0.682 (0.1284) | 0.425 (0.15)
Lasso | 0.3 (0) | 0.2 (0.1414) | 0.262 (0.1103) | 0.175 (0.15)
Median | 0.529 (0.0626) | 0.338 (0.1134) | 0.46 (0.1078) | 0.403 (0.1414)
Q-0.9 | 0.185 (0.1239) | 0.122 (0.0465) | 0.121 (0.041) | 0.097 (0.0071)
Q-0.99 | 0.102 (0.0355) | 0.092 (0.0055) | 0.092 (0.0096) | 0.097 (0.0096)
Q-0.999 | 0.328 (0.2583) | 0.202 (0.1506) | 0.093 (0.0087) | 0.093 (0.0105)
Threshold | 0 (0) | 0.05 (0.1) | 0.073 (0.0994) | 0 (0)
##### Scenario 4: Number of Dimensions
We change the number of dimensions of the model matrix to study the
performance of the different methods in relatively higher dimensional
settings. We let $P=750,1500,2250,$ and $3000$, while we hold the number of
true features constant (thus decreasing the sparsity level as we increase
$P$). Table 4 shows the results. All of the approaches do tend to decay in
accuracy. In particular, the least squares and the Extreme Lasso methods tend
to show a relatively larger decline in performance, while quantile regression
at large quantiles and thresholding appear to be more stable. Once again
though, even when performance decays in the higher dimensional settings, the
F-1 scores for the Extreme Lasso still exceeds any of the others.
Table 4: Average F-1 scores, changing number of dimensions of predictor matrix in the linear model. | P = 750 | P = 1500 | P = 2250 | P = 3000
---|---|---|---|---
ExLasso ($\gamma=4$) | 0.875 (0.05) | 0.75 (0.0577) | 0.827 (0.0848) | 0.627 (0.2906)
ExLasso ($\gamma=6$) | 0.85 (0.0577) | 0.65 (0.1) | 0.642 (0.1962) | 0.55 (0.1)
Lasso | 0.3 (0) | 0 (0) | 0 (0) | 0 (0)
Median | 0.529 (0.0626) | 0.249 (0.0671) | 0.247 (0.1579) | 0.103 (0.0269)
Q-0.9 | 0.185 (0.1239) | 0.117 (0.0553) | 0.102 (0.0119) | 0.103 (0.0269)
Q-0.99 | 0.102 (0.0355) | 0.1 (0.0041) | 0.089 (0.0051) | 0.095 (0.0037)
Q-0.999 | 0.328 (0.2583) | 0.117 (0.0495) | 0.279 (0.2265) | 0.089 (0.0051)
Threshold | 0 (0) | 0 (0) | 0 (0) | 0 (0)
### 4.2 Mixture Model Simulation Study
Next, we study a case where the extreme value linear regression model is
misspecified with respect to the data generating model, but where the data
still have extreme values. We generate data from a mixture model of the form:
$\mathbf{y}_{i}=\sum_{k=1}^{K}\mathds{1}_{ik}\mathbf{X}_{i}\beta_{k}+\epsilon_{i}$
$\boldsymbol{\epsilon}\overset{iid}{\sim}Gamma(\alpha,\beta).$
We simulate 4 different sets of predictor variables. The first set contains
features which are generated from a mean 0 Gaussian distribution with added
extreme values at several randomly selected observation points. The second set
contains variables simulated from a Gaussian distribution with no extreme
values but which has a mean shift of $2\sigma^{2}$ for half of the
observations. The third set contains variables which exhibit cross-correlation
($\rho=0.9$) to one of the variables in the first feature set, but with
different extreme value observation points. The fourth set contains
uncorrelated white noise variables. We then create a response variable using
the above mixture model with $K=2$ mixture components, where the first
component is comprised of the first set of the simulated predictor variables
with extreme values, and the second component is comprised of the second set
of the simulated predictor variables with mean shift. The first component
creates extreme values in the response variable because of their presence in
the first set of predictor variables, while the second component will be
correlated with the non-extreme values in the response because of the mean
shift of the corresponding predictors. Our goal is thus to recover as the
support set the variables associated with the first mixture component, i.e.
the ones which generate the extreme values in the response, without selecting
any variables from any of the others.
The predictor matrices $\mathbf{X}$ we simulate contain $n=1000$ observations
and $p=750$ columns; 10 features assigned to the each of the first 3 sets of
predictor variables as described above and the rest designated as part of the
last set. As in the linear regression simulation study, we analyze four
different varying simulation specifications:
1. 1.
The signal to noise ratio of the extreme events relative to baseline noise,
$\tau$.
2. 2.
The number of extreme events added to the variables in the first and third
components.
3. 3.
The distribution of the errors $\boldsymbol{\epsilon}$.
4. 4.
The number of dimensions $p$, holding the number of observations and parameter
sparsity level for each of the mixture components constant.
Again, we compare our method with regularized ordinary least squares
regression, $\ell_{1}$ quantile regression, and Lasso regression after
thresholding; we use 4 replications for each scenario, and we compare results
using average F-1 scores. Full results can be found in the Appendix.
##### Scenario 1: Magnitude of Extreme Values of Response Variable
We first vary the size of the signal to noise ratio between $\tau=6,7,9,$ and
$50.$ The results are shown in Table 5. The extreme value methods are able to
select the true features at a relatively smaller level of $\tau$. The least
squares and thresholding methods are unable to select the features associated
with the extreme values until $\tau$ is astronomically large. Meanwhile, the
quantile regression methods appear to do better than many of the other methods
when $\tau$ is relatively small, but the performance does not improve much
with larger values of $\tau$.
Table 5: Average F-1 scores, changing relative extreme value magnitudes for the mixture model. | $\tau$ = 6 | $\tau$ = 7 | $\tau$ = 9 | $\tau$ = 50
---|---|---|---|---
ExLasso ($\gamma=4$) | 0.128 (0.0986) | 0.175 (0.1708) | 0.9 (0.0816) | 1 (0)
ExLasso ($\gamma=6$) | 0.259 (0.3143) | 0.757 (0.0963) | 1 (0) | 1 (0)
Lasso | 0 (0) | 0.075 (0.15) | 0 (0) | 1 (0)
Median | 0.094 (0.0022) | 0.185 (0.1797) | 0.095 (0) | 0.095 (0)
Q0.9 | 0.094 (0.0022) | 0.14 (0.0887) | 0.095 (0) | 0.095 (0)
Q0.99 | 0.179 (0.0599) | 0.098 (0.0193) | 0.312 (0.1434) | 0.739 (0.1504)
Q0.999 | 0.348 (0.0986) | 0.369 (0.1994) | 0.394 (0.1643) | 0.474 (0.1721)
Threshold | 0 (0) | 0.123 (0.1798) | 0.384 (0.392) | 0.977 (0.0455)
##### Scenario 2: Number of Extreme Events in Response
Here, we change the number of extreme value events $E$ from 1 to 4 for
$\tau=6$. Results are shown in Table 6. As we increase the number of extreme
value events, the performance of the extreme value methods steadily increases.
Thresholding and quantile regression also tend to perform slightly better with
more extreme events, although the improvement is not as drastic. The least
squares regression methods never are able to pick any of the features
associated with the extreme events.
Table 6: Average F-1 scores, changing number of extreme events for the mixture model. | E = 1 | E = 2 | E = 3 | E = 4
---|---|---|---|---
ExLasso ($\gamma=4$) | 0.128 (0.0986) | 0.313 (0.1514) | 0.632 (0.1489) | 0.836 (0.0473)
ExLasso ($\gamma=6$) | 0.259 (0.3143) | 0.52 (0.0869) | 0.795 (0.1527) | 0.908 (0.0789)
Lasso | 0 (0) | 0 (0) | 0 (0) | 0 (0)
Median | 0.094 (0.0022) | 0.094 (0.0022) | 0.095 (0) | 0.095 (0)
Q0.9 | 0.094 (0.0022) | 0.094 (0.0022) | 0.095 (0) | 0.095 (0)
Q0.99 | 0.179 (0.0599) | 0.421 (0.1032) | 0.604 (0.093) | 0.65 (0.1238)
Q0.999 | 0.348 (0.0986) | 0.358 (0.0519) | 0.45 (0.1935) | 0.474 (0.1154)
Threshold | 0 (0) | 0.229 (0.1455) | 0.596 (0.1489) | 0.758 (0.1173)
##### Scenario 3: Error Distributions
In this scenario, we vary the distribution of the added errors by changing the
rate parameter to $\beta=0.33,0.2,0.166$, and $0.125$ at $\tau=9$. Table 7
displays the results. Once again, an increase in the rate parameter
significantly degrades the performance the extreme value methods because of
the increased presence of large magnitude errors, while other methods are not
affected nearly as much. We do eventually see a point where the extreme value
methods perform worse than quantile or thresholding.
Table 7: Average F-1 scores, changing error distribution for the mixture model. | $\beta$ = 0.33 | $\beta$ = 0.2 | $\beta$ = 0.166 | $\beta$ = 0.125
---|---|---|---|---
ExLasso ($\gamma=4$) | 0.9 (0.0816) | 0.875 (0.1258) | 0.816 (0.0526) | 0.278 (0.3587)
ExLasso ($\gamma=6$) | 1 (0) | 1 (0) | 0.582 (0.2852) | 0.184 (0.217)
Lasso | 0 (0) | 0 (0) | 0 (0) | 0 (0)
Median | 0.095 (0) | 0.095 (0) | 0.095 (0) | 0.094 (0.0022)
Q-0.9 | 0.095 (0) | 0.095 (0) | 0.095 (0) | 0.094 (0.0022)
Q-0.99 | 0.312 (0.1434) | 0.14 (0.0494) | 0.249 (0.0897) | 0.24 (0.0465)
Q-0.999 | 0.394 (0.1643) | 0.299 (0.0897) | 0.19 (0.0186) | 0.115 (0.0505)
Threshold | 0.384 (0.392) | 0.05 (0.1) | 0.64 (0.0773) | 0.508 (0.2058)
##### Scenario 4: Number of Dimensions
We change the number of dimensions of the model matrix to $P=750,1500,2250,$
and $3000$, with $\tau=9$ and holding the number of features in components 1,
2, and 3 constant. Results are in Table 8. The performance of the extreme
value and least squares methods do not change much with the increased
dimensionality. The quantile regression methods actually perform slightly
better with more dimensions, while thresholding tends to do worse.
Table 8: Average F-1 scores, changing number of dimensions of predictor matrix in the mixture model. | P = 750 | P = 1500 | P = 2250 | P = 3000
---|---|---|---|---
ExLasso ($\gamma=4$) | 0.9 (0.0816) | 0.816 (0.0526) | 0.922 (0.0673) | 0.838 (0.0062)
ExLasso ($\gamma=6$) | 1 (0) | 0.947 (0) | 0.961 (0.0263) | 0.947 (0)
Lasso | 0 (0) | 0 (0) | 0 (0) | 0 (0)
Median | 0.095 (0) | 0.095 (0) | 0.095 (0) | 0.095 (0)
Q-0.9 | 0.095 (0) | 0.095 (0) | 0.095 (0) | 0.095 (0)
Q-0.99 | 0.312 (0.1434) | 0.238 (0.0794) | 0.189 (0.0744) | 0.093 (0.0262)
Q-0.999 | 0.394 (0.1643) | 0.39 (0.0962) | 0.501 (0.0663) | 0.577 (0.2264)
Threshold | 0.384 (0.392) | 0.32 (0.1879) | 0.316 (0.087) | 0 (0)
### 4.3 Real Data Investigation: Calcium Imaging
We now study the performance of regularized extreme value linear regression on
a calcium imaging study from neuroscience, available from the Allen Brain
Atlas Brain Observatory data repository [20]. The data set contains
fluorescence traces of neuronal activity for 227 simultaneously recorded
neurons in the visual cortex of a mouse brain during periods of controlled
visual stimuli. For this study, we work with the parts of the study associated
with drifting grating movies, i.e. during time periods which the mice are
shown moving black and white gratings of various changing angles and
frequencies. Our objective is to predict the recorded fluorescence traces of
each of the neurons, with a specific focus on the large positive extreme
values that represent neuron firing activity. The predictor variables for the
data set are the visual stimulus information from the drifting grating movie,
namely the angular orientation and frequency of the drifting gratings being
shown to the mouse, as well as other recorded data about the activity of the
mouse including treadmill running speed and pupil size and location. We fit
the Lasso and 8th power Extreme Lasso regressions to each neuron independently
and compare both the chosen stimulus features and the predicted neuron
activity traces from each method. Hyperparameter selection for both methods
are performed via 5-fold cross validation.
Figure 2: Top: True (black) vs. predicted (blue) neuron trace from the Lasso.
Times of chosen angular orientation stimuli are highlighted in color. Bottom:
True (black) vs. predicted (red) neuron trace from the Extreme Lasso. Times of
chosen angular orientation stimuli are highlighted in color.
In Figure 2, we look at the results from one particular neuron. In the top of
Figure 2, we see that the prediction from the Lasso does not include any
spikes; instead, we see that the Lasso is fitting to the random baseline
fluctuations in the fluorescence trace, which are likely to be random noise or
measurement artifacts and are not especially useful for this type of data. We
also see that the times of the angular orientation stimulus feature selected
by the Lasso do not correspond with any spiking activity of the neuron,
meaning that the stimuli selected by the Lasso are not particularly
scientifically meaningful in this specific context. On the other hand, in the
bottom of Figure 2, we see that the Extreme Lasso appears to select an angular
orientation stimulus feature for which spikes in the fluorescence trace occur
during the time periods where the angle is shown in the drifting grating
movie. The predictions from the Extreme Lasso reflect the importance of the
extreme values for the estimation procedure as well, with the estimated value
of the fluorescence trace being much more sensitive to the spikes in the
observed data relative to the Lasso.
### 4.4 Real Data Investigation: Climatology
Our second real data example comes from the field of climatology. The data
used here are available from the US EPA AQS Data Mart [1] and from the MERRA-2
project on NASA MDISC [11]. Our goal is to predict and find features
associated with large spikes in the hourly measurements of total volatile
organic compound (TVOC) concentration in parts per billion (ppb) for a single
outdoor monitoring site in Deer Park, Texas. We use as predictor variables for
modeling the contemporaneous average hourly data of various atmospheric
weather conditions, including temperature, humidity, air pressure, ozone
level, wind speed, water vapor concentration, and dew point. From the raw
weather data, we also create new predictors using 1 day moving averages of all
of the aforementioned variables at time lags ranging from concurrent to 7
days. The data set we look at below contains hourly observations from January
1st, 2015 to December 31st, 2017, totaling approximately 52500 total recorded
measurements. We split this in to a training data set which spans the first
two years of our data, and a test data set which spans the final year. For
this case study, we compare the results from Lasso regression and the 10th
power Extreme Lasso regression models. We first perform feature selection with
these two methods using the training data set; for this step, hyperparameter
values are selected using cross-validation. We then fit the corresponding
unbiased regression models to the test data set. Below, we discuss the
features which were selected using the regularized regression methods on the
training data set, and the model predictions and residuals from the unbiased
models on the test data set.
Lasso | Extreme Lasso
---|---
Concurrent hourly air humidity | Concurrent hourly air pressure
Concurrent hourly vapor volume | 1 day average precipitation, 5 day lag
Concurrent hourly dew point | 1 day average precipitation, 6 day lag
Concurrent hourly wind speed | 1 day average precipitation, 7 day lag
1 day average temperature, 0 day lag | 1 day average wind speed, 6 day lag
1 day average humidity, 0 day lag |
1 day average vapor, 0 day lag |
Table 9: Selected predictors from each regularized regression model.
In Table 9, we show the predictors selected by each of the two methods. As we
can see, the Lasso tends to select predictors which are associated with
concurrent and current 1-day average atmospheric weather conditions, such as
concurrent air humidity and wind speed and 1-day average temperature and water
vapor content. On the other hand, the Extreme Lasso mainly selects features
associated with daily average weather conditions from 5-7 days prior,
particularly with respect to precipitation. Thus, we see that the two methods
pick very different sets of predictors. Scientifically, it seems that the
Lasso is finding predictors that tend to be associated with smaller common
fluctuations in TVOCs, while the Extreme Lasso selects predictors that
indicate occurrences of large rainfall events which have been linked in
previous literature to large spikes in pollutant concentrations [2].
Figure 3: Top left: True (black) vs. predicted (blue) hourly TVOC
concentration from the ordinary linear regression model. Top right: True
(black) vs. predicted (red) hourly TVOC concentration. Bottom Left: True vs.
predicted hourly TVOC concentrations from the ordinary linear regression.
Bottom Right: True vs. predicted hourly TVOC concentrations from the extreme
value linear regression.
In Figure 3, we show the predicted TVOC concentrations from the extreme value
linear regression and ordinary linear regression models with their previous
respective selected features on the test data set; in the top row, we see
these plotted over time, and in the bottom we see the predicted and actual
values plotted against each other. As we can see, the linear regression model
appears to be solely capturing the minor fluctuations which occur regularly
across time, but does not seem to capture any of the large spikes in TVOC
concentration which occur several times over the course of a year. On the
other hand, the extreme value linear regression model, while not always
accurate with respect to the smaller value of TVOC, appears to do a much
better job in predicting the instances of extreme events where TVOC
concentrations spike to irregularly high levels. While neither model is
particularly accurate with respect to predicting all of the observed TVOC
concentration values, the extreme value linear regression model actually does
predict occurrences of extreme value events, whereas the the ordinary linear
regression grossly underestimates the TVOC concentration levels when they are
above a few hundred parts per billion.
Figure 4: Average absolute value of model residuals from the extreme value
linear regression and the ordinary linear regression for values above a
concentration threshold.
We analyze more closely the residuals of the extreme of the regression model
estimate fit by both methods in Figure 4. Here, we show the average magnitude
of the regression residuals from the ordinary least squares regression and
extreme value linear regression models for observations with TVOC
concentrations above a changing threshold value. From Figure 4, we see that
the ordinary linear regression predictions are closer to the actual values on
average when we consider the entire data set, i.e. when the threshold is 0.
However, as we start looking only at data points towards the upper quantiles,
we see that the extreme value linear regression begins to outperform the
ordinary linear regression in terms of prediction accuracy. For this
particular example, the extreme value linear regression model becomes more
accurate on average than the ordinary linear regression for observations with
TVOC concentration values above 262 ppm, and the difference between the
prediction accuracy of the two models gradually increases as we consider
smaller, more extreme subsets of the observations of the response variable.
## 5 Discussion
In this paper, we have introduced the extreme value linear regression model, a
potential new methodological approach to linear regression for extreme values.
Our method is motivated by $\ell_{\gamma}$-norm regression, which gives much
more weight to the loss for large magnitude residuals relative to ordinary
least squares. This concept has several advantages over other methods
currently used in the literature, namely that it does not require using a two-
step pipeline of pre-processing the data before analysis, nor does it force
the data to be binarized as either extreme or non-extreme. Our method also
does not necessitate the a priori choice of certain model hyperparameters that
may be difficult to select. Our simulation studies provide promising results
which demonstrate that, for a response variable with rare extreme values, the
extreme value linear regression model with automatic feature selection
performs better than quantile regression, thresholding, and least squares
penalized regression in terms of selecting predictors which are correlated
with the extreme values in the response. We have also shown deterministic
finite sample performance guarantees for consistency and model selection
consistency of the Extreme Lasso regression model under the assumption of a
linear data generating model with different potential error distributions,
demonstrating that the estimates from the extreme value linear regression
model are reliable. The theoretical results here could also be of use for
other types of similar problems. In particular, the concentration bounds and
theory presented for the case of generalized normal distributed errors for
$\gamma>2$ could be applied to generate new theoretical results for other
mathematical statistics problems.
There are several potential areas for future work for the extreme value linear
regression. Our theoretical work has mainly focused on using a simple
$\ell_{1}$ Lasso penalty for regularization under the linear regression data
generating model; however, the extreme values in a response variable could
come from a variety of different data generating models. Theoretical results
for the variance of estimators in the low-dimensional case have also not been
addressed here, and could be of interest for future study. There remains
potential methodological developments for the extreme value linear regression
to explore as well. Just as ordinary regression methods are insufficient for
fitting a model for the extreme values, traditional model selection methods
may not work particularly well in this context. While we use regular cross-
validation to select $\lambda$ during model fitting in our real data examples,
we recognize that this may not be the optimal method. The model selection
problem may instead require more nuanced treatment, as naive cross-validation
methods may not work well when the extreme values are particularly rare.
Though we have described a general approach for selecting $\gamma$ for an
individual data set, additional empirical investigation may be useful for
gaining a better understanding of what values of $\gamma$ are typically useful
for analyzing real-world data. Also, while we have presented a couple of
potential applications, our method has the potential to be applied broadly to
a variety of fields, such as for signal processing or for spectral domain
data; explorations in to other applications could provide new insights in
these areas. In conclusion, we develop a novel method for extreme value
regression modeling that opens many area for future research.
## Appendix A Proofs for Section 3
### A.1 Lemma 3.3
For $t>0$,
$\displaystyle\mathbb{P}(\mathbf{Q}^{\gamma}\geq t)=\mathbb{P}(\mathbf{Q}\geq
t^{{}^{1/\gamma}})=\mathbb{P}(e^{\lambda\mathbf{Q}}\geq e^{\lambda
t^{{}^{1/\gamma}}})\leq\frac{e^{\sigma^{2}\lambda^{2}/2}}{e^{\lambda
t^{{}^{1/\gamma}}}}=\exp\bigg{\\{}\sigma^{2}\lambda^{2}/2-\lambda
t^{{}^{1/\gamma}}\bigg{\\}}.$
The right hand side is minimized by
$\lambda^{*}=\frac{t^{1/\gamma}}{\sigma^{2}}$. Hence, we have
$\displaystyle\mathbb{P}(\mathbf{Q}^{\gamma}\geq
t)\leq\exp\bigg{\\{}-\frac{t^{2/\gamma}}{2\sigma^{2}}\bigg{\\}}.$
$\square$
### A.2 Theorem 3.1
In the Extreme Lasso problem, by Lemma 3.3,
$\|\epsilon_{i}^{\gamma-1}\|_{\psi_{\gamma}}\leq K_{n,p}$ where
$\gamma=\frac{2}{\gamma-1}$. For fixed design $\mathbf{X}$, $\mathbf{X}_{i}$’s
are marginally sub-Weibull $(\infty)$ and
$\max_{1\leq i\leq n}\left\|X_{i}\right\|_{M,\psi_{2}}\leq\max_{1\leq i\leq
n}\left\|X_{i}\right\|_{M,\psi_{\infty}}=\max_{1\leq i\leq n}\max_{1\leq j\leq
p}\left|X_{i}(j)\right|.$
Applying Lemma 3.4 with $\alpha=\infty$, we have $\tau=2/(\gamma-1)$.
Therefore, by choosing $\lambda_{n}$ to be
$\lambda_{n}=2\gamma\big{(}7\sqrt{2}\sigma_{n,p}\sqrt{\frac{\log(np)}{n}}+\frac{C_{\tau}K_{n,p}^{2}(\log(2n))^{1/\tau}(2\log(np))^{1/\tau}}{n}\bigg{)},$
the Extreme Lasso estimator satisfies
$\displaystyle\|\hat{\operatorname{\boldsymbol{\beta}}}-\operatorname{\boldsymbol{\beta}}^{*}\|_{2}$
$\displaystyle\leq\frac{6\sqrt{s}}{\kappa_{\mathcal{L}}}\cdot\gamma\bigg{(}7\sqrt{2}\sigma_{n,p}\sqrt{\frac{\log(np)}{n}}+\frac{C_{\tau}K_{n,p}^{2}(\log(2n))^{1/\tau}(2\log(np))^{1/\tau}}{n}\bigg{)}$
where $\tau=2/(\gamma-1).$
$\square$
### A.3 Theorem 3.2
Similar to Theorem 3.1, we prove model selection Consistency holds by applying
Lemma 3.2 with the concentration bound demonstrated in Lemma 3.4.
$\square$
### A.4 Lemma 3.5
Suppose $\mathbf{Z}\sim$ Subbotin($\alpha$), i.e.
$\displaystyle
f_{\mathbf{Z}}(z)=\frac{\alpha}{2\Gamma(\frac{1}{\alpha})}\exp\big{[}-|z|^{\alpha}\big{]}.$
Let $\mathbf{Y}=\mathbf{Z}^{\alpha}$, then $z=\pm y^{\frac{1}{\alpha}}$,
$|\frac{dz}{dy}|=\frac{1}{\alpha}y^{\frac{1}{\alpha}-1}$ and
$\displaystyle
f_{\mathbf{Y}}(y)=\frac{\alpha}{\Gamma(\frac{1}{\alpha})}\exp[-y]\frac{1}{\alpha}y^{\frac{1}{\alpha}-1}=\frac{1}{\Gamma(\frac{1}{\alpha})}\exp[-y]y^{\frac{1}{\alpha}-1}.$
Thus, $\mathbf{Y}=\mathbf{Z}^{\alpha}\sim$ Gamma$(\frac{1}{\alpha},1)$.
$\square$
### A.5 Lemma 3.7
By Lemma 3.5, we have $\epsilon_{i}^{\theta}\sim$ Gamma$(1/\theta,1)$. Hence,
Lemma 3.6 suggests
$\displaystyle\mathbb{P}\big{(}\epsilon_{i}^{\theta}-\mathbb{E}[\epsilon_{i}^{\theta}]\geq
2\sqrt{2\frac{1}{\theta}t}+t\big{)}\leq e^{-t}.$
If $\gamma-1\leq\theta$, we can show that $\epsilon_{i}^{\gamma-1}$ is also
sub-Gamma with $(\frac{1}{\theta},1)$ as the latter one has lower tail.
$\displaystyle\mathbb{P}\big{(}\epsilon_{i}^{\gamma-1}-\frac{1}{\theta}\geq
2\sqrt{2\frac{1}{\theta}t}+t\big{)}\leq\mathbb{P}\big{(}\epsilon_{i}^{\theta}-\frac{1}{\theta}\geq
2\sqrt{2\frac{1}{\theta}t}+t\big{)}\leq e^{-t}.$
For $\|\mathbf{X}_{j}\|_{\infty}\leq 1$, we have
$\mathbf{X}_{j}^{T}\epsilon^{\gamma-1}$ is sub-Gamma with $(n/\theta,1)$ since
sum of sub-Gamma is also sub-Gamma. From this, we find
$\displaystyle\mathbb{P}\big{(}\mathbf{X}_{j}^{T}\epsilon^{\gamma-1}-\mathbb{E}[\mathbf{X}_{j}^{T}\epsilon^{\gamma-1}]\geq
2\sqrt{2\frac{n}{\theta}t}+t\big{)}\leq e^{-t}.$
By using union bounds, we thus have
$\displaystyle\mathbb{P}\big{(}\|\mathbf{X}^{T}\epsilon^{\gamma-1}\|_{\infty}\geq
2\sqrt{2\frac{n}{\theta}t}+t\big{)}\leq pe^{-t}.$
Choosing $t=\log p$, we get
$\displaystyle\|\mathbf{X}^{T}\epsilon^{\gamma-1}\|_{\infty}\leq
2\sqrt{\frac{2}{\theta}}\sqrt{n\log p}+\log p$
with probability at least $1-c_{1}\exp(-c_{2}\log p)$. This is equivalent to:
$\displaystyle\|\mathbf{X}^{T}\epsilon^{\gamma-1}\|_{\infty}/n$
$\displaystyle\leq 2\sqrt{\frac{2}{\theta}}\sqrt{\frac{\log p}{n}}+\frac{\log
p}{n}$ $\displaystyle=\sqrt{\frac{\log
p}{n}}\bigg{[}2\sqrt{\frac{2}{\theta}}+\sqrt{\frac{\log p}{n}}\bigg{]}$
$\displaystyle\leq\sqrt{\frac{\log
p}{n}}\bigg{[}2\sqrt{\frac{2}{\gamma}}+\sqrt{\frac{\log p}{n}}\bigg{]}$
with probability at least $1-c_{1}\exp(-c_{2}\log p)$.
$\square$
### A.6 Theorem 3.3
By applying Lemma 3.1 with the concentration bound demonstrated in Lemma 3.7,
we have the consistency result.
$\square$
### A.7 Theorem 3.4
By applying Lemma 3.2 with the concentration bound demonstrated in Lemma 3.7,
we have the model consistency result.
$\square$
### A.8 Theorem 3.5
By the KKT condition as required for optimality of
$\boldsymbol{\beta}_{\epsilon}^{*}$, we have:
$\displaystyle(1-\epsilon)\int\left[-\gamma\left(y-\mathbf{x}^{T}\boldsymbol{\beta}_{\epsilon}^{*}\right)^{\gamma-1}\mathbf{x}\right]\times\mathrm{d}F(\mathbf{x},y)$
$\displaystyle+\epsilon\left(-\gamma\left(y_{0}-\mathbf{x}_{0}^{T}\boldsymbol{\beta}_{\epsilon}^{*}\right)^{\gamma-1}\times\mathbf{x}_{0}\right)-v_{1}(\epsilon)=0,$
(A.1)
where
$v_{1}(\epsilon)=\left(p_{\lambda_{1}}^{\prime}\left(\left|\beta_{\epsilon
1}\right|\right)\operatorname{sign}\left(\beta_{\epsilon
1}\right),\ldots,p_{\lambda_{d}}^{\prime}\left(\left|\beta_{\epsilon
d}\right|\right)\operatorname{sign}\left(\beta_{\epsilon
d}\right)\right)^{T}$. Let $r_{0}=y_{0}-\mathbf{x}_{0}^{T}\beta_{0}^{*}$.
Differentiating with respect to $\epsilon$ in both sides of (A.8) and letting
$\epsilon\rightarrow 0$, we obtain
$\displaystyle\int\left[-\gamma(\gamma-1)\left(y-\mathbf{x}^{T}\boldsymbol{\beta}_{\epsilon}^{*}\right)^{\gamma-2}\times\frac{\partial}{\partial\epsilon}\left(y-\mathbf{x}^{T}\boldsymbol{\beta}_{\epsilon}^{*}\right)\mathbf{x}\right]$
$\displaystyle\left.\mathrm{d}F(\mathbf{x},y)\right|_{\epsilon=0}-\frac{\partial
v_{1}(\epsilon)}{\partial\epsilon}$ $\displaystyle=\gamma
r_{0}^{\gamma-1}x_{0}-v_{2},$ (A.2)
where
$v_{2}=\left(p_{\lambda_{1}}^{\prime}\left(\left|\beta_{01}^{*}\right|\right)\operatorname{sign}\left(\beta_{01}^{*}\right),\ldots,p_{\lambda_{d}}^{\prime}\left(\left|\beta_{0d}^{*}\right|\right)\operatorname{sign}\left(\beta_{0d}^{*}\right)\right)^{T}$.
Using (A.8) and (A.8), it can be shown that
$\left(A(\gamma)-B_{1}\right)\left[\operatorname{IF}\left\\{\left(\mathbf{x}_{0},y_{0}\right),\boldsymbol{\beta}_{0}^{*}\right\\}\right]=\gamma
r_{0}^{\gamma-1}x_{0}-v_{2},$
where
$A(\gamma)=\int\mathbf{x}\mathbf{x}^{T}\gamma(\gamma-1)\left(y-\mathbf{x}^{T}\boldsymbol{\beta}_{0}^{*}\right)^{\gamma-2}\times\mathrm{d}F(\mathbf{x},y),$
$\displaystyle B_{1}=$
$\displaystyle\operatorname{diag}\left\\{p_{\lambda_{1}}^{\prime\prime}\left(\left|\beta_{01}^{*}\right|\right)+p_{\lambda_{1}}^{\prime}\left(\left|\beta_{01}^{*}\right|\right)\delta\left(\beta_{01}^{*}\right),\ldots,p_{\lambda_{d}}^{\prime\prime}\left(\left|\beta_{0d}^{*}\right|\right)+p_{\lambda_{d}}^{\prime}\left(\left|\beta_{0d}^{*}\right|\right)\delta\left(\beta_{0d}^{*}\right)\right\\},$
with
$\delta(x)=\begin{cases}+\infty,&\text{ if }x=0,\\\ 0,&\text{ otherwise.
}\end{cases}$
$\square$
## Appendix B Full Tabular Results
For our full results, we also compare different regularization penalties for
the extreme value linear regression model and the linear regression model.
### B.1 Linear Model Simulation Study
Scenario 1: Changing Magnitude of Extreme Values of Response Variable
Table 10: Average F-1 score for changing extreme value magnitude. | $\tau$ = 6 | $\tau$ = 7 | $\tau$ = 11 | $\tau$ = 15
---|---|---|---|---
ExLasso ($\gamma=4$) | 0.196 (0.1382) | 0.209 (0.1778) | 0.875 (0.05) | 0.938 (0.0481)
ExLasso ($\gamma=6$) | 0.296 (0.1416) | 0.782 (0.0894) | 0.85 (0.0577) | 0.938 (0.0481)
ExSCAD 4th | 0.196 (0.1382) | 0.209 (0.1778) | 0.875 (0.05) | 0.938 (0.0481)
ExSCAD 6th | 0.296 (0.1416) | 0.782 (0.0894) | 0.85 (0.0577) | 0.938 (0.0481)
ExMCP 4th | 0.1 (0.1155) | 0.195 (0.1556) | 0.888 (0.0637) | 0.938 (0.0481)
ExMCP 6th | 0.255 (0.0662) | 0.757 (0.1606) | 0.864 (0.0474) | 0.938 (0.0481)
Lasso | 0.2 (0.1414) | 0.225 (0.15) | 0.3 (0) | 0.938 (0.0481)
SCAD | 0.2 (0.1414) | 0.225 (0.15) | 0.3 (0) | 0.938 (0.0481)
MCP | 0.2 (0.1414) | 0.225 (0.15) | 0.3 (0) | 0.963 (0.0477)
Median | 0.149 (0.0357) | 0.301 (0.2087) | 0.529 (0.0626) | 0.44 (0.1056)
Q0.9 | 0.149 (0.0357) | 0.127 (0.0429) | 0.185 (0.1239) | 0.147 (0.0508)
Q0.99 | 0.095 (0.0394) | 0.09 (0.0194) | 0.102 (0.0355) | 0.111 (0)
Q0.999 | 0.132 (0.1028) | 0.219 (0.2222) | 0.328 (0.2583) | 0.321 (0.1821)
Threshold | 0.028 (0.0556) | 0 (0) | 0 (0) | 0.893 (0.1056)
Table 11: Average true positive rates for changing extreme value magnitude. | $\tau$ = 6 | $\tau$ = 7 | $\tau$ = 11 | $\tau$ = 15
---|---|---|---|---
ExLasso ($\gamma=4$) | 0.193 (0.1355) | 0.196 (0.1571) | 0.875 (0.05) | 0.927 (0.0487)
ExLasso ($\gamma=6$) | 0.293 (0.1421) | 0.767 (0.1054) | 0.85 (0.0577) | 0.927 (0.0487)
ExSCAD 4th | 0.193 (0.1355) | 0.196 (0.1571) | 0.875 (0.05) | 0.927 (0.0487)
ExSCAD 6th | 0.293 (0.1421) | 0.767 (0.1054) | 0.85 (0.0577) | 0.927 (0.0487)
ExMCP 4th | 0.1 (0.1155) | 0.191 (0.1488) | 0.877 (0.0517) | 0.927 (0.0487)
ExMCP 6th | 0.262 (0.0828) | 0.764 (0.1467) | 0.855 (0.053) | 0.927 (0.0487)
Lasso | 0.2 (0.1414) | 0.225 (0.15) | 0.3 (0) | 0.927 (0.0487)
SCAD | 0.2 (0.1414) | 0.225 (0.15) | 0.3 (0) | 0.927 (0.0487)
MCP | 0.2 (0.1414) | 0.225 (0.15) | 0.3 (0) | 0.952 (0.0552)
Median | 0.398 (0.2045) | 0.318 (0.16) | 0.446 (0.041) | 0.435 (0.0842)
Q0.9 | 0.398 (0.2045) | 0.164 (0.1179) | 0.158 (0.1061) | 0.145 (0.0449)
Q0.99 | 0.128 (0.1367) | 0.086 (0.0384) | 0.09 (0.0362) | 0.125 (0)
Q0.999 | 0.175 (0.2165) | 0.196 (0.2072) | 0.303 (0.2673) | 0.352 (0.2432)
Threshold | 0.031 (0.0625) | 0 (0) | 0 (0) | 0.864 (0.1174)
Table 12: Average false positive rates for changing extreme value magnitude. | $\tau$ = 6 | $\tau$ = 7 | $\tau$ = 11 | $\tau$ = 15
---|---|---|---|---
ExLasso ($\gamma=4$) | 0.011 (0.0017) | 0.011 (0.0014) | 0.002 (7e-04) | 0.001 (7e-04)
ExLasso ($\gamma=6$) | 0.01 (0.002) | 0.003 (0.0017) | 0.002 (8e-04) | 0.001 (7e-04)
ExSCAD 4th | 0.011 (0.0017) | 0.011 (0.0014) | 0.002 (7e-04) | 0.001 (7e-04)
ExSCAD 6th | 0.01 (0.002) | 0.003 (0.0017) | 0.002 (8e-04) | 0.001 (7e-04)
ExMCP 4th | 0.011 (0.0014) | 0.01 (7e-04) | 0.002 (7e-04) | 0.001 (7e-04)
ExMCP 6th | 0.01 (0.002) | 0.003 (0.0017) | 0.002 (8e-04) | 0.001 (7e-04)
Lasso | 0.011 (0.0019) | 0.01 (0.002) | 0.009 (0) | 0.001 (7e-04)
SCAD | 0.011 (0.0019) | 0.01 (0.002) | 0.009 (0) | 0.001 (7e-04)
MCP | 0.01 (8e-04) | 0.01 (0.002) | 0.009 (0) | 0.001 (8e-04)
Median | 0.004 (0.0061) | 0.009 (0.0046) | 0.011 (0) | 0.008 (0.0013)
Q0.9 | 0.008 (0.0036) | 0.011 (0.0045) | 0.01 (0.003) | 0.008 (0.0016)
Q0.99 | 0.017 (0.0093) | 0.016 (0.0058) | 0.018 (0.0059) | 0.009 (0)
Q0.999 | 0.015 (0.0083) | 0.016 (0.0055) | 0.013 (0.0078) | 0.008 (0.0039)
Threshold | 0.009 (0.0019) | 0.011 (0.0017) | 0.01 (7e-04) | 0.002 (0.0017)
Scenario 2: Changing Number of Extreme Events in Response
Table 13: Average F-1 scores for changing number of extreme events. | E = 1 | E = 2 | E = 3 | E = 4
---|---|---|---|---
ExLasso ($\gamma=4$) | 0.875 (0.05) | 0.225 (0.05) | 0.79 (0.0838) | 0.913 (0.0857)
ExLasso ($\gamma=6$) | 0.85 (0.0577) | 0.788 (0.2022) | 0.779 (0.1447) | 0.85 (0.1291)
ExSCAD 4th | 0.875 (0.05) | 0.225 (0.05) | 0.79 (0.0838) | 0.913 (0.0857)
ExSCAD 6th | 0.85 (0.0577) | 0.788 (0.2022) | 0.779 (0.1447) | 0.85 (0.1291)
ExMCP 4th | 0.888 (0.0637) | 0.25 (0.1) | 0.788 (0.1192) | 0.913 (0.0857)
ExMCP 6th | 0.864 (0.0474) | 0.813 (0.1555) | 0.779 (0.1447) | 0.89 (0.0978)
Lasso | 0.3 (0) | 0.36 (0.1925) | 0.339 (0.0773) | 0.325 (0.05)
SCAD | 0.3 (0) | 0.36 (0.1925) | 0.339 (0.0773) | 0.325 (0.05)
MCP | 0.3 (0) | 0.295 (0.0741) | 0.339 (0.0773) | 0.325 (0.05)
Median | 0.529 (0.0626) | 0.513 (0.059) | 0.472 (0.1155) | 0.457 (0.1337)
Q0.9 | 0.185 (0.1239) | 0.301 (0.0809) | 0.311 (0.157) | 0.414 (0.1092)
Q0.99 | 0.102 (0.0355) | 0.107 (0.0053) | 0.099 (0.0048) | 0.126 (0.0376)
Q0.999 | 0.328 (0.2583) | 0.232 (0.0992) | 0.334 (0.0793) | 0.445 (0.2531)
Threshold | 0 (0) | 0 (0) | 0 (0) | 0.075 (0.15)
Table 14: Average true positive rates for changing number of extreme events. | E = 1 | E = 2 | E = 3 | E = 4
---|---|---|---|---
ExLasso ($\gamma=4$) | 0.875 (0.05) | 0.225 (0.05) | 0.782 (0.0894) | 0.902 (0.0818)
ExLasso ($\gamma=6$) | 0.85 (0.0577) | 0.777 (0.1913) | 0.759 (0.1297) | 0.85 (0.1291)
ExSCAD 4th | 0.875 (0.05) | 0.225 (0.05) | 0.782 (0.0894) | 0.902 (0.0818)
ExSCAD 6th | 0.85 (0.0577) | 0.777 (0.1913) | 0.759 (0.1297) | 0.85 (0.1291)
ExMCP 4th | 0.877 (0.0517) | 0.25 (0.1) | 0.777 (0.0997) | 0.902 (0.0818)
ExMCP 6th | 0.855 (0.053) | 0.802 (0.1436) | 0.759 (0.1297) | 0.882 (0.1133)
Lasso | 0.3 (0) | 0.333 (0.1414) | 0.329 (0.0583) | 0.325 (0.05)
SCAD | 0.3 (0) | 0.333 (0.1414) | 0.329 (0.0583) | 0.325 (0.05)
MCP | 0.3 (0) | 0.291 (0.0676) | 0.329 (0.0583) | 0.325 (0.05)
Median | 0.446 (0.041) | 0.484 (0.078) | 0.449 (0.0937) | 0.444 (0.1012)
Q0.9 | 0.158 (0.1061) | 0.283 (0.0754) | 0.299 (0.148) | 0.409 (0.1113)
Q0.99 | 0.09 (0.0362) | 0.115 (0.0121) | 0.098 (0.0096) | 0.129 (0.0276)
Q0.999 | 0.303 (0.2673) | 0.239 (0.1036) | 0.322 (0.1004) | 0.441 (0.2547)
Threshold | 0 (0) | 0 (0) | 0 (0) | 0.075 (0.15)
Table 15: Average false positive rates for changing number of extreme events. | E = 1 | E = 2 | E = 3 | E = 4
---|---|---|---|---
ExLasso ($\gamma=4$) | 0.002 (7e-04) | 0.01 (7e-04) | 0.003 (0.0013) | 0.001 (0.0011)
ExLasso ($\gamma=6$) | 0.002 (8e-04) | 0.003 (0.0026) | 0.003 (0.0017) | 0.002 (0.0017)
ExSCAD 4th | 0.002 (7e-04) | 0.01 (7e-04) | 0.003 (0.0013) | 0.001 (0.0011)
ExSCAD 6th | 0.002 (8e-04) | 0.003 (0.0026) | 0.003 (0.0017) | 0.002 (0.0017)
ExMCP 4th | 0.002 (7e-04) | 0.01 (0.0014) | 0.003 (0.0013) | 0.001 (0.0011)
ExMCP 6th | 0.002 (8e-04) | 0.003 (0.0019) | 0.003 (0.0017) | 0.002 (0.0017)
Lasso | 0.009 (0) | 0.01 (7e-04) | 0.009 (0) | 0.009 (7e-04)
SCAD | 0.009 (0) | 0.01 (7e-04) | 0.009 (0) | 0.009 (7e-04)
MCP | 0.009 (0) | 0.01 (7e-04) | 0.009 (0) | 0.009 (7e-04)
Median | 0.011 (0) | 0.008 (0.0022) | 0.008 (0.0011) | 0.008 (0.0013)
Q0.9 | 0.01 (0.003) | 0.011 (0.0016) | 0.012 (0.0017) | 0.009 (0.0011)
Q0.99 | 0.018 (0.0059) | 0.01 (0.0013) | 0.012 (0.0013) | 0.011 (0.002)
Q0.999 | 0.013 (0.0078) | 0.01 (0.0017) | 0.01 (0.003) | 0.008 (0.0036)
Threshold | 0.01 (7e-04) | 0.01 (7e-04) | 0.009 (0) | 0.01 (7e-04)
Scenario 3: Changing Error Distribution
Table 16: Average F-1 scores for changing residual distribution. | $\beta$ = 0.33 | $\beta$ = 0.2 | $\beta$ = 0.125 | $\beta$ = 0.083
---|---|---|---|---
ExLasso ($\gamma=4$) | 0.875 (0.05) | 0.8 (0.1155) | 0.625 (0.1258) | 0.275 (0.2217)
ExLasso ($\gamma=6$) | 0.85 (0.0577) | 0.75 (0.1732) | 0.682 (0.1284) | 0.425 (0.15)
ExSCAD 4th | 0.875 (0.05) | 0.8 (0.1155) | 0.625 (0.1258) | 0.275 (0.2217)
ExSCAD 6th | 0.85 (0.0577) | 0.75 (0.1732) | 0.682 (0.1284) | 0.425 (0.15)
ExMCP 4th | 0.888 (0.0637) | 0.825 (0.0957) | 0.625 (0.1258) | 0.275 (0.2217)
ExMCP 6th | 0.864 (0.0474) | 0.788 (0.166) | 0.575 (0.0957) | 0.425 (0.15)
Lasso | 0.3 (0) | 0.2 (0.1414) | 0.262 (0.1103) | 0.175 (0.15)
SCAD | 0.3 (0) | 0.2 (0.1414) | 0.262 (0.1103) | 0.175 (0.15)
MCP | 0.3 (0) | 0.2 (0.1414) | 0.27 (0.1197) | 0.15 (0.1732)
Median | 0.529 (0.0626) | 0.338 (0.1134) | 0.46 (0.1078) | 0.403 (0.1414)
Q-0.9 | 0.185 (0.1239) | 0.122 (0.0465) | 0.121 (0.041) | 0.097 (0.0071)
Q-0.99 | 0.102 (0.0355) | 0.092 (0.0055) | 0.092 (0.0096) | 0.097 (0.0096)
Q-0.999 | 0.328 (0.2583) | 0.202 (0.1506) | 0.093 (0.0087) | 0.093 (0.0105)
Threshold | 0 (0) | 0.05 (0.1) | 0.073 (0.0994) | 0 (0)
Table 17: Average true positive rates for changing residual distribution. | $\beta$ = 0.33 | $\beta$ = 0.2 | $\beta$ = 0.125 | $\beta$ = 0.083
---|---|---|---|---
ExLasso ($\gamma=4$) | 0.875 (0.05) | 0.8 (0.1155) | 0.625 (0.1258) | 0.275 (0.2217)
ExLasso ($\gamma=6$) | 0.85 (0.0577) | 0.75 (0.1732) | 0.667 (0.1247) | 0.425 (0.15)
ExSCAD 4th | 0.875 (0.05) | 0.8 (0.1155) | 0.625 (0.1258) | 0.275 (0.2217)
ExSCAD 6th | 0.85 (0.0577) | 0.75 (0.1732) | 0.667 (0.1247) | 0.425 (0.15)
ExMCP 4th | 0.877 (0.0517) | 0.825 (0.0957) | 0.625 (0.1258) | 0.275 (0.2217)
ExMCP 6th | 0.855 (0.053) | 0.777 (0.1526) | 0.575 (0.0957) | 0.425 (0.15)
Lasso | 0.3 (0) | 0.2 (0.1414) | 0.252 (0.1013) | 0.175 (0.15)
SCAD | 0.3 (0) | 0.2 (0.1414) | 0.252 (0.1013) | 0.175 (0.15)
MCP | 0.3 (0) | 0.2 (0.1414) | 0.266 (0.1146) | 0.15 (0.1732)
Median | 0.446 (0.041) | 0.33 (0.0991) | 0.449 (0.0937) | 0.385 (0.1168)
Q-0.9 | 0.158 (0.1061) | 0.12 (0.0449) | 0.117 (0.034) | 0.094 (0.0136)
Q-0.99 | 0.09 (0.0362) | 0.086 (0.0099) | 0.086 (0.0176) | 0.096 (0.0199)
Q-0.999 | 0.303 (0.2673) | 0.207 (0.1615) | 0.087 (0.0163) | 0.088 (0.0184)
Threshold | 0 (0) | 0.05 (0.1) | 0.072 (0.1048) | 0 (0)
Table 18: Average false positive rates for changing residual distribution. | $\beta$ = 0.33 | $\beta$ = 0.2 | $\beta$ = 0.125 | $\beta$ = 0.083
---|---|---|---|---
ExLasso ($\gamma=4$) | 0.002 (7e-04) | 0.003 (0.0016) | 0.005 (0.0017) | 0.01 (0.003)
ExLasso ($\gamma=6$) | 0.002 (8e-04) | 0.003 (0.0023) | 0.005 (0.0017) | 0.008 (0.002)
ExSCAD 4th | 0.002 (7e-04) | 0.003 (0.0016) | 0.005 (0.0017) | 0.01 (0.003)
ExSCAD 6th | 0.002 (8e-04) | 0.003 (0.0023) | 0.005 (0.0017) | 0.008 (0.002)
ExMCP 4th | 0.002 (7e-04) | 0.002 (0.0013) | 0.005 (0.0017) | 0.01 (0.003)
ExMCP 6th | 0.002 (8e-04) | 0.003 (0.002) | 0.006 (0.0013) | 0.008 (0.002)
Lasso | 0.009 (0) | 0.012 (0.0039) | 0.011 (0.0016) | 0.011 (0.002)
SCAD | 0.009 (0) | 0.012 (0.0039) | 0.011 (0.0016) | 0.011 (0.002)
MCP | 0.009 (0) | 0.011 (0.0019) | 0.01 (0.0014) | 0.011 (0.0023)
Median | 0.011 (0) | 0.009 (0.0019) | 0.008 (0.0013) | 0.009 (8e-04)
Q-0.9 | 0.01 (0.003) | 0.012 (7e-04) | 0.014 (0.0023) | 0.013 (0.0013)
Q-0.99 | 0.018 (0.0059) | 0.015 (0.0017) | 0.015 (0.0029) | 0.013 (0.0026)
Q-0.999 | 0.013 (0.0078) | 0.011 (0.0033) | 0.015 (0.0026) | 0.015 (0.0034)
Threshold | 0.01 (7e-04) | 0.014 (0.0048) | 0.016 (0.0045) | 0.014 (0.0013)
Scenario 4: Changing Number of Dimensions
Table 19: Average F-1 scores for changing number of dimensions. | P = 750 | P = 1500 | P = 2250 | P = 3000
---|---|---|---|---
ExLasso ($\gamma=4$) | 0.875 (0.05) | 0.75 (0.0577) | 0.827 (0.0848) | 0.627 (0.2906)
ExLasso ($\gamma=6$) | 0.85 (0.0577) | 0.65 (0.1) | 0.642 (0.1962) | 0.55 (0.1)
ExSCAD 4th | 0.875 (0.05) | 0.75 (0.0577) | 0.827 (0.0848) | 0.627 (0.2906)
ExSCAD 6th | 0.85 (0.0577) | 0.65 (0.1) | 0.642 (0.1962) | 0.55 (0.1)
ExMCP 4th | 0.888 (0.0637) | 0.75 (0.0577) | 0.615 (0.2091) | 0.425 (0.2062)
ExMCP 6th | 0.864 (0.0474) | 0.664 (0.1218) | 0.521 (0.1279) | 0.6 (0)
Lasso | 0.3 (0) | 0 (0) | 0 (0) | 0 (0)
SCAD | 0.3 (0) | 0 (0) | 0 (0) | 0 (0)
MCP | 0.3 (0) | 0 (0) | 0 (0) | 0 (0)
Median | 0.529 (0.0626) | 0.249 (0.0671) | 0.247 (0.1579) | 0.103 (0.0269)
Q-0.9 | 0.185 (0.1239) | 0.117 (0.0553) | 0.102 (0.0119) | 0.103 (0.0269)
Q-0.99 | 0.102 (0.0355) | 0.1 (0.0041) | 0.089 (0.0051) | 0.095 (0.0037)
Q-0.999 | 0.328 (0.2583) | 0.117 (0.0495) | 0.279 (0.2265) | 0.089 (0.0051)
Threshold | 0 (0) | 0 (0) | 0 (0) | 0 (0)
Table 20: Average true positive rates for changing number of dimensions. | P = 750 | P = 1500 | P = 2250 | P = 3000
---|---|---|---|---
ExLasso ($\gamma=4$) | 0.875 (0.05) | 0.75 (0.0577) | 0.752 (0.0346) | 0.542 (0.2378)
ExLasso ($\gamma=6$) | 0.85 (0.0577) | 0.65 (0.1) | 0.617 (0.1607) | 0.55 (0.1)
ExSCAD 4th | 0.875 (0.05) | 0.75 (0.0577) | 0.752 (0.0346) | 0.542 (0.2378)
ExSCAD 6th | 0.85 (0.0577) | 0.65 (0.1) | 0.617 (0.1607) | 0.55 (0.1)
ExMCP 4th | 0.877 (0.0517) | 0.75 (0.0577) | 0.663 (0.2358) | 0.425 (0.2062)
ExMCP 6th | 0.855 (0.053) | 0.672 (0.0713) | 0.544 (0.1423) | 0.6 (0)
Lasso | 0.3 (0) | 0 (0) | 0 (0) | 0 (0)
SCAD | 0.3 (0) | 0 (0) | 0 (0) | 0 (0)
MCP | 0.3 (0) | 0 (0) | 0 (0) | 0 (0)
Median | 0.446 (0.041) | 0.229 (0.0473) | 0.247 (0.1426) | 0.124 (0.0845)
Q-0.9 | 0.158 (0.1061) | 0.111 (0.0596) | 0.107 (0.0266) | 0.124 (0.0845)
Q-0.99 | 0.09 (0.0362) | 0.101 (0.0083) | 0.081 (0.0084) | 0.091 (0.0068)
Q-0.999 | 0.303 (0.2673) | 0.112 (0.0494) | 0.267 (0.2336) | 0.081 (0.0084)
Threshold | 0 (0) | 0 (0) | 0 (0) | 0 (0)
Table 21: Average false positive rates for changing number of dimensions. | P = 750 | P = 1500 | P = 2250 | P = 3000
---|---|---|---|---
ExLasso ($\gamma=4$) | 0.002 (7e-04) | 0.003 (8e-04) | 0.004 (0) | 0.008 (0.0028)
ExLasso ($\gamma=6$) | 0.002 (8e-04) | 0.005 (0.0014) | 0.005 (0.0019) | 0.006 (0.0014)
ExSCAD 4th | 0.002 (7e-04) | 0.003 (8e-04) | 0.004 (0) | 0.008 (0.0028)
ExSCAD 6th | 0.002 (8e-04) | 0.005 (0.0014) | 0.005 (0.0019) | 0.006 (0.0014)
ExMCP 4th | 0.002 (7e-04) | 0.003 (8e-04) | 0.004 (0.0029) | 0.008 (0.0028)
ExMCP 6th | 0.002 (8e-04) | 0.004 (0.0013) | 0.006 (0.002) | 0.005 (0)
Lasso | 0.009 (0) | 0.015 (0.002) | 0.014 (0) | 0.015 (0.002)
SCAD | 0.009 (0) | 0.015 (0.002) | 0.014 (0) | 0.015 (0.002)
MCP | 0.009 (0) | 0.014 (0.0022) | 0.013 (0.0014) | 0.014 (0.0014)
Median | 0.011 (0) | 0.012 (0.0011) | 0.01 (0.0026) | 0.012 (0.0058)
Q-0.9 | 0.01 (0.003) | 0.014 (0.0028) | 0.013 (0.0028) | 0.015 (0.0029)
Q-0.99 | 0.018 (0.0059) | 0.012 (0.0011) | 0.016 (0.0017) | 0.014 (0.0011)
Q-0.999 | 0.013 (0.0078) | 0.014 (0.0028) | 0.012 (0.0051) | 0.016 (0.0017)
Threshold | 0.01 (7e-04) | 0.021 (7e-04) | 0.02 (0.0098) | 0.016 (0.0052)
### B.2 Mixture Model Simulation Study
Scenario 1: Changing Magnitude of Extreme Values of Response Variable
Table 22: Average F-1 scores for changing magnitude of extreme value magnitude. | $\tau$ = 6 | $\tau$ = 7 | $\tau$ = 9 | $\tau$ = 50
---|---|---|---|---
ExLasso ($\gamma=4$) | 0.128 (0.0986) | 0.175 (0.1708) | 0.9 (0.0816) | 1 (0)
ExLasso ($\gamma=6$) | 0.259 (0.3143) | 0.757 (0.0963) | 1 (0) | 1 (0)
ExSCAD 4th | 0.128 (0.0986) | 0.175 (0.1708) | 0.9 (0.0816) | 1 (0)
ExSCAD 6th | 0.259 (0.3143) | 0.757 (0.0963) | 1 (0) | 1 (0)
ExMCP 4th | 0.028 (0.0556) | 0.106 (0.1222) | 0.82 (0.0688) | 0.972 (0.0556)
ExMCP 6th | 0.105 (0.2105) | 0.653 (0.1974) | 0.946 (0.0454) | 1 (0)
Lasso | 0 (0) | 0.075 (0.15) | 0 (0) | 1 (0)
SCAD | 0 (0) | 0.075 (0.15) | 0 (0) | 1 (0)
MCP | 0 (0) | 0.075 (0.15) | 0 (0) | 1 (0)
Median | 0.094 (0.0022) | 0.185 (0.1797) | 0.095 (0) | 0.095 (0)
Q0.9 | 0.094 (0.0022) | 0.14 (0.0887) | 0.095 (0) | 0.095 (0)
Q0.99 | 0.179 (0.0599) | 0.098 (0.0193) | 0.312 (0.1434) | 0.739 (0.1504)
Q0.999 | 0.348 (0.0986) | 0.369 (0.1994) | 0.394 (0.1643) | 0.474 (0.1721)
Threshold | 0 (0) | 0.123 (0.1798) | 0.384 (0.392) | 0.977 (0.0455)
Table 23: Average true positive rates for changing magnitude of extreme value magnitude. | $\tau$ = 6 | $\tau$ = 7 | $\tau$ = 9 | $\tau$ = 50
---|---|---|---|---
ExLasso ($\gamma=4$) | 0.131 (0.102) | 0.175 (0.1708) | 0.9 (0.0816) | 1 (0)
ExLasso ($\gamma=6$) | 0.246 (0.2936) | 0.742 (0.1067) | 1 (0) | 1 (0)
ExSCAD 4th | 0.131 (0.102) | 0.175 (0.1708) | 0.9 (0.0816) | 1 (0)
ExSCAD 6th | 0.246 (0.2936) | 0.742 (0.1067) | 1 (0) | 1 (0)
ExMCP 4th | 0.031 (0.0625) | 0.112 (0.1315) | 0.842 (0.0618) | 1 (0)
ExMCP 6th | 0.111 (0.2222) | 0.686 (0.1878) | 1 (0) | 1 (0)
Lasso | 0 (0) | 0.075 (0.15) | 0 (0) | 1 (0)
SCAD | 0 (0) | 0.075 (0.15) | 0 (0) | 1 (0)
MCP | 0 (0) | 0.075 (0.15) | 0 (0) | 1 (0)
Median | 0.089 (0.0038) | 0.172 (0.1629) | 0.091 (0) | 0.091 (0)
Q0.9 | 0.089 (0.0038) | 0.131 (0.0795) | 0.091 (0) | 0.091 (0)
Q0.99 | 0.166 (0.0587) | 0.1 (0.0322) | 0.303 (0.1415) | 0.671 (0.1675)
Q0.999 | 0.389 (0.1361) | 0.346 (0.188) | 0.373 (0.1713) | 0.426 (0.1602)
Threshold | 0 (0) | 0.122 (0.1714) | 0.353 (0.3505) | 0.958 (0.0833)
Table 24: Average false positive rates for changing magnitude of extreme value magnitude. | $\tau$ = 6 | $\tau$ = 7 | $\tau$ = 9 | $\tau$ = 50
---|---|---|---|---
ExLasso ($\gamma=4$) | 0.011 (0.0013) | 0.011 (0.0023) | 0.001 (0.0011) | 0 (0)
ExLasso ($\gamma=6$) | 0.01 (0.0028) | 0.004 (0.0017) | 0 (0) | 0 (0)
ExSCAD 4th | 0.011 (0.0013) | 0.011 (0.0023) | 0.001 (0.0011) | 0 (0)
ExSCAD 6th | 0.01 (0.0028) | 0.004 (0.0017) | 0 (0) | 0 (0)
ExMCP 4th | 0.01 (8e-04) | 0.01 (0.0014) | 0.002 (8e-04) | 0 (0)
ExMCP 6th | 0.01 (0.0023) | 0.004 (0.002) | 0 (0) | 0 (0)
Lasso | 0.013 (7e-04) | 0.012 (0.002) | 0.014 (0) | 0 (0)
SCAD | 0.013 (7e-04) | 0.012 (0.002) | 0.014 (0) | 0 (0)
MCP | 0.013 (7e-04) | 0.012 (0.0019) | 0.013 (0.0014) | 0 (0)
Median | 0.014 (7e-04) | 0.012 (0.002) | 0.014 (0) | 0.014 (0)
Q0.9 | 0.014 (8e-04) | 0.014 (8e-04) | 0.014 (0) | 0.014 (7e-04)
Q0.99 | 0.014 (0.004) | 0.014 (0.007) | 0.01 (0.0026) | 0.006 (0.0032)
Q0.999 | 0.007 (0.0032) | 0.011 (0.0052) | 0.01 (0.0036) | 0.01 (0.0051)
Threshold | 0.009 (0.0023) | 0.011 (0.0019) | 0.008 (0.0034) | 0.001 (0.0014)
Scenario 2: Changing Number of Extreme Events in Response
Table 25: Average F-1 scores for changing number of extreme events. | E = 1 | E = 2 | E = 3 | E = 4
---|---|---|---|---
ExLasso ($\gamma=4$) | 0.128 (0.0986) | 0.313 (0.1514) | 0.632 (0.1489) | 0.836 (0.0473)
ExLasso ($\gamma=6$) | 0.259 (0.3143) | 0.52 (0.0869) | 0.795 (0.1527) | 0.908 (0.0789)
ExSCAD 4th | 0.128 (0.0986) | 0.313 (0.1514) | 0.632 (0.1489) | 0.836 (0.0473)
ExSCAD 6th | 0.259 (0.3143) | 0.52 (0.0869) | 0.795 (0.1527) | 0.908 (0.0789)
ExMCP 4th | 0.028 (0.0556) | 0.164 (0.136) | 0.559 (0.2802) | 0.743 (0.1306)
ExMCP 6th | 0.105 (0.2105) | 0.239 (0.2046) | 0.697 (0.2623) | 0.83 (0.0989)
Lasso | 0 (0) | 0 (0) | 0 (0) | 0 (0)
SCAD | 0 (0) | 0 (0) | 0 (0) | 0 (0)
MCP | 0 (0) | 0 (0) | 0 (0) | 0 (0)
Median | 0.094 (0.0022) | 0.094 (0.0022) | 0.095 (0) | 0.095 (0)
Q0.9 | 0.094 (0.0022) | 0.094 (0.0022) | 0.095 (0) | 0.095 (0)
Q0.99 | 0.179 (0.0599) | 0.421 (0.1032) | 0.604 (0.093) | 0.65 (0.1238)
Q0.999 | 0.348 (0.0986) | 0.358 (0.0519) | 0.45 (0.1935) | 0.474 (0.1154)
Threshold | 0 (0) | 0.229 (0.1455) | 0.596 (0.1489) | 0.758 (0.1173)
Table 26: Average true positive rates for changing number of extreme events. | E = 1 | E = 2 | E = 3 | E = 4
---|---|---|---|---
ExLasso ($\gamma=4$) | 0.131 (0.102) | 0.328 (0.1627) | 0.667 (0.1571) | 0.847 (0.0547)
ExLasso ($\gamma=6$) | 0.246 (0.2936) | 0.542 (0.0949) | 0.817 (0.1599) | 0.944 (0.0642)
ExSCAD 4th | 0.131 (0.102) | 0.328 (0.1627) | 0.667 (0.1571) | 0.847 (0.0547)
ExSCAD 6th | 0.246 (0.2936) | 0.542 (0.0949) | 0.817 (0.1599) | 0.944 (0.0642)
ExMCP 4th | 0.031 (0.0625) | 0.182 (0.144) | 0.599 (0.2817) | 0.837 (0.0834)
ExMCP 6th | 0.111 (0.2222) | 0.259 (0.2049) | 0.761 (0.2223) | 0.937 (0.0745)
Lasso | 0 (0) | 0 (0) | 0 (0) | 0 (0)
SCAD | 0 (0) | 0 (0) | 0 (0) | 0 (0)
MCP | 0 (0) | 0 (0) | 0 (0) | 0 (0)
Median | 0.089 (0.0038) | 0.089 (0.0038) | 0.091 (0) | 0.091 (0)
Q0.9 | 0.089 (0.0038) | 0.089 (0.0038) | 0.091 (0) | 0.091 (0)
Q0.99 | 0.166 (0.0587) | 0.397 (0.0874) | 0.588 (0.1032) | 0.691 (0.0929)
Q0.999 | 0.389 (0.1361) | 0.369 (0.0525) | 0.479 (0.2206) | 0.479 (0.1158)
Threshold | 0 (0) | 0.234 (0.1401) | 0.653 (0.1768) | 0.767 (0.1054)
Table 27: Average false positive rates for changing number of extreme events. | E = 1 | E = 2 | E = 3 | E = 4
---|---|---|---|---
ExLasso ($\gamma=4$) | 0.011 (0.0013) | 0.008 (0.0023) | 0.004 (0.0019) | 0.002 (8e-04)
ExLasso ($\gamma=6$) | 0.01 (0.0028) | 0.006 (0.0013) | 0.002 (0.002) | 0.001 (8e-04)
ExSCAD 4th | 0.011 (0.0013) | 0.008 (0.0023) | 0.004 (0.0019) | 0.002 (8e-04)
ExSCAD 6th | 0.01 (0.0028) | 0.006 (0.0013) | 0.002 (0.002) | 0.001 (8e-04)
ExMCP 4th | 0.01 (8e-04) | 0.008 (7e-04) | 0.004 (0.0028) | 0.002 (7e-04)
ExMCP 6th | 0.01 (0.0023) | 0.008 (0.0013) | 0.002 (0.002) | 0.001 (8e-04)
Lasso | 0.013 (7e-04) | 0.013 (7e-04) | 0.013 (7e-04) | 0.013 (7e-04)
SCAD | 0.013 (7e-04) | 0.013 (7e-04) | 0.013 (7e-04) | 0.013 (7e-04)
MCP | 0.013 (7e-04) | 0.012 (0.0013) | 0.012 (0.0019) | 0.012 (0.0019)
Median | 0.014 (7e-04) | 0.014 (7e-04) | 0.014 (0) | 0.014 (0)
Q0.9 | 0.014 (8e-04) | 0.014 (0) | 0.014 (0.0014) | 0.014 (8e-04)
Q0.99 | 0.014 (0.004) | 0.009 (0.0013) | 0.006 (0.0023) | 0.004 (0.0013)
Q0.999 | 0.007 (0.0032) | 0.008 (0.0011) | 0.007 (0.0038) | 0.007 (0.0026)
Threshold | 0.009 (0.0023) | 0.009 (0.0011) | 0.004 (0.0022) | 0.003 (0.0013)
Scenario 3: Changing Error Distribution
Table 28: Average F-1 scores for changing residual distribution. | $\beta$ = 0.33 | $\beta$ = 0.2 | $\beta$ = 0.166 | $\beta$ = 0.125
---|---|---|---|---
ExLasso ($\gamma=4$) | 0.9 (0.0816) | 0.875 (0.1258) | 0.816 (0.0526) | 0.278 (0.3587)
ExLasso ($\gamma=6$) | 1 (0) | 1 (0) | 0.582 (0.2852) | 0.184 (0.217)
ExSCAD 4th | 0.9 (0.0816) | 0.875 (0.1258) | 0.816 (0.0526) | 0.278 (0.3587)
ExSCAD 6th | 1 (0) | 1 (0) | 0.582 (0.2852) | 0.184 (0.217)
ExMCP 4th | 0.82 (0.0688) | 0.818 (0.0819) | 0.735 (0.1126) | 0.288 (0.3796)
ExMCP 6th | 0.946 (0.0454) | 0.917 (0.0556) | 0.693 (0.1714) | 0.144 (0.1744)
Lasso | 0 (0) | 0 (0) | 0 (0) | 0 (0)
SCAD | 0 (0) | 0 (0) | 0 (0) | 0 (0)
MCP | 0 (0) | 0 (0) | 0 (0) | 0 (0)
Median | 0.095 (0) | 0.095 (0) | 0.095 (0) | 0.094 (0.0022)
Q-0.9 | 0.095 (0) | 0.095 (0) | 0.095 (0) | 0.094 (0.0022)
Q-0.99 | 0.312 (0.1434) | 0.14 (0.0494) | 0.249 (0.0897) | 0.24 (0.0465)
Q-0.999 | 0.394 (0.1643) | 0.299 (0.0897) | 0.19 (0.0186) | 0.115 (0.0505)
Threshold | 0.384 (0.392) | 0.05 (0.1) | 0.64 (0.0773) | 0.508 (0.2058)
Table 29: Average true positive rates for changing residual distribution. | $\beta$ = 0.33 | $\beta$ = 0.2 | $\beta$ = 0.166 | $\beta$ = 0.125
---|---|---|---|---
ExLasso ($\gamma=4$) | 0.9 (0.0816) | 0.875 (0.1258) | 0.861 (0.0556) | 0.281 (0.358)
ExLasso ($\gamma=6$) | 1 (0) | 1 (0) | 0.589 (0.2837) | 0.194 (0.2291)
ExSCAD 4th | 0.9 (0.0816) | 0.875 (0.1258) | 0.861 (0.0556) | 0.281 (0.358)
ExSCAD 6th | 1 (0) | 1 (0) | 0.589 (0.2837) | 0.194 (0.2291)
ExMCP 4th | 0.842 (0.0618) | 0.869 (0.1245) | 0.893 (0.1368) | 0.307 (0.3857)
ExMCP 6th | 1 (0) | 1 (0) | 0.821 (0.1798) | 0.17 (0.209)
Lasso | 0 (0) | 0 (0) | 0 (0) | 0 (0)
SCAD | 0 (0) | 0 (0) | 0 (0) | 0 (0)
MCP | 0 (0) | 0 (0) | 0 (0) | 0 (0)
Median | 0.091 (0) | 0.091 (0) | 0.091 (0) | 0.089 (0.0038)
Q-0.9 | 0.091 (0) | 0.091 (0) | 0.091 (0) | 0.089 (0.0038)
Q-0.99 | 0.303 (0.1415) | 0.133 (0.0438) | 0.251 (0.0829) | 0.232 (0.0391)
Q-0.999 | 0.373 (0.1713) | 0.277 (0.0854) | 0.182 (0.0343) | 0.107 (0.0505)
Threshold | 0.353 (0.3505) | 0.05 (0.1) | 0.689 (0.092) | 0.517 (0.2134)
Table 30: Average false positive rates for changing residual distribution. | $\beta$ = 0.33 | $\beta$ = 0.2 | $\beta$ = 0.166 | $\beta$ = 0.125
---|---|---|---|---
ExLasso ($\gamma=4$) | 0.001 (0.0011) | 0.002 (0.0017) | 0.002 (7e-04) | 0.009 (0.0045)
ExLasso ($\gamma=6$) | 0 (0) | 0 (0) | 0.005 (0.0038) | 0.01 (0.0032)
ExSCAD 4th | 0.001 (0.0011) | 0.002 (0.0017) | 0.002 (7e-04) | 0.009 (0.0045)
ExSCAD 6th | 0 (0) | 0 (0) | 0.005 (0.0038) | 0.01 (0.0032)
ExMCP 4th | 0.002 (8e-04) | 0.002 (0.0017) | 0.001 (0.0013) | 0.007 (0.0033)
ExMCP 6th | 0 (0) | 0 (0) | 0.002 (0.0017) | 0.009 (0.0026)
Lasso | 0.014 (0) | 0.014 (0) | 0.014 (0) | 0.013 (7e-04)
SCAD | 0.014 (0) | 0.014 (0) | 0.014 (0) | 0.013 (7e-04)
MCP | 0.013 (0.0014) | 0.013 (7e-04) | 0.012 (0.0016) | 0.012 (0.0019)
Median | 0.014 (0) | 0.014 (0) | 0.014 (0) | 0.014 (7e-04)
Q-0.9 | 0.014 (0) | 0.015 (7e-04) | 0.013 (0.0014) | 0.013 (0.0023)
Q-0.99 | 0.01 (0.0026) | 0.013 (0.002) | 0.01 (0.0023) | 0.011 (0.0013)
Q-0.999 | 0.01 (0.0036) | 0.011 (0.0017) | 0.012 (0.0028) | 0.015 (0.002)
Threshold | 0.008 (0.0034) | 0.011 (7e-04) | 0.004 (0.0013) | 0.006 (0.003)
Scenario 4: Changing Number of Dimensions
Table 31: Average F-1 scores for changing number of dimensions. | P = 750 | P = 1500 | P = 2250 | P = 3000
---|---|---|---|---
ExLasso ($\gamma=4$) | 0.9 (0.0816) | 0.816 (0.0526) | 0.922 (0.0673) | 0.838 (0.0062)
ExLasso ($\gamma=6$) | 1 (0) | 0.947 (0) | 0.961 (0.0263) | 0.947 (0)
ExSCAD 4th | 0.9 (0.0816) | 0.816 (0.0526) | 0.922 (0.0673) | 0.838 (0.0062)
ExSCAD 6th | 1 (0) | 0.947 (0) | 0.961 (0.0263) | 0.947 (0)
ExMCP 4th | 0.82 (0.0688) | 0.782 (0.0708) | 0.765 (0.0679) | 0.683 (0.134)
ExMCP 6th | 0.946 (0.0454) | 0.854 (0.0619) | 0.961 (0.0263) | 0.885 (0.0876)
Lasso | 0 (0) | 0 (0) | 0 (0) | 0 (0)
SCAD | 0 (0) | 0 (0) | 0 (0) | 0 (0)
MCP | 0 (0) | 0 (0) | 0 (0) | 0 (0)
Median | 0.095 (0) | 0.095 (0) | 0.095 (0) | 0.095 (0)
Q-0.9 | 0.095 (0) | 0.095 (0) | 0.095 (0) | 0.095 (0)
Q-0.99 | 0.312 (0.1434) | 0.238 (0.0794) | 0.189 (0.0744) | 0.093 (0.0262)
Q-0.999 | 0.394 (0.1643) | 0.39 (0.0962) | 0.501 (0.0663) | 0.577 (0.2264)
Threshold | 0.384 (0.392) | 0.32 (0.1879) | 0.316 (0.087) | 0 (0)
Table 32: Average true positive rates for changing number of dimensions. | P = 750 | P = 1500 | P = 2250 | P = 3000
---|---|---|---|---
ExLasso ($\gamma=4$) | 0.9 (0.0816) | 0.861 (0.0556) | 0.947 (0.0611) | 0.802 (0.1235)
ExLasso ($\gamma=6$) | 1 (0) | 1 (0) | 1 (0) | 1 (0)
ExSCAD 4th | 0.9 (0.0816) | 0.861 (0.0556) | 0.947 (0.0611) | 0.802 (0.1235)
ExSCAD 6th | 1 (0) | 1 (0) | 1 (0) | 1 (0)
ExMCP 4th | 0.842 (0.0618) | 0.853 (0.0524) | 0.929 (0.0825) | 0.795 (0.1136)
ExMCP 6th | 1 (0) | 1 (0) | 1 (0) | 1 (0)
Lasso | 0 (0) | 0 (0) | 0 (0) | 0 (0)
SCAD | 0 (0) | 0 (0) | 0 (0) | 0 (0)
MCP | 0 (0) | 0 (0) | 0 (0) | 0 (0)
Median | 0.091 (0) | 0.091 (0) | 0.091 (0) | 0.091 (0)
Q-0.9 | 0.091 (0) | 0.091 (0) | 0.091 (0) | 0.091 (0)
Q-0.99 | 0.303 (0.1415) | 0.207 (0.0762) | 0.18 (0.0684) | 0.092 (0.0468)
Q-0.999 | 0.373 (0.1713) | 0.371 (0.1244) | 0.486 (0.0278) | 0.567 (0.2974)
Threshold | 0.353 (0.3505) | 0.344 (0.1929) | 0.335 (0.0958) | 0 (0)
Table 33: Average false positive rates for changing number of dimensions. | P = 750 | P = 1500 | P = 2250 | P = 3000
---|---|---|---|---
ExLasso ($\gamma=4$) | 0.001 (0.0011) | 0.002 (7e-04) | 0.001 (8e-04) | 0.003 (0.0029)
ExLasso ($\gamma=6$) | 0 (0) | 0 (0) | 0 (0) | 0 (0)
ExSCAD 4th | 0.001 (0.0011) | 0.002 (7e-04) | 0.001 (8e-04) | 0.003 (0.0029)
ExSCAD 6th | 0 (0) | 0 (0) | 0 (0) | 0 (0)
ExMCP 4th | 0.002 (8e-04) | 0.002 (7e-04) | 0.001 (8e-04) | 0.002 (0.001)
ExMCP 6th | 0 (0) | 0 (0) | 0 (0) | 0 (0)
Lasso | 0.014 (0) | 0.013 (8e-04) | 0.013 (8e-04) | 0.013 (0.001)
SCAD | 0.014 (0) | 0.013 (8e-04) | 0.013 (8e-04) | 0.013 (0.001)
MCP | 0.013 (0.0014) | 0.011 (0.0013) | 0.012 (0.0017) | 0.013 (0.001)
Median | 0.014 (0) | 0.014 (0) | 0.014 (0) | 0.014 (0)
Q-0.9 | 0.014 (0) | 0.017 (0.0013) | 0.015 (0.0026) | 0.014 (0.001)
Q-0.99 | 0.01 (0.0026) | 0.017 (0.0078) | 0.012 (0.0019) | 0.016 (0.0086)
Q-0.999 | 0.01 (0.0036) | 0.01 (0.0045) | 0.007 (0.0014) | 0.007 (0.0067)
Threshold | 0.008 (0.0034) | 0.007 (0.0017) | 0.008 (0.0016) | 0.012 (0.0019)
## Acknowledgements
The authors acknowledge support from NSF DMS-1554821 and NSF NeuroNex-1707400.
We also thank Dr. Michael Weylandt for providing useful discussions.
## References
* [1] [author] US Environmental Protection Agency. Air Quality System Data Mart [internet database] available via https://www.epa.gov/airdata. Accessed March 20, 2021.
* [2] [author] US Environmental Protection Agency. Air Quality System Data Mart [internet database] available via https://www.epa.gov/airdata. Accessed March 20, 2021.
* Bali [2003] [author] Bali, T. G.T. G. (2003). The generalized extreme value distribution. Economics letters 79(3) 423-427.
* Bangare et al. [2015] [author] Bangare, S. L.S. L., Dubal, A.A., Bangare, P. S.P. S. and Patil, S. T.S. T. (2015). Reviewing Otsu’s method for image thresholding. International Journal of Applied Engineering Research 10(9) 21777-21783.
* Belloni and Chernozhukov [2011] [author] Belloni, A.A. and Chernozhukov, V.V. (2011). $\ell$1-penalized quantile regression in high-dimensional sparse models. The Annals of Statistics 39(1) 82-130.
* Boucheron, Lugosi and Massart [2013] [author] Boucheron, StéphaneS., Lugosi, GáborG. and Massart, PascalP. (2013). Concentration inequalities: A nonasymptotic theory of independence. Oxford university press.
* Calabrese and Osmetti [2013] [author] Calabrese, R.R. and Osmetti, S. A.S. A. (2013). Modelling small and medium enterprise loan defaults as rare events: the generalized extreme value regression model. Journal of Applied Statistics 40(6) 1172-1188.
* Campbell and Allen [2017] [author] Campbell, F.F. and Allen, G. I.G. I. (2017). Within group variable selection through the exclusive lasso. Electronic Journal of Statistics 11(2) 4220-4257.
* Cherkassky and Ma [2004] [author] Cherkassky, V.V. and Ma, Y.Y. (2004). Practical selection of SVM parameters and noise estimation for SVM regression. Neural networks 17(1) 113-126.
* Friedman, Hastie and Tibshirani [2010] [author] Friedman, J.J., Hastie, T.T. and Tibshirani, R.R. (2010). A note on the group lasso and a sparse group lasso. arXiv preprint arXiv:1001.0736.
* Gelaro et al. [2017] [author] Gelaro, R.R. et al. (2017). The modern-era retrospective analysis for research and applications, version 2 (MERRA-2). Journal of climate 30(14) 5419-5454.
* Hampel [1968] [author] Hampel, Frank RudolfF. R. (1968). Contributions to the theory of robust estimation. University of California, Berkeley.
* Harter [1974-1975] [author] Harter, H. L.H. L. (1974-1975). The Method of Least Squares and Some Alternatives. ” Parts I–VI, International Statistical Review 42-43.
* Holland and Welsch [1977] [author] Holland, P. W.P. W. and Welsch, R. E.R. E. (1977). Robust regression using iteratively reweighted least-squares. Communications in Statistics-theory and Methods 6(9) 813-827.
* Huber [1973] [author] Huber, P. J.P. J. (1973). Robust regression: asymptotics, conjectures and Monte Carlo. The Annals of Statistics 1(5) 799-821.
* Koenker and Hallock [2001] [author] Koenker, R.R. and Hallock, K. F.K. F. (2001). Quantile regression. Journal of economic perspectives 15(4) 143-156.
* Kuchibhotla and Chakrabortty [2018] [author] Kuchibhotla, Arun KumarA. K. and Chakrabortty, AbhishekA. (2018). Moving beyond sub-gaussianity in high-dimensional statistics: Applications in covariance estimation and linear regression. arXiv preprint arXiv:1804.02605.
* Lai and Lee [2005] [author] Lai, P. Y.P. Y. and Lee, S. M. S.S. M. S. (2005). An overview of asymptotic properties of Lp regression under general classes of error distributions. Journal of the American Statistical Association 100(470) 446-458.
* Lee et al. [2015] [author] Lee, Jason DJ. D., Sun, YuekaiY., Taylor, Jonathan EJ. E. et al. (2015). On model selection consistency of regularized M-estimators. Electronic Journal of Statistics 9 608–642.
* Lein et al. [2007] [author] Lein, E. S.E. S. et al. (2007). Genome-wide atlas of gene expression in the adult mouse brain. Nature 445(7124) 168-176.
* Leiva et al. [2016] [author] Leiva, V.V., Ferreira, M.M., Gomes, M. I.M. I. and Lillo, C.C. (2016). Extreme value Birnbaum–Saunders regression models applied to environmental data. Stochastic Environmental Research and Risk Assessment 30(3) 1045-1058.
* Li and Zhu [2008] [author] Li, Y.Y. and Zhu, J.J. (2008). L1-norm quantile regression. Journal of Computational and Graphical Statistics 17(1) 163-185.
* Loh et al. [2017] [author] Loh, P. L.P. L. et al. (2017). Statistical consistency and asymptotic normality for high-dimensional robust M-estimators. The Annals of Statistics 45(2) 866-896.
* Loh and Wainwright [2013] [author] Loh, Po-LingP.-L. and Wainwright, Martin JM. J. (2013). Regularized M-estimators with nonconvexity: Statistical and algorithmic theory for local optima. In Advances in Neural Information Processing Systems 476–484.
* Michel and Chave [2007] [author] Michel, A. P.A. P. and Chave, A. D.A. D. (2007). Analysis of laser-induced breakdown spectroscopy spectra: the case for extreme value statistics. Spectrochimica Acta Part B: Atomic Spectroscopy 62(12) .1370-1378.
* Money et al. [1982] [author] Money, A. H.A. H., Affleck-Graves, J. F.J. F., Hart, M. L.M. L. and Barr, G. D. I.G. D. I. (1982). The linear regression model: Lp norm estimation and the choice of p. Communications in Statistics-Simulation and Computation 11(1) 89-109.
* Negahban et al. [2012] [author] Negahban, Sahand NS. N., Ravikumar, PradeepP., Wainwright, Martin JM. J., Yu, BinB. et al. (2012). A unified framework for high-dimensional analysis of $M$-estimators with decomposable regularizers. Statistical Science 27 538–557.
* Nyquist [1983] [author] Nyquist, H.H. (1983). The optimal Lp norm estimator in linear regression models. Communications in Statistics-Theory and Methods 12(21) 2511-2524.
* Parikh and Boyd [2014] [author] Parikh, N.N. and Boyd, S.S. (2014). Proximal algorithms. Foundations and Trends in Optimization 1(3) 127-239.
* Pauli and Coles [2001] [author] Pauli, F.F. and Coles, S.S. (2001). Penalized likelihood inference in extreme value analyses. Journal of Applied Statistics 28(5) 547-560.
* Rousseeuw and Leroy [2005] [author] Rousseeuw, P. J.P. J. and Leroy, A. M.A. M. (2005). Robust regression and outlier detection.
* Rousseeuw and Yohai [1984] [author] Rousseeuw, P.P. and Yohai, V.V. (1984). Robust regression by means of S-estimators. Robust and nonlinear time series analysis 256-272.
* Song, Chaudhuri and Sarwate [2015] [author] Song, S.S., Chaudhuri, K.K. and Sarwate, A.A. (2015). Learning from data with heterogeneous noise using sgd. Artificial Intelligence and Statistics 894-902.
* Sposito and Hand [1983] [author] Sposito, V. A.V. A. and Hand, M. L.M. L. (1983). On the Efficiency of Using the Sample Kurtosis in Selecting Optimal Lp Estimators. Communications in Statistics, Part B—Simulation and Computation 12 265–272.
* Subbotin [1923] [author] Subbotin, M. T.M. T. (1923). On the Law of Frequency of Errors. Matematicheskii Sbornik 31 296-301.
* Theis et al. [2016] [author] Theis, L.L., Berens, P.P., Froudarakis, E.E., Reimer, J.J., Rosón, M. R.M. R., Baden, T.T., Euler, T.T., Tolias, A. S.A. S. and Bethge, M.M. (2016). Benchmarking spike rate inference in population calcium imaging. Neuron 90(3) 471-482.
* Tibshirani et al. [2005] [author] Tibshirani, R.R., Saunders, M.M., Rosset, S.S., Zhu, J.J. and Knight, K.K. (2005). Sparsity and smoothness via the fused lasso. Journal of the Royal Statistical Society: Series B (Statistical Methodology) 67(1) 91-108.
* Vladimirova et al. [2019] [author] Vladimirova, MariiaM., Girard, StéphaneS., Nguyen, HienH. and Arbel, JulyanJ. (2019). Sub-Weibull distributions: generalizing sub-Gaussian and sub-Exponential properties to heavier-tailed distributions. arXiv preprint arXiv:1905.04955.
* Wang, Li and Tsai [2007] [author] Wang, H.H., Li, R.R. and Tsai, C. L.C. L. (2007). Tuning parameter selectors for the smoothly clipped absolute deviation method. Biometrika 94(3) 553-568.
* Wang et al. [2013] [author] Wang, XueqinX., Jiang, YunluY., Huang, MianM. and Zhang, HepingH. (2013). Robust variable selection with exponential squared loss. Journal of the American Statistical Association 108 632–643.
* Wu et al. [2009] [author] Wu, Z.Z., Jiao, J.J., Yang, M.M., Liu, Y.Y. and Wang, Z.Z. (2009). An enhanced adaptive CUSUM control chart. IIE transactions 41(7) 642-653.
* Yang, Tewari and Ravikumar [2013] [author] Yang, E.E., Tewari, A.A. and Ravikumar, P.P. (2013). On robust estimation of high dimensional generalized linear models. Twenty-Third International Joint Conference on Artificial Intelligence.
* Yatsenko et al. [2015] [author] Yatsenko, D.D., Josić, K.K., Ecker, A. S.A. S., Froudarakis, E.E., Cotton, R. J.R. J. and Tolias, A. S.A. S. (2015). Improved estimation and interpretation of correlations in neural circuits. PLoS computational biology 11(3) e1004083.
* Zhang [2010] [author] Zhang, C. H.C. H. (2010). Nearly unbiased variable selection under minimax concave penalty. The Annals of statistics 38(2) 894-942.
|
# Magnetic Field at the Galactic Centre from Multi-Wavelength Dust
Polarization
M. S. Akshaya1 and Thiem Hoang1,2
1Korea Astronomy and Space Science Institute, Daejeon 34055, Republic of Korea
2Department of Astronomy and Space Science, University of Science and
Technology, 217 Gajeong-ro, Yuseong-gu, Daejeon 34113, Republic of Korea
E-mail<EMAIL_ADDRESS>(MSA)
(Accepted XXX. Received YYY; in original form ZZZ)
###### Abstract
We have mapped the magnetic field ($B$-field) for a region of about 30 pc
around the centre of our Galaxy, which encompasses the circumnuclear disk
(CND), the mini-spirals, and the 20 km s-1 and 50 km s-1 molecular clouds,
using thermal dust polarization observations obtained from SOFIA/HAWC+ and
JCMT/SCUPOL. We decompose the spectra of 12CO ($J=3\rightarrow 2$) transition
from this region into individual clustered cloud components and find the
polarization observed at different wavelengths might be tracing completely
different layers of dust along the line-of-sight (LOS). We use modified Davis-
Chandrasekhar-Fermi methods to estimate the $B$-field from the observations.
From our analysis we find the mean strength of the plane-of-sky $B$-field
($B_{{}_{\mathrm{POS}}}$) of the CND and the mini-spirals, probed at 53 µm to
be of the order of $\sim 2$ mG. The magnetic field is lowest close to the
Galactic Centre, in the region of the ionized mini-cavity within the CND with
$B_{{}_{\mathrm{POS}}}<1$ mG, and increases outwards. However, the longer
wavelength polarization at 216 µm appears to come from a dust layer that is
cooler and behind the CND and has a stronger $B$-field of about 6 mG. The
$B$-field has the least strength along the Eastern Arm of the mini-spiral,
which is also the only region with $\mathcal{M}_{\mathrm{A}}>1$ and a mass-to-
flux ratio of $\lambda\gtrsim 1$. The similarity between the
$B_{{}_{\mathrm{POS}}}$ estimates of the 53 µm and 850 µm observations might
indicate them originating from the same depth along the LOS, mostly from the
CND and its foreground cloud features, including the negative-longitude
extension.
###### keywords:
dust, extinction – Galaxy: centre – infrared: ISM – ISM: magnetic fields –
ISM: general – polarization
††pubyear: 2024††pagerange: Magnetic Field at the Galactic Centre from Multi-
Wavelength Dust Polarization–Magnetic Field at the Galactic Centre from Multi-
Wavelength Dust Polarization
## 1 Introduction
The centre of galaxies play a vital role in their evolution, right from star
formation to quenching. A significant fraction of the galactic star formation
might be driven by the inflow of material into the galactic centre, whereas
the massive outflows from these starburst activities as well as from the
supermassive blackhole at the galactic centre (like the Sgr A∗ in case of the
Milky Way) can trigger galactic quenching and hence its evolution (Oort, 1977;
Kormendy & Kennicutt, 2004; Veilleux et al., 2020, and references therenin).
Only in our Galaxy do we have the capability to study this complex region in
great detail at high resolution using various observational techniques like
imaging, spectroscopy, and polarimetry. It is well known that the star
formation rate (SFR) of the Milky Way is far below what is expected from
gravitational collapse of molecular clouds, and the SRF in Galactic Centre
(GC) is even lower than the average SFR in the Milky Way, yet the origin for
the low SRF in the GC remains elusive (Genzel et al., 2010; Barnes et al.,
2017; Bryant & Krabbe, 2021; Henshaw et al., 2022, and references therein).
Magnetic field ($B$-field) is one of the popular candidates that can play a
role in this suppressed star formation. In order to understand the role of
$B$-fields in star formation and evolution of the GC, it is crucial to map its
strength and morphology.
Significant efforts have been made to map the magnetic field of the GC in the
past, with most of these studies focused on the circum-nuclear disk which is
the closest molecular reservoir to Sgr A∗ (CND; Becklin et al., 1982; Guesten
et al., 1987; Jackson et al., 1993). The line-of-sight (LOS) strength of the
$B$-field ($B_{{}_{\mathrm{LOS}}}$) was estimated from the observations of
Zeeman splitting in spectral lines. The average $B_{{}_{\mathrm{LOS}}}$ from
the Zeeman measurements at various locations along the CND was about 3 mG
(Schwarz & Lasenby, 1990; Killeen et al., 1992; Plante et al., 1995; Marshall
et al., 1995; Yusef-Zadeh et al., 1996, 1999). Any observed variation between
the measurements was attributed to possible changes in the magnetic field
orientation.
Aitken et al. (1986) and Aitken et al. (1998) predicted the upper and lower
limits of the magnetic field in the region based on the mid-infrared thermal
dust polarization observations. They assumed paramagnetic relaxation as the
mechanism of grain alignment, and estimate the field to be between 2 – 10 mG
around the GC. However, this assumption no longer holds as recent studies show
that paramagnetic relaxation by itself is not strong enough to drive grain
alignment (Hoang & Lazarian, 2016). Regardless, polarized thermal emission
from aligned dust grains is still a popular tool to map the plane-of-sky (POS)
magnetic field ($B_{{}_{\mathrm{POS}}}$). This technique is based on the fact
that non-spherical dust grains tend to align with their longest axis
perpendicular to the $B$-fields (Lazarian, 2007; Lazarian & Hoang, 2007), so
that the polarization of thermal emission is perpendicular to the $B$-fields
(Hildebrand, 1988). As the result, by observing the thermal dust polarization
and rotating polarization vectors by 90°, we can infer the $B$-field
morphology. The GC makes an ideal target for mapping the magnetic fields using
dust polarization because the dust grains are expected to be efficiently
aligned with the $B$-fields in this environment. Indeed, if the dust grains
are not aligned with the magnetic field, then the dust polarization does not
trace the $B$-field morphology. However, from our previous study (Akshaya &
Hoang, 2023), we have found that, due to the high $B$-field strength observed
in the region by Zeeman measurements, grains could achieve perfect alignment
through Magnetically-Enhanced Radiative Torque Alignment mechanism (MRAT;
Hoang & Lazarian, 2016; Hoang, 2022). Therefore, dust polarization can be a
robust tracer of the $B$-fields in the GC.
The GC is a complex and dynamic environment known to have densities, pressure,
and $B$-field orders of magnitude greater than that observed in the diffuse
interstellar medium (ISM). Molecular spectra of the region also indicate the
presence of complex multi-component structures along the LOS (Sutton et al.,
1990; Henshaw et al., 2016; Eden et al., 2020; Hu et al., 2022). If these dust
components present at various distances along the LOS are emitting radiation
at different wavebands (due to difference in their temperatures), we can use
the different polarization morphology observed at multiple wavelengths to get
an understanding of the change in the $B$-field strength, morphology, and
orientation along the LOS.
The goal of this paper is to measure the strength of the $B$-field within a
region of about 30 pc around the supermassive blackhole Sgr A∗ at the centre
of our Galaxy. We will use polarization observations in three wavebands, the
same which were used in our previous study (Akshaya & Hoang, 2023) at 53, 216,
and 850 µm. The observations focus on the CND, which is a warm torus of gas
and dust orbiting around Sgr A∗, and the mini-spirals which are a set of
ionized gas filaments present within and interacting with the CND (Lo &
Claussen, 1983; Christopher et al., 2005). Each observations probes the region
on different scales allowing us to map the small and large scale magnetic
field in the region. The region of the CND is covered in all the three
observations while part of the 20 km s-1 and 50 km s-1 clouds (Kauffmann et
al., 2017) are seen in the 216 µm observation. The 850 µm data covers the
largest area around Sgr A∗ which includes the CND, mini-spirals, 20 km s-1,
and 50 km s-1 clouds. Recent study of the CND using Stratospheric Observatory
for Infrared Astronomy (SOFIA; Temi et al., 2018) observation by Guerra et al.
(2023) estimate the $B$-field strength of the mini-spirals within the CND to
have median values between 5 – 8 mG on a spacial scale of $\lesssim 1$ pc.
The most commonly used technique for the measurement of the POS magnetic field
strength from thermal dust polarization is the Davis-Chandrasekhar-Fermi (DCF)
method (Davis, 1951; Chandrasekhar & Fermi, 1953). It is based on the
propagation of Alfvén waves through the medium, when there is an energy
balance between the gas kinetic energy and magnetic energy. Assuming the
magnetic field lines to be frozen with the matter in the general ISM
conditions, the method attributes any small scale irregularities observed in
the polarization vectors to turbulence. The net magnetic field in the region
can be assumed to be made up of a regular component ($B_{0}$) and a turbulent
component ($\delta B$). A strong $B_{0}$ resists being perturbed by turbulence
such that $\delta B\ll B_{0}$, thus allowing us to characterise the POS
component of $B_{0}$ ($B_{{}_{\mathrm{POS}}}$) by measuring the net
irregularity of the field lines using the relation,
$B_{{}_{\mathrm{POS}}}=\sqrt{4\pi\rho}\frac{\sigma_{v}}{\sigma_{\phi}},$ (1)
where $\rho$ is the gas mass density, $\sigma_{v}$ is the turbulence-induced
velocity dispersion from non-thermal line-width measurements, and
$\sigma_{\phi}$ is the distortion in the $B$-field measured by the
polarization angle dispersion. Even though the method is known to overestimate
the magnetic field due to its restrictive initial assumptions about the
conditions of the underlying medium (Ostriker et al., 2001; Houde et al.,
2009; Chen et al., 2022; Myers et al., 2023, and references therein), it is a
useful technique to get an upper estimate of the magnetic field in simple
environments without self gravity or sheer motion.
The DCF method has been applied successfully by many earlier polarization
studies (Pillai et al., 2015; Planck Collaboration et al., 2016; Pattle et
al., 2017; Guerra et al., 2021; Ngoc et al., 2021; Hwang et al., 2021; Hoang
et al., 2022b). Some of the physical conditions that contribute to the over-
estimate of the mean $B_{{}_{\mathrm{POS}}}$ while using DCF include;
anisotropic turbulence, failure of equipartition between the magnetic and
kinetic energy, and self gravity. It is difficult to disentangle the
importance of each of them without a detailed understanding of the kinematics
of the region. The overestimation could also be a result of integration
effects from within the beam size of individual observations, as well as along
the LOS. These uncertainties arise from the polarization observation and
impact the parameter $\sigma_{\phi}$ in Equation 1 (Hildebrand et al., 2009;
Houde et al., 2009; Skalidis & Tassis, 2021; Li et al., 2022; Guerra et al.,
2023). However, Chen et al. (2022) found that the hydrodynamic properties of
gas in the region contributes equally if not more to the uncertainty of the
measured field (reflected in the parameters $\rho$ and $\sigma_{v}$).
Several attempts have been made to improve the original DCF technique. The DCF
method is a good approximation when the observed angle dispersion is small
i.e. when $\delta B\ll B_{0}$ (Ostriker et al., 2001). Falceta-Gonçalves et
al. (2008) extended the DCF method to cases where the turbulent component of
the magnetic field is comparable to the mean field, thus resulting in large
angle dispersion. Hildebrand et al. (2009) and Houde et al. (2009) further
improved the DCF estimate by using the turbulent-to-ordered magnetic field
ratio to estimate the angel dispersion, based on the second order structure
function of magnetic field position angles introduced by Falceta-Gonçalves et
al. (2008). This method incorporated the large scale structure of the magnetic
field along with the instrumental effects in the observations. The recent
modifications include those by Cho & Yoo (2016) and Lazarian et al. (2022)
where they address the anisotropic nature of MHD turbulence, by incorporating
structure function in combination with velocity centroids to estimate the
velocity fluctuations in the POS. We will apply the DCF modifications from
Houde et al. (2009) and Lazarian et al. (2022) in our study, with the goal to
understand how multi-wavelength polarization can be used to probe the 3D
strength of the $B$-field. No correction factors are used in our estimates and
they can be considered as an upper limit of the $B$-field in the region
covered by each observation.
Figure 1: The maps of polarization from SOFIA/HAWC+ observations at 53 µm
(top), 216 µm (bottom left), and JCMT/SCUPOL at 850 µm (bottom right). The
colorbars represent the intensity in respective wavebands and the beam size of
each instrument is shown on the bottom right of each figure along with the
representative scale of polarization percentage at the top.
In this work, we will measure the $B_{{}_{\mathrm{POS}}}$ from three
observations of the region around the GC spanning a physical scale of 10 – 30
pc in wavebands centered at 53, 216, and 850 µm. Our initial study of the
thermal dust polarization from these observations is discussed in Akshaya &
Hoang (2023), where we focus on the grain alignment physics. We have used the
same observations in this work to understand how the strength of the magnetic
field varies across different scales and wavelengths along the LOS, with
longer wavelengths probing deeper into the LOS compared to shorter
wavelengths. The region of the Galactic disk in general, and GC in particular
are known to have multiple structures at different distances along the LOS. We
will use recent techniques to isolate the individual components and understand
their effects on the derived magnetic field.
The rest of our paper is structured as follows; Section 2 describes the
polarization observations and their data quality assessment. The estimation of
the gas velocity dispersion is discussed in Section 3. The magnetic field
measured from the DCF modifications by Houde et al. (2009) and Lazarian et al.
(2022) is described in Section 4. A detailed discussion of the estimated
$B$-field and its implications on the kinematics of the regions are presented
in Section 5, followed by a brief summary of our results in Section 6.
Figure 2: The top map shows the HAWC+ polarization observation of the CND at
216 µm with the criteria $\Delta\phi_{\mathrm{ref}}<10\degree$, compared with
the 250 µm polarization from PILOT (red vectors). The background intensity map
is the Stokes I measurements from HAWC+. The PILOT polarization vectors are
scaled to $p=5\%$. The map at the bottom compares the HAWC+ observations of
the same region taken using the CNM (red) and OTFMAP (blue) mapping strategies
respectively. The background intensity in this case is from the Stokes I
measurements of the OTFMAP. The scale of polarization percentage is shown on
each map along with the beam size of the HAWC+ instrument.
## 2 Dust Polarization Observations
We have used the thermal dust polarization observations around the GC at 53
and 216 µm from High-resolution Airborne Wide-band Camera Plus (HAWC+; Harper
et al., 2018), which is a far-infrared imager and polarimeter for NASA’s
Stratospheric Observatory for Infrared Astronomy (SOFIA; Temi et al., 2018)
and the reprocessed data at 850 µm from SCUPOL, which was the polarimeter for
the Submillimeter Common User Bolometer Array (SCUBA) instrument on the James
Clerk Maxwell Telescope (JCMT) presented by Matthews et al. (2009). More
details about the data and the polarization cut off criteria used are
presented in Akshaya & Hoang (2023), where the grain alignment of the region
was studied with the same observations. The SOFIA/HAWC+ observations were
reduced using the latest version of the HAWC+ data reduction pipeline i.e. DRP
v3.2.0 compared to the DRP v1.3.0 used in our previous analysis. The quality
of these datasets are nominal with all the problematic files removed during
the reduction procedure.
We have also used the data from Herschel in five wavebands at 70, 160, 250,
350, and 500 µm to derive the gas column density ($N_{{}_{\mathrm{H}}}$) and
the dust temperature ($T_{\mathrm{d}}$) as described in Akshaya & Hoang
(2023). The polarization maps of each observation is shown in Fig. 1. The
magnetic field strength will be estimated for each of the polarization
observations, which we believe to be probing different components of dust
along the LOS. Throughout this work we assume the distance to the GC to be 8
kpc (Trippe et al., 2008; Genzel et al., 2010).
### 2.1 Data Quality Assessment
The observed polarization from the CND and its surrounding can be subject to
reference beam contamination due to the emission from extended features in
this region. This is more of a problem in the 216 µm observation due to the
temperature of the surrounding material. At 53 µm the dust being probed is
much hotter than its surrounding and hence the contribution from the reference
beams is not as significant, allowing us to use the Level 4 data products as
it is at this wavelength. We have estimated the level of reference beam
contamination in the SOFIA/HAWC+ 216 µm observation using the method described
by Novak et al. (1997) and Chuss et al. (2019). The chop angle and amplitude
of the observation are 60$\degree$ and 240′′, respectively with two chop
images taken symmetrically on either side of the source. Herschel 70, 160,
250, and 350 µm observations of the region were used to model the expected
intensity of the chop beams in the HAWC+ filter band-pass using the relation;
$I=A\nu^{2}B_{\nu}(T_{\mathrm{d}}),$ (2)
where $A$ is the amplitude, $\nu$ the frequency, and $B_{v}(T_{\mathrm{d}})$
is the Planck function corresponding to the dust temperature $T_{\mathrm{d}}$.
The mean intensity of the chop beams was used in combination with the
calibrated Stokes I intensity from HAWC+ to determine the contrast of the
source beam with respect to the chop beams. Due to the brightness of the GC,
the ratio of these intensities (represented by $w\equiv I_{r}/I_{m}$, where
$I_{r}$ is the average intensity from the chop beams and $I_{m}$ is the
measured Stokes I intensity of the polarization observation) was found to be
$w<6$ throughout the region. The established method is to only consider
polarization vectors where the contrast is greater than 10 (Santos et al.,
2019). However, this seems unlikely in the Galactic disk at these temperatures
due to the presence of extended emissions along most of the lines of sights.
The other approach is the quantify the level of contamination in the measured
$p$ and $\phi$. Since our goal is to measure the strength of the $B$-field, we
are only interested in the level of contamination in the polarization angle,
which is a key parameter to estimate the $B_{{}_{\mathrm{POS}}}$.
Novak et al. (1997) derived the maximum error in the measured polarization
angle with only the intensity estimates from the reference beam to be,
$\Delta\phi_{\mathrm{ref}}=\frac{1}{2}\mathrm{tan}^{-1}\left[\frac{p_{r}w}{(p_{m}^{2}-p_{r}^{2}w^{2})^{1/2}}\right],$
(3)
where $p_{m}$ is the measured polarization fraction (without debias) and
$p_{r}$ is the assumed polarization of the reference beam.
From Fig. 1 it can be seen that except the top right region of the 216 µm
polarization map, most of the map has a polarization of roughly $p\sim 1\%$.
This is also the level of the observed polarization in the regions of the chop
beams from $Planck$ observations at 350 µm (Planck Collaboration et al.,
2020a, b). Thus the reference beam polarization was set to $p_{r}=1\%$ based
on these observations to estimate the $\Delta\phi_{\mathrm{ref}}$. Following
the cut off criteria defined in Chuss et al. (2019), we use only the
polarization vectors with $\Delta\phi_{\mathrm{ref}}<10\degree$ and the
resulting polarization map is shown in Fig. 2 (top). Earlier estimates of the
contamination from reference beam use $p_{r}=10\%$ (Chuss et al., 2019; Lee et
al., 2021). However this values seems too high for our observation and results
in the rejection of all the polarization vectors.
Due to the low contrast between the source and reference beams in this region,
we also tested the agreement between the polarization angles observed by HAWC+
and those from PILOT balloon experiment presented by Mangilli et al. (2019),
which measured the polarization from an extended region of the GC at 240 µm.
The cut off based on $\Delta\phi_{\mathrm{ref}}$ removed the vectors from low
intense regions, predominantly in the top right as seen in Fig. 1. From our
previous analytical study presented in Akshaya & Hoang (2023), this was the
region where the environmental conditions favour efficient grain alignment,
and hence where we can expect the highest degree of polarization. The other
removed vectors are from the regions where the observed polarization was very
low ($p<0.1\%$). These seem to be the most affected by the reference beam
contamination. The polarization vectors which did not make it with the cut off
set by $\Delta\phi_{\mathrm{ref}}$ also do not match well with the
polarization vectors from PILOT observations, especially in the regions of low
polarization fraction. We consider the remaining vectors good enough for
further analysis and use them as is for the subsequent discussions presented
in this paper.
We have also compared the HAWC+ 216 µm observation after the quality cuts
mentioned above with the observations from the same instrument but taken with
a different observing strategy as part of the Far-Infrared Polarimetric Large
Area CMZ Exploration survey (FIREPLACE; Butterfield et al., 2023, 2024; Paré
et al., 2024). FIREPLACE survey used the on-the-fly mapping mode (OTFMAP) to
measure the polarization of the Central Molecular Zone (CMZ) to overcome some
of the challenges in observing extended emission features posed by the
standard observing strategy of the HAWC+ instrument which was Chop-Nod-Match
(CNM; Harper et al., 2018). The main difference between the two strategies
lies in the background subtraction procedure. While the CNM observations are
limited by background chop images taken close to and symmetrically on either
side of the source, the OTFMAP can choose a background region completely
devoid of any structures associated with the source, as the scans can begin at
a reasonable distance from the target location. This can be very useful when
dealing with extended emission features like in the GC as the chop images
necessary for the background subtraction tend to have some remnant features of
the extended object, resulting in significant reference beam contamination.
The polarization vectors from both these observation strategies is shown in
the bottom map of Fig. 2, where the OTFMAP map follow the standard cuts to the
data with polarization signal-to-noise $p/\sigma_{p}>3$, polarization degree
$p<50\%$, and intensity signal-to-noise (Stokes I) $I/\sigma_{I}>200$. The
main problematic region is the low intense part in the right of the figure
where both the OTFMAP and the CNM observations do not retain any polarization
vectors. The other regions where we have measurements from both strategies
seem to agree well except in the few parts where the CNM observation show low
polarization fraction. Considering the overall agreement in the cutoff
regions, we will use the CNM observation with the
$\Delta\phi_{\mathrm{ref}}<10\degree$ criteria for further analysis from hence
forth.
The 850 µm observation from SCUPOL used in this work is a mosaic created from
69 original observations from SCUPOL archive (Matthews et al., 2009). An
estimation of possible contamination in this data is beyond the scope of the
current paper. We will use this observation as is for our analysis.
Figure 3: Moment zero map of the CO ($J=3\rightarrow 2$) transition used for
our analysis taken from CHIMPS2. The data was binned to match the resolution
of the SOFIA/HAWC+ 216 µm observation.
Figure 4: Sample spectrum of the multiple components observed along the GC is
shown in the top figure along with the respective Gaussian fits to each
identified component in the bottom figure. The blue lines are the decomposed
Gaussian components from scousepy. The dotted red line indicated the best fit
spectrum from the combination of the decomposed components.
## 3 Gas Velocity Dispersion
Carbon monoxide (CO) is the second most abundant molecule in the interstellar
medium (ISM) after H2 and is ubiquitous in the Galactic plane, making it an
ideal tracer of the velocity structures and morphology of gas along this LOS.
Its rotational transitions can be observed in the submillimeter wavelengths,
with different transitions probing different densities. ${}^{12}{\rm
C}^{16}{\rm O}$ (commonly denoted as CO) is the most common isotope of the CO
and can be used to trace gas densities of the order of
$n_{{}_{\mathrm{H}}}\sim 10^{3}-10^{4}$ cm-3 depending on the transition.
Other isotopes like ${}^{13}{\rm CO}$ and ${\rm C}^{18}{\rm O}$ trace denser
regions but are relatively less abundant than CO. The CO ($J=3\rightarrow 2$)
transition is optically thin compared to its other rotational transitions
($J=1\rightarrow 0$ and $J=2\rightarrow 1$) and also has a higher resolution
due to its high frequency ($\nu=345.8$GHz). Thus CO ($J=3\rightarrow 2$) can
trace warmer and denser environments compared to its lower energy
counterparts. Due to the nature of CO ($J=3\rightarrow 2$) transition, the
observations of this line along the Galactic plane often shows multiple
complex emission features corresponding to various structures at different
distances along the LOS. This can be a drawback in cases where the focus is on
a single filament or star forming region and it might be ideal to use higher
density tracers like ${}^{13}{\rm CO}$ or ${\rm C}^{18}{\rm O}$ for these
cases. For our study we chose to use CO ($J=3\rightarrow 2$) transition
because we are interested in seeing all the components along this LOS which
might be contributing to the observed polarization.
Figure 5: Velocity components observed for the region covered by SOFIA/HAWC+
216 µm observation decomposed using scousepy and grouped using acorns. The
centroid velocity of the velocity components is along the z-axis and the data
points are scaled according to their corresponding velocity dispersion. Each
colour represents individual trees resulting from the clustering analysis.
Only the trees with number of leaves greater than four are displayed here.
We have used CO Heterodyne Inner Milky Way Plane Survey 2 (CHIMPS2) data for
the analysis and estimation of the velocity dispersion along the LOS to the
GC. The aim of the CHIMPS2 survey is to map the Inner Galaxy, the CMZ, and a
section of the Outer Galaxy in the 12CO, 13CO, and C18O ($J=3\rightarrow 2$)
emissions using the Heterodyne Array Receiver Program (HARP) on the JCMT. We
have used the first look data of the CHIMPS2 survey towards the CMZ in CO
transition $J=3\rightarrow 2$ presented by Eden et al. (2020), with a spatial
resolution of 15 arcsec and a velocity resolution of 1 km s-1. We re-binned
the spectral cube to match the resolution of SOFIA/HAWC+ 216 µm and the
JCMT/SCUPOL 850 µm observations. The 53 µm observation from HAWC+ is at a much
higher resolution than the CO data, hence we have used the derived velocity
dispersion values at the resolution of the 216 µm observation as an
approximation for its $B$-filed calculation. The moment zero map of the data
is shown in Fig. 3. From the figure it can be seen that the morphology of the
prominent features observed in the polarization maps are also observed in the
integrated intensity map of the CO spectra. This in general is a good
indicator that the chosen spectra can trace the distribution of matter
corresponding to the observed polarization along the LOS and is the basic test
before choosing any molecular species to measure the turbulence needed for the
DCF technique. The complication arises when there are more than one clearly
distinguishable Gaussian components in the spectra for most of the region,
like in the case of the GC.
This complexity of the GC environment makes it a challenging region to measure
the magnetic field using the DCF method. Recent investigation of the method by
Chen et al. (2022) show that the line-width measurement which is used to
constrain the turbulence along the LOS can be the major source of uncertainty
or overestimation of the $B$-field when using the DCF method. The measured
line-width can be a fairly reliable tracer of turbulence in regions where we
can expect a single structure along the LOS, as in the case of filaments
mostly at high Galactic latitudes. But when we look at the spectrum closer to
the Galactic disk, it becomes evident that this approach becomes insufficient.
A sample spectrum of the GC is shown in Fig. 4. The spectrum indicates matter
distributed in multiple layers along the LOS. Another factor to consider when
using these data to measure the line-width is which component/components along
the LOS contribute to the observed polarization.
### 3.1 Spectral decomposition using scousepy
The decomposition of the CO spectrum was done using the Python implementation
of the Semi-Automated multi-COmponent Universal Spectral-line fitting Engine
(scousepy; Henshaw et al., 2016,
2019)111https://github.com/jdhenshaw/scousepy. This is a routine developed in
Interactive Data Language (IDL) and later implemented in Python and is very
useful to decompose complex 3D spectral cube in a systematic and efficient
way. It is a multi-stage procedure and the important steps can be broken down
as follows: (i) specify the region of interest based on position, velocity, or
noise threshold; (ii) the routine then breaks the region into Spectral
Averaging Areas (SAAs) based on the complexity of the spectrum in the region
and extracts the spatially averaged spectrum from each; (iii) the extracted
spectrum is manually fit interactively using
pyspeckit222https://github.com/pyspeckit/pyspeckit; (iv) the best fit solution
of the extracted spectrum from each SAA is used to fit all the individual
spectra within the SAA using a fully automated fitting procedure. The
tolerance levels used in each step is described in Henshaw et al. (2016,
2019).
In our implementation of scousepy on the data re-binned to the HAWC+ 216 µm
resolution, we masked all the pixels with integrated signal below 1.5 K and
used the width of the SAAs to be 10 pixels. We obtained 258 SAAs which were
manually fit to automate the fitting for the 6362 individual spectra of the
region. Similar procedure was performed to obtain the decomposed velocity
components for the SCUPOL 850 µm observation. At the end of the routine we
obtain the number of velocity components in each pixel, their amplitude,
shift, width, and a few additional useful statistics. A sample of the spectra
decomposed using scousepy is shown in the bottom of Fig. 4. We reject the
components with signal-to-noise less than five and use only the robust
resolved components for further analysis.
Figure 6: Moment zero maps of the sub-slabs of PPV cube based on the
clustering shown in Fig. 5. The spectral cube was re-binned to match the
SOFIA/HAWC+ 216 µm observation. The Stokes I maps of HAWC+ 53 µm and SCUPOL
850 µm observation is also shown to compare the morphology of individual slabs
with the morphology of intensity (Stokes I) observed in the polarization maps.
Figure 7: Same as Fig. 6 but zoomed-in to focus on the region covered by the
HAWC+ 53 µm observation, which mostly contains the CND and the mini-spirals.
### 3.2 Hierarchical Agglomerative Clustering using acorns
In order to characterise the decomposed velocity components from scousepy and
understand its effects on the observed polarization, we have used a recently
developed analysis tool based on hierarchical agglomerative clustering called
acorns (Agglomerative Clustering for ORganising Nested Structures; Henshaw et
al., 2019)333https://github.com/jdhenshaw/acorns. The clustering procedure
here begins with the most significant data point and a hierarchy is
established by the merging of clusters based on user defined criteria. For a
PPV cube like our data set, the clustering is performed using the position
(x,y), intensity, and velocity. The criteria we used for the clustering
analysis are as follows: (i) minimum amplitude or peak intensity for the
components to be considered was five times the typical root-mean-square (rms)
value (Henshaw et al., 2019) (ii) minimum size of clusters had to be 30 arcsec
($\sim 1.5$ times the observation beam size) (iii) the difference in velocity
between linked data points cannot be greater than 15 km s-1. For visualization
purpose, we chose trees from the forest which has a minimum of 4 leaves from
the clustering procedure. This resulted in 8 prominent trees and are shown in
Fig. 5. From the figure it can be seen that though there seems to be a great
number of velocity components towards the GC, they can still be grouped into
meaningful distinct sub-structures.
The structures with negative centroid velocities centred at about -60, -40,
and -20 km s-1 are quite distinct and appear to be isolated structures without
much overlap. These components are clearly seen in most of the spectra as
shown in Fig. 4. The positive velocity region however shows significant
overlap in the observed features except for a distinction at around 30 km s-1.
This can also be seen in Fig. 4, where the peaks in the positive velocity
region are more blended together than the negative velocity peaks which are
distinct. Based on these observed velocity features, we divided the initial
spectral cube into sub-slabs with velocity ranges corresponding to the
distinct trees from acorns. The moment zero maps of these slabs are shown in
Fig. 6. We compared the morphology of the velocity features observed in the
moment zero maps of each sub-slab with the intensity maps of the polarization
observations in all the three wavebands considered. The zoom-in version of the
map for the region covered by the 53 µm observation is shown in Fig. 7.
The key take away from this analysis is the difference in morphologies of the
decomposed velocity structures and how they match the polarization
observations at different wavelengths. Probing a region at different
frequencies results in the observation of dust or material at different
temperatures along the LOS, with shorter wavelengths originating from hot
regions with high optical depth and longer wavelength optically thin emission
originating from cooler deeper dust. When we are looking at regions which are
known to have various independent structures along the LOS, we need to be
careful with our measurement of line width to characterise turbulence as the
chosen gaussian component might be probing a layer quite different from the
source of our measured polarization. This effect is demonstrated in Fig. 6 and
Fig. 7.
When we look at the morphology of the PPV sub-slabs and compare them with the
Stokes I maps of HAWC+ 53 and 216 µm observations, the 216 µm observed
intensity follows closely the morphology of the slab with the velocity range
$0<v<45$ km s-1. Though this velocity sub-slab seems to be the major
contributor, there are still traces of features from other slabs in the
intensity map of 216 µm observation. This could indicate that the net observed
emission at this wavelength might be an integrated effect from all the layers
along the line of sight. However, this does not seem to be the case for the 53
µm observation. When we compare its morphology with those of the moment maps
from individual slaps, it becomes clear that most if not all of the observed
emission at this wavelength is originating from the negative velocity
components of our decomposed spectra, and is clearly shown in Fig. 7. The 53
µm observation is dominated by the emission from the CND and the mini-spirals.
There does not seem to be any trace of the morphology observed from the
positive velocity components in the Stokes I map of the 53 µm observation.
This is very important when we are using the spectra to constrain the
turbulence in the region to measure the $B$-field. From this analysis, we
propose that the right estimate of turbulence for the CND and its streamers
comes from the negative velocity components of the CO spectra and use only the
velocity peaks in this range for the measurement of velocity dispersion for
the 53 µm observation. It is also interesting to note that the three negative
velocity components shown in Fig. 7 closely resemble the model for the
velocity of the streamers in Sgr A∗ by Zhao et al. (2009, Fig. 21).
Figure 8: Normalized auto-correlation function calculated from polarized
intensity for each of the observation. The value of $\Delta^{\prime}$ is
determined at half magnitude and corresponds to $0.47^{\prime},0.81^{\prime},$
and $0.78^{\prime}$ at at 53, 216, and 850 µm respectively.
## 4 Measurements of the plane of sky B-field
We now estimate the $B$-field strength using two modified DCF methods; (1)
based on the angle dispersion structure function by Houde et al. (2009) and
(2) the recently proposed Differential Measure Analysis technique by Lazarian
et al. (2022).
### 4.1 $\mathbf{B_{{}_{\mathrm{POS}}}}$ from Angle Dispersion Function
Houde et al. (2009) improved the original DCF technique by using structure
function of polarization angles to calculate the polarization angle dispersion
such that the dispersion due to the change in field orientation is also
accounted for along with the dispersion due to turbulence. Following Houde et
al. (2009), Equation (1) can be rewritten as;
$B_{{}_{\mathrm{POS}}}=\sqrt{4\pi\rho}\sigma_{v}\left[\frac{\langle
B_{t}^{2}\rangle}{\langle B_{0}^{2}\rangle}\right]^{-1/2},$ (4)
where $\langle B_{t}^{2}\rangle/\langle B_{0}^{2}\rangle$ is the turbulent to
large-scale magnetic field strength ratio and can be approximated as the
dispersion in the polarization vectors ($\sigma_{\phi}=\sqrt{\langle
B_{t}^{2}\rangle/\langle B_{0}^{2}\rangle}$). This quantity can be determined
using the two-point dispersion function proposed by Houde et al. (2009),
modelled as a superposition of the large-scale field structure and the small-
scale turbulence. This method takes care of the spatial changes in the
$B$-field morphology and also incorporates the signal integration within the
cloud thickness and the telescope beam. If $\Delta\phi(\ell)$ is the
difference is polarization angle between two vectors separated by an angle
$\ell$ on the sky, then the dispersion function is defined as;
$1-\langle\mathrm{cos}\left[\Delta\phi(\ell)\right]\rangle=\frac{1-e^{-\ell^{2}/2(\delta^{2}+2W^{2})}}{1+\mathcal{N}\left[\frac{\langle
B_{t}^{2}\rangle}{\langle B_{0}^{2}\rangle}\right]^{-1}}+a_{2}\ell^{2},$ (5)
where $\delta$ is the turbulence correlation length, $W$ is the observations
beam radius,
$\mathcal{N}=\Delta^{\prime}(\delta^{2}+2W^{2})/\sqrt{2\pi}\delta^{3}$ is the
number of turbulent cells along the line of sight, and $\Delta^{\prime}$ is
the effective depth of the cloud. The small-scale $B$-field contribution to
the observed dispersion is quantified by the first term in the equation and
the large scale contribution is described by the second term.
We use this angle dispersion structure function to determine $\sigma_{\phi}$
on a pixel-by-pixel basis for each of the observations we have chosen. Before
we apply the dispersion function to the whole image, we need to determine the
values of $\Delta^{\prime}$, $\delta$, and the kernel size $w$. Depth of the
cloud is determined using the width of the auto-correlation function of the
polarized intensity ($P=\sqrt{Q^{2}+U^{2}}$) given by;
$\langle\overline{P}^{2}(\ell)\rangle\equiv\langle\overline{P}(r)\overline{P}(r+\ell)\rangle,$
(6)
as described in Houde et al. (2009). Based on the assumption that the cloud
has similar characteristics across and along its depth, it is a reasonable
approximation of the effective depth of the cloud even though it is a function
of $\ell$ along the cloud surface. We calculated the normalized auto-
correlation function for each of the observations and the corresponding plot
is shown in Fig. 8. The estimated values of $\Delta^{\prime}$ at 53, 216, and
850 µm are $0.47^{\prime}$, $0.81^{\prime}$, and $0.78^{\prime}$ respectively.
Table 1: DCF method parameters.
Parameter | 53 µm | 216 µm | 850 µm
---|---|---|---
beamsize (arcmin) | $4.84$ | $18.2$ | $20$
pixel size (arcsec) | $2.42$ | $4.55$ | $10$
$\Delta^{\prime}$ (arcmin) | $0.47$ | $0.81$ | $0.78$
$w$ (pixels) | 9 | 7 | 13
$\delta_{\mathrm{fit}}$ (arcsec) | $21$ | $21.69$ | -
$[\langle B_{t}^{2}\rangle/\langle B_{0}^{2}\rangle]_{\mathrm{fit}}$ | $0.16$ | $0.019$ | -
$[\langle B_{t}^{2}\rangle/\langle B_{0}^{2}\rangle]_{\mathrm{mean}}$ | $0.29$ | $0.019$ | $0.41$
$\sigma_{v}$ (km/s) | $5.3(8.77)^{*}$ | $8.68$ | $9.75$
$\sigma_{\phi}$ (rad) | $0.42$ | $0.12$ | $0.64$
$n_{{}_{\mathrm{H}}}^{\mathrm{mean}}$ (cm-3) | $10^{4.2}$ | $10^{3.95}$ | $10^{3.96}$
$B_{{}_{\mathrm{POS}}}^{\mathrm{mean}}$ (mG) | $2.07(3.4)^{*}$ | $6.53$ | $1.39$
∗ estimates from all the velocity components.
In order to estimate the dispersion function locally around each pixel, it is
important to choose a kernel size ($w$) in pixel radius over which
$1-\langle\mathrm{cos}\left[\Delta\phi(\ell)\right]\rangle$ is estimated for
each pair of polarization angles. This symmetric two-dimensional normalized
circular kernel ensures no preferential direction is chosen when estimating
the dispersion function. The size has to be greater than the observation beam
size and the turbulence correlation length so that we have enough vectors
within the kernel to calculate the dispersion function with greater accuracy.
We used the method described by Guerra et al. (2021) to determine the optimal
kernel size for each observation. As each of our polarization maps probe a
different physical scale with quite different resolutions, the corresponding
kernel size could also be different. We fit the dispersion function to every
pixel for a range of kernel sizes ranging from $w=5-15$ pixels and calculated
the Spearman’s correlation coefficient ($\rho_{\mathrm{sp}}$) for each kernel
size. The best size was chosen based on the highest median value of estimated
$\rho_{\mathrm{sp}}$ and was found to be 9, 7, and 13 pixels for the 53, 216,
and 850 µm observations, respectively.
Figure 9: Dispersion function fit to the entire field of view for the 53 µm
(top) and the 216 µm (bottom) observations. The blue dots are the estimates
from the observations and the dashed red line is the best fit model for each
case. The first term of the dispersion function which gives an estimate of the
small-scale turbulence component is shown as the cyan line and the large-scale
mean field component is shown by the orange line. The size of the kernel
chosen ($w=9$ and 7 for 53 and 216 µm observations, respectively) for each of
the maps is shown by the black dotted line in both the plots.
Using the above described values of $\Delta^{\prime}$ and $w$, we fit the
dispersion function for each of the polarization observation to determine the
parameters $\delta$, $a_{2}$, and $\langle B_{t}^{2}\rangle/\langle
B_{0}^{2}\rangle$. However, we found most of the estimated $\delta$ values
($\sim 50$%) for the 216 and 850 µm observations to be $\delta<W\sqrt{2}$
which is the correlated beam size of the observations. This indicates that the
resolution is not good enough to resolve the local gas turbulence and might
lead to incorrect estimates of the $B$-field. Thus we tried another approach
presented in Guerra et al. (2021) where we fix the value of $\delta$ based on
the fit of a single dispersion function for the whole region (global
$\delta$). The best fit $\delta$ was found to be $21.69^{\prime\prime}$ for
the 216 µm observation. We used the same value for the 850 µm observation as
they have relatively the same beam sizes ($18.2^{\prime\prime}$ and
$20^{\prime\prime}$ respectively) and due to the lower data quality of the 850
µm observation.
Figure 10: Maps of the $B_{\mathrm{POS}}$ estimated from HAWC+ 216 µm (left)
and SCUPOL 850 µm (right) observations using the DCF method. The location of
the 50 km s-1 and 20 km s-1 clouds are shown as the red and magenta circles,
respectively. The beam size of the observations is shown at the bottom of each
map.
We tried both the approaches of a fixed $\delta$ of $21^{\prime\prime}$ and
using $\delta$ as a free parameter for the 53 µm observation. When $\delta$ is
a free parameter, as in the previous cases about 40% of the pixels still
showed $\delta<W\sqrt{2}$. Hence we chose to fix the $\delta$ value to
$21^{\prime\prime}$ obtained from a single dispersion function fit to the
whole image. There were some pixels with large error bars due to missing data
in the surrounding region within the kernel size ($w=9$). In these cases we
increased the kernel size to 11 to get better estimates of the parameters. By
using this method we were able to get reasonable values of $\langle
B_{t}^{2}\rangle/\langle B_{0}^{2}\rangle$ for most of the pixels of the
observation ($\sim 90\%$).
The plots of the single dispersion function fits to the 53 µm and 216 µm data
are shown in Fig. 9. The small-scale turbulent (first term of the dispersion
function) and the large-scale mean field (last term) components are also
shown. It can be seen from the plots that the turbulent component dominates
the dispersion when $\ell<44^{\prime\prime}$ and $\ell<1^{\prime}$ in 53 and
216 µm observations, respectively. Beyond this size there is significant
contribution from the large-scale component to the observed dispersion. These
sizes are also around the same as the kernel sizes chosen for each
observation.
The best fit values of $\langle B_{t}^{2}\rangle/\langle B_{0}^{2}\rangle$
from each observation was used to estimate $\sigma_{\phi}$ using the
approximation $\sigma_{\phi}\simeq[\langle B_{t}^{2}\rangle/\langle
B_{0}^{2}\rangle]^{1/2}$. The mean value of $\langle B_{t}^{2}\rangle/\langle
B_{0}^{2}\rangle$ was found to be 0.29, 0.019, and 0.41 for 53, 216 and 850 µm
observations respectively. The 216 µm observation shows the least dispersion
which reflects the rather uniform polarization vectors observed in Fig. 1 & 2.
This might be a result of majority of the dust emission at this wavelength
originating from a region of strong magnetic field, and will be discussed in
the following sections.
Figure 11: Maps of $\sigma_{v}$ from blue-shifted velocity components (left),
all velocity components (middle), and $\sigma_{\phi}$ (right) for the 53 µm
HAWC+ observation used to estimate the magnetic field of the region from DCF
method. Figure 12: $B_{{}_{\mathrm{POS}}}$ for the 53 µm observations
estimated using the DCF method from the blue-shifted velocity components is
shown on the left and from all the velocity components is shown in the middle.
The map on the right shows the $B$-field estimate from the DMA method. The
mean field from each map is 2.07$\pm$0.03 mG, 3.4$\pm$0.04 mG, and
2.85$\pm$0.04 mG respectively. The HAWC+ polarization vectors are overlaid on
all the maps with its scale shown at the top of the map on the left.
Figure 13: Histogram distribution of the estimated $B_{{}_{\mathrm{POS}}}$
from the 53 (left), 216 (middle), and 850 (right) µm observations.
Now, we measure the $B_{{}_{\mathrm{POS}}}$ using the modified DCF method
described in Equation 4. The mass density was determined from the gas volume
density ($n_{{}_{\mathrm{H}}}$) using the relation $\rho=\mu
m_{{}_{\mathrm{H}}}n_{{}_{\mathrm{H}}}$, where $\mu=2.8$ is the mean molecular
weight per unit mass of hydrogen (Kauffmann et al., 2008; Sadavoy et al.,
2013) and $m_{{}_{\mathrm{H}}}$ is the mass of hydrogen (Crutcher, 2004). The
volume densities were determined using the depth of the clouds estimated from
the auto-correlation function and the column density measurements from
Herschel observations. The resulting values match well with the earlier
predictions using molecular line observations ($n_{{}_{\mathrm{H}}}\simeq
10^{4.1}$ cm-3; Oka et al., 2011). More details about the estimation of
temperature and gas column density of the region can be found in Akshaya &
Hoang (2023).
The velocity dispersion described in Section. 3 was corrected for the
contribution from molecular thermal motion to extract only the dispersion due
to turbulence using the relation
$\sigma_{v}^{2}=\sigma_{v0}^{2}-k_{\mathrm{B}}T_{\mathrm{gas}}/m_{\mathrm{CO}}$,
where $\sigma_{v0}$ is the dispersion measured from the molecular
spectroscopic data, $k_{\mathrm{B}}$ is the Boltzmann constant, and
$T_{\mathrm{gas}}$ is the temperature of gas in the region. We use the same
approximation of $T_{\mathrm{gas}}=T_{\mathrm{dust}}$ as in our previous
study. We have also ignored the pixels where the best fit values of $\langle
B_{t}^{2}\rangle/\langle B_{0}^{2}\rangle<0.001$ as these are found mostly at
the boundaries of the polarization observations and might be a result of poor
fitting due to insufficient data.
The $B_{{}_{\mathrm{POS}}}$ of the 216 and 850 µm observations were estimated
using the mean $\sigma_{v}$ from all the resolved components shown in Fig. 5.
The resultant maps of the magnetic field are shown in Fig. 10 and have a mean
value of 6.53$\pm$0.08 mG and 1.39$\pm$0.02 mG for 216 and 850 µm
observations, respectively. The high magnetic field estimate of the 216 µm
observation is due to very low dispersion observed in the polarization
measurement at this wavelength, as can be seen from the rather uniform
polarization vectors in Fig. 1. A histogram of the distribution of the
magnetic field for each wavelength is shown in Fig. 13. It is important to
note that these estimates are from the DCF method without any correction
factor applied to the equation, hence the mean $B$-field strength can be
treated as an upper limit of the field at these wavelengths.
#### 4.1.1 B${}_{{}_{\mathrm{POS}}}$ from resolved velocity components
We have used a different approach to measure the $B$-field of the 53 µm
observation. As discussed in Section. 3, the morphology of the observed
emission at the 53 µm wavelength matches well with the integrated map of the
velocity components from the blue-shifted region of the CO ($J=3\rightarrow
2$) spectrum. Not taking this into account while using the $\sigma_{v}$ value
will lead to an overestimation of the $B$-field. To understand the importance
of resolving and isolating the velocity components in such complex regions we
estimate the $B$-field from two values of $\sigma_{v}$. In one case we
consider the mean $\sigma_{v}$ only from the blue-shifted Gaussian components
of the spectrum while in the other case we take the mean value from all the
resolved components. The maps of $\sigma_{v}$ for the two cases and the
$\sigma_{\phi}$ for this region is shown in Fig. 11. Similar to the previous
case, we have ignored the pixels with $\langle B_{t}^{2}\rangle/\langle
B_{0}^{2}\rangle<0.01$ at the boundaries of the polarization data. The
resulting $B_{{}_{\mathrm{POS}}}$ maps are shown in Fig. 12. There is a clear
difference in the estimated strength especially in the regions away from the
central source Sgr A∗. The mean field of the region estimated from the blue-
shifted components is about $B_{-v}^{\mathrm{DCF}}=2.07\pm 0.03$ mG and the
estimate by considering all the components is $B_{v}^{\mathrm{DCF}}=3.4\pm
0.04$ mG. The distribution of the field from both cases are shown in Fig. 13.
The field from all the components seem to show two peaks in the distribution,
at around 1 mG and 4 mG.
Figure 14: Histogram comparing the $B_{{}_{\mathrm{POS}}}$ measured from DMA
and DCF methods for the 53 µm observation, using correction factor $f=1$.
Figure 15: Correlation between the estimated difference in the
$B_{{}_{\mathrm{POS}}}$ measured from DCF and DMA method with the number of
components resolved along the line of sight (top) and the measured velocity
dispersion (bottom).
### 4.2 $\mathbf{B_{{}_{\mathrm{POS}}}}$ from Differential Measure Analysis
We also measure the $B$-field strength using another recent modification of
the original DCF method, called the Differential Measure Analysis (DMA)
technique proposed by Lazarian et al. (2022). The observed dispersion in the
polarization is an effect of the tug-off between the turbulence induced
randomisation and aligning effect due to tension in the magnetic field.
Stronger the magnetic field, lesser the randomised turbulence. The DCF method,
traditionally used for the measurement of the $B$-field from dust polarization
is based on the assumption that the turbulence in isotropic. However, MHD
turbulence is anisotropic by nature (Beresnyak & Lazarian, 2019, and
references therein) and the DMA technique was introduced by Lazarian et al.
(2022) to address this key problematic assumption of the widely used DCF
technique.
The DCF technique we have employed in the previous sections is a modified
version from the original method proposed by Houde et al. (2009), where they
constrain the dispersion in the polarization angle taking into account the
dispersion induced by the changes in the large scale orientation of the
$B$-field as well as the beam integration effects. However, as discussed in
Lazarian et al. (2022) the method still assumes the turbulence to be
isotropic, where we use the LOS velocity dispersion as an accurate description
of the POS velocity fluctuation and comparable to the observed dispersion in
the polarization angle in the POS. The DMA method corrects this assumption by
using the small scale structure function of the velocity centroids (defined as
$C(r)$) for the estimate of the velocity fluctuations instead of the line
width velocity dispersion, similar to the method proposed by Cho & Yoo (2016).
The centroid velocities give an estimate of the average of the mean velocities
from multiple individual eddies along the LOS and are also not affected by
thermal broadening, which is another parameter that contributes to the error
in the $B$-field estimates from DCF. Using the simplest form of DMA method
from Lazarian et al. (2022) and Hu & Lazarian (2023), the POS $B$-field is
given by;
$B=f\sqrt{4\pi\rho}\sqrt{\frac{D_{v}(\ell)}{D_{\phi}(\ell)}},$ (7)
where $D_{v}$ and $D_{\phi}$ are the structure functions of velocity centroid
$C(r)$ and the polarization angle within scale $\ell$ respectively. The factor
$f$ is similar to the one used in the DCF method and in our case we use $f=1$
for both DCF and DMA estimates of the $B_{{}_{\mathrm{POS}}}$. The structure
functions are estimated as,
$\begin{split}D_{v}=\langle(C(r)-C(r+\ell))^{2}\rangle,\\\
D_{\mathrm{\phi}}=\langle(\phi(r)-\phi(r+\ell))^{2}\rangle,\end{split}$ (8)
where the centroid velocity $C(r)$ is given by,
$C(r)=\frac{\int{T_{\mathrm{mb}}(r,v)vdv}}{\int{T_{\mathrm{mb}}(r,v)dv}}.$ (9)
Here $T_{\mathrm{mb}}(r,v)$ is the brightness temperature of the CO line we
have chosen with $r=(x,y)$ and $v$ its respective plane of sky position and
velocity. We have applied the structure function on the same scale as used in
the dispersion function for the DCF method. Thus we choose a kernel size of
$w=9$, keeping the estimates consistent with the DCF from the previous
section.
As discussed in Cho & Yoo (2016), the DCF method gives appropriate estimate of
the $B_{{}_{\mathrm{POS}}}$ when the number of independent fluctuations
(eddies) along the LOS is small. This is not the case for regions along the
Galactic disk, as can be seen from the spectrum shown in Fig. 4, where we
clearly have multiple features along the LOS. Thus the turbulence injection
and driving scale and the number of independent fluctuations along the LOS
become important for the DCF method, while they are taken care of by the
velocity centroid approach in the DMA method. Our approach of resolving the
different individual components which contribute to the net velocity
fluctuation along the LOS might address this problem with DCF to some degree.
We have estimated the $B_{{}_{\mathrm{POS}}}$ for the 53 µm observation from
the above described DMA method using only the blue shifted part of the
spectral cube (with $v<0$ km s-1) and the corresponding map is shown in Fig.
12. The mean magnetic field estimated from the DMA method is
$B_{\mathrm{POS}}^{\mathrm{DMA}}=2.85\pm 0.04$ mG while that from the DCF
method is $B_{\mathrm{POS}}^{\mathrm{DCF}}=2.07\pm 0.03$ mG. A histogram of
the distribution of $B_{{}_{\mathrm{POS}}}$ from both the methods is shown in
Fig. 14. Though the overall distribution of $B_{{}_{\mathrm{POS}}}$ looks
similar, there are regions where there is a large difference between them.
Fig. 15 shows the relation between the difference in the estimated
$B_{{}_{\mathrm{POS}}}$ and the number of resolved components along the LOS
($N_{\mathrm{comps}}$) and $\sigma_{v}$. The difference seems to be the
highest where the velocity dispersion is high and there was no apparent
relation between the estimated difference and the angle dispersion. The key
difference in the two methods is in the way velocity fluctuations are
measured, with the velocity centroids used in DMA being a better way to
constrain the turbulence driven fluctuations along the LOS. The overall
agreement in the measured $B_{{}_{\mathrm{POS}}}$ is a positive indication
that resolving the individual components can address some of the problems
while implementing DCF for the estimate of the magnetic field. Numerical
modelling of this approach paired with DMA can give further understanding on
how the various velocity fluctuations along the LOS resolved from molecular
spectra affect the observed polarization and will be addressed in our future
work.
## 5 Discussion
### 5.1 Alfvén Mach number ($\mathcal{M_{\mathrm{A}}}$)
We have used the 3D Alfvén Mach number ($\mathcal{M}_{\mathrm{A}}$) defined as
the ratio of the turbulent velocity and the Alfvén speed to understand the
interplay between the turbulence in the region and the magnetic field. The
relation is given by,
$\mathcal{M}_{\mathrm{A}}=\sqrt{3}\frac{\sigma_{v}}{\mathcal{V}_{\mathrm{A}}},$
(10)
where $\sigma_{v}$ is the one-dimensional non-thermal velocity dispersion
which characterises the turbulence and $\mathcal{V}_{\mathrm{A}}$ is the
Alfvén velocity given by $B_{\mathrm{tot}}/\sqrt{4\pi\rho}$, where $\rho$ is
the gas mass density and
$B_{\mathrm{tot}}=B_{{}_{\mathrm{LOS}}}+B_{{}_{\mathrm{POS}}}$ is the total
mean magnetic field which can also be approximated as
$B_{\mathrm{tot}}=(4/\pi)B_{{}_{\mathrm{POS}}}$ (Crutcher et al., 2004). The
$\sqrt{3}$ in the equation takes care of the 3D velocity dispersion
approximation from the one-dimensional estimates, assuming isotropic
turbulence in the region. A value of $\mathcal{M}_{\mathrm{A}}<1$ indicates
sub-Alfvénic condition where the magnetic field dominates the gas motion while
a value of $\mathcal{M}_{\mathrm{A}}>1$ indicates the super-Alfvénic condition
where turbulence pressure dominates over the magnetic pressure.
We have estimated $\mathcal{M}_{\mathrm{A}}$ for the HAWC+ 53 and 216 µm
observations. The JCMT 850 µm observation was not considered due to relative
low quality of the data, though can be used to get a rough estimate of the
field strength in the region are not reliable enough to draw conclusions on
the importance of turbulence on these scales. The $\mathcal{M}_{\mathrm{A}}$
estimate for 216 µm observation is sub-Alfvénic throughout the region due to
its high ordered magnetic field. The distribution of
$\mathcal{M}_{\mathrm{A}}$ is much more distinct in the 53 µm observation and
is shown in Fig. 16. $\mathcal{M}_{\mathrm{A}}>1$ along the Eastern Arm of the
mini-spiral where the $B_{{}_{\mathrm{POS}}}\lesssim 1$mG. Turbulent pressure
clearly dominates the gas motion in this region. The dust components probed by
the 53 µm observation show a distribution of sub- and super-Alfvénic
conditions, in contrast to the 216 µm observation where the regions is mostly
sub-Alfvénic.
Figure 16: Alfvén Mach number ($\mathcal{M}_{\mathrm{A}}$) estimated for the
53 µm observation. The $\mathcal{M}_{\mathrm{A}}>1$ is seen along the Eastern
Arm of the mini-spiral indicating the gas kinematics to be driven by
turbulence in the region. The HAWC+ polarization vectors are overlaid on map
with its scale shown at the top and the beamsize shown at the bottom of the
figure.
### 5.2 Mass-to-Flux ratio ($\lambda$)
The CND is known to be clumpy and is a key region to understand how the gas
mass is being fed to our central supermassive black hole (SMBH) as well as the
star formation in this region (Hsieh et al., 2018, 2021). The mini-spirals
observed within the CND are believed to be molecular gas that are in-flowing
from ambient clouds well outside the CND, and contain compact dense cores of
varies sizes as seen from Atacama Large Millimeter/submillimeter Array (ALMA)
observations (Hsieh et al., 2019). These cores need to be dense enough to
survive the tidal disruption along the path where they become part of the CND
and are eventually fed to the SMBH at the centre. An understanding of whether
these cores can lead to successful star formation in this complex and highly
dynamic environment around our SMBH has great implications for the study of
nuclear star clusters and star formation and evolution of galaxies. One of the
key parameters to probe the possible star formation in the presence of
magnetic field is the mass-to-flux ratio ($M/\phi\equiv\lambda$) which
estimates whether the magnetic field can support the cloud against
gravitational collapse. Following Crutcher et al. (2004), $\lambda$ is defined
in terms of the critical value of mass that can be supported by magnetic flux
($M_{\mathrm{crit}}=\phi_{\mathrm{crit}}/2\pi\sqrt{G}$; Nakano & Nakamura,
1978) as,
$\lambda=\frac{(M/\phi)_{\mathrm{obs}}}{(M/\phi)_{\mathrm{crit}}}=\frac{\mu
m_{{}_{\mathrm{H}}}N(H_{2})/B_{\mathrm{tot}}}{1/2\pi\sqrt{G}}=7.6\times
10^{-21}\frac{N(H_{2})}{B_{\mathrm{tot}}},$ (11)
where $\mu=2.8$ is the mean molecular weight, $m_{\mathrm{H}}$ is the mass of
hydrogen atom, $G$ is the gravitational constant, $N(H_{2}$) is the gas column
density in cm-2, and $B_{\mathrm{tot}}$ is the total magnetic field strength
in $\mu$G. A value of $\lambda>1$ indicates that the magnetic field cannot
prevent gravitational collapse of the cloud and is said to be magnetically
supercritical. If $\lambda<1$ then the cloud is magnetically supported and is
said to be magnetically sub-critical. We estimate the mass-to-flux ratio for
the 53 µm observation focused on the CND and the corresponding map is shown in
Fig. 17. Most of the region is magnetically sub-critical due to the high
magnetic field with $B_{{}_{\mathrm{POS}}}>1$ mG. Only along the Eastern Arm
where we also observe $\mathcal{M}_{\mathrm{A}}>1$ do we see $\lambda>1$
indicating the magnetic field might not be strong enough to support gravity.
The physical scale probed by the HAWC+ observation is not high enough to
resolve the dense cores observed along the CND. However, this gives us an idea
of the region of weak magnetic field, where possible star formation can be
triggered in the CND.
Figure 17: Mass-to-Flux ratio ($\lambda$) of the 53 µm observation.
$\lambda>1$ along the Easter Arm of the mini-spiral within the CND where the
estimated magnetic field is $B_{{}_{\mathrm{POS}}}\lesssim 1$ mG. The HAWC+
polarization vectors are overlaid on map with its scale shown at the top and
the beamsize shown at the bottom of the figure.
### 5.3 Grain alignment and field morphology
From our previous study of the CND and its surroundings in Akshaya & Hoang
(2023), we found that the grain alignment in this region can be perfect if the
dust grains contain even low levels of iron atoms (super-paramagnetic grains
with $\sim 20$ iron atoms per cluster). The magnetic field was assumed to be 5
mG throughout the region for the study. The results still hold with the
current detailed map of the $B_{{}_{\mathrm{POS}}}$. The least magnetic field
observed is within the CND with $B_{{}_{\mathrm{POS}}}\leq 1$ mG, where we
also observe a drop in the column density. Earlier Zeeman measurements of this
region by Plante et al. (1995) estimate similar field strength of
$B_{{}_{\mathrm{LOS}}}<1$ mG. It is interesting to note the agreement between
the two methods of $B$-field estimation. Except in the few regions at the
edges of the CND, the overall magnetic field for the 53 µm observation show
$B_{{}_{\mathrm{POS}}}\leq 5$ mG. Thus, we reinforce that if the dust grains
in the region contain even a small percent of iron atoms, we can expect
perfect alignment of dust with the magnetic field in the region where the
alignment will be driven by Magnetically Enhanced Radiative Alignment (MRAT;
Hoang et al., 2022a). But without iron atoms, the grain alignment is only
driven by radiative torques (Dolginov & Mitrofanov, 1976; Draine &
Weingartner, 1997; Lazarian & Hoang, 2007) as the field in the region is not
strong enough to enhance the alignment via magnetic relaxation. A detailed map
to the metallicity of the region can better help constrain the degree of grain
alignment and aid in the construction of a 3D morphology of the $B$-field in
this region.
Hoang et al. (2023) propose a new way to map the 3D $B$-field from the 2D dust
polarization observations based on the MRAT alignment theory, taking into
account the local physical conditions that affect the net polarization
efficiency of the dust grains. Considering the strong magnetic field in the GC
which can promote greater degree of magnetic relaxation in contrast to the
diffuse ISM, particularly in the presence of iron clusters in the dust grains,
it would be insightful to apply this method to derive a complete picture of
the variation of the $B$-field along the LOS and how it effects the transport
of material in this region and will be addressed in a future study.
### 5.4 3D magnetic field from multi-wavelength thermal dust polarization
The major focus of this work has been to map the magnetic field of the GC in
multi-wavelengths and to test if regions such as the GC with multiple
resolvable velocity fluctuations along the LOS can be used to create a picture
of how the $B_{{}_{\mathrm{POS}}}$ changes at different depths, assuming the
different wavelengths are probing different layers at varied temperatures. We
believe we have achieved this to some extent given the limitations of the DCF
method used to determine the magnetic field. The estimated
$B_{{}_{\mathrm{POS}}}$ is in good agreement with the earlier Zeeman
measurements of $B_{{}_{\mathrm{LOS}}}$ from Killeen et al. (1992) and Plante
et al. (1995). Killeen et al. (1992) obtained $B_{{}_{\mathrm{LOS}}}\sim 2$ mG
in the north and the south of the CND while Plante et al. (1995) suggested the
$B_{{}_{\mathrm{LOS}}}<1$ mG within the CND. Assuming the dust polarization
and Zeeman measurements trace similar regions along the CND, using our POS
magnetic field from DCF of $B_{{}_{\mathrm{POS}}}^{\mathrm{mean}}\sim 2.07$
mG, we can estimate an approximate mean full strength of 3D $B$-field of the
CND as;
$\displaystyle
B_{{}_{\mathrm{3D}}}=\sqrt{B_{{}_{\mathrm{POS}}}^{2}+B_{{}_{\mathrm{LOS}}}^{2}},$
(12)
which yields $B_{{}_{\mathrm{3D}}}\sim 2.87$ mG.
The observed polarization in the 53 µm appears to be independent of the deeper
velocity structures identified as shown in Fig. 5, but the 216 and 850 µm
observations seem to trace the relatively cooler dust from different depths
along the LOS. Numerical simulations can be used to investigate if combining
such multi-wavelength polarization observations with the information of the 3D
distribution of dust along the LOS can be used to create a 3D morphology of
the magnetic field. As an extension of this study, we plan to use the recently
upgraded POLArized RadIation Simulator (POLARIS; Reissl et al., 2016) by Giang
et al. (2023), which incorporates the latest grain alignment theories
discussed in our previous work (Akshaya & Hoang, 2023) to model how the
distribution of independent velocity fluctuations affect the observed
polarized emission at different wavelengths in this region. Combined with the
limited $B_{{}_{\mathrm{LOS}}}$ Zeeman measurements, this might be a step
closer to getting a comprehensive view of the elusive 3D morphology of the
magnetic field in the Galactic disk. Due to the nature of the grain alignment
and how sensitive it is to the local physical conditions like the metallicity,
temperature, dust composition, local density, and the amount of incident
radiation, we need to use the latest knowledge of the dust alignment physics
to get a full picture of the dynamical interaction between the magnetic field
and the material in this complex region.
### 5.5 Implications of B-field strength from multi-wavelength polarization
at the GC
The CND is well known to be in the influence of the gravitational potential of
the Sgr A∗ and also play a role in the accretion of material onto the inner
sub-parsec scale ionized cavity surrounding Sgr A∗ (Solanki et al., 2023).
There is also evidence of collisional interaction between the 20 km s-1 cloud
and the outer edges of the CND, which might aid further mass accretion onto
the CND and the inner cavity (Takekawa et al., 2017). The evolution studies of
the CND suggest its formation due to interaction and break down of giant
molecular clouds with the gravitational potential of the Sgr A∗, with two
molecular clouds (20 km s-1 and 50 km s-1 clouds) ideally located at the
Galactic south of the CND (Sanders, 1998; Oka et al., 2011; Mapelli & Trani,
2016). The CND is also observed to have an extended feature in the negative
Galactic longitude direction called the negative-longitude extension (NLE;
Serabyn & Guesten, 1986; Sutton et al., 1990; Takekawa et al., 2017). This is
a foreground (with respect to the CND) feature and we observe a similar
structure in the velocity range of $-50>v>-165$ km s-1 shown in Fig. 18.
Considering the similar $B_{{}_{\mathrm{POS}}}$ estimates for the 53 µm
observation covering the CND and the 850 µm observation with the CND, NLE, and
the 20 km s-1 features, the 850 µm maybe tracing the CND and its foreground
material at a lower temperature. This cannot be confirmed due to the low data
quality of the SCUPOL observation. However, it is interesting nonetheless to
disentangle the source location of the magnetic field to get a view of how it
varies with the depth along the LOS. The 216 µm observations measure the
highest magnetic field out of the three observations that we have considered.
From Fig. 5 it is evident that there is some degree of interaction between the
cloud clusters in the negative velocity region but there is no interaction
between the $v<0$ km s-1 and $v>0$ km s-1 cloud features. Also by looking at
Fig. 6, majority of the emission in the 216 µm observation matches well with
the component morphologies at $v>0$ km s-1, unlike the 53 µm intensity which
match well with the integrated morphologies from $v<0$ km s-1. If the observed
dust emissions are arising from these proposed components, then the strong
magnetic field of the 216 µm observation can be attributed to the velocity
components in the background with respect to the CND. The uniformity in the
polarization vectors at this wavelength have been noted by earlier observation
at 250 µm by the PILOT experiment as well (Mangilli et al., 2019). Considering
our data quality check and the agreement between the observations taken by
other instruments and in different modes, this uniformity of polarization can
be treated as a real physical effect, arising due to dust grains being
perfectly aligned with a strong magnetic field.
Modelling and numerical simulations are necessary to further confirm this idea
of the different field strength observed from multi-wavelength polarization
could be a result of dust being traced at different depths along the LOS and
will be addressed in our future work. The growth of the CND and subsequent
feeding of material into the central black hole is an ongoing study with most
of the simulation not yet considering the effect of the magnetic field in the
material dynamics due to the complexity of the region. An overview of the
strength and 3D morphology of the $B$-field in the region that can be
estimated from polarization observation, complemented with the current grain
alignment physics is a great way to further our understanding of the material
transport at the centre of our Galaxy.
Figure 18: The NLE (black circle) observed in CO ($J=3\rightarrow 2$) spectra
evident in the region covered by the JCMT/SCUPOL observation, right below the
CND and above the location of the 20 km s-1 cloud (shown as the magenta
circle). The component appears in the blue shifted velocity range of
$-50>v>-165$ km s-1.
## 6 Summary
We have used thermal dust polarization observations at 53, 216, and 850 µm
combined with the spectrum of the CO ($J=3\rightarrow 2$) transition to map
the POS magnetic field for a region of about 30 pc around the centre of our
Galaxy. The main conclusions from our study are as follows;
1. 1.
The velocity dispersion in the region can be decomposed into multiple
physically distinct components with an average dispersion of about $\sim 9$ km
s-1.
2. 2.
The physical morphologies of the negative velocity structures match best with
the observed morphology of the CND and the mini-spirals at 53 µm, indicating
that these structure are at a much higher temperature than those at positive
velocities. This also shows that the velocity dispersion at this wavelength
has the contribution arising only from the negative velocity components.
3. 3.
We used the DCF method to estimate the map of the $B_{{}_{\mathrm{POS}}}$ for
all the three observations and found the mean field to be $2.07\pm 0.03$,
$6.53\pm 0.08$, and $1.39\pm 0.02$ mG at 53, 216, and 850 µm respectively.
4. 4.
Most of the region encompassing the CND and the mini-spiral is sub-Alfvénic
with $\mathcal{M}_{\mathrm{A}}<1$ except along the Eastern Arm of the mini-
spiral where $\mathcal{M}_{\mathrm{A}}>1$, indicating the gas motion being
driven by turbulence.
5. 5.
The same region also has a mass-to-flux ratio of $\lambda>1$ indicating
magnetic field cannot prevent gravitational collapse. The CND is known to have
several clumps with $n_{{}_{\mathrm{H}}}>10^{7}$ cm-3, though these scales are
not probed in the current observation the map of $\lambda$ shows the likely
region for star formation in the CND.
6. 6.
We find good agreement between the $B_{{}_{\mathrm{POS}}}$ estimated from DCF
and DMA method for the 53 µm observation, indicating that the decomposing of
the spectra into its constituent multiple velocity components might overcome
the drawback of the velocity dispersion used in the DCF method, where
determining the turbulence driving scale and the number of independent
fluctuations along the LOS become crucial to overcome the overestimation of
the $B$-field.
7. 7.
The similarity in the estimated $B_{{}_{\mathrm{POS}}}$ for the 53 µm and the
850 µm observation might be due to dust emission in these wavebands tracing
the same components, mostly including the CND and its foreground, where as the
216 µm observation might have majority of its emission coming from a component
in the background with respect to the CND, with a much stronger magnetic
field.
## Acknowledgements
MSA thanks Dr. Lopez-Rodriguez and Dr. G. S. Pillai for the insightful
discussion on the polarization data quality assessment. This work was partly
supported by a grant from the Simons Foundation to IFIRSE, ICISE (916424,
N.H.). This study is based in part on observations made with the NASA/DLR
Stratospheric Observatory for Infrared Astronomy (SOFIA). SOFIA is jointly
operated by the Universities Space Research Association, Inc. (USRA), under
NASA contract NNA17BF53C , and the Deutsches SOFIA Institut (DSI) under DLR
contract 50 OK 2002 to the University of Stuttgart. This work made use of
Astropy:444http://www.astropy.org a community-developed core Python package
and an ecosystem of tools and resources for astronomy (Astropy Collaboration
et al., 2013, 2018, 2022). This research made use of APLpy, an open-source
plotting package for Python (Robitaille & Bressert, 2012).
## Data Availability
The data underlying this article will be shared on reasonable request to the
corresponding author.
## Appendix A Supplementary Figures
Figure 19: Maps of the dispersion in velocity (left) and polarization angle
(right) used for the estimation of the $B_{{}_{\mathrm{POS}}}$ from DCF method
for the HAWC+ 216 µm observation.
Figure 20: Maps of the dispersion in velocity (left) and polarization angle
(right) used for the estimation of the $B_{{}_{\mathrm{POS}}}$ from DCF method
for the SCUPOL 850 µm observation.
## References
* Aitken et al. (1986) Aitken D. K., Briggs G. P., Roche P. F., Bailey J. A., Hough J. H., 1986, MNRAS, 218, 363
* Aitken et al. (1998) Aitken D. K., Smith C. H., Moore T. J. T., Roche P. F., 1998, MNRAS, 299, 743
* Akshaya & Hoang (2023) Akshaya M. S., Hoang T., 2023, MNRAS, 522, 4196
* Astropy Collaboration et al. (2013) Astropy Collaboration et al., 2013, A&A, 558, A33
* Astropy Collaboration et al. (2018) Astropy Collaboration et al., 2018, AJ, 156, 123
* Astropy Collaboration et al. (2022) Astropy Collaboration et al., 2022, ApJ, 935, 167
* Barnes et al. (2017) Barnes A. T., Longmore S. N., Battersby C., Bally J., Kruijssen J. M. D., Henshaw J. D., Walker D. L., 2017, MNRAS, 469, 2263
* Becklin et al. (1982) Becklin E. E., Gatley I., Werner M. W., 1982, ApJ, 258, 135
* Beresnyak & Lazarian (2019) Beresnyak A., Lazarian A., 2019, Turbulence in Magnetohydrodynamics. De Gruyter, Berlin, Boston, doi:doi:10.1515/9783110263282, https://doi.org/10.1515/9783110263282
* Bryant & Krabbe (2021) Bryant A., Krabbe A., 2021, New Astron. Rev., 93, 101630
* Butterfield et al. (2023) Butterfield N. O., et al., 2023, arXiv e-prints, p. arXiv:2306.01681
* Butterfield et al. (2024) Butterfield N. O., et al., 2024, arXiv e-prints, p. arXiv:2401.01983
* Chandrasekhar & Fermi (1953) Chandrasekhar S., Fermi E., 1953, ApJ, 118, 113
* Chen et al. (2022) Chen C.-Y., Li Z.-Y., Mazzei R. R., Park J., Fissel L. M., Chen M. C. Y., Klein R. I., Li P. S., 2022, MNRAS, 514, 1575
* Cho & Yoo (2016) Cho J., Yoo H., 2016, ApJ, 821, 21
* Christopher et al. (2005) Christopher M. H., Scoville N. Z., Stolovy S. R., Yun M. S., 2005, ApJ, 622, 346
* Chuss et al. (2019) Chuss D. T., et al., 2019, ApJ, 872, 187
* Crutcher (2004) Crutcher R. M., 2004, in Uyaniker B., Reich W., Wielebinski R., eds, The Magnetized Interstellar Medium. pp 123–132
* Crutcher et al. (2004) Crutcher R. M., Nutter D. J., Ward-Thompson D., Kirk J. M., 2004, ApJ, 600, 279
* Davis (1951) Davis L., 1951, PhRv, 81, 890
* Dolginov & Mitrofanov (1976) Dolginov A. Z., Mitrofanov I. G., 1976, Ap&SS, 43, 291
* Draine & Weingartner (1997) Draine B. T., Weingartner J. C., 1997, ApJ, 480, 633
* Eden et al. (2020) Eden D. J., et al., 2020, MNRAS, 498, 5936
* Falceta-Gonçalves et al. (2008) Falceta-Gonçalves D., Lazarian A., Kowal G., 2008, ApJ, 679, 537
* Genzel et al. (2010) Genzel R., Eisenhauer F., Gillessen S., 2010, Rev. Modern Phys., 82, 3121
* Giang et al. (2023) Giang N. C., Hoang T., Kim J.-G., Tram L. N., 2023, MNRAS, 520, 3788
* Guerra et al. (2021) Guerra J. A., Chuss D. T., Dowell C. D., Houde M., Michail J. M., Siah J., Wollack E. J., 2021, ApJ, 908, 98
* Guerra et al. (2023) Guerra J. A., Lopez-Rodriguez E., Chuss D. T., Butterfield N. O., Schmelz J. T., 2023, AJ, 166, 37
* Guesten et al. (1987) Guesten R., Genzel R., Wright M. C. H., Jaffe D. T., Stutzki J., Harris A. I., 1987, ApJ, 318, 124
* Harper et al. (2018) Harper D. A., et al., 2018, J. Astron. Instr., 7, 1840008
* Henshaw et al. (2016) Henshaw J. D., et al., 2016, MNRAS, 457, 2675
* Henshaw et al. (2019) Henshaw J. D., et al., 2019, MNRAS, 485, 2457
* Henshaw et al. (2022) Henshaw J. D., Barnes A. T., Battersby C., Ginsburg A., Sormani M. C., Walker D. L., 2022, arXiv e-prints, p. arXiv:2203.11223
* Hildebrand (1988) Hildebrand R. H., 1988, QJRAS, 29, 327
* Hildebrand et al. (2009) Hildebrand R. H., Kirby L., Dotson J. L., Houde M., Vaillancourt J. E., 2009, ApJ, 696, 567
* Hoang (2022) Hoang T., 2022, ApJ, 928, 102
* Hoang & Lazarian (2016) Hoang T., Lazarian A., 2016, ApJ, 831, 159
* Hoang et al. (2022a) Hoang T., Tram L. N., Minh Phan V. H., Giang N. C., Phuong N. T., Dieu N. D., 2022a, AJ, 164, 248
* Hoang et al. (2022b) Hoang T. D., et al., 2022b, ApJ, 929, 27
* Hoang et al. (2023) Hoang T., Phan V. H. M., Tram L. N., 2023, arXiv e-prints, p. arXiv:2301.07832
* Houde et al. (2009) Houde M., Vaillancourt J. E., Hildebrand R. H., Chitsazzadeh S., Kirby L., 2009, ApJ, 706, 1504
* Hsieh et al. (2018) Hsieh P.-Y., Koch P. M., Kim W.-T., Ho P. T. P., Tang Y.-W., Wang H.-H., 2018, ApJ, 862, 150
* Hsieh et al. (2019) Hsieh P.-Y., Koch P. M., Kim W.-T., Ho P. T. P., Yen H.-W., Harada N., Tang Y.-W., 2019, ApJ, 885, L20
* Hsieh et al. (2021) Hsieh P.-Y., et al., 2021, ApJ, 913, 94
* Hu & Lazarian (2023) Hu Y., Lazarian A., 2023, MNRAS, 524, 4431
* Hu et al. (2022) Hu Y., Lazarian A., Wang Q. D., 2022, MNRAS, 513, 3493
* Hwang et al. (2021) Hwang J., et al., 2021, ApJ, 913, 85
* Jackson et al. (1993) Jackson J. M., Geis N., Genzel R., Harris A. I., Madden S., Poglitsch A., Stacey G. J., Townes C. H., 1993, ApJ, 402, 173
* Kauffmann et al. (2008) Kauffmann J., Bertoldi F., Bourke T. L., Evans N. J. I., Lee C. W., 2008, A&A, 487, 993
* Kauffmann et al. (2017) Kauffmann J., Pillai T., Zhang Q., Menten K. M., Goldsmith P. F., Lu X., Guzmán A. E., 2017, A&A, 603, A89
* Killeen et al. (1992) Killeen N. E. B., Lo K. Y., Crutcher R., 1992, ApJ, 385, 585
* Kormendy & Kennicutt (2004) Kormendy J., Kennicutt Robert C. J., 2004, ARA&A, 42, 603
* Lazarian (2007) Lazarian A., 2007, J. Quant. Spectrosc. Radiative Transfer, 106, 225
* Lazarian & Hoang (2007) Lazarian A., Hoang T., 2007, MNRAS, 378, 910
* Lazarian et al. (2022) Lazarian A., Yuen K. H., Pogosyan D., 2022, ApJ, 935, 77
* Lee et al. (2021) Lee D., et al., 2021, ApJ, 918, 39
* Li et al. (2022) Li P. S., Lopez-Rodriguez E., Ajeddig H., André P., McKee C. F., Rho J., Klein R. I., 2022, MNRAS, 510, 6085
* Lo & Claussen (1983) Lo K. Y., Claussen M. J., 1983, Nature, 306, 647
* Mangilli et al. (2019) Mangilli A., et al., 2019, A&A, 630, A74
* Mapelli & Trani (2016) Mapelli M., Trani A. A., 2016, A&A, 585, A161
* Marshall et al. (1995) Marshall J., Lasenby A. N., Yusef-Zadeh F., 1995, MNRAS, 274, 519
* Matthews et al. (2009) Matthews B. C., McPhee C. A., Fissel L. M., Curran R. L., 2009, ApJS, 182, 143
* Myers et al. (2023) Myers P. C., Stephens I. W., Coudé S., 2023, arXiv e-prints, p. arXiv:2312.09330
* Nakano & Nakamura (1978) Nakano T., Nakamura T., 1978, PASJ, 30, 671
* Ngoc et al. (2021) Ngoc N. B., et al., 2021, ApJ, 908, 10
* Novak et al. (1997) Novak G., Dotson J. L., Dowell C. D., Goldsmith P. F., Hildebrand R. H., Platt S. R., Schleuning D. A., 1997, ApJ, 487, 320
* Oka et al. (2011) Oka T., Nagai M., Kamegai K., Tanaka K., 2011, ApJ, 732, 120
* Oort (1977) Oort J. H., 1977, ARA&A, 15, 295
* Ostriker et al. (2001) Ostriker E. C., Stone J. M., Gammie C. F., 2001, ApJ, 546, 980
* Paré et al. (2024) Paré D., Butterfield N. O., Chuss D. T., Guerra J. A., Iuliano J. I., Karpovich K., Morris M. R., Wollack E., 2024, arXiv e-prints, p. arXiv:2401.05317
* Pattle et al. (2017) Pattle K., et al., 2017, ApJ, 846, 122
* Pillai et al. (2015) Pillai T., Kauffmann J., Tan J. C., Goldsmith P. F., Carey S. J., Menten K. M., 2015, ApJ, 799, 74
* Planck Collaboration et al. (2016) Planck Collaboration et al., 2016, A&A, 586, A138
* Planck Collaboration et al. (2020a) Planck Collaboration et al., 2020a, A&A, 641, A12
* Planck Collaboration et al. (2020b) Planck Collaboration et al., 2020b, A&A, 643, A42
* Plante et al. (1995) Plante R. L., Lo K. Y., Crutcher R. M., 1995, ApJ, 445, L113
* Reissl et al. (2016) Reissl S., Wolf S., Brauer R., 2016, A&A, 593, A87
* Robitaille & Bressert (2012) Robitaille T., Bressert E., 2012, APLpy: Astronomical Plotting Library in Python, Astrophysics Source Code Library, record ascl:1208.017 (ascl:1208.017)
* Sadavoy et al. (2013) Sadavoy S. I., et al., 2013, ApJ, 767, 126
* Sanders (1998) Sanders R. H., 1998, MNRAS, 294, 35
* Santos et al. (2019) Santos F. P., et al., 2019, ApJ, 882, 113
* Schwarz & Lasenby (1990) Schwarz U. J., Lasenby J., 1990, in Beck R., Kronberg P. P., Wielebinski R., eds, IAU Symp., Vol. 140, Galactic and Intergalactic Magnetic Fields. p. 383
* Serabyn & Guesten (1986) Serabyn E., Guesten R., 1986, A&A, 161, 334
* Skalidis & Tassis (2021) Skalidis R., Tassis K., 2021, A&A, 647, A186
* Solanki et al. (2023) Solanki S., Ressler S. M., Murchikova L., Stone J. M., Morris M. R., 2023, ApJ, 953, 22
* Sutton et al. (1990) Sutton E. C., Danchi W. C., Jaminet P. A., Masson C. R., 1990, ApJ, 348, 503
* Takekawa et al. (2017) Takekawa S., Oka T., Tanaka K., 2017, ApJ, 834, 121
* Temi et al. (2018) Temi P., Hoffman D., Ennico K., Le J., 2018, J. Astron. Instr., 7, 1840011
* Trippe et al. (2008) Trippe S., et al., 2008, A&A, 492, 419
* Veilleux et al. (2020) Veilleux S., Maiolino R., Bolatto A. D., Aalto S., 2020, A&ARv, 28, 2
* Yusef-Zadeh et al. (1996) Yusef-Zadeh F., Roberts D. A., Goss W. M., Frail D. A., Green A. J., 1996, ApJ, 466, L25
* Yusef-Zadeh et al. (1999) Yusef-Zadeh F., Roberts D. A., Goss W. M., Frail D. A., Green A. J., 1999, ApJ, 512, 230
* Zhao et al. (2009) Zhao J.-H., Morris M. R., Goss W. M., An T., 2009, ApJ, 699, 186
|
# Swarm Learning: A Survey of Concepts, Applications, and Trends
Elham Shammar
School of Cyber Science and Engineering,
Wuhan University, Wuhan, China
<EMAIL_ADDRESS>
Xiaohui Cui
School of Cyber Science and Engineering,
Wuhan University, Wuhan, China
<EMAIL_ADDRESS>
Mohammed A. A. Al-qaness
College of Physics and Electronic
Information Engineering
Zhejiang Normal University
Jinhua 321004, China
<EMAIL_ADDRESS>
###### Abstract
Deep learning models have raised privacy and security concerns due to their
reliance on large datasets on central servers. As the number of Internet of
Things (IoT) devices increases, artificial intelligence (AI) will be crucial
for resource management, data processing, and knowledge acquisition. To
address those issues, federated learning (FL) has introduced a novel approach
to building a versatile, large-scale machine learning framework that operates
in a decentralized and hardware-agnostic manner. However, FL faces network
bandwidth limitations and data breaches. To reduce the central dependency in
FL and increase scalability, swarm learning (SL) has been proposed in
collaboration with Hewlett Packard Enterprise (HPE). SL represents a
decentralized machine learning framework that leverages blockchain technology
for secure, scalable, and private data management. A blockchain-based network
enables the exchange and aggregation of model parameters among participants,
thus mitigating the risk of a single point of failure and eliminating
communication bottlenecks. To the best of our knowledge, this survey is the
first to introduce the principles of Swarm Learning, its architectural design,
and its fields of application. In addition, it highlights numerous research
avenues that require further exploration by academic and industry communities
to unlock the full potential and applications of SL.
_Keywords_ IoT, Blockchain, Swarm Learning; Edge Computing, Security,
Decentralized Machine Learning, Federated Learning, Privacy Preservation
## 1 Introduction
The next five years are expected to witness a significant increase in the
number of IoT devices. In 2019, the healthcare sector utilizes one-third of
all IoT devices, which are expected to climb to 40%, or $6.2 trillion, of the
total global IoT technology market value by 2025 [1]. The global adoption of
IoT devices is expected to reach 29 billion by 2030, covering a wide range of
economic sectors and disciplines [2]. Particularly, IoMT devices are poised to
save $300 billion, predominantly in the chronic illness and telemedicine
sectors. This market is considered attractive for investors, with projections
that estimate revenues of $135 billion by 2025 [3]. Moreover, the global
healthcare market is expected to grow to $6.2 trillion by 2028 [4],
necessitating advancements in AI, resource management, data processing, and
knowledge mining. The rapid advancement of the 5G standard and Multi-Access
Edge Computing (MEC) has markedly improved productivity [5].
Modern deep learning models are raising concerns about privacy and security
due to their reliance on centralized servers to store large datasets [6].
Although cloud-based local learning allows some level of collaboration and
improvement of results, it introduces several inherent challenges to this
centralized approach, such as data redundancy, increased data traffic, and
increased security and privacy risks. Two primary challenges associated with
traditional centralized learning methods are data ownership and privacy [4].
Federated learning (FL) emerges as a viable solution to these challenges,
potentially aligning with data protection standards that could conflict with
traditional centralized learning approaches [7]. FL promises notable
improvements in security, fairness, and transparency, setting a new benchmark
for digital data management and model training [8].
FL facilitates collaborative learning that preserves privacy. It addresses
central data storage issues by allowing the raw data to remain on local
devices at each participating node [5],[9],[10]. However, FL is still
vulnerable to sophisticated cyber threats, including membership inference and
data reconstruction attacks, which pose significant risks of data breach. FL
also has limitations in network bandwidth that cause delays. To mitigate these
vulnerabilities, two approaches are introduced: 1) Distributed FL (DFL [11])
and 2) a novel approach called Swarm Learning (SL) that was developed in
collaboration with Hewlett Packard Enterprise (HPE[6]).
DFL and SL are approaches to machine learning that improve privacy and reduce
reliance on centralized data storage. DFL extends the traditional federated
learning model by allowing multiple nodes to train models collaboratively
without a central server[12],[13], [14], while Swarm Learning uses blockchain
technology to create an autonomous peer-to-peer network without a central
authority. Both approaches aim to decentralize learning and enhance privacy,
but SL employs blockchain for even greater security and decentralization.
SL is a decentralized machine learning framework that combines the principles
of blockchain technology with federated learning. Instead of using a central
server to compile model updates as in standard FL, SL uses a peer-to-peer
network that is managed by blockchain to guarantee member validity, data
integrity, and security. SL trains models locally, and only parameter weights
are transmitted on a network of numerous swarm devices. The integration of
blockchain technology ensures secrecy and security, enabling effective
collaboration among disparate entities. Transactions can only be performed by
preauthorized parties through computationally efficient consensus mechanisms.
SL eliminates the need for a central server, reducing the risk of single
points of failure and centralized data breaches. Unlike FL, which ensures data
privacy through aggregating initial local gradients, SL facilitates data
sharing among registered customers via smart contracts, thus preserving data
privacy. A node in SL must undergo registration, authentication, model
retrieval, local training, gradient sharing, and finally, result aggregation
using the Federated Average method [15].
SL enhances fault tolerance, reduces vulnerability to attacks, and supports
scalability, making it ideal for applications requiring high data privacy and
system robustness, such as healthcare, the automotive industry, financial
services, smart cities, edge computing, IoT, and the metaverse. In healthcare,
SL guarantees the preservation of data privacy by allowing hospitals and
research institutions to train models collaboratively without sharing
sensitive patient data [16]. In the industry, SL enables machines and system
components to act as individual learning agents, allowing real-time decision-
making and adjustments without central oversight. It aligns well with Industry
4.0 principles, supporting advanced manufacturing technologies requiring high
levels of data integrity, flexibility, and automation [17], [18], [19], [20].
In financial services, SL can enhance fraud detection systems by learning
transaction data between different entities without compromising client
confidentiality [21]. In smart cities, SL can optimize traffic flow and public
transport management by allowing multiple sensors and nodes to learn and adapt
to real-time traffic conditions. SL supports data sovereignty and
auditability, ensuring compliance with data protection regulations. It also
offers innovation and competitive advantage, allowing faster time to market
and customization[22].
HAN et al. [23] sought to bridge the gap between the theoretical aspects of SL
and its practical application, providing empirical evidence through
experiments carried out on three public datasets. Their findings have
evidenced that SL is supposed to be suitable for most application scenarios,
no matter whether the dataset is balanced, polluted, or biased over irrelevant
features. However, challenges remain, such as backdoor attacks against SL,
managing blockchain integration complexity, and dealing with computational
overhead.
### 1.1 Paper objectives and contribution
The considerable advantages offered by SL require a detailed examination to
understand its current research landscape and practical applications, as well
as to pinpoint areas requiring further improvement. To this end, this SL
survey aims to investigate its capabilities within decentralized learning
environments. Our objectives are to assess its practical implementation,
identify both technical and operational challenges, and highlight potential
avenues for future innovations. Furthermore, the survey seeks to explore
forward-looking developments, such as the integration of advanced
cryptographic techniques to enhance security and the adaptation of SL to
support emerging technologies such as edge computing and the IoT.
This effort will consolidate existing knowledge, clarify research gaps, and
outline strategic directions to expand the adoption of SL. As a resource, this
survey will be invaluable for scholars, researchers, and practitioners. By
improving academic discourse and guiding practical implementations, it aims to
pave the way for a broader application and optimization of SL in various
industries, thus expanding its impact and utility.
To sum up, the main contribution of this paper can be presented as follows:
* •
We present the first survey paper in the field of swarm learning (SL). To the
best of our knowledge, this literature review is the first review on SL.
* •
We provide a comprehensive overview of the existing literature on Swarm
learning and its current applications to give readers a complete picture of
this new and promising research direction.
* •
We studied and analyzed the current applications of SL. We categorized them
into healthcare, transportation, industry, robotic systems, smart homes,
financial services, multimedia IoT, fake news detection, and Metaverse.
* •
We present an in-depth analysis of the current limitations and challenges
facing SL. We explore how these issues impact their development and
deployment. Additionally, we discuss potential future directions to improve SL
technologies and applications. We suggest paths for advancement and areas ripe
for further research to enhance the effectiveness and applicability of SL
technologies.
The road map of this paper, as shown in Fig.1 is outlined as follows: Section
2 provides an introduction to swarm learning and its fundamental concepts and
components. Sections 3 and 4 explore the applications of swarm learning and
its associated challenges, respectively. Section 5 highlights potential
directions for future research in SL. The paper is concluded in Section 6.
Figure 1: Paper Structure
### 1.2 Paper Selection
We conducted a comprehensive search in six databases, namely IEEE, PubMed,
Science Direct, Scopus, Springer, and Web of Science. Specifically, we
retrieved 25 papers from IEEE, 10 from PubMed, 116 from Science Direct, 72
from Scopus, 28 from Springer, and 43 from Web of Science. Subsequently, we
meticulously screened these papers, focusing on those directly related to
swarm learning, while excluding articles on swarm intelligence and swarm
optimization. Following this screening process, we identified a total of 56
papers that met our inclusion criteria.
The number of research papers has increased each year, as shown in Fig. 2.
Research on SL has steadily increased since its humble beginnings in 2020. By
2024, it experienced a significant increase, indicating the growing importance
of SL in various fields. This surge highlights the growing interest of the
academic community in exploring and maximizing the potential of this advanced
technology. The surge in SL research is driven by advances in computational
power, data availability, the proliferation of IoT devices, privacy-preserving
AI techniques, and the emergence of complex problems such as healthcare,
autonomous driving, and smart cities, which require scalable and decentralized
learning methods.
Figure 2: Annual increase in the number of Swarm Learning research papers.
### 1.3 Research Questions
1. 1.
What are Swarm Learning concepts, architecture, and components?
2. 2.
What is the difference between Swarm Learning and Federated Learning,
Distributed FL/Decentralized Federated Learning, and Swarm Intelligence?
3. 3.
What are the applications of swarm learning?
4. 4.
What challenges do we see in the adoption and implementation of swarm learning
in real-world applications?
## 2 Swarm Learning (SL)
SL is a decentralized machine learning framework that enables the training of
the on-device model without the need to transfer raw data. In the SL model,
the data is kept localized at the data owner’s site, substantially reducing
data traffic by avoiding the transmission of raw data [24]. Using blockchain
technology, SL enhances privacy and security through the exchange of only the
model parameters and weights, not the actual data itself. This approach
incorporates smart contracts to manage the training and updating of the
decentralized machine learning models using local user data, distinguishing it
significantly from traditional centralized systems or even FL frameworks that
rely on a central server for aggregating model updates [21]. Additionally, SL
incorporates advanced data privacy and security mechanisms, making it an
ideal, flexible, and secure solution for content caching within contemporary
network architectures [25].
SL employs a permissioned blockchain network and a decentralized hardware
infrastructure to facilitate secure member onboarding, dynamic leader
election, and efficient merging of model parameters. The system utilizes
standardized AI engines within a distributed machine learning context to
ensure secure and reliable operations. An SL library supports an iterative AI
learning process that leverages decentralized data, adhering rigorously to the
prevailing privacy and security standards [26]. This structured approach
secures data and also streamlines the computational process across diverse
network nodes.
### 2.1 Swarm Learning Architecture
The SL architecture encompasses two primary layers: the application layer and
the infrastructure (or hardware) layer. The application layer includes the
Machine Learning (ML) platform, blockchain, and the Swarm Learning Library
(SLL). The hardware layer consists of data sources and models relevant to
specific domains, such as datasets related to missions or geographic locations
[4].
The SL system consists of two components: Swarm edge nodes and Swarm network
(blockchain) [24]. With blockchain technology, SL has the following
characteristics and advantages: (1) storing vast amounts of data locally; (2)
reducing data traffic by not requiring the exchange of original data; (3) not
requiring a secure central network; (3) offering high-level data security and
shielding the model from attacks; and (5) allowing all members to merge
parameters with equal rights [26].
Fig. 3 [6],[16], [27] depicts the architecture of the swarm learning system.
There are several swarm edge nodes (let us say, M nodes), and each node Ci
uses local private data Di, i = 1, 2, 3,…, M, to train its model Li after
downloading an initial model from the network. Then, every node Ci distributes
its model parameters throughout the network. These nodes are recognized,
permitted, and registered with a smart contract in a peer-to-peer blockchain
network to safeguard network security. Each node Ci has an opportunity to be
chosen as a temporary leading node C for model aggregation in a training cycle
t. When the local model Li is trained to satisfy predetermined synchronization
requirements (such as a predetermined training batch), several chosen nodes
will disclose their model parameters to a storm API. As a result, each chosen
node will get the global model parameters from the leading node C, which will
then use a weighted average approach to aggregate them into a global model G
[6].
Figure 3: Swarm learning system architecture
The Swarm network diagram shows how edge nodes are set up to exchange
parameters to learn, with blockchain technology serving as a facilitator.
Private data is used at each node in combination with models provided by the
Swarm network, guaranteeing a decentralized and secure method of collaborative
learning [16]. To take part in model training, Swarm edge nodes must register
via the blockchain’s smart contract. After registering, every node uniformly
downloads the first global model from the blockchain and trains the local
model using its local data. Swarm edge nodes upload the local model parameters
of the training to the leader via the Swarm network. The smart contract on the
blockchain selects the Swarm edge node leader in real-time. The leader will
average the collected local model parameters. To continue local model
training, each Swarm edge node will download the aggregate model from the
Swarm network until the aggregation model meets the requirements of the
trained aggregation model. If not, the leader in a block generates the
aggregation model [24].
The workflow for updating the model in SL, as shown in Fig.4 [28], consists of
two primary stages. Initially, individual organizations trained their local
models and updated them using their own SL nodes. These updates are then
consolidated on their respective permissioned blockchains. In the subsequent
stage, organizations use a network of multiple blockchains to further refine
their local models and synchronize the global model’s state. This approach of
sharing models across various blockchains fosters a more decentralized SL
process and mitigates security risks from external entities [28].
Figure 4: Workflow of SL with multiple permissioned blockchains. The chains of
different colors belong to different participating organizations
In SL, model sharing is seen as a data-transfer process among participating
blockchains. The challenge lies in creating a method for blockchain data
interplay that remains consistent and secure and is adaptable to various
blockchain types without altering the core operations. The diversity in
blockchain structures and consensus protocols used by different organizations
adds to the complexity of enabling interblockchain interactions. Traditional
methods of cross-chain communication, which often rely on a third-party trust
entity, contradict the decentralized nature of SL and are therefore not
suitable. Solutions such as the Cosmos architecture, which relies on a central
hub for blockchain interoperability, also fall short of the ideal
decentralized approach required for SL [28].
### 2.2 Leader Election Algorithm (LEA)
In SL, the fairness and performance of the network are greatly affected by the
leader election process. Swarm edge nodes in SL are best placed on instances
with plenty of bandwidth and processing power to handle the demands of
decentralized decision-making. However, the unfairness of the leader election
mechanism could cause nodes to use excessive amounts of bandwidth, which would
result in inefficiencies and possibly bottlenecks. Participants may be unhappy
with this discrepancy because they believe it is unfair and because nodes with
higher data traffic may be more easily targeted by attackers[23]
The current LEA speculated to be a Proof of Stake (PoS), relies on leadership
election on nodes’ stakes or account balances. The authors in [23] recommended
switching from PoS to a Proof of Work (PoW) model, in which nodes compete to
solve cryptographic puzzles and leadership is established by meeting
predetermined hash value requirements. By equating the likelihood of becoming
a leader based on processing power, this technique seeks to guarantee a more
fair distribution of network load among nodes. Future efforts will focus on
collaborating with Hewlett Packard Enterprise (HPE) to enhance the fairness
and effectiveness of LEA in SL.
### 2.3 Concept of Swarm Learning
ML, in theory, can be carried out locally if enough data and computing
equipment are available. The data and computation existed at different,
disconnected locations (Fig. 5 (A) [16]). In cloud-based computation, data are
transported centrally (Fig. 5 (B)[16]) so that centralized computing can be
used to perform machine learning. It greatly improves the amount of data
available for training, and thereby improves machine learning outcomes.
However, there are some disadvantages, such as increased data traffic and
duplication, as well as problems with data privacy and security. In FL,
parameter settings are managed by a central parameter server, while data
remain with the data owner/contributor, and computing is performed at the
location of local data storage and availability. Dedicated parameter servers
are in charge of gathering and dispersing local learning in FL (Fig. 5 (C)
[16]). Alternatively, SL eliminates the need for a dedicated server, as shown
in Fig. 5 (D).SL distributes the parameters over the swarm network and
develops the models separately at each location using private data [16].
The integration of ML methods into the SL framework can increase training
rates. SL’s decentralized nature allows local data processing at edge nodes,
reducing latency, and potentially speeding up the training process. It also
leverages the computational power of multiple decentralized nodes, improving
training speed. SL reduces communication overhead by distributing workloads
across multiple nodes, reducing the need for frequent communication between
nodes. The blockchain component in SL manages model updates securely and
efficiently, minimizing delays. Dynamic leader elections optimize the training
process by choosing the most capable nodes for crucial tasks. SL’s approach to
handling non-IID data across different nodes can enhance model robustness and
accuracy faster than centralized approaches. SL’s ability to operate on nodes
with varying computational capacities allows for resource optimization[16].
However, integrating ML methods into SL can introduce complexities, making it
difficult to analyze training rate improvements. Traditional machine learning
methods can vary in architecture and complexity, affecting learning rates,
convergence behaviors, and efficiencies. SL’s decentralized nature and varying
computational resources may affect efficiency and scalability. Blockchain
technology for synchronization may introduce overhead, and adjusting ML
methods to fit SL could complicate performance assessment. Empirical studies
and benchmarking against traditional centralized and federated learning
systems are needed to quantify the benefits of SL in real-world scenarios.
Figure 5: Comparative overview of learning models
### 2.4 Swarm Learning Components
As shown in Fig. 6 [29], the SL framework consists of various nodes:
* •
Swarm Learning (SL) node: SL nodes run the core of SL, sharing learnings and
incorporating insights.
* •
Swarm Network (SN) node: Using the Ethereum blockchain, the SN nodes
communicate with each other to track training progress and save global state
information about the model. Additionally, during initialization, every SL
node registers with an SN node, and each SN node manages the training pipeline
for its corresponding SL nodes. Note that the model parameters are not
recorded by the blockchain; instead, it simply stores metadata such as the
model state and the training progress.
* •
Swarm Operator (SWOP) nodes: SWOP nodes manage SL operations, performing tasks
such as starting and stopping Swarm runs, building and upgrading ML
containers, and sharing models for training.
* •
Swarm Learning Command Interface (SWCI) nodes: SWCI nodes monitor the
framework and can connect to any SN node in a given framework.
* •
Swarm Learning Management User Interface (SLM-UI): SLM-UI nodes are GUI
management tools used to install the framework, deploy Swarm training, monitor
progress, and track past runs[29].
* •
SPIFFE SPIRE Server node: SPIFFE SPIRE Server node ensures the SL framework’s
security. A SPIRE Agent Workload Attestor plugin is included in each SN or SL
node, and it interacts with the SPIRE Server nodes to verify the identities of
each node and to get and maintain an SPIFFE Verifiable Identity Document
(SVID) [23].
* •
License Server (LS) node installs and manages the license to run the SL
framework[23].
SL security and digital identity are handled by X.509 certificates, which can
be generated by users or standard security software like SPIRE. SL components
communicate using TCP/IP ports, and participating nodes must be able to access
each other’s ports[29].
Figure 6: Swarm learning Components
### 2.5 Features of Swarm Learning
Swarm learning encompasses several distinct features that strengthen its
application in decentralized settings:
1. A.
Privacy Preservation: SL keeps data at each node which minimizes the risk of
privacy breaches and confidentiality.
2. B.
Decentralization: SL reduces the risk of a single point of failure or data
monopoly by eliminating the need for a central data storage or authority for
model aggregation.
3. C.
Continuous Learning: Models are continuously updated with new data available
at each node, adapting to new conditions such as emerging diseases.
4. D.
Data Diversity and Volume: SL handles larger and varied datasets from multiple
nodes, enhancing model robustness and generalization.
5. E.
Collaborative Learning: Nodes collaborate to train a shared model, benefiting
from shared insights without actual data transfer, crucial to maintaining
patient confidentiality.
### 2.6 Swarm Learning vs. Federated Learning
SL and FL are two distributed learning techniques that provide aggregation of
cooperative models from numerous participating nodes [30],[31]. Several
training rounds will result in the generation of a global model. Furthermore,
to guarantee equitable and safe model aggregation, these participating nodes
are not required to disclose their proprietary datasets. However, there are
two key distinctions between them [26], [32].
* •
Information transmission: In FL, participating nodes and the central server
exchange local model parameters as well as global model updates. However, in
SL, peer-to-peer networks based on blockchain technology and edge computing
work together to ensure that participating nodes can transmit safely and
fairly without the need for central server coordination [6], [33],[34], [35].
* •
With or without a central server: In FL, a central server is utilized to
collect model parameters from involved nodes and employ model aggregation to
generate a global model. On the other hand, SL does not make use of a central
server. During each training cycle, every participating node has the
opportunity to be randomly selected to serve as a temporary server to compile
model modifications. [6]. Using a blockchain-based Swarm network for safe and
decentralized parameter exchange and aggregation of the model, SL eliminates
the need for a central server [26].
SL addresses several key issues in FL and provides many benefits in security,
privacy, and scalability. SL can envision ways to develop more secure,
private, and faster distributed machine learning applications from different
domains.
To tackle the gradient leakage and data privacy issues in FL, Madni et al [15]
developed a secure, collaborative, and decentralized framework for machine
learning training by combining blockchain technology with SL. SL protects the
privacy of the data and the secrecy of the model parameters without unattended
accesses and guarantees data integrity, since it authenticates only trusted
nodes and deploys blockchain mechanisms. Research has been conducted against
common machine learning approaches for anomaly detection, where it is
demonstrated that the SL method gives better precision than current methods
and addresses gradient leakage, which is the current major limiter of the FL.
In their two articles [36], [37], Xu et al. addressed issues such as data
heterogeneity, security, and communication bottlenecks in FL by creating a
strong edge learning framework for smart IoT devices. They presented a new
technique called Communication-Efficient and Byzantine-Robust Distributed
Swarm Learning (CB-DSL). This work is the first thorough theoretical
examination of FL in conjunction with PSO (particle swarm optimization). It
provides a closed-form formula to assess the projected convergence rate of CB-
DSL, which makes it superior to traditional FL approaches such as Federated
Averaging (FedAvg). It also offers a model divergence analysis to assess the
possible advantages of adopting a globally shared dataset for enhancing
learning outcomes in non-IId. situations.
### 2.7 Swarm learning vs Distributed FL/Decentralized FL
Decentralized FL and SL are two approaches to distributed machine learning
that combine edge computing, blockchain technology, and peer-to-peer
networking [38]. Decentralized FL eliminates the need for a central server,
allowing for peer-to-peer communication and a more structured system, such as
blockchain technology [39]. It also includes a consensus mechanism for
updating the global model [40]. SL, developed by HPE Enterprise, integrates
blockchain technology into its core operation, ensuring data integrity, node
authenticity, and traceability. It also improves data privacy by keeping the
data localized and maintaining security through cryptographic measures.
Decentralized FL involves various nodes working together to train a global
model without a central coordinator, while Swarm Learning uses a leader
election mechanism to aggregate updates and update the blockchain. Both
approaches aim to decentralize the machine learning process and maintain data
localization, but SL incorporates blockchain for security and dynamic network
management. Both approaches are suitable for environments requiring high
levels of data integrity and auditability. SL offers advantages such as
enhanced privacy and security but may face challenges in privacy preservation
and server-centric issues. Future research could explore empirical comparisons
and develop hybrid models that combine the strengths of both SL and FL.
Beltrán et al.[12] explored the evolution of Decentralized Federated Learning
(DFL) compared to Centralized Federated Learning (CFL), highlighting its
benefits like improved fault tolerance and scalability. They compared DFL
frameworks and their implementation in various applications, including
healthcare, Industry 4.0, mobile services, military uses, and vehicular
networks. Hallaji et al.[41] explored the security and privacy of DFL,
highlighting its robustness and potential threats. They emphasized the need
for comprehensive security analyses and ongoing research to mitigate inherent
risks in DFL systems. The integration of blockchain technology with
decentralized federated learning (DFL) has been surveyed by Zhang et al.[11],
[42], highlighting its operational workflow and applications in the IoT and
Internet of Vehicles (IoV) domains. It discusses challenges like communication
overhead and system complexity, recommending further research.
The choice between DFL and SL depends on the application’s specific
requirements, such as security, trust requirements, operational complexity,
regulatory compliance, scalability, flexibility, real-time performance, data
privacy, and cost implications. SL is ideal for fields like healthcare and
finance, where data breaches or tampering could have severe consequences. DFL
is suitable for scenarios where operational complexity and resource
availability are concerns. SL is more suitable for highly regulated
environments and requires strict data provenance and audit trails. DFL offers
better scalability and flexibility, while SL may offer enhanced security
features. However, the implementation and maintenance of a blockchain for SL
can be more costly.
### 2.8 Swarm learning vs swarm intelligence
Swarm intelligence is a branch of artificial intelligence that uses the
principles of basic agent behavior research to provide algorithms for
scheduling, routing, and optimization issues. Particle Swarm Optimization
(PSO), Bee Colony Optimization (BCO), and Ant Colony Optimization (ACO) are
examples of swarm intelligence algorithms. In contrast, SL is a subset of
machine learning that focuses on distributed and dedicated learning without
sharing raw data. SL emphasizes decentralized and collaborative machine
learning in a privacy-preserving manner, while swarm intelligence focuses on
problem solving and process optimization, drawing natural influences from a
variety of systems.
Although SL and Swarm Intelligence have related names and are inspired by
natural swarm behaviors, which may be confusing, it is important to compare
them because, in computational and system design settings, they serve distinct
purposes and operate on different principles within computational and system
design contexts. Exploring the intersections and differences between SL and
swarm intelligence can lead to the development of hybrid approaches that
leverage the strengths of both. By comparing SL and swarm intelligence,
researchers can identify new application areas that may benefit from either
approach or a combination of both, aiding in educational and research
development. Ultimately, comparing SL and swarm intelligence enhances the
deployment of these technologies effectively across various domains.
SL provides an innovative set of effective solutions to the difficulties of
conventional optimization algorithms in swarm intelligence. By addressing
these issues, Swarm Learning overcomes the limitations of traditional
optimization algorithms in swarm intelligence and also opens new possibilities
for solving complex, dynamic, and large-scale optimization problems in a
secure, efficient, and privacy-preserving manner.
The Bacterial Foraging Optimization (BFO) algorithm, introduced by Kevin M.
Passino in 2002, is a nature-inspired optimization technique based on E.
coli’s natural foraging behavior. It has been applied in various fields,
including engineering, control systems, and optimization problems. However,
BFO has limitations depending on the problem’s nature and implementation
details, and its performance may not be ideal in all cases. Gan and Xiao [43]
introduced swarm learning strategies to improve convergence accuracy and
prevent premature convergence in BFO. This includes cooperative communication
with the global best bacteria and competitive learning mechanisms, improving
optimal solutions and swarm diversity, and addressing standard BFO
deficiencies.
Bolshakov et al. [44] have developed a deep reinforcement learning algorithm
called Deep Reinforcement Ant Colony Optimization (DRACO), inspired by
traditional ant colony optimization and designed for cooperative homogeneous
swarm learning. DRACO aims to shape collective behavior in decentralized
systems of independent agents, offering an alternative to centralized
learning. The algorithm’s advantages include natural parallelization, solving
collective tasks beyond the reach of single agents, increased reliability,
faster environmental exploration, and economic and energy efficiency.
### 2.9 Swarm Learning and IoT
In conventional cloud-based structures, IoT devices send data to central
servers for analysis. This approach can lead to potential bottlenecks,
compromise data privacy during transmission, and also increase latency. SL, on
the other hand, facilitates local data processing either on the device itself
or on proximate edge servers, thereby decreasing the necessity to transmit
sensitive data over the network and improving response times. SL enhances data
privacy and security by keeping data localized and using blockchain technology
for secure data sharing. This method ensures that sensitive data remain within
the local environment, complying with data protection regulations such as
GDPR. SL enables IoT devices to continuously learn and adapt in real-time,
providing real-time insights and real-time updates. The distributed nature of
SL provides excellent fault tolerance, making it suitable for IoT applications
such as healthcare monitoring systems and industrial automation. It scales
well without relying on a central server, making it suitable for sprawling IoT
networks. Implementing SL improves IoT networks’ efficiency, security, and
privacy compliance, making them better suited to handle vast amounts of data.
The next section will explore more into the applications of swarm learning in
IoT.
## 3 Applications of SL
SL is used in many fields, such as healthcare, autonomous vehicle systems,
environmental monitoring, and robotics, as shown in Fig.7, to improve
diagnostic accuracy, traffic flow, and safety. SL enables data aggregation
without compromising privacy, allows communication and learning from
experiences, and encourages cooperative robots for complex tasks. Its
potential to revolutionize distributed systems and information processing is
significant. The following subsections discuss the applications of SL in the
reviewed papers.
Figure 7: Swarm learning applications
### 3.1 Healthcare
Modern hospitals collect substantial volumes of private patient information
electronically. These data are extremely private and secret because they
pertain to both national security and individual privacy. The exchange of
medical data between institutions is restricted by legal and privacy concerns,
which impact the effectiveness of AI models trained on small datasets. While
distributed deep learning reduces communication and computing costs by making
optimal use of scattered data, it also poses privacy problems [24],[45]. SL
enables local machine learning model training using data from multiple health
nodes, such as hospitals. To maintain data privacy, the trained model
parameters are then exchanged, combined, and dispersed among nodes without the
requirement of a central collecting entity. By using blockchain, SL ensures
data security and confidentiality [45].
As shown in Fig.8[46], the SLN plays a central role by using its unique
digital identifier to train local models with private data and contribute to a
collective global model. The SNN, pivotal for consensus within the blockchain,
manages communication between the SLN and PBN, overseeing the training
process, and maintaining the model’s status. Lastly, the permissioned
blockchain network underpins the model-sharing aspect of swarm learning,
safeguarding the security and confidentiality of the process, and facilitating
effective collaboration between the SLNs.
Figure 8: The framework of metaverse swarm learning, which enables cross-
domain cooperation between metaverse and the physical world via blockchain
SL has demonstrated better performance in healthcare applications, such as
COVID-19 profiles and chest X-ray images, allowing ongoing learning and
enhancement across many data sources while closely respecting privacy laws
such as the General Data Protection Regulation (GDPR) and the Health Insurance
Portability and Accountability Act (HIPAA). It offers opportunities for the
development of cooperative research and diagnostics across hospitals and
research institutions networks and is flexible enough to fit a variety of
medical data environments. For example, German university hospitals are using
SL to evaluate COVID-19 patient data and create AI-based algorithms for the
detection of novel biomarkers. SL will develop into a crucial tool for
collaborative healthcare research and precision treatment [24],[45],[47].
For example, when hospitals use SL to manage COVID-19 data, they first gather
encrypted and anonymized patient data, including symptoms and treatments.
Every hospital sets up a separate SL node for safe local data processing. By
eliminating raw data exchange, these nodes preserve data privacy by locally
training models and sharing just the model parameters over a blockchain. These
parameters are then combined by a blockchain consensus method to update and
synchronize the global model across all nodes. Real-time deployment of this
continuously improving model enables more effective diagnosis and treatment
plans. SL improves predictive models by integrating diverse datasets from
multiple nodes, improving accuracy and treatment efficacy. It prioritizes
privacy and security by keeping sensitive patient data on-premises, reducing
reliance on central repositories. SL also increases efficiency in hospitals by
implementing personalized treatment plans. It is highly scalable, allowing
easy integration of new nodes without significant infrastructure changes.
Warnat-Herresthal et al. in their novel study [16] use SL to train AI models
on large datasets of histopathology images of more than 5,000 patients. SL was
demonstrated for disease classifier development using distributed data from
COVID-19, tuberculosis, leukemia, and lung pathologies, using over 16,400
blood transcriptomes. The study shows SL’s effectiveness in predicting
molecular alterations in colorectal cancer, demonstrating its potential for
enhancing medical imaging analysis without centralized data collection.
Fan et al. [26] was the first to examine the fairness problem in SL as it
relates to healthcare, mainly in duties related to the class of skin lesions.
They evaluated the fairness of the SL model for medical applications,
comparing performance and fairness with the single, centralized, and SL
models. The results show that SL can achieve better performance than single-
institution training and does not amplify biases. However, the study
acknowledges the high complexity of SL implementation due to the complex
configurations of the blockchain network.
To overcome the difficulties presented by non-independent and identically
distributed (non-IID) data in decentralized machine learning, notably in
clinical contexts, Wang et al. [48] introduced a generative augmentation
framework called SL-GAN, which combines a GAN in a swarm learning network to
augment non-IID data into IID data. The non-IID problem is directly addressed
for the first time in the context of SL in this paper, which is emphasized as
a significant advancement in decentralized clinical machine learning research.
The authors suggested improving synthetic data quality by introducing
differential privacy and studying synthetic data privacy.
DeMed, a decentralized privacy-preserving system for medical image processing
that uses blockchain technology, was proposed by Aggarwal et al. [49]. The
approach aggregates data into a classifier using smart contracts after using
self-supervised learning to create low-dimensional representations of medical
images. This paradigm seeks to address security and resilience concerns in
decentralized learning systems, with a particular focus on preventing
malicious or unintentional data alterations. The efficacy of the system is
demonstrated by independent medical picture classification tasks, such as
chest X-rays and pathological data.
By integrating swarm learning with homomorphic encryption. These papers [24],
[50] addressed a significant gap in distributed machine learning privacy-
preserving techniques. Swarm learning participants can securely share model
updates without disclosing sensitive data by incorporating homomorphic
encryption. To maintain participant privacy, the authors devised a partial
decryption algorithm that only required a fraction of the private key to allow
participants to decrypt aggregated model information locally. This
significantly advances the creation of machine learning applications in
domains where privacy is a concern. They recommended handling offline
participants, guarding against model poisoning, and maximizing encryption
trade-offs as areas of future research.
Gao et al.[8] proposed a unique strategy for SL that gathers local knowledge
from each center to overcome the forgetting of global knowledge during local
training. The proposed methodology demonstrates how utilizing data from
several centers can enhance medical picture segmentation while preserving data
privacy and resolving skew problems with non-IID data. The Label Skew-Aware
Loss (LaSA) is introduced to address label skew, preserving global label
information during local training. LaSA maximizes the forecast for the most
likely class determined by the global model. Feature Skew-Aware Regularization
(FeSA) is used to align local feature distributions with the global model,
mitigating the effects of feature skew caused by different imaging techniques
or demography.
Yuan et al. [51] developed a cooperative deep neural network (DNN)
partitioning system to accelerate disease diagnosis in multi-access edge
computing (MEC) networks. They used Swarm Reinforcement Learning (SRL) to
tackle the optimization problem of DNN partitioning and offloading and
blockchain technology to address challenges such as limited resources and
dynamic network environments. The algorithm allows agents to learn from local
data and generate judicious offloading actions.
A study by Saldanha et al. [33] used SL to identify molecular biomarkers in
gastric cancer from pathological images. They focused on microsatellite
instability (MSI) and Epstein-Barr virus (EBV) status. Patients cohorts from
the UK, USA, Switzerland, and Germany are included in their study. Every
dataset is kept apart from the others. However, the study was constrained by
uneven label classifications and the small number of biomarkers examined.
Future research must use a larger number of biomarkers and larger and more
diverse cohorts. To further enhance model performance and interpretability,
the authors recommended investigating attention-based deep learning
techniques.
Pan et al.[52] made a significant contribution to the field of drug
development. The study presented a "Nanonitrator," a nitrate nanoparticle made
of 3000 chitosan, sodium nitrate, and vitamin C as its main constituents. It
was produced utilizing the microencapsulation technique. The purpose of this
innovative nanoparticle is to improve nitrate’s long-circulating delivery
capability, extending its effects on the body’s duration and potency without
sacrificing safety. The authors described a novel method that uses a
combination drug prediction system driven by SL technology to improve the
bioavailability and protective effects of inorganic nitrate.
To predict molecular changes directly from hematoxylin and eosin (H&E)-stained
pathology slides of colorectal cancer, the study by Saldanha et al. [34] uses
SL to train AI models on large datasets of histopathology images from over
5,000 patients. The study shows the effectiveness of SL in predicting the BRAF
mutational status and microsatellite instability, demonstrating its potential
to improve medical imaging analysis without centralized data collection or
control.
A unique methodology based on SL was presented by Zhang et al. [46] for the
safe and equitable sharing of AI models in metaverse healthcare. The framework
addresses security, fairness, and data quality issues, improving model
accuracy and reliability. A novel parameter merging approach is devised to
maximize local models of SL nodes using lower-quality data. A permission
blockchain is used to incentivize high-quality data resources.
Mohammed et al. [53] developed a system using machine learning and SL to
diagnose diseases from nail images. The system uses transfer learning models,
InceptionV3 and VGG16, with an accuracy rate of 80%. The decentralized
approach eliminates trust and uses blockchain technology for parameter
merging. Despite limited training data, the system achieves an accuracy
comparable to or better than centralized models.
The problem of using vast amounts of medical data for cancer research,
particularly breast cancer, while adhering to privacy rules was the main focus
of the study by Shashank et al. [54]. The primary contribution lies in
showcasing SL, as a productive, privacy-preserving approach to improve
clinical research through the analysis of varied datasets from various
sources. They used 1,300 histochemical pictures of breast cancer tumors and
follow-up records to analyze diverse datasets, demonstrating how SL can
enhance clinical research, improve machine learning models, and maintain data
privacy without compromising quality.
By integrating user feedback into AI model training, Purkayastha et al.[55]
introduced a comprehensive approach that enables a more reliable and efficient
collaboration between radiologists and AI. The system uses few-shot learning
and SL, allowing continuous retraining of AI models based on active learning
strategies. The platform presents new capabilities for human-AI partnerships,
such as SL and few-shot learning methods. These techniques enable AI models
built on active learning algorithms to be continuously retrained. Through the
use of tailored model changes and collective knowledge, this approach
facilitates more accurate and repeatable radiological assessments.
Shriyan et al. [56] introduced a novel method to detect cataracts, which is
one of the most common eye conditions in the modern world, using SL. The
authors highlighted the benefits of SL over conventional FL and centralized
learning, emphasizing its effectiveness in the healthcare industry, especially
when data privacy is crucial. Hospitals can improve early cataract detection
by working together to create a global model while maintaining data privacy
through the use of SL. The method advocates for a scalable paradigm that might
include more nodes for higher data diversity and also proposes possible
applications in identifying other retinal illnesses.
Table 1 summarizes the main contributions of those articles.
Table 1: Swarm Learning in Healthcare Ref | Application | Contributions | Methodology | Used Datasets | Key Findings | Future Work
---|---|---|---|---|---|---
[16] | Medical imaging | Demonstrated the potential of SL for enhancing medical imaging analysis. | SL for large-scale pathology image analysis | Tuberculosis, leukemia, COVID-19, and lung pathologies | SL enhanced medical imaging analysis by facilitating multicentric collaboration and maintaining data privacy | Exploring additional medical fields where large, diverse datasets are crucial, possibly extending the decentralized model to more global collaboration.
[26] | Skin disease | Examined the fairness problem in SL | SL in skin lesion classification | Skin lesions | Investigated the fairness aspect of SL, showing robustness to heterogeneous data distributions and maintaining fairness without degrading performance | Future studies to improve model performance within the SL framework, focusing on managing model fairness and designing bias mitigation strategies for SL
[48] | Clinical settings | Overcame the difficulties presented by non-IID data in decentralized ML. | SL-GAN for non-IID data | Tuberculosis, Leukemia, and COVID-19 datasets | Addressed challenges posed by non-IID data in decentralized machine learning, specifically in clinical environments | Continued research to optimize decentralized clinical ML research, potentially exploring new algorithms and integration methods
[49] | Medical image analysis | A privacy-preserving decentralized framework for medical image analysis using blockchain technology. | Distributing a pre-trained Masked AutoEncoder (MAE) as a feature extractor and aggregating trained weights through smart contracts on the blockchain | Chest X-rays and pathological data | Developed a decentralized framework for medical image analysis, leveraging self-supervised learning and blockchain for privacy-preserving model training | Expanding the framework to include more complex medical imaging tasks, potentially increasing the variety of diseases that can be diagnosed using the system
[24], [50] | Privacy-preserving techniques | Including homomorphic encryption into SL. | Enhancing the Paillier homomorphic encryption using the Chinese Remainder Theorem for efficient operations and integrating a blockchain-based SL architecture for decentralized model aggregation through FedAvg | MNIST dataset | Significantly advanced machine learning applications in privacy-sensitive areas by allowing secure model updates sharing without revealing sensitive data | Enhancing defenses against model poisoning, optimizing encryption trade-offs, and handling offline participants
[8] | Medical imaging | Overcoming forgetting global knowledge during local training. Solves skew issues with Non-IID data. | Local knowledge assembly, LaSA loss, FeSA regularization | FeTS, M&Ms, MSProsMRI, MMWHS datasets | Enhanced medical image segmentation by handling Non-IID data issues, preserving data privacy | Further application to systems with unidirectional input constraints and expanding to other medical imaging tasks
[51] | Disease diagnosis | A cooperative DNN partitioning system for accelerating disease diagnosis in MEC networks. | Swarm Reinforcement Learning (SRL) in MEC networks | VGG16, AlexNet, ResNet18, NiN | Accelerated DNN-based disease diagnosis through cooperative DNN partitioning and offloading, minimizing service latency | Real-world applicability validation in clinical settings with specific constraints.
[33] | Molecular biomarker prediction | Predicting molecular biomarkers in gastric cancer from pathological images. | training MSI and EBV prediction models in individual merged cohorts and SL trained, using statistical analysis to assess prediction accuracy and explainability through pathologist-reviewed visualizations | Datasets from Bern, Leeds, TUM Cohort, TCGA | Improved prediction of molecular biomarkers in gastric cancer using multicentric data without compromising privacy | Expansion to include more biomarkers and larger datasets, exploring attention-based DL methods for improved model performance
[52] | Drug development | SL-based combination drug prediction system that identified vitamin C as the drug of choice to be combined with nitrate | AI-driven drug discovery, "Nanonitrator" nanoparticles | DPN, DDN, DTN from DrugBank, ChEMBL, UniProt | Enhanced bioavailability and therapeutic effects of inorganic nitrate for prolonged efficacy and safety | Not explicitly mentioned, but likely involve further clinical trials and detailed pharmacokinetic studies
[34] | Medical imaging | Predict molecular alterations from H&E-stained slides of colorectal cancer | A retrospective analysis of colorectal cancer patient images from five cohorts, using SL to train and validate ML models for predicting molecular features like MSI and BRAF mutations | Datasets from Northern Ireland, Germany, UK, TCGA, YCR BCIP | Demonstrated feasibility and effectiveness of SL in training AI models to predict molecular alterations in colorectal cancer using large, multicentric datasets | Expanding the SL application to other oncology areas and enhancing scalability and applicability of AI technologies in routine diagnostics.
[46] | Metaverse healthcare | Safe and equitable sharing of AI models in metaverse healthcare. A novel parameter merging approach for SL nodes. | SL nodes that train local models using private data, Swarm Network Nodes (SNN) for blockchain communication and monitoring, and a Permissioned Blockchain Network (PBN) for secure collaboration | COVID-19 dataset, PAMAP dataset | Improved accuracy and reliability of healthcare AI models in metaverse by ensuring security, fairness, and data quality distribution | Enhancing security and fairness in model-sharing processes through further integration of decentralized technologies
[53] | Disease diagnosis | Diagnose diseases from nail images. | Integrating three components: SL Node for managing insights, SNN for blockchain operations, and ML Node for training models using pre-trained bases | Four nail disease classification datasets on Google Cloud Drive | Achieved high diagnostic accuracy with a decentralized approach using transfer learning models, maintaining patient privacy | Expansion to other types of medical data and further enhancement of model training processes to maintain high accuracy with limited data
[54] | Cancer research | Using vast amounts of medical data for cancer research while adhering to privacy rules. | SL for training decentralized cancer diagnosis model across two nodes simulating different medical data sources. Data from the WDBC, WPBC, and BreakHis datasets, featuring both tumor characteristics and images, were split between nodes to reflect diverse medical scenarios | BreakHis, WDBC, WPBC | Utilized large volumes of medical data for cancer research while adhering to privacy norms showing how SL facilitates decentralized learning | Extending the decentralized model training to improve oncology research outcomes, leveraging larger and more diverse datasets
[55] | Radiology | A new capability for Human-AI partnerships. | SL with user feedback in AI model training | WDBC, WPBC, BreakHis | Introduced a system that incorporates user feedback in AI training, promoting personalized and efficient radiological assessments | Further development of Human-AI partnership capabilities, optimizing the interaction between radiologists and AI models
[56] | Eye disease detection | A novel method for detecting cataracts. | Pre-processing and data splitting, model training with the VGG-19 architecture, and Swarm Learning integration. | ODIR dataset, a collection of retinal images | Highlighted the advantages of SL over traditional centralized and federated learning systems in detecting cataracts | Expanding the model to include more diseases and larger networks for richer data diversity
### 3.2 Transportation
Innovations in communication and computing technologies have significantly
advanced the Internet-of-Vehicles (IoV). IoV is crucial to improving traffic
management, emergency responses, flow control, and efficiency in Intelligent
Transportation Systems (ITS). FL and Federated Deep Learning (FDL) have been
introduced to address privacy issues in IoV [57],[58]. Despite the benefits of
SL, there are drawbacks to using SL for collaborative Vehicle Trajectory
Prediction (VTP). For example, the need for global communication across a
large-scale network results in significant communication overhead, and the
cost of blockchain increases with the number of participants, making SL less
effective for large networks [57].
A framework that allows Vehicle Users (VUs) to cooperatively train and
aggregate models without the requirement of a central coordinator was
suggested by Lin et al. in [59]. An important consideration in the IoV
environment is the mobility of VUs, which is taken into account in the
proposed cooperative SL architecture. The authors create an incentive system
based on an iterative double auction to entice VUs to participate in the SL
process. An incentive mechanism and real-time models are included for dynamic
vehicle environments. The authors developed an optimization problem that
maximizes social welfare while achieving market equilibrium.
A novel SL approach for edge IoT contexts, communication-efficient, and
Byzantine-robust distributed swarm learning (CB-DSL). To solve issues like
data heterogeneity, communication constraints, and security concerns, CB-DSL
integrates biological intelligence and AI. To strengthen the local model and
the aggregation mechanism within the Direction Decide as a Service (DDaaS)
scheme, they used a three-layer service architecture to transfer traffic data
and control instructions, boosting forecast accuracy and real-time signal
light switching management. The CB-DSL framework is validated using real-world
healthcare datasets and simulation experiments with SUMO (Simulation of Urban
MObility) to demonstrate its effectiveness in reducing traffic congestion
compared to other existing methods.
IoV-SFDL (Internet of Vehicles-Swarm Federated Deep Learning) is a unique
framework that was presented by Wang et al.[58]. It combines SL into the FDL
framework and is specifically tailored for the IoV scenario. The goal of this
framework is to overcome the drawbacks of FDL in IoV, including significant
communication overhead, risks to data privacy, and difficulties brought on by
vehicle movement, erratic communication, and dynamic settings. The system is
more effective for IoV situations where the model training convergence speed
is accelerated through the use of an algorithm in the framework to anticipate
the credibility of weights.
The directed acyclic graph (DAG)-based Swarm Learning (DSL) framework was
created by Huang et al.[60] to address challenges such as unreliable
communications and vulnerability to malicious attacks in IoV. DSL combined
blockchain, Edge Computing (EC), and FL technologies to provide asynchronous
model training and data sharing in IoV. The authors created a Dynamic Vehicle
Association (DVA) algorithm based on DSL to handle vehicle movement and
enhance model training efficiency by maximizing the links between Vehicle
Nodes (VNs) and Road Side Units (RSUs). The DSL framework uses a method to
detect malicious attacks, ensuring security and resilience. It also introduces
a reward mechanism to encourage honest participation in model training,
promoting a collaborative and trustworthy learning environment.
Hou et al. [57] proposed Hierarchical Swarm Learning (HierSL), a novel edge-
assisted framework for Vehicle Trajectory Prediction (VTP). HierSL is proposed
to improve efficiency and security in the collaborative learning process,
particularly for large-scale edge-assisted IoV systems. HierSL reduces global
communications reliance and blockchain costs. Tests are carried out on an
actual NGSIM US-101 data set, and the outcomes demonstrate that the suggested
approach outperforms vanilla Swarm Learning and as well as centralized
learning.
Yin et al. [22] proposed a Multi-Region Asynchronous Swarm Learning (MASL)
framework. MASL is a hierarchical blockchain-powered framework for large-scale
data exchange in IoV. The blockchain, EC, and FL technologies were all merged
by MASL to ensure the anonymity and security of the sharing process. Secure
asynchronous model training and identity authentication have been accomplished
by coordinating the intra-regional (IR) and cross-regional (CR) sharing and
the non-IID data issues between regions. Furthermore, the DAG-enabled MASL is
a fully asynchronous system that is capable of responding to anomalous
vehicles on the IoVs.
Liu et al. [61] introduced a 6G-driven urban traffic congestion mitigation
solution called DDaaS. DDaaS includes a model layer for data collection,
parameter training, and congestion value prediction, a Swarm Network (SN)
layer for safe parameter transmission, and a decision-making layer for signal
light switching. Based on SUMO, simulation trials demonstrate that DDaaS can
reduce traffic congestion and achieve accurate prediction.
Autonomous driving technology has advanced significantly, but privacy concerns
arise due to the use of sensors and cameras. Mishra et al. [62] proposed an
SL-based training approach to address these concerns. By sharing model
learnings across nodes, SL protects sensitive information and reduces privacy
breaches. SL presents a promising solution to create effective and respectful
autonomous driving systems. This approach offers performance comparable to
traditional methods and outperforms other distributed machine learning
techniques like FL. Table 2 shows the main contributions of these articles.
Table 2: Swarm Learning in Transportation Ref | Application | Contributions | Methodology | Used Datasets | Key Findings | Future Work
---|---|---|---|---|---|---
[59] | IoV | A new framework that allows VUs to cooperatively train and aggregate models without the requirement for a central coordinator. An optimization problem that maximizes social welfare while achieving market equilibrium. | Cooperative SL framework with an incentive mechanism based on the mobility of vehicle users | - | Proposes a more communication-efficient method than FL; enhances social welfare and dynamic adjustment to mobility | Develop the incentive mechanism to ensure fair participation and better model aggregation methods
[63] | Edge IoT | Communication-efficient and Byzantine-robust Distributed Swarm Learning (CB-DSL) combining AI with BI principles | Evaluating the model performance under both i.i.d. and non-i.i.d. conditions and in the presence of Byzantine attacks. | CIFAR-10 and MNIST datasets | Improves local model accuracy and decision-making in traffic management; addresses local optima issues | Validate the framework in real-world settings and address more inherent challenges in edge IoT environments
[58] | IoV | IoV-SFDL: Overcomes the drawbacks of FDL in IoV, including significant communication overhead, risks to data privacy, and challenges caused by vehicle movement, erratic communication, and dynamic settings | Integrates SL into Federated Deep Learning framework | Next-Generation Simulation (NGSIM) dataset | Addresses communication overhead, improves model convergence speed in IoV contexts | Explore additional IoV-specific challenges and expand the framework to more dynamic scenarios
[60] | IoV | Improve data sharing and model training in the context of IoV | Directed Acyclic Graph-based SL (DSL) combining edge computing, FL, and blockchain | Traffic Signs Preprocessed dataset based on GTSRB | Enhances data sharing and model training; Introduces dynamic vehicle association and malicious attack detection | Develop more robust mechanisms for attack detection and introduce more adaptive algorithms for vehicle mobility
[57] | Vehicle Trajectory Prediction (VTP) in IoV | A novel edge-assisted framework for VTP | Hierarchical SL with a two-layer learning framework | NGSIM US-101 dataset | Reduces global communication reliance and blockchain costs; improves security in large-scale IoV systems | Optimize synchronization steps and system topology for better accuracy and efficiency
[22] | IoV | A secure, efficient framework for large-scale data sharing in IoVs | Multi-Region Asynchronous Swarm Learning (MASL) with hierarchical blockchain for parallel execution | Traffic Signs Pre-processed dataset based on GTSRB | Addresses scalability, security, and data heterogeneity; maintains user data privacy in large-scale data sharing | Improve the asynchronous training methods and expand blockchain integration for better data privacy and security
[61] | ITS | Direction Decide as a Service (DDaaS) to Reduce Traffic Congestion in 6G-Driven ITS. A traffic simulation and congestion prediction experiment using SUMO in Beijing, China. | Direction Decide as a Service (DDaaS) with a novel three-layer architecture incorporating SL | - | Facilitates the orderly transmission of traffic data and control instructions; improves traffic management and reduces congestion | Enhance the traffic control algorithm for more adaptive and timely decisions; expand to more complex ITS scenarios
[62] | Autonomous Driving Systems | Training autonomous driving systems | SL-based training method for privacy preservation and performance enhancement | Kitti 3d dataset | Claims superior privacy preservation and potentially better performance over traditional methods | Expand research to compare with more distributed machine learning techniques and validate in practical autonomous driving contexts
### 3.3 Industry
The Industrial Internet of Things (IIoT) is being developed using technologies
such as IoT, big data and digital twin (DT). Combining IIoT with AI algorithms
can improve productivity and interoperability, offering solutions for advanced
manufacturing systems. However, the DT technique faces challenges in capturing
dynamic industrial environments due to its data-driven nature and security and
privacy concerns[64]. SL is revolutionizing manufacturing by providing real-
time intelligent agents that improve operational efficiency by streamlining
manufacturing lines, dynamically allocating resources, and instantly resolving
problems. This approach allows for production line optimization, dynamic
resource allocation, and real-time problem solutions without centralized
control. SL is ideal for companies aiming to use Industry 4.0 and smart
manufacturing, creating more resilient and intelligent factories for the
future [65],[20]. However, there is limited research on integrating SL with
IIoT. Reliability issues in industrial systems are crucial, especially in
emergencies. Industrial environments are complex and subject to high
temperatures and noise, making them more complex than normal environments.
With automation, competition for limited communications resources increases
the unreliability of IIoT systems[64].
Pongfai et al.[17] developed a Dragonfly Swarm Learning Process (D-SLP)
algorithm for nonlinear feedback control systems, improving robustness,
performance, and stability. The D-SLP controller demonstrated superior
performance in simulations of a permanent magnet synchronous motor control
system compared to other control methods. However, the study acknowledges
limitations and suggests future work for unidirectional input constraints and
input dead zones in systems.
Using a deterministic Q-Swarm Learning Process (Q-SLP) algorithm and SL
principles, Pongfai et al.[18] created an enhanced control approach. This
method optimizes proportional integral and derivative (PID) controller
parameters, improving system performance, stability, and convergence. The
approach improves convergence time and performance by addressing shortcomings
in conventional techniques. Simulations showed superior performance and
convergence over traditional SLP, improved particle swarm optimization (IPSO),
and the whale optimization algorithm (WOA).
Pongfai and other authors created an adaptive SLP method in a different work
[66]to create the best PID controller possible for multiple-input/multiple-
output (MIMO) systems. The approach dynamically updates online weights
depending on system failures, improving PID parameter autotuning performance,
stability, and resilience. The authors evaluated the algorithm against
conventional techniques using a two-wheel inverted pendulum system as a case
study. The method could be investigated to approximate discrete-time
responses, predict behavior, and observe systems.
Sun et al.[19] proposed a new diagnostic framework for bearing faults in
rotating machinery, addressing data privacy concerns and insufficient labeled
data in factories. The framework uses convolutional neural networks and
adversarial domain networks to train local diagnostic models without sharing
data. Sun et al. in another paper [20] proposed a framework using SL to
diagnose faults in multiple components of the machinery, addressing data
privacy and insufficient data. The framework uses local diagnosis models like
AlexNet and the Chebyshev filter, enhancing efficiency and accuracy.
Xiang et al.[64] presented a ground-breaking architecture for IIoT that is
enhanced by DT technology and powered by credibility-weighted SL. Their method
tries to solve the privacy risks and significant communications costs. With
the aid of DT, they developed a DRL technique to simultaneously optimize
energy consumption and IIoT system reliability. To address issues with
operational efficiency and sustainability, they also developed and solved an
optimization problem in the recommended DT architecture to optimize system
reliability and minimize energy usage.
Wang et al.[65] have introduced a novel approach that utilizes cooperative
multi-agent SL and DT to optimize robot assembly cells and thus can be adapted
to any manufacturing environment. This model of interaction, where each
element acts as an autonomous agent, permits these agents to respond
instantaneously to issues of mechanical structure, networked software, and
hardware integration. The approach considers each component as an agent,
allowing them to interact dynamically to address mechanical structure,
software, and hardware integration changes. The framework supports dynamic
reconfiguration, ensuring efficient manufacturing systems in response to
varying product demands and production cycles.
Luo and Zhang [67] have developed a blockchain-based data management method to
ensure the integrity of the engine data, preventing tampering and deletion.
The method uses SL to verify the integrity of engine test data and protect
privacy. The integrated approach improves trustworthiness, supports
collaborative learning, and optimizes data usage while protecting privacy.
Table 3 shows the main contributions of these articles.
Table 3: Swarm Learning in Industry Ref | Application | Contributions | Methodology | Used Datasets | Key Findings | Future Work
---|---|---|---|---|---|---
[17] | Nonlinear control systems | A Dragonfly Swarm Learning Process (D-SLP) algorithm for nonlinear feedback control systems | Blendsing dragonfly algorithm behaviors with SL protocols to adaptively tune control parameters amidst system variations; A two-layer blockchain framework to ensure secure and private intra-regional and cross-regional data sharing among vehicles and base stations | - | Superior control performance in nonlinear systems compared to conventional methods | Explore application to systems with specific constraints like PAM
[18] | PID controller optimization | Use a deterministic Q-SLP algorithm to optimize and improve the PID parameter’s autotuning process | A Deterministic Q-SLP Algorithm for optimizing PID controllers, combining swarm and learning to refine control parameters KP, KI, and KD, enhancing system response and stability | CPC system | Improved convergence and performance optimization over traditional methods | Not specified
[19] | Diagnostic frameworks for rotating machinery | A new diagnostic framework for bearing faults in rotating machinery | Integrates adversarial domain networks with CNNs | CRWU, HITsz, XJTU-SY, and SCU | Increased efficiency and accuracy in fault diagnosis without compromising data privacy | Not specified
[64] | Industrial Internet of Things (IIoT) | A revolutionary architecture for IIoT powered by credibility-weighted SL and improved by DT technology | Digital Twin technology with credibility-weighted SL | real-world MNIST dataset | Enhancing IIoT system reliability and reducing energy consumption | Further address practical concerns in IIoT for operational efficiency
[65] | Reconfigurable robotic assembly cells | A method for optimizing the layout of reconfigurable robotic assembly cells in manufacturing environments | Multi-agent cooperative swarm learning with digital twin | - | Improved layout optimization and operational efficiency in manufacturing | Enhance the framework to adapt to rapid changes in manufacturing demands
[67] | Data management in engine lifecycle | A blockchain-based data management method to ensure engine data integrity | Utilizing blockchain for secure data interactions, and employs a trusted application (BCAPP) for data processing and validation | NASA open dataset | Enhanced data integrity and security throughout the engine’s lifecycle | Optimize multi-party collaborative learning and data usage
### 3.4 Robotic systems
Learning processes can be significantly accelerated when multiple robots work
together to form a swarm. Such entities could exchange learned information in
a decentralized or centralized fashion. In SL, nodes in the network pool share
locally learned models among themselves without the need for a central
authority. When using SL in networked robotic applications, a collection of
linked robots must be able to operate together or independently to complete
tasks. Rangu and Nair [68], offered a method that uses mobile agents to
execute SL on a group of robots and each learns a task. The learning process
is distributed, with a mobile agent compiling and disseminating the models
learned locally as it moves seamlessly across the network of both simulated
and actual robots. The authors demonstrated the SL approach using a mixed
group of both simulated and real robots, considering that assembling a swarm
solely of real robots would be cost-prohibitive. The application of
reinforcement learning at the local level to groups consisting of simulated,
real, and combinations of these robots has proven the viability and efficiency
of SL within a diverse network of robots.
### 3.5 Smart home
Edge Intelligence (EI) integrates edge computing and AI in smart homes, real-
time video analysis, and precision agriculture. However, centralized machine
learning models have limitations like data privacy breaches and communication
overhead[69]
SL is transforming smart home ecosystems by enabling decentralized decision-
making processes. This allows smart devices to communicate and learn from each
other’s experiences, optimizing energy consumption, security, and automation.
Smart thermostats, lighting, and appliances can adjust settings based on
occupants’ habits, ensuring comfort and energy efficiency. Swarm learning also
allows security networks to analyze data and adapt without human intervention.
Xu et al. [69] introduced a novel cooperative SL framework to overcome Central
Machine Learning issues by leveraging decentralized SL for the prediction of
thermal comfort. This approach reduces communication overhead and improves
model performance by leveraging real data from all nodes within the edge
computing network. The framework’s effectiveness was demonstrated through an
extensive empirical investigation using a Non-IID thermal comfort dataset.
Liu et al.[70] developed ADONIS, a framework for detecting abnormal behavior
in IoT devices. It uses Swarm Learning, knowledge distillation, and human-
computer interaction (HCI) to improve security and operational efficiency. The
decentralized approach reduces central node failure risk and reduces latency
and energy consumption. ADONIS can be applied to smart cities and IoVs, and
its adaptability makes it suitable for various applications. Future research
includes further enhancements and refinement of parameter aggregation methods.
### 3.6 Financial services field
By using decentralized networks for data analysis, decision-making, and risk
management, SL is completely changing the financial services industry. Swarm
learning’s decentralized nature reshapes data-driven decisions in the complex
financial landscape. Enhance investment recommendations and fraud detection
rates while protecting against single points of failure. By using SL,
financial organizations can modify their strategy in response to current
market conditions and consumer trends. John et al.[21] used SL for credit
scoring in Peer-to-Peer lending on a blockchain platform in the financial
services industry, ensuring user data privacy and secure transactions. The
decentralized model training and credit scoring process eliminate centralized
data storage risks. Future work includes testing with real-time datasets and
improving user experience.
### 3.7 Multimedia Internet of Things
By enabling the processing and dissemination of decentralized content in real-
time in environments containing IoT devices, swarm learning is transforming
the Multimedia IoT ecosystem. This method guarantees that content is
personalized for each user, minimizes latency, and maximizes network capacity
utilization. Additionally, processing data locally on devices improves
security and privacy by lowering the possibility that private information will
be hacked.
Zhang et al.[71] have improved the privacy and security of multimedia IoT
devices using Radio Frequency Fingerprinting (RFF) for identity
authentication. They integrated differential privacy, specifically the
Gaussian mechanism, into SL to protect RFF data. They also proposed a novel
node evaluation mechanism to prevent malicious nodes from affecting the
model’s accuracy and integrity. By guaranteeing the security of the underlying
IoT devices through enhanced privacy protection in SL, the research paves the
way for safe multimedia services.
### 3.8 Fake news detection
Social media has significantly impacted the distribution of information, but
the lack of systematic management has led to the spread of fake news. Machine
learning techniques like convolutional neural networks (CNN) and recurrent
neural networks (RNN)can detect fake news, but centralized detection can
violate user privacy. Decentralized methods like SL offer privacy-preserving
learning on local data, reducing hacking risks and allowing users to maintain
confidentiality without sharing data [72]. Dong et al.[72] developed Human-in-
the-loop Based Swarm Learning (HBSL), a decentralized method for detecting
fake news. HBSL uses SL and human-in-the-loop (HITL) techniques to detect fake
news across nodes, ensuring user privacy. It incorporates user feedback,
allowing models to be continuously updated. The method was validated using a
benchmark dataset (LIAR), showing its superiority over existing methods.
### 3.9 Metaverse
The Metaverse faces challenges in reliable extended reality (XR) data
transmission due to a lack of incentives and untrust among users. To address
these issues, a configurable secure resource trading mechanism based on swarm
learning is proposed in [73]. This framework includes subchains for
decentralized Intelligent Reflecting Surfaces (IRS) resource management and
intelligent allocation, a smart contract-enabled scheme, and a decentralized
federated learning-driven IRS allocation scheme. Experimental results
demonstrate the effectiveness of this configurable SL-based resource trading
for reliable XR communication.
Table 4 shows the main contributions of those articles.
Table 4: Swarm Learning Applications Field | Ref | Application | Contributions | Methodology | Used Datasets | Key Findings | Future Work
---|---|---|---|---|---|---|---
Robotic systems | [68] | Networked robotic applications | A method that uses mobile agents to execute SL on a group of robots | Mobile agents executing SL on a group of robots; Each robot learns individually, and a mobile agent facilitates the aggregation and sharing of locally learned models across the swarm | - | Demonstrated viability and efficiency of SL in a mixed robot swarm; reinforced learning applied locally to enhance task completion | Not specified
Smart home | [69] | Edge intelligent computing networks | Cooperative SL framework with cyclic ring all reduce topology for thermal comfort prediction | utilizing stochastic gradient descent within a cyclic edge intelligent computing network. | Non-IID thermal comfort dataset | Demonstrates reduced communication overhead, enhanced data privacy, and improved model performance by leveraging data from all nodes without sharing it | Extend empirical investigations, optimize model performance and handle real-world applications’ data distribution issues
| [70] | IoT, specifically abnormal behavior detection | ADONIS, a framework for detecting abnormal behavior in IoT devices | SL combined with knowledge distillation and HCI for anomaly detection in IoT devices | Traffic dataset | Enhanced security and operational efficiency in IoT networks by local data fusion and a lightweight model to accommodate resource-constrained environment | Further framework enhancement, increase communication efficiency, and refine parameter aggregation for non-IID data
Financial services field | [21] | Credit scoring | Credit scoring in Peer-to-Peer lending on a blockchain platform in the financial services industry | Lending platform on Web 3.0 that connects lenders and borrowers using blockchain technology to ensure secure, peer-to-peer transactions without intermediaries | Universal Bank dataset | Ensures data privacy and secure transactions, with model performance comparable to centralized approaches | Test model with dynamic datasets, explore other decentralized platforms (Solana, Hyperledger, Corda), and enhance user experience
Multimedia IoT | [71] | Multimedia IoT device security using RFF | Improved the privacy and security of multimedia IoT devices using RFF for identity authentication | Integration of differential privacy and a novel node evaluation mechanism in SL | RFF dataset | Enhancing privacy and security for IoT devices by protecting RFF data and making the system resilient against various cyber attacks | Future research could focus on extending these methodologies to broader IoT applications and further improving the robustness of the security measures
Fake news detection | [72] | Decentralized fake news detection | Human-in-the-loop-based swarm learning (HBSL), a decentralized method that incorporates user feedback for detecting fake news | The methodology involves local learning, collaborative model update and human feedback to enhance detection capabilities across the network through a cyclic process | LIAR dataset | Significantly improves the accuracy of fake news detection using local data and user feedback | Design detection models tailored to specific node features to enhance effectiveness
Metaverse | [73] | 6G-Metaverse XR communication | SL-based secure configurable resource trading mechanism for reliable 6G-Metaverse XR communication. | A decentralized trading framework using SL for resource management in a 6G-Metaverse environment, facilitated by IRS and blockchain technology, and Federated Learning for privacy enhancement. | Custom dataset | Effective in reliable XR communication via decentralized management and smart contract-based resource trading | Investigate customization of SL for more fine-grained communication hardware resource management and scheduling
## 4 Challenges
### 4.1 Non-IID Problem in SL
SL enables participants to register, train models, and exchange parameters
through edge nodes, ensuring data sovereignty and confidentiality. However, SL
performance is significantly affected by non-independent and identically
distributed (Non-IID ) data[48], which can lead to inconsistent model updates
and degraded aggregate performance. When data is dispersed unevenly among
various network nodes or participants, it is called the non-IID problem in SL.
This implies that distinct statistical characteristics, such as mean,
variance, and data distribution patterns, may exist in the dataset at each
node. Several factors, including variations in patient demographics, the type
of medical equipment utilized, or even the particular focus or specialization
of the medical institutes providing the data, might contribute to this
heterogeneity in the data. Non-IID data problems include quantity, label, and
feature skews. Feature skew and label skew are caused by differences in
imaging protocols or demographics, leading to inconsistencies in annotations
and Non-IID label distributions. Various strategies, including elastic weight
consolidation and batch normalization, have been proposed to address feature,
label, and quantity skew in classification tasks. However, these methods do
not fully consider label skew, which could cause suboptimal performance[8],
[74].
Two types of strategies are now being used to tackle the non-IID challenge:
algorithm-based and data-based approaches. Algorithm-based methods align local
models with global models, while data-based methods balance distribution but
require a trusted central coordinator. Furthermore, with non-IID data,
convergence problems may arise when utilizing Generative Adversarial Networks
(GAN) for data augmentation[48].
To address the non-IID problem in SL, methods must be created that can either
reduce the impact of data heterogeneity or take advantage of it to increase
the global model’s resilience and generalizability. Strategies such as
advanced aggregation techniques, personalized models, and data augmentation
can improve the robustness and generalizability of the global model[8].
Currently, effective solutions to address the non-IID problem in SL are yet to
be established[48].
### 4.2 Fairness and bias in SL
Fairness and bias in machine learning models indicate how they could perform
or reflect dominant groupings in the data in an unbalanced way. The impact of
SL on model bias and fairness has not yet been fully assessed, even though
fairness issues have been considered in the context of FL. In[26], the authors
suggested comparing SL with centralized learning and subgroup-specific model
training to investigate the fairness of SL in medical imaging tasks without
the need for additional bias mitigation techniques. To provide insight into
how SL might balance performance and fairness in healthcare applications,
their study seeks to determine if SL’s fairness features are more in line with
centralized learning or subgroup-specific training.
### 4.3 Attacks on swarm learning
SL has the potential to handle distributed large-scale data better than FL,
but it also faces significant security issues that require more scrutiny. In
the stages of SL, as shown in Fig.9 [6], different attacks can occur:
unreliable parties may compromise data during local training and before the
locally trained metadata are secured on the blockchain, it might be vulnerable
to various network attacks like Eclipse and DDoS. Furthermore, malicious
participants could introduce harmful parameters during the merging process,
potentially introducing backdoors into the global model. 1) Data poisoning
might occur in the local training phase; 2) eclipse attacks could occur in the
blockchain P2P network in the metadata upload phase; and 3) the global model
could be hacked by poisoned parameters in the parameter aggregation phase [6].
Figure 9: Attack on swarm learning
#### 4.3.1 Backdoor attacks against distributed swarm learning
Despite its privacy and decentralized training benefits, SL faces significant
security threats, such as backdoor attacks, which need to be addressed to
ensure the integrity and reliability of SL systems. Backdoor attacks in
machine learning, especially SL, manipulate data and training processes to
produce incorrect outputs. In SL, where multiple nodes collaborate, a backdoor
attack could be particularly insidious. Moreover, the decentralized nature of
SL makes detecting such attacks challenging due to the non-IID nature of real-
world data. Addressing backdoor attacks requires technological solutions,
robust security practices, and new collaborative learning approaches to ensure
integrity and trustworthiness in decentralized machine learning
environments[6],[35].
Chen et al.[6] conducted a study on security threats in SL using a pixel
pattern backdoor attack method. Their research consists of a number of studies
that evaluate the effectiveness of backdoor attacks in diverse scenarios
utilizing a variety of datasets (MNIST, CIFAR-10, SVHN). These circumstances
include varied network sizes, different data distributions (IID vs. non-IID),
distinct attack targets (single vs. multitarget), and attack continuity
policies (single-shot vs. multiple-shot). To reduce the effects of backdoor
attacks, they also suggested a number of security strategies, including L2
regularization and the addition of noise. Experimental data verify the
efficiency of these protections.
Yang et al.[35] identified a hybrid vulnerability in SL that uses backdoor and
eclipse attacks to propagate backdoors secretly. They introduced a strategy
called sample-specific eclipse (SSE) to target high data contribution nodes,
reducing attack costs and accelerating backdoor propagation. The study
investigates the use of distributed backdoor poisoning attacks in conjunction
with Eclipse assaults for the first time, showing how they can be used
together to allow backdoors to spread covertly among innocent users on the SL
network. Afterward, they suggested a fresh assault plan that concentrates on
nodes that contribute a lot of data, speeding up the spread of backdoors and
requiring fewer resources overall to be effective.
#### 4.3.2 Poisoning attack
SL faces unique challenges from poisoning attacks. Poisoning can compromise
the collective learning process, affecting model parameters and performance.
The decentralized nature of SL complicates detection, as there is no central
authority to monitor data quality or model updates. Therefore, robust
decentralized consensus mechanisms are needed to detect and mitigate poisoned
inputs [35], [71],[28]. Qi Y. et al.[28] developed strategies to prevent
poisoning attacks and ensure the integrity and security of the SL process.
Rongxuan et al.[75] introduced a Zero Trust Architecture (ZTA)-based defense
scheme for SL to combat poisoning attacks in decentralized learning
environments. It identifies a unique vulnerability where a malicious ’header’
node can compromise the model. The defense mechanism emphasizes continuous
risk calculation and anomaly detection, allowing dynamic responses to threats.
The scheme also uses Manhattan distance and accuracy differences to identify
and mitigate risks from both the header and edge nodes. The effectiveness of
the proposed defense strategy is demonstrated through systematic experiments,
proving its practical applicability in real-world scenarios.
#### 4.3.3 Eclipse attack
An Eclipse attack in SL involves an attacker controlling the network
communication between nodes. This is particularly relevant in peer-to-peer
networks where nodes share information and model updates without a centralized
authority [35]. An attacker can isolate a target node or group of nodes by
monopolizing their network connections, potentially introducing false data or
model updates [76]. This could impact the integrity of the model and degrade
performance. To protect against Eclipse attacks, robust peer discovery and
management mechanisms should be implemented, including diverse peer
connections, validating peer identities, and detecting network patterns that
might indicate control of communication channels[35].
#### 4.3.4 Inference attacks
Inference attacks aim to deduce sensitive information about the training data
used by a model, such as recovering private or sensitive attributes. They can
be used to determine if a specific data record was part of the training set,
infer specific attributes or features of data instances, or attempt to
reconstruct a model’s parameters. Inference attacks focus on extracting
information about the training data or model behavior, such as determining if
specific data were used in training or guessing private attributes based on
model outputs. Decentralized machine learning methods allow multiple nodes to
collaboratively learn a shared model without exchanging local data, typically
through blockchain technology [77]. Inference attacks exploit shared model
updates or the final model to infer properties of the training data or
identify unique characteristics of individual participants’ datasets. To
protect against inference attacks, advanced cryptographic and privacy-
preserving techniques such as homomorphic encryption, secure multi-party
computation, and differential privacy are employed. However, the balance
between privacy protection and model performance is a critical challenge in
SL[71].
#### 4.3.5 Model inversion attacks
Model inversion attacks aim directly at reconstructing the inputs used to
train the model, effectively reversing the model’s computations to approximate
or reveal the actual data. They often target models that provide detailed or
confident predictions, which can inadvertently reveal information about the
training data [78]. While inference attacks often derive indirect information
about the data or its attributes, model inversion attacks engage in a more
direct and complex effort to recreate the original training inputs themselves.
In SL, where nodes collaborate to train a model without sharing their local
datasets. The decentralized nature of SL allows each node to contribute to the
model’s learning by updating it based on local data. However, shared model
updates or predictions can leak information, potentially inferring specific
characteristics or reconstructing aspects of the original training data. To
defend against model inversion attacks, strategies such as output
perturbation, differential privacy mechanisms, access controls, and strict
query limits can be implemented[71].
## 5 Future Research
SL addresses privacy and data integration issues, but research gaps exist,
indicating potential areas for further exploration.
* •
Security and Trust: Although SL uses blockchain technology to ensure security
and trust, more investigation is required to solve potential security flaws,
such as sophisticated cyber threats and insider attacks. It is essential to
have strong trust mechanisms and security measures specifically designed for
SL networks. Swarm-FHE [79] offers a significant advancement in SL security by
integrating fully homomorphic encryption and blockchain technology. This
method ensures that collaborative model training is conducted without
compromising data, even in the presence of compromised or malicious
participants. Blockchain technology and lightweight homomorphic encryption are
also combined in a privacy-preserving SL by Li et al. [44], which promotes
model security, data privacy, and computational performance and offers a
competitive substitute for FL in remote machine learning applications [80].
* •
Dynamic Node Management: Enhancing the robustness and dependability of SL
systems may involve investigating dynamic techniques for node participation
and incentive mechanisms to guarantee nodes’ continued and productive
engagement in the swarm network.
* •
Optimizing Leader Election: The leader election process in SL can lead to
disproportionate bandwidth consumption, inefficiencies, and potential
bottlenecks, causing dissatisfaction among participants and potentially
compromising network security. To address these challenges, [23] suggested
refining the leader election mechanism for more equitable network load
distribution.
* •
Scalability and Efficiency: The ability of SL to expand across a growing
number of nodes and a variety of data formats while maintaining efficiency and
model performance should be investigated. Enhancing model aggregation
techniques and communication protocols could be the main areas of research to
facilitate widespread implementations of SL.
* •
Interoperability and Standards: For SL to succeed, standards compliance and
interoperability amongst various systems are essential. To solve issues with
data format, protocols, and compliance, research could examine methods for SL
to seamlessly integrate into existing IT systems. Qi et al.[28] developed a
blockchain twin mechanism to improve the interoperability and efficiency of SL
on different blockchains, introducing an incentive mechanism for active
participation, thus improving the overall performance and security of the SL
process.
* •
Energy Efficiency: Considering the possible magnitude of SL deployments,
especially in the context of IoT, the development of power-saving learning
algorithms is of the utmost importance. The emphasis of such research would be
on minimizing the energy usage of devices involved in the SL process, a factor
that is particularly critical for devices running on batteries or sensors
located remotely.
* •
Cross-domain Applications: Investigating the potential use of SL in diverse
sectors like healthcare, autonomous vehicles, smart cities, and manufacturing
can be extremely advantageous. Each of these areas poses distinct challenges
and demands, and customized SL approaches could result in significant
advancements in the way these sectors employ decentralized learning.
* •
Data Heterogeneity and Non-IID Data: To efficiently tackle the non-IID issue
in SL, forthcoming studies might concentrate on the creation of a hybrid model
adaptation method that merges both algorithmic innovations and robust data
management strategies. The goal of this method should be to reduce the effects
of data heterogeneity and boost the performance and unification of the global
model in a distributed environment.
* •
Advanced-Data Augmentation Techniques: Investigate the application of advanced
generative models, like variational autoencoders (VAEs) or enhanced GANs, for
the production of synthetic data samples. These samples can efficiently
supplement sparse or imbalanced datasets across different nodes, thereby
addressing the non-IID problem.
* •
Ethical AI and Fairness: As SL models become more widespread, it is crucial to
ensure that these models do not perpetuate or exacerbate biases. Research
could focus on developing fairness-sensitive algorithms that promote ethical
AI practices within SL frameworks.
* •
Resource Management: As mentioned in[23], the impact of adding more Swarm
coordinator nodes on resource overhead is negligible. However, the resource
overhead increases linearly with the number of Swarm edge nodes added,
indicating that scaling these nodes should be done with care. This observation
provides valuable guidance and actionable recommendations for developers and
researchers looking to apply SL effectively in real-world scenarios.
* •
Integrating ML into SL: The integration of ML methods into the SL framework
can introduce challenges in analyzing the specific contributions of SL to
training rate improvements. SL uses blockchain technology to synchronize model
updates amongst nodes. Although confidentiality and integrity are guaranteed,
the overhead resulting from blockchain operations (such as consensus processes
and transaction validations) may outweigh the anticipated gains in training
speed from concurrent decentralized training. Therefore, integrating ML
methods into SL may complicate the assessment of training rate improvements.
Empirical studies and benchmarking against traditional systems are needed to
assess its benefits in real-world scenarios.
## 6 Conclusion
SL is a promising advancement in decentralized machine learning that enables
efficient, secure, and privacy-preserving collaborative learning without
central data storage. This review provides invaluable information on the
advantages of SL and emphasizes how SL can facilitate safe, confidential, and
effective collaborative machine learning across dispersed networks. Highlights
the benefits of SL, such as improved data privacy, reduced risk of centralized
breaches, and the ability to learn from diverse data sources without data
transfer. SL has potential applications in healthcare, IoV, industry, etc.
However, challenges like non-IID problems, fairness, bias, and vulnerability
to attacks need to be addressed. Robust decentralized consensus mechanisms and
advanced cryptographic techniques are essential for the integrity and privacy
of SL. These research gaps offer a wide range of opportunities for researchers
interested in advancing the field of decentralized machine learning.
## References
* [1] Faisal Alsubaei, Abdullah Abuhussein, Vivek Shandilya, and Sajjan Shiva. Iomt-saf: Internet of medical things security assessment framework. Internet of Things, 8:100123, 2019.
* [2] Samira A Baho and Jemal Abawajy. Analysis of consumer iot device vulnerability quantification frameworks. Electronics, 12(5):1176, 2023.
* [3] Ali Ghubaish, Tara Salman, Maede Zolanvari, Devrim Unal, Abdulla Al-Ali, and Raj Jain. Recent advances in the internet-of-medical-things (iomt) systems security. IEEE Internet of Things Journal, 8(11):8707–8718, 2020.
* [4] Vitaly A Dovgal. Swarm learning based on the artificially intelligent edge. In CEUR Workshop Proceedingsthis, volume 3057, pages 260–265, 2021\.
* [5] Yuwei Sun, Hideya Ochiai, and Hiroshi Esaki. Decentralized deep learning for multi-access edge computing: A survey on communication efficiency and trustworthiness. IEEE Transactions on Artificial Intelligence, 3(6):963–972, 2021\.
* [6] Kongyang Chen, Huaiyuan Zhang, Xiangyu Feng, Xiaoting Zhang, Bing Mi, and Zhiping Jin. Backdoor attacks against distributed swarm learning. ISA transactions, 2023.
* [7] Chuan Ma, Jun Li, Long Shi, Ming Ding, Taotao Wang, Zhu Han, and H Vincent Poor. When federated learning meets blockchain: A new distributed learning paradigm. IEEE Computational Intelligence Magazine, 17(3):26–33, 2022.
* [8] Zheyao Gao, Fuping Wu, Weiguo Gao, and Xiahai Zhuang. A new framework of swarm learning consolidating knowledge from multi-center non-iid data for medical image segmentation. IEEE Transactions on Medical Imaging, 2022.
* [9] Feng Zhang, Yongjing Zhang, Shan Ji, and Zhaoyang Han. Secure and decentralized federated learning framework with non-iid data based on blockchain. Heliyon, 2024.
* [10] Ming Zhou, Zhen Yang, Haiyang Yu, and Shui Yu. Vdfchain: Secure and verifiable decentralized federated learning via committee-based blockchain. Journal of Network and Computer Applications, 223:103814, 2024.
* [11] Haoran Zhang, Shan Jiang, and Shichang Xuan. Decentralized federated learning based on blockchain: concepts, framework, and challenges. Computer Communications, 216:140–150, 2024.
* [12] Enrique Tomás Martínez Beltrán, Mario Quiles Pérez, Pedro Miguel Sánchez Sánchez, Sergio López Bernal, Gérôme Bovet, Manuel Gil Pérez, Gregorio Martínez Pérez, and Alberto Huertas Celdrán. Decentralized federated learning: Fundamentals, state of the art, frameworks, trends, and challenges. IEEE Communications Surveys & Tutorials, 2023.
* [13] Banhirup Sengupta, Souvik Sengupta, Susham Nandi, and Anthony Simonet-Boulogne. Blockchain and federated-learning empowered secure and trustworthy vehicular traffic. In 2022 IEEE/ACM 15th International Conference on Utility and Cloud Computing (UCC), pages 346–351. IEEE, 2022.
* [14] Mohamed Ghanem, Fadi Dawoud, Habiba Gamal, Eslam Soliman, Tamer El-Batt, and Hossam Sharara. Flobc: A decentralized blockchain-based federated learning framework. In 2022 Fourth International Conference on Blockchain Computing and Applications (BCCA), pages 85–92. IEEE, 2022.
* [15] Hussain Ahmad Madni, Rao Muhammad Umer, and Gian Luca Foresti. Blockchain-based swarm learning for the mitigation of gradient leakage in federated learning. IEEE Access, 11:16549–16556, 2023.
* [16] Stefanie Warnat-Herresthal, Hartmut Schultze, Krishnaprasad Lingadahalli Shastry, Sathyanarayanan Manamohan, Saikat Mukherjee, Vishesh Garg, Ravi Sarveswara, Kristian Händler, Peter Pickkers, N Ahmad Aziz, et al. Swarm learning for decentralized and confidential clinical machine learning. Nature, 594(7862):265–270, 2021.
* [17] Jirapun Pongfai, Wudhichai Assawinchaichote, Peng Shi, and Xiaojie Su. Novel d-slp controller design for nonlinear feedback control. IEEE Access, 8:128796–128808, 2020.
* [18] Jirapun Pongfai, Xiaojie Su, Huiyan Zhang, and Wudhichai Assawinchaichote. Pid controller autotuning design by a deterministic q-slp algorithm. IEEE Access, 8:50010–50021, 2020.
* [19] Shilong Sun, Haodong Huang, Tengyi Peng, Changqing Shen, and Dong Wang. A data privacy protection diagnosis framework for multiple machines vibration signals based on a swarm learning algorithm. IEEE Transactions on Instrumentation and Measurement, 72:1–9, 2023\.
* [20] Shilong Sun, Haodong Huang, Tengyi Peng, and Dong Wang. An improved data privacy diagnostic framework for multiple machinery components data based on swarm learning algorithm. IEEE Transactions on Instrumentation and Measurement, 2023.
* [21] Antony Prince John, Jagadhiswaran Devaraj, Lathaselvi Gandhimaruthian, and Javid Ali Liakath. Swarm learning based credit scoring for p2p lending in block chain. Peer-to-Peer Networking and Applications, 16(5):2113–2130, 2023\.
* [22] Hongbo Yin, Xiaoge Huang, Yuhang Wu, Chengchao Liang, and Qianbin Chen. Multi-region asynchronous swarm learning for data sharing in large-scale internet of vehicles. IEEE Communications Letters, 2023.
* [23] Jialiang Han, Yun Ma, and Yudong Han. Demystifying swarm learning: A new paradigm of blockchain-based decentralized federated learning. arXiv preprint arXiv:2201.05286, 2022.
* [24] Lijie Chen, Shaojing Fu, Liu Lin, Yuchuan Luo, and Wentao Zhao. Privacy-preserving swarm learning based on homomorphic encryption. In International Conference on Algorithms and Architectures for Parallel Processing, pages 509–523. Springer, 2021.
* [25] Jiajin Yang, Lixing Chen, Junhua Tang, Jianhua Li, and Wu Yang. Swarm reinforcement learning for collaborative content caching in information centric networks. In ICC 2023-IEEE International Conference on Communications, pages 384–390. IEEE, 2023.
* [26] Di Fan, Yifan Wu, and Xiaoxiao Li. On the fairness of swarm learning in skin lesion classification. In Clinical Image-Based Procedures, Distributed and Collaborative Learning, Artificial Intelligence for Combating COVID-19 and Secure and Privacy-Preserving Machine Learning: 10th Workshop, CLIP 2021, Second Workshop, DCL 2021, First Workshop, LL-COVID19 2021, and First Workshop and Tutorial, PPML 2021, Held in Conjunction with MICCAI 2021, Strasbourg, France, September 27 and October 1, 2021, Proceedings 2, pages 120–129. Springer, 2021.
* [27] Joachim L Schultze, Maren Büttner, and Matthias Becker. Swarm immunology: harnessing blockchain technology and artificial intelligence in human immunology. Nature Reviews Immunology, 22(7):401–403, 2022.
* [28] Yuxin Qi, Xi Lin, Jun Wu, and Yunyun Han. Game-aided blockchain twin for incentive and relay-free model sharing in heterogeneous chain-driven swarm learning. IEEE Systems Journal, 2023.
* [29] Hewlett Packard. Swarm learning. https://github.com/HewlettPackard/swarm-learning, 2023. Accessed: March 2023.
* [30] Caner Korkmaz, Halil Eralp Kocas, Ahmet Uysal, Ahmed Masry, Oznur Ozkasap, and Baris Akgun. Chain fl: Decentralized federated machine learning via blockchain. In 2020 Second international conference on blockchain computing and applications (BCCA), pages 140–146. IEEE, 2020.
* [31] Tian Wang, Yan Liu, Xi Zheng, Hong-Ning Dai, Weijia Jia, and Mande Xie. Edge-based communication optimization for distributed federated learning. IEEE Transactions on Network Science and Engineering, 9(4):2015–2024, 2021.
* [32] Xiaokang Zhou, Wang Huang, Wei Liang, Zheng Yan, Jianhua Ma, Yi Pan, I Kevin, and Kai Wang. Federated distillation and blockchain empowered secure knowledge sharing for internet of medical things. Information Sciences, 662:120217, 2024.
* [33] Oliver Lester Saldanha, Hannah Sophie Muti, Heike I Grabsch, Rupert Langer, Bastian Dislich, Meike Kohlruss, Gisela Keller, Marko van Treeck, Katherine Jane Hewitt, Fiona R Kolbinger, et al. Direct prediction of genetic aberrations from pathology images in gastric cancer with swarm learning. Gastric cancer, 26(2):264–274, 2023.
* [34] Oliver Lester Saldanha, Philip Quirke, Nicholas P West, Jacqueline A James, Maurice B Loughrey, Heike I Grabsch, Manuel Salto-Tellez, Elizabeth Alwers, Didem Cifci, Narmin Ghaffari Laleh, et al. Swarm learning for decentralized artificial intelligence in cancer histopathology. Nature Medicine, 28(6):1232–1239, 2022.
* [35] Zheng Yang, Gaolei Li, Jun Wu, and Wu Yang. Propagable backdoors over blockchain-based federated learning via sample-specific eclipse. In GLOBECOM 2022-2022 IEEE Global Communications Conference, pages 2579–2584. IEEE, 2022.
* [36] Xin Fan, Yue Wang, Yan Huo, and Zhi Tian. Efficient distributed swarm learning for edge computing. In ICC 2023-IEEE International Conference on Communications, pages 3627–3632. IEEE, 2023.
* [37] Xin Fan, Yue Wang, Yan Huo, and Zhi Tian. Robust distributed swarm learning for intelligent iot. In ICC 2023-IEEE International Conference on Communications, pages 973–978. IEEE, 2023.
* [38] Jinsheng Yang, Wenfeng Zhang, Zhaohui Guo, and Zhen Gao. Trustdfl: A blockchain-based verifiable and trusty decentralized federated learning framework. Electronics, 13(1):86, 2023.
* [39] Abdelaziz Salama, Syed Ali Zaidi, Des McLernon, and Mohammed MH Qazzaz. Flcc: Efficient distributed federated learning on iomt over csma/ca. In 2023 IEEE 97th Vehicular Technology Conference (VTC2023-Spring), pages 1–6. IEEE, 2023.
* [40] Laveen Bhatia and Saeed Samet. Decentralized federated learning: A comprehensive survey and a new blockchain-based data evaluation scheme. In 2022 Fourth International Conference on Blockchain Computing and Applications (BCCA), pages 289–296. IEEE, 2022.
* [41] Ehsan Hallaji, Roozbeh Razavi-Far, Mehrdad Saif, Boyu Wang, and Qiang Yang. Decentralized federated learning: A survey on security and privacy. IEEE Transactions on Big Data, 2024.
* [42] Mazin Abed Mohammed, Abdullah Lakhan, Karrar Hameed Abdulkareem, Dilovan Asaad Zebari, Jan Nedoma, Radek Martinek, Seifedine Kadry, and Begonya Garcia-Zapirain. Energy-efficient distributed federated learning offloading and scheduling healthcare system in blockchain based networks. Internet of Things, 22:100815, 2023.
* [43] Xiaobing Gan and Baoyu Xiao. Improved bacterial foraging optimization algorithm with comprehensive swarm learning strategies. In Advances in Swarm Intelligence: 11th International Conference, ICSI 2020, Belgrade, Serbia, July 14–20, 2020, Proceedings 11, pages 325–334. Springer, 2020.
* [44] Vladislav Bolshakov, Alexander Alfimtsev, Sergey Sakulin, and Nikita Bykov. Deep reinforcement ant colony optimization for swarm learning. In Advances in Neural Computation, Machine Learning, and Cognitive Research V: Selected Papers from the XXIII International Conference on Neuroinformatics, October 18-22, 2021, Moscow, Russia, pages 9–15. Springer, 2022.
* [45] Matthias Becker. Swarm learning for decentralized healthcare. Der Hautarzt, 73(4):323–325, 2022.
* [46] Guoqiang Zhang, Yueyue Dai, Jian Wu, Xiaojie Zhu, and Yunlong Lu. Swarm learning-based secure and fair model sharing for metaverse healthcare. Mobile Networks and Applications, pages 1–12, 2023.
* [47] Joachim L Schultze. Building trust in medical use of artificial intelligence-the swarm learning principle. Journal of CME, 12(1):2162202, 2023.
* [48] Zirui Wang, Shaoming Duan, Chengyue Wu, Wenhao Lin, Xinyu Zha, Peiyi Han, and Chuanyi Liu. Generative data augmentation for non-iid problem in decentralized clinical machine learning. In 2022 4th International Conference on Data Intelligence and Security (ICDIS), pages 336–343. IEEE, 2022.
* [49] Garima Aggarwal, Chun-Yin Huang, Di Fan, Xiaoxiao Li, and Zehua Wang. Demed: A novel and efficient decentralized learning framework for medical images classification on blockchain. In International Workshop on Distributed, Collaborative, and Federated Learning, pages 100–109. Springer, 2022.
* [50] Yong Li, Haichao Ling, Xianglin Ren, Chun Yu, and Tongtong Liu. Privacy-preserving swarm learning based on lightweight homomorphic encryption and blockchain technology. In 2023 IEEE 5th Eurasia Conference on IOT, Communication and Engineering (ECICE), pages 692–697. IEEE, 2023.
* [51] Xiaohan Yuan, Chuan Sun, and Shuyu Chen. Cooperative dnn partitioning for accelerating dnn-empowered disease diagnosis via swarm reinforcement learning. Applied Soft Computing, 148:110844, 2023.
* [52] Wen Pan, Geng Hu, Shaorong Li, Guoqing Li, Xiaoyu Feng, Zhifang Wu, Dong Zhang, Lizheng Qin, Xue Wang, Liang Hu, et al. Nanonitrator: novel enhancer of inorganic nitrate’s protective effects, predicated on swarm learning approach. Science Bulletin, 68(8):838–850, 2023.
* [53] Aasim Mohammed, PS Shrikanth Karthik, Razik Fatin Shariff, Tankala Sunaina, Arti Arya, and Pooja Agarwal. Privacy preserving early disease diagnosis in human nails using swarm learning. In International Conference on Information and Communication Technology for Intelligent Systems, pages 117–130. Springer, 2023.
* [54] HS Shashank, Anirudh B Sathyanarayana, Aniruddh Acharya, MR Akhil, and Sujatha R Upadhyaya. Swarm learning for oncology research. In International Conference on Multi-disciplinary Trends in Artificial Intelligence, pages 159–168. Springer, 2023.
* [55] Saptarshi Purkayastha, Rohan Isaac, Sharon Anthony, Shikhar Shukla, Elizabeth A Krupinski, Joshua A Danish, and Judy Wawira Gichoya. A general-purpose ai assistant embedded in an open-source radiology information system. In International Conference on Artificial Intelligence in Medicine, pages 373–377. Springer, 2023.
* [56] Thilak Shekhar Shriyan, Samyak Maurya, Janavi Srinivasan, Vaibhav Guruprasad Achar, Pooja Agarwal, and Arti Arya. An empirical study on privacy-preserving swarm learning for cataract detection. In International Conference on Computational & Experimental Engineering and Sciences, pages 1–13. Springer, 2023.
* [57] Xuewei Hou, Lixing Chen, Junhua Tang, Jianhua Li, and Wu Yang. Hierarchical swarm learning for edge-assisted collaborative vehicle trajectory prediction. In ICC 2023-IEEE International Conference on Communications, pages 4144–4149. IEEE, 2023.
* [58] Zhe Wang, Xinhang Li, Tianhao Wu, Chen Xu, and Lin Zhang. A credibility-aware swarm-federated deep learning framework in internet of vehicles. Digital Communications and Networks, 2023.
* [59] Shangjing Lin, Yueying Li, Bei Zhuang, Tao Ning, Ziyi Li, Chunhong Zhang, and Zheng Hu. Double auction mechanism for cooperative swarm learning in internet of vehicles. In 2022 IEEE Globecom Workshops (GC Wkshps), pages 1102–1108. IEEE, 2022.
* [60] Xiaoge Huang, Hongbo Yin, Qianbin Chen, Yu Zeng, and Jianfeng Yao. Dag-based swarm learning: a secure asynchronous learning framework for internet of vehicles. Digital Communications and Networks, 2023.
* [61] Yibing Liu, Lijun Huo, Jun Wu, and Ali Kashif Bashir. Swarm learning-based dynamic optimal management for traffic congestion in 6g-driven intelligent transportation system. IEEE Transactions on Intelligent Transportation Systems, 2023.
* [62] Abhishek Mishra, OP Joy Jefferson, Pradish Kapur, Kiran Kannur, Pooja Agarwal, and Arti Arya. Swarm learning in autonomous driving: A privacy preserving approach. In Proceedings of the 2023 15th International Conference on Computer Modeling and Simulation, pages 271–277, 2023.
* [63] Xin Fan, Yue Wang, Yan Huo, and Zhi Tian. Cb-dsl: Communication-efficient and byzantine-robust distributed swarm learning on non-iid data. IEEE Transactions on Cognitive Communications and Networking, 2023\.
* [64] Wei Xiang, Jie Li, Yuan Zhou, Peng Cheng, Jiong Jin, and Kan Yu. Digital twin empowered industrial iot based on credibility-weighted swarm learning. IEEE Transactions on Industrial Informatics, 2023.
* [65] Likun Wang, Zi Wang, Kevin Gumma, Alison Turner, and Svetan Ratchev. Multi-agent cooperative swarm learning for dynamic layout optimisation of reconfigurable robotic assembly cells based on digital twin. Journal of Intelligent Manufacturing, pages 1–24, 2024.
* [66] Jirapun Pongfai, Chrissanthi Angeli, Peng Shi, Xiaojie Su, and Wudhichai Assawinchaichote. Optimal pid controller autotuning design for mimo nonlinear systems based on the adaptive slp algorithm. International Journal of Control, Automation and Systems, 19:392–403, 2021.
* [67] Zhenjie Luo and Hui Zhang. Blockchain-based engine data trustworthy swarm learning management method. Blockchain: Research and Applications, page 100185, 2024.
* [68] Gayathri Rangu and Shivashankar B Nair. On mobile-agent-based swarm reinforcement learning in a heterogeneous robotic network. In Proceedings of the 2023 6th International Conference on Advances in Robotics, pages 1–6, 2023.
* [69] Rongxu Xu, Wenquan Jin, Anam Nawaz Khan, Sunhwan Lim, and Do-Hyeun Kim. Cooperative swarm learning for distributed cyclic edge intelligent computing. Internet of Things, 22:100783, 2023.
* [70] Yibing Liu, Xiongtao Zhang, Lijun Huo, Jun Wu, and Mohsen Guizani. Swarm learning and knowledge distillation empowered self-driving detection against threat behavior for intelligent iot. IEEE Transactions on Mobile Computing, 2023.
* [71] Lei Zhang, Lei Feng, Yue Liu, Fanqin Zhou, Boyu Liu, Zheng Jia, Yanru Wang, Hui Liu, and Wenjie Ma. Improved swarm learning with differential privacy for radio frequency fingerprinting. In 2023 IEEE International Symposium on Broadband Multimedia Systems and Broadcasting (BMSB), pages 1–5. IEEE, 2023.
* [72] Xishuang Dong, Shouvon Sarker, and Lijun Qian. Integrating human-in-the-loop into swarm learning for decentralized fake news detection. In 2022 International Conference on Intelligent Data Science Technologies and Applications (IDSTA), pages 46–53. IEEE, 2022.
* [73] Qianqian Pan, Jun Wu, Xinping Guan, and M Jamal Deen. Swarm learning irs in 6g-metaverse: Secure configurable resources trading for reliable xr communications. In GLOBECOM 2023-2023 IEEE Global Communications Conference, pages 74–79. IEEE, 2023.
* [74] Siyuan Liu, Zhiqiang Liu, Zhiwei Xu, Wenjing Liu, and Jie Tian. Hierarchical decentralized federated learning framework with adaptive clustering: Bloom-filter-based companions choice for learning non-iid data in iov. Electronics, 12(18):3811, 2023.
* [75] Rongxuan Song, Jun Wu, Qianqian Pan, Muhammad Imran, Niddal Naser, Rebet Jones, and Christos Verikoukis. Zero-trust empowered decentralized security defense against poisoning attacks in sl-iot: Joint distance-accuracy detection approach. In GLOBECOM 2023-2023 IEEE Global Communications Conference, pages 2766–2771. IEEE, 2023.
* [76] Anurag Gupta and Brian Sadler. Eclipse attack detection on a blockchain network as a non-parametric change detection problem. arXiv preprint arXiv:2404.00538, 2024.
* [77] Bosen Rao, Jiale Zhang, Di Wu, Chengcheng Zhu, Xiaobing Sun, and Bing Chen. Privacy inference attack and defense in centralized and federated learning: A comprehensive survey. IEEE Transactions on Artificial Intelligence, 2024.
* [78] Hao Fang, Yixiang Qiu, Hongyao Yu, Wenbo Yu, Jiawei Kong, Baoli Chong, Bin Chen, Xuan Wang, and Shu-Tao Xia. Privacy leakage on dnns: A survey of model inversion attacks and defenses. arXiv preprint arXiv:2402.04013, 2024.
* [79] Hussain Ahmad Madni, Rao Muhammad Umer, and Gian Luca Foresti. Swarm-fhe: Fully homomorphic encryption-based swarm learning for malicious clients. International journal of neural systems, page 2350033, 2023.
* [80] Mengxue Shang, Dandan Zhang, and Fengyin Li. Decentralized distributed federated learning based on multi-key homomorphic encryption. In 2023 International Conference on Data Security and Privacy Protection (DSPP), pages 260–265. IEEE, 2023.
|
# Learning New Tasks from a Few Examples with Soft-Label Prototypes
Avyav Kumar Singh
Department of Informatics
King’s College London
<EMAIL_ADDRESS>&Ekaterina Shutova
ILLC
University of Amsterdam
<EMAIL_ADDRESS>&Helen Yannakoudakis
Department of Informatics
King’s College London
<EMAIL_ADDRESS>
###### Abstract
It has been experimentally demonstrated that humans are able to learn in a
manner that allows them to make predictions on categories for which they have
not seen any examples Malaviya et al. (2022). Sucholutsky and Schonlau (2020)
have recently presented a machine learning approach that aims to do the same.
They utilise synthetically generated data and demonstrate that it is possible
to achieve sub-linear scaling and develop models that can learn to recognise
$N$ classes from $M<N$ training samples – aka less-than-one shot learning.
Their method was, however, defined for univariate or simple multivariate data.
We extend it to work on large, high-dimensional and real-world datasets and
empirically validate it in this new and challenging setting. We apply this
method to learn previously unseen NLP tasks from very few examples (4, 8 or
16). We first generate compact less-than-one shot representations called soft-
label prototypes which are fitted on training data, capturing the distribution
of different classes across the input domain space. We then use a modified
k-Nearest Neighbours classifier to demonstrate that soft-label prototypes can
classify data competitively, even outperforming much more computationally
complex few-shot learning methods.
## 1 Introduction
Humans have the remarkable ability to adapt knowledge gained in one domain for
use in another setting and to identify or disambiguate objects after observing
only a few examples Lake et al. (2015). This has inspired machine learning
researchers to build models able to do the same. In the rapidly growing area
of few-shot learning, researchers aim to enable correct classification in a
new task after exposure to only a few examples per class.
In natural language processing (NLP), few-shot learning techniques first
relied on interventions at the data level, such as dataset augmentation Clark
et al. (2018) or generation of adversarial examples from few-shot datasets
Miyato et al. (2016). The advent of large language models Devlin et al. (2018)
led to fine-tuning models on different target tasks, creating prompt-enhanced
few-shot datasets for training Gao et al. (2020); Schick and Schütze (2020b);
Lester et al. (2021) and the application of meta-learning algorithms Finn et
al. (2017); Snell et al. (2017) which optimise model parameters in a specific
manner so as to adapt quickly to new tasks using past experiences as a
reference.
Meta-learning has been successfully applied across a range of NLP tasks. These
included learning new classes from a few examples, e.g. in relation
classification Obamuyide and Vlachos (2019a, b) or word sense disambiguation
Holla et al. (2020) tasks; learning to quickly adapt to new domains
Nooralahzadeh et al. (2020) and to new languages Langedijk et al. (2021); van
der Heijden et al. (2021); Wang et al. (2020). An ambitious recent effort
focused on learning entirely new, previously unseen NLP tasks from as few as
4, 8 or 16 examples Bansal et al. (2019).
Despite many successes, it is important to note that meta-learning algorithms
come with limitations. While the amount of data required for training a model
for a specific few-shot task can be very little, a large amount of data from
diverse few-shot tasks is still needed to learn to generalise well to new
tasks. In practice, training time for meta-training algorithms can still run
in the order of days as they are computationally complex and resource-heavy.
In this paper, we explore and extend an alternative, more simple and yet
powerful, approach to few-shot learning, proposed by Sucholutsky et al.
(2021), which aims to offset the disadvantages mentioned above. The method
relies on representing input data sub-linearly (i.e.: representing $M$ classes
using $N$ points with $M\geq N$) and using a simple machine learning
classifier to classify data. We first generate compact, sophisticated less-
than-one shot representations called soft-label prototypes which are fitted on
the training data of the task at hand, capturing the distribution of different
classes across the input domain space. While Sucholutsky et al.’s original
approach was designed and applied to simpler univariate and multivariate data,
we extend it to work on real-world, high dimensional data. We apply the method
to learning previously unseen NLP tasks from only a few examples of that task.
We experimentally demonstrate that classification with soft-label prototypes
using a simple, modified version of k-Nearest Neighbours Cunningham and Delany
(2022) outperforms a range of strong baselines as well as the best performing
existing few-shot learning approaches in 9/16 investigated tasks.
## 2 Related Work
### 2.1 Few-shot learning in NLP
Few-shot learning techniques in NLP not exclusively requiring large language
models include data augmentation and semi-supervised learning techniques such
as augmentation with adversarial examples Miyato et al. (2016), interpolation
of training data into hidden space Chen et al. (2020) and consistency training
to make models more resistant to noise Xie et al. (2019). Semi-supervised
learning with cloze-style tasks and further fine-tuning large language models
outperformed Schick and Schütze (2020a, b) fine-tuning a large language model
with few-shot datasets using a classifier head as described by Devlin et al.
(2018). GPT-2 Radford et al. (2019) focused on achieving competitive results
with zero-shot training in multi-task settings, however, it’s successor, GPT-3
Brown et al. (2020), outperformed existing state-of-the-art models on machine
translation, cloze-style tasks and question-answering tasks with few-shot
training. Adding prompts, either in the vocabulary space Gao et al. (2020) or
the embedding space Li and Liang (2021) Lester et al. (2021), resulted in the
formation of masked language modeling tasks over the training dataset and led
to enhanced performance in few-shot settings compared to fine-tuning over a
language model.
### 2.2 Meta-Learning
In the meta-learning paradigm, the training and test sets, referred to as
$\mathcal{D}$meta-train and $\mathcal{D}$meta-test, are split into episodes.
Conceptually, each episode encompasses a task $\mathcal{T}$i and consists of a
support set $\mathcal{D}^{(i)}$support and a query set
$\mathcal{D}^{(i)}$query. Meta-learning algorithms initially fit the model on
the support set of the episode and then achieve generalisation across episodes
by optimising performance on the query sets of the episodes. For evaluation,
the model trains on the support set and checks performance on the query set
for each task $\mathcal{T}$i $\in$ $\mathcal{D}$meta-test. Meta-learning
algorithms are divided into three different areas based on their learning
methodology - model-based, metric-based and optimisation-based meta-learning
methods.
#### 2.2.1 Model Agnostic Meta-Learning
MAML Finn et al. (2017) is an optimisation-based meta-learning approach which
principally focuses on adding a generalisability objective in it’s cost
function. Optimising this generalisability objective results in a model which
learns to combine new information on top of existing knowledge efficiently.
Let us parameterise the model $f_{\theta}$ by $\theta$. As defined before,
each task $\mathcal{T}$i consists of a support set $\mathcal{D}^{(i)}$support
and a query set $\mathcal{D}^{(i)}$query. Initially, $\theta$ in
$f_{\theta}(x)$ is optimised to $\theta^{\prime}_{i}$ for a task
$\mathcal{T}$i. This initial step is referred to as inner-loop optimisation.
Mathematically, for a single step
$\mathbf{\theta^{\prime}_{i}}=\mathbf{\theta}-\alpha\nabla_{\theta}\mathcal{L}^{s}_{\mathcal{T}_{i}}(f_{\theta})$
where $\alpha$ is the learning rate and $\mathcal{L}^{s}_{\mathcal{T}_{i}}$
denotes the loss on $\mathcal{D}^{(i)}$support. As a result, each task
$\mathcal{T}$i results in an episode-specific optimised model
$f_{\theta_{i}^{\prime}}$. We can therefore capture the overall objective,
which is to have the model $f_{\theta_{i}^{\prime}}$ generalise across all
tasks in the distribution.
$\underset{\theta}{m}in\sum_{\mathcal{T}_{i}\sim
p(\mathcal{T})}\mathcal{L}^{q}_{\mathcal{T}_{i}}(f_{\theta_{i}^{\prime}})$
$=\underset{\theta}{m}in\sum_{\mathcal{T}_{i}\sim
p(\mathcal{T})}\mathcal{L}^{q}_{\mathcal{T}_{i}}(f_{\mathbf{\theta}-\alpha\nabla_{\theta}\mathcal{L}^{s}_{\mathcal{T}_{i}}(f_{\theta})})$
The losses $\mathcal{L}^{q}_{\mathcal{T}_{i}}$ are defined by the mis-
classifications on the query examples in the episode. The primary idea is to
optimise with respect to $\theta$ even though the losses are computed with
respect to $\theta_{i}$ for each episode — this enables the model to optimise
the initial parameters $\theta$ in a way as to improve generalisability. This
process is called outer-loop optimisation.
$\mathbf{\theta}\leftarrow\mathbf{\theta}-\beta\>\nabla_{\theta}\sum_{\mathcal{T}_{i}\sim
p(\mathcal{T})}\mathcal{L}^{q}_{\mathcal{T}_{i}}(f_{\theta_{i}^{\prime}})$
#### 2.2.2 Prototypical Networks
Prototypical Networks is an example of metric-based meta-learning methods.
Metric-based methods use an embedding function to get a complex, high-
dimensional representation for each instance in the support set of an episode
and then optimise the embedding function such that query examples are
correctly classified — the goal is to minimise the loss measured on the query
set between the true label and classification based on a similarity metric.
Prototypical networks, initially proposed by Snell et al. (2017), use an
embedding network, given by $f(\theta)$, to output a high-dimensional vector
that is the arithmetic mean of the data points per class in the support set of
an episode.
If we denote the sum of the instances belonging to class $c$ in the support
set of an episode by $S_{c}$ then the prototype $\mu_{c}$ is given by
$\mu_{c}=\frac{1}{|S_{c}|}\sum_{(x_{i},y_{i})\in S_{c}}f_{\theta}(x_{i})$
We can use a distance function as our similarity metric and take a softmax
over that function to get the probability of a test point beloning to a
particular class as
$\displaystyle p(y=c|x)=softmax(-d(f_{\theta}(x),\mu_{c})$
$\displaystyle=\frac{exp(-d(f_{\theta}(x),\mu_{c}))}{\sum_{c^{\prime}\in
C}exp(-d(f_{\theta}(x),\mu_{c}^{\prime}))}$
We further define the loss function as
$J(\mathbf{\theta})=-log(p(y=c^{*}|x,\theta))$
During training, each episode contains a support set and a query set
consisting of instances belonging to the same classes in the meta-training
set.
#### 2.2.3 Meta-Learning in NLP
Meta-learning has been applied in a range of natural language processing tasks
in recent years. Obamuyide and Vlachos (2019a) used meta-learning for relation
extraction in a lifelong learning setting. Obamuyide and Vlachos (2019b) also
used meta-learning for relation classification. Holla et al. (2020) use
various meta-learning algorithms to solve the problem of word-sense
disambiguation. Bansal et al. (2020) use a combination of semi-supervised and
supervised tasks to evaluate performance of their models (using meta-learning
algorithms) on tasks such as relation classification, sentiment
classification, entity-typing and natural language inference. Meta-learning
algorithms like MAML have also been applied to cross-lingual classification
tasks van der Heijden et al. (2021) and dependency parsing Langedijk et al.
(2021).
### 2.3 LEOPARD
LEOPARD Bansal et al. (2019) is a meta-learned large language model built by
producing a task-dependent final layer, typically a softmax, based on the
inputs. Given a meta-training set $\mathcal{D}_{i}^{tr}=\\{(x_{i},y_{i})\\}$
containing a task $\mathcal{T}_{i}$, the algorithm initially divides the data
with the task into it’s respective classes given by
$\mathcal{C}_{i}^{m}=\\{x_{i}|y_{i}=n\\}$ where $n\in N_{i}$ and $N_{i}$
represents the number of classes in the task at hand. It then calculates the
weights and biases of the output softmax layer as
$w_{i}^{n},b_{i}^{n}=\frac{1}{|\mathcal{C}_{i}^{n}|}\sum_{x_{j}\in\mathcal{C}_{i}^{n}}g_{\psi}(f_{\theta}(x_{j}))$
where $g_{\psi}$ is a multi-layer perceptron with two layers using a symmetric
sigmoid activation function for each layer, $w_{i}^{n}$ is a $l$-dimensional
vector and $b_{i}^{n}$ is a scalar. Combining the weights and biases for each
class, we get
$\mathbf{W_{i}}=[w_{i}^{1};...;w_{i}^{N_{i}}]\;and\;\mathbf{b_{i}}=[b_{i}^{1};...;b_{i}^{N_{i}}]$
The final equation which is used for classification is given as
$p(y|x^{*})=softmax(\mathbf{W_{i}}h_{\phi}(f_{\theta}(x^{*}))+\mathbf{b_{i}})$
where $h_{\phi}(.)$ is a separate multi-layer perceptron parameterised by
$\phi$ using all classes, given by $N_{i}$, in the task at hand. The primary
idea, as with most meta-learning approaches, is that one should not introduce
task-dependent parameters during any stage of the training process as the
model can learn task-relevant data and use it for inference rather than learn
a good initialisation point that allows the model to generalise well.
Therefore, the softmax parameters vary episodically and the existing model is
used to generate the initial parameters.
Meta-training with LEOPARD involves sampling multiple episodes at a time,
given by $G>1$, to produce a good initialisation point for the softmax
parameters. The remaining episodes are used for meta-adaptation of the model.
Mathematically, the update happens as follows, starting with
$\phi_{i}^{(0)}\leftarrow\phi_{i}$ for $s\in 0,1,...,G-1$
$\phi_{i}^{s+1}=\phi_{i}^{s}-\alpha_{s}\mathbb{E}_{\mathcal{D}_{i}^{tr}\sim
T_{i}}[\nabla_{\phi}\mathcal{L}_{i}(\\{\Theta,\phi_{i}\\},\mathcal{D}_{i}^{tr})]$
Here, $\Theta$ denote the parameters of the encoder which are fine-tuned on
meta-training, $\phi$ are the parameters of the softmax function previously
generated and $\alpha_{s}$ is the learning rate on the support set of an
episode.
Bansal et al. (2019) fine-tuned few-shot datasets over the meta-trained
LEOPARD model to evaluate performance on test datasets over NLP tasks such as
natural language inference, review classification and entity typing. They
reported that LEOPARD outperformed the baselines used by them which included
non meta-trained models as well as a prototypical network based meta-learning
model.
## 3 Method
We encode training and test instances using a large language model. We then
generate soft-label prototypes as defined by Sucholutsky et al. (2021) and
classify test encodings using those soft-label prototypes as input features.
### 3.1 Encoders
Our default base encoder is a BERT Devlin et al. (2018) model - provided by
the HuggingFace Transformers111https://huggingface.co/docs/transformers/index
library Wolf et al. (2019). We use the base cased BERT model as our default
encoder to generate input features which are used to generate and classify
with soft-label prototypes. To facilitate comparison with LEOPARD, which was
additionally meta-trained using GLUE tasks, in a separate experiment we also
use the MT-DNN model Liu et al. (2019) which has been pretrained in a multi-
task learning setting on the set of GLUE tasks. We use MT-DNN to provide the
input features for classification with soft-label prototypes. We use feature-
based encodings derived from BERT and MT-DNN without further fine-tuning.
### 3.2 Defining Soft-Label Prototypes
Sucholutsky and Schonlau (2020) define a soft label as a vector representing a
point’s "simultaneous membership to several classes". It is used in cases
where we can denote a point’s partial association to different classes. Using
this definition, they further define a soft-label prototype given by
$(\vec{X},Y)$ where $\vec{X}$ is a point in the input space (in machine
learning, an input feature vector) and $Y$ is the corresponding soft-label.
Thus, it is possible to represent $N$ classes using $M$ data points where
$M<N$ and each soft-label prototype contains class information of multiple
classes.
### 3.3 Generating Soft-Label Prototypes
As soft-label prototypes assign soft labels to every point in the input domain
(denoted in Figure 1(a)), a soft-label prototype at point $\vec{X}$
essentially represents the class distribution (determined from the training
data) at $\vec{X}$. The process of generating soft-label prototypes from
training data is a two step process detailed below.
#### 3.3.1 Finding Lines Connecting all Centroids
First, we compute the centroid of each class in the input dataset. The next
step involves finding and fitting class centroids on the minimum number of
lines - with the aim to create soft-label prototypes at the ends of each line
which capture the class distribution of all classes along that line.
Sucholutsky et al. (2021) propose three algorithms to generate the optimal set
of lines. We provide a high level overview of the two methods we used to
generate lines.
Brute Force This method first generates combinations of all possible lines
between class centroids. It then evaluates their quality by generating a score
proportional to the perpendicular distance of all centroids from the nearest
line that do not lie on it and finally uses up to $l$ lines which score the
lowest. The complexity is given by $O(n^{2l})$ where $n$ denotes the total
number of classes and $l$ denotes the total number of lines. Practically, this
method is suitable only in cases where the number of classes is small. As it
is independent of the dimensions of the input space, it also works well for
high-dimensional input space given $n$ is small, typically $\leq$10.
Recursive Regression This method clusters centroids hierarchically to group
similar centroids together. The similarity of centroids within a single
cluster is judged by how well all the centroids fit on a regression line. If
the distance of a particular centroid is beyond a pre-defined tolerance
threshold $\epsilon$ from a line, it is removed from that cluster and assigned
to another cluster. We used this method to run all our experiments. It is
computationally less expensive and performs well for high-dimensional data
spread across many more classes relative to the Brute Force method - thus
being more suitable for large, real-world data.
We depict the process of finding lines connecting all centroids in Figure
1(b).
#### 3.3.2 Defining Constraints
Once we find the lines, we use the endpoints of each line as the location of
soft-label prototypes. Therefore, we would have $2l$ prototypes if there are
$l$ lines fitted on $n$ centroids. Sucholutsky et al. (2021) mathematically
define finding the class distribution at each soft-label prototype as a
constrained optimisation problem.
Each class centroid is assigned an interval on the line segment connecting it
to other centroids on the same line. The start of this interval is marked at
the midpoint between the class centroid and the centroid of the class
preceding it along the line. The end of the interval is marked at the midpoint
between the class centroid and the succeeding class’s centroid. At any point
on the line segment, the class influences at that particular point is the sum
of all soft-label prototype class distributions weighted by the inverse of the
distance from each prototype. For this point to lie in the decision region of
one class, the weighted magnitude of that class’s influence must be higher
than that of other classes. The optimisation problem therefore consists
conceptually of two main constraints (1) the desired class has the maximum
influence amongst all classes and (2) the difference between the influence of
the desired class and the sum of the influences of all other classes is
maximised. Therefore, two soft-label prototypes are opened at the ends of each
line as denoted in Figure 1(c).
### 3.4 Classification with Soft-Label Prototypes
We use k prototypes to classify a point using a modified version of k-Nearest
Neighbours called Soft-Label Prototype k-Nearest Neighbors (SLP) Sucholutsky
and Schonlau (2020). Given $M$ soft-label prototypes which represent the input
distribution of $N$ classes, we define
$S=(X_{1},Y_{1}),...,(X_{M},Y_{M})$
to be the set of prototypes used in training where $X_{i}$ is the location of
the $i^{th}$ prototype in the input feature space and $Y_{i}$ is a matrix of
size $[N\times 1]$ denoting the soft label. We suppose that the location of
the point we want to classify is $x$. We calculate the distances from each
prototype to $x$ and denote them by the set
$D={(X_{i},x)}_{i=1,2...M}$
We then sort $S$ in ascending order of distances using $D$. Finally, we weigh
the probability distribution of the $i^{th}$ nearest prototype inversely by
it’s distance from the point and obtain
$Y^{*}=\sum_{i=1}^{k}\frac{Y_{i}}{d(X_{i},x)}$
$x$ is then assigned to the class
$C^{SLP}(x)=argmax_{j}Y^{*}_{j}$
where $Y^{*}_{j}$ is the jth element of $Y^{*}$. The decision boundaries are
described diagrammatically in Figure 1(d).
(a)
(b)
(c)
(d)
Figure 1: Generating soft-label prototypes and representing the final decision
landscape.
## 4 Tasks and Datasets
For comparison purposes, we use the same set of tasks and datasets as Bansal
et al. (2019). This set covers a variety of text classification tasks with a
large number of classes, ranging from 2 to 13. The tasks include (a) Entity
typing \- CoNLL-2003 Sang and De Meulder (2003) and MIT-Restaurant Liu et al.
(2013) datasets; (b) Review rating classification \- review ratings from
Amazon Reviews Blitzer et al. (2007) and three-way classification on the data;
(c) Text classification \- scraped social media data from
crowdflower222https://www.figure-eight.com/data-for-everyone/ which comprises
sentiment and emotion classification in a range of domains, as well as
political bias detection; and (d) Natural language inference in the scientific
domain – the SciTail dataset. Khot et al. (2018).
We use the same datasets and splits as Bansal et al. to ensure direct
comparability of approaches. At evaluation time, Bansal et al. sample an N-way
k-shot dataset $\forall\;k\in\\{4,8,16\\}$ for few-shot fine-tuning and then
evaluate the model performance on the test set of the respective task. As
performance is dependent on the data used for fine-tuning, ten test episodes
per task are sampled in total and each model is fine-tuned and evaluated on
each episode. We use accuracy as the primary metric to compare models, similar
to Bansal et al. (2019). Final performance per classification task is reported
using the mean accuracy across ten episodes as well as the standard deviation.
We ensure that the test episodes are the same as Bansal et al. by directly
using their few-shot splits available
publicly333https://github.com/iesl/leopard/. For entity typing tasks, rather
than generating a full sentence (or text) encoding as done in ratings and text
classification tasks, we obtain token-level encodings for classification.
## 5 Baselines
Our baseline selection is two-fold — essentially (1) we want to determine how
well the soft-label prototype classifier performs compared to other
classifiers given the same few-shot training data and (2) how classification
with soft-label prototypes compares against other models when both are trained
on a larger training set.
### 5.1 1-Nearest Neighbour Classifier
Our first baseline is the simple 1-Nearest Neighbour classifier for each task.
We first encode all of the data points in the episode using BERT. The test
instances are then assigned the class of their closest support example in this
embedding space. We use Euclidean distance as the distance metric.
### 5.2 Fine-tuning BERT
Another baseline we consider is fine-tuning BERT on the support set of the
test task and then evaluating the performance of that fine-tuned model on the
test set. Since this method relies on fitting a large language model on only a
few support instances, there is a higher risk of overfitting. Since 1-NN and
this baseline do not rely on additional training data, we use them to evaluate
performance in a scenario where fine-tuning is only limited to few-shot
datasets.
### 5.3 Multi-Task Learning on BERT
Bansal et al. (2019) further fine-tune the BERT model on different tasks from
the GLUE benchmark - classification, natural language inference, question
answering etc (Wang et al., 2018) prior to applying it to learn an unseen task
in a few shot fashion. Bansal et al.’s multi-task trained BERT (MT-BERT) is
similar to another implementation of a multi-task trained BERT called MT-DNN
Liu et al. (2019) which has been trained on the same set of GLUE tasks. Since
Bansal et al.’s multitask-trained BERT model is not publicly available, we use
MT-DNN as a baseline. We train a classifier head on top of MT-DNN using the
few-shot episodes and compute accuracy on the test set.
### 5.4 Prototypical Networks
We also compare to Prototypical Networks, a meta-learning algorithm, used to
meta-train the BERT encoder on the set of GLUE tasks. We use Euclidean
distance as the distance metric. Classification with soft-label prototypes
also uses Euclidean distance to weigh soft-label probabilities. This baseline
allows us to compare SLP to learning deterministic class prototypes.
### 5.5 LEOPARD
Bansal et al. (2019) report LEOPARD as the meta-learning algorithm which
presented the highest scores across most tasks for entity typing, ratings
classification as well as text classification. We compare our method to theirs
since (to the best of our knowledge) it represents the current state-of-the-
art in few-shot learning of previously unseen tasks.
## 6 Experimental setup
Sucholutsky et al.’s less-than-one-shot learning framework needed some
additional adaptation to work on real-world, complex datasets. They use the
CVXPY Diamond and Boyd (2016) library to perform computations for generating
the class distribution of a soft-label prototype, however, as their
experiments are low-dimensional ($\leq 5$), they are able to perform their
computations using a lightweight optimisation solver called ECOS Domahidi et
al. (2013). As language models have a higher output dimension (for example
BERT outputs a $768$ dimensional vector for a token) we use a more powerful,
commercially available solver called MOSEK MOSEK ApS (2019) capable of
integration within the CVXPY library to perform the required optimisations for
getting the soft-label prototypes — we found that ECOS was unable to handle
the computational complexity which resulted in our experiment program
terminating erroneously. We share the code used to run our experiments and the
detailed descriptions of our extensions to Sucholutsky et al.’s framework
publicly444https://github.com/avyavkumar/few-shot-learning-notebooks.
### 6.1 Hyperparameters
We used a held-out validation task (sentiment classification in the
electronics domain) to determine the ideal hyperparameters for soft-label
prototype classification. The hyperparameters with a major influence on
accuracy are listed below:
1) $k$ which denotes the number of nearest neighbour prototypes. We searched
among the values $k\in\\{1...10\\}$ and we found that $k=1$ provided the
highest accuracy on the validation set.
2) $\epsilon$ which is a control factor used to denote the maximum tolerance
between a centroid and the line assigned to it. The tolerance is measured as
the Euclidean distance between the line and the centroid. We use a tolerance
value of $1e-1$.
3) $l$ which denotes the maximum number of lines used to connect all
centroids. We experimented with $l\in\\{0.2n,0.4n,0.6n,0.8n,n-1\\}$ where n is
the number of centroids. For cases where it was not possible to connect $n$
centroids with $l$ lines, the optimisation process failed. Experimentally, we
found that $l=n-1$ corresponded to the highest accuracy in the validation set.
The result can also be understood intuitively - more prototype centroids can
be connected with more lines. In cases where $n$ points can be represented
with $l^{\prime}\leq l$ lines such that the tolerance is $\leq\epsilon$,
$l^{\prime}$ lines are used.
When training MT-DNN on the few-shot test episodes, we use a learning rate of
$2e-5$, a batch size of 4 and train for 100, 125 and 150 epochs for
$shots\in\\{4,8,16\\}$ respectively.
## 7 Results and discussion
Category (Classes) | Shot | 1NN | BERTbase* | SLPBERT
---|---|---|---|---
Political Bias (2) | 4 | 52.245 ± 4.348 | 54.57 ± 5.02 | 53.447 ± 3.281
| 8 | 54.568 ± 3.015 | 56.15 ± 3.75 | 55.824 ± 3.725
| 16 | 55.884 ± 2.436 | 60.96 ± 4.25 | 58.277 ± 4.128
Emotion (13) | 4 | 8.474 ± 1.028 | 09.20 ± 3.22 | 9.235 ± 2.169
| 8 | 8.418 ± 0.792 | 08.21 ± 2.12 | 8.423 ± 3.36
| 16 | 8.758 ± 0.854 | 13.43 ± 2.51 | 9.414 ± 3.398
Sentiment Books (2) | 4 | 57.56 ± 4.571 | 54.81 ± 3.75 | 59.89 ± 5.385
| 8 | 60.76 ± 4.177 | 53.54 ± 5.17 | 64.34 ± 2.565
| 16 | 60.98 ± 2.857 | 65.56 ± 4.12 | 66.36 ± 2.183
Rating DVD (3) | 4 | 37.289 ± 6.846 | 32.22 ± 08.72 | 38.158 ± 10.056
| 8 | 37.598 ± 5.364 | 36.35 ± 12.50 | 38.504 ± 9.973
| 16 | 37.461 ± 4.506 | 42.79 ± 10.18 | 36.778 ± 9.852
Rating Electronics (3) | 4 | 37.001 ± 5.164 | 39.27 ± 10.15 | 35.554 ± 9.373
| 8 | 37.646 ± 4.223 | 28.74 ± 08.22 | 43.193 ± 9.391
| 16 | 38.633 ± 3.103 | 45.48 ± 06.13 | 45.133 ± 9.754
Rating Kitchen (3) | 4 | 36.769 ± 7.738 | 34.76 ± 11.20 | 38.671 ± 9.775
| 8 | 37.435 ± 6.347 | 34.49 ± 08.72 | 45.142 ± 11.026
| 16 | 38.047 ± 3.226 | 47.94 ± 08.28 | 45.253 ± 13.455
Political Audience (2) | 4 | 51.827 ± 1.754 | 51.02 ± 1.23 | 51.305 ± 2.68
| 8 | 53.113 ± 2.314 | 50.87 ± 1.88 | 53.104 ± 3.669
| 16 | 53.287 ± 1.874 | 53.09 ± 1.93 | 53.888 ± 3.305
Sentiment Kitchen (2) | 4 | 60.17 ± 3.197 | 56.93 ± 7.10 | 61.96 ± 4.594
| 8 | 59.82 ± 2.703 | 57.13 ± 6.60 | 64.83 ± 3.983
| 16 | 61.85 ± 2.65 | 68.88 ± 3.39 | 68.21 ± 3.298
Disaster (2) | 4 | 53.629 ± 9.391 | 55.73 ± 10.29 | 52.77 ± 10.803
| 8 | 57.625 ± 7.638 | 56.31 ± 09.57 | 56.888 ± 11.139
| 16 | 60.93 ± 4.997 | 64.52 ± 08.93 | 65.907 ± 3.691
Airline (3) | 4 | 46.175 ± 4.922 | 42.76 ± 13.50 | 43.653 ± 17.948
| 8 | 48.039 ± 5.677 | 38.00 ± 17.06 | 36.343 ± 17.598
| 16 | 51.754 ± 2.954 | 58.01 ± 08.23 | 54.806 ± 16.779
Rating Books (3) | 4 | 39.324 ± 4.985 | 39.42 ± 07.22 | 45.588 ± 10.329
| 8 | 38.447 ± 4.07 | 39.55 ± 10.01 | 38.77 ± 12.485
| 16 | 41.107 ± 2.683 | 43.08 ± 11.78 | 42.105 ± 11.289
Political Message (9) | 4 | 15.044 ± 1.556 | 15.64 ± 2.73 | 15.424 ± 1.281
| 8 | 15.666 ± 1.053 | 13.38 ± 1.74 | 16.332 ± 2.056
| 16 | 14.844 ± 1.426 | 20.67 ± 3.89 | 18.185 ± 1.885
Scitail (2) | 4 | 53.66 ± 4.594 | 58.53 ± 09.74 | 52.296 ± 4.366
| 8 | 53.212 ± 3.028 | 57.93 ± 10.70 | 55.964 ± 5.705
| 16 | 54.53 ± 4.399 | 65.66 ± 06.82 | 59.675 ± 4.033
Sentiment DVD (2) | 4 | 53.78 ± 1.316 | 54.98 ± 3.96 | 56.06 ± 2.408
| 8 | 53.8 ± 2.843 | 55.63 ± 4.34 | 56.98 ± 3.299
| 16 | 54.05 ± 1.659 | 58.69 ± 6.08 | 58.95 ± 2.813
Restaurant (8) | 4 | 48.194 ± 4.881 | 49.37 ± 4.28 | 46.407 ± 7.062
| 8 | 57.36 ± 3.681 | 49.38 ± 7.76 | 57.434 ± 7.745
| 16 | 64.706 ± 2.913 | 69.24 ± 3.68 | 59.074 ± 8.34
CoNLL (4) | 4 | 45.104 ± 7.724 | 50.44 ± 08.57 | 39.35 ± 5.798
| 8 | 46.897 ± 4.183 | 50.06 ± 11.30 | 47.892 ± 14.31
| 16 | 53.969 ± 3.252 | 74.47 ± 03.10 | 54.397 ± 8.991
Table 1: Classification using SLP is indicated by SLP indexed by the encoder
used. Entries in bold highlight the highest scores. All models were only fine-
tuned and they did not rely on the meta-training set. * represents the
performance of the baselines as reported by Bansal et al. (2019) in their
paper – note that our experimental setting is identical to theirs.
Category (Classes) | Shot | LEOPARD* | MTLMT-BERT* | Proto-Net* | MTLMT-DNN | SLPMT-DNN
---|---|---|---|---|---|---
Political Bias (2) | 4 | 60.49 ± 6.66 | 54.66 ± 3.74 | 56.33 ± 4.37 | 62.46 ± 4.09 | 60.743 ± 10.656
| 8 | 61.74 ± 6.73 | 54.79 ± 4.19 | 58.87 ± 3.79 | 66.37 ± 0.36 | 65.34 ± 2.886
| 16 | 65.08 ± 2.14 | 60.30 ± 3.26 | 57.01 ± 4.44 | 66.10 ± 3.79 | 66.337 ± 0.552
Emotion (13) | 4 | 11.71 ± 2.16 | 09.84 ± 2.14 | 09.18 ± 3.14 | 12.88 ± 0.73 | 11.503 ± 1.28
| 8 | 12.90 ± 1.63 | 11.21 ± 2.11 | 11.18 ± 2.95 | 15.39 ± 1.37 | 14.151 ± 3.366
| 16 | 13.38 ± 2.20 | 12.75 ± 2.04 | 12.32 ± 3.73 | 17.45 ± 0.56 | 13.654 ± 1.943
Sentiment Books (2) | 4 | 82.54 ± 1.33 | 64.93 ± 8.65 | 73.15 ± 5.85 | 85.53 ± 0.99 | 86.07 ± 0.313
| 8 | 83.03 ± 1.28 | 67.38 ± 9.78 | 75.46 ± 6.87 | 85.43 ± 0.34 | 86.18 ± 0.326
| 16 | 83.33 ± 0.79 | 69.65 ± 8.94 | 77.26 ± 3.27 | 85.73 ± 0.71 | 86.13 ± 0.414
Rating DVD (3) | 4 | 49.76 ± 9.80 | 41.23 ± 10.98 | 47.73 ± 6.20 | 51.26 ± 8.3 | 57.183 ± 15.254
| 8 | 53.28 ± 4.66 | 45.24 ± 9.76 | 47.11 ± 4.00 | 55.39 ± 10.07 | 60.192 ± 13.438
| 16 | 53.52 ± 4.77 | 45.19 ± 11.56 | 48.39 ± 3.74 | 60.02 ± 1.8 | 66.987 ± 0.654
Rating Electronics (3) | 4 | 51.71 ± 7.20 | 41.20 ± 10.69 | 37.40 ± 3.72 | 59.82 ± 4.24 | 61.938 ± 9.047
| 8 | 54.78 ± 6.48 | 45.41 ± 09.49 | 43.64 ± 7.31 | 63.37 ± 0.78 | 65.023 ± 1.013
| 16 | 58.69 ± 2.41 | 47.29 ± 10.55 | 44.83 ± 5.96 | 62.85 ± 0.62 | 64.911 ± 0.969
Rating Kitchen (3) | 4 | 50.21 ± 09.63 | 36.77 ± 10.62 | 44.72 ± 9.13 | 54.80 ± 13.02 | 55.56 ± 18.927
| 8 | 53.72 ± 10.31 | 47.98 ± 09.73 | 46.03 ± 8.57 | 57.08 ± 11.95 | 59.631 ± 17.894
| 16 | 57.00 ± 08.69 | 53.79 ± 09.47 | 49.85 ± 9.31 | 63.95 ± 4.34 | 63.285 ± 15.583
Political Audience (2) | 4 | 52.60 ± 3.51 | 51.53 ± 1.80 | 51.47 ± 3.68 | 51.03 ± 7.22 | 53.85 ± 7.353
| 8 | 54.31 ± 3.95 | 54.34 ± 2.88 | 51.83 ± 3.77 | 56.74 ± 5.30 | 56.759 ± 5.072
| 16 | 57.71 ± 3.52 | 55.14 ± 4.57 | 53.53 ± 3.25 | 58.70 ± 2.15 | 58.529 ± 2.3
Sentiment Kitchen (2) | 4 | 78.35 ± 18.36 | 60.53 ± 9.25 | 62.71 ± 9.53 | 82.50 ± 3.49 | 86.57 ± 0.508
| 8 | 84.88 ± 01.12 | 69.66 ± 8.05 | 70.19 ± 6.42 | 86.67 ± 0.60 | 86.67 ± 0.25
| 16 | 85.27 ± 01.31 | 77.37 ± 6.74 | 71.83 ± 5.94 | 86.63 ± 0.50 | 86.85 ± 0.19
Disaster (2) | 4 | 51.45 ± 4.25 | 50.61 ± 8.33 | 50.87 ± 1.12 | 50.29 ± 3.60 | 52.17 ± 4.258
| 8 | 55.96 ± 3.58 | 54.93 ± 7.88 | 51.30 ± 2.30 | 51.61 ± 5.07 | 53.549 ± 4.306
| 16 | 61.32 ± 2.83 | 60.70 ± 6.05 | 52.76 ± 2.92 | 56.08 ± 4.77 | 54.504 ± 4.544
Airline (3) | 4 | 54.95 ± 11.81 | 46.29 ± 12.26 | 40.27 ± 8.19 | 70.41 ± 2.45 | 61.305 ± 17.572
| 8 | 61.44 ± 03.90 | 49.81 ± 10.86 | 51.16 ± 7.60 | 70.22 ± 2.47 | 70.585 ± 0.813
| 16 | 62.15 ± 05.56 | 57.25 ± 09.90 | 48.73 ± 6.79 | 71.04 ± 0.64 | 70.553 ± 0.695
Rating Books (3) | 4 | 54.92 ± 6.18 | 38.97 ± 13.27 | 48.44 ± 7.43 | 65.52 ± 8.96 | 69.728 ± 12.633
| 8 | 59.16 ± 4.13 | 46.77 ± 14.12 | 52.13 ± 4.79 | 69.28 ± 0.27 | 72.825 ± 9.53
| 16 | 61.02 ± 4.19 | 51.68 ± 11.27 | 57.28 ± 4.57 | 68.94 ± 0.78 | 72.998 ± 9.566
Political Message (9) | 4 | 15.69 ± 1.57 | 14.49 ± 1.75 | 14.22 ± 1.25 | 19.78 ± 1.10 | 20.091 ± 2.605
| 8 | 18.02 ± 2.32 | 15.24 ± 2.81 | 15.67 ± 1.96 | 21.56 ± 0.65 | 21.8 ± 1.272
| 16 | 18.07 ± 2.41 | 19.20 ± 2.20 | 16.49 ± 1.96 | 23.63 ± 2.47 | 23.046 ± 1.706
Scitail (2) | 4 | 69.50 ± 9.56 | 63.97 ± 14.36 | 76.27 ± 4.26 | 56.11 ± 3.52 | 62.546 ± 7.193
| 8 | 75.00 ± 2.42 | 68.24 ± 10.33 | 78.27 ± 0.98 | 69.36 ± 3.41 | 68.726 ± 5.825
| 16 | 77.03 ± 1.82 | 75.35 ± 04.80 | 78.59 ± 0.48 | 74.18 ± 3.38 | 73.909 ± 3.531
Sentiment DVD (2) | 4 | 80.32 ± 1.02 | 66.36 ± 7.46 | 74.38 ± 2.44 | 84.30 ± 0.96 | 84.88 ± 0.492
| 8 | 80.85 ± 1.23 | 68.37 ± 6.51 | 75.19 ± 2.56 | 84.90 ± 0.75 | 85.02 ± 0.312
| 16 | 81.25 ± 1.41 | 70.29 ± 7.40 | 75.26 ± 1.07 | 84.73 ± 0.52 | 85.05 ± 0.276
Restaurant (8) | 4 | 49.84 ± 3.31 | 50.49 ± 4.40 | 17.36 ± 2.75 | 28.37 ± 2.65 | 16.584 ± 2.874
| 8 | 62.99 ± 3.28 | 58.01 ± 3.54 | 18.70 ± 2.38 | 46.65 ± 3.80 | 18.866 ± 2.513
| 16 | 70.44 ± 2.89 | 66.16 ± 3.46 | 16.41 ± 1.87 | 58.66 ± 1.24 | 19.332 ± 2.326
CoNLL (4) | 4 | 54.16 ± 6.32 | 55.63 ± 4.99 | 32.23 ± 5.10 | 35.97 ± 2.38 | 30.434 ± 4.716
| 8 | 67.38 ± 4.33 | 58.32 ± 3.77 | 34.49 ± 5.15 | 49.04 ± 0.63 | 35.458 ± 2.937
| 16 | 76.37 ± 3.08 | 71.29 ± 3.30 | 33.75 ± 6.05 | 64.55 ± 1.14 | 35.599 ± 2.617
Table 2: Classification using soft-label prototypes is indicated by SLP
indexed by the encoder used. Entries in bold highlight the highest scores. All
models were either trained or meta-trained on the set of GLUE tasks. MTL
denotes multi-task models. * represents the performance of the baselines as
reported by Bansal et al. (2019) in their paper — note that our experimental
setting is identical to theirs.
Table 1 shows the performance of our SLP models and the respective baselines
in the setting where only the few-shot training data for the test task is
available. Classification with soft-label prototypes matches or outperforms
other baselines for the majority of the tasks considered - 22/48 tasks across
all categories report the highest performance with SLPBERT, 22/48 tasks for
BERTbase and 4/48 tasks for 1-NN. Table 2 shows that these percentages
increase for SLPMT-DNN when using an encoder trained on GLUE tasks with SLPMT-
DNN outperforming it’s counterparts on 27/48 tasks. Classification with MT-
DNN, referred to as MTLMT-DNN, is second-best, attaining superior performance
on 10/48 tasks. LEOPARD and Prototypical Networks register the highest
accuracy on 8/48 and 3/48 tasks respectively.
The performance of SLPMT-DNN contrasts strongly with that of LEOPARD – SLPMT-
DNN outperforms the latter in 36/48 tasks. SLPMT-DNN also outperforms MTLMT-
DNN in 28/48 tasks. SLPMT-DNN generally performs poorly in scientific natural
language inference tasks (SciTail) and entity typing tasks (Restaurant and
CoNLL), registering the lowest accuracy scores of the three baselines
(LEOPARD, SLPMT-DNN and MTLMT-DNN) in these sets of tasks.
In the highly-limited training data setting (Table 1), it can be observed (in
$7/16$ tasks) that if there is a consistent increase in performance from 4
shots to 8 shots, a comparable increase in performance from 8 shots to 16
shots is, however, not achieved. This effect is even more pronounced after
meta-training or training on GLUE tasks — this phenomenon is observed in
$(13/16)$ categories and a few categories within these even show a slight
decline in performance $(4/13)$.
A point to note is how closely related the performance of SLPMT-DNN and
prototypical networks is for entity typing tasks in Table 2. ProtoNet also
suffers from similar performance lapses in entity typing tasks and the
performance of ProtoNet and SLPMT-DNN is almost the same - which points to a
common factor between the two approaches which makes it unsuitable for entity
typing tasks. Both these methods use Euclidean distance as a metric for
measuring similarity — prototypical networks optimise the model by evaluating
the performance of (deterministic) support prototypes on the query set and use
Euclidean distance as the similarity measure. Classification with soft-label
prototypes similarly weighs the class distribution of the nearest $k$
prototypes by the Euclidean distance between the test point and the
prototypes. This seems to suggest that Euclidean distance is not the best
similarity measure for tasks which involve multi-dimensional natural language
embeddings. One potential reason is that for fine-grained linguistic
encodings, it is possible to find different classes within the locus
characterised by the same Euclidean distance from a point (i.e.: a circle in
two dimensions). As the number of dimensions increase, it becomes more
important to consider projection in individual dimensions as opposed to a
generalised measure of similarity - for example, using cosine distance versus
Euclidean distance for higher dimensions - for entity typing tasks.
(a)
(b)
(c)
Figure 2: Representing three classes using two soft-label prototypes
##### Depicting Three Classes using Soft Labels with Two Points
Soft-label prototypes have the ability to synthesis new data points, based on
the input data distribution, with a probability distribution of the class
labels. For instance, consider the probability distribution of two prototypes
of the category airline (derived from one episode of few-shot training data
using shots=16 using MT-DNN as the pre-trained encoder). Using the constraints
highlighted in previous sections, we generate two soft-label prototypes which
contain information about three classes depicted in Figure 2. A key point to
note from this is that we only use two soft-label prototypes to represent
three classes — therefore, this is an example of less-than-one shot
representation with respect to the class labels.
##### Training Time
Training time reduces to three operations - obtaining training encodings from
few-shot training datasets, composing lines from class centroids and
constructing soft-label prototypes at the ends of those lines. The total times
taken for each operation are different for different categories due to the
varying number of data points (given by $shots*classes$). As an example, we
take the average of time taken in each operation in all episodes of the
category airline and depict our results in Figure 3.
(a)
(b)
(c)
Figure 3: Average time taken for different training operations for the
category airline. We use MT-DNN as the encoder.
It should be noted nonetheless that times taken were very small for processes
involving the generation of lines and obtaining soft-label prototypes.
Training time, as a whole, was exceptionally low compared to meta-learning
approaches. Additionally, we did not use a GPU for our experiments.
##### Limitations
Classification with soft-label prototypes relies on creating centroids based
on the mean of the data points of a class in a training episode. The class
distribution at a particular point in the input feature space is therefore
highly dependant on the quality of encodings per class which makes the
generation of soft-label prototypes particularly susceptible to noise. This
can also lead to overfitting as the arithmetic mean of the class attempts to
incorporate all points per class, including outliers.
## 8 Conclusion
Soft-label prototypes generate a sublinear representation of data and perform
classification using a simple, computationally efficient version of k-nearest
neighbours classifier competitively with state-of-the-art optimisation-based
methods and even in very complex scenarios. Soft-label prototypes are
generated on few-shot datasets and directly utilise the knowledge captured
within a large language model. They scale well with the complexity of the
language encoder and are able to capture knowledge of large language models
and represent it well in a new target setting — providing comparable
performance with state-of-the-art approaches or improving over them. Sub-
linear representation and classification of data is also a further step
towards algorithmic less-than-one shot learning, wherein we disambiguate $M$
classes using only $N$ instances in total where $M\geq N$.
## References
* Bansal et al. (2019) Trapit Bansal, Rishikesh Jha, and Andrew McCallum. 2019. Learning to few-shot learn across diverse natural language classification tasks.
* Bansal et al. (2020) Trapit Bansal, Rishikesh Jha, Tsendsuren Munkhdalai, and Andrew McCallum. 2020. Self-supervised meta-learning for few-shot natural language classification tasks.
* Blitzer et al. (2007) John Blitzer, Mark Dredze, and Fernando Pereira. 2007. Biographies, Bollywood, boom-boxes and blenders: Domain adaptation for sentiment classification. In _Proceedings of the 45th Annual Meeting of the Association of Computational Linguistics_ , pages 440–447, Prague, Czech Republic. Association for Computational Linguistics.
* Brown et al. (2020) Tom B. Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, Sandhini Agarwal, Ariel Herbert-Voss, Gretchen Krueger, Tom Henighan, Rewon Child, Aditya Ramesh, Daniel M. Ziegler, Jeffrey Wu, Clemens Winter, Christopher Hesse, Mark Chen, Eric Sigler, Mateusz Litwin, Scott Gray, Benjamin Chess, Jack Clark, Christopher Berner, Sam McCandlish, Alec Radford, Ilya Sutskever, and Dario Amodei. 2020. Language models are few-shot learners.
* Chen et al. (2020) Jiaao Chen, Zichao Yang, and Diyi Yang. 2020. Mixtext: Linguistically-informed interpolation of hidden space for semi-supervised text classification.
* Clark et al. (2018) Kevin Clark, Minh-Thang Luong, Christopher D. Manning, and Quoc V. Le. 2018. Semi-supervised sequence modeling with cross-view training.
* Cunningham and Delany (2022) Pá draig Cunningham and Sarah Jane Delany. 2022. k-nearest neighbour classifiers - a tutorial. _ACM Computing Surveys_ , 54(6):1–25.
* Devlin et al. (2018) Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2018. Bert: Pre-training of deep bidirectional transformers for language understanding.
* Diamond and Boyd (2016) Steven Diamond and Stephen Boyd. 2016. CVXPY: A Python-embedded modeling language for convex optimization. _Journal of Machine Learning Research_ , 17(83):1–5.
* Domahidi et al. (2013) Alexander Domahidi, Eric Chu, and Stephen Boyd. 2013. Ecos: An socp solver for embedded systems. In _2013 European Control Conference (ECC)_ , pages 3071–3076.
* Finn et al. (2017) Chelsea Finn, Pieter Abbeel, and Sergey Levine. 2017. Model-agnostic meta-learning for fast adaptation of deep networks.
* Gao et al. (2020) Tianyu Gao, Adam Fisch, and Danqi Chen. 2020. Making pre-trained language models better few-shot learners.
* van der Heijden et al. (2021) Niels van der Heijden, Helen Yannakoudakis, Pushkar Mishra, and Ekaterina Shutova. 2021. Multilingual and cross-lingual document classification: A meta-learning approach.
* Holla et al. (2020) Nithin Holla, Pushkar Mishra, Helen Yannakoudakis, and Ekaterina Shutova. 2020. Learning to learn to disambiguate: Meta-learning for few-shot word sense disambiguation.
* Khot et al. (2018) Tushar Khot, Ashish Sabharwal, and Peter Clark. 2018. Scitail: A textual entailment dataset from science question answering. In _AAAI_.
* Lake et al. (2015) Brenden M. Lake, Ruslan Salakhutdinov, and Joshua B. Tenenbaum. 2015. Human-level concept learning through probabilistic program induction. _Science_ , 350(6266):1332–1338.
* Langedijk et al. (2021) Anna Langedijk, Verna Dankers, Phillip Lippe, Sander Bos, Bryan Cardenas Guevara, Helen Yannakoudakis, and Ekaterina Shutova. 2021. Meta-learning for fast cross-lingual adaptation in dependency parsing.
* Lester et al. (2021) Brian Lester, Rami Al-Rfou, and Noah Constant. 2021. The power of scale for parameter-efficient prompt tuning.
* Li and Liang (2021) Xiang Lisa Li and Percy Liang. 2021. Prefix-tuning: Optimizing continuous prompts for generation.
* Liu et al. (2013) Jingjing Liu, Panupong Pasupat, Scott Cyphers, and Jim Glass. 2013. Asgard: A portable architecture for multilingual dialogue systems. In _2013 IEEE International Conference on Acoustics, Speech and Signal Processing_ , pages 8386–8390.
* Liu et al. (2019) Xiaodong Liu, Pengcheng He, Weizhu Chen, and Jianfeng Gao. 2019. Multi-task deep neural networks for natural language understanding.
* Malaviya et al. (2022) Maya Malaviya, Ilia Sucholutsky, Kerem Oktar, and Thomas L Griffiths. 2022. Can humans do less-than-one-shot learning? _arXiv:2202.04670_.
* Miyato et al. (2016) Takeru Miyato, Andrew M. Dai, and Ian Goodfellow. 2016. Adversarial training methods for semi-supervised text classification.
* MOSEK ApS (2019) MOSEK ApS. 2019. _The MOSEK optimization toolbox for MATLAB manual. Version 9.3._
* Nooralahzadeh et al. (2020) Farhad Nooralahzadeh, Giannis Bekoulis, Johannes Bjerva, and Isabelle Augenstein. 2020. Zero-shot cross-lingual transfer with meta learning.
* Obamuyide and Vlachos (2019a) Abiola Obamuyide and Andreas Vlachos. 2019a. Meta-learning improves lifelong relation extraction. In _Proceedings of the 4th Workshop on Representation Learning for NLP (RepL4NLP-2019)_ , pages 224–229, Florence, Italy. Association for Computational Linguistics.
* Obamuyide and Vlachos (2019b) Abiola Obamuyide and Andreas Vlachos. 2019b. Model-agnostic meta-learning for relation classification with limited supervision. In _Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics_ , pages 5873–5879, Florence, Italy. Association for Computational Linguistics.
* Radford et al. (2019) Alec Radford, Jeff Wu, Rewon Child, David Luan, Dario Amodei, and Ilya Sutskever. 2019. Language models are unsupervised multitask learners.
* Sang and De Meulder (2003) Erik F. Tjong Kim Sang and Fien De Meulder. 2003. Introduction to the conll-2003 shared task: Language-independent named entity recognition.
* Schick and Schütze (2020a) Timo Schick and Hinrich Schütze. 2020a. Exploiting cloze questions for few shot text classification and natural language inference.
* Schick and Schütze (2020b) Timo Schick and Hinrich Schütze. 2020b. It’s not just size that matters: Small language models are also few-shot learners.
* Snell et al. (2017) Jake Snell, Kevin Swersky, and Richard S. Zemel. 2017. Prototypical networks for few-shot learning.
* Sucholutsky et al. (2021) Ilia Sucholutsky, Nam-Hwui Kim, Ryan P. Browne, and Matthias Schonlau. 2021. One line to rule them all: Generating LO-shot soft-label prototypes. In _2021 International Joint Conference on Neural Networks (IJCNN)_. IEEE.
* Sucholutsky and Schonlau (2020) Ilia Sucholutsky and Matthias Schonlau. 2020. ’less than one’-shot learning: Learning n classes from m.
* Wang et al. (2018) Alex Wang, Amanpreet Singh, Julian Michael, Felix Hill, Omer Levy, and Samuel Bowman. 2018. GLUE: A multi-task benchmark and analysis platform for natural language understanding. In _Proceedings of the 2018 EMNLP Workshop BlackboxNLP: Analyzing and Interpreting Neural Networks for NLP_ , pages 353–355, Brussels, Belgium. Association for Computational Linguistics.
* Wang et al. (2020) Zirui Wang, Zachary C. Lipton, and Yulia Tsvetkov. 2020. On negative interference in multilingual models: Findings and a meta-learning treatment. In _Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)_ , pages 4438–4450, Online. Association for Computational Linguistics.
* Wolf et al. (2019) Thomas Wolf, Lysandre Debut, Victor Sanh, Julien Chaumond, Clement Delangue, Anthony Moi, Pierric Cistac, Tim Rault, Rémi Louf, Morgan Funtowicz, Joe Davison, Sam Shleifer, Patrick von Platen, Clara Ma, Yacine Jernite, Julien Plu, Canwen Xu, Teven Le Scao, Sylvain Gugger, Mariama Drame, Quentin Lhoest, and Alexander M. Rush. 2019. Huggingface’s transformers: State-of-the-art natural language processing.
* Xie et al. (2019) Qizhe Xie, Zihang Dai, Eduard Hovy, Minh-Thang Luong, and Quoc V. Le. 2019. Unsupervised data augmentation for consistency training.
|
# From Questions to Insightful Answers: Building an Informed Chatbot for
University Resources
Subash Neupane1, Elias Hossain1, Jason Keith2, Himanshu Tripathi1,
Farbod Ghiasi1, Noorbakhsh Amiri Golilarz1Amin Amirlatifi2, Sudip Mittal1,
Shahram Rahimi1 1 Department of Computer Science & Engineering
Mississippi State University
{sn922, mh3511, ht557, fg289}@msstate.edu,{amiri, mittal,
<EMAIL_ADDRESS>2 Dave C. Swalm School of Chemical Engineering
Mississippi State University
{keith<EMAIL_ADDRESS>
###### Abstract
This paper presents BarkPlug v.2, a Large Language Model (LLM)-based chatbot
system built using Retrieval Augmented Generation (RAG) pipelines to enhance
the user experience and access to information within academic settings. The
objective of BarkPlug v.2 is to provide information to users about various
campus resources, including academic departments, programs, campus facilities,
and student resources at a university setting in an interactive fashion. Our
system leverages university data as an external data corpus and ingests it
into our RAG pipelines for domain-specific question-answering tasks. We
evaluate the effectiveness of our system in generating accurate and pertinent
responses for Mississippi State University, as a case study, using
quantitative measures, employing frameworks such as Retrieval Augmented
Generation Assessment (RAGAS). Furthermore, we evaluate the usability of this
system via subjective satisfaction surveys using the System Usability Scale
(SUS). Our system demonstrates impressive quantitative performance, with a
mean RAGAS score of 0.96, and satisfactory user experience, as validated by
usability assessments.
###### Index Terms:
Chatbot, LLM, RAG, University resources, information access
## I Introduction
Colleges and universities invest significant time and resources into enhancing
their websites to effectively communicate crucial information about the
institution and available campus resources. The institutional website serves
as its “virtual face”, the face it has chosen to present to the online world,
including potential and current students, faculty, parents, alumni and general
users [1]. Although these websites offer comprehensive information, they lack
the capability to provide personalized responses to user queries. For
instance, when a prospective student needs details about submitting ACT
scores, wants to know their tuition and fees, or is unsure which parent’s
information to use on their FAFSA application, they must navigate through
multiple webpages to find answers. This process frequently requires a
considerable amount of time. Yet, at times, users’ queries are left unanswered
due to either unclear information or lack of personal interaction.
Various campus resources and services, such as academic departments, career
centers, admissions, registration, scholarships, and financial aid, are
available to assist students with both academic and non-academic queries.
These resources are equipped with dedicated officers who provide guidance to
students. However, they are constrained by service-time limitation (may only
be available during specific working hours) and may require an appointment,
which might not always accommodate busy student schedules. Additionally,
delays in responses and longer wait times, particularly during application
periods, can diminish prospective students’ interest in the institution,
ultimately affecting university revenue.
Figure 1: Comparative example of completion (response generation) without
using the RAG approach versus using the RAG approach for a given user prompt
related to specific individual at Mississippi State University.
To address these challenges, universities are currently employing
conversational agents, also known as chatbots, to offer support to users.
Chatbots are _“software systems that mimic interactions with real people”_[2]
by engaging in conversation through natural language using machine learning
technology, specifically Natural Language Processing (NLP)[3]. For instance,
Arizona State University (ASU) developed a chatbot named _Sunny_ [4] with the
intentional design to offer emotional support to students, alongside providing
information regarding ASU. Sunny efficiently addresses frequently asked
questions such as inquiries about financial aid, academic advisors, and
accessing ASU email accounts. Similarly, Georgia State University (GSU)
introduced a virtual assistant named _Pounce_ [5] to tackle obstacles to
enrollment faced by students transitioning from high school to college. These
obstacles encompassed tasks like financial aid applications, document
submissions, immunization records, placement exams, and class registration.
GSU reported a notable 22% reduction in summer melt due to Pounce’s
assistance. Beyond admissions and enrollment, universities are increasingly
deploying chatbots to aid students in their academic pursuits and campus life.
One notable example is the chatbot _Beacon_ [6] developed by Staffordshire
University. Beacon offers personalized and responsive support, including
information on timetables and answers to frequently asked questions.
Apart from the higher education sector, chatbots are being increasingly
adopted across a diverse range of industries and contexts including healthcare
[7, 8], cybersecurity[9, 10], retail[11], and hospitality[12] among others due
to their ability to emulate human conversations, automate services, and reduce
human workload. The meteoric rise in interest in using chatbots by industries
at present is attributed to the overwhelming success of
ChatGPT.111https://chatgpt.com/. In fact, the global chatbot market size was
valued at 5.39 billion dollars in 2023 which is expected to reach 42.83
billion dollars by 2033, according to a market research report [13] published
by Spherical Insights & Consulting.
In this paper, we introduce BarkPlug v.2\- the second iteration of a chatbot
system built for Mississippi State University (MSU), with an architecture that
can be applied to any university setting. This system serves as an assistive
tool, capable of leveraging all university resources to provide more
intelligent analyses of university related content. It responds interactively,
making access to relevant information easier for users. Compared to the other
educational chatbots, BarkPlug v.2 is more comprehensive and covers several
aspects of university functions and services. The development of our chatbot
utilizes Retrieval Augmented Generation (RAG) [14] techniques for response
generation. RAG pipelines consist of two vital components: a retriever and a
generator based on a Large Language Model (LLM). We opt for the RAG approach
because pretrained LLMs, such as _gpt-3.5-turbo_ , alone cannot adequately
answer domain-specific questions or perform well on data outside their
training dataset, often resulting in hallucinated outputs. Figure 1 provides a
comparative overview of the response generation for a given user prompt
without RAG and with RAG. As is evident from Figure 1, ChatGPT clearly fails
to answer domain-specific questions, while BarkPlug v.2, which uses the RAG
approach, can accurately answer a user prompt. In our pipeline we utilize
various campus resources such as information on _academic departments,
financial aid, admission, scholarships, dining, housing, and health center_ as
a corpus of external data source for retrieval.
BarkPlug v.2 project’s key contributions include:
* •
Design and development of a comprehensive chatbot system proficient in
responding to a wide spectrum of queries pertaining to the diverse array of
campus resources available at Mississippi State University.
* •
Demonstrating the possibility of promptly providing personalized, real-time
information, thereby augmenting user engagement through the continuous
availability of the chatbot.
* •
Showcasing the application’s effectiveness through rigorous evaluation,
validating its performance and user satisfaction.
The rest of this article is divided into five connected sections. In Section
II, we present the background and related work. Following that, in Section
III, we explain the architecture and methodology. Section IV provides a
detailed analysis of experimental results. Moving on to Section V, we provide
implementation details and discuss the limitations and future works. Finally,
we conclude our paper.
## II BACKGROUND AND RELATED WORK
In this section, we briefly look into the pre-requisite background followed by
exploring related research that focuses on development of chatbot applications
in educational context.
### II-A Large Language Models (LLMs)
Large Language Models (LLMs) like GPT-4, LLAMA3, and PaLM are at the forefront
of computational linguistics, powered by Transformer-based architectures [15]
with vast parameter spaces, often exceeding hundreds of billions. These models
rely on the self-attention mechanism within Transformers. LLMs excel in
understanding and generating human language, reshaping the Natural Language
Processing (NLP) landscape. They leverage various Transformer architectures
and pre-training objectives, including decoder-only models (e.g., GPT2, GPT3),
encoder-only models (e.g., BERT, RoBERTa), and encoder-decoder architectures
like BART.
These architectures efficiently process sequential data, capturing intricate
dependencies within text while enabling effective parallelization. LLMs
integrate prompting or in-context learning, enhancing text generation by
incorporating contextual information. This capability facilitates coherent and
contextually relevant responses, fostering interactive question-and-answer
engagements [16].
### II-B Retrieval Augmented Generation (RAG)
Pre-trained Large Language Models (LLMs) are proficient at acquiring extensive
knowledge but lack memory expansion or revision capabilities, leading to
errors like hallucinations. To address this, hybrid approaches like Retrieval
Augmented Generation (RAG) have emerged [17, 18, 14].
RAG integrates input sequences with information retrieved from corpus of an
external data source, enriching context for sequence generation. The retriever
component selects the top $k$ text passages relevant to the input query,
augmenting the model’s understanding and enhancing output sequence generation.
This process is governed by the equation: $p_{n}(z|x)$ where $p_{n}$
represents the retriever component with parameters $n$ (number of documents or
passages a user wants to retrieve), selecting relevant passages $z$ from the
knowledge database given input $x$.
### II-C Related Works
Recent research on educational chatbots explores various areas such as
application fields, objectives, learning experiences, design approaches,
technology, evaluation methods, and challenges. Studies have shown that
educational chatbots are used in health advocacy, language learning, and self-
advocacy. They can be flow-based or powered by AI, facilitating answering
Frequently Asked Questions (FAQs), performing quizzes, recommending
activities, and informing users about various events [19][20]. Chatbots have
been found to improve students’ learning experiences by motivating them,
keeping them engaged, and providing immediate online assistance [21].
Additionally, chatbots make education more accessible and available [20].
Design aspects such as the role and appearance of chatbots are significant
factors that affect their effectiveness as educational tools [22]. Chatbots
are designed using various methods, including flow-based and AI-based
approaches, and can incorporate speech recognition capabilities [23].
Technologies used to implement chatbots include Dialogflow
222https://cloud.google.com/dialogflow and ChatFuel333https://chatfuel.com/
among others. These technologies impact chatbot performance and quality,
necessitating careful selection during design and development [24]. Flow-based
chatbots, such as those powered by Dialogflow, can provide structured
interactions based on predetermined scripts, while AI-based chatbots leverage
machine learning and NLP to offer more flexible and dynamic interactions.
In regards to assessment of the effectiveness of educational chatbots,
evaluation methods such as surveys, experiments, and evaluation studies are
used, measuring acceptance, motivation, and usability [25][24][26]. Surveys
gather feedback from students and educators regarding their experiences with
chatbots, while experiments may involve testing chatbots in controlled
settings to measure their impact on learning outcomes. Evaluation studies
provide deeper insights into how chatbots perform in various educational
scenarios and how users perceive their usefulness. In terms of interaction
styles, research examines whether chatbots are user-driven or chatbot-driven,
depending on who controls the conversation [23][19]. Chatbot-driven
interactions often involve more automated and guided conversations, while
user-driven interactions prioritize user input. Striking a balance between
these approaches can result in more natural and effective communication.
However, it’s important to acknowledge that achieving this balance
necessitates addressing substantive challenges to optimize the chatbot’s
applicability across diverse contexts, including the field of education.
Ethical considerations, such as compliance with educational norms and
safeguarding user data, assume paramount importance [21, 27]. Leveraging novel
methodologies in their development, we aim to navigate these issues more
effectively. Moreover, we confront persistent programming complexities and the
importance of sustaining chatbot utility amidst educational evolution [28,
29]. By harnessing advancements in technology, we endeavor to bolster our
chatbots’ resilience to these challenges. These collaborative endeavors offer
a strategic direction, utilizing technological advancements to refine
educational chatbots. Furthermore, the language model (conversational chatbot)
contends with conceptual challenges essential for its operational efficacy,
requiring careful research focus.
Insights from studies such as [30] reveal how language models such as BERT
establish relationships between expressions and queries, shedding light on
chatbot interaction styles and response quality. This study contributes to
understanding how advanced language models can be integrated into chatbots for
more nuanced and context-aware responses. [31] discusses the gap between
chatbot responses and user intent, which can be more pronounced in complex
university settings. Chatbots in academic environments often encounter
questions that require a deep understanding of the subject matter and context.
This necessitates the use of sophisticated models that can handle intricate
queries and provide accurate and relevant responses. [30, 31] underscores the
importance of understanding and controlling the context of language models,
thereby guiding our efforts to integrate advanced language models into
chatbots for more nuanced and context-aware responses. Their context-aware
approach has been instrumental in shaping our chatbot’s unique capabilities.
The integration of chatbots within university platforms and metaverses offers
promising avenues for enhancing user experience and facilitating learning. For
instance, [32] demonstrate how chatbots in metaverse-based university
platforms offer instant, personalized support for tasks such as course
navigation and answering FAQs, leveraging NLP and machine learning to
streamline information dissemination and reduce administrative burdens. This
kind of integration not only facilitates academic processes but also helps in
addressing students’ concerns promptly, ensuring smoother academic
experiences. In specific university contexts, [33] develops a question-
answering system for an Indonesian university admissions using Sequence-to-
sequence learning. This system demonstrates how chatbots can be employed in
specialized areas to address particular challenges, such as providing guidance
during the admissions process. Similarly, [34] introduce a dynamic chatbot
enhancing student interaction by covering admissions, academic assistance, and
event information, prioritizing user feedback for accuracy, reliability, and
safety. Frequent updates ensure that chatbots maintain relevance and continue
to serve as effective tools for student support. Moreover, [35] presents
TutorBot+, which employs LLMs like ChatGPT to offer feedback in programming
courses. Their quasi-experimental research shows positive impacts on students’
computational reasoning abilities, illustrating the potential of such
interventions in education. TutorBot+ demonstrates the benefits of integrating
advanced AI models to support students in understanding complex programming
concepts, potentially transforming how computational subjects are taught.
## III BarkPlug v.2 Architecture & Methodology
Figure 2: Overview of BarkPlug v.2’s two phase architecture. The first phase
_Context retrieval_ is responsible to retrieve relevant documents based on the
user prompt. The second phase, _Completion_ responsible of generating
personalized responses utlilizing retrieved documents as context along with
user prompt.
This section describes the architecture of BarkPlug v.2 consisting two main
phases: _context retrieval_ and _completion_ as shown in Fig. 2. The first
phase retrieves documents relevant to the user prompt. The second phase
utilizes these retrieved documents and user prompts to generate contextual
responses referred to as completions. The subsequent subsections will provide
a comprehensive breakdown of each phase, discussing their functionalities and
methodologies.
Figure 3: Similarity score threshold retrieval.
### III-A Context Retrieval
Retrieval in BarkPlug v.2 involves obtaining pertinent information from an
external data source to establish context for completions. This phase takes a
prompt (query) as an input and produces chunks of documents relevant to the
prompt. In our context, the external data is the university resources
available through Mississippi State University’s Website
444https://www.msstate.edu/. We curate data of 42 different department within
the university using web crawlers. These include _academic departments,
financial aid, admissions, housing, dinning services, library, health center_
etc. Inclusion of these campus resources as external data source is to ensure
BarkPlug v.2 is comprehensive enough to answer diverse question. For example,
a user might ask a question such as _“What are the funding opportunities
available for graduate students in the CSE department?"_. Followed by the
question _“Who do I contact if I have additional questions about majors or
attending MSU?"_ To answer the first question the system should have
information about funding opportunities within CSE, whereas to answer the
second question, information about _academic counselors_ should be present in
the external data source. Please refer to Section IV-A for detailed
explanation on how we curate these data, prepare and process for this phase.
The first step in this phase is to transform the external data source. This
step relies on two important components: an _embedding model_ and a _vector
database_. Embeddings refers to functions that map or transforms raw input
data to low-dimensional vector representations while retaining important
semantic information about the inputs [36]. On the other hand, the vector
database is a type of database that stores data as high-dimensional vectors
that are usually generated by applying embedding functions to the raw data
[37], such as text in our case. It supports complex and unstructured data and
allows fast and accurate similarity search and retrieval. BarkPlug v.2
utilizes an embedding model to vectorize the external data sources, in
particular, we leverage a _text-embedding-3-large_ model managed through API
calls. These vectors are subsequently stored in Chroma DB [38] an in-memory
vector database.
For efficient context retrieval process we use vector store-backed retriever
technique provided by LangChain [39]. It utilizes vector store to retrieve
documents. In general the vector store retriever uses two types of search
methods including Maximum Marginal Relevancy (MMR) and Similarity Search. In
this work, we have leveraged _similarity score threshold retrieval_ , depicted
in Fig.3 as our searching mechanism. This searching strategy returns all the
possible results to a user prompt based on the minimum similarity threshold we
define. The output of this phase is the relevant documents that serves as
context for subsequent completions phase.
### III-B Completion
The second phase is the _completion_ which is also referred to as _response
generation_. We utilize a gpt based LLM for completions, in particular, we
leverage OpenAI’s _gpt-3.5-turbo_ as our base generator model. The input to
the generator consists of retrieved document chunks and the user prompt. Then,
the generator model, or LLM, uses this information as a guideline to produce
acurate and relevant responses (completions). Response generation in BarkPlug
v.2 is managed through OpenAI’s API calls.
An example of completion for the user prompts _"What are the funding
opportunities available for graduate students in the CSE department?"_ and
_"Who do I contact if I have questions about attending MSU_ can be seen in
Fig. 4. In the first example, a user seeks information about funding
opportunities within the CSE department for graduate studies. The completion,
in this case, is a contextualized response generated by combining relevant
information retrieved from campus resources, such as the academic department
(CSE), with the user prompt. In the second example, the user is seeking
information about admission to Mississippi State University. To answer this,
the system must possess knowledge about campus resources, such as admissions
and scholarships. As mentioned earlier, BarkPlug v.2 not only has knowledge
about academic departments but also other campus resources, including
admissions and scholarships. By utilizing admissions and scholarships as
external data sources, the system returns the relevant document reflecting
information about admissions and scholarships, along with contact details, as
a completion for the user prompt.
Prompt: What are the funding opportunities available for graduate students in
the CSE department?
Completion: Graduate students in the Computer Science and Engineering (CSE)
department at Mississippi State University have access…funding
opportunities…graduate assistantships, research grants, fellowships,
scholarships, and industry partnerships…research grants cover project
expenses.
Prompt: Who do I contact if I have questions about attending MSU?
Completion: For queries on admissions at Mississippi State University, reach
out to the Office of Admissions and Scholarships at 662-325-2224 or
<EMAIL_ADDRESS>
Figure 4: A graphical depiction of response generation. Red highlights
indicate BarkPlug v.2’s ability to contextualize and generate factually
grounded response.
## IV Experiments and Results
To evaluate the performance of our system, we adopt a two-fold approach
including both _quantitative_ and _usability assessment_ methods. For the
quantitative evaluation (See Section IV-B), we utilize the RAGAS [40]
framework, while the SUS is adopted for usability assessment (See Section
IV-C). In the following subsection, we first discuss the dataset and steps we
took to prepossess them and then provide a detailed explanation of our
evaluation approaches.
### IV-A Dataset Description & Preparation
To ensure a comprehensive chatbot system capable of answering diverse
questions—whether academic or non-academic—we initially developed a web
scraper to gather information on various campus resources at Mississippi State
University. This collection would then serve as an external data source in our
pipeline. We scraped various campus resources including academic departments,
financial aid, scholarships, housing, dining, parking, and police. In total,
we scraped 42 campus resources into a JSON file. Each JSON file includes the
following information: the URL, title, and content of the scraped webpage, all
wrapped into a JSON object. We consolidated the individual files into a master
JSON file which serves as an external data source and is ingested into our RAG
pipeline. A subset of the data utilized by BarkPlug v.2 can be observed in
Table I.
To enhance retrieval accuracy, we first preprocess the JSON file. This
preprocessing step involves removing noise, such as undesirable Unicode
characters, redundant, and unnecessary information. We then implement a
recursive chunking strategy, with a chunk size of 8000 and an overlap of 1200
characters. This step is crucial for optimizing the performance of RAG chatbot
systems with the objective of ensuring that our chatbot generates an accurate
response that is contextually appropriate. Subsequently, we transformed the
textual data into vectorized representations utilizing an _embedding model_
(Refer to Section III-A to learn for more details on embedding models.).
TABLE I: A subset of an external data source containing campus resources, including both academic and non-academic departments, indicating the total number of tokens associated with each. | Departments | # of Tokens
---|---|---
Campus Resources | Computer Science and Engineering | 200623
Chemical Engineering | 118271
Electrical and Computer Engineering | 328558
Industrial and Systems Engineering | 22390
Agricultural and Biological Engineering | 79978
Civil and Environmental Engineering | 61071
Aerospace Engineering | 37812
Biomedical Engineering | 256761
Housing | 132193
Admission | 276972
MSU Police | 16629
### IV-B Quantitative Evaluation
To evaluate BarkPlug v.2’s ability to produce contextually appropriate
responses, we utilize the RAGAS framework [40]. We choose this framework
because it is specifically designed to assess RAG pipelines. Other popular
evaluation metrics such as ROUGE [41] and BLEU [42] are not suitable in our
context. This is because ROUGE is generally used to evaluate summarization
tasks, while BLEU is designed to evaluate language translation tasks.
TABLE II: Overview of results: Retrieval scores pertain to the _context retrieval_ phase of the architecture, where _prec._ refers to context precision, and recall refers to context recall. Generation scores pertain to the _completion phase_ , where _faith_ stands for faithfulness and _rel._ for answer relevancy. The end-to-end evaluation showcases BarkPlug v.2’s efficiency in generating contextually relevant and accurate answers through metrics such as answer similarity and answer correctness. | Retrieval | Generation | RAGAS Score | End-to-End Evaluation
---|---|---|---|---
Category | Prec. | Recall | Faith. | Rel. | Harmonic Mean | Answer Similarity | Answer Correctness
Engineering Programs | 0.98 | 0.96 | 0.99 | 0.97 | 0.97 | 0.8434 | 0.8620
General Inquiry | 0.95 | 0.97 | 0.98 | 0.96 | 0.96 | 0.7764 | 0.8123
Research Opportunities | 0.97 | 0.98 | 0.96 | 0.99 | 0.97 | 0.8245 | 0.8841
University Resources | 0.96 | 0.99 | 0.97 | 0.98 | 0.97 | 0.8317 | 0.8923
We evaluate both phase of BarkPlug v.2 architecture (See section III) i.e.
_context retrieval_ and _completion_. To evaluate the retrieval, we employ two
metrics such as _context precision and context recall_. The first metric
represents the Signal-to-Noise Ratio (SNR) of retrieved context, while the
second metric evaluates whether the retriever has the ability to retrieve all
the relevant evidence to answer a question. Similarly, to evalaute _completion
or generation_ we employ _faithfullness_ and _answer relevance_ metrics.
Faithfulness evaluates how factually accurate the generated answer is while
answer relevance evaluates how relevant the generated answer is to the
question. The final RAGAS score, representing the harmonic mean of these four
metrics, falls within a range of 0 to 1, with 1 denoting optimal generation.
This score serves as a singular measure of a QA system’s performance.
Therefore, the RAGAS score is essential for assessing the overall performance
and relevance of BarkPlug v.2 in its targeted educational environments.
To conduct phase wise evaluation, we first crafted a set of questions and
their ground truth pertaining to _engineering programs, general inquiries,
research opportunities_ , and _other university resources_. We report RAGAS
score of 0.97, 0.96, 0.97 and 0.97 for these categories respectively in Table
II. These score underlines both retrieval and completion component are
efficient.
We also conduct end-to-end evaluation to measure overall performance of
BarkPlug v.2, as it directly affects the user experience. Metrics such as
_answer similarity_ and _answer correctness_ are employed to assess the
overall performance, ensuring a comprehensive evaluation. In particular,
_answer similarity_ scores that reflect strong alignment with ideal responses
are reported to be high in cases when questions about engineering programs and
research opportunities are asked, with scores of 0.8434 and 0.8317
respectively. Moreover, _answer correctness_ , which indicates high factual
accuracy, is reported to be high when the system is asked questions about
university resources and research opportunities, at 0.8923 and 0.8841
respectively. Overall, these metrics suggest that BarkPlug v.2 effectively
retrieves relevant and accurate answers.
### IV-C Usability Assessment
To further understand the user experience when using BarkPlug v.2, we perform
a subjective satisfaction survey using the System Usability Scale (SUS) [43]
\- a widely reliable method that accesses systems usability through set of
questionnaire. Given the expensive nature of this evaluation we engage a panel
of 50 graduate and undergraduate students undertaking CSE8011 (Seminar course)
at Mississippi State University. The participants were tasked to answer set of
10 questions as depicted in 5, each offering five response options ranging
from “strongly agree” to “strongly disagree”. We then collected their feedback
and calculated an average SUS score of 67.75. The feedback results indicated
satisfactory usability with a room for improvement for future iterations of
our system.
Figure 5: Distribution of average System Usability Scale (SUS) scores.
## V Detailed Analysis and Insights
In this section, we discuss implementation details, where we explain the
technical process behind developing BarkPlug v.2. Then, we discuss constraints
and shortcomings encountered, and provide our plan for the future.
### V-A Implementation Detail
For data curation we employed a multi-thread web crawler with the Scrapy
Python library to collect data from over 42 campus resources (See Section IV-A
for details). We carefully selected important HTML div tags that comprised of
relevant information about a topic. This process was semi-automatic in nature
because every HTML pages were differently formatted with different div ids.
Manual div selection also allowed us to remove noise to some extent. The data
was exported to JSON file format with url, topic and content. Individual JSON
files for each of the campus resources was then consolidated into a master
JSON file for comprehensive retrieval.
We predominantly use LangChain framework to develop BarkPlug v.2. First, we
preprocess master JSON into smaller chunks using Recursive Character Text
Splitter splitting strategy. Given the nature of our data we opted for 8000
chunk size with 1200 overlap. We then apply an embedding function on these
chunks utilizing OpenAI’s _text-embedding-3-large_ model and store the vectors
in Chroma DB. This step allowed us to retrieve documents relevant to specific
user prompts. In our case, we utilize _vectorstore_ for _context retrieval_
with a similarity search threshold as our search strategy (See Section III-A
for more details). For completion or response generation we leverage OpenAI’s
_gpt-3.5-turbo_ model. Both embedding and response generation is managed
through API calls.
BarkPlug v.2 is built with Django framework using python. For front-end we
utilize HTML, CSS and Javascript. The current version of our system has not
only question-answering functionality but also user sign up and log in
feature. Once a user is registered they can ask queries, they can see previous
conversations, delete conversations, and email conversations. Our application
is deployed through a third-party cloud service for accessibility.
### V-B Limitations & Future Direction
Despite the achievements in developing our educational chatbot, several
significant challenges currently limit its broader application.BarkPlug v.2
does not currently have Automatic Speech Recognition (ASR) capability, which
might hinder its use among visually impaired, disabled, or elderly users.
Additionally, given that Mississippi State University hosts a number of
international students annually from non-English speaking countries, it
currently lacks multi-lingual support. In terms of technical limitations, our
retrieval system sometimes fails to provide accurate or relevant results,
occasionally producing incorrect information, a phenomenon known as
‘hallucinations’. We are also limited by a maximum number of output tokens,
which is 4096, and a context window of 16k. This sometimes hinders system’s
ability to capture the full length of the conversation in the memory buffer.
To address the limitations discussed above and enhance BarkPlug v.2’s
functionality and usability, we are planning several key upgrades. These
include adding support for multiple languages to cater to a diverse user base,
integrating ASR and text conversion features to enable various interaction
modes, and improving the retrieval algorithms to boost the accuracy and
relevance of the information provided. Moreover, in response to the token
limitations of the OpenAI API, we aim to apply the map-reduced document chain
approach from LangChain. Through these improvements, we aim to transform
BarkPlug v.2 into a more reliable and accessible educational tool.
## VI Conclusion
This study highlights the significant potential of AI-based chat systems in
improving communication and access to information regarding university
resources. Our system, BarkPlug v.2 integrates large amounts of university
data, including academic programs, campus facilities, student service as
external data corpus into its RAG pipelines for domain-specific question and
answering tasks. By incorporating this external data corpus, our system
ensures the delivery of precise and contextually relevant responses to both
academic and non-academic user inquiries. The comprehensive end-to-end
evaluation process demonstrated BarkPlug v.2’s efficiency in generating
contextually relevant and accurate answers as measured by metrics such as
answer similarity and correctness. Furthermore, system usability experiments
employing the SUS indicated that BarkPlug v.2 is practical and effective for
real-world usage, affirming its reliability and the positive user experience
it offers. The positive outcomes of using BarkPlug v.2 at Mississippi State
University suggest promising opportunities for broader implementation. This
system could be adapted for use in other universities or different sectors and
can be viewed as enterprise document retrieval systems that enhance user
engagement and information access.
## Declaration of Competing Interest
The authors declare that the research was conducted in the absence of any
commercial or financial relationships that could be construed as a potential
conflict of interest.
## Acknowledgement
This work was supported by the PATENT Lab (Predictive Analytics and Technology
Integration Laboratory) at the Department of Computer Science and Engineering,
Mississippi State University.
## References
* [1] K. A. Meyer and S. Jones, “Information found and not found: what university websites tell students,” _Online journal of distance learning administration_ , vol. 14, no. 3, pp. 1–10, 2011.
* [2] C. Khatri, A. Venkatesh, B. Hedayatnia, R. Gabriel, A. Ram, and R. Prasad, “Alexa prize—state of the art in conversational ai,” _AI Magazine_ , vol. 39, no. 3, pp. 40–55, 2018.
* [3] L. Bradeško and D. Mladenić, “A survey of chatbot systems through a loebner prize competition,” in _Proceedings of Slovenian language technologies society eighth conference of language technologies_ , vol. 2. sn, 2012, pp. 34–37.
* [4] ASU. Introducing sunny. [Online]. Available: https://heysunny.asu.edu/about
* [5] GSU. Reduction of summer melt. [Online]. Available: https://success.gsu.edu/initiatives/reduction-of-summer-melt/
* [6] S. University. Beacon - your digital guide. [Online]. Available: https://www.staffs.ac.uk/students/digital-services/beacon
* [7] L. Laranjo, A. G. Dunn, H. L. Tong, A. B. Kocaballi, J. Chen, R. Bashir, D. Surian, B. Gallego, F. Magrabi, A. Y. Lau _et al._ , “Conversational agents in healthcare: a systematic review,” _Journal of the American Medical Informatics Association_ , vol. 25, no. 9, pp. 1248–1258, 2018.
* [8] S. Neupane, S. Mitra, S. Mittal, N. A. Golilarz, S. Rahimi, and A. Amirlatifi, “Medinsight: A multi-source context augmentation framework for generating patient-centric medical responses using large language models,” _arXiv preprint arXiv:2403.08607_ , 2024.
* [9] M. F. Franco, B. Rodrigues, E. J. Scheid, A. Jacobs, C. Killer, L. Z. Granville, and B. Stiller, “Secbot: a business-driven conversational agent for cybersecurity planning and management,” in _2020 16th international conference on network and service management (CNSM)_. IEEE, 2020, pp. 1–7.
* [10] S. Mitra, S. Neupane, T. Chakraborty, S. Mittal, A. Piplai, M. Gaur, and S. Rahimi, “Localintel: Generating organizational threat intelligence from global and local cyber knowledge,” _arXiv preprint arXiv:2401.10036_ , 2024.
* [11] M. Chung, E. Ko, H. Joung, and S. J. Kim, “Chatbot e-service and customer satisfaction regarding luxury brands,” _Journal of Business Research_ , vol. 117, pp. 587–595, 2020.
* [12] X. Y. Leung and H. Wen, “Chatbot usage in restaurant takeout orders: A comparison study of three ordering methods,” _Journal of Hospitality and Tourism Management_ , vol. 45, pp. 377–386, 2020.
* [13] S. I. LLP. Global chatbot market size. [Online]. Available: https://finance.yahoo.com/news/global-chatbot-market-size-exceed-080000758.html
* [14] P. Lewis, E. Perez, A. Piktus, F. Petroni, V. Karpukhin, N. Goyal, H. Küttler, M. Lewis, W.-t. Yih, T. Rocktäschel _et al._ , “Retrieval-augmented generation for knowledge-intensive nlp tasks,” _Advances in Neural Information Processing Systems_ , vol. 33, pp. 9459–9474, 2020.
* [15] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, Ł. Kaiser, and I. Polosukhin, “Attention is all you need,” _Advances in neural information processing systems_ , vol. 30, 2017.
* [16] Y. Chang, X. Wang, J. Wang, Y. Wu, L. Yang, K. Zhu, H. Chen, X. Yi, C. Wang, Y. Wang _et al._ , “A survey on evaluation of large language models,” _ACM Transactions on Intelligent Systems and Technology_ , vol. 15, no. 3, pp. 1–45, 2024.
* [17] F. Petroni, T. Rocktäschel, P. Lewis, A. Bakhtin, Y. Wu, A. H. Miller, and S. Riedel, “Language models as knowledge bases?” _arXiv preprint arXiv:1909.01066_ , 2019.
* [18] Z. Ji, N. Lee, R. Frieske, T. Yu, D. Su, Y. Xu, E. Ishii, Y. J. Bang, A. Madotto, and P. Fung, “Survey of hallucination in natural language generation,” _ACM Computing Surveys_ , vol. 55, no. 12, pp. 1–38, 2023.
* [19] S. Cunningham-Nelson, W. Boles, L. Trouton, and E. Margerison, “A review of chatbots in education: practical steps forward,” in _30th annual conference for the australasian association for engineering education (AAEE 2019): educators becoming agents of change: innovate, integrate, motivate_. Engineers Australia, 2019, pp. 299–306.
* [20] S. Wollny, J. Schneider, D. Di Mitri, J. Weidlich, M. Rittberger, and H. Drachsler, “Are we there yet?-a systematic literature review on chatbots in education,” _Frontiers in artificial intelligence_ , vol. 4, p. 654924, 2021.
* [21] C. W. Okonkwo and A. Ade-Ibijola, “Chatbots applications in education: A systematic review,” _Computers and Education: Artificial Intelligence_ , vol. 2, p. 100033, 2021.
* [22] A. S. D. Martha and H. B. Santoso, “The design and impact of the pedagogical agent: A systematic literature review.” _Journal of educators Online_ , vol. 16, no. 1, p. n1, 2019.
* [23] R. Winkler, S. Hobert, A. Salovaara, M. Söllner, and J. M. Leimeister, “Sara, the lecturer: Improving learning in online education with a scaffolding-based conversational agent,” in _Proceedings of the 2020 CHI conference on human factors in computing systems_ , 2020, pp. 1–14.
* [24] J. Q. Pérez, T. Daradoumis, and J. M. M. Puig, “Rediscovering the use of chatbots in education: A systematic literature review,” _Computer Applications in Engineering Education_ , vol. 28, no. 6, pp. 1549–1565, 2020.
* [25] S. Hobert and R. Meyer von Wolff, “Say hello to your new automated tutor–a structured literature review on pedagogical conversational agents,” 2019.
* [26] G.-J. Hwang and C.-Y. Chang, “A review of opportunities and challenges of chatbots in education,” _Interactive Learning Environments_ , vol. 31, no. 7, pp. 4099–4112, 2023.
* [27] K.-J. Tokayev, “Ethical implications of large language models a multidimensional exploration of societal, economic, and technical concerns,” _International Journal of Social Analytics_ , vol. 8, no. 9, pp. 17–33, 2023.
* [28] E. Adamopoulou and L. Moussiades, “An overview of chatbot technology,” in _IFIP international conference on artificial intelligence applications and innovations_. Springer, 2020, pp. 373–383.
* [29] ——, “Chatbots: History, technology, and applications,” _Machine Learning with applications_ , vol. 2, p. 100006, 2020.
* [30] H. Tripathi, “Experimental approach toward training and analysing siamese deep neural network for sentence with no repeated expressions,” in _2023 14th International Conference on Computing Communication and Networking Technologies (ICCCNT)_. IEEE, 2023, pp. 1–5.
* [31] M. A. Kuhail, N. Alturki, S. Alramlawi, and K. Alhejori, “Interacting with educational chatbots: A systematic review,” _Education and Information Technologies_ , vol. 28, no. 1, pp. 973–1018, 2023.
* [32] Q. Xie, W. Lu, Q. Zhang, L. Zhang, T. Zhu, and J. Wang, “Chatbot integration for metaverse-a university platform prototype,” in _2023 IEEE International Conference on Omni-layer Intelligent Systems (COINS)_. IEEE, 2023, pp. 1–6.
* [33] Y. W. Chandra and S. Suyanto, “Indonesian chatbot of university admission using a question answering system based on sequence-to-sequence model,” _Procedia Computer Science_ , vol. 157, pp. 367–374, 2019.
* [34] P. F. Oliveira and P. Matos, “Introducing a chatbot to the web portal of a higher education institution to enhance student interaction,” _Engineering Proceedings_ , vol. 56, no. 1, p. 128, 2023.
* [35] C. Martinez-Araneda, M. Gutiérrez, D. Maldonado, P. Gómez, A. Segura, and C. Vidal-Castro, “Designing a chatbot to support problem-solving in a programming course,” in _INTED2024 Proceedings_. IATED, 2024, pp. 966–975.
* [36] C. Song and A. Raghunathan, “Information leakage in embedding models,” in _Proceedings of the 2020 ACM SIGSAC conference on computer and communications security_ , 2020, pp. 377–390.
* [37] Y. Han, C. Liu, and P. Wang, “A comprehensive survey on vector database: Storage and retrieval technique, challenge,” _arXiv preprint arXiv:2310.11703_ , 2023.
* [38] Chroma. Chroma: The ai-native open-source embedding database. [Online]. Available: https://www.trychroma.com/
* [39] LangChain. Applications that can reason. powered by langchain. [Online]. Available: https://www.langchain.com/
* [40] S. Es, J. James, L. Espinosa-Anke, and S. Schockaert, “Ragas: Automated evaluation of retrieval augmented generation,” _arXiv preprint arXiv:2309.15217_ , 2023.
* [41] C.-Y. Lin, “Rouge: A package for automatic evaluation of summaries,” in _Text summarization branches out_ , 2004, pp. 74–81.
* [42] K. Papineni, S. Roukos, T. Ward, and W.-J. Zhu, “Bleu: a method for automatic evaluation of machine translation,” in _Proceedings of the 40th annual meeting of the Association for Computational Linguistics_ , 2002, pp. 311–318.
* [43] P. Vlachogianni and N. Tselios, “Perceived usability evaluation of educational technology using the system usability scale (sus): A systematic review,” _Journal of Research on Technology in Education_ , vol. 54, no. 3, pp. 392–409, 2022.
|
mythm[theorem]Theorem mylmm[theorem]Lemma University of Southern California,
USA and https://jaredraycoleman.com
[email protected]://orcid.org/0000-0003-1227-2962 Carleton University,
Ottawa, Ontario, Canada and https://people.scs.carleton.ca/~kranakis/
[email protected]://orcid.org/0000-0002-8959-4428Research
supported in part by NSERC Discovery grant. Wesleyan University, Middletown
CT, USA and http://dkrizanc.web.wesleyan.edu/<EMAIL_ADDRESS>California
State University, Long Beach, CA, USA and https://home.csulb.edu/~omorales/
[email protected]://orcid.org/0000-0002-9645-1257 Jared
Coleman and Evangelos Kranakis and Danny Krizanc and Oscar Morales Ponce
<ccs2012> <concept> <concept_id>10003752.10003809.10010047</concept_id>
<concept_desc>Theory of computation Online algorithms</concept_desc>
<concept_significance>300</concept_significance> </concept> <concept>
<concept_id>10003752.10003809.10010047.10010051</concept_id>
<concept_desc>Theory of computation Adversary models</concept_desc>
<concept_significance>300</concept_significance> </concept> </ccs2012>
[300]Theory of computation Online algorithms [300]Theory of computation
Adversary models John Q. Open and Joan R. Access 2 42nd Conference on Very
Important Topics (CVIT 2016) CVIT 2016 CVIT 2016 December 24–27, 2016 Little
Whinging, United Kingdom 42 23
# Line Search for an Oblivious Moving Target
Jared Coleman Evangelos Kranakis Danny Krizanc Oscar Morales-Ponce
###### Abstract
Consider search on an infinite line involving an autonomous robot starting at
the origin of the line and an oblivious moving target at initial distance
$d\geq 1$ from it. The robot can change direction and move anywhere on the
line with constant maximum speed $1$ while the target is also moving on the
line with constant speed $v>0$ but is unable to change its speed or direction.
The goal is for the robot to catch up to the target in as little time as
possible.
The classic case where $v=0$ and the target’s initial distance $d$ is unknown
to the robot is the well-studied “cow-path problem”. Alpert and Gal
[alpern2003theory] gave an optimal algorithm for the case where a target with
unknown initial distance $d$ is moving away from the robot with a known speed
$v<1$. In this paper we design and analyze search algorithms for the remaining
possible knowledge situations, namely, when $d$ and $v$ are known, when $v$ is
known but $d$ is unknown, when $d$ is known but $v$ is unknown, and when both
$v$ and $d$ are unknown. Furthermore, for each of these knowledge models we
consider separately the case where the target is moving away from the origin
and the case where it is moving toward the origin. We design algorithms and
analyze competitive ratios for all eight cases above. The resulting
competitive ratios are shown to be optimal when the target is moving towards
the origin as well as when $v$ is known and the target is moving away from the
origin.
###### keywords:
Infinite Line, Knowledge, Oblivious, Robot, Search, Search-Time, Speed, Target
## 1 Introduction
Search is important to many areas of computer science and mathematics and has
received the attention of numerous studies. In the simplest search scenario,
one is interested in the optimal trajectory of a single autonomous mobile
agent (also referred to simply as a robot) tasked with finding a target placed
at an unknown location on the infinite line. The line search problem is to
give an algorithm for the agent so as to minimize the competitive ratio
defined as the supremum over all possible target locations of the ratio of the
time the agent takes to find the target and the time it would take if the
target’s initial position was known to the robot ahead of time. This classic
problem has led to many variations (see [alpern2003theory] for more on its
history).
In this paper we consider an extension of the line search problem involving an
autonomous robot and an oblivious moving target. The search is again performed
on an infinite line and concerns an autonomous robot starting at the origin of
the line but differs from the previously studied case in that the search is
for a moving target whose speed and direction are not necessarily known to the
searching robot. The robot starts at the origin and the target at an arbitrary
distance $d$ from the origin. The target is moving with constant speed and is
oblivious in that it cannot change its speed and/or direction of movement. We
consider and analyze several alternative knowledge-based scenarios in which
the target’s speed and initial distance from the origin may be known or
unknown to the searching robot. The case where a target with unknown initial
distance from the origin is moving away from the origin was solved by Alpern
and Gal [alpern2003theory]. As far as we are aware, these are the first
results for the remaining cases.
### 1.1 Notation and terminology
On the infinite real line, consider an autonomous robot which is initially
placed at the origin whose maximum speed is $1$ and an oblivious robot (also
referred to as the moving target) initially placed at a distance $d$ to the
right or left of the origin and moving with constant speed $v>0$. As is
usually done in linear search and in order to avoid trivial considerations on
the competitive ratio by adversarially placing the target very close to the
robot, we assume that $d$ is not smaller than the unit distance, i.e., $d\geq
1$.
The target may be moving away from or toward the origin. If it is moving away,
we assume its speed is strictly less than $1$ as otherwise the problem can not
be solved. Further, we assume that the autonomous robot knows the direction
the target is moving (away from or toward the origin). The search is completed
as soon as the robot and target are co-located.
The movement of the autonomous robot is determined by a trajectory which is
defined as a continuous function $t\to f(t)$, with $f(t)$ denoting the
location of the robot at time $t$. Moreover, it is true that
$|f(t)-f(t^{\prime})|\leq u|t-t^{\prime}|$, for all $t,t^{\prime}$, where $u$
is the speed of the agent (be that the searching robot or the oblivious
target). The autonomous robot can move with its own constant speed and during
the traversal of its trajectory it may stop and/or change direction
instantaneously and at any time as specified by the search algorithm.
A search strategy is a sequence of movements followed by the robot. The
competitive ratio of a search strategy $X$, denoted $CR_{X}$, is defined as
the supremum over all possible initial target locations and speeds of the
ratio of the time the agent takes to find the target and the time it would
take if the target’s initial position was known to the robot ahead of time.
The competitive ratio of a certain type of search problem is the infimum of
$CR_{X}$ taken over all possible strategies $X$ for this problem. By abuse of
notation we may drop mention of $X$ when this is easily implied from the
context.
Our goal in this paper is to prove bounds on the competitive ratios of
algorithms under four different knowledge models:
1. 1.
FullKnowledge: The robot knows both the target’s speed $v$ and its initial
distance $d$.
2. 2.
NoDistance: The robot knows the target’s speed $v$ but not its initial
distance $d$.
3. 3.
NoSpeed: The robot knows the target’s initial distance $d$, but not its speed
$v$.
4. 4.
NoKnowledge: The robot knows neither the target’s speed $v$ nor its initial
distance $d$.
For all knowledge models, the robot does not know the target’s initial
position. We study each of the above knowledge models for the case where the
target is moving toward the origin (Toward) and where it is moving away (Away)
from the origin. In each case, we assume the robot knows the direction of
travel of the target.
### 1.2 Related Work
Several research papers have considered the search problem for a robot
searching for a static (fixed) target placed at an unknown location on the
real line, see [BCR93, schuierer2001lower]. The problem was first
independently considered in a stochastic setting by Bellman and Beck in the
1960’s (cf. [beck1964linear, bellman1963optimal] as well as [BCR93,
schuierer2001lower]). In a deterministic setting it is now well known that the
optimal trajectory for this single agent search uses a doubling strategy whose
trajectory attains a competitive ratio of $9$. Linear search has attracted
much attention and been the focus of books including [ahlswede1987search,
alpern2003theory, stone1975theory].
The case of a moving target appears to have been first considered by McCabe
[mccabe1974]. In that paper, the problem of searching for an oblivious target
that follows a Bernoulli random walk on the integers is considered. For the
case of a deterministic oblivious searcher, the only result we are aware of us
is found in Alpern and Gal [alpern2003theory]. There they consider the case
where the target is moving away from the origin at a constant speed $v<1$
which is known to the searching robot. Only the initial distance of the target
is unknown. They give an algorithm with optimal competitive ratio for this
case.
Our problem is reminiscent of the problem of catching a fugitive in a given
domain which is generally referred to as the cops and robbers problem
[anthony2011game]. The main difference is that in those problems, the target
(robber) is itself an autonomous agent. As a result, the techniques considered
there do not apply to our case.
Our problem is also related to rendezvous (of two robots) on an infinite line
but it differs because in our case only one of the robots is autonomous while
the other is oblivious. Related studies on the infinite line include
rendezvous with asymmetric clocks [czyzowicz2018linear] and asynchrnous
deterministic rendezvous [de2006asynchronous]. More recent work on linear
search concerns searching for a static target by a group of cooperating
robots, some of which may have suffered either crash
[czyzowicz2019searchcrash] or Byzantine [czyzowicz2021searchbyz] faults.
### 1.3 Results of the paper
In all situations considered it is unknown to the robot whether the target is
initially to the left or to the right of the origin. We analyze the
competitive ratio in four situations which reflect what knowledge the robot
has about the target. We present results on the FullKnowledge model (the robot
knows $v$ and $d$) in Section 2, the NoDistance model (the robot knows $v$ but
not $d$) in Section 3, the NoSpeed model (the robot knows $d$ but not $v$) in
Section 4, and the NoKnowledge model (the robot knows neither $v$ nor $d$) in
Section 5. For each of these models we study separately the case when the
target is moving away or toward the origin (this knowledge being available to
the robot). The results are summarized in Table 1. We conclude with a summary
and additional open problems.
Knowledge | Movement | Competitive Ratio | Section
---|---|---|---
$v,d$ | Away | $CR=1+\frac{2}{1-v}$ | 2.1
Toward | | $CR=1+\frac{2}{1+v}$ if $v<1$
---
$CR=1+\frac{1}{v}$ otherwise
2.2
$v$ | Away | $CR=1+8\frac{1+v}{(1-v)^{2}}$ | 3.1 [alpern2003theory]
Toward | | $CR=1+\frac{1}{v}$ if $v\geq\frac{1}{3}$
---
$CR=1+8\frac{1-v}{(1+v)^{2}}$ otherwise
3.2
$d$ | Away | | $CR\leq 5$ if $v\leq\frac{1}{2}$
---
$CR\leq 1+16\frac{\left(\log\frac{1}{1-v}\right)^{2}}{(1-v)^{4}}$ otherwise
4.1
Toward | $CR=3$ | 4.2
$\emptyset$ | Away | | $CR\leq 1+\frac{16}{d}\left[\log\log\left(\max\left(d,\frac{1}{1-v}\right)\right)+3\right]$
---
$\cdot\max\left(d,\frac{1}{1-v}\right)^{8}\cdot\log^{2}\left[\max\left(d,\frac{1}{1-v}\right)\right]$
5.1
Toward | $CR=1+\frac{1}{v}$ | 5.2
Table 1: Table of competitive ratio bounds proven for each knowledge model for
cases with the target moving away from or towards the origin with speed $v$
and initial distance $d$ from the robot which is moving with speed $1$.
Equalities indicate that tight upper and lower bounds are proven.
## 2 The FullKnowledge Model
We first study the model where the robot knows the target’s speed $v$ and its
initial distance from the origin $d$.
### 2.1 The FullKnowledge/Away Model
For the case when the target is moving away from the origin, clearly if $v\geq
1$ then the robot can never catch the target. Thus, for this model (and all
other Away models), we assume $v<1$. In this section, we will analyze an
algorithm where the robot chooses a direction and moves for time
$\frac{d}{1-v}$. If the robot does not find the target after moving for time
$\frac{d}{1-v}$ in one direction, then it changes direction and continues
moving until it does.
Algorithm 1 Online Algorithm for FullKnowledge/Away Model
1:input: target speed $v$ and initial distance $d$
2:choose any direction and go for time $\frac{d}{1-v}$
3:if target not found then
4: change direction and go until target is found
###### Theorem 2.1.
For the FullKnowledge/Away model, Algorithm 1 has an optimal competitive ratio
of
$1+\frac{2}{1-v}.$ (1)
###### Proof 2.2.
By Algorithm 1, the robot goes in one direction for a time $\frac{d}{1-v}$.
Observe that if the robot does not encounter the target after this amount of
time, it must be on the opposite side of the origin (in the other direction).
At the time the robot changes direction, its distance to the target will be
equal to $\frac{d}{1-v}+d+\frac{dv}{1-v}=\frac{2d}{1-v}$. Thus, the total time
required until the robot catches up to the target is at most
$\frac{d}{1-v}+\frac{2d}{(1-v)^{2}}.$
Clearly then, the competitive ratio is at most
$\frac{\frac{d}{1-v}+\frac{2d}{(1-v)^{2}}}{\frac{d}{1-v}}=1+\frac{2}{1-v}$
which is as claimed in Equation (1) above.
Optimality follows from the fact that regardless of which direction the robot
chooses to travel, the adversary can place the target in the opposite
direction. Moreover, for the robot to catch up to the target it must visit one
of the points $\pm\frac{d}{1-v}$. If the robot visits location $\frac{d}{1-v}$
to the right (resp. left) the adversary places the target on the left (resp.
right). Therefore the completion time will be at least
$\frac{d}{1-v}+\frac{2d}{(1-v)^{2}}$. This shows the upper bound is tight and
completes the proof of Theorem 2.1.
### 2.2 The FullKnowledge/Toward Model
Consider the following algorithm which is similar to Algorithm 1.
Algorithm 2 Online Algorithm for the FullKnowledge/Toward Model
1:input: target speed $v$ and initial distance $d$
2:choose any direction and go for time $\frac{d}{1+v}$
3:if target not found then
4: change direction and go until target is found
###### Theorem 2.3.
For the FullKnowledge/Toward model, Algorithm 2 has competitive ratio at most
$1+\frac{2}{1+v}.$ (2)
###### Proof 2.4.
The robot goes in one direction for a time $\frac{d}{1+v}$. If the robot finds
the target in this time, the algorithm is clearly optimal. If, however, the
robot does not find the target, then it must be on the opposite side of the
origin (in the other direction). If this is the case, then by time
$\frac{d}{1+v}$ the target has moved a distance $\frac{dv}{1+v}$ and is at
distance $d-\frac{dv}{1+v}=\frac{d}{1+v}$ from the origin. Therefore at the
time the robot changes direction, the distance between robot and target is
$\frac{2d}{1+v}$. Thus, the robot will encounter the target in additional time
$\frac{2d}{(1+v)^{2}}$. It follows that the total time required for the robot
to meet the target is $\frac{d}{1+v}+\frac{2d}{(1+v)^{2}}$ and the resulting
competitive ratio satisfies
$CR\leq\frac{\frac{d}{1+v}+\frac{2d}{(1+v)^{2}}}{\frac{d}{v+1}}=1+\frac{2}{1+v}.$
This completes the proof of Theorem 2.3.
###### Theorem 2.5.
For the FullKnowledge/Toward model, the competitive ratio of any online
algorithm is at least $1+\frac{2}{1+v}$, provided that $v<1$. In particular,
Algorithm 2 is optimal for $v<1$.
###### Proof 2.6.
Consider any algorithm for a robot starting at the origin to meet a target
initially placed at an unknown location distance $d$ from the origin. For any
point at distance $a$ from the origin, the target takes exactly
$\frac{d-a}{v}$ time to reach $a$. Then, let $t$ denote the time the robot
first passes through a point at distance $a$ from the origin. If
$t<\frac{d-a}{v}$, then the robot cannot know whether the target is on the
same or opposite side of the origin. On the other hand, if
$t\geq\frac{d-a}{v}$ and it has not encountered the target, then the target
must be on the opposite side of the origin. Thus, given a trajectory, let $\pm
a$ be the first point such that the robot is at position $\pm a$ at time
exactly $\frac{d-a}{v}$. Clearly such a point must exist for any trajectory
since the target is moving toward the origin. Then whichever side of the
origin the robot is on, consider the instance where the target started on the
opposite side. Clearly then, the robot takes an additional time at least
$\frac{2a}{1+v}$ to reach the target. Thus, the competitive ratio is given by:
$\displaystyle\frac{\frac{d-a}{v}+\frac{2a}{1+v}}{\frac{d}{1+v}}$
$\displaystyle=\frac{(d-a)(1+v)+2av}{vd}=1+\frac{1}{v}+\frac{a(v-1)}{dv}.$ (3)
Observe that whenever $v<1$, the right-hand side of Equation (3) satisfies
$1+\frac{1}{v}+\frac{a(v-1)}{dv}\geq
1+\frac{1}{v}+\frac{\frac{d}{1+v}(v-1)}{dv}=1+\frac{2}{1+v}$
which completes the proof of Theorem 2.5.
With Theorem 2.5 proved, we know Algorithm 2 is optimal for any value of $v$
between $0$ and $1$, but what about when $v>1$? In this case, we’ll prove the
following algorithm is optimal: the robot waits at the origin forever. We call
this algorithm “the waiting algorithm”.
###### Theorem 2.7.
Whenever the target is moving toward the origin with speed $v\geq 1$, the
waiting algorithm has an optimal competitive ratio of $1+\frac{1}{v}$.
###### Proof 2.8.
Clearly the algorithm takes exactly time $d/v$ to complete and so the upper
bound follows trivially. For the lower bound, we build upon the proof of
Theorem 2.5. It simply remains to consider Equation (3) for $v>1$. In this
case, the right-hand side of Equation (3) is increasing with respect to $a\geq
0$, so
$1+\frac{1}{v}+\frac{a(v-1)}{dv}\geq 1+\frac{1}{v}.$
This completes the proof of Theorem 2.7.
###### Remark 2.9.
Observe that the waiting algorithm makes no use of the target’s speed or
initial distance and therefore, as long as the target is moving toward the
origin, applies directly to the other knowledge models.
## 3 The NoDistance Model
In this section we assume that the robot knows $v$ but not $d$. Consider the
following zig-zag algorithm with “expansion ratio” $a>0$ (with the value of
$a$ to be determined).
Algorithm 3 Online Algorithm for NoDistance/Away and NoDistance/Toward Models
1:input: target speed $v$ and expansion ratio $a$
2:$i\leftarrow 0$
3:while target not found do
4: if at origin then
5: $d\leftarrow(-a)^{i}$
6: $i\leftarrow i+1$
7: else if at $d$ then
8: $d\leftarrow 0$
9: move toward $d$
### 3.1 The NoDistance/Away Model
The following result was shown by Alpern and Gal [alpern2003theory].
###### Theorem 3.1.
For the NoDistance/Away model, Algorithm 3 with $a=2\frac{1+v}{1-v}$ has an
optimal competitive ratio of
$\displaystyle 1+8\frac{1+v}{(1-v)^{2}}.$
### 3.2 The NoDistance/Toward Model
Recall first the statement made in Remark 2.9, that the optimality of the
waiting algorithm (which makes no use of any knowledge of $d$) holds for any
$d$ as long as $v\geq 1$. Thus, we need only consider scenarios where $0\leq
v<1$. As we will see, however, when the target is moving toward the origin,
the waiting algorithm is optimal for far slower targets! In general, since the
target is moving toward the origin, the robot need not search ever-increasing
distances away from the origin (i.e. execute Algorithm 3 with an expansion
ratio $a>1$). We call any algorithm which involves the robot never traveling
further than some finite distance from the origin (in one or both directions)
a contracting algorithm. Note that Algorithm 3 for $0<a\leq 1$ is a
contracting algorithm and $a=0$ is exactly the waiting algorithm. We’ll start
by showing that any contracting algorithm cannot have a better competitive
ratio than the waiting algorithm:
###### Theorem 3.2.
The competitive ratio of Algorithm 3 for any $0\leq a\leq 1$ is
$1+\frac{1}{v}$.
###### Proof 3.3.
Let $d^{\prime}$ be the finite distance further than which the robot will
never travel in at least one direction. Then consider the scenario where the
target is initially a distance $d=c\cdot d^{\prime}>>d^{\prime}$ from the
origin in the same direction. Then the competitive ratio is at least
$\displaystyle\sup_{c}\frac{\frac{cd^{\prime}-d^{\prime}}{v}}{\frac{cd^{\prime}}{v+1}}$
$\displaystyle=\sup_{c}\frac{c-1}{c}\frac{1+v}{v}=\lim_{c\rightarrow\infty}\frac{c-1}{c}\left(1+\frac{1}{v}\right)=1+\frac{1}{v}$
which proves Theorem 3.2.
By Theorem 3.2, any algorithm which hopes to out-perform the waiting algorithm
must be expanding. Now we show that the following hybrid algorithm, Algorithm
4, is optimal.
Algorithm 4 Wait or Zig-Zag Search Algorithm for NoDistance/Toward model
1:input: target speed $v$
2:if $v\geq\frac{1}{3}$ then
3: execute waiting algorithm
4:else
5: execute Algorithm 3 with $a=2\frac{1-v}{1+v}$
###### Theorem 3.4.
For the NoDistance/Toward model, the competitive ratio of Algorithm 4 is at
most
$\displaystyle\begin{cases}1+\frac{1}{v}&\mbox{if $v\geq\frac{1}{3}$}\\\
1+8\frac{1-v}{(1+v)^{2}}&\mbox{if $v<\frac{1}{3}$}\end{cases}$ (4)
###### Proof 3.5.
The first case is trivial: the competitive ratio of the waiting algorithm is
exactly $1+\frac{1}{v}$ by Theorem 2.7. The second case, however, is a bit
more complicated. First, observe that if $v<\frac{1}{3}$ then $a$ must be less
than $3$. Indeed, consider the scenario where the robot “just misses” the
target on the very first round of the algorithm (after traveling a distance
$1$ in some direction and then turning around). Then the competitive ratio of
the algorithm is
$\displaystyle 1+\frac{2a+2}{1+v}$
which is greater than $1+8\frac{1-v}{(1+v)^{2}}$ for any $a>3$ and $v>0$:
$\displaystyle 1+\frac{2a+2}{1+v}$
$\displaystyle>1+\frac{8}{1+v}>1+\frac{8}{1+v}\cdot\frac{1-v}{1+v}$
since $\frac{1-v}{1+v}<1$.
Now, consider the round $k$ when the robot catches up to the target and
observe that
$\displaystyle a^{k-2}$
$\displaystyle<d-\left(2\sum_{i=0}^{k-3}a^{i}+a^{k-2}\right)v=d-\left(2\frac{a^{k-2}-1}{a-1}+a^{k-2}\right)v$
since otherwise, the robot would have caught up to the target in round $k-2$.
This yields the following inequality which will prove useful in analyzing the
competitive ratio below:
$\displaystyle a^{k-2}$
$\displaystyle<d-\left(2\frac{a^{k-2}}{a-1}+a^{k-2}\right)v$
$\displaystyle\leq d\frac{a-1}{a-1+v(a+1)}+\frac{v}{a-1+v(a+1)}$
$\displaystyle\leq d\frac{a-1}{a-1+v(a+1)}+\frac{1}{4a-2}$ (5)
Observe the worst competitive ratio, then, is given by the situation where the
robot “just misses” the target on the $(k-2)^{\text{th}}$ round and catches up
to it only on round $k$. It follows the competitive ratio of Algorithm 4 is
$\displaystyle\frac{2\sum_{i=0}^{k-3}a^{i}+a^{k-2}+\frac{2(a^{k-2}+a^{k-1})}{1+v}}{\frac{d}{1-v}}\leq\frac{2\frac{a^{k-2}-1}{a-1}+a^{k-2}+\frac{2(a^{k-2}+a^{k-1})}{1+v}}{\frac{d}{1-v}}$
which, by Inequality (5) (and by substituting each $a^{k-2}$ with the right-
hand side of Inequality (5)), is less than or equal to
$\displaystyle CR$ $\displaystyle\leq
1+\frac{1}{2}\left[\frac{1}{d}\left(\frac{a-3}{a-1}\cdot\frac{v(5-7a)}{1-3a+2a^{2}}\right)+\frac{4a^{2}}{(a-1)+v(a+1)}\right]$
(6)
$\displaystyle\leq\lim_{d\rightarrow\infty}1+\frac{1}{2}\left[\frac{1}{d}\left(\frac{a-3}{a-1}\cdot\frac{v(5-7a)}{1-3a+2a^{2}}\right)+\frac{4a^{2}}{(a-1)+v(a+1)}\right]$
$\displaystyle=1+\frac{2a^{2}}{(a-1)+v(a+1)}$ (7)
which follows since the right-hand side of Inequality (6) is increasing with
respect to $d$ on $1<a\leq 3$. Finally, the right-hand side of Inequality (7)
is minimized at $a=2\frac{v-1}{v+1}$ with a value of
$1+8\frac{1-v}{(1+v)^{2}}$, which proves Theorem 3.4.
Now we show that Algorithm 4 is optimal by proving a tight lower bound on the
competitive ratio for any online algorithm. Our proof is based on techniques
developed in [killick2022cone]. Let $X(t)$ denote be the robot’s position at
time $t$ according to a given strategy.
###### Theorem 3.6.
For the NoDistance/Toward model, any strategy $X$ has a competitive ratio of
at least
$\displaystyle\begin{cases}1+\frac{1}{v}&\mbox{if $v\geq\frac{1}{3}$}\\\
1+8\frac{1-v}{(1+v)^{2}}&\mbox{otherwise}\end{cases}$
###### Proof 3.7.
Let $\beta_{t}=\inf_{t^{\prime}>t}\frac{t^{\prime}}{|X(t^{\prime})|}$.
Clearly, then, if $t_{1}\leq t_{2}$ then $\beta_{t_{1}}\leq\beta_{t_{2}}$.
Furthermore, $\beta_{t}\geq 1$ for all $t$ since the maximum speed of the
robot is $1$. Now let $\beta=\lim_{t\rightarrow\infty}\beta_{t}$. By
definition of the limit infimum, there must exist a finite time $t$ such that
$\beta\leq\frac{t^{\prime}}{|X(t^{\prime})|}$ for all $t^{\prime}\geq t$ and
thus there must exist a time $t_{1}>\frac{\beta(\beta+1)}{\beta-1}t$ such that
the robot reaches a point (without loss of generality, on the right side of
the origin) $X(t_{1})=\frac{t_{1}}{\beta+\epsilon_{1}}$ for any arbitrarily
small $\epsilon_{1}>0$. Consider such a time and observe that, by
construction, the robot could not have reached any point to the left of
$x_{0}=-\frac{t_{1}-X(t_{1})}{1+\beta}$ after time
$t_{0}=\frac{\beta(t_{1}-X(t_{1}))}{1+\beta}$ since $x_{0}\leq-t$ and
$t_{0}>t$ (see Figure 1).
Figure 1: The cone-bounded trajectory of the robot and worst-case placement
$p$ of the target. The small gray triangle is to remind the reader that, by
the definition of $\beta$, the robot trajectory is only guaranteed to be
contained by the cone after some finite time $t$. Thus, in order to maximize
the competitive ratio, we (as the adversary) should place the target so that
its trajectory does not intersect $(x_{0},t_{0})$ or the gray triangle.
Now, consider a target starting at initial positon $p$ (to be determined)
moving at speed $v>0$ toward a robot which starts at the origin and has a
speed of $1$. Thus, by placing the target at a starting location so that the
farthest right the robot could have reached is $x_{0}-\epsilon_{0}$ for any
arbitrarily small $\epsilon_{0}$, the robot can not have reached the target by
time $t_{1}$. Such a target has an initial position of
$\displaystyle p=-\frac{(1+\beta v)(t_{1}-X(t_{1}))}{1+\beta}-\epsilon_{0}$
and follows the trajectory
$\displaystyle X_{\text{target}}(t)=vt+p$ (8)
where $X_{\text{target}}(t)$ denotes the robot’s position at time $t$. Observe
also, if the robot moves directly toward the target after $t_{1}$, then its
trajectory after time $t_{1}$ is given by
$\displaystyle X(t)=X(t_{1})+t_{1}-t$ (9)
Thus, the earliest time the robot could possibly encounter the target can be
computed by finding the intersection between the robot trajectory (Equation
(9)) and the target’s trajectory (Equation (8)) and solving for $t$:
$\displaystyle vt+p$ $\displaystyle=X(t_{1})+t_{1}-t$ $\displaystyle t$
$\displaystyle=\frac{X(t_{1})+t_{1}-p}{1+v}.$ (10)
Then the competitive ratio (Equation (10) divided by $-p/(1+v)$, the optimal
search time) is
$\displaystyle CR$
$\displaystyle\geq\sup_{\epsilon_{0},\epsilon_{1}}\frac{(X(t_{1})+t_{1}-p)/(1+v)}{-p/(1+v)}=\sup_{\epsilon_{0},\epsilon_{1}}\frac{X(t_{1})+t_{1}-p}{-p}=\sup_{\epsilon_{0},\epsilon_{1}}\left[1-\frac{X(t_{1})+t_{1}}{p}\right]$
$\displaystyle=\sup_{\epsilon_{1}}\left[1+\frac{(1+\beta)(t_{1}+X(t_{1}))}{(1+\beta
v)(t_{1}-X(t_{1}))}\right]$ (11)
$\displaystyle=1+\frac{(1+\beta)^{2}}{(1+\beta v)(\beta-1)}$ (12)
where Inequality (11) follows since $p=-\frac{(1+\beta
v)(t_{1}-X(t_{1}))}{1+\beta}-\epsilon_{0}$ for arbitrarily small
$\epsilon_{0}>0$ and Inequality (12) follows since
$X(t_{1})=\frac{t_{1}}{\beta+\epsilon_{1}}$ for arbitrarily small
$\epsilon_{1}>0$. Finally, observe that if $v<\frac{1}{3}$, then the right-
hand side of Equality (12) has a single minimum of $1+8\frac{1-v}{(1+v)^{2}}$
at $\beta=\frac{v-3}{3v-1}$. On the other hand, if $v\geq\frac{1}{3}$, then
the right-hand side of Equality (12) is decreasing with respect to $\beta$ and
thus the competitive ratio satisfies
$\displaystyle
CR\geq\lim_{\beta\rightarrow\infty}\left[1+\frac{(1+\beta)^{2}}{(1+\beta
v)(\beta-1)}\right]=1+\frac{1}{v}.$
## 4 The NoSpeed Model
In this section we assume that the robot knows $d$ but not $v$.
### 4.1 The NoSpeed/Away
For this model, it is clear that the robot cannot execute an algorithm like
Algorithm 1 since no upper bound on the target’s speed is known. Note that, if
any upper bound $\hat{v}<1$ on the target’s speed were known, the robot could
execute Algorithm 1 by assuming the target speed to be equal to $\hat{v}$,
resulting in a competitive ratio of at most $1+\frac{2}{1-\hat{v}}$. Since the
target speed $v$ is unknown (and potentially very close to $1$), however, we
propose another strategy. Consider a monotone increasing non-negative integer
sequence $\\{f_{i}:i\geq 0\\}$ such that $f_{0}=1$ and $f_{i}<f_{i+1}$, for
all $i\geq 0$. The idea of the algorithm is to search for the target by making
a guess about its speed in rounds as follows. We start from the origin and
alternate searching right and left. On the $i$-th round, we use the guess
$v_{i}=1-2^{-f_{i}}$ and search the necessary distance away from the origin
such that, if the target’s speed is less than or equal to $v_{i}$ and the
target’s initial position is in the same direction from the origin that the
robot moves in round $i$, then the target will be found in round $i$.
Otherwise, we can conclude that either the target is moving with a speed
greater than $v_{i}$ or else it is on the opposite side of the origin. In this
case the robot returns to the origin and repeats the algorithm in the opposite
direction.
Later in the analysis we will show how to select the integer sequence
$\\{f_{i}:i\geq 0\\}$ so as to obtain bounds on the competitive ratio. The
algorithm explained above is formalized as Algorithm 5.
Algorithm 5 Online Algorithm for NoSpeed/Away Model
1:input: target initial distance $d$
2: integer sequence $\\{f_{i}:i\geq 0\\}$ such that $f_{i}<f_{i+1}$, for
$i\geq 0$ and $f_{0}=1$;
3:$t\leftarrow 0$
4:for $i\leftarrow 0,1,2,\ldots$ until target found do
5: $v_{i}\leftarrow 1-2^{-f_{i}}$
6: $x_{i}\leftarrow(-1)^{i}\cdot\frac{d+tv_{i}}{1-v_{i}}$
7: move to $x_{i}$ and back to the origin
8: $t\leftarrow t+|x_{i}|$
To compensate for the fact that the starting speed of the robot in the
algorithm is $v_{0}=1-2^{-1}=1/2$ we first need to consider the case
$v\leq\frac{1}{2}$.
###### Lemma 4.1.
For the NoSpeed/Away model, if the unknown speed $v$ of the target is less
than or equal to $\frac{1}{2}$ then the competitive ratio of Algorithm 5 is at
most $5$.
###### Proof 4.2.
According to Algorithm 5 and since $v<1/2$, the robot will find the target
either on its first trip away from the origin, after time at most $2d$, or
after the first time it changes direction of movement. In the worst case it
will spend time $2d$ in one direction and then additional time
$\frac{2d+d+2dv}{1-v}$. It follows that the competitive ratio is at most
$\displaystyle\frac{2d+\frac{2d+d+2dv}{1-v}}{\frac{d}{1-v}}=5$
which proves Lemma 4.1.
Next we analyze the competitive ratio of the algorithm for $v>\frac{1}{2}$.
###### Lemma 4.3.
For the NoSpeed/Away model, if the unknown speed $v$ of the target is greater
than $\frac{1}{2}$ then the competitive ratio of Algorithm 5 is at most
$1+2^{1+\sum_{j=0}^{k}f_{j}}\cdot 4^{k+1}$ where $k$ is the first $k$ such
that $v_{k}\geq v$.
###### Proof 4.4.
Let $d_{i}$ be the distance from the origin the target would be if its speed
was equal to $v_{i}$, where $v_{i}=1-2^{-f_{i}}$ at time
$\sum_{j=0}^{i-1}\left(1-2^{-f_{i}}\right)$. In other words, if $v_{i}\geq v$,
then $d_{i}$ is the maximum distance of the target from the origin (and thus,
the robot) at the beginning of round $i$ of the algorithm. Thus, if the speed
of the target is less than or equal to $v_{i}$ and the robot moves toward it
in round $i$, then it would take at most
$x_{i}=\frac{d_{i}}{1-v_{i}}=2^{f_{i}}d_{i}$ additional time for the robot to
catch up to the target, for $i\geq 0$. Recall the algorithm involves the robot
moving a distance $x_{i}$ (in time $x_{i}$, since the robot’s speed is $1$)
away from the origin and back in round $i$. Observe then that $d_{0}=d$,
$v_{0}=1/2$, and
$\displaystyle d_{i}=d+2v_{i}\sum_{j=0}^{i-1}x_{j}.$ (13)
Therefore, it follows from the definition of $x_{i}$ that
$\displaystyle x_{i}=2^{f_{i}}\left(d+2v_{i}\sum_{j=0}^{i-1}x_{j}\right).$
(14)
As a consequence
$\displaystyle\sum_{j=0}^{i-1}x_{j}=\frac{x_{i}-2^{f_{i}}d}{2^{f_{i}}\cdot
2\cdot v_{i}}$ (15)
Similarly, if we replace $i$ with $i+1$ we have that
$\displaystyle\sum_{j=0}^{i}x_{j}=\frac{x_{i+1}-2^{f_{i+1}}d}{2^{f_{i+1}}\cdot
2\cdot v_{i+1}}$ (16)
Subtracting Equation (15) from Equation (16), we derive the recurrence
$\displaystyle x_{i}=\frac{x_{i+1}-2^{f_{i+1}}d}{2^{f_{i+1}}\cdot 2\cdot
v_{i+1}}-\frac{x_{i}-2^{f_{i}}d}{2^{f_{i}}\cdot 2\cdot v_{i}}$ (17)
Collecting similar terms and simplifying Equation (17) yields
$\displaystyle\frac{x_{i+1}}{2^{f_{i+1}}\cdot 2\cdot v_{i+1}}$
$\displaystyle=\left(1+\frac{1}{2^{f_{i}}\cdot 2\cdot
v_{i}}\right)x_{i}+\left(\frac{2^{f_{i+1}}}{2^{f_{i+1}}\cdot 2\cdot
v_{i+1}}-\frac{2^{f_{i}}}{2^{f_{i}}\cdot 2\cdot v_{i}}\right)d$
$\displaystyle=x_{i}\left(1+\frac{1}{2^{f}_{i}\cdot 2\cdot
v_{i}}\right)+\frac{d}{2v_{i+1}}-\frac{d}{2v_{i}}$ (18)
$\displaystyle\leq\left(1+\frac{1}{2^{f_{i}}\cdot 2\cdot v_{i}}\right)x_{i}$
(19)
following from the fact the sum of the last two terms in Inequality 18 is less
than or equal to $0$.
If we simplify the right-hand side of Equation (19), we derive the following
recursive inequalities
$\displaystyle x_{i+1}$ $\displaystyle\leq 2^{f_{i+1}}\cdot 2\cdot
v_{i+1}\left(1+\frac{1}{2^{f_{i}}\cdot 2\cdot v_{i}}\right)x_{i}$
$\displaystyle\leq 2^{f_{i+1}}\cdot
2\cdot\left(1+\frac{1}{2^{f_{i}}}\right)x_{i}$ $\displaystyle\leq\left(2\cdot
2^{f_{i+1}}+2\cdot 2^{f_{i+1}-f_{i}}\right)x_{i}$ $\displaystyle\leq
2^{f_{i+1}}\cdot 4\cdot x_{i},$ (20)
which follows since $\frac{1}{2}\leq v_{i}<1$ and $f_{i}<f_{i+1}$ for all $i$.
By repeated application of the last Recurrence (20) above and using the fact
that by definition $x_{0}=2^{f_{0}}d$, it follows easily by induction that
$\displaystyle x_{i+1}$ $\displaystyle\leq 2^{f_{i+1}}\cdot 4\cdot x_{i}$
$\displaystyle\leq 2^{f_{i+1}+f_{i}}\cdot 4^{2}\cdot x_{i-1}$
$\displaystyle\leavevmode\nobreak\ \vdots$ $\displaystyle\leq
2^{f_{i+1}+f_{i}+f_{i-1}+\cdots+f_{1}}\cdot 4^{i+1}\cdot x_{0}$
$\displaystyle\leq 2^{\sum_{j=0}^{i+1}f_{j}}\cdot 4^{i+1}\cdot d$ (21)
Consider the first $i$ such that $v_{i}\geq v$. It follows that and
$v_{i-1}<v$ which yields $1-v<1-v_{i-1}=2^{-f_{i-1}}$ and implies that
$2^{f_{i-1}}<\frac{1}{1-v}$. Note that although $v_{i}\geq v$, the robot may
not find the target in round $i$ because it is located in the opposite
direction. It is guaranteed, however, to find the target by round $i+1$.
Moreover the total time that has elapsed from the start until round $i$ is
$2\sum_{j=0}^{i}x_{j}$ at which time the target is at distance
$d+v2\sum_{j=0}^{i}x_{j}$ from the origin.
As a consequence the competitive ratio of Algorithm 5 is at most
$\displaystyle\frac{2\sum_{j=0}^{i}x_{j}+\frac{d+2v\sum_{j=0}^{i}x_{j}}{1-v}}{\frac{d}{1-v}}$
$\displaystyle=1+\frac{2(v+1-v)}{d}\sum_{j=0}^{i}x_{j}$
$\displaystyle=1+\frac{x_{i+1}-2^{f_{i+1}}d}{d}\cdot\frac{2}{2^{f_{i+1}}\cdot
2\cdot v_{i+1}}$ (By (16)) $\displaystyle\leq
1+\frac{x_{i+1}}{d}\cdot\frac{2}{2^{f_{i+1}}\cdot 2\cdot v_{i+1}}$
$\displaystyle\leq 1+\frac{1}{v_{i+1}}2^{\sum_{j=0}^{i}f_{j}}\cdot 4^{i+1}$
(By (21))
Since $v_{i+1}\geq 1/2$ we conclude with an upper bound on the competitive
ratio of Algorithm 5 of
$\displaystyle 1+2^{1+\sum_{j=0}^{i}f_{j}}\cdot 4^{i+1}$ (22)
which proves Lemma 4.3.
We are now ready to prove the main theorem about the competitive ratio of
Algorithm 5.
###### Theorem 4.5.
For the NoSpeed/Away model, the competitive ratio of Algorithm 5 when applied
to the sequence $f_{j}=2^{j}$, for all $j\geq 0$, is at most
$\displaystyle\begin{cases}5&\text{if }v\leq\frac{1}{2}\\\
1+\frac{16\left(\log\frac{1}{1-v}\right)^{2}}{(1-v)^{4}}&\text{otherwise}\end{cases}$
where $\log$ is the base-2 logarithm.
###### Proof 4.6.
Consider the first index $i$ such that $v_{i}\geq v$. It follows that and
$v_{i-1}<v$, and so
$\displaystyle 1-2^{-2^{i-1}}<v\Rightarrow 2^{2^{i-1}}<\frac{1}{1-v}.$
Then, by Lemma 4.3, the competitive ratio of is at most
$\displaystyle 1+2^{1+\sum_{j=0}^{i}f_{j}}\cdot 4^{i+1}$
$\displaystyle=1+2^{1+\sum_{j=0}^{i}2^{j}}\cdot 4^{i+1}$
$\displaystyle=1+2^{2^{i+1}}\cdot 4^{i+1}\leq
1+\left(\frac{1}{1-v}\right)^{4}\cdot\left(4\cdot\log\left(\frac{1}{1-v}\right)\right)^{2}$
$\displaystyle\leq 1+\frac{16\left(\log\frac{1}{1-v}\right)^{2}}{(1-v)^{4}}$
This proves Theorem 4.5.
###### Remark 4.7.
By Theorem 2.1, $1+\frac{2}{1-v}$ is a lower bound on any algorithm when both
$v,d$ are known. As a consequence it must also be a lower bound when $d$ is
known but $v$ is not.
### 4.2 The NoSpeed/Toward Model
Now we consider the case where the target is moving toward the origin.
Algorithm 6 Online Algorithm for NoSpeed/Toward Model
1:input: target initial distance $d$
2:choose any direction and go for time $d$
3:if target not found then
4: change direction and go until target is found
###### Theorem 4.8.
For the NoSpeed/Toward model, Algorithm 6 achieves an optimal competitive
ratio at most $3$.
###### Proof 4.9.
The robot chooses a direction (without loss of generality, say to the right)
and goes for a time $d$ (this is where the robot makes use of its knowledge of
the distance $d$). If it does not find the target it changes direction. In the
meantime the target has moved for a distance $dv$ and now must be at location
$-d+dv$. Therefore at the time the robot changes direction the distance
between robot and target is equal to $d-(-d+dv)=2d-dv$, and hence the meeting
will take place in additional time $\frac{2d-dv}{1+v}$. It follows that the
total time required for the robot to meet the target must be equal to
$d+\frac{2d-dv}{1+v}$. The resulting competitive ratio satisfies
$\displaystyle CR\leq\frac{d+\frac{2d-dv}{1+v}}{\frac{d}{v+1}}=3.$
This proves the upper bound.
To prove the lower bound we argue as follows. If the searcher never visits
either of the points $\pm d$ then the competitive ratio is arbitrarily large
for very small values of $v$. Let $\epsilon>0$ be sufficiently small and let
the speed of the target be $v=\epsilon/3$. Consider the first time, say $t$,
that the robot reaches one of the points $\pm(d-\epsilon)$. Without loss of
generality let this point be $d-\epsilon$ and suppose the target is
adversarially placed at $-d$. Then at time $t$ it will be located at $-d+tv$.
Therefore the distance between the robot and the target at time $t$ will be
$d-\epsilon-(-d+tv)=2d-tv-\epsilon$. The time it takes for robot to find the
target, then, is at least $d-\epsilon+\frac{2d-tv-\epsilon}{1+v}$ and the
competitive ratio is at least
$\displaystyle\frac{d-\epsilon+\frac{2d-tv-\epsilon}{1+v}}{\frac{d}{1+v}}\geq
3-\frac{2\epsilon+(t+\epsilon)v}{d}$
It follows easily that if $t\geq 3d-\epsilon$ then
$CR\geq\frac{t}{d/(1+v)}\geq 3-3\epsilon$. However, if $t\leq 3d-\epsilon$
then $\frac{2\epsilon+(t+\epsilon)v}{d}\leq 2\epsilon+3v\leq 3\epsilon$, since
by assumption $v=\epsilon/3$. Therefore again $CR\geq 3-3\epsilon$. This
completes the proof of Theorem 4.8.
## 5 The NoKnowledge Model
In this section we assume that neither the initial distance $d$ nor the speed
$v$ of the target is known to the robot.
### 5.1 The NoKnowledge/Away Model
We now describe an approximation strategy resembling that described in Section
4.1. For this strategy though, the robot will need to guess both the target’s
speed and its initial distance.
Consider the situation where neither the distance $d$ to the target nor its
speed $v<1$ is known to the robot. Also consider two monotone increasing non-
negative integer sequences $\\{f_{i},g_{i}:i\geq 0\\}$ such that $f_{0}=1$ and
$g_{0}=0$ and $f_{i}<f_{i+1}$ and $g_{i}<g_{i+1}$, for all $i\geq 0$. The idea
of the algorithm is to search for the target by making a guess for its speed
and starting distance in rounds as follows. The robot, starting from the
origin, alternates searching to the right and left. On the $i$-th round, it
guesses that the target’s speed does not exceed $v_{i}=1-2^{-f_{i}}$ and that
it’s initial distance from the origin does not exceed $2^{g_{i}}$. Using these
guesses, the robot searches exactly the distance required (which we will later
denote $d_{i}$) to catch the target, given its guesses are correct and that
the target is in the direction the robot searches in round $i$. If robot does
not find the target after searching this distance, it returns to the origin
and begins the next round. Later in the analysis we will show how to select
the integer sequences $\\{f_{i},g_{i}:i\geq 0\\}$ so as to obtain bounds on
the competitive ratio. We formalize the algorithm described above as Algorithm
7.
Algorithm 7 Online Algorithm for NoKnowledge/Away Model
1:Inputs; Integer sequences $\\{f_{i},g_{i}:i\geq 0\\}$ such that
$f_{i}<f_{i+1}$ and $g_{i}<g_{i+1}$, for $i\geq 0$ and $f_{0}=1$ and
$g_{0}=0$;
2:$t\leftarrow 0$
3:for $i\leftarrow 0,1,2,\ldots$ until target found do
4: $d_{i}\leftarrow 2^{g_{i}}$
5: $v_{i}\leftarrow 1-2^{-f_{i}}$
6: $x_{i}\leftarrow(-1)^{i}\cdot\frac{d_{i}+tv_{i}}{1-v_{i}}$
7: move to $x_{i}$ and back to the origin
8: $t\leftarrow t+|x_{i}|$
Since there is always an integer $i\ \geq 1$ such that both
$v_{i}=1-2^{-f_{i}}\geq v$ and $2^{g_{i}}\geq d$, it is clear that the robot
will eventually succeed in catching the target. Next we analyze the
competitive ratio of the algorithm.
For the NoKnowledge/Away model, if Algorithm 7 terminates successfully in
round $i+1$ then its competitive ratio must satisfy
$CR\leq 1+\frac{2(i+2)}{d}\cdot 2^{g_{i+1}}\cdot 2^{\sum_{j=0}^{i}f_{j}}\cdot
4^{i+1}.$ (23)
###### Proof 5.1.
We call each iteration of the loop in Algorithm 7 a round. For any round $i$,
let $d_{i}$ be the distance from the origin to where the target would be if
its speed was equal to $v_{i}=1-2^{-f_{i}}$ and its starting position
$2^{g_{i}}$. Recall that during the first $i-1$ unsuccessful rounds, the taret
is moving further and further away from the origin. If the robot is at the
origin and the speed of the target is $v_{i}$ then it takes time at most
$x_{i}=\frac{d_{i}}{1-v_{i}}=2^{f_{i}}d_{i}$ for the robot to catch up to the
target, for $i\geq 0$. Observe from the algorithm that $d_{0}=1$ and
$v_{0}=1/2$ and
$\displaystyle d_{i}=2^{g_{i}}+2v_{i}\sum_{j=0}^{i-1}x_{j}.$ (24)
Therefore, it follows from the definition of $x_{i}$ that
$\displaystyle x_{i}$
$\displaystyle=2^{f_{i}}\left(2^{g_{i}}+2v_{i}\sum_{j=9}^{i-1}x_{j}\right).$
(25)
As a consequence
$\displaystyle\sum_{j=0}^{i-1}x_{j}$
$\displaystyle=\frac{x_{i}-2^{f_{i}+g_{i}}}{2^{f_{i}}\cdot 2\cdot v_{i}}$ (26)
Similarly, if we replace $i$ with $i+1$ we have that
$\displaystyle\sum_{j=0}^{i}x_{j}$
$\displaystyle=\frac{x_{i+1}-2^{f_{i+1}+g_{i+1}}}{2^{f_{i+1}}\cdot 2\cdot
v_{i+1}}.$ (27)
Subtracting Equation (26) from Equation (27) we derive the recurrence
$\displaystyle x_{i}$
$\displaystyle=\frac{x_{i+1}-2^{f_{i+1}+g_{i+1}}}{2^{f_{i+1}}\cdot 2\cdot
v_{i+1}}-\frac{x_{i}-2^{f_{i}+g_{i}}}{2^{f_{i}}\cdot 2\cdot v_{i}}$ (28)
Collecting similar terms and simplifying Equation (28) yields
$\displaystyle\frac{x_{i+1}}{2^{f_{i+1}}\cdot 2\cdot v_{i+1}}$
$\displaystyle=\left(1+\frac{1}{2^{f_{i}}\cdot 2\cdot
v_{i}}\right)x_{i}+\left(\frac{2^{f_{i+1}+g_{i+1}}}{2^{f_{i+1}}\cdot 2\cdot
v_{i+1}}-\frac{2^{f_{i}+g_{i}}}{2^{f_{i}}\cdot 2\cdot v_{i}}\right)$
$\displaystyle=\left(1+\frac{1}{2^{f_{i}}\cdot 2\cdot
v_{i}}\right)x_{i}+2^{g_{i+1}-1}\left(\frac{1}{v_{i+1}}-\frac{2^{g_{i}}}{2^{g_{i+1}}\cdot
v_{i}}\right)$ $\displaystyle\leq\left(1+\frac{1}{2^{f_{i}}\cdot 2\cdot
v_{i}}\right)x_{i}+2^{g_{i+1}-1}$ (29)
where Inequality (29) follows since
$\frac{1}{v_{i+1}}-\frac{2^{g_{i}}}{2^{g_{i+1}}\cdot v_{i}}\leq 1$.
If we multiply out with the denominator in the lefthand side of Inequality
(29) and simplify the righthand side we derive the following recursive
inequalities
$\displaystyle x_{i+1}$ $\displaystyle\leq 2^{f_{i+1}}\cdot 2\cdot
v_{i+1}\left(1+\frac{1}{2^{f_{i}}\cdot 2\cdot
v_{i}}\right)x_{i}+2^{f_{i+1}+g_{i+1}}v_{i+1}$
$\displaystyle\leq\left(2(2^{f_{i+1}}-1)+2^{f_{i+1}-f_{i}}\cdot\frac{v_{i+1}}{v_{i}}\right)x_{i}+2^{f_{i+1}+g_{i+1}}$
(30) $\displaystyle\leq 2^{f_{i+1}}\cdot 4\cdot x_{i}+2^{f_{i+1}+g_{i+1}},$
(31)
where in the derivation of Inequality (31) from the previous Inequality (30)
we used the fact that $\frac{v_{i+1}}{v_{i}}\leq 2$.
By repeated application of the last Recurrence (31) above and using the fact
that by definition $x_{0}=2^{f_{0}}d$, it follows easily by induction that
$\displaystyle x_{i+1}$ $\displaystyle\leq 2^{f_{i+1}}\cdot 4\cdot
x_{i}+2^{f_{i+1}+g_{i+1}}$ $\displaystyle\leq 2^{f_{i+1}+f_{i}}\cdot
4^{2}\cdot x_{i-1}+2^{f_{i+1}+f_{i}+g_{i}}\cdot 4^{1}+2^{f_{i+1}+g_{i+1}}$
$\displaystyle\leavevmode\nobreak\ \vdots$ $\displaystyle\leq
2^{g_{0}+\sum_{j=0}^{i+1}f_{j}}\cdot 4^{i+1}\cdot
x_{0}+2^{g_{1}+\sum_{j=1}^{i+1}f_{j}}\cdot
4^{i}+2^{g_{2}+\sum_{j=2}^{i+1}f_{j}}\cdot 4^{i-1}$
$\displaystyle\leavevmode\nobreak\ \leavevmode\nobreak\ \leavevmode\nobreak\
+\cdots+2^{f_{i+1}+g_{i+1}}$
$\displaystyle=\sum_{k=0}^{i+1}2^{g_{k}+\sum_{j=k}^{i+1}f_{j}}\cdot 4^{i-k+1}$
(32)
since $x_{0}=1$.
The total time that has elapsed from the start until the beginning of last
round $i$ (when the robot visits the origin for the last time before catching
the target) will be $\sum_{j=0}^{i}2x_{j}$ at which time the target is at
distance $d+v\sum_{j=0}^{i}2x_{j}$ from the origin. As a consequence the
competitive ratio of Algorithm 7 must satisfy the inequality
$\displaystyle CR$
$\displaystyle\leq\frac{2\sum_{j=0}^{i}x_{j}+\frac{d+2v\sum_{j=0}^{i}x_{j}}{1-v}}{\frac{d}{1-v}}.$
(33)
Simplifying the righthand side of Inequality (33) and using Identity (26)
yields
$\displaystyle CR$ $\displaystyle\leq 1+\frac{2}{d}\sum_{j=0}^{i}x_{j}$
$\displaystyle\leq 1+\frac{x_{i+1}}{d}\cdot\frac{1}{2^{f_{i+1}}\cdot
v_{i+1}}\mbox{\leavevmode\nobreak\ (Use Equation\leavevmode\nobreak\
\eqref{eq:no_knowledge__upper_guess_3})}$ $\displaystyle\leq
1+\frac{1}{v_{i+1}d2^{f_{i+1}}}\sum_{k=0}^{i+1}2^{g_{k}+\sum_{j=k}^{i+1}f_{j}}\cdot
4^{i-k+1}\mbox{\leavevmode\nobreak\ (Use Equation\leavevmode\nobreak\
\eqref{eq:no_knowledge_guess})}$ $\displaystyle\leq
1+\frac{1}{v_{i+1}d}\sum_{k=0}^{i+1}2^{g_{k}+\sum_{j=k}^{i}f_{j}}\cdot
4^{i-k+1}.$
Since $v_{i+1}\geq 1/2$ we conclude with
$\displaystyle CR$ $\displaystyle\leq
1+\frac{2}{d}\sum_{k=0}^{i+1}2^{g_{k}+\sum_{j=k}^{i}f_{j}}\cdot 4^{i-k+1}$
$\displaystyle\leq
1+\frac{2}{d}\sum_{k=0}^{i+1}2^{g_{k}+\sum_{j=k}^{i}f_{j}}\cdot 4^{i-k+1}$
$\displaystyle\leq 1+\frac{2(i+2)}{d}\cdot 2^{g_{i+1}}\cdot
2^{\sum_{j=0}^{i}f_{j}}\cdot 4^{i+1}$ (34)
This completes the proof of Lemma 5.1.
We now prove the following theorem.
###### Theorem 5.2.
For the NoKnowledge/Away model, Algorithm 7 with the sequences
$g_{i}=f_{i}=2^{i}$ has a competitive ratio of at most
$\displaystyle
1+\frac{16}{d}\left[\log\log\max\left(d,\frac{1}{1-v}\right)+3\right]\cdot\max\left(d,\frac{1}{1-v}\right)^{8}\cdot\log^{2}\max\left(d,\frac{1}{1-v}\right)$
where $\log$ is the base-2 logarithm.
###### Proof 5.3.
Observe that if the robot finds the target in round $i+1$, then by design, one
or both of the robot’s round $i-1$ guesses for the target’s speed
($1-2^{-2^{i-1}}$) or initial distance ($2^{2^{i-1}}$) must have been too low,
otherwise the robot would have found the target in an earlier round. In other
words, either $1-2^{-2^{i-1}}<v$ or $2^{2^{i-1}}<d$. It follows, then that
$i-1<\log\log\max\left(d,\frac{1}{1-v}\right)$. Then by Lemma 5.1, an upper
bound on the competitive ratio is given by
$\displaystyle CR$ $\displaystyle\leq 1+\frac{2(i+2)}{d}\cdot 2^{g_{i+1}}\cdot
2^{\sum_{j=0}^{i}f_{j}}\cdot 4^{i+1}$ $\displaystyle=1+\frac{2(i+2)}{d}\cdot
2^{2^{i+1}}\cdot 2^{2^{i+1}-1}\cdot 4^{i+1}$
$\displaystyle=1+\frac{(i-1)+3}{d}\cdot\left(2^{2^{i-1}}\right)^{8}\cdot
16\left(2^{i-1}\right)^{2}$
$\displaystyle=1+\frac{16}{d}\left[\log\log\max\left(d,\frac{1}{1-v}\right)+3\right]\cdot\max\left(d,\frac{1}{1-v}\right)^{8}\cdot\log^{2}\max\left(d,\frac{1}{1-v}\right)$
which proves Theorem 5.2.
###### Remark 5.4.
Observe that a lower bound of $1+8\frac{1+v}{(1-v)^{2}}$ follows directly from
the NoDistance/Away model.
### 5.2 The NoKnowledge/Toward Model
We can prove the following theorem.
The optimal competitive ratio is $1+\frac{1}{v}$ and is given by the waiting
Algorithm.
###### Proof 5.5.
The upper bound follows directly from Theorem 2.7. For the lower bound,
consider an algorithm where the robot does not wait forever and instead moves
a distance $d^{\prime}>0$ to the right (without loss of generality – a
symmetric argument for the case where the robot moves to the left follows
trivially) after waiting at the origin for time $t\geq 0$. Then consider the
scenario where the target with speed $v=\frac{d}{t+d^{\prime}}$ is initially
at $-d$ for any distance $d\geq 1$. Thus, the target reaches the origin at
exactly the time the robot reaches $d^{\prime}$ and so their earliest possible
meeting time is
$\displaystyle
t+d^{\prime}+\frac{d^{\prime}}{1+v}=\frac{d}{v}+\frac{d^{\prime}}{1+v}\geq\frac{d}{v}$
Thus, the competitive ratio is at least
$\displaystyle\frac{d/v}{d/(1+v)}=1+\frac{1}{v}$
This proves Theorem 5.2.
## 6 Conclusion
We considered linear search for an autonomous robot searching for an oblivious
moving target on an infinite line. Two scenarios were analyzed depending on
whether the target is moving towards or away from the origin (and this is
known to the robot). In either of these two scenarios we considered the
knowledge the robot has about the speed and starting distance of the target.
For each scenario we gave search algorithms and analyzed their competitive
ratio for the four possible cases arising. Our bounds are tight in all cases
when the target is moving towards the origin. They are also shown to be tight
when the target is moving away from the origin and its speed $0<v<1$ is known
to the robot; for this scenario we also obtain upper bounds when $v$ is not
known to the robot. It remains an open problem to prove tight bounds for the
case when $v$ is unknown to the robot and the target is moving away from the
origin. It also remains open to find tight bounds for the case where the
direction of movement of the target is unknown.
## References
* [1] R. Ahlswede and I. Wegener. Search problems. Wiley-Interscience, 1987.
* [2] Steve Alpern and Shmuel Gal. The theory of search games and rendezvous, volume 55 of International series in operations research and management science. Kluwer, 2003.
* [3] Ricardo A. Baeza-Yates, Joseph C. Culberson, and Gregory J. E. Rawlins. Searching in the plane. Inf. Comput., 106(2):234–252, 1993. doi:10.1006/inco.1993.1054.
* [4] A. Beck. On the linear search problem. Israel Journal of Mathematics, 2(4):221–228, 1964.
* [5] R. Bellman. An optimal search. Siam Review, 5(3):274–274, 1963.
* [6] A. Bonato and R. Nowakowski. The game of cops and robbers on graphs. American Mathematical Soc., 2011.
* [7] Jurek Czyzowicz, Konstantinos Georgiou, Evangelos Kranakis, Danny Krizanc, Lata Narayanan, Jaroslav Opatrny, and Sunil M. Shende. Search on a line by byzantine robots. Int. J. Found. Comput. Sci., 32(4):369–387, 2021. doi:10.1142/S0129054121500209.
* [8] Jurek Czyzowicz, Ryan Killick, and Evangelos Kranakis. Linear rendezvous with asymmetric clocks. In Jiannong Cao, Faith Ellen, Luis Rodrigues, and Bernardo Ferreira, editors, 22nd International Conference on Principles of Distributed Systems, OPODIS 2018, December 17-19, 2018, Hong Kong, China, volume 125 of LIPIcs, pages 25:1–25:16. Schloss Dagstuhl - Leibniz-Zentrum für Informatik, 2018. doi:10.4230/LIPIcs.OPODIS.2018.25.
* [9] Jurek Czyzowicz, Evangelos Kranakis, Danny Krizanc, Lata Narayanan, and Jaroslav Opatrny. Search on a line with faulty robots. Distributed Comput., 32(6):493–504, 2019. doi:10.1007/s00446-017-0296-0.
* [10] Ryan Killick. Lower bound for linear search. Unpublished Manuscript, 2022.
* [11] Gianluca De Marco, Luisa Gargano, Evangelos Kranakis, Danny Krizanc, Andrzej Pelc, and Ugo Vaccaro. Asynchronous deterministic rendezvous in graphs. Theor. Comput. Sci., 355(3):315–326, 2006. doi:10.1016/j.tcs.2005.12.016.
* [12] B. J. McCabe. Searching for a one-dimensional random walker. J. Applied Probability, pages 86–93, 1974.
* [13] Sven Schuierer. Lower bounds in on-line geometric searching. Comput. Geom., 18(1):37–53, 2001. doi:10.1016/S0925-7721(00)00030-4.
* [14] L. Stone. Theory of optimal search. Academic Press New York, 1975.
|
goodness | $\chi^{2}_{\mathrm{cp}}$ | goodness of fit for closure phases | 1.8
of fit | $\chi^{2}_{\mathrm{lca}}$ | goodness of fit for log closure amplitudes | 2.3
| $\mathcal{A}^{\mathrm{(sc)}}_{\mathrm{intra}}$ | mean self-calibration gain for co-located stations | 0.98
| $\mathcal{A}^{\mathrm{(sc)}}_{\mathrm{\tiny{\acs{lm}}}}$ | LMT mean self-calibration gain | 1.13
| $\mathcal{A}^{\mathrm{(sc)}}_{\mathrm{\tiny{\acs{az}}}}$ | SMT mean self-calibration gain | 1.01
| $\mathcal{A}^{\mathrm{(sc)}}_{\mathrm{\tiny{\acs{sp}}}}$ | SPT mean self-calibration gain | 1.03
EHT-HOPS image | $\chi^{2}_{\mathrm{amp}}$ | goodness of fit for amplitudes | 1.2
goodness | $\chi^{2}_{\mathrm{cp}}$ | goodness of fit for closure phases | 2.1
of fit | $\chi^{2}_{\mathrm{lca}}$ | goodness of fit for log closure amplitudes | 1.2
| $\mathcal{A}^{\mathrm{(sc)}}_{\mathrm{intra}}$ | mean self-calibration gain for co-located stations | 0.98
| $\mathcal{A}^{\mathrm{(sc)}}_{\mathrm{\tiny{\acs{lm}}}}$ | LMT mean self-calibration gain | 1.15
| $\mathcal{A}^{\mathrm{(sc)}}_{\mathrm{\tiny{\acs{az}}}}$ | SMT mean self-calibration gain | 1.01
| $\mathcal{A}^{\mathrm{(sc)}}_{\mathrm{\tiny{\acs{sp}}}}$ | SPT mean self-calibration gain | 1.00
Note – $\mathcal{A}^{\mathrm{(sc)}}_{\mathrm{intra}}$ corresponds to the mean
gain of ALMA, APEX, JCMT, and SMA.
Figure 7: Cen A data properties from April 2017. The top left panel shows the
$(u,v)$ coverage. A priori calibrated amplitude (before self-calibration) and
closure phase data points are shown in the top right and bottom panel,
respectively, overplotted with lines from the final image model as a function
of $(u,v)$ distances. The error bars indicate thermal noise and $5\,\%$ non-
closing error uncertainties added in quadrature, which are smaller than the
plotted symbols in some cases. The color-coding shows different baselines.
Amplitudes projected along and perpendicular to the jet position angle are
given in Supplementary Fig. 8.
Figure 8: Source structure along specific position angles on the sky. A priori
calibrated amplitudes are shown projected along the jet position angle (PA) on
the sky in the left panel and perpendicular to the PA in the right panel. The
color coding and error bars shown are the same as in Supplementary Fig. 7.
Figure 9: Determination of the jet apex location. A zoomed-in version of the
final image model is shown. The solid blue lines show simple linear
extrapolations of the inner NW and SE jet arms, which would place the jet apex
well within the counterjet region. The dashed white lines mark the certain
edges of the approaching jet and the counterjet. The quadrangle enclosed by
the solid and dashed lines is the region where the jet apex is located. Inside
this quadrangle, a tentative convergence of the two streamlines can be seen.
The apex position assumed in this work is indicated with a white cross. The
surrounding blue dashed circle corresponds to the
$z_{\mathrm{col}}=32\,\acs{muas}$ distance. Vertical black bars mark the
brightest regions along each jet arm, which correspond to the assumed location
of the radio core.
*[VLBI]: very long baseline interferometry
*[SMBH]: supermassive black hole
*[Cen A]: Centaurus A
*[EHT]: Event Horizon Telescope
*[$\mu$as]: microarcseconds
*[GRMHD]: general relativistic magnetohydrodynamics
*[Jy]: jansky
*[PA]: position angle
*[LLAGN]: low-luminosity AGN
*[AGN]: active galactic nuclei
*[pc]: parsec
*[ALMA]: Atacama Large Millimeter/submillimeter Array
*[APEX]: Atacama Pathfinder Experiment
*[JCMT]: James Clerk Maxwell Telescope
*[LMT]: Large Millimeter Telescope Alfonso Serrano
*[SPT]: South Pole Telescope
*[SMA]: Submillimeter Array
*[SMT]: Submillimeter Telescope
*[S/N]: signal-to-noise ratio
|
Subsets and Splits
Filtered Text Samples
Retrieves 100 samples of text containing the specific phrase "You are a helpful assistant", providing limited insight into the dataset.
Helpful Assistant Text Samples
Returns a limited set of rows containing the phrase 'helpful assistant' in the text, providing basic filtering of relevant entries.